@clinebot/llms 0.0.11 → 0.0.13
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/index.browser.js +14 -14
- package/dist/index.js +13 -13
- package/dist/providers/handlers/base.d.ts +4 -0
- package/package.json +2 -2
- package/src/models/generated.ts +223 -641
- package/src/models/models-dev-catalog.test.ts +2 -2
- package/src/models/models-dev-catalog.ts +9 -2
- package/src/providers/handlers/ai-sdk-community.ts +8 -2
- package/src/providers/handlers/base.test.ts +119 -0
- package/src/providers/handlers/base.ts +93 -0
- package/src/providers/handlers/bedrock-base.ts +4 -4
- package/src/providers/handlers/community-sdk.test.ts +1 -1
- package/src/providers/handlers/openai-base.ts +14 -6
- package/src/providers/handlers/openai-responses.test.ts +2 -2
- package/src/providers/handlers/openai-responses.ts +10 -2
- package/src/providers/handlers/r1-base.ts +5 -2
- package/src/providers/transform/gemini-format.ts +77 -1
package/src/models/generated.ts
CHANGED
|
@@ -14,7 +14,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
14
14
|
version: number;
|
|
15
15
|
providers: Record<string, Record<string, ModelInfo>>;
|
|
16
16
|
} = {
|
|
17
|
-
version:
|
|
17
|
+
version: 1774365323810,
|
|
18
18
|
providers: {
|
|
19
19
|
aihubmix: {
|
|
20
20
|
"claude-sonnet-4-6": {
|
|
@@ -147,7 +147,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
147
147
|
id: "kimi-k2.5",
|
|
148
148
|
name: "Kimi K2.5",
|
|
149
149
|
contextWindow: 262144,
|
|
150
|
-
maxTokens:
|
|
150
|
+
maxTokens: 13107,
|
|
151
151
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
152
152
|
pricing: {
|
|
153
153
|
input: 0.6,
|
|
@@ -545,7 +545,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
545
545
|
id: "Kimi-K2-0905",
|
|
546
546
|
name: "Kimi K2 0905",
|
|
547
547
|
contextWindow: 262144,
|
|
548
|
-
maxTokens:
|
|
548
|
+
maxTokens: 13107,
|
|
549
549
|
capabilities: ["tools", "temperature"],
|
|
550
550
|
pricing: {
|
|
551
551
|
input: 0.55,
|
|
@@ -587,7 +587,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
587
587
|
id: "qwen3-235b-a22b-instruct-2507",
|
|
588
588
|
name: "Qwen3 235B A22B Instruct 2507",
|
|
589
589
|
contextWindow: 262144,
|
|
590
|
-
maxTokens:
|
|
590
|
+
maxTokens: 13107,
|
|
591
591
|
capabilities: ["tools", "temperature"],
|
|
592
592
|
pricing: {
|
|
593
593
|
input: 0.28,
|
|
@@ -601,7 +601,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
601
601
|
id: "qwen3-235b-a22b-thinking-2507",
|
|
602
602
|
name: "Qwen3 235B A22B Thinking 2507",
|
|
603
603
|
contextWindow: 262144,
|
|
604
|
-
maxTokens:
|
|
604
|
+
maxTokens: 13107,
|
|
605
605
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
606
606
|
pricing: {
|
|
607
607
|
input: 0.28,
|
|
@@ -993,8 +993,8 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
993
993
|
},
|
|
994
994
|
},
|
|
995
995
|
baseten: {
|
|
996
|
-
"nvidia/Nemotron-
|
|
997
|
-
id: "nvidia/Nemotron-
|
|
996
|
+
"nvidia/Nemotron-120B-A12B": {
|
|
997
|
+
id: "nvidia/Nemotron-120B-A12B",
|
|
998
998
|
name: "Nemotron 3 Super",
|
|
999
999
|
contextWindow: 262144,
|
|
1000
1000
|
maxTokens: 32678,
|
|
@@ -1011,7 +1011,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1011
1011
|
id: "MiniMaxAI/MiniMax-M2.5",
|
|
1012
1012
|
name: "MiniMax-M2.5",
|
|
1013
1013
|
contextWindow: 204000,
|
|
1014
|
-
maxTokens:
|
|
1014
|
+
maxTokens: 10200,
|
|
1015
1015
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
1016
1016
|
pricing: {
|
|
1017
1017
|
input: 0.3,
|
|
@@ -1067,7 +1067,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1067
1067
|
id: "zai-org/GLM-4.6",
|
|
1068
1068
|
name: "GLM 4.6",
|
|
1069
1069
|
contextWindow: 200000,
|
|
1070
|
-
maxTokens:
|
|
1070
|
+
maxTokens: 10000,
|
|
1071
1071
|
capabilities: ["tools", "temperature"],
|
|
1072
1072
|
pricing: {
|
|
1073
1073
|
input: 0.6,
|
|
@@ -1095,7 +1095,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1095
1095
|
id: "openai/gpt-oss-120b",
|
|
1096
1096
|
name: "GPT OSS 120B",
|
|
1097
1097
|
contextWindow: 128000,
|
|
1098
|
-
maxTokens:
|
|
1098
|
+
maxTokens: 6400,
|
|
1099
1099
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
1100
1100
|
pricing: {
|
|
1101
1101
|
input: 0.1,
|
|
@@ -1191,6 +1191,34 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1191
1191
|
},
|
|
1192
1192
|
releaseDate: "2026-02-17",
|
|
1193
1193
|
},
|
|
1194
|
+
"minimax.minimax-m2.5": {
|
|
1195
|
+
id: "minimax.minimax-m2.5",
|
|
1196
|
+
name: "MiniMax M2.5",
|
|
1197
|
+
contextWindow: 1000000,
|
|
1198
|
+
maxTokens: 131072,
|
|
1199
|
+
capabilities: ["tools", "reasoning", "temperature"],
|
|
1200
|
+
pricing: {
|
|
1201
|
+
input: 0.3,
|
|
1202
|
+
output: 1.2,
|
|
1203
|
+
cacheRead: 0,
|
|
1204
|
+
cacheWrite: 0,
|
|
1205
|
+
},
|
|
1206
|
+
releaseDate: "2026-02-12",
|
|
1207
|
+
},
|
|
1208
|
+
"zai.glm-5": {
|
|
1209
|
+
id: "zai.glm-5",
|
|
1210
|
+
name: "GLM-5",
|
|
1211
|
+
contextWindow: 200000,
|
|
1212
|
+
maxTokens: 131072,
|
|
1213
|
+
capabilities: ["tools", "reasoning", "temperature"],
|
|
1214
|
+
pricing: {
|
|
1215
|
+
input: 1,
|
|
1216
|
+
output: 3.2,
|
|
1217
|
+
cacheRead: 0,
|
|
1218
|
+
cacheWrite: 0,
|
|
1219
|
+
},
|
|
1220
|
+
releaseDate: "2026-02-11",
|
|
1221
|
+
},
|
|
1194
1222
|
"deepseek.v3.2": {
|
|
1195
1223
|
id: "deepseek.v3.2",
|
|
1196
1224
|
name: "DeepSeek-V3.2",
|
|
@@ -1209,7 +1237,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1209
1237
|
id: "moonshotai.kimi-k2.5",
|
|
1210
1238
|
name: "Kimi K2.5",
|
|
1211
1239
|
contextWindow: 256000,
|
|
1212
|
-
maxTokens:
|
|
1240
|
+
maxTokens: 12800,
|
|
1213
1241
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
1214
1242
|
pricing: {
|
|
1215
1243
|
input: 0.6,
|
|
@@ -1377,7 +1405,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1377
1405
|
id: "moonshot.kimi-k2-thinking",
|
|
1378
1406
|
name: "Kimi K2 Thinking",
|
|
1379
1407
|
contextWindow: 256000,
|
|
1380
|
-
maxTokens:
|
|
1408
|
+
maxTokens: 12800,
|
|
1381
1409
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
1382
1410
|
pricing: {
|
|
1383
1411
|
input: 0.6,
|
|
@@ -1517,7 +1545,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1517
1545
|
id: "qwen.qwen3-vl-235b-a22b",
|
|
1518
1546
|
name: "Qwen/Qwen3-VL-235B-A22B-Instruct",
|
|
1519
1547
|
contextWindow: 262000,
|
|
1520
|
-
maxTokens:
|
|
1548
|
+
maxTokens: 13100,
|
|
1521
1549
|
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
1522
1550
|
pricing: {
|
|
1523
1551
|
input: 0.3,
|
|
@@ -1615,7 +1643,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1615
1643
|
id: "qwen.qwen3-32b-v1:0",
|
|
1616
1644
|
name: "Qwen3 32B (dense)",
|
|
1617
1645
|
contextWindow: 16384,
|
|
1618
|
-
maxTokens:
|
|
1646
|
+
maxTokens: 819,
|
|
1619
1647
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
1620
1648
|
pricing: {
|
|
1621
1649
|
input: 0.15,
|
|
@@ -1657,7 +1685,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
1657
1685
|
id: "qwen.qwen3-next-80b-a3b",
|
|
1658
1686
|
name: "Qwen/Qwen3-Next-80B-A3B-Instruct",
|
|
1659
1687
|
contextWindow: 262000,
|
|
1660
|
-
maxTokens:
|
|
1688
|
+
maxTokens: 13100,
|
|
1661
1689
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
1662
1690
|
pricing: {
|
|
1663
1691
|
input: 0.14,
|
|
@@ -2377,7 +2405,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2377
2405
|
id: "accounts/fireworks/models/minimax-m2p5",
|
|
2378
2406
|
name: "MiniMax-M2.5",
|
|
2379
2407
|
contextWindow: 196608,
|
|
2380
|
-
maxTokens:
|
|
2408
|
+
maxTokens: 9830,
|
|
2381
2409
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2382
2410
|
pricing: {
|
|
2383
2411
|
input: 0.3,
|
|
@@ -2405,7 +2433,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2405
2433
|
id: "accounts/fireworks/models/kimi-k2p5",
|
|
2406
2434
|
name: "Kimi K2.5",
|
|
2407
2435
|
contextWindow: 256000,
|
|
2408
|
-
maxTokens:
|
|
2436
|
+
maxTokens: 12800,
|
|
2409
2437
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
2410
2438
|
pricing: {
|
|
2411
2439
|
input: 0.6,
|
|
@@ -2415,11 +2443,25 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2415
2443
|
},
|
|
2416
2444
|
releaseDate: "2026-01-27",
|
|
2417
2445
|
},
|
|
2446
|
+
"accounts/fireworks/routers/kimi-k2p5-turbo": {
|
|
2447
|
+
id: "accounts/fireworks/routers/kimi-k2p5-turbo",
|
|
2448
|
+
name: "Kimi K2.5 Turbo",
|
|
2449
|
+
contextWindow: 256000,
|
|
2450
|
+
maxTokens: 12800,
|
|
2451
|
+
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
2452
|
+
pricing: {
|
|
2453
|
+
input: 0,
|
|
2454
|
+
output: 0,
|
|
2455
|
+
cacheRead: 0,
|
|
2456
|
+
cacheWrite: 0,
|
|
2457
|
+
},
|
|
2458
|
+
releaseDate: "2026-01-27",
|
|
2459
|
+
},
|
|
2418
2460
|
"accounts/fireworks/models/minimax-m2p1": {
|
|
2419
2461
|
id: "accounts/fireworks/models/minimax-m2p1",
|
|
2420
2462
|
name: "MiniMax-M2.1",
|
|
2421
2463
|
contextWindow: 200000,
|
|
2422
|
-
maxTokens:
|
|
2464
|
+
maxTokens: 10000,
|
|
2423
2465
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2424
2466
|
pricing: {
|
|
2425
2467
|
input: 0.3,
|
|
@@ -2433,7 +2475,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2433
2475
|
id: "accounts/fireworks/models/glm-4p7",
|
|
2434
2476
|
name: "GLM 4.7",
|
|
2435
2477
|
contextWindow: 198000,
|
|
2436
|
-
maxTokens:
|
|
2478
|
+
maxTokens: 9900,
|
|
2437
2479
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2438
2480
|
pricing: {
|
|
2439
2481
|
input: 0.6,
|
|
@@ -2447,7 +2489,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2447
2489
|
id: "accounts/fireworks/models/deepseek-v3p2",
|
|
2448
2490
|
name: "DeepSeek V3.2",
|
|
2449
2491
|
contextWindow: 160000,
|
|
2450
|
-
maxTokens:
|
|
2492
|
+
maxTokens: 8000,
|
|
2451
2493
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2452
2494
|
pricing: {
|
|
2453
2495
|
input: 0.56,
|
|
@@ -2461,7 +2503,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2461
2503
|
id: "accounts/fireworks/models/kimi-k2-thinking",
|
|
2462
2504
|
name: "Kimi K2 Thinking",
|
|
2463
2505
|
contextWindow: 256000,
|
|
2464
|
-
maxTokens:
|
|
2506
|
+
maxTokens: 12800,
|
|
2465
2507
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2466
2508
|
pricing: {
|
|
2467
2509
|
input: 0.6,
|
|
@@ -2475,7 +2517,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2475
2517
|
id: "accounts/fireworks/models/deepseek-v3p1",
|
|
2476
2518
|
name: "DeepSeek V3.1",
|
|
2477
2519
|
contextWindow: 163840,
|
|
2478
|
-
maxTokens:
|
|
2520
|
+
maxTokens: 8192,
|
|
2479
2521
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2480
2522
|
pricing: {
|
|
2481
2523
|
input: 0.56,
|
|
@@ -2517,7 +2559,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2517
2559
|
id: "accounts/fireworks/models/glm-4p5-air",
|
|
2518
2560
|
name: "GLM 4.5 Air",
|
|
2519
2561
|
contextWindow: 131072,
|
|
2520
|
-
maxTokens:
|
|
2562
|
+
maxTokens: 6553,
|
|
2521
2563
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2522
2564
|
pricing: {
|
|
2523
2565
|
input: 0.22,
|
|
@@ -2531,7 +2573,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
2531
2573
|
id: "accounts/fireworks/models/glm-4p5",
|
|
2532
2574
|
name: "GLM 4.5",
|
|
2533
2575
|
contextWindow: 131072,
|
|
2534
|
-
maxTokens:
|
|
2576
|
+
maxTokens: 6553,
|
|
2535
2577
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
2536
2578
|
pricing: {
|
|
2537
2579
|
input: 0.55,
|
|
@@ -3124,7 +3166,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3124
3166
|
id: "llama-3.1-8b-instant",
|
|
3125
3167
|
name: "Llama 3.1 8B Instant",
|
|
3126
3168
|
contextWindow: 131072,
|
|
3127
|
-
maxTokens:
|
|
3169
|
+
maxTokens: 6553,
|
|
3128
3170
|
capabilities: ["tools", "temperature"],
|
|
3129
3171
|
pricing: {
|
|
3130
3172
|
input: 0.05,
|
|
@@ -3196,7 +3238,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3196
3238
|
id: "moonshotai/Kimi-K2.5",
|
|
3197
3239
|
name: "Kimi-K2.5",
|
|
3198
3240
|
contextWindow: 262144,
|
|
3199
|
-
maxTokens:
|
|
3241
|
+
maxTokens: 13107,
|
|
3200
3242
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
3201
3243
|
pricing: {
|
|
3202
3244
|
input: 0.6,
|
|
@@ -3266,7 +3308,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3266
3308
|
id: "moonshotai/Kimi-K2-Thinking",
|
|
3267
3309
|
name: "Kimi-K2-Thinking",
|
|
3268
3310
|
contextWindow: 262144,
|
|
3269
|
-
maxTokens:
|
|
3311
|
+
maxTokens: 13107,
|
|
3270
3312
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
3271
3313
|
pricing: {
|
|
3272
3314
|
input: 0.6,
|
|
@@ -3378,7 +3420,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3378
3420
|
id: "deepseek-ai/DeepSeek-R1-0528",
|
|
3379
3421
|
name: "DeepSeek-R1-0528",
|
|
3380
3422
|
contextWindow: 163840,
|
|
3381
|
-
maxTokens:
|
|
3423
|
+
maxTokens: 8192,
|
|
3382
3424
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
3383
3425
|
pricing: {
|
|
3384
3426
|
input: 3,
|
|
@@ -3524,7 +3566,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3524
3566
|
id: "kimi-k2.5",
|
|
3525
3567
|
name: "Kimi K2.5",
|
|
3526
3568
|
contextWindow: 262144,
|
|
3527
|
-
maxTokens:
|
|
3569
|
+
maxTokens: 13107,
|
|
3528
3570
|
capabilities: ["images", "tools", "reasoning", "structured_output"],
|
|
3529
3571
|
pricing: {
|
|
3530
3572
|
input: 0.6,
|
|
@@ -3538,7 +3580,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3538
3580
|
id: "kimi-k2-thinking",
|
|
3539
3581
|
name: "Kimi K2 Thinking",
|
|
3540
3582
|
contextWindow: 262144,
|
|
3541
|
-
maxTokens:
|
|
3583
|
+
maxTokens: 13107,
|
|
3542
3584
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
3543
3585
|
pricing: {
|
|
3544
3586
|
input: 0.6,
|
|
@@ -3552,7 +3594,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3552
3594
|
id: "kimi-k2-thinking-turbo",
|
|
3553
3595
|
name: "Kimi K2 Thinking Turbo",
|
|
3554
3596
|
contextWindow: 262144,
|
|
3555
|
-
maxTokens:
|
|
3597
|
+
maxTokens: 13107,
|
|
3556
3598
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
3557
3599
|
pricing: {
|
|
3558
3600
|
input: 1.15,
|
|
@@ -3566,7 +3608,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3566
3608
|
id: "kimi-k2-0905-preview",
|
|
3567
3609
|
name: "Kimi K2 0905",
|
|
3568
3610
|
contextWindow: 262144,
|
|
3569
|
-
maxTokens:
|
|
3611
|
+
maxTokens: 13107,
|
|
3570
3612
|
capabilities: ["tools", "temperature"],
|
|
3571
3613
|
pricing: {
|
|
3572
3614
|
input: 0.6,
|
|
@@ -3580,7 +3622,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
3580
3622
|
id: "kimi-k2-turbo-preview",
|
|
3581
3623
|
name: "Kimi K2 Turbo",
|
|
3582
3624
|
contextWindow: 262144,
|
|
3583
|
-
maxTokens:
|
|
3625
|
+
maxTokens: 13107,
|
|
3584
3626
|
capabilities: ["tools", "temperature"],
|
|
3585
3627
|
pricing: {
|
|
3586
3628
|
input: 2.4,
|
|
@@ -4323,7 +4365,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4323
4365
|
id: "kimi-k2.5",
|
|
4324
4366
|
name: "kimi-k2.5",
|
|
4325
4367
|
contextWindow: 262144,
|
|
4326
|
-
maxTokens:
|
|
4368
|
+
maxTokens: 13107,
|
|
4327
4369
|
capabilities: ["images", "tools", "reasoning"],
|
|
4328
4370
|
pricing: {
|
|
4329
4371
|
input: 0,
|
|
@@ -4393,7 +4435,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4393
4435
|
id: "devstral-2:123b",
|
|
4394
4436
|
name: "devstral-2:123b",
|
|
4395
4437
|
contextWindow: 262144,
|
|
4396
|
-
maxTokens:
|
|
4438
|
+
maxTokens: 13107,
|
|
4397
4439
|
capabilities: ["tools"],
|
|
4398
4440
|
pricing: {
|
|
4399
4441
|
input: 0,
|
|
@@ -4407,7 +4449,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4407
4449
|
id: "devstral-small-2:24b",
|
|
4408
4450
|
name: "devstral-small-2:24b",
|
|
4409
4451
|
contextWindow: 262144,
|
|
4410
|
-
maxTokens:
|
|
4452
|
+
maxTokens: 13107,
|
|
4411
4453
|
capabilities: ["images", "tools"],
|
|
4412
4454
|
pricing: {
|
|
4413
4455
|
input: 0,
|
|
@@ -4435,7 +4477,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4435
4477
|
id: "mistral-large-3:675b",
|
|
4436
4478
|
name: "mistral-large-3:675b",
|
|
4437
4479
|
contextWindow: 262144,
|
|
4438
|
-
maxTokens:
|
|
4480
|
+
maxTokens: 13107,
|
|
4439
4481
|
capabilities: ["images", "tools"],
|
|
4440
4482
|
pricing: {
|
|
4441
4483
|
input: 0,
|
|
@@ -4463,7 +4505,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4463
4505
|
id: "kimi-k2-thinking",
|
|
4464
4506
|
name: "kimi-k2-thinking",
|
|
4465
4507
|
contextWindow: 262144,
|
|
4466
|
-
maxTokens:
|
|
4508
|
+
maxTokens: 13107,
|
|
4467
4509
|
capabilities: ["tools", "reasoning"],
|
|
4468
4510
|
pricing: {
|
|
4469
4511
|
input: 0,
|
|
@@ -4547,7 +4589,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4547
4589
|
id: "deepseek-v3.1:671b",
|
|
4548
4590
|
name: "deepseek-v3.1:671b",
|
|
4549
4591
|
contextWindow: 163840,
|
|
4550
|
-
maxTokens:
|
|
4592
|
+
maxTokens: 8192,
|
|
4551
4593
|
capabilities: ["tools", "reasoning"],
|
|
4552
4594
|
pricing: {
|
|
4553
4595
|
input: 0,
|
|
@@ -4603,7 +4645,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
4603
4645
|
id: "kimi-k2:1t",
|
|
4604
4646
|
name: "kimi-k2:1t",
|
|
4605
4647
|
contextWindow: 262144,
|
|
4606
|
-
maxTokens:
|
|
4648
|
+
maxTokens: 13107,
|
|
4607
4649
|
capabilities: ["tools"],
|
|
4608
4650
|
pricing: {
|
|
4609
4651
|
input: 0,
|
|
@@ -5217,7 +5259,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5217
5259
|
id: "gpt-4",
|
|
5218
5260
|
name: "GPT-4",
|
|
5219
5261
|
contextWindow: 8192,
|
|
5220
|
-
maxTokens:
|
|
5262
|
+
maxTokens: 409,
|
|
5221
5263
|
capabilities: ["tools", "temperature"],
|
|
5222
5264
|
pricing: {
|
|
5223
5265
|
input: 30,
|
|
@@ -5351,40 +5393,26 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5351
5393
|
},
|
|
5352
5394
|
releaseDate: "2026-03-12",
|
|
5353
5395
|
},
|
|
5354
|
-
"
|
|
5355
|
-
id: "
|
|
5356
|
-
name: "
|
|
5396
|
+
"nvidia/nemotron-3-super-120b-a12b": {
|
|
5397
|
+
id: "nvidia/nemotron-3-super-120b-a12b",
|
|
5398
|
+
name: "Nemotron 3 Super",
|
|
5357
5399
|
contextWindow: 262144,
|
|
5358
|
-
maxTokens:
|
|
5359
|
-
capabilities: [
|
|
5360
|
-
"images",
|
|
5361
|
-
"files",
|
|
5362
|
-
"tools",
|
|
5363
|
-
"reasoning",
|
|
5364
|
-
"structured_output",
|
|
5365
|
-
"temperature",
|
|
5366
|
-
],
|
|
5400
|
+
maxTokens: 13107,
|
|
5401
|
+
capabilities: ["tools", "reasoning", "temperature"],
|
|
5367
5402
|
pricing: {
|
|
5368
|
-
input: 0,
|
|
5369
|
-
output: 0,
|
|
5403
|
+
input: 0.1,
|
|
5404
|
+
output: 0.5,
|
|
5370
5405
|
cacheRead: 0,
|
|
5371
5406
|
cacheWrite: 0,
|
|
5372
5407
|
},
|
|
5373
5408
|
releaseDate: "2026-03-11",
|
|
5374
5409
|
},
|
|
5375
|
-
"
|
|
5376
|
-
id: "
|
|
5377
|
-
name: "
|
|
5378
|
-
contextWindow:
|
|
5379
|
-
maxTokens:
|
|
5380
|
-
capabilities: [
|
|
5381
|
-
"images",
|
|
5382
|
-
"files",
|
|
5383
|
-
"tools",
|
|
5384
|
-
"reasoning",
|
|
5385
|
-
"structured_output",
|
|
5386
|
-
"temperature",
|
|
5387
|
-
],
|
|
5410
|
+
"nvidia/nemotron-3-super-120b-a12b-free": {
|
|
5411
|
+
id: "nvidia/nemotron-3-super-120b-a12b-free",
|
|
5412
|
+
name: "Nemotron 3 Super (free)",
|
|
5413
|
+
contextWindow: 262144,
|
|
5414
|
+
maxTokens: 13107,
|
|
5415
|
+
capabilities: ["tools", "reasoning", "temperature"],
|
|
5388
5416
|
pricing: {
|
|
5389
5417
|
input: 0,
|
|
5390
5418
|
output: 0,
|
|
@@ -5627,25 +5655,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5627
5655
|
},
|
|
5628
5656
|
releaseDate: "2026-02-12",
|
|
5629
5657
|
},
|
|
5630
|
-
"openrouter/aurora-alpha": {
|
|
5631
|
-
id: "openrouter/aurora-alpha",
|
|
5632
|
-
name: "Aurora Alpha",
|
|
5633
|
-
contextWindow: 128000,
|
|
5634
|
-
maxTokens: 50000,
|
|
5635
|
-
capabilities: [
|
|
5636
|
-
"tools",
|
|
5637
|
-
"reasoning",
|
|
5638
|
-
"structured_output",
|
|
5639
|
-
"temperature",
|
|
5640
|
-
],
|
|
5641
|
-
pricing: {
|
|
5642
|
-
input: 0,
|
|
5643
|
-
output: 0,
|
|
5644
|
-
cacheRead: 0,
|
|
5645
|
-
cacheWrite: 0,
|
|
5646
|
-
},
|
|
5647
|
-
releaseDate: "2026-02-09",
|
|
5648
|
-
},
|
|
5649
5658
|
"anthropic/claude-opus-4.6": {
|
|
5650
5659
|
id: "anthropic/claude-opus-4.6",
|
|
5651
5660
|
name: "Claude Opus 4.6",
|
|
@@ -5691,7 +5700,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5691
5700
|
id: "stepfun/step-3.5-flash",
|
|
5692
5701
|
name: "Step 3.5 Flash",
|
|
5693
5702
|
contextWindow: 256000,
|
|
5694
|
-
maxTokens:
|
|
5703
|
+
maxTokens: 12800,
|
|
5695
5704
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
5696
5705
|
pricing: {
|
|
5697
5706
|
input: 0.1,
|
|
@@ -5705,7 +5714,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5705
5714
|
id: "stepfun/step-3.5-flash:free",
|
|
5706
5715
|
name: "Step 3.5 Flash (free)",
|
|
5707
5716
|
contextWindow: 256000,
|
|
5708
|
-
maxTokens:
|
|
5717
|
+
maxTokens: 12800,
|
|
5709
5718
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
5710
5719
|
pricing: {
|
|
5711
5720
|
input: 0,
|
|
@@ -5719,7 +5728,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5719
5728
|
id: "arcee-ai/trinity-large-preview:free",
|
|
5720
5729
|
name: "Trinity Large Preview",
|
|
5721
5730
|
contextWindow: 131072,
|
|
5722
|
-
maxTokens:
|
|
5731
|
+
maxTokens: 6553,
|
|
5723
5732
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
5724
5733
|
pricing: {
|
|
5725
5734
|
input: 0,
|
|
@@ -5733,7 +5742,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5733
5742
|
id: "arcee-ai/trinity-mini:free",
|
|
5734
5743
|
name: "Trinity Mini",
|
|
5735
5744
|
contextWindow: 131072,
|
|
5736
|
-
maxTokens:
|
|
5745
|
+
maxTokens: 6553,
|
|
5737
5746
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
5738
5747
|
pricing: {
|
|
5739
5748
|
input: 0,
|
|
@@ -5747,7 +5756,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5747
5756
|
id: "moonshotai/kimi-k2.5",
|
|
5748
5757
|
name: "Kimi K2.5",
|
|
5749
5758
|
contextWindow: 262144,
|
|
5750
|
-
maxTokens:
|
|
5759
|
+
maxTokens: 13107,
|
|
5751
5760
|
capabilities: [
|
|
5752
5761
|
"images",
|
|
5753
5762
|
"tools",
|
|
@@ -5865,7 +5874,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5865
5874
|
id: "nvidia/nemotron-3-nano-30b-a3b:free",
|
|
5866
5875
|
name: "Nemotron 3 Nano 30B A3B (free)",
|
|
5867
5876
|
contextWindow: 256000,
|
|
5868
|
-
maxTokens:
|
|
5877
|
+
maxTokens: 12800,
|
|
5869
5878
|
capabilities: [
|
|
5870
5879
|
"tools",
|
|
5871
5880
|
"reasoning",
|
|
@@ -5979,25 +5988,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
5979
5988
|
},
|
|
5980
5989
|
releaseDate: "2025-12-01",
|
|
5981
5990
|
},
|
|
5982
|
-
"tngtech/tng-r1t-chimera:free": {
|
|
5983
|
-
id: "tngtech/tng-r1t-chimera:free",
|
|
5984
|
-
name: "R1T Chimera (free)",
|
|
5985
|
-
contextWindow: 163840,
|
|
5986
|
-
maxTokens: 163840,
|
|
5987
|
-
capabilities: [
|
|
5988
|
-
"tools",
|
|
5989
|
-
"reasoning",
|
|
5990
|
-
"structured_output",
|
|
5991
|
-
"temperature",
|
|
5992
|
-
],
|
|
5993
|
-
pricing: {
|
|
5994
|
-
input: 0,
|
|
5995
|
-
output: 0,
|
|
5996
|
-
cacheRead: 0,
|
|
5997
|
-
cacheWrite: 0,
|
|
5998
|
-
},
|
|
5999
|
-
releaseDate: "2025-11-26",
|
|
6000
|
-
},
|
|
6001
5991
|
"anthropic/claude-opus-4.5": {
|
|
6002
5992
|
id: "anthropic/claude-opus-4.5",
|
|
6003
5993
|
name: "Claude Opus 4.5",
|
|
@@ -6060,34 +6050,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6060
6050
|
},
|
|
6061
6051
|
releaseDate: "2025-11-18",
|
|
6062
6052
|
},
|
|
6063
|
-
"openrouter/sherlock-dash-alpha": {
|
|
6064
|
-
id: "openrouter/sherlock-dash-alpha",
|
|
6065
|
-
name: "Sherlock Dash Alpha",
|
|
6066
|
-
contextWindow: 1840000,
|
|
6067
|
-
maxTokens: 0,
|
|
6068
|
-
capabilities: ["images", "tools", "temperature"],
|
|
6069
|
-
pricing: {
|
|
6070
|
-
input: 0,
|
|
6071
|
-
output: 0,
|
|
6072
|
-
cacheRead: 0,
|
|
6073
|
-
cacheWrite: 0,
|
|
6074
|
-
},
|
|
6075
|
-
releaseDate: "2025-11-15",
|
|
6076
|
-
},
|
|
6077
|
-
"openrouter/sherlock-think-alpha": {
|
|
6078
|
-
id: "openrouter/sherlock-think-alpha",
|
|
6079
|
-
name: "Sherlock Think Alpha",
|
|
6080
|
-
contextWindow: 1840000,
|
|
6081
|
-
maxTokens: 0,
|
|
6082
|
-
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
6083
|
-
pricing: {
|
|
6084
|
-
input: 0,
|
|
6085
|
-
output: 0,
|
|
6086
|
-
cacheRead: 0,
|
|
6087
|
-
cacheWrite: 0,
|
|
6088
|
-
},
|
|
6089
|
-
releaseDate: "2025-11-15",
|
|
6090
|
-
},
|
|
6091
6053
|
"openai/gpt-5.1": {
|
|
6092
6054
|
id: "openai/gpt-5.1",
|
|
6093
6055
|
name: "GPT-5.1",
|
|
@@ -6188,25 +6150,11 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6188
6150
|
},
|
|
6189
6151
|
releaseDate: "2025-11-13",
|
|
6190
6152
|
},
|
|
6191
|
-
"kwaipilot/kat-coder-pro:free": {
|
|
6192
|
-
id: "kwaipilot/kat-coder-pro:free",
|
|
6193
|
-
name: "Kat Coder Pro (free)",
|
|
6194
|
-
contextWindow: 256000,
|
|
6195
|
-
maxTokens: 65536,
|
|
6196
|
-
capabilities: ["tools", "structured_output", "temperature"],
|
|
6197
|
-
pricing: {
|
|
6198
|
-
input: 0,
|
|
6199
|
-
output: 0,
|
|
6200
|
-
cacheRead: 0,
|
|
6201
|
-
cacheWrite: 0,
|
|
6202
|
-
},
|
|
6203
|
-
releaseDate: "2025-11-10",
|
|
6204
|
-
},
|
|
6205
6153
|
"moonshotai/kimi-k2-thinking": {
|
|
6206
6154
|
id: "moonshotai/kimi-k2-thinking",
|
|
6207
6155
|
name: "Kimi K2 Thinking",
|
|
6208
6156
|
contextWindow: 262144,
|
|
6209
|
-
maxTokens:
|
|
6157
|
+
maxTokens: 13107,
|
|
6210
6158
|
capabilities: [
|
|
6211
6159
|
"tools",
|
|
6212
6160
|
"reasoning",
|
|
@@ -6239,7 +6187,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6239
6187
|
id: "nvidia/nemotron-nano-12b-v2-vl:free",
|
|
6240
6188
|
name: "Nemotron Nano 12B 2 VL (free)",
|
|
6241
6189
|
contextWindow: 128000,
|
|
6242
|
-
maxTokens:
|
|
6190
|
+
maxTokens: 6400,
|
|
6243
6191
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
6244
6192
|
pricing: {
|
|
6245
6193
|
input: 0,
|
|
@@ -6487,7 +6435,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6487
6435
|
id: "mistralai/devstral-2512",
|
|
6488
6436
|
name: "Devstral 2 2512",
|
|
6489
6437
|
contextWindow: 262144,
|
|
6490
|
-
maxTokens:
|
|
6438
|
+
maxTokens: 13107,
|
|
6491
6439
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
6492
6440
|
pricing: {
|
|
6493
6441
|
input: 0.15,
|
|
@@ -6497,25 +6445,11 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6497
6445
|
},
|
|
6498
6446
|
releaseDate: "2025-09-12",
|
|
6499
6447
|
},
|
|
6500
|
-
"mistralai/devstral-2512:free": {
|
|
6501
|
-
id: "mistralai/devstral-2512:free",
|
|
6502
|
-
name: "Devstral 2 2512 (free)",
|
|
6503
|
-
contextWindow: 262144,
|
|
6504
|
-
maxTokens: 262144,
|
|
6505
|
-
capabilities: ["tools", "temperature"],
|
|
6506
|
-
pricing: {
|
|
6507
|
-
input: 0,
|
|
6508
|
-
output: 0,
|
|
6509
|
-
cacheRead: 0,
|
|
6510
|
-
cacheWrite: 0,
|
|
6511
|
-
},
|
|
6512
|
-
releaseDate: "2025-09-12",
|
|
6513
|
-
},
|
|
6514
6448
|
"qwen/qwen3-next-80b-a3b-instruct": {
|
|
6515
6449
|
id: "qwen/qwen3-next-80b-a3b-instruct",
|
|
6516
6450
|
name: "Qwen3 Next 80B A3B Instruct",
|
|
6517
6451
|
contextWindow: 262144,
|
|
6518
|
-
maxTokens:
|
|
6452
|
+
maxTokens: 13107,
|
|
6519
6453
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
6520
6454
|
pricing: {
|
|
6521
6455
|
input: 0.14,
|
|
@@ -6529,7 +6463,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6529
6463
|
id: "qwen/qwen3-next-80b-a3b-instruct:free",
|
|
6530
6464
|
name: "Qwen3 Next 80B A3B Instruct (free)",
|
|
6531
6465
|
contextWindow: 262144,
|
|
6532
|
-
maxTokens:
|
|
6466
|
+
maxTokens: 13107,
|
|
6533
6467
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
6534
6468
|
pricing: {
|
|
6535
6469
|
input: 0,
|
|
@@ -6543,7 +6477,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6543
6477
|
id: "qwen/qwen3-next-80b-a3b-thinking",
|
|
6544
6478
|
name: "Qwen3 Next 80B A3B Thinking",
|
|
6545
6479
|
contextWindow: 262144,
|
|
6546
|
-
maxTokens:
|
|
6480
|
+
maxTokens: 13107,
|
|
6547
6481
|
capabilities: [
|
|
6548
6482
|
"tools",
|
|
6549
6483
|
"reasoning",
|
|
@@ -6590,7 +6524,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6590
6524
|
id: "nvidia/nemotron-nano-9b-v2:free",
|
|
6591
6525
|
name: "Nemotron Nano 9B V2 (free)",
|
|
6592
6526
|
contextWindow: 128000,
|
|
6593
|
-
maxTokens:
|
|
6527
|
+
maxTokens: 6400,
|
|
6594
6528
|
capabilities: [
|
|
6595
6529
|
"tools",
|
|
6596
6530
|
"reasoning",
|
|
@@ -6642,7 +6576,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6642
6576
|
id: "nousresearch/hermes-4-405b",
|
|
6643
6577
|
name: "Hermes 4 405B",
|
|
6644
6578
|
contextWindow: 131072,
|
|
6645
|
-
maxTokens:
|
|
6579
|
+
maxTokens: 6553,
|
|
6646
6580
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
6647
6581
|
pricing: {
|
|
6648
6582
|
input: 1,
|
|
@@ -6656,7 +6590,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6656
6590
|
id: "nousresearch/hermes-4-70b",
|
|
6657
6591
|
name: "Hermes 4 70B",
|
|
6658
6592
|
contextWindow: 131072,
|
|
6659
|
-
maxTokens:
|
|
6593
|
+
maxTokens: 6553,
|
|
6660
6594
|
capabilities: [
|
|
6661
6595
|
"tools",
|
|
6662
6596
|
"reasoning",
|
|
@@ -6675,7 +6609,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6675
6609
|
id: "deepseek/deepseek-chat-v3.1",
|
|
6676
6610
|
name: "DeepSeek-V3.1",
|
|
6677
6611
|
contextWindow: 163840,
|
|
6678
|
-
maxTokens:
|
|
6612
|
+
maxTokens: 8192,
|
|
6679
6613
|
capabilities: [
|
|
6680
6614
|
"tools",
|
|
6681
6615
|
"reasoning",
|
|
@@ -6714,7 +6648,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6714
6648
|
id: "nvidia/nemotron-nano-9b-v2",
|
|
6715
6649
|
name: "nvidia-nemotron-nano-9b-v2",
|
|
6716
6650
|
contextWindow: 131072,
|
|
6717
|
-
maxTokens:
|
|
6651
|
+
maxTokens: 6553,
|
|
6718
6652
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
6719
6653
|
pricing: {
|
|
6720
6654
|
input: 0.04,
|
|
@@ -6728,7 +6662,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6728
6662
|
id: "mistralai/mistral-medium-3.1",
|
|
6729
6663
|
name: "Mistral Medium 3.1",
|
|
6730
6664
|
contextWindow: 262144,
|
|
6731
|
-
maxTokens:
|
|
6665
|
+
maxTokens: 13107,
|
|
6732
6666
|
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
6733
6667
|
pricing: {
|
|
6734
6668
|
input: 0.4,
|
|
@@ -6928,7 +6862,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6928
6862
|
id: "mistralai/codestral-2508",
|
|
6929
6863
|
name: "Codestral 2508",
|
|
6930
6864
|
contextWindow: 256000,
|
|
6931
|
-
maxTokens:
|
|
6865
|
+
maxTokens: 12800,
|
|
6932
6866
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
6933
6867
|
pricing: {
|
|
6934
6868
|
input: 0.3,
|
|
@@ -6956,7 +6890,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6956
6890
|
id: "qwen/qwen3-30b-a3b-instruct-2507",
|
|
6957
6891
|
name: "Qwen3 30B A3B Instruct 2507",
|
|
6958
6892
|
contextWindow: 262000,
|
|
6959
|
-
maxTokens:
|
|
6893
|
+
maxTokens: 13100,
|
|
6960
6894
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
6961
6895
|
pricing: {
|
|
6962
6896
|
input: 0.2,
|
|
@@ -6970,7 +6904,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
6970
6904
|
id: "qwen/qwen3-30b-a3b-thinking-2507",
|
|
6971
6905
|
name: "Qwen3 30B A3B Thinking 2507",
|
|
6972
6906
|
contextWindow: 262000,
|
|
6973
|
-
maxTokens:
|
|
6907
|
+
maxTokens: 13100,
|
|
6974
6908
|
capabilities: [
|
|
6975
6909
|
"tools",
|
|
6976
6910
|
"reasoning",
|
|
@@ -7137,7 +7071,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7137
7071
|
id: "moonshotai/kimi-k2:free",
|
|
7138
7072
|
name: "Kimi K2 (free)",
|
|
7139
7073
|
contextWindow: 32800,
|
|
7140
|
-
maxTokens:
|
|
7074
|
+
maxTokens: 1640,
|
|
7141
7075
|
capabilities: ["tools", "temperature"],
|
|
7142
7076
|
pricing: {
|
|
7143
7077
|
input: 0,
|
|
@@ -7151,7 +7085,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7151
7085
|
id: "mistralai/devstral-medium-2507",
|
|
7152
7086
|
name: "Devstral Medium",
|
|
7153
7087
|
contextWindow: 131072,
|
|
7154
|
-
maxTokens:
|
|
7088
|
+
maxTokens: 6553,
|
|
7155
7089
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
7156
7090
|
pricing: {
|
|
7157
7091
|
input: 0.4,
|
|
@@ -7165,7 +7099,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7165
7099
|
id: "mistralai/devstral-small-2507",
|
|
7166
7100
|
name: "Devstral Small 1.1",
|
|
7167
7101
|
contextWindow: 131072,
|
|
7168
|
-
maxTokens:
|
|
7102
|
+
maxTokens: 6553,
|
|
7169
7103
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
7170
7104
|
pricing: {
|
|
7171
7105
|
input: 0.1,
|
|
@@ -7222,20 +7156,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7222
7156
|
},
|
|
7223
7157
|
releaseDate: "2025-06-20",
|
|
7224
7158
|
},
|
|
7225
|
-
"mistralai/mistral-small-3.2-24b-instruct:free": {
|
|
7226
|
-
id: "mistralai/mistral-small-3.2-24b-instruct:free",
|
|
7227
|
-
name: "Mistral Small 3.2 24B (free)",
|
|
7228
|
-
contextWindow: 96000,
|
|
7229
|
-
maxTokens: 96000,
|
|
7230
|
-
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
7231
|
-
pricing: {
|
|
7232
|
-
input: 0,
|
|
7233
|
-
output: 0,
|
|
7234
|
-
cacheRead: 0,
|
|
7235
|
-
cacheWrite: 0,
|
|
7236
|
-
},
|
|
7237
|
-
releaseDate: "2025-06-20",
|
|
7238
|
-
},
|
|
7239
7159
|
"google/gemini-2.5-flash-lite": {
|
|
7240
7160
|
id: "google/gemini-2.5-flash-lite",
|
|
7241
7161
|
name: "Gemini 2.5 Flash Lite",
|
|
@@ -7271,20 +7191,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7271
7191
|
},
|
|
7272
7192
|
releaseDate: "2025-06-17",
|
|
7273
7193
|
},
|
|
7274
|
-
"moonshotai/kimi-dev-72b:free": {
|
|
7275
|
-
id: "moonshotai/kimi-dev-72b:free",
|
|
7276
|
-
name: "Kimi Dev 72b (free)",
|
|
7277
|
-
contextWindow: 131072,
|
|
7278
|
-
maxTokens: 131072,
|
|
7279
|
-
capabilities: ["tools", "temperature"],
|
|
7280
|
-
pricing: {
|
|
7281
|
-
input: 0,
|
|
7282
|
-
output: 0,
|
|
7283
|
-
cacheRead: 0,
|
|
7284
|
-
cacheWrite: 0,
|
|
7285
|
-
},
|
|
7286
|
-
releaseDate: "2025-06-16",
|
|
7287
|
-
},
|
|
7288
7194
|
"google/gemini-2.5-pro-preview-06-05": {
|
|
7289
7195
|
id: "google/gemini-2.5-pro-preview-06-05",
|
|
7290
7196
|
name: "Gemini 2.5 Pro Preview 06-05",
|
|
@@ -7306,34 +7212,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7306
7212
|
},
|
|
7307
7213
|
releaseDate: "2025-06-05",
|
|
7308
7214
|
},
|
|
7309
|
-
"deepseek/deepseek-r1-0528-qwen3-8b:free": {
|
|
7310
|
-
id: "deepseek/deepseek-r1-0528-qwen3-8b:free",
|
|
7311
|
-
name: "Deepseek R1 0528 Qwen3 8B (free)",
|
|
7312
|
-
contextWindow: 131072,
|
|
7313
|
-
maxTokens: 131072,
|
|
7314
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7315
|
-
pricing: {
|
|
7316
|
-
input: 0,
|
|
7317
|
-
output: 0,
|
|
7318
|
-
cacheRead: 0,
|
|
7319
|
-
cacheWrite: 0,
|
|
7320
|
-
},
|
|
7321
|
-
releaseDate: "2025-05-29",
|
|
7322
|
-
},
|
|
7323
|
-
"sarvamai/sarvam-m:free": {
|
|
7324
|
-
id: "sarvamai/sarvam-m:free",
|
|
7325
|
-
name: "Sarvam-M (free)",
|
|
7326
|
-
contextWindow: 32768,
|
|
7327
|
-
maxTokens: 32768,
|
|
7328
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7329
|
-
pricing: {
|
|
7330
|
-
input: 0,
|
|
7331
|
-
output: 0,
|
|
7332
|
-
cacheRead: 0,
|
|
7333
|
-
cacheWrite: 0,
|
|
7334
|
-
},
|
|
7335
|
-
releaseDate: "2025-05-25",
|
|
7336
|
-
},
|
|
7337
7215
|
"anthropic/claude-opus-4": {
|
|
7338
7216
|
id: "anthropic/claude-opus-4",
|
|
7339
7217
|
name: "Claude Opus 4",
|
|
@@ -7362,25 +7240,11 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7362
7240
|
},
|
|
7363
7241
|
releaseDate: "2025-05-22",
|
|
7364
7242
|
},
|
|
7365
|
-
"mistralai/devstral-small-2505:free": {
|
|
7366
|
-
id: "mistralai/devstral-small-2505:free",
|
|
7367
|
-
name: "Devstral Small 2505 (free)",
|
|
7368
|
-
contextWindow: 32768,
|
|
7369
|
-
maxTokens: 32768,
|
|
7370
|
-
capabilities: ["tools", "temperature"],
|
|
7371
|
-
pricing: {
|
|
7372
|
-
input: 0,
|
|
7373
|
-
output: 0,
|
|
7374
|
-
cacheRead: 0,
|
|
7375
|
-
cacheWrite: 0,
|
|
7376
|
-
},
|
|
7377
|
-
releaseDate: "2025-05-21",
|
|
7378
|
-
},
|
|
7379
7243
|
"mistralai/devstral-small-2505": {
|
|
7380
7244
|
id: "mistralai/devstral-small-2505",
|
|
7381
7245
|
name: "Devstral Small",
|
|
7382
7246
|
contextWindow: 128000,
|
|
7383
|
-
maxTokens:
|
|
7247
|
+
maxTokens: 6400,
|
|
7384
7248
|
capabilities: ["tools", "temperature"],
|
|
7385
7249
|
pricing: {
|
|
7386
7250
|
input: 0.06,
|
|
@@ -7394,7 +7258,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7394
7258
|
id: "mistralai/mistral-medium-3",
|
|
7395
7259
|
name: "Mistral Medium 3",
|
|
7396
7260
|
contextWindow: 131072,
|
|
7397
|
-
maxTokens:
|
|
7261
|
+
maxTokens: 6553,
|
|
7398
7262
|
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
7399
7263
|
pricing: {
|
|
7400
7264
|
input: 0.4,
|
|
@@ -7443,7 +7307,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7443
7307
|
id: "qwen/qwen3-4b:free",
|
|
7444
7308
|
name: "Qwen3 4B (free)",
|
|
7445
7309
|
contextWindow: 40960,
|
|
7446
|
-
maxTokens:
|
|
7310
|
+
maxTokens: 2048,
|
|
7447
7311
|
capabilities: [
|
|
7448
7312
|
"tools",
|
|
7449
7313
|
"reasoning",
|
|
@@ -7458,25 +7322,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7458
7322
|
},
|
|
7459
7323
|
releaseDate: "2025-04-30",
|
|
7460
7324
|
},
|
|
7461
|
-
"qwen/qwen3-14b:free": {
|
|
7462
|
-
id: "qwen/qwen3-14b:free",
|
|
7463
|
-
name: "Qwen3 14B (free)",
|
|
7464
|
-
contextWindow: 40960,
|
|
7465
|
-
maxTokens: 40960,
|
|
7466
|
-
capabilities: [
|
|
7467
|
-
"tools",
|
|
7468
|
-
"reasoning",
|
|
7469
|
-
"structured_output",
|
|
7470
|
-
"temperature",
|
|
7471
|
-
],
|
|
7472
|
-
pricing: {
|
|
7473
|
-
input: 0,
|
|
7474
|
-
output: 0,
|
|
7475
|
-
cacheRead: 0,
|
|
7476
|
-
cacheWrite: 0,
|
|
7477
|
-
},
|
|
7478
|
-
releaseDate: "2025-04-28",
|
|
7479
|
-
},
|
|
7480
7325
|
"qwen/qwen3-235b-a22b-07-25": {
|
|
7481
7326
|
id: "qwen/qwen3-235b-a22b-07-25",
|
|
7482
7327
|
name: "Qwen3 235B A22B Instruct 2507",
|
|
@@ -7491,140 +7336,22 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7491
7336
|
},
|
|
7492
7337
|
releaseDate: "2025-04-28",
|
|
7493
7338
|
},
|
|
7494
|
-
"
|
|
7495
|
-
id: "
|
|
7496
|
-
name: "
|
|
7497
|
-
contextWindow:
|
|
7498
|
-
maxTokens:
|
|
7499
|
-
capabilities: ["tools", "temperature"],
|
|
7500
|
-
pricing: {
|
|
7501
|
-
input: 0,
|
|
7502
|
-
output: 0,
|
|
7503
|
-
cacheRead: 0,
|
|
7504
|
-
cacheWrite: 0,
|
|
7505
|
-
},
|
|
7506
|
-
releaseDate: "2025-04-28",
|
|
7507
|
-
},
|
|
7508
|
-
"qwen/qwen3-235b-a22b:free": {
|
|
7509
|
-
id: "qwen/qwen3-235b-a22b:free",
|
|
7510
|
-
name: "Qwen3 235B A22B (free)",
|
|
7511
|
-
contextWindow: 131072,
|
|
7512
|
-
maxTokens: 131072,
|
|
7339
|
+
"openai/o4-mini": {
|
|
7340
|
+
id: "openai/o4-mini",
|
|
7341
|
+
name: "o4 Mini",
|
|
7342
|
+
contextWindow: 200000,
|
|
7343
|
+
maxTokens: 100000,
|
|
7513
7344
|
capabilities: [
|
|
7345
|
+
"images",
|
|
7514
7346
|
"tools",
|
|
7515
7347
|
"reasoning",
|
|
7516
7348
|
"structured_output",
|
|
7517
7349
|
"temperature",
|
|
7518
7350
|
],
|
|
7519
7351
|
pricing: {
|
|
7520
|
-
input:
|
|
7521
|
-
output:
|
|
7522
|
-
cacheRead: 0,
|
|
7523
|
-
cacheWrite: 0,
|
|
7524
|
-
},
|
|
7525
|
-
releaseDate: "2025-04-28",
|
|
7526
|
-
},
|
|
7527
|
-
"qwen/qwen3-30b-a3b:free": {
|
|
7528
|
-
id: "qwen/qwen3-30b-a3b:free",
|
|
7529
|
-
name: "Qwen3 30B A3B (free)",
|
|
7530
|
-
contextWindow: 40960,
|
|
7531
|
-
maxTokens: 40960,
|
|
7532
|
-
capabilities: [
|
|
7533
|
-
"tools",
|
|
7534
|
-
"reasoning",
|
|
7535
|
-
"structured_output",
|
|
7536
|
-
"temperature",
|
|
7537
|
-
],
|
|
7538
|
-
pricing: {
|
|
7539
|
-
input: 0,
|
|
7540
|
-
output: 0,
|
|
7541
|
-
cacheRead: 0,
|
|
7542
|
-
cacheWrite: 0,
|
|
7543
|
-
},
|
|
7544
|
-
releaseDate: "2025-04-28",
|
|
7545
|
-
},
|
|
7546
|
-
"qwen/qwen3-32b:free": {
|
|
7547
|
-
id: "qwen/qwen3-32b:free",
|
|
7548
|
-
name: "Qwen3 32B (free)",
|
|
7549
|
-
contextWindow: 40960,
|
|
7550
|
-
maxTokens: 40960,
|
|
7551
|
-
capabilities: [
|
|
7552
|
-
"tools",
|
|
7553
|
-
"reasoning",
|
|
7554
|
-
"structured_output",
|
|
7555
|
-
"temperature",
|
|
7556
|
-
],
|
|
7557
|
-
pricing: {
|
|
7558
|
-
input: 0,
|
|
7559
|
-
output: 0,
|
|
7560
|
-
cacheRead: 0,
|
|
7561
|
-
cacheWrite: 0,
|
|
7562
|
-
},
|
|
7563
|
-
releaseDate: "2025-04-28",
|
|
7564
|
-
},
|
|
7565
|
-
"qwen/qwen3-8b:free": {
|
|
7566
|
-
id: "qwen/qwen3-8b:free",
|
|
7567
|
-
name: "Qwen3 8B (free)",
|
|
7568
|
-
contextWindow: 40960,
|
|
7569
|
-
maxTokens: 40960,
|
|
7570
|
-
capabilities: [
|
|
7571
|
-
"tools",
|
|
7572
|
-
"reasoning",
|
|
7573
|
-
"structured_output",
|
|
7574
|
-
"temperature",
|
|
7575
|
-
],
|
|
7576
|
-
pricing: {
|
|
7577
|
-
input: 0,
|
|
7578
|
-
output: 0,
|
|
7579
|
-
cacheRead: 0,
|
|
7580
|
-
cacheWrite: 0,
|
|
7581
|
-
},
|
|
7582
|
-
releaseDate: "2025-04-28",
|
|
7583
|
-
},
|
|
7584
|
-
"microsoft/mai-ds-r1:free": {
|
|
7585
|
-
id: "microsoft/mai-ds-r1:free",
|
|
7586
|
-
name: "MAI DS R1 (free)",
|
|
7587
|
-
contextWindow: 163840,
|
|
7588
|
-
maxTokens: 163840,
|
|
7589
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7590
|
-
pricing: {
|
|
7591
|
-
input: 0,
|
|
7592
|
-
output: 0,
|
|
7593
|
-
cacheRead: 0,
|
|
7594
|
-
cacheWrite: 0,
|
|
7595
|
-
},
|
|
7596
|
-
releaseDate: "2025-04-21",
|
|
7597
|
-
},
|
|
7598
|
-
"thudm/glm-z1-32b:free": {
|
|
7599
|
-
id: "thudm/glm-z1-32b:free",
|
|
7600
|
-
name: "GLM Z1 32B (free)",
|
|
7601
|
-
contextWindow: 32768,
|
|
7602
|
-
maxTokens: 32768,
|
|
7603
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7604
|
-
pricing: {
|
|
7605
|
-
input: 0,
|
|
7606
|
-
output: 0,
|
|
7607
|
-
cacheRead: 0,
|
|
7608
|
-
cacheWrite: 0,
|
|
7609
|
-
},
|
|
7610
|
-
releaseDate: "2025-04-17",
|
|
7611
|
-
},
|
|
7612
|
-
"openai/o4-mini": {
|
|
7613
|
-
id: "openai/o4-mini",
|
|
7614
|
-
name: "o4 Mini",
|
|
7615
|
-
contextWindow: 200000,
|
|
7616
|
-
maxTokens: 100000,
|
|
7617
|
-
capabilities: [
|
|
7618
|
-
"images",
|
|
7619
|
-
"tools",
|
|
7620
|
-
"reasoning",
|
|
7621
|
-
"structured_output",
|
|
7622
|
-
"temperature",
|
|
7623
|
-
],
|
|
7624
|
-
pricing: {
|
|
7625
|
-
input: 1.1,
|
|
7626
|
-
output: 4.4,
|
|
7627
|
-
cacheRead: 0.28,
|
|
7352
|
+
input: 1.1,
|
|
7353
|
+
output: 4.4,
|
|
7354
|
+
cacheRead: 0.28,
|
|
7628
7355
|
cacheWrite: 0,
|
|
7629
7356
|
},
|
|
7630
7357
|
releaseDate: "2025-04-16",
|
|
@@ -7657,34 +7384,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7657
7384
|
},
|
|
7658
7385
|
releaseDate: "2025-04-14",
|
|
7659
7386
|
},
|
|
7660
|
-
"meta-llama/llama-4-scout:free": {
|
|
7661
|
-
id: "meta-llama/llama-4-scout:free",
|
|
7662
|
-
name: "Llama 4 Scout (free)",
|
|
7663
|
-
contextWindow: 64000,
|
|
7664
|
-
maxTokens: 64000,
|
|
7665
|
-
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
7666
|
-
pricing: {
|
|
7667
|
-
input: 0,
|
|
7668
|
-
output: 0,
|
|
7669
|
-
cacheRead: 0,
|
|
7670
|
-
cacheWrite: 0,
|
|
7671
|
-
},
|
|
7672
|
-
releaseDate: "2025-04-05",
|
|
7673
|
-
},
|
|
7674
|
-
"qwen/qwen2.5-vl-32b-instruct:free": {
|
|
7675
|
-
id: "qwen/qwen2.5-vl-32b-instruct:free",
|
|
7676
|
-
name: "Qwen2.5 VL 32B Instruct (free)",
|
|
7677
|
-
contextWindow: 8192,
|
|
7678
|
-
maxTokens: 8192,
|
|
7679
|
-
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
7680
|
-
pricing: {
|
|
7681
|
-
input: 0,
|
|
7682
|
-
output: 0,
|
|
7683
|
-
cacheRead: 0,
|
|
7684
|
-
cacheWrite: 0,
|
|
7685
|
-
},
|
|
7686
|
-
releaseDate: "2025-03-24",
|
|
7687
|
-
},
|
|
7688
7387
|
"google/gemini-2.5-pro": {
|
|
7689
7388
|
id: "google/gemini-2.5-pro",
|
|
7690
7389
|
name: "Gemini 2.5 Pro",
|
|
@@ -7724,7 +7423,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7724
7423
|
id: "google/gemma-3-27b-it",
|
|
7725
7424
|
name: "Gemma 3 27B",
|
|
7726
7425
|
contextWindow: 96000,
|
|
7727
|
-
maxTokens:
|
|
7426
|
+
maxTokens: 4800,
|
|
7728
7427
|
capabilities: ["images", "tools", "structured_output", "temperature"],
|
|
7729
7428
|
pricing: {
|
|
7730
7429
|
input: 0.04,
|
|
@@ -7748,53 +7447,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7748
7447
|
},
|
|
7749
7448
|
releaseDate: "2025-03-12",
|
|
7750
7449
|
},
|
|
7751
|
-
"rekaai/reka-flash-3": {
|
|
7752
|
-
id: "rekaai/reka-flash-3",
|
|
7753
|
-
name: "Reka Flash 3",
|
|
7754
|
-
contextWindow: 32768,
|
|
7755
|
-
maxTokens: 8192,
|
|
7756
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7757
|
-
pricing: {
|
|
7758
|
-
input: 0,
|
|
7759
|
-
output: 0,
|
|
7760
|
-
cacheRead: 0,
|
|
7761
|
-
cacheWrite: 0,
|
|
7762
|
-
},
|
|
7763
|
-
releaseDate: "2025-03-12",
|
|
7764
|
-
},
|
|
7765
|
-
"qwen/qwq-32b:free": {
|
|
7766
|
-
id: "qwen/qwq-32b:free",
|
|
7767
|
-
name: "QwQ 32B (free)",
|
|
7768
|
-
contextWindow: 32768,
|
|
7769
|
-
maxTokens: 32768,
|
|
7770
|
-
capabilities: [
|
|
7771
|
-
"tools",
|
|
7772
|
-
"reasoning",
|
|
7773
|
-
"structured_output",
|
|
7774
|
-
"temperature",
|
|
7775
|
-
],
|
|
7776
|
-
pricing: {
|
|
7777
|
-
input: 0,
|
|
7778
|
-
output: 0,
|
|
7779
|
-
cacheRead: 0,
|
|
7780
|
-
cacheWrite: 0,
|
|
7781
|
-
},
|
|
7782
|
-
releaseDate: "2025-03-05",
|
|
7783
|
-
},
|
|
7784
|
-
"nousresearch/deephermes-3-llama-3-8b-preview": {
|
|
7785
|
-
id: "nousresearch/deephermes-3-llama-3-8b-preview",
|
|
7786
|
-
name: "DeepHermes 3 Llama 3 8B Preview",
|
|
7787
|
-
contextWindow: 131072,
|
|
7788
|
-
maxTokens: 8192,
|
|
7789
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7790
|
-
pricing: {
|
|
7791
|
-
input: 0,
|
|
7792
|
-
output: 0,
|
|
7793
|
-
cacheRead: 0,
|
|
7794
|
-
cacheWrite: 0,
|
|
7795
|
-
},
|
|
7796
|
-
releaseDate: "2025-02-28",
|
|
7797
|
-
},
|
|
7798
7450
|
"anthropic/claude-3.7-sonnet": {
|
|
7799
7451
|
id: "anthropic/claude-3.7-sonnet",
|
|
7800
7452
|
name: "Claude Sonnet 3.7",
|
|
@@ -7870,67 +7522,11 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7870
7522
|
},
|
|
7871
7523
|
releaseDate: "2025-02-17",
|
|
7872
7524
|
},
|
|
7873
|
-
"cognitivecomputations/dolphin3.0-mistral-24b": {
|
|
7874
|
-
id: "cognitivecomputations/dolphin3.0-mistral-24b",
|
|
7875
|
-
name: "Dolphin3.0 Mistral 24B",
|
|
7876
|
-
contextWindow: 32768,
|
|
7877
|
-
maxTokens: 8192,
|
|
7878
|
-
capabilities: ["tools", "temperature"],
|
|
7879
|
-
pricing: {
|
|
7880
|
-
input: 0,
|
|
7881
|
-
output: 0,
|
|
7882
|
-
cacheRead: 0,
|
|
7883
|
-
cacheWrite: 0,
|
|
7884
|
-
},
|
|
7885
|
-
releaseDate: "2025-02-13",
|
|
7886
|
-
},
|
|
7887
|
-
"cognitivecomputations/dolphin3.0-r1-mistral-24b": {
|
|
7888
|
-
id: "cognitivecomputations/dolphin3.0-r1-mistral-24b",
|
|
7889
|
-
name: "Dolphin3.0 R1 Mistral 24B",
|
|
7890
|
-
contextWindow: 32768,
|
|
7891
|
-
maxTokens: 8192,
|
|
7892
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7893
|
-
pricing: {
|
|
7894
|
-
input: 0,
|
|
7895
|
-
output: 0,
|
|
7896
|
-
cacheRead: 0,
|
|
7897
|
-
cacheWrite: 0,
|
|
7898
|
-
},
|
|
7899
|
-
releaseDate: "2025-02-13",
|
|
7900
|
-
},
|
|
7901
|
-
"qwen/qwen2.5-vl-72b-instruct:free": {
|
|
7902
|
-
id: "qwen/qwen2.5-vl-72b-instruct:free",
|
|
7903
|
-
name: "Qwen2.5 VL 72B Instruct (free)",
|
|
7904
|
-
contextWindow: 32768,
|
|
7905
|
-
maxTokens: 32768,
|
|
7906
|
-
capabilities: ["images", "tools", "temperature"],
|
|
7907
|
-
pricing: {
|
|
7908
|
-
input: 0,
|
|
7909
|
-
output: 0,
|
|
7910
|
-
cacheRead: 0,
|
|
7911
|
-
cacheWrite: 0,
|
|
7912
|
-
},
|
|
7913
|
-
releaseDate: "2025-02-01",
|
|
7914
|
-
},
|
|
7915
|
-
"deepseek/deepseek-r1:free": {
|
|
7916
|
-
id: "deepseek/deepseek-r1:free",
|
|
7917
|
-
name: "R1 (free)",
|
|
7918
|
-
contextWindow: 163840,
|
|
7919
|
-
maxTokens: 163840,
|
|
7920
|
-
capabilities: ["tools", "reasoning", "temperature"],
|
|
7921
|
-
pricing: {
|
|
7922
|
-
input: 0,
|
|
7923
|
-
output: 0,
|
|
7924
|
-
cacheRead: 0,
|
|
7925
|
-
cacheWrite: 0,
|
|
7926
|
-
},
|
|
7927
|
-
releaseDate: "2025-01-20",
|
|
7928
|
-
},
|
|
7929
7525
|
"minimax/minimax-01": {
|
|
7930
7526
|
id: "minimax/minimax-01",
|
|
7931
7527
|
name: "MiniMax-01",
|
|
7932
7528
|
contextWindow: 1000000,
|
|
7933
|
-
maxTokens:
|
|
7529
|
+
maxTokens: 50000,
|
|
7934
7530
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
7935
7531
|
pricing: {
|
|
7936
7532
|
input: 0.2,
|
|
@@ -7979,25 +7575,11 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
7979
7575
|
},
|
|
7980
7576
|
releaseDate: "2024-12-11",
|
|
7981
7577
|
},
|
|
7982
|
-
"google/gemini-2.0-flash-exp:free": {
|
|
7983
|
-
id: "google/gemini-2.0-flash-exp:free",
|
|
7984
|
-
name: "Gemini 2.0 Flash Experimental (free)",
|
|
7985
|
-
contextWindow: 1048576,
|
|
7986
|
-
maxTokens: 1048576,
|
|
7987
|
-
capabilities: ["images", "tools", "temperature"],
|
|
7988
|
-
pricing: {
|
|
7989
|
-
input: 0,
|
|
7990
|
-
output: 0,
|
|
7991
|
-
cacheRead: 0,
|
|
7992
|
-
cacheWrite: 0,
|
|
7993
|
-
},
|
|
7994
|
-
releaseDate: "2024-12-11",
|
|
7995
|
-
},
|
|
7996
7578
|
"meta-llama/llama-3.3-70b-instruct:free": {
|
|
7997
7579
|
id: "meta-llama/llama-3.3-70b-instruct:free",
|
|
7998
7580
|
name: "Llama 3.3 70B Instruct (free)",
|
|
7999
7581
|
contextWindow: 131072,
|
|
8000
|
-
maxTokens:
|
|
7582
|
+
maxTokens: 6553,
|
|
8001
7583
|
capabilities: ["tools", "structured_output", "temperature"],
|
|
8002
7584
|
pricing: {
|
|
8003
7585
|
input: 0,
|
|
@@ -8021,34 +7603,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
8021
7603
|
},
|
|
8022
7604
|
releaseDate: "2024-10-22",
|
|
8023
7605
|
},
|
|
8024
|
-
"qwen/qwen-2.5-vl-7b-instruct:free": {
|
|
8025
|
-
id: "qwen/qwen-2.5-vl-7b-instruct:free",
|
|
8026
|
-
name: "Qwen2.5-VL 7B Instruct (free)",
|
|
8027
|
-
contextWindow: 32768,
|
|
8028
|
-
maxTokens: 32768,
|
|
8029
|
-
capabilities: ["images", "tools", "temperature"],
|
|
8030
|
-
pricing: {
|
|
8031
|
-
input: 0,
|
|
8032
|
-
output: 0,
|
|
8033
|
-
cacheRead: 0,
|
|
8034
|
-
cacheWrite: 0,
|
|
8035
|
-
},
|
|
8036
|
-
releaseDate: "2024-08-28",
|
|
8037
|
-
},
|
|
8038
|
-
"mistralai/mistral-nemo:free": {
|
|
8039
|
-
id: "mistralai/mistral-nemo:free",
|
|
8040
|
-
name: "Mistral Nemo (free)",
|
|
8041
|
-
contextWindow: 131072,
|
|
8042
|
-
maxTokens: 131072,
|
|
8043
|
-
capabilities: ["tools", "structured_output", "temperature"],
|
|
8044
|
-
pricing: {
|
|
8045
|
-
input: 0,
|
|
8046
|
-
output: 0,
|
|
8047
|
-
cacheRead: 0,
|
|
8048
|
-
cacheWrite: 0,
|
|
8049
|
-
},
|
|
8050
|
-
releaseDate: "2024-07-19",
|
|
8051
|
-
},
|
|
8052
7606
|
"openai/gpt-4o-mini": {
|
|
8053
7607
|
id: "openai/gpt-4o-mini",
|
|
8054
7608
|
name: "GPT-4o-mini",
|
|
@@ -8063,20 +7617,6 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
8063
7617
|
},
|
|
8064
7618
|
releaseDate: "2024-07-18",
|
|
8065
7619
|
},
|
|
8066
|
-
"mistralai/mistral-7b-instruct:free": {
|
|
8067
|
-
id: "mistralai/mistral-7b-instruct:free",
|
|
8068
|
-
name: "Mistral 7B Instruct (free)",
|
|
8069
|
-
contextWindow: 32768,
|
|
8070
|
-
maxTokens: 32768,
|
|
8071
|
-
capabilities: ["tools", "temperature"],
|
|
8072
|
-
pricing: {
|
|
8073
|
-
input: 0,
|
|
8074
|
-
output: 0,
|
|
8075
|
-
cacheRead: 0,
|
|
8076
|
-
cacheWrite: 0,
|
|
8077
|
-
},
|
|
8078
|
-
releaseDate: "2024-05-27",
|
|
8079
|
-
},
|
|
8080
7620
|
},
|
|
8081
7621
|
requesty: {
|
|
8082
7622
|
"openai/gpt-5.4": {
|
|
@@ -9027,7 +8567,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9027
8567
|
id: "Qwen/Qwen3-Coder-Next-FP8",
|
|
9028
8568
|
name: "Qwen3 Coder Next FP8",
|
|
9029
8569
|
contextWindow: 262144,
|
|
9030
|
-
maxTokens:
|
|
8570
|
+
maxTokens: 13107,
|
|
9031
8571
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9032
8572
|
pricing: {
|
|
9033
8573
|
input: 0.5,
|
|
@@ -9041,7 +8581,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9041
8581
|
id: "moonshotai/Kimi-K2.5",
|
|
9042
8582
|
name: "Kimi K2.5",
|
|
9043
8583
|
contextWindow: 262144,
|
|
9044
|
-
maxTokens:
|
|
8584
|
+
maxTokens: 13107,
|
|
9045
8585
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
9046
8586
|
pricing: {
|
|
9047
8587
|
input: 0.5,
|
|
@@ -9055,7 +8595,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9055
8595
|
id: "essentialai/Rnj-1-Instruct",
|
|
9056
8596
|
name: "Rnj-1 Instruct",
|
|
9057
8597
|
contextWindow: 32768,
|
|
9058
|
-
maxTokens:
|
|
8598
|
+
maxTokens: 1638,
|
|
9059
8599
|
capabilities: ["tools", "temperature"],
|
|
9060
8600
|
pricing: {
|
|
9061
8601
|
input: 0.15,
|
|
@@ -9069,7 +8609,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9069
8609
|
id: "zai-org/GLM-4.6",
|
|
9070
8610
|
name: "GLM 4.6",
|
|
9071
8611
|
contextWindow: 200000,
|
|
9072
|
-
maxTokens:
|
|
8612
|
+
maxTokens: 10000,
|
|
9073
8613
|
capabilities: ["tools", "temperature"],
|
|
9074
8614
|
pricing: {
|
|
9075
8615
|
input: 0.6,
|
|
@@ -9083,7 +8623,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9083
8623
|
id: "deepseek-ai/DeepSeek-V3-1",
|
|
9084
8624
|
name: "DeepSeek V3.1",
|
|
9085
8625
|
contextWindow: 131072,
|
|
9086
|
-
maxTokens:
|
|
8626
|
+
maxTokens: 6553,
|
|
9087
8627
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9088
8628
|
pricing: {
|
|
9089
8629
|
input: 0.6,
|
|
@@ -9097,7 +8637,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9097
8637
|
id: "openai/gpt-oss-120b",
|
|
9098
8638
|
name: "GPT OSS 120B",
|
|
9099
8639
|
contextWindow: 131072,
|
|
9100
|
-
maxTokens:
|
|
8640
|
+
maxTokens: 6553,
|
|
9101
8641
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9102
8642
|
pricing: {
|
|
9103
8643
|
input: 0.15,
|
|
@@ -9111,7 +8651,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9111
8651
|
id: "Qwen/Qwen3-235B-A22B-Instruct-2507-tput",
|
|
9112
8652
|
name: "Qwen3 235B A22B Instruct 2507 FP8",
|
|
9113
8653
|
contextWindow: 262144,
|
|
9114
|
-
maxTokens:
|
|
8654
|
+
maxTokens: 13107,
|
|
9115
8655
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9116
8656
|
pricing: {
|
|
9117
8657
|
input: 0.2,
|
|
@@ -9125,7 +8665,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9125
8665
|
id: "Qwen/Qwen3-Next-80B-A3B-Instruct",
|
|
9126
8666
|
name: "Qwen3-Next-80B-A3B-Instruct",
|
|
9127
8667
|
contextWindow: 262144,
|
|
9128
|
-
maxTokens:
|
|
8668
|
+
maxTokens: 13107,
|
|
9129
8669
|
capabilities: ["tools", "temperature"],
|
|
9130
8670
|
pricing: {
|
|
9131
8671
|
input: 0.15,
|
|
@@ -9139,7 +8679,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9139
8679
|
id: "zai-org/GLM-4.7",
|
|
9140
8680
|
name: "GLM-4.7",
|
|
9141
8681
|
contextWindow: 200000,
|
|
9142
|
-
maxTokens:
|
|
8682
|
+
maxTokens: 10000,
|
|
9143
8683
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9144
8684
|
pricing: {
|
|
9145
8685
|
input: 0.45,
|
|
@@ -9153,7 +8693,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9153
8693
|
id: "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8",
|
|
9154
8694
|
name: "Qwen3 Coder 480B A35B Instruct",
|
|
9155
8695
|
contextWindow: 262144,
|
|
9156
|
-
maxTokens:
|
|
8696
|
+
maxTokens: 13107,
|
|
9157
8697
|
capabilities: ["tools", "temperature"],
|
|
9158
8698
|
pricing: {
|
|
9159
8699
|
input: 2,
|
|
@@ -9167,7 +8707,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9167
8707
|
id: "moonshotai/Kimi-K2-Instruct",
|
|
9168
8708
|
name: "Kimi K2 Instruct",
|
|
9169
8709
|
contextWindow: 131072,
|
|
9170
|
-
maxTokens:
|
|
8710
|
+
maxTokens: 6553,
|
|
9171
8711
|
capabilities: ["tools", "temperature"],
|
|
9172
8712
|
pricing: {
|
|
9173
8713
|
input: 1,
|
|
@@ -9181,7 +8721,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9181
8721
|
id: "deepseek-ai/DeepSeek-V3",
|
|
9182
8722
|
name: "DeepSeek V3",
|
|
9183
8723
|
contextWindow: 131072,
|
|
9184
|
-
maxTokens:
|
|
8724
|
+
maxTokens: 6553,
|
|
9185
8725
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9186
8726
|
pricing: {
|
|
9187
8727
|
input: 1.25,
|
|
@@ -9195,7 +8735,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9195
8735
|
id: "meta-llama/Llama-3.3-70B-Instruct-Turbo",
|
|
9196
8736
|
name: "Llama 3.3 70B",
|
|
9197
8737
|
contextWindow: 131072,
|
|
9198
|
-
maxTokens:
|
|
8738
|
+
maxTokens: 6553,
|
|
9199
8739
|
capabilities: ["tools", "temperature"],
|
|
9200
8740
|
pricing: {
|
|
9201
8741
|
input: 0.88,
|
|
@@ -9309,7 +8849,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9309
8849
|
id: "xai/grok-4.20-multi-agent-beta",
|
|
9310
8850
|
name: "Grok 4.20 Multi Agent Beta",
|
|
9311
8851
|
contextWindow: 2000000,
|
|
9312
|
-
maxTokens:
|
|
8852
|
+
maxTokens: 100000,
|
|
9313
8853
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9314
8854
|
pricing: {
|
|
9315
8855
|
input: 2,
|
|
@@ -9323,7 +8863,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9323
8863
|
id: "xai/grok-4.20-non-reasoning-beta",
|
|
9324
8864
|
name: "Grok 4.20 Beta Non-Reasoning",
|
|
9325
8865
|
contextWindow: 2000000,
|
|
9326
|
-
maxTokens:
|
|
8866
|
+
maxTokens: 100000,
|
|
9327
8867
|
capabilities: ["images", "files", "tools", "temperature"],
|
|
9328
8868
|
pricing: {
|
|
9329
8869
|
input: 2,
|
|
@@ -9337,7 +8877,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9337
8877
|
id: "xai/grok-4.20-reasoning-beta",
|
|
9338
8878
|
name: "Grok 4.20 Beta Reasoning",
|
|
9339
8879
|
contextWindow: 2000000,
|
|
9340
|
-
maxTokens:
|
|
8880
|
+
maxTokens: 100000,
|
|
9341
8881
|
capabilities: ["images", "files", "tools", "reasoning", "temperature"],
|
|
9342
8882
|
pricing: {
|
|
9343
8883
|
input: 2,
|
|
@@ -9347,6 +8887,48 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9347
8887
|
},
|
|
9348
8888
|
releaseDate: "2026-03-11",
|
|
9349
8889
|
},
|
|
8890
|
+
"xai/grok-4.20-multi-agent": {
|
|
8891
|
+
id: "xai/grok-4.20-multi-agent",
|
|
8892
|
+
name: "Grok 4.20 Multi-Agent",
|
|
8893
|
+
contextWindow: 2000000,
|
|
8894
|
+
maxTokens: 100000,
|
|
8895
|
+
capabilities: ["tools", "reasoning", "temperature"],
|
|
8896
|
+
pricing: {
|
|
8897
|
+
input: 2,
|
|
8898
|
+
output: 6,
|
|
8899
|
+
cacheRead: 0.19999999999999998,
|
|
8900
|
+
cacheWrite: 0,
|
|
8901
|
+
},
|
|
8902
|
+
releaseDate: "2026-03-09",
|
|
8903
|
+
},
|
|
8904
|
+
"xai/grok-4.20-non-reasoning": {
|
|
8905
|
+
id: "xai/grok-4.20-non-reasoning",
|
|
8906
|
+
name: "Grok 4.20 Non-Reasoning",
|
|
8907
|
+
contextWindow: 2000000,
|
|
8908
|
+
maxTokens: 100000,
|
|
8909
|
+
capabilities: ["images", "files", "tools", "temperature"],
|
|
8910
|
+
pricing: {
|
|
8911
|
+
input: 2,
|
|
8912
|
+
output: 6,
|
|
8913
|
+
cacheRead: 0.19999999999999998,
|
|
8914
|
+
cacheWrite: 0,
|
|
8915
|
+
},
|
|
8916
|
+
releaseDate: "2026-03-09",
|
|
8917
|
+
},
|
|
8918
|
+
"xai/grok-4.20-reasoning": {
|
|
8919
|
+
id: "xai/grok-4.20-reasoning",
|
|
8920
|
+
name: "Grok 4.20 Reasoning",
|
|
8921
|
+
contextWindow: 2000000,
|
|
8922
|
+
maxTokens: 100000,
|
|
8923
|
+
capabilities: ["images", "files", "tools", "reasoning", "temperature"],
|
|
8924
|
+
pricing: {
|
|
8925
|
+
input: 2,
|
|
8926
|
+
output: 6,
|
|
8927
|
+
cacheRead: 0.19999999999999998,
|
|
8928
|
+
cacheWrite: 0,
|
|
8929
|
+
},
|
|
8930
|
+
releaseDate: "2026-03-09",
|
|
8931
|
+
},
|
|
9350
8932
|
"openai/gpt-5.4": {
|
|
9351
8933
|
id: "openai/gpt-5.4",
|
|
9352
8934
|
name: "GPT 5.4",
|
|
@@ -9421,7 +9003,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9421
9003
|
id: "inception/mercury-2",
|
|
9422
9004
|
name: "Mercury 2",
|
|
9423
9005
|
contextWindow: 128000,
|
|
9424
|
-
maxTokens:
|
|
9006
|
+
maxTokens: 6400,
|
|
9425
9007
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9426
9008
|
pricing: {
|
|
9427
9009
|
input: 0.25,
|
|
@@ -9547,7 +9129,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9547
9129
|
id: "moonshotai/kimi-k2.5",
|
|
9548
9130
|
name: "Kimi K2.5",
|
|
9549
9131
|
contextWindow: 262144,
|
|
9550
|
-
maxTokens:
|
|
9132
|
+
maxTokens: 13107,
|
|
9551
9133
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
9552
9134
|
pricing: {
|
|
9553
9135
|
input: 0.6,
|
|
@@ -9603,7 +9185,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9603
9185
|
id: "mistral/devstral-2",
|
|
9604
9186
|
name: "Devstral 2",
|
|
9605
9187
|
contextWindow: 256000,
|
|
9606
|
-
maxTokens:
|
|
9188
|
+
maxTokens: 12800,
|
|
9607
9189
|
capabilities: ["tools", "temperature"],
|
|
9608
9190
|
pricing: {
|
|
9609
9191
|
input: 0,
|
|
@@ -9645,7 +9227,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9645
9227
|
id: "prime-intellect/intellect-3",
|
|
9646
9228
|
name: "INTELLECT 3",
|
|
9647
9229
|
contextWindow: 131072,
|
|
9648
|
-
maxTokens:
|
|
9230
|
+
maxTokens: 6553,
|
|
9649
9231
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9650
9232
|
pricing: {
|
|
9651
9233
|
input: 0.2,
|
|
@@ -9687,7 +9269,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9687
9269
|
id: "moonshotai/kimi-k2-thinking",
|
|
9688
9270
|
name: "Kimi K2 Thinking",
|
|
9689
9271
|
contextWindow: 216144,
|
|
9690
|
-
maxTokens:
|
|
9272
|
+
maxTokens: 10807,
|
|
9691
9273
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9692
9274
|
pricing: {
|
|
9693
9275
|
input: 0.47,
|
|
@@ -9701,7 +9283,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9701
9283
|
id: "moonshotai/kimi-k2-thinking-turbo",
|
|
9702
9284
|
name: "Kimi K2 Thinking Turbo",
|
|
9703
9285
|
contextWindow: 262114,
|
|
9704
|
-
maxTokens:
|
|
9286
|
+
maxTokens: 13105,
|
|
9705
9287
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9706
9288
|
pricing: {
|
|
9707
9289
|
input: 1.15,
|
|
@@ -9715,7 +9297,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9715
9297
|
id: "minimax/minimax-m2",
|
|
9716
9298
|
name: "MiniMax M2",
|
|
9717
9299
|
contextWindow: 262114,
|
|
9718
|
-
maxTokens:
|
|
9300
|
+
maxTokens: 13105,
|
|
9719
9301
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9720
9302
|
pricing: {
|
|
9721
9303
|
input: 0.27,
|
|
@@ -9841,7 +9423,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
9841
9423
|
id: "deepseek/deepseek-v3.2-exp",
|
|
9842
9424
|
name: "DeepSeek V3.2 Exp",
|
|
9843
9425
|
contextWindow: 163840,
|
|
9844
|
-
maxTokens:
|
|
9426
|
+
maxTokens: 8192,
|
|
9845
9427
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
9846
9428
|
pricing: {
|
|
9847
9429
|
input: 0.27,
|
|
@@ -10093,7 +9675,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10093
9675
|
id: "nvidia/nemotron-nano-9b-v2",
|
|
10094
9676
|
name: "Nvidia Nemotron Nano 9B V2",
|
|
10095
9677
|
contextWindow: 131072,
|
|
10096
|
-
maxTokens:
|
|
9678
|
+
maxTokens: 6553,
|
|
10097
9679
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
10098
9680
|
pricing: {
|
|
10099
9681
|
input: 0.04,
|
|
@@ -10107,7 +9689,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10107
9689
|
id: "zai/glm-4.5v",
|
|
10108
9690
|
name: "GLM 4.5V",
|
|
10109
9691
|
contextWindow: 66000,
|
|
10110
|
-
maxTokens:
|
|
9692
|
+
maxTokens: 3300,
|
|
10111
9693
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
10112
9694
|
pricing: {
|
|
10113
9695
|
input: 0.6,
|
|
@@ -10289,7 +9871,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10289
9871
|
id: "openai/gpt-oss-120b",
|
|
10290
9872
|
name: "GPT OSS 120B",
|
|
10291
9873
|
contextWindow: 131072,
|
|
10292
|
-
maxTokens:
|
|
9874
|
+
maxTokens: 6553,
|
|
10293
9875
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
10294
9876
|
pricing: {
|
|
10295
9877
|
input: 0.1,
|
|
@@ -10317,7 +9899,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10317
9899
|
id: "zai/glm-4.5",
|
|
10318
9900
|
name: "GLM 4.5",
|
|
10319
9901
|
contextWindow: 131072,
|
|
10320
|
-
maxTokens:
|
|
9902
|
+
maxTokens: 6553,
|
|
10321
9903
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
10322
9904
|
pricing: {
|
|
10323
9905
|
input: 0.6,
|
|
@@ -10345,7 +9927,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10345
9927
|
id: "alibaba/qwen3-coder-plus",
|
|
10346
9928
|
name: "Qwen3 Coder Plus",
|
|
10347
9929
|
contextWindow: 1000000,
|
|
10348
|
-
maxTokens:
|
|
9930
|
+
maxTokens: 50000,
|
|
10349
9931
|
capabilities: ["tools", "temperature"],
|
|
10350
9932
|
pricing: {
|
|
10351
9933
|
input: 1,
|
|
@@ -10359,7 +9941,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10359
9941
|
id: "alibaba/qwen3-coder-next",
|
|
10360
9942
|
name: "Qwen3 Coder Next",
|
|
10361
9943
|
contextWindow: 256000,
|
|
10362
|
-
maxTokens:
|
|
9944
|
+
maxTokens: 12800,
|
|
10363
9945
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
10364
9946
|
pricing: {
|
|
10365
9947
|
input: 0.5,
|
|
@@ -10555,7 +10137,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10555
10137
|
id: "mistral/devstral-small-2",
|
|
10556
10138
|
name: "Devstral Small 2",
|
|
10557
10139
|
contextWindow: 256000,
|
|
10558
|
-
maxTokens:
|
|
10140
|
+
maxTokens: 12800,
|
|
10559
10141
|
capabilities: ["tools", "temperature"],
|
|
10560
10142
|
pricing: {
|
|
10561
10143
|
input: 0,
|
|
@@ -10751,7 +10333,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10751
10333
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
10752
10334
|
name: "Qwen3 235B A22B Thinking 2507",
|
|
10753
10335
|
contextWindow: 262114,
|
|
10754
|
-
maxTokens:
|
|
10336
|
+
maxTokens: 13105,
|
|
10755
10337
|
capabilities: ["images", "files", "tools", "reasoning", "temperature"],
|
|
10756
10338
|
pricing: {
|
|
10757
10339
|
input: 0.3,
|
|
@@ -10849,7 +10431,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
10849
10431
|
id: "mistral/magistral-small",
|
|
10850
10432
|
name: "Magistral Small",
|
|
10851
10433
|
contextWindow: 128000,
|
|
10852
|
-
maxTokens:
|
|
10434
|
+
maxTokens: 6400,
|
|
10853
10435
|
capabilities: ["tools", "reasoning", "temperature"],
|
|
10854
10436
|
pricing: {
|
|
10855
10437
|
input: 0.5,
|
|
@@ -11017,7 +10599,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11017
10599
|
id: "arcee-ai/trinity-large-preview",
|
|
11018
10600
|
name: "Trinity Large Preview",
|
|
11019
10601
|
contextWindow: 131000,
|
|
11020
|
-
maxTokens:
|
|
10602
|
+
maxTokens: 6550,
|
|
11021
10603
|
capabilities: ["tools", "temperature"],
|
|
11022
10604
|
pricing: {
|
|
11023
10605
|
input: 0.25,
|
|
@@ -11183,7 +10765,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11183
10765
|
id: "nvidia/nemotron-nano-12b-v2-vl",
|
|
11184
10766
|
name: "Nvidia Nemotron Nano 12B V2 VL",
|
|
11185
10767
|
contextWindow: 131072,
|
|
11186
|
-
maxTokens:
|
|
10768
|
+
maxTokens: 6553,
|
|
11187
10769
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|
|
11188
10770
|
pricing: {
|
|
11189
10771
|
input: 0.2,
|
|
@@ -11211,7 +10793,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11211
10793
|
id: "mistral/pixtral-large",
|
|
11212
10794
|
name: "Pixtral Large (latest)",
|
|
11213
10795
|
contextWindow: 128000,
|
|
11214
|
-
maxTokens:
|
|
10796
|
+
maxTokens: 6400,
|
|
11215
10797
|
capabilities: ["images", "tools", "temperature"],
|
|
11216
10798
|
pricing: {
|
|
11217
10799
|
input: 2,
|
|
@@ -11253,7 +10835,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11253
10835
|
id: "mistral/ministral-3b",
|
|
11254
10836
|
name: "Ministral 3B (latest)",
|
|
11255
10837
|
contextWindow: 128000,
|
|
11256
|
-
maxTokens:
|
|
10838
|
+
maxTokens: 6400,
|
|
11257
10839
|
capabilities: ["tools", "temperature"],
|
|
11258
10840
|
pricing: {
|
|
11259
10841
|
input: 0.04,
|
|
@@ -11267,7 +10849,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11267
10849
|
id: "mistral/ministral-8b",
|
|
11268
10850
|
name: "Ministral 8B (latest)",
|
|
11269
10851
|
contextWindow: 128000,
|
|
11270
|
-
maxTokens:
|
|
10852
|
+
maxTokens: 6400,
|
|
11271
10853
|
capabilities: ["tools", "temperature"],
|
|
11272
10854
|
pricing: {
|
|
11273
10855
|
input: 0.1,
|
|
@@ -11323,7 +10905,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11323
10905
|
id: "mistral/pixtral-12b",
|
|
11324
10906
|
name: "Pixtral 12B",
|
|
11325
10907
|
contextWindow: 128000,
|
|
11326
|
-
maxTokens:
|
|
10908
|
+
maxTokens: 6400,
|
|
11327
10909
|
capabilities: ["images", "tools", "temperature"],
|
|
11328
10910
|
pricing: {
|
|
11329
10911
|
input: 0.15,
|
|
@@ -11463,7 +11045,7 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11463
11045
|
id: "mistral/mixtral-8x22b-instruct",
|
|
11464
11046
|
name: "Mixtral 8x22B",
|
|
11465
11047
|
contextWindow: 64000,
|
|
11466
|
-
maxTokens:
|
|
11048
|
+
maxTokens: 3200,
|
|
11467
11049
|
capabilities: ["tools", "temperature"],
|
|
11468
11050
|
pricing: {
|
|
11469
11051
|
input: 2,
|
|
@@ -11673,9 +11255,9 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11673
11255
|
},
|
|
11674
11256
|
},
|
|
11675
11257
|
xai: {
|
|
11676
|
-
"grok-4.20-
|
|
11677
|
-
id: "grok-4.20-
|
|
11678
|
-
name: "Grok 4.20
|
|
11258
|
+
"grok-4.20-0309-non-reasoning": {
|
|
11259
|
+
id: "grok-4.20-0309-non-reasoning",
|
|
11260
|
+
name: "Grok 4.20 (Non-Reasoning)",
|
|
11679
11261
|
contextWindow: 2000000,
|
|
11680
11262
|
maxTokens: 30000,
|
|
11681
11263
|
capabilities: ["images", "tools", "temperature"],
|
|
@@ -11687,9 +11269,9 @@ export const GENERATED_PROVIDER_MODELS: {
|
|
|
11687
11269
|
},
|
|
11688
11270
|
releaseDate: "2026-03-09",
|
|
11689
11271
|
},
|
|
11690
|
-
"grok-4.20-
|
|
11691
|
-
id: "grok-4.20-
|
|
11692
|
-
name: "Grok 4.20
|
|
11272
|
+
"grok-4.20-0309-reasoning": {
|
|
11273
|
+
id: "grok-4.20-0309-reasoning",
|
|
11274
|
+
name: "Grok 4.20 (Reasoning)",
|
|
11693
11275
|
contextWindow: 2000000,
|
|
11694
11276
|
maxTokens: 30000,
|
|
11695
11277
|
capabilities: ["images", "tools", "reasoning", "temperature"],
|