@clinebot/llms 0.0.10 → 0.0.12

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -14,7 +14,7 @@ export const GENERATED_PROVIDER_MODELS: {
14
14
  version: number;
15
15
  providers: Record<string, Record<string, ModelInfo>>;
16
16
  } = {
17
- version: 1774031158558,
17
+ version: 1774365323810,
18
18
  providers: {
19
19
  aihubmix: {
20
20
  "claude-sonnet-4-6": {
@@ -147,7 +147,7 @@ export const GENERATED_PROVIDER_MODELS: {
147
147
  id: "kimi-k2.5",
148
148
  name: "Kimi K2.5",
149
149
  contextWindow: 262144,
150
- maxTokens: 262144,
150
+ maxTokens: 13107,
151
151
  capabilities: ["images", "tools", "reasoning", "temperature"],
152
152
  pricing: {
153
153
  input: 0.6,
@@ -545,7 +545,7 @@ export const GENERATED_PROVIDER_MODELS: {
545
545
  id: "Kimi-K2-0905",
546
546
  name: "Kimi K2 0905",
547
547
  contextWindow: 262144,
548
- maxTokens: 262144,
548
+ maxTokens: 13107,
549
549
  capabilities: ["tools", "temperature"],
550
550
  pricing: {
551
551
  input: 0.55,
@@ -587,7 +587,7 @@ export const GENERATED_PROVIDER_MODELS: {
587
587
  id: "qwen3-235b-a22b-instruct-2507",
588
588
  name: "Qwen3 235B A22B Instruct 2507",
589
589
  contextWindow: 262144,
590
- maxTokens: 262144,
590
+ maxTokens: 13107,
591
591
  capabilities: ["tools", "temperature"],
592
592
  pricing: {
593
593
  input: 0.28,
@@ -601,7 +601,7 @@ export const GENERATED_PROVIDER_MODELS: {
601
601
  id: "qwen3-235b-a22b-thinking-2507",
602
602
  name: "Qwen3 235B A22B Thinking 2507",
603
603
  contextWindow: 262144,
604
- maxTokens: 262144,
604
+ maxTokens: 13107,
605
605
  capabilities: ["tools", "reasoning", "temperature"],
606
606
  pricing: {
607
607
  input: 0.28,
@@ -993,8 +993,8 @@ export const GENERATED_PROVIDER_MODELS: {
993
993
  },
994
994
  },
995
995
  baseten: {
996
- "nvidia/Nemotron-3-Super": {
997
- id: "nvidia/Nemotron-3-Super",
996
+ "nvidia/Nemotron-120B-A12B": {
997
+ id: "nvidia/Nemotron-120B-A12B",
998
998
  name: "Nemotron 3 Super",
999
999
  contextWindow: 262144,
1000
1000
  maxTokens: 32678,
@@ -1011,7 +1011,7 @@ export const GENERATED_PROVIDER_MODELS: {
1011
1011
  id: "MiniMaxAI/MiniMax-M2.5",
1012
1012
  name: "MiniMax-M2.5",
1013
1013
  contextWindow: 204000,
1014
- maxTokens: 204000,
1014
+ maxTokens: 10200,
1015
1015
  capabilities: ["tools", "reasoning", "temperature"],
1016
1016
  pricing: {
1017
1017
  input: 0.3,
@@ -1067,7 +1067,7 @@ export const GENERATED_PROVIDER_MODELS: {
1067
1067
  id: "zai-org/GLM-4.6",
1068
1068
  name: "GLM 4.6",
1069
1069
  contextWindow: 200000,
1070
- maxTokens: 200000,
1070
+ maxTokens: 10000,
1071
1071
  capabilities: ["tools", "temperature"],
1072
1072
  pricing: {
1073
1073
  input: 0.6,
@@ -1095,7 +1095,7 @@ export const GENERATED_PROVIDER_MODELS: {
1095
1095
  id: "openai/gpt-oss-120b",
1096
1096
  name: "GPT OSS 120B",
1097
1097
  contextWindow: 128000,
1098
- maxTokens: 128000,
1098
+ maxTokens: 6400,
1099
1099
  capabilities: ["tools", "reasoning", "temperature"],
1100
1100
  pricing: {
1101
1101
  input: 0.1,
@@ -1191,6 +1191,34 @@ export const GENERATED_PROVIDER_MODELS: {
1191
1191
  },
1192
1192
  releaseDate: "2026-02-17",
1193
1193
  },
1194
+ "minimax.minimax-m2.5": {
1195
+ id: "minimax.minimax-m2.5",
1196
+ name: "MiniMax M2.5",
1197
+ contextWindow: 1000000,
1198
+ maxTokens: 131072,
1199
+ capabilities: ["tools", "reasoning", "temperature"],
1200
+ pricing: {
1201
+ input: 0.3,
1202
+ output: 1.2,
1203
+ cacheRead: 0,
1204
+ cacheWrite: 0,
1205
+ },
1206
+ releaseDate: "2026-02-12",
1207
+ },
1208
+ "zai.glm-5": {
1209
+ id: "zai.glm-5",
1210
+ name: "GLM-5",
1211
+ contextWindow: 200000,
1212
+ maxTokens: 131072,
1213
+ capabilities: ["tools", "reasoning", "temperature"],
1214
+ pricing: {
1215
+ input: 1,
1216
+ output: 3.2,
1217
+ cacheRead: 0,
1218
+ cacheWrite: 0,
1219
+ },
1220
+ releaseDate: "2026-02-11",
1221
+ },
1194
1222
  "deepseek.v3.2": {
1195
1223
  id: "deepseek.v3.2",
1196
1224
  name: "DeepSeek-V3.2",
@@ -1209,7 +1237,7 @@ export const GENERATED_PROVIDER_MODELS: {
1209
1237
  id: "moonshotai.kimi-k2.5",
1210
1238
  name: "Kimi K2.5",
1211
1239
  contextWindow: 256000,
1212
- maxTokens: 256000,
1240
+ maxTokens: 12800,
1213
1241
  capabilities: ["images", "tools", "reasoning", "temperature"],
1214
1242
  pricing: {
1215
1243
  input: 0.6,
@@ -1377,7 +1405,7 @@ export const GENERATED_PROVIDER_MODELS: {
1377
1405
  id: "moonshot.kimi-k2-thinking",
1378
1406
  name: "Kimi K2 Thinking",
1379
1407
  contextWindow: 256000,
1380
- maxTokens: 256000,
1408
+ maxTokens: 12800,
1381
1409
  capabilities: ["tools", "reasoning", "temperature"],
1382
1410
  pricing: {
1383
1411
  input: 0.6,
@@ -1517,7 +1545,7 @@ export const GENERATED_PROVIDER_MODELS: {
1517
1545
  id: "qwen.qwen3-vl-235b-a22b",
1518
1546
  name: "Qwen/Qwen3-VL-235B-A22B-Instruct",
1519
1547
  contextWindow: 262000,
1520
- maxTokens: 262000,
1548
+ maxTokens: 13100,
1521
1549
  capabilities: ["images", "tools", "structured_output", "temperature"],
1522
1550
  pricing: {
1523
1551
  input: 0.3,
@@ -1615,7 +1643,7 @@ export const GENERATED_PROVIDER_MODELS: {
1615
1643
  id: "qwen.qwen3-32b-v1:0",
1616
1644
  name: "Qwen3 32B (dense)",
1617
1645
  contextWindow: 16384,
1618
- maxTokens: 16384,
1646
+ maxTokens: 819,
1619
1647
  capabilities: ["tools", "reasoning", "temperature"],
1620
1648
  pricing: {
1621
1649
  input: 0.15,
@@ -1657,7 +1685,7 @@ export const GENERATED_PROVIDER_MODELS: {
1657
1685
  id: "qwen.qwen3-next-80b-a3b",
1658
1686
  name: "Qwen/Qwen3-Next-80B-A3B-Instruct",
1659
1687
  contextWindow: 262000,
1660
- maxTokens: 262000,
1688
+ maxTokens: 13100,
1661
1689
  capabilities: ["tools", "structured_output", "temperature"],
1662
1690
  pricing: {
1663
1691
  input: 0.14,
@@ -2377,7 +2405,7 @@ export const GENERATED_PROVIDER_MODELS: {
2377
2405
  id: "accounts/fireworks/models/minimax-m2p5",
2378
2406
  name: "MiniMax-M2.5",
2379
2407
  contextWindow: 196608,
2380
- maxTokens: 196608,
2408
+ maxTokens: 9830,
2381
2409
  capabilities: ["tools", "reasoning", "temperature"],
2382
2410
  pricing: {
2383
2411
  input: 0.3,
@@ -2405,7 +2433,7 @@ export const GENERATED_PROVIDER_MODELS: {
2405
2433
  id: "accounts/fireworks/models/kimi-k2p5",
2406
2434
  name: "Kimi K2.5",
2407
2435
  contextWindow: 256000,
2408
- maxTokens: 256000,
2436
+ maxTokens: 12800,
2409
2437
  capabilities: ["images", "tools", "reasoning", "temperature"],
2410
2438
  pricing: {
2411
2439
  input: 0.6,
@@ -2415,11 +2443,25 @@ export const GENERATED_PROVIDER_MODELS: {
2415
2443
  },
2416
2444
  releaseDate: "2026-01-27",
2417
2445
  },
2446
+ "accounts/fireworks/routers/kimi-k2p5-turbo": {
2447
+ id: "accounts/fireworks/routers/kimi-k2p5-turbo",
2448
+ name: "Kimi K2.5 Turbo",
2449
+ contextWindow: 256000,
2450
+ maxTokens: 12800,
2451
+ capabilities: ["images", "tools", "reasoning", "temperature"],
2452
+ pricing: {
2453
+ input: 0,
2454
+ output: 0,
2455
+ cacheRead: 0,
2456
+ cacheWrite: 0,
2457
+ },
2458
+ releaseDate: "2026-01-27",
2459
+ },
2418
2460
  "accounts/fireworks/models/minimax-m2p1": {
2419
2461
  id: "accounts/fireworks/models/minimax-m2p1",
2420
2462
  name: "MiniMax-M2.1",
2421
2463
  contextWindow: 200000,
2422
- maxTokens: 200000,
2464
+ maxTokens: 10000,
2423
2465
  capabilities: ["tools", "reasoning", "temperature"],
2424
2466
  pricing: {
2425
2467
  input: 0.3,
@@ -2433,7 +2475,7 @@ export const GENERATED_PROVIDER_MODELS: {
2433
2475
  id: "accounts/fireworks/models/glm-4p7",
2434
2476
  name: "GLM 4.7",
2435
2477
  contextWindow: 198000,
2436
- maxTokens: 198000,
2478
+ maxTokens: 9900,
2437
2479
  capabilities: ["tools", "reasoning", "temperature"],
2438
2480
  pricing: {
2439
2481
  input: 0.6,
@@ -2447,7 +2489,7 @@ export const GENERATED_PROVIDER_MODELS: {
2447
2489
  id: "accounts/fireworks/models/deepseek-v3p2",
2448
2490
  name: "DeepSeek V3.2",
2449
2491
  contextWindow: 160000,
2450
- maxTokens: 160000,
2492
+ maxTokens: 8000,
2451
2493
  capabilities: ["tools", "reasoning", "temperature"],
2452
2494
  pricing: {
2453
2495
  input: 0.56,
@@ -2461,7 +2503,7 @@ export const GENERATED_PROVIDER_MODELS: {
2461
2503
  id: "accounts/fireworks/models/kimi-k2-thinking",
2462
2504
  name: "Kimi K2 Thinking",
2463
2505
  contextWindow: 256000,
2464
- maxTokens: 256000,
2506
+ maxTokens: 12800,
2465
2507
  capabilities: ["tools", "reasoning", "temperature"],
2466
2508
  pricing: {
2467
2509
  input: 0.6,
@@ -2475,7 +2517,7 @@ export const GENERATED_PROVIDER_MODELS: {
2475
2517
  id: "accounts/fireworks/models/deepseek-v3p1",
2476
2518
  name: "DeepSeek V3.1",
2477
2519
  contextWindow: 163840,
2478
- maxTokens: 163840,
2520
+ maxTokens: 8192,
2479
2521
  capabilities: ["tools", "reasoning", "temperature"],
2480
2522
  pricing: {
2481
2523
  input: 0.56,
@@ -2517,7 +2559,7 @@ export const GENERATED_PROVIDER_MODELS: {
2517
2559
  id: "accounts/fireworks/models/glm-4p5-air",
2518
2560
  name: "GLM 4.5 Air",
2519
2561
  contextWindow: 131072,
2520
- maxTokens: 131072,
2562
+ maxTokens: 6553,
2521
2563
  capabilities: ["tools", "reasoning", "temperature"],
2522
2564
  pricing: {
2523
2565
  input: 0.22,
@@ -2531,7 +2573,7 @@ export const GENERATED_PROVIDER_MODELS: {
2531
2573
  id: "accounts/fireworks/models/glm-4p5",
2532
2574
  name: "GLM 4.5",
2533
2575
  contextWindow: 131072,
2534
- maxTokens: 131072,
2576
+ maxTokens: 6553,
2535
2577
  capabilities: ["tools", "reasoning", "temperature"],
2536
2578
  pricing: {
2537
2579
  input: 0.55,
@@ -3124,7 +3166,7 @@ export const GENERATED_PROVIDER_MODELS: {
3124
3166
  id: "llama-3.1-8b-instant",
3125
3167
  name: "Llama 3.1 8B Instant",
3126
3168
  contextWindow: 131072,
3127
- maxTokens: 131072,
3169
+ maxTokens: 6553,
3128
3170
  capabilities: ["tools", "temperature"],
3129
3171
  pricing: {
3130
3172
  input: 0.05,
@@ -3196,7 +3238,7 @@ export const GENERATED_PROVIDER_MODELS: {
3196
3238
  id: "moonshotai/Kimi-K2.5",
3197
3239
  name: "Kimi-K2.5",
3198
3240
  contextWindow: 262144,
3199
- maxTokens: 262144,
3241
+ maxTokens: 13107,
3200
3242
  capabilities: ["images", "tools", "reasoning", "temperature"],
3201
3243
  pricing: {
3202
3244
  input: 0.6,
@@ -3266,7 +3308,7 @@ export const GENERATED_PROVIDER_MODELS: {
3266
3308
  id: "moonshotai/Kimi-K2-Thinking",
3267
3309
  name: "Kimi-K2-Thinking",
3268
3310
  contextWindow: 262144,
3269
- maxTokens: 262144,
3311
+ maxTokens: 13107,
3270
3312
  capabilities: ["tools", "reasoning", "temperature"],
3271
3313
  pricing: {
3272
3314
  input: 0.6,
@@ -3378,7 +3420,7 @@ export const GENERATED_PROVIDER_MODELS: {
3378
3420
  id: "deepseek-ai/DeepSeek-R1-0528",
3379
3421
  name: "DeepSeek-R1-0528",
3380
3422
  contextWindow: 163840,
3381
- maxTokens: 163840,
3423
+ maxTokens: 8192,
3382
3424
  capabilities: ["tools", "reasoning", "temperature"],
3383
3425
  pricing: {
3384
3426
  input: 3,
@@ -3524,7 +3566,7 @@ export const GENERATED_PROVIDER_MODELS: {
3524
3566
  id: "kimi-k2.5",
3525
3567
  name: "Kimi K2.5",
3526
3568
  contextWindow: 262144,
3527
- maxTokens: 262144,
3569
+ maxTokens: 13107,
3528
3570
  capabilities: ["images", "tools", "reasoning", "structured_output"],
3529
3571
  pricing: {
3530
3572
  input: 0.6,
@@ -3538,7 +3580,7 @@ export const GENERATED_PROVIDER_MODELS: {
3538
3580
  id: "kimi-k2-thinking",
3539
3581
  name: "Kimi K2 Thinking",
3540
3582
  contextWindow: 262144,
3541
- maxTokens: 262144,
3583
+ maxTokens: 13107,
3542
3584
  capabilities: ["tools", "reasoning", "temperature"],
3543
3585
  pricing: {
3544
3586
  input: 0.6,
@@ -3552,7 +3594,7 @@ export const GENERATED_PROVIDER_MODELS: {
3552
3594
  id: "kimi-k2-thinking-turbo",
3553
3595
  name: "Kimi K2 Thinking Turbo",
3554
3596
  contextWindow: 262144,
3555
- maxTokens: 262144,
3597
+ maxTokens: 13107,
3556
3598
  capabilities: ["tools", "reasoning", "temperature"],
3557
3599
  pricing: {
3558
3600
  input: 1.15,
@@ -3566,7 +3608,7 @@ export const GENERATED_PROVIDER_MODELS: {
3566
3608
  id: "kimi-k2-0905-preview",
3567
3609
  name: "Kimi K2 0905",
3568
3610
  contextWindow: 262144,
3569
- maxTokens: 262144,
3611
+ maxTokens: 13107,
3570
3612
  capabilities: ["tools", "temperature"],
3571
3613
  pricing: {
3572
3614
  input: 0.6,
@@ -3580,7 +3622,7 @@ export const GENERATED_PROVIDER_MODELS: {
3580
3622
  id: "kimi-k2-turbo-preview",
3581
3623
  name: "Kimi K2 Turbo",
3582
3624
  contextWindow: 262144,
3583
- maxTokens: 262144,
3625
+ maxTokens: 13107,
3584
3626
  capabilities: ["tools", "temperature"],
3585
3627
  pricing: {
3586
3628
  input: 2.4,
@@ -4323,7 +4365,7 @@ export const GENERATED_PROVIDER_MODELS: {
4323
4365
  id: "kimi-k2.5",
4324
4366
  name: "kimi-k2.5",
4325
4367
  contextWindow: 262144,
4326
- maxTokens: 262144,
4368
+ maxTokens: 13107,
4327
4369
  capabilities: ["images", "tools", "reasoning"],
4328
4370
  pricing: {
4329
4371
  input: 0,
@@ -4393,7 +4435,7 @@ export const GENERATED_PROVIDER_MODELS: {
4393
4435
  id: "devstral-2:123b",
4394
4436
  name: "devstral-2:123b",
4395
4437
  contextWindow: 262144,
4396
- maxTokens: 262144,
4438
+ maxTokens: 13107,
4397
4439
  capabilities: ["tools"],
4398
4440
  pricing: {
4399
4441
  input: 0,
@@ -4407,7 +4449,7 @@ export const GENERATED_PROVIDER_MODELS: {
4407
4449
  id: "devstral-small-2:24b",
4408
4450
  name: "devstral-small-2:24b",
4409
4451
  contextWindow: 262144,
4410
- maxTokens: 262144,
4452
+ maxTokens: 13107,
4411
4453
  capabilities: ["images", "tools"],
4412
4454
  pricing: {
4413
4455
  input: 0,
@@ -4435,7 +4477,7 @@ export const GENERATED_PROVIDER_MODELS: {
4435
4477
  id: "mistral-large-3:675b",
4436
4478
  name: "mistral-large-3:675b",
4437
4479
  contextWindow: 262144,
4438
- maxTokens: 262144,
4480
+ maxTokens: 13107,
4439
4481
  capabilities: ["images", "tools"],
4440
4482
  pricing: {
4441
4483
  input: 0,
@@ -4463,7 +4505,7 @@ export const GENERATED_PROVIDER_MODELS: {
4463
4505
  id: "kimi-k2-thinking",
4464
4506
  name: "kimi-k2-thinking",
4465
4507
  contextWindow: 262144,
4466
- maxTokens: 262144,
4508
+ maxTokens: 13107,
4467
4509
  capabilities: ["tools", "reasoning"],
4468
4510
  pricing: {
4469
4511
  input: 0,
@@ -4547,7 +4589,7 @@ export const GENERATED_PROVIDER_MODELS: {
4547
4589
  id: "deepseek-v3.1:671b",
4548
4590
  name: "deepseek-v3.1:671b",
4549
4591
  contextWindow: 163840,
4550
- maxTokens: 163840,
4592
+ maxTokens: 8192,
4551
4593
  capabilities: ["tools", "reasoning"],
4552
4594
  pricing: {
4553
4595
  input: 0,
@@ -4603,7 +4645,7 @@ export const GENERATED_PROVIDER_MODELS: {
4603
4645
  id: "kimi-k2:1t",
4604
4646
  name: "kimi-k2:1t",
4605
4647
  contextWindow: 262144,
4606
- maxTokens: 262144,
4648
+ maxTokens: 13107,
4607
4649
  capabilities: ["tools"],
4608
4650
  pricing: {
4609
4651
  input: 0,
@@ -5217,7 +5259,7 @@ export const GENERATED_PROVIDER_MODELS: {
5217
5259
  id: "gpt-4",
5218
5260
  name: "GPT-4",
5219
5261
  contextWindow: 8192,
5220
- maxTokens: 8192,
5262
+ maxTokens: 409,
5221
5263
  capabilities: ["tools", "temperature"],
5222
5264
  pricing: {
5223
5265
  input: 30,
@@ -5351,40 +5393,26 @@ export const GENERATED_PROVIDER_MODELS: {
5351
5393
  },
5352
5394
  releaseDate: "2026-03-12",
5353
5395
  },
5354
- "openrouter/healer-alpha": {
5355
- id: "openrouter/healer-alpha",
5356
- name: "Healer Alpha",
5396
+ "nvidia/nemotron-3-super-120b-a12b": {
5397
+ id: "nvidia/nemotron-3-super-120b-a12b",
5398
+ name: "Nemotron 3 Super",
5357
5399
  contextWindow: 262144,
5358
- maxTokens: 64000,
5359
- capabilities: [
5360
- "images",
5361
- "files",
5362
- "tools",
5363
- "reasoning",
5364
- "structured_output",
5365
- "temperature",
5366
- ],
5400
+ maxTokens: 13107,
5401
+ capabilities: ["tools", "reasoning", "temperature"],
5367
5402
  pricing: {
5368
- input: 0,
5369
- output: 0,
5403
+ input: 0.1,
5404
+ output: 0.5,
5370
5405
  cacheRead: 0,
5371
5406
  cacheWrite: 0,
5372
5407
  },
5373
5408
  releaseDate: "2026-03-11",
5374
5409
  },
5375
- "openrouter/hunter-alpha": {
5376
- id: "openrouter/hunter-alpha",
5377
- name: "Hunter Alpha",
5378
- contextWindow: 1048576,
5379
- maxTokens: 64000,
5380
- capabilities: [
5381
- "images",
5382
- "files",
5383
- "tools",
5384
- "reasoning",
5385
- "structured_output",
5386
- "temperature",
5387
- ],
5410
+ "nvidia/nemotron-3-super-120b-a12b-free": {
5411
+ id: "nvidia/nemotron-3-super-120b-a12b-free",
5412
+ name: "Nemotron 3 Super (free)",
5413
+ contextWindow: 262144,
5414
+ maxTokens: 13107,
5415
+ capabilities: ["tools", "reasoning", "temperature"],
5388
5416
  pricing: {
5389
5417
  input: 0,
5390
5418
  output: 0,
@@ -5627,25 +5655,6 @@ export const GENERATED_PROVIDER_MODELS: {
5627
5655
  },
5628
5656
  releaseDate: "2026-02-12",
5629
5657
  },
5630
- "openrouter/aurora-alpha": {
5631
- id: "openrouter/aurora-alpha",
5632
- name: "Aurora Alpha",
5633
- contextWindow: 128000,
5634
- maxTokens: 50000,
5635
- capabilities: [
5636
- "tools",
5637
- "reasoning",
5638
- "structured_output",
5639
- "temperature",
5640
- ],
5641
- pricing: {
5642
- input: 0,
5643
- output: 0,
5644
- cacheRead: 0,
5645
- cacheWrite: 0,
5646
- },
5647
- releaseDate: "2026-02-09",
5648
- },
5649
5658
  "anthropic/claude-opus-4.6": {
5650
5659
  id: "anthropic/claude-opus-4.6",
5651
5660
  name: "Claude Opus 4.6",
@@ -5691,7 +5700,7 @@ export const GENERATED_PROVIDER_MODELS: {
5691
5700
  id: "stepfun/step-3.5-flash",
5692
5701
  name: "Step 3.5 Flash",
5693
5702
  contextWindow: 256000,
5694
- maxTokens: 256000,
5703
+ maxTokens: 12800,
5695
5704
  capabilities: ["tools", "reasoning", "temperature"],
5696
5705
  pricing: {
5697
5706
  input: 0.1,
@@ -5705,7 +5714,7 @@ export const GENERATED_PROVIDER_MODELS: {
5705
5714
  id: "stepfun/step-3.5-flash:free",
5706
5715
  name: "Step 3.5 Flash (free)",
5707
5716
  contextWindow: 256000,
5708
- maxTokens: 256000,
5717
+ maxTokens: 12800,
5709
5718
  capabilities: ["tools", "reasoning", "temperature"],
5710
5719
  pricing: {
5711
5720
  input: 0,
@@ -5719,7 +5728,7 @@ export const GENERATED_PROVIDER_MODELS: {
5719
5728
  id: "arcee-ai/trinity-large-preview:free",
5720
5729
  name: "Trinity Large Preview",
5721
5730
  contextWindow: 131072,
5722
- maxTokens: 131072,
5731
+ maxTokens: 6553,
5723
5732
  capabilities: ["tools", "structured_output", "temperature"],
5724
5733
  pricing: {
5725
5734
  input: 0,
@@ -5733,7 +5742,7 @@ export const GENERATED_PROVIDER_MODELS: {
5733
5742
  id: "arcee-ai/trinity-mini:free",
5734
5743
  name: "Trinity Mini",
5735
5744
  contextWindow: 131072,
5736
- maxTokens: 131072,
5745
+ maxTokens: 6553,
5737
5746
  capabilities: ["tools", "structured_output", "temperature"],
5738
5747
  pricing: {
5739
5748
  input: 0,
@@ -5747,7 +5756,7 @@ export const GENERATED_PROVIDER_MODELS: {
5747
5756
  id: "moonshotai/kimi-k2.5",
5748
5757
  name: "Kimi K2.5",
5749
5758
  contextWindow: 262144,
5750
- maxTokens: 262144,
5759
+ maxTokens: 13107,
5751
5760
  capabilities: [
5752
5761
  "images",
5753
5762
  "tools",
@@ -5865,7 +5874,7 @@ export const GENERATED_PROVIDER_MODELS: {
5865
5874
  id: "nvidia/nemotron-3-nano-30b-a3b:free",
5866
5875
  name: "Nemotron 3 Nano 30B A3B (free)",
5867
5876
  contextWindow: 256000,
5868
- maxTokens: 256000,
5877
+ maxTokens: 12800,
5869
5878
  capabilities: [
5870
5879
  "tools",
5871
5880
  "reasoning",
@@ -5979,25 +5988,6 @@ export const GENERATED_PROVIDER_MODELS: {
5979
5988
  },
5980
5989
  releaseDate: "2025-12-01",
5981
5990
  },
5982
- "tngtech/tng-r1t-chimera:free": {
5983
- id: "tngtech/tng-r1t-chimera:free",
5984
- name: "R1T Chimera (free)",
5985
- contextWindow: 163840,
5986
- maxTokens: 163840,
5987
- capabilities: [
5988
- "tools",
5989
- "reasoning",
5990
- "structured_output",
5991
- "temperature",
5992
- ],
5993
- pricing: {
5994
- input: 0,
5995
- output: 0,
5996
- cacheRead: 0,
5997
- cacheWrite: 0,
5998
- },
5999
- releaseDate: "2025-11-26",
6000
- },
6001
5991
  "anthropic/claude-opus-4.5": {
6002
5992
  id: "anthropic/claude-opus-4.5",
6003
5993
  name: "Claude Opus 4.5",
@@ -6060,34 +6050,6 @@ export const GENERATED_PROVIDER_MODELS: {
6060
6050
  },
6061
6051
  releaseDate: "2025-11-18",
6062
6052
  },
6063
- "openrouter/sherlock-dash-alpha": {
6064
- id: "openrouter/sherlock-dash-alpha",
6065
- name: "Sherlock Dash Alpha",
6066
- contextWindow: 1840000,
6067
- maxTokens: 0,
6068
- capabilities: ["images", "tools", "temperature"],
6069
- pricing: {
6070
- input: 0,
6071
- output: 0,
6072
- cacheRead: 0,
6073
- cacheWrite: 0,
6074
- },
6075
- releaseDate: "2025-11-15",
6076
- },
6077
- "openrouter/sherlock-think-alpha": {
6078
- id: "openrouter/sherlock-think-alpha",
6079
- name: "Sherlock Think Alpha",
6080
- contextWindow: 1840000,
6081
- maxTokens: 0,
6082
- capabilities: ["images", "tools", "reasoning", "temperature"],
6083
- pricing: {
6084
- input: 0,
6085
- output: 0,
6086
- cacheRead: 0,
6087
- cacheWrite: 0,
6088
- },
6089
- releaseDate: "2025-11-15",
6090
- },
6091
6053
  "openai/gpt-5.1": {
6092
6054
  id: "openai/gpt-5.1",
6093
6055
  name: "GPT-5.1",
@@ -6188,25 +6150,11 @@ export const GENERATED_PROVIDER_MODELS: {
6188
6150
  },
6189
6151
  releaseDate: "2025-11-13",
6190
6152
  },
6191
- "kwaipilot/kat-coder-pro:free": {
6192
- id: "kwaipilot/kat-coder-pro:free",
6193
- name: "Kat Coder Pro (free)",
6194
- contextWindow: 256000,
6195
- maxTokens: 65536,
6196
- capabilities: ["tools", "structured_output", "temperature"],
6197
- pricing: {
6198
- input: 0,
6199
- output: 0,
6200
- cacheRead: 0,
6201
- cacheWrite: 0,
6202
- },
6203
- releaseDate: "2025-11-10",
6204
- },
6205
6153
  "moonshotai/kimi-k2-thinking": {
6206
6154
  id: "moonshotai/kimi-k2-thinking",
6207
6155
  name: "Kimi K2 Thinking",
6208
6156
  contextWindow: 262144,
6209
- maxTokens: 262144,
6157
+ maxTokens: 13107,
6210
6158
  capabilities: [
6211
6159
  "tools",
6212
6160
  "reasoning",
@@ -6239,7 +6187,7 @@ export const GENERATED_PROVIDER_MODELS: {
6239
6187
  id: "nvidia/nemotron-nano-12b-v2-vl:free",
6240
6188
  name: "Nemotron Nano 12B 2 VL (free)",
6241
6189
  contextWindow: 128000,
6242
- maxTokens: 128000,
6190
+ maxTokens: 6400,
6243
6191
  capabilities: ["images", "tools", "reasoning", "temperature"],
6244
6192
  pricing: {
6245
6193
  input: 0,
@@ -6487,7 +6435,7 @@ export const GENERATED_PROVIDER_MODELS: {
6487
6435
  id: "mistralai/devstral-2512",
6488
6436
  name: "Devstral 2 2512",
6489
6437
  contextWindow: 262144,
6490
- maxTokens: 262144,
6438
+ maxTokens: 13107,
6491
6439
  capabilities: ["tools", "structured_output", "temperature"],
6492
6440
  pricing: {
6493
6441
  input: 0.15,
@@ -6497,25 +6445,11 @@ export const GENERATED_PROVIDER_MODELS: {
6497
6445
  },
6498
6446
  releaseDate: "2025-09-12",
6499
6447
  },
6500
- "mistralai/devstral-2512:free": {
6501
- id: "mistralai/devstral-2512:free",
6502
- name: "Devstral 2 2512 (free)",
6503
- contextWindow: 262144,
6504
- maxTokens: 262144,
6505
- capabilities: ["tools", "temperature"],
6506
- pricing: {
6507
- input: 0,
6508
- output: 0,
6509
- cacheRead: 0,
6510
- cacheWrite: 0,
6511
- },
6512
- releaseDate: "2025-09-12",
6513
- },
6514
6448
  "qwen/qwen3-next-80b-a3b-instruct": {
6515
6449
  id: "qwen/qwen3-next-80b-a3b-instruct",
6516
6450
  name: "Qwen3 Next 80B A3B Instruct",
6517
6451
  contextWindow: 262144,
6518
- maxTokens: 262144,
6452
+ maxTokens: 13107,
6519
6453
  capabilities: ["tools", "structured_output", "temperature"],
6520
6454
  pricing: {
6521
6455
  input: 0.14,
@@ -6529,7 +6463,7 @@ export const GENERATED_PROVIDER_MODELS: {
6529
6463
  id: "qwen/qwen3-next-80b-a3b-instruct:free",
6530
6464
  name: "Qwen3 Next 80B A3B Instruct (free)",
6531
6465
  contextWindow: 262144,
6532
- maxTokens: 262144,
6466
+ maxTokens: 13107,
6533
6467
  capabilities: ["tools", "structured_output", "temperature"],
6534
6468
  pricing: {
6535
6469
  input: 0,
@@ -6543,7 +6477,7 @@ export const GENERATED_PROVIDER_MODELS: {
6543
6477
  id: "qwen/qwen3-next-80b-a3b-thinking",
6544
6478
  name: "Qwen3 Next 80B A3B Thinking",
6545
6479
  contextWindow: 262144,
6546
- maxTokens: 262144,
6480
+ maxTokens: 13107,
6547
6481
  capabilities: [
6548
6482
  "tools",
6549
6483
  "reasoning",
@@ -6590,7 +6524,7 @@ export const GENERATED_PROVIDER_MODELS: {
6590
6524
  id: "nvidia/nemotron-nano-9b-v2:free",
6591
6525
  name: "Nemotron Nano 9B V2 (free)",
6592
6526
  contextWindow: 128000,
6593
- maxTokens: 128000,
6527
+ maxTokens: 6400,
6594
6528
  capabilities: [
6595
6529
  "tools",
6596
6530
  "reasoning",
@@ -6642,7 +6576,7 @@ export const GENERATED_PROVIDER_MODELS: {
6642
6576
  id: "nousresearch/hermes-4-405b",
6643
6577
  name: "Hermes 4 405B",
6644
6578
  contextWindow: 131072,
6645
- maxTokens: 131072,
6579
+ maxTokens: 6553,
6646
6580
  capabilities: ["tools", "reasoning", "temperature"],
6647
6581
  pricing: {
6648
6582
  input: 1,
@@ -6656,7 +6590,7 @@ export const GENERATED_PROVIDER_MODELS: {
6656
6590
  id: "nousresearch/hermes-4-70b",
6657
6591
  name: "Hermes 4 70B",
6658
6592
  contextWindow: 131072,
6659
- maxTokens: 131072,
6593
+ maxTokens: 6553,
6660
6594
  capabilities: [
6661
6595
  "tools",
6662
6596
  "reasoning",
@@ -6675,7 +6609,7 @@ export const GENERATED_PROVIDER_MODELS: {
6675
6609
  id: "deepseek/deepseek-chat-v3.1",
6676
6610
  name: "DeepSeek-V3.1",
6677
6611
  contextWindow: 163840,
6678
- maxTokens: 163840,
6612
+ maxTokens: 8192,
6679
6613
  capabilities: [
6680
6614
  "tools",
6681
6615
  "reasoning",
@@ -6714,7 +6648,7 @@ export const GENERATED_PROVIDER_MODELS: {
6714
6648
  id: "nvidia/nemotron-nano-9b-v2",
6715
6649
  name: "nvidia-nemotron-nano-9b-v2",
6716
6650
  contextWindow: 131072,
6717
- maxTokens: 131072,
6651
+ maxTokens: 6553,
6718
6652
  capabilities: ["tools", "reasoning", "temperature"],
6719
6653
  pricing: {
6720
6654
  input: 0.04,
@@ -6728,7 +6662,7 @@ export const GENERATED_PROVIDER_MODELS: {
6728
6662
  id: "mistralai/mistral-medium-3.1",
6729
6663
  name: "Mistral Medium 3.1",
6730
6664
  contextWindow: 262144,
6731
- maxTokens: 262144,
6665
+ maxTokens: 13107,
6732
6666
  capabilities: ["images", "tools", "structured_output", "temperature"],
6733
6667
  pricing: {
6734
6668
  input: 0.4,
@@ -6928,7 +6862,7 @@ export const GENERATED_PROVIDER_MODELS: {
6928
6862
  id: "mistralai/codestral-2508",
6929
6863
  name: "Codestral 2508",
6930
6864
  contextWindow: 256000,
6931
- maxTokens: 256000,
6865
+ maxTokens: 12800,
6932
6866
  capabilities: ["tools", "structured_output", "temperature"],
6933
6867
  pricing: {
6934
6868
  input: 0.3,
@@ -6956,7 +6890,7 @@ export const GENERATED_PROVIDER_MODELS: {
6956
6890
  id: "qwen/qwen3-30b-a3b-instruct-2507",
6957
6891
  name: "Qwen3 30B A3B Instruct 2507",
6958
6892
  contextWindow: 262000,
6959
- maxTokens: 262000,
6893
+ maxTokens: 13100,
6960
6894
  capabilities: ["tools", "structured_output", "temperature"],
6961
6895
  pricing: {
6962
6896
  input: 0.2,
@@ -6970,7 +6904,7 @@ export const GENERATED_PROVIDER_MODELS: {
6970
6904
  id: "qwen/qwen3-30b-a3b-thinking-2507",
6971
6905
  name: "Qwen3 30B A3B Thinking 2507",
6972
6906
  contextWindow: 262000,
6973
- maxTokens: 262000,
6907
+ maxTokens: 13100,
6974
6908
  capabilities: [
6975
6909
  "tools",
6976
6910
  "reasoning",
@@ -7137,7 +7071,7 @@ export const GENERATED_PROVIDER_MODELS: {
7137
7071
  id: "moonshotai/kimi-k2:free",
7138
7072
  name: "Kimi K2 (free)",
7139
7073
  contextWindow: 32800,
7140
- maxTokens: 32800,
7074
+ maxTokens: 1640,
7141
7075
  capabilities: ["tools", "temperature"],
7142
7076
  pricing: {
7143
7077
  input: 0,
@@ -7151,7 +7085,7 @@ export const GENERATED_PROVIDER_MODELS: {
7151
7085
  id: "mistralai/devstral-medium-2507",
7152
7086
  name: "Devstral Medium",
7153
7087
  contextWindow: 131072,
7154
- maxTokens: 131072,
7088
+ maxTokens: 6553,
7155
7089
  capabilities: ["tools", "structured_output", "temperature"],
7156
7090
  pricing: {
7157
7091
  input: 0.4,
@@ -7165,7 +7099,7 @@ export const GENERATED_PROVIDER_MODELS: {
7165
7099
  id: "mistralai/devstral-small-2507",
7166
7100
  name: "Devstral Small 1.1",
7167
7101
  contextWindow: 131072,
7168
- maxTokens: 131072,
7102
+ maxTokens: 6553,
7169
7103
  capabilities: ["tools", "structured_output", "temperature"],
7170
7104
  pricing: {
7171
7105
  input: 0.1,
@@ -7222,20 +7156,6 @@ export const GENERATED_PROVIDER_MODELS: {
7222
7156
  },
7223
7157
  releaseDate: "2025-06-20",
7224
7158
  },
7225
- "mistralai/mistral-small-3.2-24b-instruct:free": {
7226
- id: "mistralai/mistral-small-3.2-24b-instruct:free",
7227
- name: "Mistral Small 3.2 24B (free)",
7228
- contextWindow: 96000,
7229
- maxTokens: 96000,
7230
- capabilities: ["images", "tools", "structured_output", "temperature"],
7231
- pricing: {
7232
- input: 0,
7233
- output: 0,
7234
- cacheRead: 0,
7235
- cacheWrite: 0,
7236
- },
7237
- releaseDate: "2025-06-20",
7238
- },
7239
7159
  "google/gemini-2.5-flash-lite": {
7240
7160
  id: "google/gemini-2.5-flash-lite",
7241
7161
  name: "Gemini 2.5 Flash Lite",
@@ -7271,20 +7191,6 @@ export const GENERATED_PROVIDER_MODELS: {
7271
7191
  },
7272
7192
  releaseDate: "2025-06-17",
7273
7193
  },
7274
- "moonshotai/kimi-dev-72b:free": {
7275
- id: "moonshotai/kimi-dev-72b:free",
7276
- name: "Kimi Dev 72b (free)",
7277
- contextWindow: 131072,
7278
- maxTokens: 131072,
7279
- capabilities: ["tools", "temperature"],
7280
- pricing: {
7281
- input: 0,
7282
- output: 0,
7283
- cacheRead: 0,
7284
- cacheWrite: 0,
7285
- },
7286
- releaseDate: "2025-06-16",
7287
- },
7288
7194
  "google/gemini-2.5-pro-preview-06-05": {
7289
7195
  id: "google/gemini-2.5-pro-preview-06-05",
7290
7196
  name: "Gemini 2.5 Pro Preview 06-05",
@@ -7306,34 +7212,6 @@ export const GENERATED_PROVIDER_MODELS: {
7306
7212
  },
7307
7213
  releaseDate: "2025-06-05",
7308
7214
  },
7309
- "deepseek/deepseek-r1-0528-qwen3-8b:free": {
7310
- id: "deepseek/deepseek-r1-0528-qwen3-8b:free",
7311
- name: "Deepseek R1 0528 Qwen3 8B (free)",
7312
- contextWindow: 131072,
7313
- maxTokens: 131072,
7314
- capabilities: ["tools", "reasoning", "temperature"],
7315
- pricing: {
7316
- input: 0,
7317
- output: 0,
7318
- cacheRead: 0,
7319
- cacheWrite: 0,
7320
- },
7321
- releaseDate: "2025-05-29",
7322
- },
7323
- "sarvamai/sarvam-m:free": {
7324
- id: "sarvamai/sarvam-m:free",
7325
- name: "Sarvam-M (free)",
7326
- contextWindow: 32768,
7327
- maxTokens: 32768,
7328
- capabilities: ["tools", "reasoning", "temperature"],
7329
- pricing: {
7330
- input: 0,
7331
- output: 0,
7332
- cacheRead: 0,
7333
- cacheWrite: 0,
7334
- },
7335
- releaseDate: "2025-05-25",
7336
- },
7337
7215
  "anthropic/claude-opus-4": {
7338
7216
  id: "anthropic/claude-opus-4",
7339
7217
  name: "Claude Opus 4",
@@ -7362,25 +7240,11 @@ export const GENERATED_PROVIDER_MODELS: {
7362
7240
  },
7363
7241
  releaseDate: "2025-05-22",
7364
7242
  },
7365
- "mistralai/devstral-small-2505:free": {
7366
- id: "mistralai/devstral-small-2505:free",
7367
- name: "Devstral Small 2505 (free)",
7368
- contextWindow: 32768,
7369
- maxTokens: 32768,
7370
- capabilities: ["tools", "temperature"],
7371
- pricing: {
7372
- input: 0,
7373
- output: 0,
7374
- cacheRead: 0,
7375
- cacheWrite: 0,
7376
- },
7377
- releaseDate: "2025-05-21",
7378
- },
7379
7243
  "mistralai/devstral-small-2505": {
7380
7244
  id: "mistralai/devstral-small-2505",
7381
7245
  name: "Devstral Small",
7382
7246
  contextWindow: 128000,
7383
- maxTokens: 128000,
7247
+ maxTokens: 6400,
7384
7248
  capabilities: ["tools", "temperature"],
7385
7249
  pricing: {
7386
7250
  input: 0.06,
@@ -7394,7 +7258,7 @@ export const GENERATED_PROVIDER_MODELS: {
7394
7258
  id: "mistralai/mistral-medium-3",
7395
7259
  name: "Mistral Medium 3",
7396
7260
  contextWindow: 131072,
7397
- maxTokens: 131072,
7261
+ maxTokens: 6553,
7398
7262
  capabilities: ["images", "tools", "structured_output", "temperature"],
7399
7263
  pricing: {
7400
7264
  input: 0.4,
@@ -7443,7 +7307,7 @@ export const GENERATED_PROVIDER_MODELS: {
7443
7307
  id: "qwen/qwen3-4b:free",
7444
7308
  name: "Qwen3 4B (free)",
7445
7309
  contextWindow: 40960,
7446
- maxTokens: 40960,
7310
+ maxTokens: 2048,
7447
7311
  capabilities: [
7448
7312
  "tools",
7449
7313
  "reasoning",
@@ -7458,25 +7322,6 @@ export const GENERATED_PROVIDER_MODELS: {
7458
7322
  },
7459
7323
  releaseDate: "2025-04-30",
7460
7324
  },
7461
- "qwen/qwen3-14b:free": {
7462
- id: "qwen/qwen3-14b:free",
7463
- name: "Qwen3 14B (free)",
7464
- contextWindow: 40960,
7465
- maxTokens: 40960,
7466
- capabilities: [
7467
- "tools",
7468
- "reasoning",
7469
- "structured_output",
7470
- "temperature",
7471
- ],
7472
- pricing: {
7473
- input: 0,
7474
- output: 0,
7475
- cacheRead: 0,
7476
- cacheWrite: 0,
7477
- },
7478
- releaseDate: "2025-04-28",
7479
- },
7480
7325
  "qwen/qwen3-235b-a22b-07-25": {
7481
7326
  id: "qwen/qwen3-235b-a22b-07-25",
7482
7327
  name: "Qwen3 235B A22B Instruct 2507",
@@ -7491,140 +7336,22 @@ export const GENERATED_PROVIDER_MODELS: {
7491
7336
  },
7492
7337
  releaseDate: "2025-04-28",
7493
7338
  },
7494
- "qwen/qwen3-235b-a22b-07-25:free": {
7495
- id: "qwen/qwen3-235b-a22b-07-25:free",
7496
- name: "Qwen3 235B A22B Instruct 2507 (free)",
7497
- contextWindow: 262144,
7498
- maxTokens: 131072,
7499
- capabilities: ["tools", "temperature"],
7500
- pricing: {
7501
- input: 0,
7502
- output: 0,
7503
- cacheRead: 0,
7504
- cacheWrite: 0,
7505
- },
7506
- releaseDate: "2025-04-28",
7507
- },
7508
- "qwen/qwen3-235b-a22b:free": {
7509
- id: "qwen/qwen3-235b-a22b:free",
7510
- name: "Qwen3 235B A22B (free)",
7511
- contextWindow: 131072,
7512
- maxTokens: 131072,
7339
+ "openai/o4-mini": {
7340
+ id: "openai/o4-mini",
7341
+ name: "o4 Mini",
7342
+ contextWindow: 200000,
7343
+ maxTokens: 100000,
7513
7344
  capabilities: [
7345
+ "images",
7514
7346
  "tools",
7515
7347
  "reasoning",
7516
7348
  "structured_output",
7517
7349
  "temperature",
7518
7350
  ],
7519
7351
  pricing: {
7520
- input: 0,
7521
- output: 0,
7522
- cacheRead: 0,
7523
- cacheWrite: 0,
7524
- },
7525
- releaseDate: "2025-04-28",
7526
- },
7527
- "qwen/qwen3-30b-a3b:free": {
7528
- id: "qwen/qwen3-30b-a3b:free",
7529
- name: "Qwen3 30B A3B (free)",
7530
- contextWindow: 40960,
7531
- maxTokens: 40960,
7532
- capabilities: [
7533
- "tools",
7534
- "reasoning",
7535
- "structured_output",
7536
- "temperature",
7537
- ],
7538
- pricing: {
7539
- input: 0,
7540
- output: 0,
7541
- cacheRead: 0,
7542
- cacheWrite: 0,
7543
- },
7544
- releaseDate: "2025-04-28",
7545
- },
7546
- "qwen/qwen3-32b:free": {
7547
- id: "qwen/qwen3-32b:free",
7548
- name: "Qwen3 32B (free)",
7549
- contextWindow: 40960,
7550
- maxTokens: 40960,
7551
- capabilities: [
7552
- "tools",
7553
- "reasoning",
7554
- "structured_output",
7555
- "temperature",
7556
- ],
7557
- pricing: {
7558
- input: 0,
7559
- output: 0,
7560
- cacheRead: 0,
7561
- cacheWrite: 0,
7562
- },
7563
- releaseDate: "2025-04-28",
7564
- },
7565
- "qwen/qwen3-8b:free": {
7566
- id: "qwen/qwen3-8b:free",
7567
- name: "Qwen3 8B (free)",
7568
- contextWindow: 40960,
7569
- maxTokens: 40960,
7570
- capabilities: [
7571
- "tools",
7572
- "reasoning",
7573
- "structured_output",
7574
- "temperature",
7575
- ],
7576
- pricing: {
7577
- input: 0,
7578
- output: 0,
7579
- cacheRead: 0,
7580
- cacheWrite: 0,
7581
- },
7582
- releaseDate: "2025-04-28",
7583
- },
7584
- "microsoft/mai-ds-r1:free": {
7585
- id: "microsoft/mai-ds-r1:free",
7586
- name: "MAI DS R1 (free)",
7587
- contextWindow: 163840,
7588
- maxTokens: 163840,
7589
- capabilities: ["tools", "reasoning", "temperature"],
7590
- pricing: {
7591
- input: 0,
7592
- output: 0,
7593
- cacheRead: 0,
7594
- cacheWrite: 0,
7595
- },
7596
- releaseDate: "2025-04-21",
7597
- },
7598
- "thudm/glm-z1-32b:free": {
7599
- id: "thudm/glm-z1-32b:free",
7600
- name: "GLM Z1 32B (free)",
7601
- contextWindow: 32768,
7602
- maxTokens: 32768,
7603
- capabilities: ["tools", "reasoning", "temperature"],
7604
- pricing: {
7605
- input: 0,
7606
- output: 0,
7607
- cacheRead: 0,
7608
- cacheWrite: 0,
7609
- },
7610
- releaseDate: "2025-04-17",
7611
- },
7612
- "openai/o4-mini": {
7613
- id: "openai/o4-mini",
7614
- name: "o4 Mini",
7615
- contextWindow: 200000,
7616
- maxTokens: 100000,
7617
- capabilities: [
7618
- "images",
7619
- "tools",
7620
- "reasoning",
7621
- "structured_output",
7622
- "temperature",
7623
- ],
7624
- pricing: {
7625
- input: 1.1,
7626
- output: 4.4,
7627
- cacheRead: 0.28,
7352
+ input: 1.1,
7353
+ output: 4.4,
7354
+ cacheRead: 0.28,
7628
7355
  cacheWrite: 0,
7629
7356
  },
7630
7357
  releaseDate: "2025-04-16",
@@ -7657,34 +7384,6 @@ export const GENERATED_PROVIDER_MODELS: {
7657
7384
  },
7658
7385
  releaseDate: "2025-04-14",
7659
7386
  },
7660
- "meta-llama/llama-4-scout:free": {
7661
- id: "meta-llama/llama-4-scout:free",
7662
- name: "Llama 4 Scout (free)",
7663
- contextWindow: 64000,
7664
- maxTokens: 64000,
7665
- capabilities: ["images", "tools", "structured_output", "temperature"],
7666
- pricing: {
7667
- input: 0,
7668
- output: 0,
7669
- cacheRead: 0,
7670
- cacheWrite: 0,
7671
- },
7672
- releaseDate: "2025-04-05",
7673
- },
7674
- "qwen/qwen2.5-vl-32b-instruct:free": {
7675
- id: "qwen/qwen2.5-vl-32b-instruct:free",
7676
- name: "Qwen2.5 VL 32B Instruct (free)",
7677
- contextWindow: 8192,
7678
- maxTokens: 8192,
7679
- capabilities: ["images", "tools", "structured_output", "temperature"],
7680
- pricing: {
7681
- input: 0,
7682
- output: 0,
7683
- cacheRead: 0,
7684
- cacheWrite: 0,
7685
- },
7686
- releaseDate: "2025-03-24",
7687
- },
7688
7387
  "google/gemini-2.5-pro": {
7689
7388
  id: "google/gemini-2.5-pro",
7690
7389
  name: "Gemini 2.5 Pro",
@@ -7724,7 +7423,7 @@ export const GENERATED_PROVIDER_MODELS: {
7724
7423
  id: "google/gemma-3-27b-it",
7725
7424
  name: "Gemma 3 27B",
7726
7425
  contextWindow: 96000,
7727
- maxTokens: 96000,
7426
+ maxTokens: 4800,
7728
7427
  capabilities: ["images", "tools", "structured_output", "temperature"],
7729
7428
  pricing: {
7730
7429
  input: 0.04,
@@ -7748,53 +7447,6 @@ export const GENERATED_PROVIDER_MODELS: {
7748
7447
  },
7749
7448
  releaseDate: "2025-03-12",
7750
7449
  },
7751
- "rekaai/reka-flash-3": {
7752
- id: "rekaai/reka-flash-3",
7753
- name: "Reka Flash 3",
7754
- contextWindow: 32768,
7755
- maxTokens: 8192,
7756
- capabilities: ["tools", "reasoning", "temperature"],
7757
- pricing: {
7758
- input: 0,
7759
- output: 0,
7760
- cacheRead: 0,
7761
- cacheWrite: 0,
7762
- },
7763
- releaseDate: "2025-03-12",
7764
- },
7765
- "qwen/qwq-32b:free": {
7766
- id: "qwen/qwq-32b:free",
7767
- name: "QwQ 32B (free)",
7768
- contextWindow: 32768,
7769
- maxTokens: 32768,
7770
- capabilities: [
7771
- "tools",
7772
- "reasoning",
7773
- "structured_output",
7774
- "temperature",
7775
- ],
7776
- pricing: {
7777
- input: 0,
7778
- output: 0,
7779
- cacheRead: 0,
7780
- cacheWrite: 0,
7781
- },
7782
- releaseDate: "2025-03-05",
7783
- },
7784
- "nousresearch/deephermes-3-llama-3-8b-preview": {
7785
- id: "nousresearch/deephermes-3-llama-3-8b-preview",
7786
- name: "DeepHermes 3 Llama 3 8B Preview",
7787
- contextWindow: 131072,
7788
- maxTokens: 8192,
7789
- capabilities: ["tools", "reasoning", "temperature"],
7790
- pricing: {
7791
- input: 0,
7792
- output: 0,
7793
- cacheRead: 0,
7794
- cacheWrite: 0,
7795
- },
7796
- releaseDate: "2025-02-28",
7797
- },
7798
7450
  "anthropic/claude-3.7-sonnet": {
7799
7451
  id: "anthropic/claude-3.7-sonnet",
7800
7452
  name: "Claude Sonnet 3.7",
@@ -7870,67 +7522,11 @@ export const GENERATED_PROVIDER_MODELS: {
7870
7522
  },
7871
7523
  releaseDate: "2025-02-17",
7872
7524
  },
7873
- "cognitivecomputations/dolphin3.0-mistral-24b": {
7874
- id: "cognitivecomputations/dolphin3.0-mistral-24b",
7875
- name: "Dolphin3.0 Mistral 24B",
7876
- contextWindow: 32768,
7877
- maxTokens: 8192,
7878
- capabilities: ["tools", "temperature"],
7879
- pricing: {
7880
- input: 0,
7881
- output: 0,
7882
- cacheRead: 0,
7883
- cacheWrite: 0,
7884
- },
7885
- releaseDate: "2025-02-13",
7886
- },
7887
- "cognitivecomputations/dolphin3.0-r1-mistral-24b": {
7888
- id: "cognitivecomputations/dolphin3.0-r1-mistral-24b",
7889
- name: "Dolphin3.0 R1 Mistral 24B",
7890
- contextWindow: 32768,
7891
- maxTokens: 8192,
7892
- capabilities: ["tools", "reasoning", "temperature"],
7893
- pricing: {
7894
- input: 0,
7895
- output: 0,
7896
- cacheRead: 0,
7897
- cacheWrite: 0,
7898
- },
7899
- releaseDate: "2025-02-13",
7900
- },
7901
- "qwen/qwen2.5-vl-72b-instruct:free": {
7902
- id: "qwen/qwen2.5-vl-72b-instruct:free",
7903
- name: "Qwen2.5 VL 72B Instruct (free)",
7904
- contextWindow: 32768,
7905
- maxTokens: 32768,
7906
- capabilities: ["images", "tools", "temperature"],
7907
- pricing: {
7908
- input: 0,
7909
- output: 0,
7910
- cacheRead: 0,
7911
- cacheWrite: 0,
7912
- },
7913
- releaseDate: "2025-02-01",
7914
- },
7915
- "deepseek/deepseek-r1:free": {
7916
- id: "deepseek/deepseek-r1:free",
7917
- name: "R1 (free)",
7918
- contextWindow: 163840,
7919
- maxTokens: 163840,
7920
- capabilities: ["tools", "reasoning", "temperature"],
7921
- pricing: {
7922
- input: 0,
7923
- output: 0,
7924
- cacheRead: 0,
7925
- cacheWrite: 0,
7926
- },
7927
- releaseDate: "2025-01-20",
7928
- },
7929
7525
  "minimax/minimax-01": {
7930
7526
  id: "minimax/minimax-01",
7931
7527
  name: "MiniMax-01",
7932
7528
  contextWindow: 1000000,
7933
- maxTokens: 1000000,
7529
+ maxTokens: 50000,
7934
7530
  capabilities: ["images", "tools", "reasoning", "temperature"],
7935
7531
  pricing: {
7936
7532
  input: 0.2,
@@ -7979,25 +7575,11 @@ export const GENERATED_PROVIDER_MODELS: {
7979
7575
  },
7980
7576
  releaseDate: "2024-12-11",
7981
7577
  },
7982
- "google/gemini-2.0-flash-exp:free": {
7983
- id: "google/gemini-2.0-flash-exp:free",
7984
- name: "Gemini 2.0 Flash Experimental (free)",
7985
- contextWindow: 1048576,
7986
- maxTokens: 1048576,
7987
- capabilities: ["images", "tools", "temperature"],
7988
- pricing: {
7989
- input: 0,
7990
- output: 0,
7991
- cacheRead: 0,
7992
- cacheWrite: 0,
7993
- },
7994
- releaseDate: "2024-12-11",
7995
- },
7996
7578
  "meta-llama/llama-3.3-70b-instruct:free": {
7997
7579
  id: "meta-llama/llama-3.3-70b-instruct:free",
7998
7580
  name: "Llama 3.3 70B Instruct (free)",
7999
7581
  contextWindow: 131072,
8000
- maxTokens: 131072,
7582
+ maxTokens: 6553,
8001
7583
  capabilities: ["tools", "structured_output", "temperature"],
8002
7584
  pricing: {
8003
7585
  input: 0,
@@ -8021,34 +7603,6 @@ export const GENERATED_PROVIDER_MODELS: {
8021
7603
  },
8022
7604
  releaseDate: "2024-10-22",
8023
7605
  },
8024
- "qwen/qwen-2.5-vl-7b-instruct:free": {
8025
- id: "qwen/qwen-2.5-vl-7b-instruct:free",
8026
- name: "Qwen2.5-VL 7B Instruct (free)",
8027
- contextWindow: 32768,
8028
- maxTokens: 32768,
8029
- capabilities: ["images", "tools", "temperature"],
8030
- pricing: {
8031
- input: 0,
8032
- output: 0,
8033
- cacheRead: 0,
8034
- cacheWrite: 0,
8035
- },
8036
- releaseDate: "2024-08-28",
8037
- },
8038
- "mistralai/mistral-nemo:free": {
8039
- id: "mistralai/mistral-nemo:free",
8040
- name: "Mistral Nemo (free)",
8041
- contextWindow: 131072,
8042
- maxTokens: 131072,
8043
- capabilities: ["tools", "structured_output", "temperature"],
8044
- pricing: {
8045
- input: 0,
8046
- output: 0,
8047
- cacheRead: 0,
8048
- cacheWrite: 0,
8049
- },
8050
- releaseDate: "2024-07-19",
8051
- },
8052
7606
  "openai/gpt-4o-mini": {
8053
7607
  id: "openai/gpt-4o-mini",
8054
7608
  name: "GPT-4o-mini",
@@ -8063,20 +7617,6 @@ export const GENERATED_PROVIDER_MODELS: {
8063
7617
  },
8064
7618
  releaseDate: "2024-07-18",
8065
7619
  },
8066
- "mistralai/mistral-7b-instruct:free": {
8067
- id: "mistralai/mistral-7b-instruct:free",
8068
- name: "Mistral 7B Instruct (free)",
8069
- contextWindow: 32768,
8070
- maxTokens: 32768,
8071
- capabilities: ["tools", "temperature"],
8072
- pricing: {
8073
- input: 0,
8074
- output: 0,
8075
- cacheRead: 0,
8076
- cacheWrite: 0,
8077
- },
8078
- releaseDate: "2024-05-27",
8079
- },
8080
7620
  },
8081
7621
  requesty: {
8082
7622
  "openai/gpt-5.4": {
@@ -9027,7 +8567,7 @@ export const GENERATED_PROVIDER_MODELS: {
9027
8567
  id: "Qwen/Qwen3-Coder-Next-FP8",
9028
8568
  name: "Qwen3 Coder Next FP8",
9029
8569
  contextWindow: 262144,
9030
- maxTokens: 262144,
8570
+ maxTokens: 13107,
9031
8571
  capabilities: ["tools", "reasoning", "temperature"],
9032
8572
  pricing: {
9033
8573
  input: 0.5,
@@ -9041,7 +8581,7 @@ export const GENERATED_PROVIDER_MODELS: {
9041
8581
  id: "moonshotai/Kimi-K2.5",
9042
8582
  name: "Kimi K2.5",
9043
8583
  contextWindow: 262144,
9044
- maxTokens: 262144,
8584
+ maxTokens: 13107,
9045
8585
  capabilities: ["images", "tools", "reasoning", "temperature"],
9046
8586
  pricing: {
9047
8587
  input: 0.5,
@@ -9055,7 +8595,7 @@ export const GENERATED_PROVIDER_MODELS: {
9055
8595
  id: "essentialai/Rnj-1-Instruct",
9056
8596
  name: "Rnj-1 Instruct",
9057
8597
  contextWindow: 32768,
9058
- maxTokens: 32768,
8598
+ maxTokens: 1638,
9059
8599
  capabilities: ["tools", "temperature"],
9060
8600
  pricing: {
9061
8601
  input: 0.15,
@@ -9069,7 +8609,7 @@ export const GENERATED_PROVIDER_MODELS: {
9069
8609
  id: "zai-org/GLM-4.6",
9070
8610
  name: "GLM 4.6",
9071
8611
  contextWindow: 200000,
9072
- maxTokens: 200000,
8612
+ maxTokens: 10000,
9073
8613
  capabilities: ["tools", "temperature"],
9074
8614
  pricing: {
9075
8615
  input: 0.6,
@@ -9083,7 +8623,7 @@ export const GENERATED_PROVIDER_MODELS: {
9083
8623
  id: "deepseek-ai/DeepSeek-V3-1",
9084
8624
  name: "DeepSeek V3.1",
9085
8625
  contextWindow: 131072,
9086
- maxTokens: 131072,
8626
+ maxTokens: 6553,
9087
8627
  capabilities: ["tools", "reasoning", "temperature"],
9088
8628
  pricing: {
9089
8629
  input: 0.6,
@@ -9097,7 +8637,7 @@ export const GENERATED_PROVIDER_MODELS: {
9097
8637
  id: "openai/gpt-oss-120b",
9098
8638
  name: "GPT OSS 120B",
9099
8639
  contextWindow: 131072,
9100
- maxTokens: 131072,
8640
+ maxTokens: 6553,
9101
8641
  capabilities: ["tools", "reasoning", "temperature"],
9102
8642
  pricing: {
9103
8643
  input: 0.15,
@@ -9111,7 +8651,7 @@ export const GENERATED_PROVIDER_MODELS: {
9111
8651
  id: "Qwen/Qwen3-235B-A22B-Instruct-2507-tput",
9112
8652
  name: "Qwen3 235B A22B Instruct 2507 FP8",
9113
8653
  contextWindow: 262144,
9114
- maxTokens: 262144,
8654
+ maxTokens: 13107,
9115
8655
  capabilities: ["tools", "reasoning", "temperature"],
9116
8656
  pricing: {
9117
8657
  input: 0.2,
@@ -9125,7 +8665,7 @@ export const GENERATED_PROVIDER_MODELS: {
9125
8665
  id: "Qwen/Qwen3-Next-80B-A3B-Instruct",
9126
8666
  name: "Qwen3-Next-80B-A3B-Instruct",
9127
8667
  contextWindow: 262144,
9128
- maxTokens: 262144,
8668
+ maxTokens: 13107,
9129
8669
  capabilities: ["tools", "temperature"],
9130
8670
  pricing: {
9131
8671
  input: 0.15,
@@ -9139,7 +8679,7 @@ export const GENERATED_PROVIDER_MODELS: {
9139
8679
  id: "zai-org/GLM-4.7",
9140
8680
  name: "GLM-4.7",
9141
8681
  contextWindow: 200000,
9142
- maxTokens: 200000,
8682
+ maxTokens: 10000,
9143
8683
  capabilities: ["tools", "reasoning", "temperature"],
9144
8684
  pricing: {
9145
8685
  input: 0.45,
@@ -9153,7 +8693,7 @@ export const GENERATED_PROVIDER_MODELS: {
9153
8693
  id: "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8",
9154
8694
  name: "Qwen3 Coder 480B A35B Instruct",
9155
8695
  contextWindow: 262144,
9156
- maxTokens: 262144,
8696
+ maxTokens: 13107,
9157
8697
  capabilities: ["tools", "temperature"],
9158
8698
  pricing: {
9159
8699
  input: 2,
@@ -9167,7 +8707,7 @@ export const GENERATED_PROVIDER_MODELS: {
9167
8707
  id: "moonshotai/Kimi-K2-Instruct",
9168
8708
  name: "Kimi K2 Instruct",
9169
8709
  contextWindow: 131072,
9170
- maxTokens: 131072,
8710
+ maxTokens: 6553,
9171
8711
  capabilities: ["tools", "temperature"],
9172
8712
  pricing: {
9173
8713
  input: 1,
@@ -9181,7 +8721,7 @@ export const GENERATED_PROVIDER_MODELS: {
9181
8721
  id: "deepseek-ai/DeepSeek-V3",
9182
8722
  name: "DeepSeek V3",
9183
8723
  contextWindow: 131072,
9184
- maxTokens: 131072,
8724
+ maxTokens: 6553,
9185
8725
  capabilities: ["tools", "reasoning", "temperature"],
9186
8726
  pricing: {
9187
8727
  input: 1.25,
@@ -9195,7 +8735,7 @@ export const GENERATED_PROVIDER_MODELS: {
9195
8735
  id: "meta-llama/Llama-3.3-70B-Instruct-Turbo",
9196
8736
  name: "Llama 3.3 70B",
9197
8737
  contextWindow: 131072,
9198
- maxTokens: 131072,
8738
+ maxTokens: 6553,
9199
8739
  capabilities: ["tools", "temperature"],
9200
8740
  pricing: {
9201
8741
  input: 0.88,
@@ -9309,7 +8849,7 @@ export const GENERATED_PROVIDER_MODELS: {
9309
8849
  id: "xai/grok-4.20-multi-agent-beta",
9310
8850
  name: "Grok 4.20 Multi Agent Beta",
9311
8851
  contextWindow: 2000000,
9312
- maxTokens: 2000000,
8852
+ maxTokens: 100000,
9313
8853
  capabilities: ["tools", "reasoning", "temperature"],
9314
8854
  pricing: {
9315
8855
  input: 2,
@@ -9323,7 +8863,7 @@ export const GENERATED_PROVIDER_MODELS: {
9323
8863
  id: "xai/grok-4.20-non-reasoning-beta",
9324
8864
  name: "Grok 4.20 Beta Non-Reasoning",
9325
8865
  contextWindow: 2000000,
9326
- maxTokens: 2000000,
8866
+ maxTokens: 100000,
9327
8867
  capabilities: ["images", "files", "tools", "temperature"],
9328
8868
  pricing: {
9329
8869
  input: 2,
@@ -9337,7 +8877,7 @@ export const GENERATED_PROVIDER_MODELS: {
9337
8877
  id: "xai/grok-4.20-reasoning-beta",
9338
8878
  name: "Grok 4.20 Beta Reasoning",
9339
8879
  contextWindow: 2000000,
9340
- maxTokens: 2000000,
8880
+ maxTokens: 100000,
9341
8881
  capabilities: ["images", "files", "tools", "reasoning", "temperature"],
9342
8882
  pricing: {
9343
8883
  input: 2,
@@ -9347,6 +8887,48 @@ export const GENERATED_PROVIDER_MODELS: {
9347
8887
  },
9348
8888
  releaseDate: "2026-03-11",
9349
8889
  },
8890
+ "xai/grok-4.20-multi-agent": {
8891
+ id: "xai/grok-4.20-multi-agent",
8892
+ name: "Grok 4.20 Multi-Agent",
8893
+ contextWindow: 2000000,
8894
+ maxTokens: 100000,
8895
+ capabilities: ["tools", "reasoning", "temperature"],
8896
+ pricing: {
8897
+ input: 2,
8898
+ output: 6,
8899
+ cacheRead: 0.19999999999999998,
8900
+ cacheWrite: 0,
8901
+ },
8902
+ releaseDate: "2026-03-09",
8903
+ },
8904
+ "xai/grok-4.20-non-reasoning": {
8905
+ id: "xai/grok-4.20-non-reasoning",
8906
+ name: "Grok 4.20 Non-Reasoning",
8907
+ contextWindow: 2000000,
8908
+ maxTokens: 100000,
8909
+ capabilities: ["images", "files", "tools", "temperature"],
8910
+ pricing: {
8911
+ input: 2,
8912
+ output: 6,
8913
+ cacheRead: 0.19999999999999998,
8914
+ cacheWrite: 0,
8915
+ },
8916
+ releaseDate: "2026-03-09",
8917
+ },
8918
+ "xai/grok-4.20-reasoning": {
8919
+ id: "xai/grok-4.20-reasoning",
8920
+ name: "Grok 4.20 Reasoning",
8921
+ contextWindow: 2000000,
8922
+ maxTokens: 100000,
8923
+ capabilities: ["images", "files", "tools", "reasoning", "temperature"],
8924
+ pricing: {
8925
+ input: 2,
8926
+ output: 6,
8927
+ cacheRead: 0.19999999999999998,
8928
+ cacheWrite: 0,
8929
+ },
8930
+ releaseDate: "2026-03-09",
8931
+ },
9350
8932
  "openai/gpt-5.4": {
9351
8933
  id: "openai/gpt-5.4",
9352
8934
  name: "GPT 5.4",
@@ -9421,7 +9003,7 @@ export const GENERATED_PROVIDER_MODELS: {
9421
9003
  id: "inception/mercury-2",
9422
9004
  name: "Mercury 2",
9423
9005
  contextWindow: 128000,
9424
- maxTokens: 128000,
9006
+ maxTokens: 6400,
9425
9007
  capabilities: ["tools", "reasoning", "temperature"],
9426
9008
  pricing: {
9427
9009
  input: 0.25,
@@ -9547,7 +9129,7 @@ export const GENERATED_PROVIDER_MODELS: {
9547
9129
  id: "moonshotai/kimi-k2.5",
9548
9130
  name: "Kimi K2.5",
9549
9131
  contextWindow: 262144,
9550
- maxTokens: 262144,
9132
+ maxTokens: 13107,
9551
9133
  capabilities: ["images", "tools", "reasoning", "temperature"],
9552
9134
  pricing: {
9553
9135
  input: 0.6,
@@ -9603,7 +9185,7 @@ export const GENERATED_PROVIDER_MODELS: {
9603
9185
  id: "mistral/devstral-2",
9604
9186
  name: "Devstral 2",
9605
9187
  contextWindow: 256000,
9606
- maxTokens: 256000,
9188
+ maxTokens: 12800,
9607
9189
  capabilities: ["tools", "temperature"],
9608
9190
  pricing: {
9609
9191
  input: 0,
@@ -9645,7 +9227,7 @@ export const GENERATED_PROVIDER_MODELS: {
9645
9227
  id: "prime-intellect/intellect-3",
9646
9228
  name: "INTELLECT 3",
9647
9229
  contextWindow: 131072,
9648
- maxTokens: 131072,
9230
+ maxTokens: 6553,
9649
9231
  capabilities: ["tools", "reasoning", "temperature"],
9650
9232
  pricing: {
9651
9233
  input: 0.2,
@@ -9687,7 +9269,7 @@ export const GENERATED_PROVIDER_MODELS: {
9687
9269
  id: "moonshotai/kimi-k2-thinking",
9688
9270
  name: "Kimi K2 Thinking",
9689
9271
  contextWindow: 216144,
9690
- maxTokens: 216144,
9272
+ maxTokens: 10807,
9691
9273
  capabilities: ["tools", "reasoning", "temperature"],
9692
9274
  pricing: {
9693
9275
  input: 0.47,
@@ -9701,7 +9283,7 @@ export const GENERATED_PROVIDER_MODELS: {
9701
9283
  id: "moonshotai/kimi-k2-thinking-turbo",
9702
9284
  name: "Kimi K2 Thinking Turbo",
9703
9285
  contextWindow: 262114,
9704
- maxTokens: 262114,
9286
+ maxTokens: 13105,
9705
9287
  capabilities: ["tools", "reasoning", "temperature"],
9706
9288
  pricing: {
9707
9289
  input: 1.15,
@@ -9715,7 +9297,7 @@ export const GENERATED_PROVIDER_MODELS: {
9715
9297
  id: "minimax/minimax-m2",
9716
9298
  name: "MiniMax M2",
9717
9299
  contextWindow: 262114,
9718
- maxTokens: 262114,
9300
+ maxTokens: 13105,
9719
9301
  capabilities: ["tools", "reasoning", "temperature"],
9720
9302
  pricing: {
9721
9303
  input: 0.27,
@@ -9841,7 +9423,7 @@ export const GENERATED_PROVIDER_MODELS: {
9841
9423
  id: "deepseek/deepseek-v3.2-exp",
9842
9424
  name: "DeepSeek V3.2 Exp",
9843
9425
  contextWindow: 163840,
9844
- maxTokens: 163840,
9426
+ maxTokens: 8192,
9845
9427
  capabilities: ["tools", "reasoning", "temperature"],
9846
9428
  pricing: {
9847
9429
  input: 0.27,
@@ -10093,7 +9675,7 @@ export const GENERATED_PROVIDER_MODELS: {
10093
9675
  id: "nvidia/nemotron-nano-9b-v2",
10094
9676
  name: "Nvidia Nemotron Nano 9B V2",
10095
9677
  contextWindow: 131072,
10096
- maxTokens: 131072,
9678
+ maxTokens: 6553,
10097
9679
  capabilities: ["tools", "reasoning", "temperature"],
10098
9680
  pricing: {
10099
9681
  input: 0.04,
@@ -10107,7 +9689,7 @@ export const GENERATED_PROVIDER_MODELS: {
10107
9689
  id: "zai/glm-4.5v",
10108
9690
  name: "GLM 4.5V",
10109
9691
  contextWindow: 66000,
10110
- maxTokens: 66000,
9692
+ maxTokens: 3300,
10111
9693
  capabilities: ["images", "tools", "reasoning", "temperature"],
10112
9694
  pricing: {
10113
9695
  input: 0.6,
@@ -10289,7 +9871,7 @@ export const GENERATED_PROVIDER_MODELS: {
10289
9871
  id: "openai/gpt-oss-120b",
10290
9872
  name: "GPT OSS 120B",
10291
9873
  contextWindow: 131072,
10292
- maxTokens: 131072,
9874
+ maxTokens: 6553,
10293
9875
  capabilities: ["tools", "reasoning", "temperature"],
10294
9876
  pricing: {
10295
9877
  input: 0.1,
@@ -10317,7 +9899,7 @@ export const GENERATED_PROVIDER_MODELS: {
10317
9899
  id: "zai/glm-4.5",
10318
9900
  name: "GLM 4.5",
10319
9901
  contextWindow: 131072,
10320
- maxTokens: 131072,
9902
+ maxTokens: 6553,
10321
9903
  capabilities: ["tools", "reasoning", "temperature"],
10322
9904
  pricing: {
10323
9905
  input: 0.6,
@@ -10345,7 +9927,7 @@ export const GENERATED_PROVIDER_MODELS: {
10345
9927
  id: "alibaba/qwen3-coder-plus",
10346
9928
  name: "Qwen3 Coder Plus",
10347
9929
  contextWindow: 1000000,
10348
- maxTokens: 1000000,
9930
+ maxTokens: 50000,
10349
9931
  capabilities: ["tools", "temperature"],
10350
9932
  pricing: {
10351
9933
  input: 1,
@@ -10359,7 +9941,7 @@ export const GENERATED_PROVIDER_MODELS: {
10359
9941
  id: "alibaba/qwen3-coder-next",
10360
9942
  name: "Qwen3 Coder Next",
10361
9943
  contextWindow: 256000,
10362
- maxTokens: 256000,
9944
+ maxTokens: 12800,
10363
9945
  capabilities: ["tools", "reasoning", "temperature"],
10364
9946
  pricing: {
10365
9947
  input: 0.5,
@@ -10555,7 +10137,7 @@ export const GENERATED_PROVIDER_MODELS: {
10555
10137
  id: "mistral/devstral-small-2",
10556
10138
  name: "Devstral Small 2",
10557
10139
  contextWindow: 256000,
10558
- maxTokens: 256000,
10140
+ maxTokens: 12800,
10559
10141
  capabilities: ["tools", "temperature"],
10560
10142
  pricing: {
10561
10143
  input: 0,
@@ -10751,7 +10333,7 @@ export const GENERATED_PROVIDER_MODELS: {
10751
10333
  id: "alibaba/qwen3-235b-a22b-thinking",
10752
10334
  name: "Qwen3 235B A22B Thinking 2507",
10753
10335
  contextWindow: 262114,
10754
- maxTokens: 262114,
10336
+ maxTokens: 13105,
10755
10337
  capabilities: ["images", "files", "tools", "reasoning", "temperature"],
10756
10338
  pricing: {
10757
10339
  input: 0.3,
@@ -10849,7 +10431,7 @@ export const GENERATED_PROVIDER_MODELS: {
10849
10431
  id: "mistral/magistral-small",
10850
10432
  name: "Magistral Small",
10851
10433
  contextWindow: 128000,
10852
- maxTokens: 128000,
10434
+ maxTokens: 6400,
10853
10435
  capabilities: ["tools", "reasoning", "temperature"],
10854
10436
  pricing: {
10855
10437
  input: 0.5,
@@ -11017,7 +10599,7 @@ export const GENERATED_PROVIDER_MODELS: {
11017
10599
  id: "arcee-ai/trinity-large-preview",
11018
10600
  name: "Trinity Large Preview",
11019
10601
  contextWindow: 131000,
11020
- maxTokens: 131000,
10602
+ maxTokens: 6550,
11021
10603
  capabilities: ["tools", "temperature"],
11022
10604
  pricing: {
11023
10605
  input: 0.25,
@@ -11183,7 +10765,7 @@ export const GENERATED_PROVIDER_MODELS: {
11183
10765
  id: "nvidia/nemotron-nano-12b-v2-vl",
11184
10766
  name: "Nvidia Nemotron Nano 12B V2 VL",
11185
10767
  contextWindow: 131072,
11186
- maxTokens: 131072,
10768
+ maxTokens: 6553,
11187
10769
  capabilities: ["images", "tools", "reasoning", "temperature"],
11188
10770
  pricing: {
11189
10771
  input: 0.2,
@@ -11211,7 +10793,7 @@ export const GENERATED_PROVIDER_MODELS: {
11211
10793
  id: "mistral/pixtral-large",
11212
10794
  name: "Pixtral Large (latest)",
11213
10795
  contextWindow: 128000,
11214
- maxTokens: 128000,
10796
+ maxTokens: 6400,
11215
10797
  capabilities: ["images", "tools", "temperature"],
11216
10798
  pricing: {
11217
10799
  input: 2,
@@ -11253,7 +10835,7 @@ export const GENERATED_PROVIDER_MODELS: {
11253
10835
  id: "mistral/ministral-3b",
11254
10836
  name: "Ministral 3B (latest)",
11255
10837
  contextWindow: 128000,
11256
- maxTokens: 128000,
10838
+ maxTokens: 6400,
11257
10839
  capabilities: ["tools", "temperature"],
11258
10840
  pricing: {
11259
10841
  input: 0.04,
@@ -11267,7 +10849,7 @@ export const GENERATED_PROVIDER_MODELS: {
11267
10849
  id: "mistral/ministral-8b",
11268
10850
  name: "Ministral 8B (latest)",
11269
10851
  contextWindow: 128000,
11270
- maxTokens: 128000,
10852
+ maxTokens: 6400,
11271
10853
  capabilities: ["tools", "temperature"],
11272
10854
  pricing: {
11273
10855
  input: 0.1,
@@ -11323,7 +10905,7 @@ export const GENERATED_PROVIDER_MODELS: {
11323
10905
  id: "mistral/pixtral-12b",
11324
10906
  name: "Pixtral 12B",
11325
10907
  contextWindow: 128000,
11326
- maxTokens: 128000,
10908
+ maxTokens: 6400,
11327
10909
  capabilities: ["images", "tools", "temperature"],
11328
10910
  pricing: {
11329
10911
  input: 0.15,
@@ -11463,7 +11045,7 @@ export const GENERATED_PROVIDER_MODELS: {
11463
11045
  id: "mistral/mixtral-8x22b-instruct",
11464
11046
  name: "Mixtral 8x22B",
11465
11047
  contextWindow: 64000,
11466
- maxTokens: 64000,
11048
+ maxTokens: 3200,
11467
11049
  capabilities: ["tools", "temperature"],
11468
11050
  pricing: {
11469
11051
  input: 2,
@@ -11673,9 +11255,9 @@ export const GENERATED_PROVIDER_MODELS: {
11673
11255
  },
11674
11256
  },
11675
11257
  xai: {
11676
- "grok-4.20-beta-latest-non-reasoning": {
11677
- id: "grok-4.20-beta-latest-non-reasoning",
11678
- name: "Grok 4.20 Beta (Non-Reasoning)",
11258
+ "grok-4.20-0309-non-reasoning": {
11259
+ id: "grok-4.20-0309-non-reasoning",
11260
+ name: "Grok 4.20 (Non-Reasoning)",
11679
11261
  contextWindow: 2000000,
11680
11262
  maxTokens: 30000,
11681
11263
  capabilities: ["images", "tools", "temperature"],
@@ -11687,9 +11269,9 @@ export const GENERATED_PROVIDER_MODELS: {
11687
11269
  },
11688
11270
  releaseDate: "2026-03-09",
11689
11271
  },
11690
- "grok-4.20-beta-latest-reasoning": {
11691
- id: "grok-4.20-beta-latest-reasoning",
11692
- name: "Grok 4.20 Beta (Reasoning)",
11272
+ "grok-4.20-0309-reasoning": {
11273
+ id: "grok-4.20-0309-reasoning",
11274
+ name: "Grok 4.20 (Reasoning)",
11693
11275
  contextWindow: 2000000,
11694
11276
  maxTokens: 30000,
11695
11277
  capabilities: ["images", "tools", "reasoning", "temperature"],