@draht/ai 2026.3.14 → 2026.3.25-1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (82) hide show
  1. package/README.md +28 -0
  2. package/dist/cli.js +0 -0
  3. package/dist/index.d.ts +10 -8
  4. package/dist/index.d.ts.map +1 -1
  5. package/dist/index.js +0 -8
  6. package/dist/index.js.map +1 -1
  7. package/dist/models.d.ts +1 -1
  8. package/dist/models.d.ts.map +1 -1
  9. package/dist/models.generated.d.ts +593 -175
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +781 -371
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/models.js +3 -3
  14. package/dist/models.js.map +1 -1
  15. package/dist/providers/amazon-bedrock.d.ts +5 -0
  16. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  17. package/dist/providers/amazon-bedrock.js +32 -9
  18. package/dist/providers/amazon-bedrock.js.map +1 -1
  19. package/dist/providers/anthropic.d.ts +7 -0
  20. package/dist/providers/anthropic.d.ts.map +1 -1
  21. package/dist/providers/anthropic.js +42 -25
  22. package/dist/providers/anthropic.js.map +1 -1
  23. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  24. package/dist/providers/azure-openai-responses.js +1 -12
  25. package/dist/providers/azure-openai-responses.js.map +1 -1
  26. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  27. package/dist/providers/google-gemini-cli.js +19 -0
  28. package/dist/providers/google-gemini-cli.js.map +1 -1
  29. package/dist/providers/google-shared.d.ts.map +1 -1
  30. package/dist/providers/google-shared.js +20 -8
  31. package/dist/providers/google-shared.js.map +1 -1
  32. package/dist/providers/google-vertex.d.ts.map +1 -1
  33. package/dist/providers/google-vertex.js +28 -1
  34. package/dist/providers/google-vertex.js.map +1 -1
  35. package/dist/providers/google.d.ts.map +1 -1
  36. package/dist/providers/google.js +19 -0
  37. package/dist/providers/google.js.map +1 -1
  38. package/dist/providers/mistral.d.ts.map +1 -1
  39. package/dist/providers/mistral.js +3 -0
  40. package/dist/providers/mistral.js.map +1 -1
  41. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  42. package/dist/providers/openai-codex-responses.js +72 -35
  43. package/dist/providers/openai-codex-responses.js.map +1 -1
  44. package/dist/providers/openai-completions.d.ts.map +1 -1
  45. package/dist/providers/openai-completions.js +53 -16
  46. package/dist/providers/openai-completions.js.map +1 -1
  47. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  48. package/dist/providers/openai-responses-shared.js +43 -39
  49. package/dist/providers/openai-responses-shared.js.map +1 -1
  50. package/dist/providers/openai-responses.d.ts.map +1 -1
  51. package/dist/providers/openai-responses.js +2 -13
  52. package/dist/providers/openai-responses.js.map +1 -1
  53. package/dist/providers/register-builtins.d.ts +28 -1
  54. package/dist/providers/register-builtins.d.ts.map +1 -1
  55. package/dist/providers/register-builtins.js +170 -47
  56. package/dist/providers/register-builtins.js.map +1 -1
  57. package/dist/types.d.ts +11 -2
  58. package/dist/types.d.ts.map +1 -1
  59. package/dist/types.js.map +1 -1
  60. package/dist/utils/oauth/anthropic.d.ts +14 -6
  61. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  62. package/dist/utils/oauth/anthropic.js +288 -57
  63. package/dist/utils/oauth/anthropic.js.map +1 -1
  64. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  65. package/dist/utils/oauth/google-antigravity.js +22 -19
  66. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  67. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  68. package/dist/utils/oauth/google-gemini-cli.js +22 -19
  69. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  70. package/dist/utils/oauth/oauth-page.d.ts +3 -0
  71. package/dist/utils/oauth/oauth-page.d.ts.map +1 -0
  72. package/dist/utils/oauth/oauth-page.js +105 -0
  73. package/dist/utils/oauth/oauth-page.js.map +1 -0
  74. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  75. package/dist/utils/oauth/openai-codex.js +24 -31
  76. package/dist/utils/oauth/openai-codex.js.map +1 -1
  77. package/dist/utils/validation.d.ts.map +1 -1
  78. package/dist/utils/validation.js +17 -9
  79. package/dist/utils/validation.js.map +1 -1
  80. package/package.json +39 -5
  81. package/bedrock-provider.d.ts +0 -1
  82. package/bedrock-provider.js +0 -1
@@ -799,6 +799,23 @@ export declare const MODELS: {
799
799
  contextWindow: number;
800
800
  maxTokens: number;
801
801
  };
802
+ readonly "minimax.minimax-m2.5": {
803
+ id: string;
804
+ name: string;
805
+ api: "bedrock-converse-stream";
806
+ provider: string;
807
+ baseUrl: string;
808
+ reasoning: true;
809
+ input: "text"[];
810
+ cost: {
811
+ input: number;
812
+ output: number;
813
+ cacheRead: number;
814
+ cacheWrite: number;
815
+ };
816
+ contextWindow: number;
817
+ maxTokens: number;
818
+ };
802
819
  readonly "mistral.devstral-2-123b": {
803
820
  id: string;
804
821
  name: string;
@@ -1411,6 +1428,23 @@ export declare const MODELS: {
1411
1428
  contextWindow: number;
1412
1429
  maxTokens: number;
1413
1430
  };
1431
+ readonly "zai.glm-5": {
1432
+ id: string;
1433
+ name: string;
1434
+ api: "bedrock-converse-stream";
1435
+ provider: string;
1436
+ baseUrl: string;
1437
+ reasoning: true;
1438
+ input: "text"[];
1439
+ cost: {
1440
+ input: number;
1441
+ output: number;
1442
+ cacheRead: number;
1443
+ cacheWrite: number;
1444
+ };
1445
+ contextWindow: number;
1446
+ maxTokens: number;
1447
+ };
1414
1448
  };
1415
1449
  readonly anthropic: {
1416
1450
  readonly "claude-3-5-haiku-20241022": {
@@ -2299,6 +2333,40 @@ export declare const MODELS: {
2299
2333
  contextWindow: number;
2300
2334
  maxTokens: number;
2301
2335
  };
2336
+ readonly "gpt-5.4-mini": {
2337
+ id: string;
2338
+ name: string;
2339
+ api: "azure-openai-responses";
2340
+ provider: string;
2341
+ baseUrl: string;
2342
+ reasoning: true;
2343
+ input: ("image" | "text")[];
2344
+ cost: {
2345
+ input: number;
2346
+ output: number;
2347
+ cacheRead: number;
2348
+ cacheWrite: number;
2349
+ };
2350
+ contextWindow: number;
2351
+ maxTokens: number;
2352
+ };
2353
+ readonly "gpt-5.4-nano": {
2354
+ id: string;
2355
+ name: string;
2356
+ api: "azure-openai-responses";
2357
+ provider: string;
2358
+ baseUrl: string;
2359
+ reasoning: true;
2360
+ input: ("image" | "text")[];
2361
+ cost: {
2362
+ input: number;
2363
+ output: number;
2364
+ cacheRead: number;
2365
+ cacheWrite: number;
2366
+ };
2367
+ contextWindow: number;
2368
+ maxTokens: number;
2369
+ };
2302
2370
  readonly "gpt-5.4-pro": {
2303
2371
  id: string;
2304
2372
  name: string;
@@ -3060,6 +3128,29 @@ export declare const MODELS: {
3060
3128
  contextWindow: number;
3061
3129
  maxTokens: number;
3062
3130
  };
3131
+ readonly "gpt-5.4-mini": {
3132
+ id: string;
3133
+ name: string;
3134
+ api: "openai-responses";
3135
+ provider: string;
3136
+ baseUrl: string;
3137
+ headers: {
3138
+ "User-Agent": string;
3139
+ "Editor-Version": string;
3140
+ "Editor-Plugin-Version": string;
3141
+ "Copilot-Integration-Id": string;
3142
+ };
3143
+ reasoning: true;
3144
+ input: ("image" | "text")[];
3145
+ cost: {
3146
+ input: number;
3147
+ output: number;
3148
+ cacheRead: number;
3149
+ cacheWrite: number;
3150
+ };
3151
+ contextWindow: number;
3152
+ maxTokens: number;
3153
+ };
3063
3154
  readonly "grok-code-fast-1": {
3064
3155
  id: string;
3065
3156
  name: string;
@@ -4654,6 +4745,23 @@ export declare const MODELS: {
4654
4745
  contextWindow: number;
4655
4746
  maxTokens: number;
4656
4747
  };
4748
+ readonly "MiniMax-M2.1-highspeed": {
4749
+ id: string;
4750
+ name: string;
4751
+ api: "anthropic-messages";
4752
+ provider: string;
4753
+ baseUrl: string;
4754
+ reasoning: true;
4755
+ input: "text"[];
4756
+ cost: {
4757
+ input: number;
4758
+ output: number;
4759
+ cacheRead: number;
4760
+ cacheWrite: number;
4761
+ };
4762
+ contextWindow: number;
4763
+ maxTokens: number;
4764
+ };
4657
4765
  readonly "MiniMax-M2.5": {
4658
4766
  id: string;
4659
4767
  name: string;
@@ -4688,6 +4796,40 @@ export declare const MODELS: {
4688
4796
  contextWindow: number;
4689
4797
  maxTokens: number;
4690
4798
  };
4799
+ readonly "MiniMax-M2.7": {
4800
+ id: string;
4801
+ name: string;
4802
+ api: "anthropic-messages";
4803
+ provider: string;
4804
+ baseUrl: string;
4805
+ reasoning: true;
4806
+ input: "text"[];
4807
+ cost: {
4808
+ input: number;
4809
+ output: number;
4810
+ cacheRead: number;
4811
+ cacheWrite: number;
4812
+ };
4813
+ contextWindow: number;
4814
+ maxTokens: number;
4815
+ };
4816
+ readonly "MiniMax-M2.7-highspeed": {
4817
+ id: string;
4818
+ name: string;
4819
+ api: "anthropic-messages";
4820
+ provider: string;
4821
+ baseUrl: string;
4822
+ reasoning: true;
4823
+ input: "text"[];
4824
+ cost: {
4825
+ input: number;
4826
+ output: number;
4827
+ cacheRead: number;
4828
+ cacheWrite: number;
4829
+ };
4830
+ contextWindow: number;
4831
+ maxTokens: number;
4832
+ };
4691
4833
  };
4692
4834
  readonly "minimax-cn": {
4693
4835
  readonly "MiniMax-M2": {
@@ -4724,6 +4866,23 @@ export declare const MODELS: {
4724
4866
  contextWindow: number;
4725
4867
  maxTokens: number;
4726
4868
  };
4869
+ readonly "MiniMax-M2.1-highspeed": {
4870
+ id: string;
4871
+ name: string;
4872
+ api: "anthropic-messages";
4873
+ provider: string;
4874
+ baseUrl: string;
4875
+ reasoning: true;
4876
+ input: "text"[];
4877
+ cost: {
4878
+ input: number;
4879
+ output: number;
4880
+ cacheRead: number;
4881
+ cacheWrite: number;
4882
+ };
4883
+ contextWindow: number;
4884
+ maxTokens: number;
4885
+ };
4727
4886
  readonly "MiniMax-M2.5": {
4728
4887
  id: string;
4729
4888
  name: string;
@@ -4758,6 +4917,40 @@ export declare const MODELS: {
4758
4917
  contextWindow: number;
4759
4918
  maxTokens: number;
4760
4919
  };
4920
+ readonly "MiniMax-M2.7": {
4921
+ id: string;
4922
+ name: string;
4923
+ api: "anthropic-messages";
4924
+ provider: string;
4925
+ baseUrl: string;
4926
+ reasoning: true;
4927
+ input: "text"[];
4928
+ cost: {
4929
+ input: number;
4930
+ output: number;
4931
+ cacheRead: number;
4932
+ cacheWrite: number;
4933
+ };
4934
+ contextWindow: number;
4935
+ maxTokens: number;
4936
+ };
4937
+ readonly "MiniMax-M2.7-highspeed": {
4938
+ id: string;
4939
+ name: string;
4940
+ api: "anthropic-messages";
4941
+ provider: string;
4942
+ baseUrl: string;
4943
+ reasoning: true;
4944
+ input: "text"[];
4945
+ cost: {
4946
+ input: number;
4947
+ output: number;
4948
+ cacheRead: number;
4949
+ cacheWrite: number;
4950
+ };
4951
+ contextWindow: number;
4952
+ maxTokens: number;
4953
+ };
4761
4954
  };
4762
4955
  readonly mistral: {
4763
4956
  readonly "codestral-latest": {
@@ -5680,6 +5873,40 @@ export declare const MODELS: {
5680
5873
  contextWindow: number;
5681
5874
  maxTokens: number;
5682
5875
  };
5876
+ readonly "gpt-5.4-mini": {
5877
+ id: string;
5878
+ name: string;
5879
+ api: "openai-responses";
5880
+ provider: string;
5881
+ baseUrl: string;
5882
+ reasoning: true;
5883
+ input: ("image" | "text")[];
5884
+ cost: {
5885
+ input: number;
5886
+ output: number;
5887
+ cacheRead: number;
5888
+ cacheWrite: number;
5889
+ };
5890
+ contextWindow: number;
5891
+ maxTokens: number;
5892
+ };
5893
+ readonly "gpt-5.4-nano": {
5894
+ id: string;
5895
+ name: string;
5896
+ api: "openai-responses";
5897
+ provider: string;
5898
+ baseUrl: string;
5899
+ reasoning: true;
5900
+ input: ("image" | "text")[];
5901
+ cost: {
5902
+ input: number;
5903
+ output: number;
5904
+ cacheRead: number;
5905
+ cacheWrite: number;
5906
+ };
5907
+ contextWindow: number;
5908
+ maxTokens: number;
5909
+ };
5683
5910
  readonly "gpt-5.4-pro": {
5684
5911
  id: string;
5685
5912
  name: string;
@@ -5971,6 +6198,23 @@ export declare const MODELS: {
5971
6198
  contextWindow: number;
5972
6199
  maxTokens: number;
5973
6200
  };
6201
+ readonly "gpt-5.4-mini": {
6202
+ id: string;
6203
+ name: string;
6204
+ api: "openai-codex-responses";
6205
+ provider: string;
6206
+ baseUrl: string;
6207
+ reasoning: true;
6208
+ input: ("image" | "text")[];
6209
+ cost: {
6210
+ input: number;
6211
+ output: number;
6212
+ cacheRead: number;
6213
+ cacheWrite: number;
6214
+ };
6215
+ contextWindow: number;
6216
+ maxTokens: number;
6217
+ };
5974
6218
  };
5975
6219
  readonly opencode: {
5976
6220
  readonly "big-pickle": {
@@ -6143,7 +6387,7 @@ export declare const MODELS: {
6143
6387
  contextWindow: number;
6144
6388
  maxTokens: number;
6145
6389
  };
6146
- readonly "gemini-3-pro": {
6390
+ readonly "gemini-3.1-pro": {
6147
6391
  id: string;
6148
6392
  name: string;
6149
6393
  api: "google-generative-ai";
@@ -6160,14 +6404,14 @@ export declare const MODELS: {
6160
6404
  contextWindow: number;
6161
6405
  maxTokens: number;
6162
6406
  };
6163
- readonly "gemini-3.1-pro": {
6407
+ readonly "glm-5": {
6164
6408
  id: string;
6165
6409
  name: string;
6166
- api: "google-generative-ai";
6410
+ api: "openai-completions";
6167
6411
  provider: string;
6168
6412
  baseUrl: string;
6169
6413
  reasoning: true;
6170
- input: ("image" | "text")[];
6414
+ input: "text"[];
6171
6415
  cost: {
6172
6416
  input: number;
6173
6417
  output: number;
@@ -6177,14 +6421,14 @@ export declare const MODELS: {
6177
6421
  contextWindow: number;
6178
6422
  maxTokens: number;
6179
6423
  };
6180
- readonly "glm-4.6": {
6424
+ readonly "gpt-5": {
6181
6425
  id: string;
6182
6426
  name: string;
6183
- api: "openai-completions";
6427
+ api: "openai-responses";
6184
6428
  provider: string;
6185
6429
  baseUrl: string;
6186
6430
  reasoning: true;
6187
- input: "text"[];
6431
+ input: ("image" | "text")[];
6188
6432
  cost: {
6189
6433
  input: number;
6190
6434
  output: number;
@@ -6194,14 +6438,14 @@ export declare const MODELS: {
6194
6438
  contextWindow: number;
6195
6439
  maxTokens: number;
6196
6440
  };
6197
- readonly "glm-4.7": {
6441
+ readonly "gpt-5-codex": {
6198
6442
  id: string;
6199
6443
  name: string;
6200
- api: "openai-completions";
6444
+ api: "openai-responses";
6201
6445
  provider: string;
6202
6446
  baseUrl: string;
6203
6447
  reasoning: true;
6204
- input: "text"[];
6448
+ input: ("image" | "text")[];
6205
6449
  cost: {
6206
6450
  input: number;
6207
6451
  output: number;
@@ -6211,14 +6455,14 @@ export declare const MODELS: {
6211
6455
  contextWindow: number;
6212
6456
  maxTokens: number;
6213
6457
  };
6214
- readonly "glm-5": {
6458
+ readonly "gpt-5-nano": {
6215
6459
  id: string;
6216
6460
  name: string;
6217
- api: "openai-completions";
6461
+ api: "openai-responses";
6218
6462
  provider: string;
6219
6463
  baseUrl: string;
6220
6464
  reasoning: true;
6221
- input: "text"[];
6465
+ input: ("image" | "text")[];
6222
6466
  cost: {
6223
6467
  input: number;
6224
6468
  output: number;
@@ -6228,7 +6472,7 @@ export declare const MODELS: {
6228
6472
  contextWindow: number;
6229
6473
  maxTokens: number;
6230
6474
  };
6231
- readonly "gpt-5": {
6475
+ readonly "gpt-5.1": {
6232
6476
  id: string;
6233
6477
  name: string;
6234
6478
  api: "openai-responses";
@@ -6245,7 +6489,7 @@ export declare const MODELS: {
6245
6489
  contextWindow: number;
6246
6490
  maxTokens: number;
6247
6491
  };
6248
- readonly "gpt-5-codex": {
6492
+ readonly "gpt-5.1-codex": {
6249
6493
  id: string;
6250
6494
  name: string;
6251
6495
  api: "openai-responses";
@@ -6262,7 +6506,7 @@ export declare const MODELS: {
6262
6506
  contextWindow: number;
6263
6507
  maxTokens: number;
6264
6508
  };
6265
- readonly "gpt-5-nano": {
6509
+ readonly "gpt-5.1-codex-max": {
6266
6510
  id: string;
6267
6511
  name: string;
6268
6512
  api: "openai-responses";
@@ -6279,7 +6523,7 @@ export declare const MODELS: {
6279
6523
  contextWindow: number;
6280
6524
  maxTokens: number;
6281
6525
  };
6282
- readonly "gpt-5.1": {
6526
+ readonly "gpt-5.1-codex-mini": {
6283
6527
  id: string;
6284
6528
  name: string;
6285
6529
  api: "openai-responses";
@@ -6296,7 +6540,7 @@ export declare const MODELS: {
6296
6540
  contextWindow: number;
6297
6541
  maxTokens: number;
6298
6542
  };
6299
- readonly "gpt-5.1-codex": {
6543
+ readonly "gpt-5.2": {
6300
6544
  id: string;
6301
6545
  name: string;
6302
6546
  api: "openai-responses";
@@ -6313,7 +6557,7 @@ export declare const MODELS: {
6313
6557
  contextWindow: number;
6314
6558
  maxTokens: number;
6315
6559
  };
6316
- readonly "gpt-5.1-codex-max": {
6560
+ readonly "gpt-5.2-codex": {
6317
6561
  id: string;
6318
6562
  name: string;
6319
6563
  api: "openai-responses";
@@ -6330,7 +6574,7 @@ export declare const MODELS: {
6330
6574
  contextWindow: number;
6331
6575
  maxTokens: number;
6332
6576
  };
6333
- readonly "gpt-5.1-codex-mini": {
6577
+ readonly "gpt-5.3-codex": {
6334
6578
  id: string;
6335
6579
  name: string;
6336
6580
  api: "openai-responses";
@@ -6347,7 +6591,7 @@ export declare const MODELS: {
6347
6591
  contextWindow: number;
6348
6592
  maxTokens: number;
6349
6593
  };
6350
- readonly "gpt-5.2": {
6594
+ readonly "gpt-5.4": {
6351
6595
  id: string;
6352
6596
  name: string;
6353
6597
  api: "openai-responses";
@@ -6364,7 +6608,7 @@ export declare const MODELS: {
6364
6608
  contextWindow: number;
6365
6609
  maxTokens: number;
6366
6610
  };
6367
- readonly "gpt-5.2-codex": {
6611
+ readonly "gpt-5.4-mini": {
6368
6612
  id: string;
6369
6613
  name: string;
6370
6614
  api: "openai-responses";
@@ -6381,7 +6625,7 @@ export declare const MODELS: {
6381
6625
  contextWindow: number;
6382
6626
  maxTokens: number;
6383
6627
  };
6384
- readonly "gpt-5.3-codex": {
6628
+ readonly "gpt-5.4-nano": {
6385
6629
  id: string;
6386
6630
  name: string;
6387
6631
  api: "openai-responses";
@@ -6398,7 +6642,7 @@ export declare const MODELS: {
6398
6642
  contextWindow: number;
6399
6643
  maxTokens: number;
6400
6644
  };
6401
- readonly "gpt-5.4": {
6645
+ readonly "gpt-5.4-pro": {
6402
6646
  id: string;
6403
6647
  name: string;
6404
6648
  api: "openai-responses";
@@ -6415,10 +6659,10 @@ export declare const MODELS: {
6415
6659
  contextWindow: number;
6416
6660
  maxTokens: number;
6417
6661
  };
6418
- readonly "gpt-5.4-pro": {
6662
+ readonly "kimi-k2.5": {
6419
6663
  id: string;
6420
6664
  name: string;
6421
- api: "openai-responses";
6665
+ api: "openai-completions";
6422
6666
  provider: string;
6423
6667
  baseUrl: string;
6424
6668
  reasoning: true;
@@ -6432,7 +6676,7 @@ export declare const MODELS: {
6432
6676
  contextWindow: number;
6433
6677
  maxTokens: number;
6434
6678
  };
6435
- readonly "kimi-k2.5": {
6679
+ readonly "mimo-v2-omni-free": {
6436
6680
  id: string;
6437
6681
  name: string;
6438
6682
  api: "openai-completions";
@@ -6449,7 +6693,7 @@ export declare const MODELS: {
6449
6693
  contextWindow: number;
6450
6694
  maxTokens: number;
6451
6695
  };
6452
- readonly "mimo-v2-flash-free": {
6696
+ readonly "mimo-v2-pro-free": {
6453
6697
  id: string;
6454
6698
  name: string;
6455
6699
  api: "openai-completions";
@@ -6466,7 +6710,7 @@ export declare const MODELS: {
6466
6710
  contextWindow: number;
6467
6711
  maxTokens: number;
6468
6712
  };
6469
- readonly "minimax-m2.1": {
6713
+ readonly "minimax-m2.5": {
6470
6714
  id: string;
6471
6715
  name: string;
6472
6716
  api: "openai-completions";
@@ -6483,10 +6727,10 @@ export declare const MODELS: {
6483
6727
  contextWindow: number;
6484
6728
  maxTokens: number;
6485
6729
  };
6486
- readonly "minimax-m2.5": {
6730
+ readonly "minimax-m2.5-free": {
6487
6731
  id: string;
6488
6732
  name: string;
6489
- api: "openai-completions";
6733
+ api: "anthropic-messages";
6490
6734
  provider: string;
6491
6735
  baseUrl: string;
6492
6736
  reasoning: true;
@@ -6500,10 +6744,10 @@ export declare const MODELS: {
6500
6744
  contextWindow: number;
6501
6745
  maxTokens: number;
6502
6746
  };
6503
- readonly "minimax-m2.5-free": {
6747
+ readonly "nemotron-3-super-free": {
6504
6748
  id: string;
6505
6749
  name: string;
6506
- api: "anthropic-messages";
6750
+ api: "openai-completions";
6507
6751
  provider: string;
6508
6752
  baseUrl: string;
6509
6753
  reasoning: true;
@@ -6517,7 +6761,9 @@ export declare const MODELS: {
6517
6761
  contextWindow: number;
6518
6762
  maxTokens: number;
6519
6763
  };
6520
- readonly "nemotron-3-super-free": {
6764
+ };
6765
+ readonly "opencode-go": {
6766
+ readonly "glm-5": {
6521
6767
  id: string;
6522
6768
  name: string;
6523
6769
  api: "openai-completions";
@@ -6534,16 +6780,14 @@ export declare const MODELS: {
6534
6780
  contextWindow: number;
6535
6781
  maxTokens: number;
6536
6782
  };
6537
- };
6538
- readonly "opencode-go": {
6539
- readonly "glm-5": {
6783
+ readonly "kimi-k2.5": {
6540
6784
  id: string;
6541
6785
  name: string;
6542
6786
  api: "openai-completions";
6543
6787
  provider: string;
6544
6788
  baseUrl: string;
6545
6789
  reasoning: true;
6546
- input: "text"[];
6790
+ input: ("image" | "text")[];
6547
6791
  cost: {
6548
6792
  input: number;
6549
6793
  output: number;
@@ -6553,14 +6797,14 @@ export declare const MODELS: {
6553
6797
  contextWindow: number;
6554
6798
  maxTokens: number;
6555
6799
  };
6556
- readonly "kimi-k2.5": {
6800
+ readonly "minimax-m2.5": {
6557
6801
  id: string;
6558
6802
  name: string;
6559
- api: "openai-completions";
6803
+ api: "anthropic-messages";
6560
6804
  provider: string;
6561
6805
  baseUrl: string;
6562
6806
  reasoning: true;
6563
- input: ("image" | "text")[];
6807
+ input: "text"[];
6564
6808
  cost: {
6565
6809
  input: number;
6566
6810
  output: number;
@@ -6570,7 +6814,7 @@ export declare const MODELS: {
6570
6814
  contextWindow: number;
6571
6815
  maxTokens: number;
6572
6816
  };
6573
- readonly "minimax-m2.5": {
6817
+ readonly "minimax-m2.7": {
6574
6818
  id: string;
6575
6819
  name: string;
6576
6820
  api: "anthropic-messages";
@@ -7541,14 +7785,14 @@ export declare const MODELS: {
7541
7785
  contextWindow: number;
7542
7786
  maxTokens: number;
7543
7787
  };
7544
- readonly "google/gemma-3-27b-it": {
7788
+ readonly "inception/mercury": {
7545
7789
  id: string;
7546
7790
  name: string;
7547
7791
  api: "openai-completions";
7548
7792
  provider: string;
7549
7793
  baseUrl: string;
7550
7794
  reasoning: false;
7551
- input: ("image" | "text")[];
7795
+ input: "text"[];
7552
7796
  cost: {
7553
7797
  input: number;
7554
7798
  output: number;
@@ -7558,14 +7802,14 @@ export declare const MODELS: {
7558
7802
  contextWindow: number;
7559
7803
  maxTokens: number;
7560
7804
  };
7561
- readonly "google/gemma-3-27b-it:free": {
7805
+ readonly "inception/mercury-2": {
7562
7806
  id: string;
7563
7807
  name: string;
7564
7808
  api: "openai-completions";
7565
7809
  provider: string;
7566
7810
  baseUrl: string;
7567
- reasoning: false;
7568
- input: ("image" | "text")[];
7811
+ reasoning: true;
7812
+ input: "text"[];
7569
7813
  cost: {
7570
7814
  input: number;
7571
7815
  output: number;
@@ -7575,7 +7819,7 @@ export declare const MODELS: {
7575
7819
  contextWindow: number;
7576
7820
  maxTokens: number;
7577
7821
  };
7578
- readonly "inception/mercury": {
7822
+ readonly "inception/mercury-coder": {
7579
7823
  id: string;
7580
7824
  name: string;
7581
7825
  api: "openai-completions";
@@ -7592,13 +7836,13 @@ export declare const MODELS: {
7592
7836
  contextWindow: number;
7593
7837
  maxTokens: number;
7594
7838
  };
7595
- readonly "inception/mercury-2": {
7839
+ readonly "kwaipilot/kat-coder-pro": {
7596
7840
  id: string;
7597
7841
  name: string;
7598
7842
  api: "openai-completions";
7599
7843
  provider: string;
7600
7844
  baseUrl: string;
7601
- reasoning: true;
7845
+ reasoning: false;
7602
7846
  input: "text"[];
7603
7847
  cost: {
7604
7848
  input: number;
@@ -7609,7 +7853,7 @@ export declare const MODELS: {
7609
7853
  contextWindow: number;
7610
7854
  maxTokens: number;
7611
7855
  };
7612
- readonly "inception/mercury-coder": {
7856
+ readonly "meituan/longcat-flash-chat": {
7613
7857
  id: string;
7614
7858
  name: string;
7615
7859
  api: "openai-completions";
@@ -7626,7 +7870,7 @@ export declare const MODELS: {
7626
7870
  contextWindow: number;
7627
7871
  maxTokens: number;
7628
7872
  };
7629
- readonly "kwaipilot/kat-coder-pro": {
7873
+ readonly "meta-llama/llama-3-8b-instruct": {
7630
7874
  id: string;
7631
7875
  name: string;
7632
7876
  api: "openai-completions";
@@ -7643,7 +7887,7 @@ export declare const MODELS: {
7643
7887
  contextWindow: number;
7644
7888
  maxTokens: number;
7645
7889
  };
7646
- readonly "meituan/longcat-flash-chat": {
7890
+ readonly "meta-llama/llama-3.1-70b-instruct": {
7647
7891
  id: string;
7648
7892
  name: string;
7649
7893
  api: "openai-completions";
@@ -7660,7 +7904,7 @@ export declare const MODELS: {
7660
7904
  contextWindow: number;
7661
7905
  maxTokens: number;
7662
7906
  };
7663
- readonly "meta-llama/llama-3-8b-instruct": {
7907
+ readonly "meta-llama/llama-3.1-8b-instruct": {
7664
7908
  id: string;
7665
7909
  name: string;
7666
7910
  api: "openai-completions";
@@ -7677,7 +7921,7 @@ export declare const MODELS: {
7677
7921
  contextWindow: number;
7678
7922
  maxTokens: number;
7679
7923
  };
7680
- readonly "meta-llama/llama-3.1-70b-instruct": {
7924
+ readonly "meta-llama/llama-3.3-70b-instruct": {
7681
7925
  id: string;
7682
7926
  name: string;
7683
7927
  api: "openai-completions";
@@ -7694,7 +7938,7 @@ export declare const MODELS: {
7694
7938
  contextWindow: number;
7695
7939
  maxTokens: number;
7696
7940
  };
7697
- readonly "meta-llama/llama-3.1-8b-instruct": {
7941
+ readonly "meta-llama/llama-3.3-70b-instruct:free": {
7698
7942
  id: string;
7699
7943
  name: string;
7700
7944
  api: "openai-completions";
@@ -7711,14 +7955,14 @@ export declare const MODELS: {
7711
7955
  contextWindow: number;
7712
7956
  maxTokens: number;
7713
7957
  };
7714
- readonly "meta-llama/llama-3.3-70b-instruct": {
7958
+ readonly "meta-llama/llama-4-maverick": {
7715
7959
  id: string;
7716
7960
  name: string;
7717
7961
  api: "openai-completions";
7718
7962
  provider: string;
7719
7963
  baseUrl: string;
7720
7964
  reasoning: false;
7721
- input: "text"[];
7965
+ input: ("image" | "text")[];
7722
7966
  cost: {
7723
7967
  input: number;
7724
7968
  output: number;
@@ -7728,14 +7972,14 @@ export declare const MODELS: {
7728
7972
  contextWindow: number;
7729
7973
  maxTokens: number;
7730
7974
  };
7731
- readonly "meta-llama/llama-3.3-70b-instruct:free": {
7975
+ readonly "meta-llama/llama-4-scout": {
7732
7976
  id: string;
7733
7977
  name: string;
7734
7978
  api: "openai-completions";
7735
7979
  provider: string;
7736
7980
  baseUrl: string;
7737
7981
  reasoning: false;
7738
- input: "text"[];
7982
+ input: ("image" | "text")[];
7739
7983
  cost: {
7740
7984
  input: number;
7741
7985
  output: number;
@@ -7745,14 +7989,14 @@ export declare const MODELS: {
7745
7989
  contextWindow: number;
7746
7990
  maxTokens: number;
7747
7991
  };
7748
- readonly "meta-llama/llama-4-maverick": {
7992
+ readonly "minimax/minimax-m1": {
7749
7993
  id: string;
7750
7994
  name: string;
7751
7995
  api: "openai-completions";
7752
7996
  provider: string;
7753
7997
  baseUrl: string;
7754
- reasoning: false;
7755
- input: ("image" | "text")[];
7998
+ reasoning: true;
7999
+ input: "text"[];
7756
8000
  cost: {
7757
8001
  input: number;
7758
8002
  output: number;
@@ -7762,14 +8006,14 @@ export declare const MODELS: {
7762
8006
  contextWindow: number;
7763
8007
  maxTokens: number;
7764
8008
  };
7765
- readonly "meta-llama/llama-4-scout": {
8009
+ readonly "minimax/minimax-m2": {
7766
8010
  id: string;
7767
8011
  name: string;
7768
8012
  api: "openai-completions";
7769
8013
  provider: string;
7770
8014
  baseUrl: string;
7771
- reasoning: false;
7772
- input: ("image" | "text")[];
8015
+ reasoning: true;
8016
+ input: "text"[];
7773
8017
  cost: {
7774
8018
  input: number;
7775
8019
  output: number;
@@ -7779,7 +8023,7 @@ export declare const MODELS: {
7779
8023
  contextWindow: number;
7780
8024
  maxTokens: number;
7781
8025
  };
7782
- readonly "minimax/minimax-m1": {
8026
+ readonly "minimax/minimax-m2.1": {
7783
8027
  id: string;
7784
8028
  name: string;
7785
8029
  api: "openai-completions";
@@ -7796,7 +8040,7 @@ export declare const MODELS: {
7796
8040
  contextWindow: number;
7797
8041
  maxTokens: number;
7798
8042
  };
7799
- readonly "minimax/minimax-m2": {
8043
+ readonly "minimax/minimax-m2.5": {
7800
8044
  id: string;
7801
8045
  name: string;
7802
8046
  api: "openai-completions";
@@ -7813,7 +8057,7 @@ export declare const MODELS: {
7813
8057
  contextWindow: number;
7814
8058
  maxTokens: number;
7815
8059
  };
7816
- readonly "minimax/minimax-m2.1": {
8060
+ readonly "minimax/minimax-m2.5:free": {
7817
8061
  id: string;
7818
8062
  name: string;
7819
8063
  api: "openai-completions";
@@ -7830,7 +8074,7 @@ export declare const MODELS: {
7830
8074
  contextWindow: number;
7831
8075
  maxTokens: number;
7832
8076
  };
7833
- readonly "minimax/minimax-m2.5": {
8077
+ readonly "minimax/minimax-m2.7": {
7834
8078
  id: string;
7835
8079
  name: string;
7836
8080
  api: "openai-completions";
@@ -8119,6 +8363,23 @@ export declare const MODELS: {
8119
8363
  contextWindow: number;
8120
8364
  maxTokens: number;
8121
8365
  };
8366
+ readonly "mistralai/mistral-small-2603": {
8367
+ id: string;
8368
+ name: string;
8369
+ api: "openai-completions";
8370
+ provider: string;
8371
+ baseUrl: string;
8372
+ reasoning: true;
8373
+ input: ("image" | "text")[];
8374
+ cost: {
8375
+ input: number;
8376
+ output: number;
8377
+ cacheRead: number;
8378
+ cacheWrite: number;
8379
+ };
8380
+ contextWindow: number;
8381
+ maxTokens: number;
8382
+ };
8122
8383
  readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
8123
8384
  id: string;
8124
8385
  name: string;
@@ -8391,6 +8652,23 @@ export declare const MODELS: {
8391
8652
  contextWindow: number;
8392
8653
  maxTokens: number;
8393
8654
  };
8655
+ readonly "nvidia/nemotron-3-super-120b-a12b": {
8656
+ id: string;
8657
+ name: string;
8658
+ api: "openai-completions";
8659
+ provider: string;
8660
+ baseUrl: string;
8661
+ reasoning: true;
8662
+ input: "text"[];
8663
+ cost: {
8664
+ input: number;
8665
+ output: number;
8666
+ cacheRead: number;
8667
+ cacheWrite: number;
8668
+ };
8669
+ contextWindow: number;
8670
+ maxTokens: number;
8671
+ };
8394
8672
  readonly "nvidia/nemotron-3-super-120b-a12b:free": {
8395
8673
  id: string;
8396
8674
  name: string;
@@ -9105,7 +9383,7 @@ export declare const MODELS: {
9105
9383
  contextWindow: number;
9106
9384
  maxTokens: number;
9107
9385
  };
9108
- readonly "openai/gpt-5.4-pro": {
9386
+ readonly "openai/gpt-5.4-mini": {
9109
9387
  id: string;
9110
9388
  name: string;
9111
9389
  api: "openai-completions";
@@ -9122,14 +9400,14 @@ export declare const MODELS: {
9122
9400
  contextWindow: number;
9123
9401
  maxTokens: number;
9124
9402
  };
9125
- readonly "openai/gpt-oss-120b": {
9403
+ readonly "openai/gpt-5.4-nano": {
9126
9404
  id: string;
9127
9405
  name: string;
9128
9406
  api: "openai-completions";
9129
9407
  provider: string;
9130
9408
  baseUrl: string;
9131
9409
  reasoning: true;
9132
- input: "text"[];
9410
+ input: ("image" | "text")[];
9133
9411
  cost: {
9134
9412
  input: number;
9135
9413
  output: number;
@@ -9139,14 +9417,14 @@ export declare const MODELS: {
9139
9417
  contextWindow: number;
9140
9418
  maxTokens: number;
9141
9419
  };
9142
- readonly "openai/gpt-oss-120b:free": {
9420
+ readonly "openai/gpt-5.4-pro": {
9143
9421
  id: string;
9144
9422
  name: string;
9145
9423
  api: "openai-completions";
9146
9424
  provider: string;
9147
9425
  baseUrl: string;
9148
9426
  reasoning: true;
9149
- input: "text"[];
9427
+ input: ("image" | "text")[];
9150
9428
  cost: {
9151
9429
  input: number;
9152
9430
  output: number;
@@ -9156,7 +9434,7 @@ export declare const MODELS: {
9156
9434
  contextWindow: number;
9157
9435
  maxTokens: number;
9158
9436
  };
9159
- readonly "openai/gpt-oss-20b": {
9437
+ readonly "openai/gpt-oss-120b": {
9160
9438
  id: string;
9161
9439
  name: string;
9162
9440
  api: "openai-completions";
@@ -9173,7 +9451,7 @@ export declare const MODELS: {
9173
9451
  contextWindow: number;
9174
9452
  maxTokens: number;
9175
9453
  };
9176
- readonly "openai/gpt-oss-20b:free": {
9454
+ readonly "openai/gpt-oss-120b:free": {
9177
9455
  id: string;
9178
9456
  name: string;
9179
9457
  api: "openai-completions";
@@ -9190,7 +9468,7 @@ export declare const MODELS: {
9190
9468
  contextWindow: number;
9191
9469
  maxTokens: number;
9192
9470
  };
9193
- readonly "openai/gpt-oss-safeguard-20b": {
9471
+ readonly "openai/gpt-oss-20b": {
9194
9472
  id: string;
9195
9473
  name: string;
9196
9474
  api: "openai-completions";
@@ -9207,14 +9485,14 @@ export declare const MODELS: {
9207
9485
  contextWindow: number;
9208
9486
  maxTokens: number;
9209
9487
  };
9210
- readonly "openai/o1": {
9488
+ readonly "openai/gpt-oss-20b:free": {
9211
9489
  id: string;
9212
9490
  name: string;
9213
9491
  api: "openai-completions";
9214
9492
  provider: string;
9215
9493
  baseUrl: string;
9216
- reasoning: false;
9217
- input: ("image" | "text")[];
9494
+ reasoning: true;
9495
+ input: "text"[];
9218
9496
  cost: {
9219
9497
  input: number;
9220
9498
  output: number;
@@ -9224,14 +9502,14 @@ export declare const MODELS: {
9224
9502
  contextWindow: number;
9225
9503
  maxTokens: number;
9226
9504
  };
9227
- readonly "openai/o3": {
9505
+ readonly "openai/gpt-oss-safeguard-20b": {
9228
9506
  id: string;
9229
9507
  name: string;
9230
9508
  api: "openai-completions";
9231
9509
  provider: string;
9232
9510
  baseUrl: string;
9233
9511
  reasoning: true;
9234
- input: ("image" | "text")[];
9512
+ input: "text"[];
9235
9513
  cost: {
9236
9514
  input: number;
9237
9515
  output: number;
@@ -9241,7 +9519,7 @@ export declare const MODELS: {
9241
9519
  contextWindow: number;
9242
9520
  maxTokens: number;
9243
9521
  };
9244
- readonly "openai/o3-deep-research": {
9522
+ readonly "openai/o1": {
9245
9523
  id: string;
9246
9524
  name: string;
9247
9525
  api: "openai-completions";
@@ -9258,14 +9536,14 @@ export declare const MODELS: {
9258
9536
  contextWindow: number;
9259
9537
  maxTokens: number;
9260
9538
  };
9261
- readonly "openai/o3-mini": {
9539
+ readonly "openai/o3": {
9262
9540
  id: string;
9263
9541
  name: string;
9264
9542
  api: "openai-completions";
9265
9543
  provider: string;
9266
9544
  baseUrl: string;
9267
- reasoning: false;
9268
- input: "text"[];
9545
+ reasoning: true;
9546
+ input: ("image" | "text")[];
9269
9547
  cost: {
9270
9548
  input: number;
9271
9549
  output: number;
@@ -9275,14 +9553,14 @@ export declare const MODELS: {
9275
9553
  contextWindow: number;
9276
9554
  maxTokens: number;
9277
9555
  };
9278
- readonly "openai/o3-mini-high": {
9556
+ readonly "openai/o3-deep-research": {
9279
9557
  id: string;
9280
9558
  name: string;
9281
9559
  api: "openai-completions";
9282
9560
  provider: string;
9283
9561
  baseUrl: string;
9284
- reasoning: false;
9285
- input: "text"[];
9562
+ reasoning: true;
9563
+ input: ("image" | "text")[];
9286
9564
  cost: {
9287
9565
  input: number;
9288
9566
  output: number;
@@ -9292,14 +9570,14 @@ export declare const MODELS: {
9292
9570
  contextWindow: number;
9293
9571
  maxTokens: number;
9294
9572
  };
9295
- readonly "openai/o3-pro": {
9573
+ readonly "openai/o3-mini": {
9296
9574
  id: string;
9297
9575
  name: string;
9298
9576
  api: "openai-completions";
9299
9577
  provider: string;
9300
9578
  baseUrl: string;
9301
9579
  reasoning: true;
9302
- input: ("image" | "text")[];
9580
+ input: "text"[];
9303
9581
  cost: {
9304
9582
  input: number;
9305
9583
  output: number;
@@ -9309,14 +9587,14 @@ export declare const MODELS: {
9309
9587
  contextWindow: number;
9310
9588
  maxTokens: number;
9311
9589
  };
9312
- readonly "openai/o4-mini": {
9590
+ readonly "openai/o3-mini-high": {
9313
9591
  id: string;
9314
9592
  name: string;
9315
9593
  api: "openai-completions";
9316
9594
  provider: string;
9317
9595
  baseUrl: string;
9318
9596
  reasoning: true;
9319
- input: ("image" | "text")[];
9597
+ input: "text"[];
9320
9598
  cost: {
9321
9599
  input: number;
9322
9600
  output: number;
@@ -9326,7 +9604,7 @@ export declare const MODELS: {
9326
9604
  contextWindow: number;
9327
9605
  maxTokens: number;
9328
9606
  };
9329
- readonly "openai/o4-mini-deep-research": {
9607
+ readonly "openai/o3-pro": {
9330
9608
  id: string;
9331
9609
  name: string;
9332
9610
  api: "openai-completions";
@@ -9343,7 +9621,7 @@ export declare const MODELS: {
9343
9621
  contextWindow: number;
9344
9622
  maxTokens: number;
9345
9623
  };
9346
- readonly "openai/o4-mini-high": {
9624
+ readonly "openai/o4-mini": {
9347
9625
  id: string;
9348
9626
  name: string;
9349
9627
  api: "openai-completions";
@@ -9360,7 +9638,7 @@ export declare const MODELS: {
9360
9638
  contextWindow: number;
9361
9639
  maxTokens: number;
9362
9640
  };
9363
- readonly "openrouter/auto": {
9641
+ readonly "openai/o4-mini-deep-research": {
9364
9642
  id: string;
9365
9643
  name: string;
9366
9644
  api: "openai-completions";
@@ -9377,7 +9655,7 @@ export declare const MODELS: {
9377
9655
  contextWindow: number;
9378
9656
  maxTokens: number;
9379
9657
  };
9380
- readonly "openrouter/free": {
9658
+ readonly "openai/o4-mini-high": {
9381
9659
  id: string;
9382
9660
  name: string;
9383
9661
  api: "openai-completions";
@@ -9394,7 +9672,7 @@ export declare const MODELS: {
9394
9672
  contextWindow: number;
9395
9673
  maxTokens: number;
9396
9674
  };
9397
- readonly "openrouter/healer-alpha": {
9675
+ readonly "openrouter/auto": {
9398
9676
  id: string;
9399
9677
  name: string;
9400
9678
  api: "openai-completions";
@@ -9411,14 +9689,14 @@ export declare const MODELS: {
9411
9689
  contextWindow: number;
9412
9690
  maxTokens: number;
9413
9691
  };
9414
- readonly "openrouter/hunter-alpha": {
9692
+ readonly "openrouter/free": {
9415
9693
  id: string;
9416
9694
  name: string;
9417
9695
  api: "openai-completions";
9418
9696
  provider: string;
9419
9697
  baseUrl: string;
9420
9698
  reasoning: true;
9421
- input: "text"[];
9699
+ input: ("image" | "text")[];
9422
9700
  cost: {
9423
9701
  input: number;
9424
9702
  output: number;
@@ -10516,6 +10794,40 @@ export declare const MODELS: {
10516
10794
  contextWindow: number;
10517
10795
  maxTokens: number;
10518
10796
  };
10797
+ readonly "xiaomi/mimo-v2-omni": {
10798
+ id: string;
10799
+ name: string;
10800
+ api: "openai-completions";
10801
+ provider: string;
10802
+ baseUrl: string;
10803
+ reasoning: true;
10804
+ input: ("image" | "text")[];
10805
+ cost: {
10806
+ input: number;
10807
+ output: number;
10808
+ cacheRead: number;
10809
+ cacheWrite: number;
10810
+ };
10811
+ contextWindow: number;
10812
+ maxTokens: number;
10813
+ };
10814
+ readonly "xiaomi/mimo-v2-pro": {
10815
+ id: string;
10816
+ name: string;
10817
+ api: "openai-completions";
10818
+ provider: string;
10819
+ baseUrl: string;
10820
+ reasoning: true;
10821
+ input: "text"[];
10822
+ cost: {
10823
+ input: number;
10824
+ output: number;
10825
+ cacheRead: number;
10826
+ cacheWrite: number;
10827
+ };
10828
+ contextWindow: number;
10829
+ maxTokens: number;
10830
+ };
10519
10831
  readonly "z-ai/glm-4-32b": {
10520
10832
  id: string;
10521
10833
  name: string;
@@ -10686,6 +10998,23 @@ export declare const MODELS: {
10686
10998
  contextWindow: number;
10687
10999
  maxTokens: number;
10688
11000
  };
11001
+ readonly "z-ai/glm-5-turbo": {
11002
+ id: string;
11003
+ name: string;
11004
+ api: "openai-completions";
11005
+ provider: string;
11006
+ baseUrl: string;
11007
+ reasoning: true;
11008
+ input: "text"[];
11009
+ cost: {
11010
+ input: number;
11011
+ output: number;
11012
+ cacheRead: number;
11013
+ cacheWrite: number;
11014
+ };
11015
+ contextWindow: number;
11016
+ maxTokens: number;
11017
+ };
10689
11018
  };
10690
11019
  readonly "vercel-ai-gateway": {
10691
11020
  readonly "alibaba/qwen-3-14b": {
@@ -11385,40 +11714,6 @@ export declare const MODELS: {
11385
11714
  contextWindow: number;
11386
11715
  maxTokens: number;
11387
11716
  };
11388
- readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
11389
- id: string;
11390
- name: string;
11391
- api: "anthropic-messages";
11392
- provider: string;
11393
- baseUrl: string;
11394
- reasoning: true;
11395
- input: ("image" | "text")[];
11396
- cost: {
11397
- input: number;
11398
- output: number;
11399
- cacheRead: number;
11400
- cacheWrite: number;
11401
- };
11402
- contextWindow: number;
11403
- maxTokens: number;
11404
- };
11405
- readonly "google/gemini-2.5-flash-preview-09-2025": {
11406
- id: string;
11407
- name: string;
11408
- api: "anthropic-messages";
11409
- provider: string;
11410
- baseUrl: string;
11411
- reasoning: true;
11412
- input: ("image" | "text")[];
11413
- cost: {
11414
- input: number;
11415
- output: number;
11416
- cacheRead: number;
11417
- cacheWrite: number;
11418
- };
11419
- contextWindow: number;
11420
- maxTokens: number;
11421
- };
11422
11717
  readonly "google/gemini-2.5-pro": {
11423
11718
  id: string;
11424
11719
  name: string;
@@ -11776,6 +12071,40 @@ export declare const MODELS: {
11776
12071
  contextWindow: number;
11777
12072
  maxTokens: number;
11778
12073
  };
12074
+ readonly "minimax/minimax-m2.7": {
12075
+ id: string;
12076
+ name: string;
12077
+ api: "anthropic-messages";
12078
+ provider: string;
12079
+ baseUrl: string;
12080
+ reasoning: true;
12081
+ input: ("image" | "text")[];
12082
+ cost: {
12083
+ input: number;
12084
+ output: number;
12085
+ cacheRead: number;
12086
+ cacheWrite: number;
12087
+ };
12088
+ contextWindow: number;
12089
+ maxTokens: number;
12090
+ };
12091
+ readonly "minimax/minimax-m2.7-highspeed": {
12092
+ id: string;
12093
+ name: string;
12094
+ api: "anthropic-messages";
12095
+ provider: string;
12096
+ baseUrl: string;
12097
+ reasoning: true;
12098
+ input: ("image" | "text")[];
12099
+ cost: {
12100
+ input: number;
12101
+ output: number;
12102
+ cacheRead: number;
12103
+ cacheWrite: number;
12104
+ };
12105
+ contextWindow: number;
12106
+ maxTokens: number;
12107
+ };
11779
12108
  readonly "mistral/codestral": {
11780
12109
  id: string;
11781
12110
  name: string;
@@ -12490,6 +12819,40 @@ export declare const MODELS: {
12490
12819
  contextWindow: number;
12491
12820
  maxTokens: number;
12492
12821
  };
12822
+ readonly "openai/gpt-5.4-mini": {
12823
+ id: string;
12824
+ name: string;
12825
+ api: "anthropic-messages";
12826
+ provider: string;
12827
+ baseUrl: string;
12828
+ reasoning: true;
12829
+ input: ("image" | "text")[];
12830
+ cost: {
12831
+ input: number;
12832
+ output: number;
12833
+ cacheRead: number;
12834
+ cacheWrite: number;
12835
+ };
12836
+ contextWindow: number;
12837
+ maxTokens: number;
12838
+ };
12839
+ readonly "openai/gpt-5.4-nano": {
12840
+ id: string;
12841
+ name: string;
12842
+ api: "anthropic-messages";
12843
+ provider: string;
12844
+ baseUrl: string;
12845
+ reasoning: true;
12846
+ input: ("image" | "text")[];
12847
+ cost: {
12848
+ input: number;
12849
+ output: number;
12850
+ cacheRead: number;
12851
+ cacheWrite: number;
12852
+ };
12853
+ contextWindow: number;
12854
+ maxTokens: number;
12855
+ };
12493
12856
  readonly "openai/gpt-5.4-pro": {
12494
12857
  id: string;
12495
12858
  name: string;
@@ -12694,7 +13057,7 @@ export declare const MODELS: {
12694
13057
  contextWindow: number;
12695
13058
  maxTokens: number;
12696
13059
  };
12697
- readonly "vercel/v0-1.0-md": {
13060
+ readonly "xai/grok-2-vision": {
12698
13061
  id: string;
12699
13062
  name: string;
12700
13063
  api: "anthropic-messages";
@@ -12711,14 +13074,14 @@ export declare const MODELS: {
12711
13074
  contextWindow: number;
12712
13075
  maxTokens: number;
12713
13076
  };
12714
- readonly "vercel/v0-1.5-md": {
13077
+ readonly "xai/grok-3": {
12715
13078
  id: string;
12716
13079
  name: string;
12717
13080
  api: "anthropic-messages";
12718
13081
  provider: string;
12719
13082
  baseUrl: string;
12720
13083
  reasoning: false;
12721
- input: ("image" | "text")[];
13084
+ input: "text"[];
12722
13085
  cost: {
12723
13086
  input: number;
12724
13087
  output: number;
@@ -12728,14 +13091,14 @@ export declare const MODELS: {
12728
13091
  contextWindow: number;
12729
13092
  maxTokens: number;
12730
13093
  };
12731
- readonly "xai/grok-2-vision": {
13094
+ readonly "xai/grok-3-fast": {
12732
13095
  id: string;
12733
13096
  name: string;
12734
13097
  api: "anthropic-messages";
12735
13098
  provider: string;
12736
13099
  baseUrl: string;
12737
13100
  reasoning: false;
12738
- input: ("image" | "text")[];
13101
+ input: "text"[];
12739
13102
  cost: {
12740
13103
  input: number;
12741
13104
  output: number;
@@ -12745,7 +13108,7 @@ export declare const MODELS: {
12745
13108
  contextWindow: number;
12746
13109
  maxTokens: number;
12747
13110
  };
12748
- readonly "xai/grok-3": {
13111
+ readonly "xai/grok-3-mini": {
12749
13112
  id: string;
12750
13113
  name: string;
12751
13114
  api: "anthropic-messages";
@@ -12762,7 +13125,7 @@ export declare const MODELS: {
12762
13125
  contextWindow: number;
12763
13126
  maxTokens: number;
12764
13127
  };
12765
- readonly "xai/grok-3-fast": {
13128
+ readonly "xai/grok-3-mini-fast": {
12766
13129
  id: string;
12767
13130
  name: string;
12768
13131
  api: "anthropic-messages";
@@ -12779,14 +13142,14 @@ export declare const MODELS: {
12779
13142
  contextWindow: number;
12780
13143
  maxTokens: number;
12781
13144
  };
12782
- readonly "xai/grok-3-mini": {
13145
+ readonly "xai/grok-4": {
12783
13146
  id: string;
12784
13147
  name: string;
12785
13148
  api: "anthropic-messages";
12786
13149
  provider: string;
12787
13150
  baseUrl: string;
12788
- reasoning: false;
12789
- input: "text"[];
13151
+ reasoning: true;
13152
+ input: ("image" | "text")[];
12790
13153
  cost: {
12791
13154
  input: number;
12792
13155
  output: number;
@@ -12796,7 +13159,7 @@ export declare const MODELS: {
12796
13159
  contextWindow: number;
12797
13160
  maxTokens: number;
12798
13161
  };
12799
- readonly "xai/grok-3-mini-fast": {
13162
+ readonly "xai/grok-4-fast-non-reasoning": {
12800
13163
  id: string;
12801
13164
  name: string;
12802
13165
  api: "anthropic-messages";
@@ -12813,14 +13176,14 @@ export declare const MODELS: {
12813
13176
  contextWindow: number;
12814
13177
  maxTokens: number;
12815
13178
  };
12816
- readonly "xai/grok-4": {
13179
+ readonly "xai/grok-4-fast-reasoning": {
12817
13180
  id: string;
12818
13181
  name: string;
12819
13182
  api: "anthropic-messages";
12820
13183
  provider: string;
12821
13184
  baseUrl: string;
12822
13185
  reasoning: true;
12823
- input: ("image" | "text")[];
13186
+ input: "text"[];
12824
13187
  cost: {
12825
13188
  input: number;
12826
13189
  output: number;
@@ -12830,7 +13193,7 @@ export declare const MODELS: {
12830
13193
  contextWindow: number;
12831
13194
  maxTokens: number;
12832
13195
  };
12833
- readonly "xai/grok-4-fast-non-reasoning": {
13196
+ readonly "xai/grok-4.1-fast-non-reasoning": {
12834
13197
  id: string;
12835
13198
  name: string;
12836
13199
  api: "anthropic-messages";
@@ -12847,7 +13210,7 @@ export declare const MODELS: {
12847
13210
  contextWindow: number;
12848
13211
  maxTokens: number;
12849
13212
  };
12850
- readonly "xai/grok-4-fast-reasoning": {
13213
+ readonly "xai/grok-4.1-fast-reasoning": {
12851
13214
  id: string;
12852
13215
  name: string;
12853
13216
  api: "anthropic-messages";
@@ -12864,13 +13227,13 @@ export declare const MODELS: {
12864
13227
  contextWindow: number;
12865
13228
  maxTokens: number;
12866
13229
  };
12867
- readonly "xai/grok-4.1-fast-non-reasoning": {
13230
+ readonly "xai/grok-4.20-multi-agent": {
12868
13231
  id: string;
12869
13232
  name: string;
12870
13233
  api: "anthropic-messages";
12871
13234
  provider: string;
12872
13235
  baseUrl: string;
12873
- reasoning: false;
13236
+ reasoning: true;
12874
13237
  input: "text"[];
12875
13238
  cost: {
12876
13239
  input: number;
@@ -12881,7 +13244,7 @@ export declare const MODELS: {
12881
13244
  contextWindow: number;
12882
13245
  maxTokens: number;
12883
13246
  };
12884
- readonly "xai/grok-4.1-fast-reasoning": {
13247
+ readonly "xai/grok-4.20-multi-agent-beta": {
12885
13248
  id: string;
12886
13249
  name: string;
12887
13250
  api: "anthropic-messages";
@@ -12898,14 +13261,14 @@ export declare const MODELS: {
12898
13261
  contextWindow: number;
12899
13262
  maxTokens: number;
12900
13263
  };
12901
- readonly "xai/grok-4.20-multi-agent-beta": {
13264
+ readonly "xai/grok-4.20-non-reasoning": {
12902
13265
  id: string;
12903
13266
  name: string;
12904
13267
  api: "anthropic-messages";
12905
13268
  provider: string;
12906
13269
  baseUrl: string;
12907
- reasoning: true;
12908
- input: "text"[];
13270
+ reasoning: false;
13271
+ input: ("image" | "text")[];
12909
13272
  cost: {
12910
13273
  input: number;
12911
13274
  output: number;
@@ -12932,6 +13295,23 @@ export declare const MODELS: {
12932
13295
  contextWindow: number;
12933
13296
  maxTokens: number;
12934
13297
  };
13298
+ readonly "xai/grok-4.20-reasoning": {
13299
+ id: string;
13300
+ name: string;
13301
+ api: "anthropic-messages";
13302
+ provider: string;
13303
+ baseUrl: string;
13304
+ reasoning: true;
13305
+ input: ("image" | "text")[];
13306
+ cost: {
13307
+ input: number;
13308
+ output: number;
13309
+ cacheRead: number;
13310
+ cacheWrite: number;
13311
+ };
13312
+ contextWindow: number;
13313
+ maxTokens: number;
13314
+ };
12935
13315
  readonly "xai/grok-4.20-reasoning-beta": {
12936
13316
  id: string;
12937
13317
  name: string;
@@ -12983,6 +13363,23 @@ export declare const MODELS: {
12983
13363
  contextWindow: number;
12984
13364
  maxTokens: number;
12985
13365
  };
13366
+ readonly "xiaomi/mimo-v2-pro": {
13367
+ id: string;
13368
+ name: string;
13369
+ api: "anthropic-messages";
13370
+ provider: string;
13371
+ baseUrl: string;
13372
+ reasoning: true;
13373
+ input: "text"[];
13374
+ cost: {
13375
+ input: number;
13376
+ output: number;
13377
+ cacheRead: number;
13378
+ cacheWrite: number;
13379
+ };
13380
+ contextWindow: number;
13381
+ maxTokens: number;
13382
+ };
12986
13383
  readonly "zai/glm-4.5": {
12987
13384
  id: string;
12988
13385
  name: string;
@@ -13153,6 +13550,23 @@ export declare const MODELS: {
13153
13550
  contextWindow: number;
13154
13551
  maxTokens: number;
13155
13552
  };
13553
+ readonly "zai/glm-5-turbo": {
13554
+ id: string;
13555
+ name: string;
13556
+ api: "anthropic-messages";
13557
+ provider: string;
13558
+ baseUrl: string;
13559
+ reasoning: true;
13560
+ input: "text"[];
13561
+ cost: {
13562
+ input: number;
13563
+ output: number;
13564
+ cacheRead: number;
13565
+ cacheWrite: number;
13566
+ };
13567
+ contextWindow: number;
13568
+ maxTokens: number;
13569
+ };
13156
13570
  };
13157
13571
  readonly xai: {
13158
13572
  readonly "grok-2": {
@@ -13478,7 +13892,7 @@ export declare const MODELS: {
13478
13892
  contextWindow: number;
13479
13893
  maxTokens: number;
13480
13894
  };
13481
- readonly "grok-4.20-beta-latest-non-reasoning": {
13895
+ readonly "grok-4.20-0309-non-reasoning": {
13482
13896
  id: string;
13483
13897
  name: string;
13484
13898
  api: "openai-completions";
@@ -13495,24 +13909,7 @@ export declare const MODELS: {
13495
13909
  contextWindow: number;
13496
13910
  maxTokens: number;
13497
13911
  };
13498
- readonly "grok-4.20-beta-latest-reasoning": {
13499
- id: string;
13500
- name: string;
13501
- api: "openai-completions";
13502
- provider: string;
13503
- baseUrl: string;
13504
- reasoning: true;
13505
- input: ("image" | "text")[];
13506
- cost: {
13507
- input: number;
13508
- output: number;
13509
- cacheRead: number;
13510
- cacheWrite: number;
13511
- };
13512
- contextWindow: number;
13513
- maxTokens: number;
13514
- };
13515
- readonly "grok-4.20-multi-agent-beta-latest": {
13912
+ readonly "grok-4.20-0309-reasoning": {
13516
13913
  id: string;
13517
13914
  name: string;
13518
13915
  api: "openai-completions";
@@ -13771,6 +14168,27 @@ export declare const MODELS: {
13771
14168
  contextWindow: number;
13772
14169
  maxTokens: number;
13773
14170
  };
14171
+ readonly "glm-5-turbo": {
14172
+ id: string;
14173
+ name: string;
14174
+ api: "openai-completions";
14175
+ provider: string;
14176
+ baseUrl: string;
14177
+ compat: {
14178
+ supportsDeveloperRole: false;
14179
+ thinkingFormat: "zai";
14180
+ };
14181
+ reasoning: true;
14182
+ input: "text"[];
14183
+ cost: {
14184
+ input: number;
14185
+ output: number;
14186
+ cacheRead: number;
14187
+ cacheWrite: number;
14188
+ };
14189
+ contextWindow: number;
14190
+ maxTokens: number;
14191
+ };
13774
14192
  };
13775
14193
  };
13776
14194
  //# sourceMappingURL=models.generated.d.ts.map