@draht/ai 2026.3.25 → 2026.4.23

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (51) hide show
  1. package/README.md +107 -12
  2. package/dist/cli.d.ts.map +1 -1
  3. package/dist/cli.js +1 -1
  4. package/dist/cli.js.map +1 -1
  5. package/dist/index.d.ts +1 -0
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -0
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.generated.d.ts +1163 -498
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +1392 -756
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +37 -9
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +38 -16
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/faux.d.ts +56 -0
  20. package/dist/providers/faux.d.ts.map +1 -0
  21. package/dist/providers/faux.js +367 -0
  22. package/dist/providers/faux.js.map +1 -0
  23. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  24. package/dist/providers/google-gemini-cli.js +2 -2
  25. package/dist/providers/google-gemini-cli.js.map +1 -1
  26. package/dist/providers/google-vertex.d.ts.map +1 -1
  27. package/dist/providers/google-vertex.js +1 -1
  28. package/dist/providers/google-vertex.js.map +1 -1
  29. package/dist/providers/google.d.ts.map +1 -1
  30. package/dist/providers/google.js +29 -4
  31. package/dist/providers/google.js.map +1 -1
  32. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  33. package/dist/providers/openai-codex-responses.js +1 -1
  34. package/dist/providers/openai-codex-responses.js.map +1 -1
  35. package/dist/providers/openai-completions.d.ts.map +1 -1
  36. package/dist/providers/openai-completions.js +18 -6
  37. package/dist/providers/openai-completions.js.map +1 -1
  38. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  39. package/dist/providers/openai-responses-shared.js +20 -1
  40. package/dist/providers/openai-responses-shared.js.map +1 -1
  41. package/dist/providers/openai-responses.d.ts.map +1 -1
  42. package/dist/providers/openai-responses.js +2 -2
  43. package/dist/providers/openai-responses.js.map +1 -1
  44. package/dist/types.d.ts +64 -5
  45. package/dist/types.d.ts.map +1 -1
  46. package/dist/types.js.map +1 -1
  47. package/dist/utils/overflow.d.ts +4 -3
  48. package/dist/utils/overflow.d.ts.map +1 -1
  49. package/dist/utils/overflow.js +28 -13
  50. package/dist/utils/overflow.js.map +1 -1
  51. package/package.json +5 -8
@@ -255,6 +255,23 @@ export declare const MODELS: {
255
255
  contextWindow: number;
256
256
  maxTokens: number;
257
257
  };
258
+ readonly "anthropic.claude-opus-4-7": {
259
+ id: string;
260
+ name: string;
261
+ api: "bedrock-converse-stream";
262
+ provider: string;
263
+ baseUrl: string;
264
+ reasoning: true;
265
+ input: ("image" | "text")[];
266
+ cost: {
267
+ input: number;
268
+ output: number;
269
+ cacheRead: number;
270
+ cacheWrite: number;
271
+ };
272
+ contextWindow: number;
273
+ maxTokens: number;
274
+ };
258
275
  readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
259
276
  id: string;
260
277
  name: string;
@@ -408,6 +425,23 @@ export declare const MODELS: {
408
425
  contextWindow: number;
409
426
  maxTokens: number;
410
427
  };
428
+ readonly "eu.anthropic.claude-opus-4-7": {
429
+ id: string;
430
+ name: string;
431
+ api: "bedrock-converse-stream";
432
+ provider: string;
433
+ baseUrl: string;
434
+ reasoning: true;
435
+ input: ("image" | "text")[];
436
+ cost: {
437
+ input: number;
438
+ output: number;
439
+ cacheRead: number;
440
+ cacheWrite: number;
441
+ };
442
+ contextWindow: number;
443
+ maxTokens: number;
444
+ };
411
445
  readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
412
446
  id: string;
413
447
  name: string;
@@ -510,6 +544,23 @@ export declare const MODELS: {
510
544
  contextWindow: number;
511
545
  maxTokens: number;
512
546
  };
547
+ readonly "global.anthropic.claude-opus-4-7": {
548
+ id: string;
549
+ name: string;
550
+ api: "bedrock-converse-stream";
551
+ provider: string;
552
+ baseUrl: string;
553
+ reasoning: true;
554
+ input: ("image" | "text")[];
555
+ cost: {
556
+ input: number;
557
+ output: number;
558
+ cacheRead: number;
559
+ cacheWrite: number;
560
+ };
561
+ contextWindow: number;
562
+ maxTokens: number;
563
+ };
513
564
  readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
514
565
  id: string;
515
566
  name: string;
@@ -1054,6 +1105,23 @@ export declare const MODELS: {
1054
1105
  contextWindow: number;
1055
1106
  maxTokens: number;
1056
1107
  };
1108
+ readonly "nvidia.nemotron-super-3-120b": {
1109
+ id: string;
1110
+ name: string;
1111
+ api: "bedrock-converse-stream";
1112
+ provider: string;
1113
+ baseUrl: string;
1114
+ reasoning: true;
1115
+ input: "text"[];
1116
+ cost: {
1117
+ input: number;
1118
+ output: number;
1119
+ cacheRead: number;
1120
+ cacheWrite: number;
1121
+ };
1122
+ contextWindow: number;
1123
+ maxTokens: number;
1124
+ };
1057
1125
  readonly "openai.gpt-oss-120b-1:0": {
1058
1126
  id: string;
1059
1127
  name: string;
@@ -1190,6 +1258,23 @@ export declare const MODELS: {
1190
1258
  contextWindow: number;
1191
1259
  maxTokens: number;
1192
1260
  };
1261
+ readonly "qwen.qwen3-coder-next": {
1262
+ id: string;
1263
+ name: string;
1264
+ api: "bedrock-converse-stream";
1265
+ provider: string;
1266
+ baseUrl: string;
1267
+ reasoning: true;
1268
+ input: "text"[];
1269
+ cost: {
1270
+ input: number;
1271
+ output: number;
1272
+ cacheRead: number;
1273
+ cacheWrite: number;
1274
+ };
1275
+ contextWindow: number;
1276
+ maxTokens: number;
1277
+ };
1193
1278
  readonly "qwen.qwen3-next-80b-a3b": {
1194
1279
  id: string;
1195
1280
  name: string;
@@ -1309,6 +1394,23 @@ export declare const MODELS: {
1309
1394
  contextWindow: number;
1310
1395
  maxTokens: number;
1311
1396
  };
1397
+ readonly "us.anthropic.claude-opus-4-7": {
1398
+ id: string;
1399
+ name: string;
1400
+ api: "bedrock-converse-stream";
1401
+ provider: string;
1402
+ baseUrl: string;
1403
+ reasoning: true;
1404
+ input: ("image" | "text")[];
1405
+ cost: {
1406
+ input: number;
1407
+ output: number;
1408
+ cacheRead: number;
1409
+ cacheWrite: number;
1410
+ };
1411
+ contextWindow: number;
1412
+ maxTokens: number;
1413
+ };
1312
1414
  readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
1313
1415
  id: string;
1314
1416
  name: string;
@@ -1532,23 +1634,6 @@ export declare const MODELS: {
1532
1634
  contextWindow: number;
1533
1635
  maxTokens: number;
1534
1636
  };
1535
- readonly "claude-3-7-sonnet-latest": {
1536
- id: string;
1537
- name: string;
1538
- api: "anthropic-messages";
1539
- provider: string;
1540
- baseUrl: string;
1541
- reasoning: true;
1542
- input: ("image" | "text")[];
1543
- cost: {
1544
- input: number;
1545
- output: number;
1546
- cacheRead: number;
1547
- cacheWrite: number;
1548
- };
1549
- contextWindow: number;
1550
- maxTokens: number;
1551
- };
1552
1637
  readonly "claude-3-haiku-20240307": {
1553
1638
  id: string;
1554
1639
  name: string;
@@ -1753,7 +1838,7 @@ export declare const MODELS: {
1753
1838
  contextWindow: number;
1754
1839
  maxTokens: number;
1755
1840
  };
1756
- readonly "claude-sonnet-4-0": {
1841
+ readonly "claude-opus-4-7": {
1757
1842
  id: string;
1758
1843
  name: string;
1759
1844
  api: "anthropic-messages";
@@ -1770,7 +1855,7 @@ export declare const MODELS: {
1770
1855
  contextWindow: number;
1771
1856
  maxTokens: number;
1772
1857
  };
1773
- readonly "claude-sonnet-4-20250514": {
1858
+ readonly "claude-sonnet-4-0": {
1774
1859
  id: string;
1775
1860
  name: string;
1776
1861
  api: "anthropic-messages";
@@ -1787,7 +1872,7 @@ export declare const MODELS: {
1787
1872
  contextWindow: number;
1788
1873
  maxTokens: number;
1789
1874
  };
1790
- readonly "claude-sonnet-4-5": {
1875
+ readonly "claude-sonnet-4-20250514": {
1791
1876
  id: string;
1792
1877
  name: string;
1793
1878
  api: "anthropic-messages";
@@ -1804,7 +1889,7 @@ export declare const MODELS: {
1804
1889
  contextWindow: number;
1805
1890
  maxTokens: number;
1806
1891
  };
1807
- readonly "claude-sonnet-4-5-20250929": {
1892
+ readonly "claude-sonnet-4-5": {
1808
1893
  id: string;
1809
1894
  name: string;
1810
1895
  api: "anthropic-messages";
@@ -1821,7 +1906,7 @@ export declare const MODELS: {
1821
1906
  contextWindow: number;
1822
1907
  maxTokens: number;
1823
1908
  };
1824
- readonly "claude-sonnet-4-6": {
1909
+ readonly "claude-sonnet-4-5-20250929": {
1825
1910
  id: string;
1826
1911
  name: string;
1827
1912
  api: "anthropic-messages";
@@ -1838,16 +1923,14 @@ export declare const MODELS: {
1838
1923
  contextWindow: number;
1839
1924
  maxTokens: number;
1840
1925
  };
1841
- };
1842
- readonly "azure-openai-responses": {
1843
- readonly "codex-mini-latest": {
1926
+ readonly "claude-sonnet-4-6": {
1844
1927
  id: string;
1845
1928
  name: string;
1846
- api: "azure-openai-responses";
1929
+ api: "anthropic-messages";
1847
1930
  provider: string;
1848
1931
  baseUrl: string;
1849
1932
  reasoning: true;
1850
- input: "text"[];
1933
+ input: ("image" | "text")[];
1851
1934
  cost: {
1852
1935
  input: number;
1853
1936
  output: number;
@@ -1857,6 +1940,8 @@ export declare const MODELS: {
1857
1940
  contextWindow: number;
1858
1941
  maxTokens: number;
1859
1942
  };
1943
+ };
1944
+ readonly "azure-openai-responses": {
1860
1945
  readonly "gpt-4": {
1861
1946
  id: string;
1862
1947
  name: string;
@@ -2282,6 +2367,23 @@ export declare const MODELS: {
2282
2367
  contextWindow: number;
2283
2368
  maxTokens: number;
2284
2369
  };
2370
+ readonly "gpt-5.3-chat-latest": {
2371
+ id: string;
2372
+ name: string;
2373
+ api: "azure-openai-responses";
2374
+ provider: string;
2375
+ baseUrl: string;
2376
+ reasoning: false;
2377
+ input: ("image" | "text")[];
2378
+ cost: {
2379
+ input: number;
2380
+ output: number;
2381
+ cacheRead: number;
2382
+ cacheWrite: number;
2383
+ };
2384
+ contextWindow: number;
2385
+ maxTokens: number;
2386
+ };
2285
2387
  readonly "gpt-5.3-codex": {
2286
2388
  id: string;
2287
2389
  name: string;
@@ -2661,6 +2763,29 @@ export declare const MODELS: {
2661
2763
  contextWindow: number;
2662
2764
  maxTokens: number;
2663
2765
  };
2766
+ readonly "claude-opus-4.7": {
2767
+ id: string;
2768
+ name: string;
2769
+ api: "anthropic-messages";
2770
+ provider: string;
2771
+ baseUrl: string;
2772
+ headers: {
2773
+ "User-Agent": string;
2774
+ "Editor-Version": string;
2775
+ "Editor-Plugin-Version": string;
2776
+ "Copilot-Integration-Id": string;
2777
+ };
2778
+ reasoning: true;
2779
+ input: ("image" | "text")[];
2780
+ cost: {
2781
+ input: number;
2782
+ output: number;
2783
+ cacheRead: number;
2784
+ cacheWrite: number;
2785
+ };
2786
+ contextWindow: number;
2787
+ maxTokens: number;
2788
+ };
2664
2789
  readonly "claude-sonnet-4": {
2665
2790
  id: string;
2666
2791
  name: string;
@@ -3589,6 +3714,57 @@ export declare const MODELS: {
3589
3714
  contextWindow: number;
3590
3715
  maxTokens: number;
3591
3716
  };
3717
+ readonly "gemma-3-27b-it": {
3718
+ id: string;
3719
+ name: string;
3720
+ api: "google-generative-ai";
3721
+ provider: string;
3722
+ baseUrl: string;
3723
+ reasoning: false;
3724
+ input: ("image" | "text")[];
3725
+ cost: {
3726
+ input: number;
3727
+ output: number;
3728
+ cacheRead: number;
3729
+ cacheWrite: number;
3730
+ };
3731
+ contextWindow: number;
3732
+ maxTokens: number;
3733
+ };
3734
+ readonly "gemma-4-26b-a4b-it": {
3735
+ id: string;
3736
+ name: string;
3737
+ api: "google-generative-ai";
3738
+ provider: string;
3739
+ baseUrl: string;
3740
+ reasoning: true;
3741
+ input: ("image" | "text")[];
3742
+ cost: {
3743
+ input: number;
3744
+ output: number;
3745
+ cacheRead: number;
3746
+ cacheWrite: number;
3747
+ };
3748
+ contextWindow: number;
3749
+ maxTokens: number;
3750
+ };
3751
+ readonly "gemma-4-31b-it": {
3752
+ id: string;
3753
+ name: string;
3754
+ api: "google-generative-ai";
3755
+ provider: string;
3756
+ baseUrl: string;
3757
+ reasoning: true;
3758
+ input: ("image" | "text")[];
3759
+ cost: {
3760
+ input: number;
3761
+ output: number;
3762
+ cacheRead: number;
3763
+ cacheWrite: number;
3764
+ };
3765
+ contextWindow: number;
3766
+ maxTokens: number;
3767
+ };
3592
3768
  };
3593
3769
  readonly "google-antigravity": {
3594
3770
  readonly "claude-opus-4-5-thinking": {
@@ -4054,16 +4230,14 @@ export declare const MODELS: {
4054
4230
  contextWindow: number;
4055
4231
  maxTokens: number;
4056
4232
  };
4057
- };
4058
- readonly groq: {
4059
- readonly "deepseek-r1-distill-llama-70b": {
4233
+ readonly "gemini-3.1-pro-preview-customtools": {
4060
4234
  id: string;
4061
4235
  name: string;
4062
- api: "openai-completions";
4236
+ api: "google-vertex";
4063
4237
  provider: string;
4064
4238
  baseUrl: string;
4065
4239
  reasoning: true;
4066
- input: "text"[];
4240
+ input: ("image" | "text")[];
4067
4241
  cost: {
4068
4242
  input: number;
4069
4243
  output: number;
@@ -4073,13 +4247,15 @@ export declare const MODELS: {
4073
4247
  contextWindow: number;
4074
4248
  maxTokens: number;
4075
4249
  };
4076
- readonly "gemma2-9b-it": {
4250
+ };
4251
+ readonly groq: {
4252
+ readonly "deepseek-r1-distill-llama-70b": {
4077
4253
  id: string;
4078
4254
  name: string;
4079
4255
  api: "openai-completions";
4080
4256
  provider: string;
4081
4257
  baseUrl: string;
4082
- reasoning: false;
4258
+ reasoning: true;
4083
4259
  input: "text"[];
4084
4260
  cost: {
4085
4261
  input: number;
@@ -4090,7 +4266,7 @@ export declare const MODELS: {
4090
4266
  contextWindow: number;
4091
4267
  maxTokens: number;
4092
4268
  };
4093
- readonly "llama-3.1-8b-instant": {
4269
+ readonly "gemma2-9b-it": {
4094
4270
  id: string;
4095
4271
  name: string;
4096
4272
  api: "openai-completions";
@@ -4107,13 +4283,13 @@ export declare const MODELS: {
4107
4283
  contextWindow: number;
4108
4284
  maxTokens: number;
4109
4285
  };
4110
- readonly "llama-3.3-70b-versatile": {
4286
+ readonly "groq/compound": {
4111
4287
  id: string;
4112
4288
  name: string;
4113
4289
  api: "openai-completions";
4114
4290
  provider: string;
4115
4291
  baseUrl: string;
4116
- reasoning: false;
4292
+ reasoning: true;
4117
4293
  input: "text"[];
4118
4294
  cost: {
4119
4295
  input: number;
@@ -4124,13 +4300,13 @@ export declare const MODELS: {
4124
4300
  contextWindow: number;
4125
4301
  maxTokens: number;
4126
4302
  };
4127
- readonly "llama3-70b-8192": {
4303
+ readonly "groq/compound-mini": {
4128
4304
  id: string;
4129
4305
  name: string;
4130
4306
  api: "openai-completions";
4131
4307
  provider: string;
4132
4308
  baseUrl: string;
4133
- reasoning: false;
4309
+ reasoning: true;
4134
4310
  input: "text"[];
4135
4311
  cost: {
4136
4312
  input: number;
@@ -4141,7 +4317,7 @@ export declare const MODELS: {
4141
4317
  contextWindow: number;
4142
4318
  maxTokens: number;
4143
4319
  };
4144
- readonly "llama3-8b-8192": {
4320
+ readonly "llama-3.1-8b-instant": {
4145
4321
  id: string;
4146
4322
  name: string;
4147
4323
  api: "openai-completions";
@@ -4158,14 +4334,14 @@ export declare const MODELS: {
4158
4334
  contextWindow: number;
4159
4335
  maxTokens: number;
4160
4336
  };
4161
- readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
4337
+ readonly "llama-3.3-70b-versatile": {
4162
4338
  id: string;
4163
4339
  name: string;
4164
4340
  api: "openai-completions";
4165
4341
  provider: string;
4166
4342
  baseUrl: string;
4167
4343
  reasoning: false;
4168
- input: ("image" | "text")[];
4344
+ input: "text"[];
4169
4345
  cost: {
4170
4346
  input: number;
4171
4347
  output: number;
@@ -4175,14 +4351,14 @@ export declare const MODELS: {
4175
4351
  contextWindow: number;
4176
4352
  maxTokens: number;
4177
4353
  };
4178
- readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
4354
+ readonly "llama3-70b-8192": {
4179
4355
  id: string;
4180
4356
  name: string;
4181
4357
  api: "openai-completions";
4182
4358
  provider: string;
4183
4359
  baseUrl: string;
4184
4360
  reasoning: false;
4185
- input: ("image" | "text")[];
4361
+ input: "text"[];
4186
4362
  cost: {
4187
4363
  input: number;
4188
4364
  output: number;
@@ -4192,7 +4368,7 @@ export declare const MODELS: {
4192
4368
  contextWindow: number;
4193
4369
  maxTokens: number;
4194
4370
  };
4195
- readonly "mistral-saba-24b": {
4371
+ readonly "llama3-8b-8192": {
4196
4372
  id: string;
4197
4373
  name: string;
4198
4374
  api: "openai-completions";
@@ -4209,14 +4385,14 @@ export declare const MODELS: {
4209
4385
  contextWindow: number;
4210
4386
  maxTokens: number;
4211
4387
  };
4212
- readonly "moonshotai/kimi-k2-instruct": {
4388
+ readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
4213
4389
  id: string;
4214
4390
  name: string;
4215
4391
  api: "openai-completions";
4216
4392
  provider: string;
4217
4393
  baseUrl: string;
4218
4394
  reasoning: false;
4219
- input: "text"[];
4395
+ input: ("image" | "text")[];
4220
4396
  cost: {
4221
4397
  input: number;
4222
4398
  output: number;
@@ -4226,14 +4402,14 @@ export declare const MODELS: {
4226
4402
  contextWindow: number;
4227
4403
  maxTokens: number;
4228
4404
  };
4229
- readonly "moonshotai/kimi-k2-instruct-0905": {
4405
+ readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
4230
4406
  id: string;
4231
4407
  name: string;
4232
4408
  api: "openai-completions";
4233
4409
  provider: string;
4234
4410
  baseUrl: string;
4235
4411
  reasoning: false;
4236
- input: "text"[];
4412
+ input: ("image" | "text")[];
4237
4413
  cost: {
4238
4414
  input: number;
4239
4415
  output: number;
@@ -4243,13 +4419,13 @@ export declare const MODELS: {
4243
4419
  contextWindow: number;
4244
4420
  maxTokens: number;
4245
4421
  };
4246
- readonly "openai/gpt-oss-120b": {
4422
+ readonly "mistral-saba-24b": {
4247
4423
  id: string;
4248
4424
  name: string;
4249
4425
  api: "openai-completions";
4250
4426
  provider: string;
4251
4427
  baseUrl: string;
4252
- reasoning: true;
4428
+ reasoning: false;
4253
4429
  input: "text"[];
4254
4430
  cost: {
4255
4431
  input: number;
@@ -4260,13 +4436,13 @@ export declare const MODELS: {
4260
4436
  contextWindow: number;
4261
4437
  maxTokens: number;
4262
4438
  };
4263
- readonly "openai/gpt-oss-20b": {
4439
+ readonly "moonshotai/kimi-k2-instruct": {
4264
4440
  id: string;
4265
4441
  name: string;
4266
4442
  api: "openai-completions";
4267
4443
  provider: string;
4268
4444
  baseUrl: string;
4269
- reasoning: true;
4445
+ reasoning: false;
4270
4446
  input: "text"[];
4271
4447
  cost: {
4272
4448
  input: number;
@@ -4277,13 +4453,13 @@ export declare const MODELS: {
4277
4453
  contextWindow: number;
4278
4454
  maxTokens: number;
4279
4455
  };
4280
- readonly "qwen-qwq-32b": {
4456
+ readonly "moonshotai/kimi-k2-instruct-0905": {
4281
4457
  id: string;
4282
4458
  name: string;
4283
4459
  api: "openai-completions";
4284
4460
  provider: string;
4285
4461
  baseUrl: string;
4286
- reasoning: true;
4462
+ reasoning: false;
4287
4463
  input: "text"[];
4288
4464
  cost: {
4289
4465
  input: number;
@@ -4294,7 +4470,7 @@ export declare const MODELS: {
4294
4470
  contextWindow: number;
4295
4471
  maxTokens: number;
4296
4472
  };
4297
- readonly "qwen/qwen3-32b": {
4473
+ readonly "openai/gpt-oss-120b": {
4298
4474
  id: string;
4299
4475
  name: string;
4300
4476
  api: "openai-completions";
@@ -4311,17 +4487,12 @@ export declare const MODELS: {
4311
4487
  contextWindow: number;
4312
4488
  maxTokens: number;
4313
4489
  };
4314
- };
4315
- readonly huggingface: {
4316
- readonly "MiniMaxAI/MiniMax-M2.1": {
4490
+ readonly "openai/gpt-oss-20b": {
4317
4491
  id: string;
4318
4492
  name: string;
4319
4493
  api: "openai-completions";
4320
4494
  provider: string;
4321
4495
  baseUrl: string;
4322
- compat: {
4323
- supportsDeveloperRole: false;
4324
- };
4325
4496
  reasoning: true;
4326
4497
  input: "text"[];
4327
4498
  cost: {
@@ -4333,15 +4504,12 @@ export declare const MODELS: {
4333
4504
  contextWindow: number;
4334
4505
  maxTokens: number;
4335
4506
  };
4336
- readonly "MiniMaxAI/MiniMax-M2.5": {
4507
+ readonly "openai/gpt-oss-safeguard-20b": {
4337
4508
  id: string;
4338
4509
  name: string;
4339
4510
  api: "openai-completions";
4340
4511
  provider: string;
4341
4512
  baseUrl: string;
4342
- compat: {
4343
- supportsDeveloperRole: false;
4344
- };
4345
4513
  reasoning: true;
4346
4514
  input: "text"[];
4347
4515
  cost: {
@@ -4353,15 +4521,12 @@ export declare const MODELS: {
4353
4521
  contextWindow: number;
4354
4522
  maxTokens: number;
4355
4523
  };
4356
- readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4524
+ readonly "qwen-qwq-32b": {
4357
4525
  id: string;
4358
4526
  name: string;
4359
4527
  api: "openai-completions";
4360
4528
  provider: string;
4361
4529
  baseUrl: string;
4362
- compat: {
4363
- supportsDeveloperRole: false;
4364
- };
4365
4530
  reasoning: true;
4366
4531
  input: "text"[];
4367
4532
  cost: {
@@ -4373,16 +4538,13 @@ export declare const MODELS: {
4373
4538
  contextWindow: number;
4374
4539
  maxTokens: number;
4375
4540
  };
4376
- readonly "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
4541
+ readonly "qwen/qwen3-32b": {
4377
4542
  id: string;
4378
4543
  name: string;
4379
4544
  api: "openai-completions";
4380
4545
  provider: string;
4381
4546
  baseUrl: string;
4382
- compat: {
4383
- supportsDeveloperRole: false;
4384
- };
4385
- reasoning: false;
4547
+ reasoning: true;
4386
4548
  input: "text"[];
4387
4549
  cost: {
4388
4550
  input: number;
@@ -4393,7 +4555,9 @@ export declare const MODELS: {
4393
4555
  contextWindow: number;
4394
4556
  maxTokens: number;
4395
4557
  };
4396
- readonly "Qwen/Qwen3-Coder-Next": {
4558
+ };
4559
+ readonly huggingface: {
4560
+ readonly "MiniMaxAI/MiniMax-M2.1": {
4397
4561
  id: string;
4398
4562
  name: string;
4399
4563
  api: "openai-completions";
@@ -4402,7 +4566,7 @@ export declare const MODELS: {
4402
4566
  compat: {
4403
4567
  supportsDeveloperRole: false;
4404
4568
  };
4405
- reasoning: false;
4569
+ reasoning: true;
4406
4570
  input: "text"[];
4407
4571
  cost: {
4408
4572
  input: number;
@@ -4413,7 +4577,7 @@ export declare const MODELS: {
4413
4577
  contextWindow: number;
4414
4578
  maxTokens: number;
4415
4579
  };
4416
- readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
4580
+ readonly "MiniMaxAI/MiniMax-M2.5": {
4417
4581
  id: string;
4418
4582
  name: string;
4419
4583
  api: "openai-completions";
@@ -4422,7 +4586,7 @@ export declare const MODELS: {
4422
4586
  compat: {
4423
4587
  supportsDeveloperRole: false;
4424
4588
  };
4425
- reasoning: false;
4589
+ reasoning: true;
4426
4590
  input: "text"[];
4427
4591
  cost: {
4428
4592
  input: number;
@@ -4433,7 +4597,7 @@ export declare const MODELS: {
4433
4597
  contextWindow: number;
4434
4598
  maxTokens: number;
4435
4599
  };
4436
- readonly "Qwen/Qwen3-Next-80B-A3B-Thinking": {
4600
+ readonly "MiniMaxAI/MiniMax-M2.7": {
4437
4601
  id: string;
4438
4602
  name: string;
4439
4603
  api: "openai-completions";
@@ -4442,7 +4606,7 @@ export declare const MODELS: {
4442
4606
  compat: {
4443
4607
  supportsDeveloperRole: false;
4444
4608
  };
4445
- reasoning: false;
4609
+ reasoning: true;
4446
4610
  input: "text"[];
4447
4611
  cost: {
4448
4612
  input: number;
@@ -4453,7 +4617,7 @@ export declare const MODELS: {
4453
4617
  contextWindow: number;
4454
4618
  maxTokens: number;
4455
4619
  };
4456
- readonly "Qwen/Qwen3.5-397B-A17B": {
4620
+ readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4457
4621
  id: string;
4458
4622
  name: string;
4459
4623
  api: "openai-completions";
@@ -4463,7 +4627,7 @@ export declare const MODELS: {
4463
4627
  supportsDeveloperRole: false;
4464
4628
  };
4465
4629
  reasoning: true;
4466
- input: ("image" | "text")[];
4630
+ input: "text"[];
4467
4631
  cost: {
4468
4632
  input: number;
4469
4633
  output: number;
@@ -4473,7 +4637,7 @@ export declare const MODELS: {
4473
4637
  contextWindow: number;
4474
4638
  maxTokens: number;
4475
4639
  };
4476
- readonly "XiaomiMiMo/MiMo-V2-Flash": {
4640
+ readonly "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
4477
4641
  id: string;
4478
4642
  name: string;
4479
4643
  api: "openai-completions";
@@ -4482,7 +4646,7 @@ export declare const MODELS: {
4482
4646
  compat: {
4483
4647
  supportsDeveloperRole: false;
4484
4648
  };
4485
- reasoning: true;
4649
+ reasoning: false;
4486
4650
  input: "text"[];
4487
4651
  cost: {
4488
4652
  input: number;
@@ -4493,7 +4657,7 @@ export declare const MODELS: {
4493
4657
  contextWindow: number;
4494
4658
  maxTokens: number;
4495
4659
  };
4496
- readonly "deepseek-ai/DeepSeek-R1-0528": {
4660
+ readonly "Qwen/Qwen3-Coder-Next": {
4497
4661
  id: string;
4498
4662
  name: string;
4499
4663
  api: "openai-completions";
@@ -4502,7 +4666,7 @@ export declare const MODELS: {
4502
4666
  compat: {
4503
4667
  supportsDeveloperRole: false;
4504
4668
  };
4505
- reasoning: true;
4669
+ reasoning: false;
4506
4670
  input: "text"[];
4507
4671
  cost: {
4508
4672
  input: number;
@@ -4513,7 +4677,7 @@ export declare const MODELS: {
4513
4677
  contextWindow: number;
4514
4678
  maxTokens: number;
4515
4679
  };
4516
- readonly "deepseek-ai/DeepSeek-V3.2": {
4680
+ readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
4517
4681
  id: string;
4518
4682
  name: string;
4519
4683
  api: "openai-completions";
@@ -4522,7 +4686,7 @@ export declare const MODELS: {
4522
4686
  compat: {
4523
4687
  supportsDeveloperRole: false;
4524
4688
  };
4525
- reasoning: true;
4689
+ reasoning: false;
4526
4690
  input: "text"[];
4527
4691
  cost: {
4528
4692
  input: number;
@@ -4533,7 +4697,7 @@ export declare const MODELS: {
4533
4697
  contextWindow: number;
4534
4698
  maxTokens: number;
4535
4699
  };
4536
- readonly "moonshotai/Kimi-K2-Instruct": {
4700
+ readonly "Qwen/Qwen3-Next-80B-A3B-Thinking": {
4537
4701
  id: string;
4538
4702
  name: string;
4539
4703
  api: "openai-completions";
@@ -4553,7 +4717,7 @@ export declare const MODELS: {
4553
4717
  contextWindow: number;
4554
4718
  maxTokens: number;
4555
4719
  };
4556
- readonly "moonshotai/Kimi-K2-Instruct-0905": {
4720
+ readonly "Qwen/Qwen3.5-397B-A17B": {
4557
4721
  id: string;
4558
4722
  name: string;
4559
4723
  api: "openai-completions";
@@ -4562,8 +4726,8 @@ export declare const MODELS: {
4562
4726
  compat: {
4563
4727
  supportsDeveloperRole: false;
4564
4728
  };
4565
- reasoning: false;
4566
- input: "text"[];
4729
+ reasoning: true;
4730
+ input: ("image" | "text")[];
4567
4731
  cost: {
4568
4732
  input: number;
4569
4733
  output: number;
@@ -4573,7 +4737,7 @@ export declare const MODELS: {
4573
4737
  contextWindow: number;
4574
4738
  maxTokens: number;
4575
4739
  };
4576
- readonly "moonshotai/Kimi-K2-Thinking": {
4740
+ readonly "XiaomiMiMo/MiMo-V2-Flash": {
4577
4741
  id: string;
4578
4742
  name: string;
4579
4743
  api: "openai-completions";
@@ -4593,7 +4757,7 @@ export declare const MODELS: {
4593
4757
  contextWindow: number;
4594
4758
  maxTokens: number;
4595
4759
  };
4596
- readonly "moonshotai/Kimi-K2.5": {
4760
+ readonly "deepseek-ai/DeepSeek-R1-0528": {
4597
4761
  id: string;
4598
4762
  name: string;
4599
4763
  api: "openai-completions";
@@ -4603,7 +4767,7 @@ export declare const MODELS: {
4603
4767
  supportsDeveloperRole: false;
4604
4768
  };
4605
4769
  reasoning: true;
4606
- input: ("image" | "text")[];
4770
+ input: "text"[];
4607
4771
  cost: {
4608
4772
  input: number;
4609
4773
  output: number;
@@ -4613,7 +4777,7 @@ export declare const MODELS: {
4613
4777
  contextWindow: number;
4614
4778
  maxTokens: number;
4615
4779
  };
4616
- readonly "zai-org/GLM-4.7": {
4780
+ readonly "deepseek-ai/DeepSeek-V3.2": {
4617
4781
  id: string;
4618
4782
  name: string;
4619
4783
  api: "openai-completions";
@@ -4633,7 +4797,7 @@ export declare const MODELS: {
4633
4797
  contextWindow: number;
4634
4798
  maxTokens: number;
4635
4799
  };
4636
- readonly "zai-org/GLM-4.7-Flash": {
4800
+ readonly "moonshotai/Kimi-K2-Instruct": {
4637
4801
  id: string;
4638
4802
  name: string;
4639
4803
  api: "openai-completions";
@@ -4642,7 +4806,7 @@ export declare const MODELS: {
4642
4806
  compat: {
4643
4807
  supportsDeveloperRole: false;
4644
4808
  };
4645
- reasoning: true;
4809
+ reasoning: false;
4646
4810
  input: "text"[];
4647
4811
  cost: {
4648
4812
  input: number;
@@ -4653,7 +4817,7 @@ export declare const MODELS: {
4653
4817
  contextWindow: number;
4654
4818
  maxTokens: number;
4655
4819
  };
4656
- readonly "zai-org/GLM-5": {
4820
+ readonly "moonshotai/Kimi-K2-Instruct-0905": {
4657
4821
  id: string;
4658
4822
  name: string;
4659
4823
  api: "openai-completions";
@@ -4662,7 +4826,7 @@ export declare const MODELS: {
4662
4826
  compat: {
4663
4827
  supportsDeveloperRole: false;
4664
4828
  };
4665
- reasoning: true;
4829
+ reasoning: false;
4666
4830
  input: "text"[];
4667
4831
  cost: {
4668
4832
  input: number;
@@ -4673,31 +4837,15 @@ export declare const MODELS: {
4673
4837
  contextWindow: number;
4674
4838
  maxTokens: number;
4675
4839
  };
4676
- };
4677
- readonly "kimi-coding": {
4678
- readonly k2p5: {
4840
+ readonly "moonshotai/Kimi-K2-Thinking": {
4679
4841
  id: string;
4680
4842
  name: string;
4681
- api: "anthropic-messages";
4843
+ api: "openai-completions";
4682
4844
  provider: string;
4683
4845
  baseUrl: string;
4684
- reasoning: true;
4685
- input: ("image" | "text")[];
4686
- cost: {
4687
- input: number;
4688
- output: number;
4689
- cacheRead: number;
4690
- cacheWrite: number;
4846
+ compat: {
4847
+ supportsDeveloperRole: false;
4691
4848
  };
4692
- contextWindow: number;
4693
- maxTokens: number;
4694
- };
4695
- readonly "kimi-k2-thinking": {
4696
- id: string;
4697
- name: string;
4698
- api: "anthropic-messages";
4699
- provider: string;
4700
- baseUrl: string;
4701
4849
  reasoning: true;
4702
4850
  input: "text"[];
4703
4851
  cost: {
@@ -4709,33 +4857,17 @@ export declare const MODELS: {
4709
4857
  contextWindow: number;
4710
4858
  maxTokens: number;
4711
4859
  };
4712
- };
4713
- readonly minimax: {
4714
- readonly "MiniMax-M2": {
4860
+ readonly "moonshotai/Kimi-K2.5": {
4715
4861
  id: string;
4716
4862
  name: string;
4717
- api: "anthropic-messages";
4863
+ api: "openai-completions";
4718
4864
  provider: string;
4719
4865
  baseUrl: string;
4720
- reasoning: true;
4721
- input: "text"[];
4722
- cost: {
4723
- input: number;
4724
- output: number;
4725
- cacheRead: number;
4726
- cacheWrite: number;
4866
+ compat: {
4867
+ supportsDeveloperRole: false;
4727
4868
  };
4728
- contextWindow: number;
4729
- maxTokens: number;
4730
- };
4731
- readonly "MiniMax-M2.1": {
4732
- id: string;
4733
- name: string;
4734
- api: "anthropic-messages";
4735
- provider: string;
4736
- baseUrl: string;
4737
4869
  reasoning: true;
4738
- input: "text"[];
4870
+ input: ("image" | "text")[];
4739
4871
  cost: {
4740
4872
  input: number;
4741
4873
  output: number;
@@ -4745,14 +4877,17 @@ export declare const MODELS: {
4745
4877
  contextWindow: number;
4746
4878
  maxTokens: number;
4747
4879
  };
4748
- readonly "MiniMax-M2.1-highspeed": {
4880
+ readonly "moonshotai/Kimi-K2.6": {
4749
4881
  id: string;
4750
4882
  name: string;
4751
- api: "anthropic-messages";
4883
+ api: "openai-completions";
4752
4884
  provider: string;
4753
4885
  baseUrl: string;
4886
+ compat: {
4887
+ supportsDeveloperRole: false;
4888
+ };
4754
4889
  reasoning: true;
4755
- input: "text"[];
4890
+ input: ("image" | "text")[];
4756
4891
  cost: {
4757
4892
  input: number;
4758
4893
  output: number;
@@ -4762,12 +4897,15 @@ export declare const MODELS: {
4762
4897
  contextWindow: number;
4763
4898
  maxTokens: number;
4764
4899
  };
4765
- readonly "MiniMax-M2.5": {
4900
+ readonly "zai-org/GLM-4.7": {
4766
4901
  id: string;
4767
4902
  name: string;
4768
- api: "anthropic-messages";
4903
+ api: "openai-completions";
4769
4904
  provider: string;
4770
4905
  baseUrl: string;
4906
+ compat: {
4907
+ supportsDeveloperRole: false;
4908
+ };
4771
4909
  reasoning: true;
4772
4910
  input: "text"[];
4773
4911
  cost: {
@@ -4779,12 +4917,15 @@ export declare const MODELS: {
4779
4917
  contextWindow: number;
4780
4918
  maxTokens: number;
4781
4919
  };
4782
- readonly "MiniMax-M2.5-highspeed": {
4920
+ readonly "zai-org/GLM-4.7-Flash": {
4783
4921
  id: string;
4784
4922
  name: string;
4785
- api: "anthropic-messages";
4923
+ api: "openai-completions";
4786
4924
  provider: string;
4787
4925
  baseUrl: string;
4926
+ compat: {
4927
+ supportsDeveloperRole: false;
4928
+ };
4788
4929
  reasoning: true;
4789
4930
  input: "text"[];
4790
4931
  cost: {
@@ -4796,12 +4937,15 @@ export declare const MODELS: {
4796
4937
  contextWindow: number;
4797
4938
  maxTokens: number;
4798
4939
  };
4799
- readonly "MiniMax-M2.7": {
4940
+ readonly "zai-org/GLM-5": {
4800
4941
  id: string;
4801
4942
  name: string;
4802
- api: "anthropic-messages";
4943
+ api: "openai-completions";
4803
4944
  provider: string;
4804
4945
  baseUrl: string;
4946
+ compat: {
4947
+ supportsDeveloperRole: false;
4948
+ };
4805
4949
  reasoning: true;
4806
4950
  input: "text"[];
4807
4951
  cost: {
@@ -4813,12 +4957,15 @@ export declare const MODELS: {
4813
4957
  contextWindow: number;
4814
4958
  maxTokens: number;
4815
4959
  };
4816
- readonly "MiniMax-M2.7-highspeed": {
4960
+ readonly "zai-org/GLM-5.1": {
4817
4961
  id: string;
4818
4962
  name: string;
4819
- api: "anthropic-messages";
4963
+ api: "openai-completions";
4820
4964
  provider: string;
4821
4965
  baseUrl: string;
4966
+ compat: {
4967
+ supportsDeveloperRole: false;
4968
+ };
4822
4969
  reasoning: true;
4823
4970
  input: "text"[];
4824
4971
  cost: {
@@ -4831,15 +4978,15 @@ export declare const MODELS: {
4831
4978
  maxTokens: number;
4832
4979
  };
4833
4980
  };
4834
- readonly "minimax-cn": {
4835
- readonly "MiniMax-M2": {
4981
+ readonly "kimi-coding": {
4982
+ readonly k2p5: {
4836
4983
  id: string;
4837
4984
  name: string;
4838
4985
  api: "anthropic-messages";
4839
4986
  provider: string;
4840
4987
  baseUrl: string;
4841
4988
  reasoning: true;
4842
- input: "text"[];
4989
+ input: ("image" | "text")[];
4843
4990
  cost: {
4844
4991
  input: number;
4845
4992
  output: number;
@@ -4849,14 +4996,14 @@ export declare const MODELS: {
4849
4996
  contextWindow: number;
4850
4997
  maxTokens: number;
4851
4998
  };
4852
- readonly "MiniMax-M2.1": {
4999
+ readonly k2p6: {
4853
5000
  id: string;
4854
5001
  name: string;
4855
5002
  api: "anthropic-messages";
4856
5003
  provider: string;
4857
5004
  baseUrl: string;
4858
5005
  reasoning: true;
4859
- input: "text"[];
5006
+ input: ("image" | "text")[];
4860
5007
  cost: {
4861
5008
  input: number;
4862
5009
  output: number;
@@ -4866,7 +5013,7 @@ export declare const MODELS: {
4866
5013
  contextWindow: number;
4867
5014
  maxTokens: number;
4868
5015
  };
4869
- readonly "MiniMax-M2.1-highspeed": {
5016
+ readonly "kimi-k2-thinking": {
4870
5017
  id: string;
4871
5018
  name: string;
4872
5019
  api: "anthropic-messages";
@@ -4883,7 +5030,9 @@ export declare const MODELS: {
4883
5030
  contextWindow: number;
4884
5031
  maxTokens: number;
4885
5032
  };
4886
- readonly "MiniMax-M2.5": {
5033
+ };
5034
+ readonly minimax: {
5035
+ readonly "MiniMax-M2.7": {
4887
5036
  id: string;
4888
5037
  name: string;
4889
5038
  api: "anthropic-messages";
@@ -4900,7 +5049,7 @@ export declare const MODELS: {
4900
5049
  contextWindow: number;
4901
5050
  maxTokens: number;
4902
5051
  };
4903
- readonly "MiniMax-M2.5-highspeed": {
5052
+ readonly "MiniMax-M2.7-highspeed": {
4904
5053
  id: string;
4905
5054
  name: string;
4906
5055
  api: "anthropic-messages";
@@ -4917,6 +5066,8 @@ export declare const MODELS: {
4917
5066
  contextWindow: number;
4918
5067
  maxTokens: number;
4919
5068
  };
5069
+ };
5070
+ readonly "minimax-cn": {
4920
5071
  readonly "MiniMax-M2.7": {
4921
5072
  id: string;
4922
5073
  name: string;
@@ -4956,7 +5107,7 @@ export declare const MODELS: {
4956
5107
  readonly "codestral-latest": {
4957
5108
  id: string;
4958
5109
  name: string;
4959
- api: "mistral-conversations";
5110
+ api: "openai-completions";
4960
5111
  provider: string;
4961
5112
  baseUrl: string;
4962
5113
  reasoning: false;
@@ -4973,7 +5124,7 @@ export declare const MODELS: {
4973
5124
  readonly "devstral-2512": {
4974
5125
  id: string;
4975
5126
  name: string;
4976
- api: "mistral-conversations";
5127
+ api: "openai-completions";
4977
5128
  provider: string;
4978
5129
  baseUrl: string;
4979
5130
  reasoning: false;
@@ -4990,7 +5141,7 @@ export declare const MODELS: {
4990
5141
  readonly "devstral-medium-2507": {
4991
5142
  id: string;
4992
5143
  name: string;
4993
- api: "mistral-conversations";
5144
+ api: "openai-completions";
4994
5145
  provider: string;
4995
5146
  baseUrl: string;
4996
5147
  reasoning: false;
@@ -5007,7 +5158,7 @@ export declare const MODELS: {
5007
5158
  readonly "devstral-medium-latest": {
5008
5159
  id: string;
5009
5160
  name: string;
5010
- api: "mistral-conversations";
5161
+ api: "openai-completions";
5011
5162
  provider: string;
5012
5163
  baseUrl: string;
5013
5164
  reasoning: false;
@@ -5024,7 +5175,7 @@ export declare const MODELS: {
5024
5175
  readonly "devstral-small-2505": {
5025
5176
  id: string;
5026
5177
  name: string;
5027
- api: "mistral-conversations";
5178
+ api: "openai-completions";
5028
5179
  provider: string;
5029
5180
  baseUrl: string;
5030
5181
  reasoning: false;
@@ -5041,7 +5192,7 @@ export declare const MODELS: {
5041
5192
  readonly "devstral-small-2507": {
5042
5193
  id: string;
5043
5194
  name: string;
5044
- api: "mistral-conversations";
5195
+ api: "openai-completions";
5045
5196
  provider: string;
5046
5197
  baseUrl: string;
5047
5198
  reasoning: false;
@@ -5058,7 +5209,7 @@ export declare const MODELS: {
5058
5209
  readonly "labs-devstral-small-2512": {
5059
5210
  id: string;
5060
5211
  name: string;
5061
- api: "mistral-conversations";
5212
+ api: "openai-completions";
5062
5213
  provider: string;
5063
5214
  baseUrl: string;
5064
5215
  reasoning: false;
@@ -5075,7 +5226,7 @@ export declare const MODELS: {
5075
5226
  readonly "magistral-medium-latest": {
5076
5227
  id: string;
5077
5228
  name: string;
5078
- api: "mistral-conversations";
5229
+ api: "openai-completions";
5079
5230
  provider: string;
5080
5231
  baseUrl: string;
5081
5232
  reasoning: true;
@@ -5092,7 +5243,7 @@ export declare const MODELS: {
5092
5243
  readonly "magistral-small": {
5093
5244
  id: string;
5094
5245
  name: string;
5095
- api: "mistral-conversations";
5246
+ api: "openai-completions";
5096
5247
  provider: string;
5097
5248
  baseUrl: string;
5098
5249
  reasoning: true;
@@ -5109,7 +5260,7 @@ export declare const MODELS: {
5109
5260
  readonly "ministral-3b-latest": {
5110
5261
  id: string;
5111
5262
  name: string;
5112
- api: "mistral-conversations";
5263
+ api: "openai-completions";
5113
5264
  provider: string;
5114
5265
  baseUrl: string;
5115
5266
  reasoning: false;
@@ -5126,7 +5277,7 @@ export declare const MODELS: {
5126
5277
  readonly "ministral-8b-latest": {
5127
5278
  id: string;
5128
5279
  name: string;
5129
- api: "mistral-conversations";
5280
+ api: "openai-completions";
5130
5281
  provider: string;
5131
5282
  baseUrl: string;
5132
5283
  reasoning: false;
@@ -5143,7 +5294,7 @@ export declare const MODELS: {
5143
5294
  readonly "mistral-large-2411": {
5144
5295
  id: string;
5145
5296
  name: string;
5146
- api: "mistral-conversations";
5297
+ api: "openai-completions";
5147
5298
  provider: string;
5148
5299
  baseUrl: string;
5149
5300
  reasoning: false;
@@ -5160,7 +5311,7 @@ export declare const MODELS: {
5160
5311
  readonly "mistral-large-2512": {
5161
5312
  id: string;
5162
5313
  name: string;
5163
- api: "mistral-conversations";
5314
+ api: "openai-completions";
5164
5315
  provider: string;
5165
5316
  baseUrl: string;
5166
5317
  reasoning: false;
@@ -5177,7 +5328,7 @@ export declare const MODELS: {
5177
5328
  readonly "mistral-large-latest": {
5178
5329
  id: string;
5179
5330
  name: string;
5180
- api: "mistral-conversations";
5331
+ api: "openai-completions";
5181
5332
  provider: string;
5182
5333
  baseUrl: string;
5183
5334
  reasoning: false;
@@ -5194,7 +5345,7 @@ export declare const MODELS: {
5194
5345
  readonly "mistral-medium-2505": {
5195
5346
  id: string;
5196
5347
  name: string;
5197
- api: "mistral-conversations";
5348
+ api: "openai-completions";
5198
5349
  provider: string;
5199
5350
  baseUrl: string;
5200
5351
  reasoning: false;
@@ -5211,7 +5362,7 @@ export declare const MODELS: {
5211
5362
  readonly "mistral-medium-2508": {
5212
5363
  id: string;
5213
5364
  name: string;
5214
- api: "mistral-conversations";
5365
+ api: "openai-completions";
5215
5366
  provider: string;
5216
5367
  baseUrl: string;
5217
5368
  reasoning: false;
@@ -5228,7 +5379,7 @@ export declare const MODELS: {
5228
5379
  readonly "mistral-medium-latest": {
5229
5380
  id: string;
5230
5381
  name: string;
5231
- api: "mistral-conversations";
5382
+ api: "openai-completions";
5232
5383
  provider: string;
5233
5384
  baseUrl: string;
5234
5385
  reasoning: false;
@@ -5245,7 +5396,7 @@ export declare const MODELS: {
5245
5396
  readonly "mistral-nemo": {
5246
5397
  id: string;
5247
5398
  name: string;
5248
- api: "mistral-conversations";
5399
+ api: "openai-completions";
5249
5400
  provider: string;
5250
5401
  baseUrl: string;
5251
5402
  reasoning: false;
@@ -5262,7 +5413,7 @@ export declare const MODELS: {
5262
5413
  readonly "mistral-small-2506": {
5263
5414
  id: string;
5264
5415
  name: string;
5265
- api: "mistral-conversations";
5416
+ api: "openai-completions";
5266
5417
  provider: string;
5267
5418
  baseUrl: string;
5268
5419
  reasoning: false;
@@ -5276,13 +5427,13 @@ export declare const MODELS: {
5276
5427
  contextWindow: number;
5277
5428
  maxTokens: number;
5278
5429
  };
5279
- readonly "mistral-small-latest": {
5430
+ readonly "mistral-small-2603": {
5280
5431
  id: string;
5281
5432
  name: string;
5282
- api: "mistral-conversations";
5433
+ api: "openai-completions";
5283
5434
  provider: string;
5284
5435
  baseUrl: string;
5285
- reasoning: false;
5436
+ reasoning: true;
5286
5437
  input: ("image" | "text")[];
5287
5438
  cost: {
5288
5439
  input: number;
@@ -5293,14 +5444,14 @@ export declare const MODELS: {
5293
5444
  contextWindow: number;
5294
5445
  maxTokens: number;
5295
5446
  };
5296
- readonly "open-mistral-7b": {
5447
+ readonly "mistral-small-latest": {
5297
5448
  id: string;
5298
5449
  name: string;
5299
- api: "mistral-conversations";
5450
+ api: "openai-completions";
5300
5451
  provider: string;
5301
5452
  baseUrl: string;
5302
- reasoning: false;
5303
- input: "text"[];
5453
+ reasoning: true;
5454
+ input: ("image" | "text")[];
5304
5455
  cost: {
5305
5456
  input: number;
5306
5457
  output: number;
@@ -5310,10 +5461,10 @@ export declare const MODELS: {
5310
5461
  contextWindow: number;
5311
5462
  maxTokens: number;
5312
5463
  };
5313
- readonly "open-mixtral-8x22b": {
5464
+ readonly "open-mistral-7b": {
5314
5465
  id: string;
5315
5466
  name: string;
5316
- api: "mistral-conversations";
5467
+ api: "openai-completions";
5317
5468
  provider: string;
5318
5469
  baseUrl: string;
5319
5470
  reasoning: false;
@@ -5327,10 +5478,10 @@ export declare const MODELS: {
5327
5478
  contextWindow: number;
5328
5479
  maxTokens: number;
5329
5480
  };
5330
- readonly "open-mixtral-8x7b": {
5481
+ readonly "open-mixtral-8x22b": {
5331
5482
  id: string;
5332
5483
  name: string;
5333
- api: "mistral-conversations";
5484
+ api: "openai-completions";
5334
5485
  provider: string;
5335
5486
  baseUrl: string;
5336
5487
  reasoning: false;
@@ -5344,14 +5495,14 @@ export declare const MODELS: {
5344
5495
  contextWindow: number;
5345
5496
  maxTokens: number;
5346
5497
  };
5347
- readonly "pixtral-12b": {
5498
+ readonly "open-mixtral-8x7b": {
5348
5499
  id: string;
5349
5500
  name: string;
5350
- api: "mistral-conversations";
5501
+ api: "openai-completions";
5351
5502
  provider: string;
5352
5503
  baseUrl: string;
5353
5504
  reasoning: false;
5354
- input: ("image" | "text")[];
5505
+ input: "text"[];
5355
5506
  cost: {
5356
5507
  input: number;
5357
5508
  output: number;
@@ -5361,10 +5512,10 @@ export declare const MODELS: {
5361
5512
  contextWindow: number;
5362
5513
  maxTokens: number;
5363
5514
  };
5364
- readonly "pixtral-large-latest": {
5515
+ readonly "pixtral-12b": {
5365
5516
  id: string;
5366
5517
  name: string;
5367
- api: "mistral-conversations";
5518
+ api: "openai-completions";
5368
5519
  provider: string;
5369
5520
  baseUrl: string;
5370
5521
  reasoning: false;
@@ -5378,16 +5529,14 @@ export declare const MODELS: {
5378
5529
  contextWindow: number;
5379
5530
  maxTokens: number;
5380
5531
  };
5381
- };
5382
- readonly openai: {
5383
- readonly "codex-mini-latest": {
5532
+ readonly "pixtral-large-latest": {
5384
5533
  id: string;
5385
5534
  name: string;
5386
- api: "openai-responses";
5535
+ api: "openai-completions";
5387
5536
  provider: string;
5388
5537
  baseUrl: string;
5389
- reasoning: true;
5390
- input: "text"[];
5538
+ reasoning: false;
5539
+ input: ("image" | "text")[];
5391
5540
  cost: {
5392
5541
  input: number;
5393
5542
  output: number;
@@ -5397,6 +5546,8 @@ export declare const MODELS: {
5397
5546
  contextWindow: number;
5398
5547
  maxTokens: number;
5399
5548
  };
5549
+ };
5550
+ readonly openai: {
5400
5551
  readonly "gpt-4": {
5401
5552
  id: string;
5402
5553
  name: string;
@@ -5822,6 +5973,23 @@ export declare const MODELS: {
5822
5973
  contextWindow: number;
5823
5974
  maxTokens: number;
5824
5975
  };
5976
+ readonly "gpt-5.3-chat-latest": {
5977
+ id: string;
5978
+ name: string;
5979
+ api: "openai-responses";
5980
+ provider: string;
5981
+ baseUrl: string;
5982
+ reasoning: false;
5983
+ input: ("image" | "text")[];
5984
+ cost: {
5985
+ input: number;
5986
+ output: number;
5987
+ cacheRead: number;
5988
+ cacheWrite: number;
5989
+ };
5990
+ contextWindow: number;
5991
+ maxTokens: number;
5992
+ };
5825
5993
  readonly "gpt-5.3-codex": {
5826
5994
  id: string;
5827
5995
  name: string;
@@ -6046,7 +6214,249 @@ export declare const MODELS: {
6046
6214
  readonly "o4-mini-deep-research": {
6047
6215
  id: string;
6048
6216
  name: string;
6049
- api: "openai-responses";
6217
+ api: "openai-responses";
6218
+ provider: string;
6219
+ baseUrl: string;
6220
+ reasoning: true;
6221
+ input: ("image" | "text")[];
6222
+ cost: {
6223
+ input: number;
6224
+ output: number;
6225
+ cacheRead: number;
6226
+ cacheWrite: number;
6227
+ };
6228
+ contextWindow: number;
6229
+ maxTokens: number;
6230
+ };
6231
+ };
6232
+ readonly "openai-codex": {
6233
+ readonly "gpt-5.1": {
6234
+ id: string;
6235
+ name: string;
6236
+ api: "openai-codex-responses";
6237
+ provider: string;
6238
+ baseUrl: string;
6239
+ reasoning: true;
6240
+ input: ("image" | "text")[];
6241
+ cost: {
6242
+ input: number;
6243
+ output: number;
6244
+ cacheRead: number;
6245
+ cacheWrite: number;
6246
+ };
6247
+ contextWindow: number;
6248
+ maxTokens: number;
6249
+ };
6250
+ readonly "gpt-5.1-codex-max": {
6251
+ id: string;
6252
+ name: string;
6253
+ api: "openai-codex-responses";
6254
+ provider: string;
6255
+ baseUrl: string;
6256
+ reasoning: true;
6257
+ input: ("image" | "text")[];
6258
+ cost: {
6259
+ input: number;
6260
+ output: number;
6261
+ cacheRead: number;
6262
+ cacheWrite: number;
6263
+ };
6264
+ contextWindow: number;
6265
+ maxTokens: number;
6266
+ };
6267
+ readonly "gpt-5.1-codex-mini": {
6268
+ id: string;
6269
+ name: string;
6270
+ api: "openai-codex-responses";
6271
+ provider: string;
6272
+ baseUrl: string;
6273
+ reasoning: true;
6274
+ input: ("image" | "text")[];
6275
+ cost: {
6276
+ input: number;
6277
+ output: number;
6278
+ cacheRead: number;
6279
+ cacheWrite: number;
6280
+ };
6281
+ contextWindow: number;
6282
+ maxTokens: number;
6283
+ };
6284
+ readonly "gpt-5.2": {
6285
+ id: string;
6286
+ name: string;
6287
+ api: "openai-codex-responses";
6288
+ provider: string;
6289
+ baseUrl: string;
6290
+ reasoning: true;
6291
+ input: ("image" | "text")[];
6292
+ cost: {
6293
+ input: number;
6294
+ output: number;
6295
+ cacheRead: number;
6296
+ cacheWrite: number;
6297
+ };
6298
+ contextWindow: number;
6299
+ maxTokens: number;
6300
+ };
6301
+ readonly "gpt-5.2-codex": {
6302
+ id: string;
6303
+ name: string;
6304
+ api: "openai-codex-responses";
6305
+ provider: string;
6306
+ baseUrl: string;
6307
+ reasoning: true;
6308
+ input: ("image" | "text")[];
6309
+ cost: {
6310
+ input: number;
6311
+ output: number;
6312
+ cacheRead: number;
6313
+ cacheWrite: number;
6314
+ };
6315
+ contextWindow: number;
6316
+ maxTokens: number;
6317
+ };
6318
+ readonly "gpt-5.3-codex": {
6319
+ id: string;
6320
+ name: string;
6321
+ api: "openai-codex-responses";
6322
+ provider: string;
6323
+ baseUrl: string;
6324
+ reasoning: true;
6325
+ input: ("image" | "text")[];
6326
+ cost: {
6327
+ input: number;
6328
+ output: number;
6329
+ cacheRead: number;
6330
+ cacheWrite: number;
6331
+ };
6332
+ contextWindow: number;
6333
+ maxTokens: number;
6334
+ };
6335
+ readonly "gpt-5.3-codex-spark": {
6336
+ id: string;
6337
+ name: string;
6338
+ api: "openai-codex-responses";
6339
+ provider: string;
6340
+ baseUrl: string;
6341
+ reasoning: true;
6342
+ input: "text"[];
6343
+ cost: {
6344
+ input: number;
6345
+ output: number;
6346
+ cacheRead: number;
6347
+ cacheWrite: number;
6348
+ };
6349
+ contextWindow: number;
6350
+ maxTokens: number;
6351
+ };
6352
+ readonly "gpt-5.4": {
6353
+ id: string;
6354
+ name: string;
6355
+ api: "openai-codex-responses";
6356
+ provider: string;
6357
+ baseUrl: string;
6358
+ reasoning: true;
6359
+ input: ("image" | "text")[];
6360
+ cost: {
6361
+ input: number;
6362
+ output: number;
6363
+ cacheRead: number;
6364
+ cacheWrite: number;
6365
+ };
6366
+ contextWindow: number;
6367
+ maxTokens: number;
6368
+ };
6369
+ readonly "gpt-5.4-mini": {
6370
+ id: string;
6371
+ name: string;
6372
+ api: "openai-codex-responses";
6373
+ provider: string;
6374
+ baseUrl: string;
6375
+ reasoning: true;
6376
+ input: ("image" | "text")[];
6377
+ cost: {
6378
+ input: number;
6379
+ output: number;
6380
+ cacheRead: number;
6381
+ cacheWrite: number;
6382
+ };
6383
+ contextWindow: number;
6384
+ maxTokens: number;
6385
+ };
6386
+ };
6387
+ readonly opencode: {
6388
+ readonly "big-pickle": {
6389
+ id: string;
6390
+ name: string;
6391
+ api: "anthropic-messages";
6392
+ provider: string;
6393
+ baseUrl: string;
6394
+ reasoning: true;
6395
+ input: "text"[];
6396
+ cost: {
6397
+ input: number;
6398
+ output: number;
6399
+ cacheRead: number;
6400
+ cacheWrite: number;
6401
+ };
6402
+ contextWindow: number;
6403
+ maxTokens: number;
6404
+ };
6405
+ readonly "claude-3-5-haiku": {
6406
+ id: string;
6407
+ name: string;
6408
+ api: "anthropic-messages";
6409
+ provider: string;
6410
+ baseUrl: string;
6411
+ reasoning: false;
6412
+ input: ("image" | "text")[];
6413
+ cost: {
6414
+ input: number;
6415
+ output: number;
6416
+ cacheRead: number;
6417
+ cacheWrite: number;
6418
+ };
6419
+ contextWindow: number;
6420
+ maxTokens: number;
6421
+ };
6422
+ readonly "claude-haiku-4-5": {
6423
+ id: string;
6424
+ name: string;
6425
+ api: "anthropic-messages";
6426
+ provider: string;
6427
+ baseUrl: string;
6428
+ reasoning: true;
6429
+ input: ("image" | "text")[];
6430
+ cost: {
6431
+ input: number;
6432
+ output: number;
6433
+ cacheRead: number;
6434
+ cacheWrite: number;
6435
+ };
6436
+ contextWindow: number;
6437
+ maxTokens: number;
6438
+ };
6439
+ readonly "claude-opus-4-1": {
6440
+ id: string;
6441
+ name: string;
6442
+ api: "anthropic-messages";
6443
+ provider: string;
6444
+ baseUrl: string;
6445
+ reasoning: true;
6446
+ input: ("image" | "text")[];
6447
+ cost: {
6448
+ input: number;
6449
+ output: number;
6450
+ cacheRead: number;
6451
+ cacheWrite: number;
6452
+ };
6453
+ contextWindow: number;
6454
+ maxTokens: number;
6455
+ };
6456
+ readonly "claude-opus-4-5": {
6457
+ id: string;
6458
+ name: string;
6459
+ api: "anthropic-messages";
6050
6460
  provider: string;
6051
6461
  baseUrl: string;
6052
6462
  reasoning: true;
@@ -6060,12 +6470,10 @@ export declare const MODELS: {
6060
6470
  contextWindow: number;
6061
6471
  maxTokens: number;
6062
6472
  };
6063
- };
6064
- readonly "openai-codex": {
6065
- readonly "gpt-5.1": {
6473
+ readonly "claude-opus-4-6": {
6066
6474
  id: string;
6067
6475
  name: string;
6068
- api: "openai-codex-responses";
6476
+ api: "anthropic-messages";
6069
6477
  provider: string;
6070
6478
  baseUrl: string;
6071
6479
  reasoning: true;
@@ -6079,10 +6487,10 @@ export declare const MODELS: {
6079
6487
  contextWindow: number;
6080
6488
  maxTokens: number;
6081
6489
  };
6082
- readonly "gpt-5.1-codex-max": {
6490
+ readonly "claude-opus-4-7": {
6083
6491
  id: string;
6084
6492
  name: string;
6085
- api: "openai-codex-responses";
6493
+ api: "anthropic-messages";
6086
6494
  provider: string;
6087
6495
  baseUrl: string;
6088
6496
  reasoning: true;
@@ -6096,10 +6504,10 @@ export declare const MODELS: {
6096
6504
  contextWindow: number;
6097
6505
  maxTokens: number;
6098
6506
  };
6099
- readonly "gpt-5.1-codex-mini": {
6507
+ readonly "claude-sonnet-4": {
6100
6508
  id: string;
6101
6509
  name: string;
6102
- api: "openai-codex-responses";
6510
+ api: "anthropic-messages";
6103
6511
  provider: string;
6104
6512
  baseUrl: string;
6105
6513
  reasoning: true;
@@ -6113,10 +6521,10 @@ export declare const MODELS: {
6113
6521
  contextWindow: number;
6114
6522
  maxTokens: number;
6115
6523
  };
6116
- readonly "gpt-5.2": {
6524
+ readonly "claude-sonnet-4-5": {
6117
6525
  id: string;
6118
6526
  name: string;
6119
- api: "openai-codex-responses";
6527
+ api: "anthropic-messages";
6120
6528
  provider: string;
6121
6529
  baseUrl: string;
6122
6530
  reasoning: true;
@@ -6130,10 +6538,10 @@ export declare const MODELS: {
6130
6538
  contextWindow: number;
6131
6539
  maxTokens: number;
6132
6540
  };
6133
- readonly "gpt-5.2-codex": {
6541
+ readonly "claude-sonnet-4-6": {
6134
6542
  id: string;
6135
6543
  name: string;
6136
- api: "openai-codex-responses";
6544
+ api: "anthropic-messages";
6137
6545
  provider: string;
6138
6546
  baseUrl: string;
6139
6547
  reasoning: true;
@@ -6147,10 +6555,10 @@ export declare const MODELS: {
6147
6555
  contextWindow: number;
6148
6556
  maxTokens: number;
6149
6557
  };
6150
- readonly "gpt-5.3-codex": {
6558
+ readonly "gemini-3-flash": {
6151
6559
  id: string;
6152
6560
  name: string;
6153
- api: "openai-codex-responses";
6561
+ api: "google-generative-ai";
6154
6562
  provider: string;
6155
6563
  baseUrl: string;
6156
6564
  reasoning: true;
@@ -6164,14 +6572,14 @@ export declare const MODELS: {
6164
6572
  contextWindow: number;
6165
6573
  maxTokens: number;
6166
6574
  };
6167
- readonly "gpt-5.3-codex-spark": {
6575
+ readonly "gemini-3.1-pro": {
6168
6576
  id: string;
6169
6577
  name: string;
6170
- api: "openai-codex-responses";
6578
+ api: "google-generative-ai";
6171
6579
  provider: string;
6172
6580
  baseUrl: string;
6173
6581
  reasoning: true;
6174
- input: "text"[];
6582
+ input: ("image" | "text")[];
6175
6583
  cost: {
6176
6584
  input: number;
6177
6585
  output: number;
@@ -6181,14 +6589,14 @@ export declare const MODELS: {
6181
6589
  contextWindow: number;
6182
6590
  maxTokens: number;
6183
6591
  };
6184
- readonly "gpt-5.4": {
6592
+ readonly "glm-5": {
6185
6593
  id: string;
6186
6594
  name: string;
6187
- api: "openai-codex-responses";
6595
+ api: "openai-completions";
6188
6596
  provider: string;
6189
6597
  baseUrl: string;
6190
6598
  reasoning: true;
6191
- input: ("image" | "text")[];
6599
+ input: "text"[];
6192
6600
  cost: {
6193
6601
  input: number;
6194
6602
  output: number;
@@ -6198,14 +6606,14 @@ export declare const MODELS: {
6198
6606
  contextWindow: number;
6199
6607
  maxTokens: number;
6200
6608
  };
6201
- readonly "gpt-5.4-mini": {
6609
+ readonly "glm-5.1": {
6202
6610
  id: string;
6203
6611
  name: string;
6204
- api: "openai-codex-responses";
6612
+ api: "openai-completions";
6205
6613
  provider: string;
6206
6614
  baseUrl: string;
6207
6615
  reasoning: true;
6208
- input: ("image" | "text")[];
6616
+ input: "text"[];
6209
6617
  cost: {
6210
6618
  input: number;
6211
6619
  output: number;
@@ -6215,16 +6623,14 @@ export declare const MODELS: {
6215
6623
  contextWindow: number;
6216
6624
  maxTokens: number;
6217
6625
  };
6218
- };
6219
- readonly opencode: {
6220
- readonly "big-pickle": {
6626
+ readonly "gpt-5": {
6221
6627
  id: string;
6222
6628
  name: string;
6223
- api: "anthropic-messages";
6629
+ api: "openai-responses";
6224
6630
  provider: string;
6225
6631
  baseUrl: string;
6226
6632
  reasoning: true;
6227
- input: "text"[];
6633
+ input: ("image" | "text")[];
6228
6634
  cost: {
6229
6635
  input: number;
6230
6636
  output: number;
@@ -6234,13 +6640,13 @@ export declare const MODELS: {
6234
6640
  contextWindow: number;
6235
6641
  maxTokens: number;
6236
6642
  };
6237
- readonly "claude-3-5-haiku": {
6643
+ readonly "gpt-5-codex": {
6238
6644
  id: string;
6239
6645
  name: string;
6240
- api: "anthropic-messages";
6646
+ api: "openai-responses";
6241
6647
  provider: string;
6242
6648
  baseUrl: string;
6243
- reasoning: false;
6649
+ reasoning: true;
6244
6650
  input: ("image" | "text")[];
6245
6651
  cost: {
6246
6652
  input: number;
@@ -6251,10 +6657,10 @@ export declare const MODELS: {
6251
6657
  contextWindow: number;
6252
6658
  maxTokens: number;
6253
6659
  };
6254
- readonly "claude-haiku-4-5": {
6660
+ readonly "gpt-5-nano": {
6255
6661
  id: string;
6256
6662
  name: string;
6257
- api: "anthropic-messages";
6663
+ api: "openai-responses";
6258
6664
  provider: string;
6259
6665
  baseUrl: string;
6260
6666
  reasoning: true;
@@ -6268,10 +6674,10 @@ export declare const MODELS: {
6268
6674
  contextWindow: number;
6269
6675
  maxTokens: number;
6270
6676
  };
6271
- readonly "claude-opus-4-1": {
6677
+ readonly "gpt-5.1": {
6272
6678
  id: string;
6273
6679
  name: string;
6274
- api: "anthropic-messages";
6680
+ api: "openai-responses";
6275
6681
  provider: string;
6276
6682
  baseUrl: string;
6277
6683
  reasoning: true;
@@ -6285,10 +6691,10 @@ export declare const MODELS: {
6285
6691
  contextWindow: number;
6286
6692
  maxTokens: number;
6287
6693
  };
6288
- readonly "claude-opus-4-5": {
6694
+ readonly "gpt-5.1-codex": {
6289
6695
  id: string;
6290
6696
  name: string;
6291
- api: "anthropic-messages";
6697
+ api: "openai-responses";
6292
6698
  provider: string;
6293
6699
  baseUrl: string;
6294
6700
  reasoning: true;
@@ -6302,10 +6708,10 @@ export declare const MODELS: {
6302
6708
  contextWindow: number;
6303
6709
  maxTokens: number;
6304
6710
  };
6305
- readonly "claude-opus-4-6": {
6711
+ readonly "gpt-5.1-codex-max": {
6306
6712
  id: string;
6307
6713
  name: string;
6308
- api: "anthropic-messages";
6714
+ api: "openai-responses";
6309
6715
  provider: string;
6310
6716
  baseUrl: string;
6311
6717
  reasoning: true;
@@ -6319,10 +6725,10 @@ export declare const MODELS: {
6319
6725
  contextWindow: number;
6320
6726
  maxTokens: number;
6321
6727
  };
6322
- readonly "claude-sonnet-4": {
6728
+ readonly "gpt-5.1-codex-mini": {
6323
6729
  id: string;
6324
6730
  name: string;
6325
- api: "anthropic-messages";
6731
+ api: "openai-responses";
6326
6732
  provider: string;
6327
6733
  baseUrl: string;
6328
6734
  reasoning: true;
@@ -6336,10 +6742,10 @@ export declare const MODELS: {
6336
6742
  contextWindow: number;
6337
6743
  maxTokens: number;
6338
6744
  };
6339
- readonly "claude-sonnet-4-5": {
6745
+ readonly "gpt-5.2": {
6340
6746
  id: string;
6341
6747
  name: string;
6342
- api: "anthropic-messages";
6748
+ api: "openai-responses";
6343
6749
  provider: string;
6344
6750
  baseUrl: string;
6345
6751
  reasoning: true;
@@ -6353,10 +6759,10 @@ export declare const MODELS: {
6353
6759
  contextWindow: number;
6354
6760
  maxTokens: number;
6355
6761
  };
6356
- readonly "claude-sonnet-4-6": {
6762
+ readonly "gpt-5.2-codex": {
6357
6763
  id: string;
6358
6764
  name: string;
6359
- api: "anthropic-messages";
6765
+ api: "openai-responses";
6360
6766
  provider: string;
6361
6767
  baseUrl: string;
6362
6768
  reasoning: true;
@@ -6370,10 +6776,10 @@ export declare const MODELS: {
6370
6776
  contextWindow: number;
6371
6777
  maxTokens: number;
6372
6778
  };
6373
- readonly "gemini-3-flash": {
6779
+ readonly "gpt-5.3-codex": {
6374
6780
  id: string;
6375
6781
  name: string;
6376
- api: "google-generative-ai";
6782
+ api: "openai-responses";
6377
6783
  provider: string;
6378
6784
  baseUrl: string;
6379
6785
  reasoning: true;
@@ -6387,10 +6793,10 @@ export declare const MODELS: {
6387
6793
  contextWindow: number;
6388
6794
  maxTokens: number;
6389
6795
  };
6390
- readonly "gemini-3.1-pro": {
6796
+ readonly "gpt-5.4": {
6391
6797
  id: string;
6392
6798
  name: string;
6393
- api: "google-generative-ai";
6799
+ api: "openai-responses";
6394
6800
  provider: string;
6395
6801
  baseUrl: string;
6396
6802
  reasoning: true;
@@ -6404,14 +6810,14 @@ export declare const MODELS: {
6404
6810
  contextWindow: number;
6405
6811
  maxTokens: number;
6406
6812
  };
6407
- readonly "glm-5": {
6813
+ readonly "gpt-5.4-mini": {
6408
6814
  id: string;
6409
6815
  name: string;
6410
- api: "openai-completions";
6816
+ api: "openai-responses";
6411
6817
  provider: string;
6412
6818
  baseUrl: string;
6413
6819
  reasoning: true;
6414
- input: "text"[];
6820
+ input: ("image" | "text")[];
6415
6821
  cost: {
6416
6822
  input: number;
6417
6823
  output: number;
@@ -6421,7 +6827,7 @@ export declare const MODELS: {
6421
6827
  contextWindow: number;
6422
6828
  maxTokens: number;
6423
6829
  };
6424
- readonly "gpt-5": {
6830
+ readonly "gpt-5.4-nano": {
6425
6831
  id: string;
6426
6832
  name: string;
6427
6833
  api: "openai-responses";
@@ -6438,7 +6844,7 @@ export declare const MODELS: {
6438
6844
  contextWindow: number;
6439
6845
  maxTokens: number;
6440
6846
  };
6441
- readonly "gpt-5-codex": {
6847
+ readonly "gpt-5.4-pro": {
6442
6848
  id: string;
6443
6849
  name: string;
6444
6850
  api: "openai-responses";
@@ -6455,14 +6861,14 @@ export declare const MODELS: {
6455
6861
  contextWindow: number;
6456
6862
  maxTokens: number;
6457
6863
  };
6458
- readonly "gpt-5-nano": {
6864
+ readonly "hy3-preview-free": {
6459
6865
  id: string;
6460
6866
  name: string;
6461
- api: "openai-responses";
6867
+ api: "openai-completions";
6462
6868
  provider: string;
6463
6869
  baseUrl: string;
6464
6870
  reasoning: true;
6465
- input: ("image" | "text")[];
6871
+ input: "text"[];
6466
6872
  cost: {
6467
6873
  input: number;
6468
6874
  output: number;
@@ -6472,10 +6878,10 @@ export declare const MODELS: {
6472
6878
  contextWindow: number;
6473
6879
  maxTokens: number;
6474
6880
  };
6475
- readonly "gpt-5.1": {
6881
+ readonly "kimi-k2.5": {
6476
6882
  id: string;
6477
6883
  name: string;
6478
- api: "openai-responses";
6884
+ api: "openai-completions";
6479
6885
  provider: string;
6480
6886
  baseUrl: string;
6481
6887
  reasoning: true;
@@ -6489,10 +6895,10 @@ export declare const MODELS: {
6489
6895
  contextWindow: number;
6490
6896
  maxTokens: number;
6491
6897
  };
6492
- readonly "gpt-5.1-codex": {
6898
+ readonly "kimi-k2.6": {
6493
6899
  id: string;
6494
6900
  name: string;
6495
- api: "openai-responses";
6901
+ api: "openai-completions";
6496
6902
  provider: string;
6497
6903
  baseUrl: string;
6498
6904
  reasoning: true;
@@ -6506,14 +6912,14 @@ export declare const MODELS: {
6506
6912
  contextWindow: number;
6507
6913
  maxTokens: number;
6508
6914
  };
6509
- readonly "gpt-5.1-codex-max": {
6915
+ readonly "ling-2.6-flash-free": {
6510
6916
  id: string;
6511
6917
  name: string;
6512
- api: "openai-responses";
6918
+ api: "openai-completions";
6513
6919
  provider: string;
6514
6920
  baseUrl: string;
6515
- reasoning: true;
6516
- input: ("image" | "text")[];
6921
+ reasoning: false;
6922
+ input: "text"[];
6517
6923
  cost: {
6518
6924
  input: number;
6519
6925
  output: number;
@@ -6523,14 +6929,14 @@ export declare const MODELS: {
6523
6929
  contextWindow: number;
6524
6930
  maxTokens: number;
6525
6931
  };
6526
- readonly "gpt-5.1-codex-mini": {
6932
+ readonly "minimax-m2.5": {
6527
6933
  id: string;
6528
6934
  name: string;
6529
- api: "openai-responses";
6935
+ api: "openai-completions";
6530
6936
  provider: string;
6531
6937
  baseUrl: string;
6532
6938
  reasoning: true;
6533
- input: ("image" | "text")[];
6939
+ input: "text"[];
6534
6940
  cost: {
6535
6941
  input: number;
6536
6942
  output: number;
@@ -6540,14 +6946,14 @@ export declare const MODELS: {
6540
6946
  contextWindow: number;
6541
6947
  maxTokens: number;
6542
6948
  };
6543
- readonly "gpt-5.2": {
6949
+ readonly "minimax-m2.5-free": {
6544
6950
  id: string;
6545
6951
  name: string;
6546
- api: "openai-responses";
6952
+ api: "anthropic-messages";
6547
6953
  provider: string;
6548
6954
  baseUrl: string;
6549
6955
  reasoning: true;
6550
- input: ("image" | "text")[];
6956
+ input: "text"[];
6551
6957
  cost: {
6552
6958
  input: number;
6553
6959
  output: number;
@@ -6557,14 +6963,14 @@ export declare const MODELS: {
6557
6963
  contextWindow: number;
6558
6964
  maxTokens: number;
6559
6965
  };
6560
- readonly "gpt-5.2-codex": {
6966
+ readonly "minimax-m2.7": {
6561
6967
  id: string;
6562
6968
  name: string;
6563
- api: "openai-responses";
6969
+ api: "openai-completions";
6564
6970
  provider: string;
6565
6971
  baseUrl: string;
6566
6972
  reasoning: true;
6567
- input: ("image" | "text")[];
6973
+ input: "text"[];
6568
6974
  cost: {
6569
6975
  input: number;
6570
6976
  output: number;
@@ -6574,14 +6980,14 @@ export declare const MODELS: {
6574
6980
  contextWindow: number;
6575
6981
  maxTokens: number;
6576
6982
  };
6577
- readonly "gpt-5.3-codex": {
6983
+ readonly "nemotron-3-super-free": {
6578
6984
  id: string;
6579
6985
  name: string;
6580
- api: "openai-responses";
6986
+ api: "openai-completions";
6581
6987
  provider: string;
6582
6988
  baseUrl: string;
6583
6989
  reasoning: true;
6584
- input: ("image" | "text")[];
6990
+ input: "text"[];
6585
6991
  cost: {
6586
6992
  input: number;
6587
6993
  output: number;
@@ -6591,10 +6997,10 @@ export declare const MODELS: {
6591
6997
  contextWindow: number;
6592
6998
  maxTokens: number;
6593
6999
  };
6594
- readonly "gpt-5.4": {
7000
+ readonly "qwen3.5-plus": {
6595
7001
  id: string;
6596
7002
  name: string;
6597
- api: "openai-responses";
7003
+ api: "anthropic-messages";
6598
7004
  provider: string;
6599
7005
  baseUrl: string;
6600
7006
  reasoning: true;
@@ -6608,10 +7014,10 @@ export declare const MODELS: {
6608
7014
  contextWindow: number;
6609
7015
  maxTokens: number;
6610
7016
  };
6611
- readonly "gpt-5.4-mini": {
7017
+ readonly "qwen3.6-plus": {
6612
7018
  id: string;
6613
7019
  name: string;
6614
- api: "openai-responses";
7020
+ api: "anthropic-messages";
6615
7021
  provider: string;
6616
7022
  baseUrl: string;
6617
7023
  reasoning: true;
@@ -6625,14 +7031,16 @@ export declare const MODELS: {
6625
7031
  contextWindow: number;
6626
7032
  maxTokens: number;
6627
7033
  };
6628
- readonly "gpt-5.4-nano": {
7034
+ };
7035
+ readonly "opencode-go": {
7036
+ readonly "glm-5": {
6629
7037
  id: string;
6630
7038
  name: string;
6631
- api: "openai-responses";
7039
+ api: "openai-completions";
6632
7040
  provider: string;
6633
7041
  baseUrl: string;
6634
7042
  reasoning: true;
6635
- input: ("image" | "text")[];
7043
+ input: "text"[];
6636
7044
  cost: {
6637
7045
  input: number;
6638
7046
  output: number;
@@ -6642,14 +7050,14 @@ export declare const MODELS: {
6642
7050
  contextWindow: number;
6643
7051
  maxTokens: number;
6644
7052
  };
6645
- readonly "gpt-5.4-pro": {
7053
+ readonly "glm-5.1": {
6646
7054
  id: string;
6647
7055
  name: string;
6648
- api: "openai-responses";
7056
+ api: "openai-completions";
6649
7057
  provider: string;
6650
7058
  baseUrl: string;
6651
7059
  reasoning: true;
6652
- input: ("image" | "text")[];
7060
+ input: "text"[];
6653
7061
  cost: {
6654
7062
  input: number;
6655
7063
  output: number;
@@ -6676,7 +7084,7 @@ export declare const MODELS: {
6676
7084
  contextWindow: number;
6677
7085
  maxTokens: number;
6678
7086
  };
6679
- readonly "mimo-v2-omni-free": {
7087
+ readonly "kimi-k2.6": {
6680
7088
  id: string;
6681
7089
  name: string;
6682
7090
  api: "openai-completions";
@@ -6693,14 +7101,14 @@ export declare const MODELS: {
6693
7101
  contextWindow: number;
6694
7102
  maxTokens: number;
6695
7103
  };
6696
- readonly "mimo-v2-pro-free": {
7104
+ readonly "mimo-v2-omni": {
6697
7105
  id: string;
6698
7106
  name: string;
6699
7107
  api: "openai-completions";
6700
7108
  provider: string;
6701
7109
  baseUrl: string;
6702
7110
  reasoning: true;
6703
- input: "text"[];
7111
+ input: ("image" | "text")[];
6704
7112
  cost: {
6705
7113
  input: number;
6706
7114
  output: number;
@@ -6710,7 +7118,7 @@ export declare const MODELS: {
6710
7118
  contextWindow: number;
6711
7119
  maxTokens: number;
6712
7120
  };
6713
- readonly "minimax-m2.5": {
7121
+ readonly "mimo-v2-pro": {
6714
7122
  id: string;
6715
7123
  name: string;
6716
7124
  api: "openai-completions";
@@ -6727,14 +7135,14 @@ export declare const MODELS: {
6727
7135
  contextWindow: number;
6728
7136
  maxTokens: number;
6729
7137
  };
6730
- readonly "minimax-m2.5-free": {
7138
+ readonly "mimo-v2.5": {
6731
7139
  id: string;
6732
7140
  name: string;
6733
- api: "anthropic-messages";
7141
+ api: "openai-completions";
6734
7142
  provider: string;
6735
7143
  baseUrl: string;
6736
7144
  reasoning: true;
6737
- input: "text"[];
7145
+ input: ("image" | "text")[];
6738
7146
  cost: {
6739
7147
  input: number;
6740
7148
  output: number;
@@ -6744,7 +7152,7 @@ export declare const MODELS: {
6744
7152
  contextWindow: number;
6745
7153
  maxTokens: number;
6746
7154
  };
6747
- readonly "nemotron-3-super-free": {
7155
+ readonly "mimo-v2.5-pro": {
6748
7156
  id: string;
6749
7157
  name: string;
6750
7158
  api: "openai-completions";
@@ -6761,9 +7169,7 @@ export declare const MODELS: {
6761
7169
  contextWindow: number;
6762
7170
  maxTokens: number;
6763
7171
  };
6764
- };
6765
- readonly "opencode-go": {
6766
- readonly "glm-5": {
7172
+ readonly "minimax-m2.5": {
6767
7173
  id: string;
6768
7174
  name: string;
6769
7175
  api: "openai-completions";
@@ -6780,14 +7186,14 @@ export declare const MODELS: {
6780
7186
  contextWindow: number;
6781
7187
  maxTokens: number;
6782
7188
  };
6783
- readonly "kimi-k2.5": {
7189
+ readonly "minimax-m2.7": {
6784
7190
  id: string;
6785
7191
  name: string;
6786
- api: "openai-completions";
7192
+ api: "anthropic-messages";
6787
7193
  provider: string;
6788
7194
  baseUrl: string;
6789
7195
  reasoning: true;
6790
- input: ("image" | "text")[];
7196
+ input: "text"[];
6791
7197
  cost: {
6792
7198
  input: number;
6793
7199
  output: number;
@@ -6797,14 +7203,14 @@ export declare const MODELS: {
6797
7203
  contextWindow: number;
6798
7204
  maxTokens: number;
6799
7205
  };
6800
- readonly "minimax-m2.5": {
7206
+ readonly "qwen3.5-plus": {
6801
7207
  id: string;
6802
7208
  name: string;
6803
7209
  api: "anthropic-messages";
6804
7210
  provider: string;
6805
7211
  baseUrl: string;
6806
7212
  reasoning: true;
6807
- input: "text"[];
7213
+ input: ("image" | "text")[];
6808
7214
  cost: {
6809
7215
  input: number;
6810
7216
  output: number;
@@ -6814,14 +7220,14 @@ export declare const MODELS: {
6814
7220
  contextWindow: number;
6815
7221
  maxTokens: number;
6816
7222
  };
6817
- readonly "minimax-m2.7": {
7223
+ readonly "qwen3.6-plus": {
6818
7224
  id: string;
6819
7225
  name: string;
6820
7226
  api: "anthropic-messages";
6821
7227
  provider: string;
6822
7228
  baseUrl: string;
6823
7229
  reasoning: true;
6824
- input: "text"[];
7230
+ input: ("image" | "text")[];
6825
7231
  cost: {
6826
7232
  input: number;
6827
7233
  output: number;
@@ -7003,13 +7409,30 @@ export declare const MODELS: {
7003
7409
  contextWindow: number;
7004
7410
  maxTokens: number;
7005
7411
  };
7006
- readonly "anthropic/claude-3.5-sonnet": {
7412
+ readonly "anthropic/claude-3.7-sonnet": {
7413
+ id: string;
7414
+ name: string;
7415
+ api: "openai-completions";
7416
+ provider: string;
7417
+ baseUrl: string;
7418
+ reasoning: true;
7419
+ input: ("image" | "text")[];
7420
+ cost: {
7421
+ input: number;
7422
+ output: number;
7423
+ cacheRead: number;
7424
+ cacheWrite: number;
7425
+ };
7426
+ contextWindow: number;
7427
+ maxTokens: number;
7428
+ };
7429
+ readonly "anthropic/claude-3.7-sonnet:thinking": {
7007
7430
  id: string;
7008
7431
  name: string;
7009
7432
  api: "openai-completions";
7010
7433
  provider: string;
7011
7434
  baseUrl: string;
7012
- reasoning: false;
7435
+ reasoning: true;
7013
7436
  input: ("image" | "text")[];
7014
7437
  cost: {
7015
7438
  input: number;
@@ -7020,7 +7443,7 @@ export declare const MODELS: {
7020
7443
  contextWindow: number;
7021
7444
  maxTokens: number;
7022
7445
  };
7023
- readonly "anthropic/claude-3.7-sonnet": {
7446
+ readonly "anthropic/claude-haiku-4.5": {
7024
7447
  id: string;
7025
7448
  name: string;
7026
7449
  api: "openai-completions";
@@ -7037,7 +7460,7 @@ export declare const MODELS: {
7037
7460
  contextWindow: number;
7038
7461
  maxTokens: number;
7039
7462
  };
7040
- readonly "anthropic/claude-3.7-sonnet:thinking": {
7463
+ readonly "anthropic/claude-opus-4": {
7041
7464
  id: string;
7042
7465
  name: string;
7043
7466
  api: "openai-completions";
@@ -7054,7 +7477,7 @@ export declare const MODELS: {
7054
7477
  contextWindow: number;
7055
7478
  maxTokens: number;
7056
7479
  };
7057
- readonly "anthropic/claude-haiku-4.5": {
7480
+ readonly "anthropic/claude-opus-4.1": {
7058
7481
  id: string;
7059
7482
  name: string;
7060
7483
  api: "openai-completions";
@@ -7071,7 +7494,7 @@ export declare const MODELS: {
7071
7494
  contextWindow: number;
7072
7495
  maxTokens: number;
7073
7496
  };
7074
- readonly "anthropic/claude-opus-4": {
7497
+ readonly "anthropic/claude-opus-4.5": {
7075
7498
  id: string;
7076
7499
  name: string;
7077
7500
  api: "openai-completions";
@@ -7088,7 +7511,7 @@ export declare const MODELS: {
7088
7511
  contextWindow: number;
7089
7512
  maxTokens: number;
7090
7513
  };
7091
- readonly "anthropic/claude-opus-4.1": {
7514
+ readonly "anthropic/claude-opus-4.6": {
7092
7515
  id: string;
7093
7516
  name: string;
7094
7517
  api: "openai-completions";
@@ -7105,7 +7528,7 @@ export declare const MODELS: {
7105
7528
  contextWindow: number;
7106
7529
  maxTokens: number;
7107
7530
  };
7108
- readonly "anthropic/claude-opus-4.5": {
7531
+ readonly "anthropic/claude-opus-4.6-fast": {
7109
7532
  id: string;
7110
7533
  name: string;
7111
7534
  api: "openai-completions";
@@ -7122,7 +7545,7 @@ export declare const MODELS: {
7122
7545
  contextWindow: number;
7123
7546
  maxTokens: number;
7124
7547
  };
7125
- readonly "anthropic/claude-opus-4.6": {
7548
+ readonly "anthropic/claude-opus-4.7": {
7126
7549
  id: string;
7127
7550
  name: string;
7128
7551
  api: "openai-completions";
@@ -7190,7 +7613,7 @@ export declare const MODELS: {
7190
7613
  contextWindow: number;
7191
7614
  maxTokens: number;
7192
7615
  };
7193
- readonly "arcee-ai/trinity-large-preview:free": {
7616
+ readonly "arcee-ai/trinity-large-preview": {
7194
7617
  id: string;
7195
7618
  name: string;
7196
7619
  api: "openai-completions";
@@ -7207,7 +7630,7 @@ export declare const MODELS: {
7207
7630
  contextWindow: number;
7208
7631
  maxTokens: number;
7209
7632
  };
7210
- readonly "arcee-ai/trinity-mini": {
7633
+ readonly "arcee-ai/trinity-large-thinking": {
7211
7634
  id: string;
7212
7635
  name: string;
7213
7636
  api: "openai-completions";
@@ -7224,7 +7647,7 @@ export declare const MODELS: {
7224
7647
  contextWindow: number;
7225
7648
  maxTokens: number;
7226
7649
  };
7227
- readonly "arcee-ai/trinity-mini:free": {
7650
+ readonly "arcee-ai/trinity-mini": {
7228
7651
  id: string;
7229
7652
  name: string;
7230
7653
  api: "openai-completions";
@@ -7434,7 +7857,7 @@ export declare const MODELS: {
7434
7857
  api: "openai-completions";
7435
7858
  provider: string;
7436
7859
  baseUrl: string;
7437
- reasoning: true;
7860
+ reasoning: false;
7438
7861
  input: "text"[];
7439
7862
  cost: {
7440
7863
  input: number;
@@ -7717,7 +8140,7 @@ export declare const MODELS: {
7717
8140
  contextWindow: number;
7718
8141
  maxTokens: number;
7719
8142
  };
7720
- readonly "google/gemini-3-pro-preview": {
8143
+ readonly "google/gemini-3.1-flash-lite-preview": {
7721
8144
  id: string;
7722
8145
  name: string;
7723
8146
  api: "openai-completions";
@@ -7734,7 +8157,7 @@ export declare const MODELS: {
7734
8157
  contextWindow: number;
7735
8158
  maxTokens: number;
7736
8159
  };
7737
- readonly "google/gemini-3.1-flash-lite-preview": {
8160
+ readonly "google/gemini-3.1-pro-preview": {
7738
8161
  id: string;
7739
8162
  name: string;
7740
8163
  api: "openai-completions";
@@ -7751,7 +8174,7 @@ export declare const MODELS: {
7751
8174
  contextWindow: number;
7752
8175
  maxTokens: number;
7753
8176
  };
7754
- readonly "google/gemini-3.1-pro-preview": {
8177
+ readonly "google/gemini-3.1-pro-preview-customtools": {
7755
8178
  id: string;
7756
8179
  name: string;
7757
8180
  api: "openai-completions";
@@ -7768,7 +8191,7 @@ export declare const MODELS: {
7768
8191
  contextWindow: number;
7769
8192
  maxTokens: number;
7770
8193
  };
7771
- readonly "google/gemini-3.1-pro-preview-customtools": {
8194
+ readonly "google/gemma-4-26b-a4b-it": {
7772
8195
  id: string;
7773
8196
  name: string;
7774
8197
  api: "openai-completions";
@@ -7785,14 +8208,14 @@ export declare const MODELS: {
7785
8208
  contextWindow: number;
7786
8209
  maxTokens: number;
7787
8210
  };
7788
- readonly "inception/mercury": {
8211
+ readonly "google/gemma-4-26b-a4b-it:free": {
7789
8212
  id: string;
7790
8213
  name: string;
7791
8214
  api: "openai-completions";
7792
8215
  provider: string;
7793
8216
  baseUrl: string;
7794
- reasoning: false;
7795
- input: "text"[];
8217
+ reasoning: true;
8218
+ input: ("image" | "text")[];
7796
8219
  cost: {
7797
8220
  input: number;
7798
8221
  output: number;
@@ -7802,14 +8225,14 @@ export declare const MODELS: {
7802
8225
  contextWindow: number;
7803
8226
  maxTokens: number;
7804
8227
  };
7805
- readonly "inception/mercury-2": {
8228
+ readonly "google/gemma-4-31b-it": {
7806
8229
  id: string;
7807
8230
  name: string;
7808
8231
  api: "openai-completions";
7809
8232
  provider: string;
7810
8233
  baseUrl: string;
7811
8234
  reasoning: true;
7812
- input: "text"[];
8235
+ input: ("image" | "text")[];
7813
8236
  cost: {
7814
8237
  input: number;
7815
8238
  output: number;
@@ -7819,14 +8242,14 @@ export declare const MODELS: {
7819
8242
  contextWindow: number;
7820
8243
  maxTokens: number;
7821
8244
  };
7822
- readonly "inception/mercury-coder": {
8245
+ readonly "google/gemma-4-31b-it:free": {
7823
8246
  id: string;
7824
8247
  name: string;
7825
8248
  api: "openai-completions";
7826
8249
  provider: string;
7827
8250
  baseUrl: string;
7828
- reasoning: false;
7829
- input: "text"[];
8251
+ reasoning: true;
8252
+ input: ("image" | "text")[];
7830
8253
  cost: {
7831
8254
  input: number;
7832
8255
  output: number;
@@ -7836,13 +8259,13 @@ export declare const MODELS: {
7836
8259
  contextWindow: number;
7837
8260
  maxTokens: number;
7838
8261
  };
7839
- readonly "kwaipilot/kat-coder-pro": {
8262
+ readonly "inception/mercury-2": {
7840
8263
  id: string;
7841
8264
  name: string;
7842
8265
  api: "openai-completions";
7843
8266
  provider: string;
7844
8267
  baseUrl: string;
7845
- reasoning: false;
8268
+ reasoning: true;
7846
8269
  input: "text"[];
7847
8270
  cost: {
7848
8271
  input: number;
@@ -7853,7 +8276,7 @@ export declare const MODELS: {
7853
8276
  contextWindow: number;
7854
8277
  maxTokens: number;
7855
8278
  };
7856
- readonly "meituan/longcat-flash-chat": {
8279
+ readonly "inclusionai/ling-2.6-flash:free": {
7857
8280
  id: string;
7858
8281
  name: string;
7859
8282
  api: "openai-completions";
@@ -7870,7 +8293,7 @@ export declare const MODELS: {
7870
8293
  contextWindow: number;
7871
8294
  maxTokens: number;
7872
8295
  };
7873
- readonly "meta-llama/llama-3-8b-instruct": {
8296
+ readonly "kwaipilot/kat-coder-pro-v2": {
7874
8297
  id: string;
7875
8298
  name: string;
7876
8299
  api: "openai-completions";
@@ -7887,7 +8310,7 @@ export declare const MODELS: {
7887
8310
  contextWindow: number;
7888
8311
  maxTokens: number;
7889
8312
  };
7890
- readonly "meta-llama/llama-3.1-70b-instruct": {
8313
+ readonly "meta-llama/llama-3-8b-instruct": {
7891
8314
  id: string;
7892
8315
  name: string;
7893
8316
  api: "openai-completions";
@@ -7904,7 +8327,7 @@ export declare const MODELS: {
7904
8327
  contextWindow: number;
7905
8328
  maxTokens: number;
7906
8329
  };
7907
- readonly "meta-llama/llama-3.1-8b-instruct": {
8330
+ readonly "meta-llama/llama-3.1-70b-instruct": {
7908
8331
  id: string;
7909
8332
  name: string;
7910
8333
  api: "openai-completions";
@@ -7921,7 +8344,7 @@ export declare const MODELS: {
7921
8344
  contextWindow: number;
7922
8345
  maxTokens: number;
7923
8346
  };
7924
- readonly "meta-llama/llama-3.3-70b-instruct": {
8347
+ readonly "meta-llama/llama-3.1-8b-instruct": {
7925
8348
  id: string;
7926
8349
  name: string;
7927
8350
  api: "openai-completions";
@@ -7938,7 +8361,7 @@ export declare const MODELS: {
7938
8361
  contextWindow: number;
7939
8362
  maxTokens: number;
7940
8363
  };
7941
- readonly "meta-llama/llama-3.3-70b-instruct:free": {
8364
+ readonly "meta-llama/llama-3.3-70b-instruct": {
7942
8365
  id: string;
7943
8366
  name: string;
7944
8367
  api: "openai-completions";
@@ -7955,14 +8378,14 @@ export declare const MODELS: {
7955
8378
  contextWindow: number;
7956
8379
  maxTokens: number;
7957
8380
  };
7958
- readonly "meta-llama/llama-4-maverick": {
8381
+ readonly "meta-llama/llama-3.3-70b-instruct:free": {
7959
8382
  id: string;
7960
8383
  name: string;
7961
8384
  api: "openai-completions";
7962
8385
  provider: string;
7963
8386
  baseUrl: string;
7964
8387
  reasoning: false;
7965
- input: ("image" | "text")[];
8388
+ input: "text"[];
7966
8389
  cost: {
7967
8390
  input: number;
7968
8391
  output: number;
@@ -8346,23 +8769,6 @@ export declare const MODELS: {
8346
8769
  contextWindow: number;
8347
8770
  maxTokens: number;
8348
8771
  };
8349
- readonly "mistralai/mistral-small-24b-instruct-2501": {
8350
- id: string;
8351
- name: string;
8352
- api: "openai-completions";
8353
- provider: string;
8354
- baseUrl: string;
8355
- reasoning: false;
8356
- input: "text"[];
8357
- cost: {
8358
- input: number;
8359
- output: number;
8360
- cacheRead: number;
8361
- cacheWrite: number;
8362
- };
8363
- contextWindow: number;
8364
- maxTokens: number;
8365
- };
8366
8772
  readonly "mistralai/mistral-small-2603": {
8367
8773
  id: string;
8368
8774
  name: string;
@@ -8380,7 +8786,7 @@ export declare const MODELS: {
8380
8786
  contextWindow: number;
8381
8787
  maxTokens: number;
8382
8788
  };
8383
- readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
8789
+ readonly "mistralai/mistral-small-3.2-24b-instruct": {
8384
8790
  id: string;
8385
8791
  name: string;
8386
8792
  api: "openai-completions";
@@ -8397,14 +8803,14 @@ export declare const MODELS: {
8397
8803
  contextWindow: number;
8398
8804
  maxTokens: number;
8399
8805
  };
8400
- readonly "mistralai/mistral-small-3.2-24b-instruct": {
8806
+ readonly "mistralai/mistral-small-creative": {
8401
8807
  id: string;
8402
8808
  name: string;
8403
8809
  api: "openai-completions";
8404
8810
  provider: string;
8405
8811
  baseUrl: string;
8406
8812
  reasoning: false;
8407
- input: ("image" | "text")[];
8813
+ input: "text"[];
8408
8814
  cost: {
8409
8815
  input: number;
8410
8816
  output: number;
@@ -8414,7 +8820,7 @@ export declare const MODELS: {
8414
8820
  contextWindow: number;
8415
8821
  maxTokens: number;
8416
8822
  };
8417
- readonly "mistralai/mistral-small-creative": {
8823
+ readonly "mistralai/mixtral-8x22b-instruct": {
8418
8824
  id: string;
8419
8825
  name: string;
8420
8826
  api: "openai-completions";
@@ -8431,7 +8837,7 @@ export declare const MODELS: {
8431
8837
  contextWindow: number;
8432
8838
  maxTokens: number;
8433
8839
  };
8434
- readonly "mistralai/mixtral-8x22b-instruct": {
8840
+ readonly "mistralai/mixtral-8x7b-instruct": {
8435
8841
  id: string;
8436
8842
  name: string;
8437
8843
  api: "openai-completions";
@@ -8448,14 +8854,14 @@ export declare const MODELS: {
8448
8854
  contextWindow: number;
8449
8855
  maxTokens: number;
8450
8856
  };
8451
- readonly "mistralai/mixtral-8x7b-instruct": {
8857
+ readonly "mistralai/pixtral-large-2411": {
8452
8858
  id: string;
8453
8859
  name: string;
8454
8860
  api: "openai-completions";
8455
8861
  provider: string;
8456
8862
  baseUrl: string;
8457
8863
  reasoning: false;
8458
- input: "text"[];
8864
+ input: ("image" | "text")[];
8459
8865
  cost: {
8460
8866
  input: number;
8461
8867
  output: number;
@@ -8465,14 +8871,14 @@ export declare const MODELS: {
8465
8871
  contextWindow: number;
8466
8872
  maxTokens: number;
8467
8873
  };
8468
- readonly "mistralai/pixtral-large-2411": {
8874
+ readonly "mistralai/voxtral-small-24b-2507": {
8469
8875
  id: string;
8470
8876
  name: string;
8471
8877
  api: "openai-completions";
8472
8878
  provider: string;
8473
8879
  baseUrl: string;
8474
8880
  reasoning: false;
8475
- input: ("image" | "text")[];
8881
+ input: "text"[];
8476
8882
  cost: {
8477
8883
  input: number;
8478
8884
  output: number;
@@ -8482,7 +8888,7 @@ export declare const MODELS: {
8482
8888
  contextWindow: number;
8483
8889
  maxTokens: number;
8484
8890
  };
8485
- readonly "mistralai/voxtral-small-24b-2507": {
8891
+ readonly "moonshotai/kimi-k2": {
8486
8892
  id: string;
8487
8893
  name: string;
8488
8894
  api: "openai-completions";
@@ -8499,7 +8905,7 @@ export declare const MODELS: {
8499
8905
  contextWindow: number;
8500
8906
  maxTokens: number;
8501
8907
  };
8502
- readonly "moonshotai/kimi-k2": {
8908
+ readonly "moonshotai/kimi-k2-0905": {
8503
8909
  id: string;
8504
8910
  name: string;
8505
8911
  api: "openai-completions";
@@ -8516,13 +8922,13 @@ export declare const MODELS: {
8516
8922
  contextWindow: number;
8517
8923
  maxTokens: number;
8518
8924
  };
8519
- readonly "moonshotai/kimi-k2-0905": {
8925
+ readonly "moonshotai/kimi-k2-thinking": {
8520
8926
  id: string;
8521
8927
  name: string;
8522
8928
  api: "openai-completions";
8523
8929
  provider: string;
8524
8930
  baseUrl: string;
8525
- reasoning: false;
8931
+ reasoning: true;
8526
8932
  input: "text"[];
8527
8933
  cost: {
8528
8934
  input: number;
@@ -8533,14 +8939,14 @@ export declare const MODELS: {
8533
8939
  contextWindow: number;
8534
8940
  maxTokens: number;
8535
8941
  };
8536
- readonly "moonshotai/kimi-k2-thinking": {
8942
+ readonly "moonshotai/kimi-k2.5": {
8537
8943
  id: string;
8538
8944
  name: string;
8539
8945
  api: "openai-completions";
8540
8946
  provider: string;
8541
8947
  baseUrl: string;
8542
8948
  reasoning: true;
8543
- input: "text"[];
8949
+ input: ("image" | "text")[];
8544
8950
  cost: {
8545
8951
  input: number;
8546
8952
  output: number;
@@ -8550,7 +8956,7 @@ export declare const MODELS: {
8550
8956
  contextWindow: number;
8551
8957
  maxTokens: number;
8552
8958
  };
8553
- readonly "moonshotai/kimi-k2.5": {
8959
+ readonly "moonshotai/kimi-k2.6": {
8554
8960
  id: string;
8555
8961
  name: string;
8556
8962
  api: "openai-completions";
@@ -9043,23 +9449,6 @@ export declare const MODELS: {
9043
9449
  contextWindow: number;
9044
9450
  maxTokens: number;
9045
9451
  };
9046
- readonly "openai/gpt-4o:extended": {
9047
- id: string;
9048
- name: string;
9049
- api: "openai-completions";
9050
- provider: string;
9051
- baseUrl: string;
9052
- reasoning: false;
9053
- input: ("image" | "text")[];
9054
- cost: {
9055
- input: number;
9056
- output: number;
9057
- cacheRead: number;
9058
- cacheWrite: number;
9059
- };
9060
- contextWindow: number;
9061
- maxTokens: number;
9062
- };
9063
9452
  readonly "openai/gpt-5": {
9064
9453
  id: string;
9065
9454
  name: string;
@@ -9094,7 +9483,7 @@ export declare const MODELS: {
9094
9483
  contextWindow: number;
9095
9484
  maxTokens: number;
9096
9485
  };
9097
- readonly "openai/gpt-5-image": {
9486
+ readonly "openai/gpt-5-mini": {
9098
9487
  id: string;
9099
9488
  name: string;
9100
9489
  api: "openai-completions";
@@ -9111,7 +9500,7 @@ export declare const MODELS: {
9111
9500
  contextWindow: number;
9112
9501
  maxTokens: number;
9113
9502
  };
9114
- readonly "openai/gpt-5-image-mini": {
9503
+ readonly "openai/gpt-5-nano": {
9115
9504
  id: string;
9116
9505
  name: string;
9117
9506
  api: "openai-completions";
@@ -9128,7 +9517,7 @@ export declare const MODELS: {
9128
9517
  contextWindow: number;
9129
9518
  maxTokens: number;
9130
9519
  };
9131
- readonly "openai/gpt-5-mini": {
9520
+ readonly "openai/gpt-5-pro": {
9132
9521
  id: string;
9133
9522
  name: string;
9134
9523
  api: "openai-completions";
@@ -9145,7 +9534,7 @@ export declare const MODELS: {
9145
9534
  contextWindow: number;
9146
9535
  maxTokens: number;
9147
9536
  };
9148
- readonly "openai/gpt-5-nano": {
9537
+ readonly "openai/gpt-5.1": {
9149
9538
  id: string;
9150
9539
  name: string;
9151
9540
  api: "openai-completions";
@@ -9162,13 +9551,13 @@ export declare const MODELS: {
9162
9551
  contextWindow: number;
9163
9552
  maxTokens: number;
9164
9553
  };
9165
- readonly "openai/gpt-5-pro": {
9554
+ readonly "openai/gpt-5.1-chat": {
9166
9555
  id: string;
9167
9556
  name: string;
9168
9557
  api: "openai-completions";
9169
9558
  provider: string;
9170
9559
  baseUrl: string;
9171
- reasoning: true;
9560
+ reasoning: false;
9172
9561
  input: ("image" | "text")[];
9173
9562
  cost: {
9174
9563
  input: number;
@@ -9179,7 +9568,7 @@ export declare const MODELS: {
9179
9568
  contextWindow: number;
9180
9569
  maxTokens: number;
9181
9570
  };
9182
- readonly "openai/gpt-5.1": {
9571
+ readonly "openai/gpt-5.1-codex": {
9183
9572
  id: string;
9184
9573
  name: string;
9185
9574
  api: "openai-completions";
@@ -9196,13 +9585,13 @@ export declare const MODELS: {
9196
9585
  contextWindow: number;
9197
9586
  maxTokens: number;
9198
9587
  };
9199
- readonly "openai/gpt-5.1-chat": {
9588
+ readonly "openai/gpt-5.1-codex-max": {
9200
9589
  id: string;
9201
9590
  name: string;
9202
9591
  api: "openai-completions";
9203
9592
  provider: string;
9204
9593
  baseUrl: string;
9205
- reasoning: false;
9594
+ reasoning: true;
9206
9595
  input: ("image" | "text")[];
9207
9596
  cost: {
9208
9597
  input: number;
@@ -9213,7 +9602,7 @@ export declare const MODELS: {
9213
9602
  contextWindow: number;
9214
9603
  maxTokens: number;
9215
9604
  };
9216
- readonly "openai/gpt-5.1-codex": {
9605
+ readonly "openai/gpt-5.1-codex-mini": {
9217
9606
  id: string;
9218
9607
  name: string;
9219
9608
  api: "openai-completions";
@@ -9230,7 +9619,7 @@ export declare const MODELS: {
9230
9619
  contextWindow: number;
9231
9620
  maxTokens: number;
9232
9621
  };
9233
- readonly "openai/gpt-5.1-codex-max": {
9622
+ readonly "openai/gpt-5.2": {
9234
9623
  id: string;
9235
9624
  name: string;
9236
9625
  api: "openai-completions";
@@ -9247,13 +9636,13 @@ export declare const MODELS: {
9247
9636
  contextWindow: number;
9248
9637
  maxTokens: number;
9249
9638
  };
9250
- readonly "openai/gpt-5.1-codex-mini": {
9639
+ readonly "openai/gpt-5.2-chat": {
9251
9640
  id: string;
9252
9641
  name: string;
9253
9642
  api: "openai-completions";
9254
9643
  provider: string;
9255
9644
  baseUrl: string;
9256
- reasoning: true;
9645
+ reasoning: false;
9257
9646
  input: ("image" | "text")[];
9258
9647
  cost: {
9259
9648
  input: number;
@@ -9264,7 +9653,7 @@ export declare const MODELS: {
9264
9653
  contextWindow: number;
9265
9654
  maxTokens: number;
9266
9655
  };
9267
- readonly "openai/gpt-5.2": {
9656
+ readonly "openai/gpt-5.2-codex": {
9268
9657
  id: string;
9269
9658
  name: string;
9270
9659
  api: "openai-completions";
@@ -9281,13 +9670,13 @@ export declare const MODELS: {
9281
9670
  contextWindow: number;
9282
9671
  maxTokens: number;
9283
9672
  };
9284
- readonly "openai/gpt-5.2-chat": {
9673
+ readonly "openai/gpt-5.2-pro": {
9285
9674
  id: string;
9286
9675
  name: string;
9287
9676
  api: "openai-completions";
9288
9677
  provider: string;
9289
9678
  baseUrl: string;
9290
- reasoning: false;
9679
+ reasoning: true;
9291
9680
  input: ("image" | "text")[];
9292
9681
  cost: {
9293
9682
  input: number;
@@ -9298,13 +9687,13 @@ export declare const MODELS: {
9298
9687
  contextWindow: number;
9299
9688
  maxTokens: number;
9300
9689
  };
9301
- readonly "openai/gpt-5.2-codex": {
9690
+ readonly "openai/gpt-5.3-chat": {
9302
9691
  id: string;
9303
9692
  name: string;
9304
9693
  api: "openai-completions";
9305
9694
  provider: string;
9306
9695
  baseUrl: string;
9307
- reasoning: true;
9696
+ reasoning: false;
9308
9697
  input: ("image" | "text")[];
9309
9698
  cost: {
9310
9699
  input: number;
@@ -9315,7 +9704,7 @@ export declare const MODELS: {
9315
9704
  contextWindow: number;
9316
9705
  maxTokens: number;
9317
9706
  };
9318
- readonly "openai/gpt-5.2-pro": {
9707
+ readonly "openai/gpt-5.3-codex": {
9319
9708
  id: string;
9320
9709
  name: string;
9321
9710
  api: "openai-completions";
@@ -9332,13 +9721,13 @@ export declare const MODELS: {
9332
9721
  contextWindow: number;
9333
9722
  maxTokens: number;
9334
9723
  };
9335
- readonly "openai/gpt-5.3-chat": {
9724
+ readonly "openai/gpt-5.4": {
9336
9725
  id: string;
9337
9726
  name: string;
9338
9727
  api: "openai-completions";
9339
9728
  provider: string;
9340
9729
  baseUrl: string;
9341
- reasoning: false;
9730
+ reasoning: true;
9342
9731
  input: ("image" | "text")[];
9343
9732
  cost: {
9344
9733
  input: number;
@@ -9349,7 +9738,7 @@ export declare const MODELS: {
9349
9738
  contextWindow: number;
9350
9739
  maxTokens: number;
9351
9740
  };
9352
- readonly "openai/gpt-5.3-codex": {
9741
+ readonly "openai/gpt-5.4-mini": {
9353
9742
  id: string;
9354
9743
  name: string;
9355
9744
  api: "openai-completions";
@@ -9366,7 +9755,7 @@ export declare const MODELS: {
9366
9755
  contextWindow: number;
9367
9756
  maxTokens: number;
9368
9757
  };
9369
- readonly "openai/gpt-5.4": {
9758
+ readonly "openai/gpt-5.4-nano": {
9370
9759
  id: string;
9371
9760
  name: string;
9372
9761
  api: "openai-completions";
@@ -9383,7 +9772,7 @@ export declare const MODELS: {
9383
9772
  contextWindow: number;
9384
9773
  maxTokens: number;
9385
9774
  };
9386
- readonly "openai/gpt-5.4-mini": {
9775
+ readonly "openai/gpt-5.4-pro": {
9387
9776
  id: string;
9388
9777
  name: string;
9389
9778
  api: "openai-completions";
@@ -9400,14 +9789,14 @@ export declare const MODELS: {
9400
9789
  contextWindow: number;
9401
9790
  maxTokens: number;
9402
9791
  };
9403
- readonly "openai/gpt-5.4-nano": {
9792
+ readonly "openai/gpt-audio": {
9404
9793
  id: string;
9405
9794
  name: string;
9406
9795
  api: "openai-completions";
9407
9796
  provider: string;
9408
9797
  baseUrl: string;
9409
- reasoning: true;
9410
- input: ("image" | "text")[];
9798
+ reasoning: false;
9799
+ input: "text"[];
9411
9800
  cost: {
9412
9801
  input: number;
9413
9802
  output: number;
@@ -9417,14 +9806,14 @@ export declare const MODELS: {
9417
9806
  contextWindow: number;
9418
9807
  maxTokens: number;
9419
9808
  };
9420
- readonly "openai/gpt-5.4-pro": {
9809
+ readonly "openai/gpt-audio-mini": {
9421
9810
  id: string;
9422
9811
  name: string;
9423
9812
  api: "openai-completions";
9424
9813
  provider: string;
9425
9814
  baseUrl: string;
9426
- reasoning: true;
9427
- input: ("image" | "text")[];
9815
+ reasoning: false;
9816
+ input: "text"[];
9428
9817
  cost: {
9429
9818
  input: number;
9430
9819
  output: number;
@@ -9899,7 +10288,7 @@ export declare const MODELS: {
9899
10288
  api: "openai-completions";
9900
10289
  provider: string;
9901
10290
  baseUrl: string;
9902
- reasoning: true;
10291
+ reasoning: false;
9903
10292
  input: "text"[];
9904
10293
  cost: {
9905
10294
  input: number;
@@ -9995,23 +10384,6 @@ export declare const MODELS: {
9995
10384
  contextWindow: number;
9996
10385
  maxTokens: number;
9997
10386
  };
9998
- readonly "qwen/qwen3-4b:free": {
9999
- id: string;
10000
- name: string;
10001
- api: "openai-completions";
10002
- provider: string;
10003
- baseUrl: string;
10004
- reasoning: true;
10005
- input: "text"[];
10006
- cost: {
10007
- input: number;
10008
- output: number;
10009
- cacheRead: number;
10010
- cacheWrite: number;
10011
- };
10012
- contextWindow: number;
10013
- maxTokens: number;
10014
- };
10015
10387
  readonly "qwen/qwen3-8b": {
10016
10388
  id: string;
10017
10389
  name: string;
@@ -10454,6 +10826,23 @@ export declare const MODELS: {
10454
10826
  contextWindow: number;
10455
10827
  maxTokens: number;
10456
10828
  };
10829
+ readonly "qwen/qwen3.6-plus": {
10830
+ id: string;
10831
+ name: string;
10832
+ api: "openai-completions";
10833
+ provider: string;
10834
+ baseUrl: string;
10835
+ reasoning: true;
10836
+ input: ("image" | "text")[];
10837
+ cost: {
10838
+ input: number;
10839
+ output: number;
10840
+ cacheRead: number;
10841
+ cacheWrite: number;
10842
+ };
10843
+ contextWindow: number;
10844
+ maxTokens: number;
10845
+ };
10457
10846
  readonly "qwen/qwq-32b": {
10458
10847
  id: string;
10459
10848
  name: string;
@@ -10471,6 +10860,23 @@ export declare const MODELS: {
10471
10860
  contextWindow: number;
10472
10861
  maxTokens: number;
10473
10862
  };
10863
+ readonly "rekaai/reka-edge": {
10864
+ id: string;
10865
+ name: string;
10866
+ api: "openai-completions";
10867
+ provider: string;
10868
+ baseUrl: string;
10869
+ reasoning: false;
10870
+ input: ("image" | "text")[];
10871
+ cost: {
10872
+ input: number;
10873
+ output: number;
10874
+ cacheRead: number;
10875
+ cacheWrite: number;
10876
+ };
10877
+ contextWindow: number;
10878
+ maxTokens: number;
10879
+ };
10474
10880
  readonly "relace/relace-search": {
10475
10881
  id: string;
10476
10882
  name: string;
@@ -10539,7 +10945,7 @@ export declare const MODELS: {
10539
10945
  contextWindow: number;
10540
10946
  maxTokens: number;
10541
10947
  };
10542
- readonly "stepfun/step-3.5-flash:free": {
10948
+ readonly "tencent/hy3-preview:free": {
10543
10949
  id: string;
10544
10950
  name: string;
10545
10951
  api: "openai-completions";
@@ -10743,7 +11149,7 @@ export declare const MODELS: {
10743
11149
  contextWindow: number;
10744
11150
  maxTokens: number;
10745
11151
  };
10746
- readonly "x-ai/grok-4.20-beta": {
11152
+ readonly "x-ai/grok-4.20": {
10747
11153
  id: string;
10748
11154
  name: string;
10749
11155
  api: "openai-completions";
@@ -10828,6 +11234,40 @@ export declare const MODELS: {
10828
11234
  contextWindow: number;
10829
11235
  maxTokens: number;
10830
11236
  };
11237
+ readonly "xiaomi/mimo-v2.5": {
11238
+ id: string;
11239
+ name: string;
11240
+ api: "openai-completions";
11241
+ provider: string;
11242
+ baseUrl: string;
11243
+ reasoning: true;
11244
+ input: ("image" | "text")[];
11245
+ cost: {
11246
+ input: number;
11247
+ output: number;
11248
+ cacheRead: number;
11249
+ cacheWrite: number;
11250
+ };
11251
+ contextWindow: number;
11252
+ maxTokens: number;
11253
+ };
11254
+ readonly "xiaomi/mimo-v2.5-pro": {
11255
+ id: string;
11256
+ name: string;
11257
+ api: "openai-completions";
11258
+ provider: string;
11259
+ baseUrl: string;
11260
+ reasoning: true;
11261
+ input: "text"[];
11262
+ cost: {
11263
+ input: number;
11264
+ output: number;
11265
+ cacheRead: number;
11266
+ cacheWrite: number;
11267
+ };
11268
+ contextWindow: number;
11269
+ maxTokens: number;
11270
+ };
10831
11271
  readonly "z-ai/glm-4-32b": {
10832
11272
  id: string;
10833
11273
  name: string;
@@ -10845,14 +11285,65 @@ export declare const MODELS: {
10845
11285
  contextWindow: number;
10846
11286
  maxTokens: number;
10847
11287
  };
10848
- readonly "z-ai/glm-4.5": {
11288
+ readonly "z-ai/glm-4.5": {
11289
+ id: string;
11290
+ name: string;
11291
+ api: "openai-completions";
11292
+ provider: string;
11293
+ baseUrl: string;
11294
+ reasoning: true;
11295
+ input: "text"[];
11296
+ cost: {
11297
+ input: number;
11298
+ output: number;
11299
+ cacheRead: number;
11300
+ cacheWrite: number;
11301
+ };
11302
+ contextWindow: number;
11303
+ maxTokens: number;
11304
+ };
11305
+ readonly "z-ai/glm-4.5-air": {
11306
+ id: string;
11307
+ name: string;
11308
+ api: "openai-completions";
11309
+ provider: string;
11310
+ baseUrl: string;
11311
+ reasoning: true;
11312
+ input: "text"[];
11313
+ cost: {
11314
+ input: number;
11315
+ output: number;
11316
+ cacheRead: number;
11317
+ cacheWrite: number;
11318
+ };
11319
+ contextWindow: number;
11320
+ maxTokens: number;
11321
+ };
11322
+ readonly "z-ai/glm-4.5-air:free": {
11323
+ id: string;
11324
+ name: string;
11325
+ api: "openai-completions";
11326
+ provider: string;
11327
+ baseUrl: string;
11328
+ reasoning: true;
11329
+ input: "text"[];
11330
+ cost: {
11331
+ input: number;
11332
+ output: number;
11333
+ cacheRead: number;
11334
+ cacheWrite: number;
11335
+ };
11336
+ contextWindow: number;
11337
+ maxTokens: number;
11338
+ };
11339
+ readonly "z-ai/glm-4.5v": {
10849
11340
  id: string;
10850
11341
  name: string;
10851
11342
  api: "openai-completions";
10852
11343
  provider: string;
10853
11344
  baseUrl: string;
10854
11345
  reasoning: true;
10855
- input: "text"[];
11346
+ input: ("image" | "text")[];
10856
11347
  cost: {
10857
11348
  input: number;
10858
11349
  output: number;
@@ -10862,7 +11353,7 @@ export declare const MODELS: {
10862
11353
  contextWindow: number;
10863
11354
  maxTokens: number;
10864
11355
  };
10865
- readonly "z-ai/glm-4.5-air": {
11356
+ readonly "z-ai/glm-4.6": {
10866
11357
  id: string;
10867
11358
  name: string;
10868
11359
  api: "openai-completions";
@@ -10879,14 +11370,14 @@ export declare const MODELS: {
10879
11370
  contextWindow: number;
10880
11371
  maxTokens: number;
10881
11372
  };
10882
- readonly "z-ai/glm-4.5-air:free": {
11373
+ readonly "z-ai/glm-4.6v": {
10883
11374
  id: string;
10884
11375
  name: string;
10885
11376
  api: "openai-completions";
10886
11377
  provider: string;
10887
11378
  baseUrl: string;
10888
11379
  reasoning: true;
10889
- input: "text"[];
11380
+ input: ("image" | "text")[];
10890
11381
  cost: {
10891
11382
  input: number;
10892
11383
  output: number;
@@ -10896,14 +11387,14 @@ export declare const MODELS: {
10896
11387
  contextWindow: number;
10897
11388
  maxTokens: number;
10898
11389
  };
10899
- readonly "z-ai/glm-4.5v": {
11390
+ readonly "z-ai/glm-4.7": {
10900
11391
  id: string;
10901
11392
  name: string;
10902
11393
  api: "openai-completions";
10903
11394
  provider: string;
10904
11395
  baseUrl: string;
10905
11396
  reasoning: true;
10906
- input: ("image" | "text")[];
11397
+ input: "text"[];
10907
11398
  cost: {
10908
11399
  input: number;
10909
11400
  output: number;
@@ -10913,7 +11404,7 @@ export declare const MODELS: {
10913
11404
  contextWindow: number;
10914
11405
  maxTokens: number;
10915
11406
  };
10916
- readonly "z-ai/glm-4.6": {
11407
+ readonly "z-ai/glm-4.7-flash": {
10917
11408
  id: string;
10918
11409
  name: string;
10919
11410
  api: "openai-completions";
@@ -10930,14 +11421,14 @@ export declare const MODELS: {
10930
11421
  contextWindow: number;
10931
11422
  maxTokens: number;
10932
11423
  };
10933
- readonly "z-ai/glm-4.6v": {
11424
+ readonly "z-ai/glm-5": {
10934
11425
  id: string;
10935
11426
  name: string;
10936
11427
  api: "openai-completions";
10937
11428
  provider: string;
10938
11429
  baseUrl: string;
10939
11430
  reasoning: true;
10940
- input: ("image" | "text")[];
11431
+ input: "text"[];
10941
11432
  cost: {
10942
11433
  input: number;
10943
11434
  output: number;
@@ -10947,7 +11438,7 @@ export declare const MODELS: {
10947
11438
  contextWindow: number;
10948
11439
  maxTokens: number;
10949
11440
  };
10950
- readonly "z-ai/glm-4.7": {
11441
+ readonly "z-ai/glm-5-turbo": {
10951
11442
  id: string;
10952
11443
  name: string;
10953
11444
  api: "openai-completions";
@@ -10964,7 +11455,7 @@ export declare const MODELS: {
10964
11455
  contextWindow: number;
10965
11456
  maxTokens: number;
10966
11457
  };
10967
- readonly "z-ai/glm-4.7-flash": {
11458
+ readonly "z-ai/glm-5.1": {
10968
11459
  id: string;
10969
11460
  name: string;
10970
11461
  api: "openai-completions";
@@ -10981,14 +11472,14 @@ export declare const MODELS: {
10981
11472
  contextWindow: number;
10982
11473
  maxTokens: number;
10983
11474
  };
10984
- readonly "z-ai/glm-5": {
11475
+ readonly "z-ai/glm-5v-turbo": {
10985
11476
  id: string;
10986
11477
  name: string;
10987
11478
  api: "openai-completions";
10988
11479
  provider: string;
10989
11480
  baseUrl: string;
10990
11481
  reasoning: true;
10991
- input: "text"[];
11482
+ input: ("image" | "text")[];
10992
11483
  cost: {
10993
11484
  input: number;
10994
11485
  output: number;
@@ -10998,14 +11489,14 @@ export declare const MODELS: {
10998
11489
  contextWindow: number;
10999
11490
  maxTokens: number;
11000
11491
  };
11001
- readonly "z-ai/glm-5-turbo": {
11492
+ readonly "~anthropic/claude-opus-latest": {
11002
11493
  id: string;
11003
11494
  name: string;
11004
11495
  api: "openai-completions";
11005
11496
  provider: string;
11006
11497
  baseUrl: string;
11007
11498
  reasoning: true;
11008
- input: "text"[];
11499
+ input: ("image" | "text")[];
11009
11500
  cost: {
11010
11501
  input: number;
11011
11502
  output: number;
@@ -11085,6 +11576,23 @@ export declare const MODELS: {
11085
11576
  contextWindow: number;
11086
11577
  maxTokens: number;
11087
11578
  };
11579
+ readonly "alibaba/qwen-3.6-max-preview": {
11580
+ id: string;
11581
+ name: string;
11582
+ api: "anthropic-messages";
11583
+ provider: string;
11584
+ baseUrl: string;
11585
+ reasoning: true;
11586
+ input: ("image" | "text")[];
11587
+ cost: {
11588
+ input: number;
11589
+ output: number;
11590
+ cacheRead: number;
11591
+ cacheWrite: number;
11592
+ };
11593
+ contextWindow: number;
11594
+ maxTokens: number;
11595
+ };
11088
11596
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11089
11597
  id: string;
11090
11598
  name: string;
@@ -11272,13 +11780,13 @@ export declare const MODELS: {
11272
11780
  contextWindow: number;
11273
11781
  maxTokens: number;
11274
11782
  };
11275
- readonly "anthropic/claude-3-haiku": {
11783
+ readonly "alibaba/qwen3.6-plus": {
11276
11784
  id: string;
11277
11785
  name: string;
11278
11786
  api: "anthropic-messages";
11279
11787
  provider: string;
11280
11788
  baseUrl: string;
11281
- reasoning: false;
11789
+ reasoning: true;
11282
11790
  input: ("image" | "text")[];
11283
11791
  cost: {
11284
11792
  input: number;
@@ -11289,7 +11797,7 @@ export declare const MODELS: {
11289
11797
  contextWindow: number;
11290
11798
  maxTokens: number;
11291
11799
  };
11292
- readonly "anthropic/claude-3.5-haiku": {
11800
+ readonly "anthropic/claude-3-haiku": {
11293
11801
  id: string;
11294
11802
  name: string;
11295
11803
  api: "anthropic-messages";
@@ -11306,7 +11814,7 @@ export declare const MODELS: {
11306
11814
  contextWindow: number;
11307
11815
  maxTokens: number;
11308
11816
  };
11309
- readonly "anthropic/claude-3.5-sonnet": {
11817
+ readonly "anthropic/claude-3.5-haiku": {
11310
11818
  id: string;
11311
11819
  name: string;
11312
11820
  api: "anthropic-messages";
@@ -11323,13 +11831,13 @@ export declare const MODELS: {
11323
11831
  contextWindow: number;
11324
11832
  maxTokens: number;
11325
11833
  };
11326
- readonly "anthropic/claude-3.5-sonnet-20240620": {
11834
+ readonly "anthropic/claude-3.7-sonnet": {
11327
11835
  id: string;
11328
11836
  name: string;
11329
11837
  api: "anthropic-messages";
11330
11838
  provider: string;
11331
11839
  baseUrl: string;
11332
- reasoning: false;
11840
+ reasoning: true;
11333
11841
  input: ("image" | "text")[];
11334
11842
  cost: {
11335
11843
  input: number;
@@ -11340,7 +11848,7 @@ export declare const MODELS: {
11340
11848
  contextWindow: number;
11341
11849
  maxTokens: number;
11342
11850
  };
11343
- readonly "anthropic/claude-3.7-sonnet": {
11851
+ readonly "anthropic/claude-haiku-4.5": {
11344
11852
  id: string;
11345
11853
  name: string;
11346
11854
  api: "anthropic-messages";
@@ -11357,7 +11865,7 @@ export declare const MODELS: {
11357
11865
  contextWindow: number;
11358
11866
  maxTokens: number;
11359
11867
  };
11360
- readonly "anthropic/claude-haiku-4.5": {
11868
+ readonly "anthropic/claude-opus-4": {
11361
11869
  id: string;
11362
11870
  name: string;
11363
11871
  api: "anthropic-messages";
@@ -11374,7 +11882,7 @@ export declare const MODELS: {
11374
11882
  contextWindow: number;
11375
11883
  maxTokens: number;
11376
11884
  };
11377
- readonly "anthropic/claude-opus-4": {
11885
+ readonly "anthropic/claude-opus-4.1": {
11378
11886
  id: string;
11379
11887
  name: string;
11380
11888
  api: "anthropic-messages";
@@ -11391,7 +11899,7 @@ export declare const MODELS: {
11391
11899
  contextWindow: number;
11392
11900
  maxTokens: number;
11393
11901
  };
11394
- readonly "anthropic/claude-opus-4.1": {
11902
+ readonly "anthropic/claude-opus-4.5": {
11395
11903
  id: string;
11396
11904
  name: string;
11397
11905
  api: "anthropic-messages";
@@ -11408,7 +11916,7 @@ export declare const MODELS: {
11408
11916
  contextWindow: number;
11409
11917
  maxTokens: number;
11410
11918
  };
11411
- readonly "anthropic/claude-opus-4.5": {
11919
+ readonly "anthropic/claude-opus-4.6": {
11412
11920
  id: string;
11413
11921
  name: string;
11414
11922
  api: "anthropic-messages";
@@ -11425,7 +11933,7 @@ export declare const MODELS: {
11425
11933
  contextWindow: number;
11426
11934
  maxTokens: number;
11427
11935
  };
11428
- readonly "anthropic/claude-opus-4.6": {
11936
+ readonly "anthropic/claude-opus-4.7": {
11429
11937
  id: string;
11430
11938
  name: string;
11431
11939
  api: "anthropic-messages";
@@ -11510,6 +12018,23 @@ export declare const MODELS: {
11510
12018
  contextWindow: number;
11511
12019
  maxTokens: number;
11512
12020
  };
12021
+ readonly "arcee-ai/trinity-large-thinking": {
12022
+ id: string;
12023
+ name: string;
12024
+ api: "anthropic-messages";
12025
+ provider: string;
12026
+ baseUrl: string;
12027
+ reasoning: true;
12028
+ input: "text"[];
12029
+ cost: {
12030
+ input: number;
12031
+ output: number;
12032
+ cacheRead: number;
12033
+ cacheWrite: number;
12034
+ };
12035
+ contextWindow: number;
12036
+ maxTokens: number;
12037
+ };
11513
12038
  readonly "bytedance/seed-1.6": {
11514
12039
  id: string;
11515
12040
  name: string;
@@ -11799,6 +12324,40 @@ export declare const MODELS: {
11799
12324
  contextWindow: number;
11800
12325
  maxTokens: number;
11801
12326
  };
12327
+ readonly "google/gemma-4-26b-a4b-it": {
12328
+ id: string;
12329
+ name: string;
12330
+ api: "anthropic-messages";
12331
+ provider: string;
12332
+ baseUrl: string;
12333
+ reasoning: false;
12334
+ input: ("image" | "text")[];
12335
+ cost: {
12336
+ input: number;
12337
+ output: number;
12338
+ cacheRead: number;
12339
+ cacheWrite: number;
12340
+ };
12341
+ contextWindow: number;
12342
+ maxTokens: number;
12343
+ };
12344
+ readonly "google/gemma-4-31b-it": {
12345
+ id: string;
12346
+ name: string;
12347
+ api: "anthropic-messages";
12348
+ provider: string;
12349
+ baseUrl: string;
12350
+ reasoning: false;
12351
+ input: ("image" | "text")[];
12352
+ cost: {
12353
+ input: number;
12354
+ output: number;
12355
+ cacheRead: number;
12356
+ cacheWrite: number;
12357
+ };
12358
+ contextWindow: number;
12359
+ maxTokens: number;
12360
+ };
11802
12361
  readonly "inception/mercury-2": {
11803
12362
  id: string;
11804
12363
  name: string;
@@ -11833,13 +12392,13 @@ export declare const MODELS: {
11833
12392
  contextWindow: number;
11834
12393
  maxTokens: number;
11835
12394
  };
11836
- readonly "meituan/longcat-flash-chat": {
12395
+ readonly "kwaipilot/kat-coder-pro-v2": {
11837
12396
  id: string;
11838
12397
  name: string;
11839
12398
  api: "anthropic-messages";
11840
12399
  provider: string;
11841
12400
  baseUrl: string;
11842
- reasoning: false;
12401
+ reasoning: true;
11843
12402
  input: "text"[];
11844
12403
  cost: {
11845
12404
  input: number;
@@ -11850,13 +12409,13 @@ export declare const MODELS: {
11850
12409
  contextWindow: number;
11851
12410
  maxTokens: number;
11852
12411
  };
11853
- readonly "meituan/longcat-flash-thinking": {
12412
+ readonly "meituan/longcat-flash-chat": {
11854
12413
  id: string;
11855
12414
  name: string;
11856
12415
  api: "anthropic-messages";
11857
12416
  provider: string;
11858
12417
  baseUrl: string;
11859
- reasoning: true;
12418
+ reasoning: false;
11860
12419
  input: "text"[];
11861
12420
  cost: {
11862
12421
  input: number;
@@ -12377,6 +12936,23 @@ export declare const MODELS: {
12377
12936
  contextWindow: number;
12378
12937
  maxTokens: number;
12379
12938
  };
12939
+ readonly "moonshotai/kimi-k2.6": {
12940
+ id: string;
12941
+ name: string;
12942
+ api: "anthropic-messages";
12943
+ provider: string;
12944
+ baseUrl: string;
12945
+ reasoning: true;
12946
+ input: ("image" | "text")[];
12947
+ cost: {
12948
+ input: number;
12949
+ output: number;
12950
+ cacheRead: number;
12951
+ cacheWrite: number;
12952
+ };
12953
+ contextWindow: number;
12954
+ maxTokens: number;
12955
+ };
12380
12956
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12381
12957
  id: string;
12382
12958
  name: string;
@@ -13057,23 +13633,6 @@ export declare const MODELS: {
13057
13633
  contextWindow: number;
13058
13634
  maxTokens: number;
13059
13635
  };
13060
- readonly "xai/grok-2-vision": {
13061
- id: string;
13062
- name: string;
13063
- api: "anthropic-messages";
13064
- provider: string;
13065
- baseUrl: string;
13066
- reasoning: false;
13067
- input: ("image" | "text")[];
13068
- cost: {
13069
- input: number;
13070
- output: number;
13071
- cacheRead: number;
13072
- cacheWrite: number;
13073
- };
13074
- contextWindow: number;
13075
- maxTokens: number;
13076
- };
13077
13636
  readonly "xai/grok-3": {
13078
13637
  id: string;
13079
13638
  name: string;
@@ -13166,7 +13725,7 @@ export declare const MODELS: {
13166
13725
  provider: string;
13167
13726
  baseUrl: string;
13168
13727
  reasoning: false;
13169
- input: "text"[];
13728
+ input: ("image" | "text")[];
13170
13729
  cost: {
13171
13730
  input: number;
13172
13731
  output: number;
@@ -13183,7 +13742,7 @@ export declare const MODELS: {
13183
13742
  provider: string;
13184
13743
  baseUrl: string;
13185
13744
  reasoning: true;
13186
- input: "text"[];
13745
+ input: ("image" | "text")[];
13187
13746
  cost: {
13188
13747
  input: number;
13189
13748
  output: number;
@@ -13200,7 +13759,7 @@ export declare const MODELS: {
13200
13759
  provider: string;
13201
13760
  baseUrl: string;
13202
13761
  reasoning: false;
13203
- input: "text"[];
13762
+ input: ("image" | "text")[];
13204
13763
  cost: {
13205
13764
  input: number;
13206
13765
  output: number;
@@ -13217,7 +13776,7 @@ export declare const MODELS: {
13217
13776
  provider: string;
13218
13777
  baseUrl: string;
13219
13778
  reasoning: true;
13220
- input: "text"[];
13779
+ input: ("image" | "text")[];
13221
13780
  cost: {
13222
13781
  input: number;
13223
13782
  output: number;
@@ -13234,7 +13793,7 @@ export declare const MODELS: {
13234
13793
  provider: string;
13235
13794
  baseUrl: string;
13236
13795
  reasoning: true;
13237
- input: "text"[];
13796
+ input: ("image" | "text")[];
13238
13797
  cost: {
13239
13798
  input: number;
13240
13799
  output: number;
@@ -13251,7 +13810,7 @@ export declare const MODELS: {
13251
13810
  provider: string;
13252
13811
  baseUrl: string;
13253
13812
  reasoning: true;
13254
- input: "text"[];
13813
+ input: ("image" | "text")[];
13255
13814
  cost: {
13256
13815
  input: number;
13257
13816
  output: number;
@@ -13567,6 +14126,40 @@ export declare const MODELS: {
13567
14126
  contextWindow: number;
13568
14127
  maxTokens: number;
13569
14128
  };
14129
+ readonly "zai/glm-5.1": {
14130
+ id: string;
14131
+ name: string;
14132
+ api: "anthropic-messages";
14133
+ provider: string;
14134
+ baseUrl: string;
14135
+ reasoning: true;
14136
+ input: "text"[];
14137
+ cost: {
14138
+ input: number;
14139
+ output: number;
14140
+ cacheRead: number;
14141
+ cacheWrite: number;
14142
+ };
14143
+ contextWindow: number;
14144
+ maxTokens: number;
14145
+ };
14146
+ readonly "zai/glm-5v-turbo": {
14147
+ id: string;
14148
+ name: string;
14149
+ api: "anthropic-messages";
14150
+ provider: string;
14151
+ baseUrl: string;
14152
+ reasoning: true;
14153
+ input: ("image" | "text")[];
14154
+ cost: {
14155
+ input: number;
14156
+ output: number;
14157
+ cacheRead: number;
14158
+ cacheWrite: number;
14159
+ };
14160
+ contextWindow: number;
14161
+ maxTokens: number;
14162
+ };
13570
14163
  };
13571
14164
  readonly xai: {
13572
14165
  readonly "grok-2": {
@@ -14072,6 +14665,7 @@ export declare const MODELS: {
14072
14665
  compat: {
14073
14666
  supportsDeveloperRole: false;
14074
14667
  thinkingFormat: "zai";
14668
+ zaiToolStream: true;
14075
14669
  };
14076
14670
  reasoning: true;
14077
14671
  input: "text"[];
@@ -14093,6 +14687,7 @@ export declare const MODELS: {
14093
14687
  compat: {
14094
14688
  supportsDeveloperRole: false;
14095
14689
  thinkingFormat: "zai";
14690
+ zaiToolStream: true;
14096
14691
  };
14097
14692
  reasoning: true;
14098
14693
  input: ("image" | "text")[];
@@ -14114,6 +14709,7 @@ export declare const MODELS: {
14114
14709
  compat: {
14115
14710
  supportsDeveloperRole: false;
14116
14711
  thinkingFormat: "zai";
14712
+ zaiToolStream: true;
14117
14713
  };
14118
14714
  reasoning: true;
14119
14715
  input: "text"[];
@@ -14135,6 +14731,29 @@ export declare const MODELS: {
14135
14731
  compat: {
14136
14732
  supportsDeveloperRole: false;
14137
14733
  thinkingFormat: "zai";
14734
+ zaiToolStream: true;
14735
+ };
14736
+ reasoning: true;
14737
+ input: "text"[];
14738
+ cost: {
14739
+ input: number;
14740
+ output: number;
14741
+ cacheRead: number;
14742
+ cacheWrite: number;
14743
+ };
14744
+ contextWindow: number;
14745
+ maxTokens: number;
14746
+ };
14747
+ readonly "glm-4.7-flashx": {
14748
+ id: string;
14749
+ name: string;
14750
+ api: "openai-completions";
14751
+ provider: string;
14752
+ baseUrl: string;
14753
+ compat: {
14754
+ supportsDeveloperRole: false;
14755
+ thinkingFormat: "zai";
14756
+ zaiToolStream: true;
14138
14757
  };
14139
14758
  reasoning: true;
14140
14759
  input: "text"[];
@@ -14156,6 +14775,7 @@ export declare const MODELS: {
14156
14775
  compat: {
14157
14776
  supportsDeveloperRole: false;
14158
14777
  thinkingFormat: "zai";
14778
+ zaiToolStream: true;
14159
14779
  };
14160
14780
  reasoning: true;
14161
14781
  input: "text"[];
@@ -14177,6 +14797,29 @@ export declare const MODELS: {
14177
14797
  compat: {
14178
14798
  supportsDeveloperRole: false;
14179
14799
  thinkingFormat: "zai";
14800
+ zaiToolStream: true;
14801
+ };
14802
+ reasoning: true;
14803
+ input: "text"[];
14804
+ cost: {
14805
+ input: number;
14806
+ output: number;
14807
+ cacheRead: number;
14808
+ cacheWrite: number;
14809
+ };
14810
+ contextWindow: number;
14811
+ maxTokens: number;
14812
+ };
14813
+ readonly "glm-5.1": {
14814
+ id: string;
14815
+ name: string;
14816
+ api: "openai-completions";
14817
+ provider: string;
14818
+ baseUrl: string;
14819
+ compat: {
14820
+ supportsDeveloperRole: false;
14821
+ thinkingFormat: "zai";
14822
+ zaiToolStream: true;
14180
14823
  };
14181
14824
  reasoning: true;
14182
14825
  input: "text"[];
@@ -14189,6 +14832,28 @@ export declare const MODELS: {
14189
14832
  contextWindow: number;
14190
14833
  maxTokens: number;
14191
14834
  };
14835
+ readonly "glm-5v-turbo": {
14836
+ id: string;
14837
+ name: string;
14838
+ api: "openai-completions";
14839
+ provider: string;
14840
+ baseUrl: string;
14841
+ compat: {
14842
+ supportsDeveloperRole: false;
14843
+ thinkingFormat: "zai";
14844
+ zaiToolStream: true;
14845
+ };
14846
+ reasoning: true;
14847
+ input: ("image" | "text")[];
14848
+ cost: {
14849
+ input: number;
14850
+ output: number;
14851
+ cacheRead: number;
14852
+ cacheWrite: number;
14853
+ };
14854
+ contextWindow: number;
14855
+ maxTokens: number;
14856
+ };
14192
14857
  };
14193
14858
  };
14194
14859
  //# sourceMappingURL=models.generated.d.ts.map