@draht/ai 2026.4.5 → 2026.4.25

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (62) hide show
  1. package/README.md +17 -8
  2. package/dist/cli.d.ts.map +1 -1
  3. package/dist/cli.js +1 -1
  4. package/dist/cli.js.map +1 -1
  5. package/dist/models.d.ts +7 -1
  6. package/dist/models.d.ts.map +1 -1
  7. package/dist/models.generated.d.ts +996 -519
  8. package/dist/models.generated.d.ts.map +1 -1
  9. package/dist/models.generated.js +1255 -765
  10. package/dist/models.generated.js.map +1 -1
  11. package/dist/models.js +20 -2
  12. package/dist/models.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +46 -14
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +44 -19
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  20. package/dist/providers/azure-openai-responses.js +2 -2
  21. package/dist/providers/azure-openai-responses.js.map +1 -1
  22. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  23. package/dist/providers/google-gemini-cli.js +2 -2
  24. package/dist/providers/google-gemini-cli.js.map +1 -1
  25. package/dist/providers/google-vertex.d.ts.map +1 -1
  26. package/dist/providers/google-vertex.js +1 -1
  27. package/dist/providers/google-vertex.js.map +1 -1
  28. package/dist/providers/google.d.ts.map +1 -1
  29. package/dist/providers/google.js +29 -4
  30. package/dist/providers/google.js.map +1 -1
  31. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  32. package/dist/providers/openai-codex-responses.js +2 -2
  33. package/dist/providers/openai-codex-responses.js.map +1 -1
  34. package/dist/providers/openai-completions.d.ts.map +1 -1
  35. package/dist/providers/openai-completions.js +29 -16
  36. package/dist/providers/openai-completions.js.map +1 -1
  37. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  38. package/dist/providers/openai-responses-shared.js +8 -1
  39. package/dist/providers/openai-responses-shared.js.map +1 -1
  40. package/dist/providers/openai-responses.d.ts.map +1 -1
  41. package/dist/providers/openai-responses.js +4 -4
  42. package/dist/providers/openai-responses.js.map +1 -1
  43. package/dist/providers/simple-options.d.ts +6 -1
  44. package/dist/providers/simple-options.d.ts.map +1 -1
  45. package/dist/providers/simple-options.js +10 -1
  46. package/dist/providers/simple-options.js.map +1 -1
  47. package/dist/types.d.ts +63 -6
  48. package/dist/types.d.ts.map +1 -1
  49. package/dist/types.js.map +1 -1
  50. package/dist/utils/oauth/index.d.ts +1 -0
  51. package/dist/utils/oauth/index.d.ts.map +1 -1
  52. package/dist/utils/oauth/index.js +4 -0
  53. package/dist/utils/oauth/index.js.map +1 -1
  54. package/dist/utils/oauth/opencode-go.d.ts +23 -0
  55. package/dist/utils/oauth/opencode-go.d.ts.map +1 -0
  56. package/dist/utils/oauth/opencode-go.js +46 -0
  57. package/dist/utils/oauth/opencode-go.js.map +1 -0
  58. package/dist/utils/overflow.d.ts +3 -2
  59. package/dist/utils/overflow.d.ts.map +1 -1
  60. package/dist/utils/overflow.js +24 -11
  61. package/dist/utils/overflow.js.map +1 -1
  62. package/package.json +5 -8
@@ -255,6 +255,23 @@ export declare const MODELS: {
255
255
  contextWindow: number;
256
256
  maxTokens: number;
257
257
  };
258
+ readonly "anthropic.claude-opus-4-7": {
259
+ id: string;
260
+ name: string;
261
+ api: "bedrock-converse-stream";
262
+ provider: string;
263
+ baseUrl: string;
264
+ reasoning: true;
265
+ input: ("image" | "text")[];
266
+ cost: {
267
+ input: number;
268
+ output: number;
269
+ cacheRead: number;
270
+ cacheWrite: number;
271
+ };
272
+ contextWindow: number;
273
+ maxTokens: number;
274
+ };
258
275
  readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
259
276
  id: string;
260
277
  name: string;
@@ -306,6 +323,40 @@ export declare const MODELS: {
306
323
  contextWindow: number;
307
324
  maxTokens: number;
308
325
  };
326
+ readonly "au.anthropic.claude-opus-4-6-v1": {
327
+ id: string;
328
+ name: string;
329
+ api: "bedrock-converse-stream";
330
+ provider: string;
331
+ baseUrl: string;
332
+ reasoning: true;
333
+ input: ("image" | "text")[];
334
+ cost: {
335
+ input: number;
336
+ output: number;
337
+ cacheRead: number;
338
+ cacheWrite: number;
339
+ };
340
+ contextWindow: number;
341
+ maxTokens: number;
342
+ };
343
+ readonly "au.anthropic.claude-sonnet-4-6": {
344
+ id: string;
345
+ name: string;
346
+ api: "bedrock-converse-stream";
347
+ provider: string;
348
+ baseUrl: string;
349
+ reasoning: true;
350
+ input: ("image" | "text")[];
351
+ cost: {
352
+ input: number;
353
+ output: number;
354
+ cacheRead: number;
355
+ cacheWrite: number;
356
+ };
357
+ contextWindow: number;
358
+ maxTokens: number;
359
+ };
309
360
  readonly "deepseek.r1-v1:0": {
310
361
  id: string;
311
362
  name: string;
@@ -408,6 +459,23 @@ export declare const MODELS: {
408
459
  contextWindow: number;
409
460
  maxTokens: number;
410
461
  };
462
+ readonly "eu.anthropic.claude-opus-4-7": {
463
+ id: string;
464
+ name: string;
465
+ api: "bedrock-converse-stream";
466
+ provider: string;
467
+ baseUrl: string;
468
+ reasoning: true;
469
+ input: ("image" | "text")[];
470
+ cost: {
471
+ input: number;
472
+ output: number;
473
+ cacheRead: number;
474
+ cacheWrite: number;
475
+ };
476
+ contextWindow: number;
477
+ maxTokens: number;
478
+ };
411
479
  readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
412
480
  id: string;
413
481
  name: string;
@@ -510,6 +578,23 @@ export declare const MODELS: {
510
578
  contextWindow: number;
511
579
  maxTokens: number;
512
580
  };
581
+ readonly "global.anthropic.claude-opus-4-7": {
582
+ id: string;
583
+ name: string;
584
+ api: "bedrock-converse-stream";
585
+ provider: string;
586
+ baseUrl: string;
587
+ reasoning: true;
588
+ input: ("image" | "text")[];
589
+ cost: {
590
+ input: number;
591
+ output: number;
592
+ cacheRead: number;
593
+ cacheWrite: number;
594
+ };
595
+ contextWindow: number;
596
+ maxTokens: number;
597
+ };
513
598
  readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
514
599
  id: string;
515
600
  name: string;
@@ -1207,6 +1292,23 @@ export declare const MODELS: {
1207
1292
  contextWindow: number;
1208
1293
  maxTokens: number;
1209
1294
  };
1295
+ readonly "qwen.qwen3-coder-next": {
1296
+ id: string;
1297
+ name: string;
1298
+ api: "bedrock-converse-stream";
1299
+ provider: string;
1300
+ baseUrl: string;
1301
+ reasoning: true;
1302
+ input: "text"[];
1303
+ cost: {
1304
+ input: number;
1305
+ output: number;
1306
+ cacheRead: number;
1307
+ cacheWrite: number;
1308
+ };
1309
+ contextWindow: number;
1310
+ maxTokens: number;
1311
+ };
1210
1312
  readonly "qwen.qwen3-next-80b-a3b": {
1211
1313
  id: string;
1212
1314
  name: string;
@@ -1326,6 +1428,23 @@ export declare const MODELS: {
1326
1428
  contextWindow: number;
1327
1429
  maxTokens: number;
1328
1430
  };
1431
+ readonly "us.anthropic.claude-opus-4-7": {
1432
+ id: string;
1433
+ name: string;
1434
+ api: "bedrock-converse-stream";
1435
+ provider: string;
1436
+ baseUrl: string;
1437
+ reasoning: true;
1438
+ input: ("image" | "text")[];
1439
+ cost: {
1440
+ input: number;
1441
+ output: number;
1442
+ cacheRead: number;
1443
+ cacheWrite: number;
1444
+ };
1445
+ contextWindow: number;
1446
+ maxTokens: number;
1447
+ };
1329
1448
  readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
1330
1449
  id: string;
1331
1450
  name: string;
@@ -1753,7 +1872,7 @@ export declare const MODELS: {
1753
1872
  contextWindow: number;
1754
1873
  maxTokens: number;
1755
1874
  };
1756
- readonly "claude-sonnet-4-0": {
1875
+ readonly "claude-opus-4-7": {
1757
1876
  id: string;
1758
1877
  name: string;
1759
1878
  api: "anthropic-messages";
@@ -1770,7 +1889,7 @@ export declare const MODELS: {
1770
1889
  contextWindow: number;
1771
1890
  maxTokens: number;
1772
1891
  };
1773
- readonly "claude-sonnet-4-20250514": {
1892
+ readonly "claude-sonnet-4-0": {
1774
1893
  id: string;
1775
1894
  name: string;
1776
1895
  api: "anthropic-messages";
@@ -1787,7 +1906,7 @@ export declare const MODELS: {
1787
1906
  contextWindow: number;
1788
1907
  maxTokens: number;
1789
1908
  };
1790
- readonly "claude-sonnet-4-5": {
1909
+ readonly "claude-sonnet-4-20250514": {
1791
1910
  id: string;
1792
1911
  name: string;
1793
1912
  api: "anthropic-messages";
@@ -1804,7 +1923,7 @@ export declare const MODELS: {
1804
1923
  contextWindow: number;
1805
1924
  maxTokens: number;
1806
1925
  };
1807
- readonly "claude-sonnet-4-5-20250929": {
1926
+ readonly "claude-sonnet-4-5": {
1808
1927
  id: string;
1809
1928
  name: string;
1810
1929
  api: "anthropic-messages";
@@ -1821,7 +1940,7 @@ export declare const MODELS: {
1821
1940
  contextWindow: number;
1822
1941
  maxTokens: number;
1823
1942
  };
1824
- readonly "claude-sonnet-4-6": {
1943
+ readonly "claude-sonnet-4-5-20250929": {
1825
1944
  id: string;
1826
1945
  name: string;
1827
1946
  api: "anthropic-messages";
@@ -1838,16 +1957,14 @@ export declare const MODELS: {
1838
1957
  contextWindow: number;
1839
1958
  maxTokens: number;
1840
1959
  };
1841
- };
1842
- readonly "azure-openai-responses": {
1843
- readonly "codex-mini-latest": {
1960
+ readonly "claude-sonnet-4-6": {
1844
1961
  id: string;
1845
1962
  name: string;
1846
- api: "azure-openai-responses";
1963
+ api: "anthropic-messages";
1847
1964
  provider: string;
1848
1965
  baseUrl: string;
1849
1966
  reasoning: true;
1850
- input: "text"[];
1967
+ input: ("image" | "text")[];
1851
1968
  cost: {
1852
1969
  input: number;
1853
1970
  output: number;
@@ -1857,6 +1974,8 @@ export declare const MODELS: {
1857
1974
  contextWindow: number;
1858
1975
  maxTokens: number;
1859
1976
  };
1977
+ };
1978
+ readonly "azure-openai-responses": {
1860
1979
  readonly "gpt-4": {
1861
1980
  id: string;
1862
1981
  name: string;
@@ -2401,6 +2520,23 @@ export declare const MODELS: {
2401
2520
  contextWindow: number;
2402
2521
  maxTokens: number;
2403
2522
  };
2523
+ readonly "gpt-5.5": {
2524
+ id: string;
2525
+ name: string;
2526
+ api: "azure-openai-responses";
2527
+ provider: string;
2528
+ baseUrl: string;
2529
+ reasoning: true;
2530
+ input: ("image" | "text")[];
2531
+ cost: {
2532
+ input: number;
2533
+ output: number;
2534
+ cacheRead: number;
2535
+ cacheWrite: number;
2536
+ };
2537
+ contextWindow: number;
2538
+ maxTokens: number;
2539
+ };
2404
2540
  readonly o1: {
2405
2541
  id: string;
2406
2542
  name: string;
@@ -2678,6 +2814,29 @@ export declare const MODELS: {
2678
2814
  contextWindow: number;
2679
2815
  maxTokens: number;
2680
2816
  };
2817
+ readonly "claude-opus-4.7": {
2818
+ id: string;
2819
+ name: string;
2820
+ api: "anthropic-messages";
2821
+ provider: string;
2822
+ baseUrl: string;
2823
+ headers: {
2824
+ "User-Agent": string;
2825
+ "Editor-Version": string;
2826
+ "Editor-Plugin-Version": string;
2827
+ "Copilot-Integration-Id": string;
2828
+ };
2829
+ reasoning: true;
2830
+ input: ("image" | "text")[];
2831
+ cost: {
2832
+ input: number;
2833
+ output: number;
2834
+ cacheRead: number;
2835
+ cacheWrite: number;
2836
+ };
2837
+ contextWindow: number;
2838
+ maxTokens: number;
2839
+ };
2681
2840
  readonly "claude-sonnet-4": {
2682
2841
  id: string;
2683
2842
  name: string;
@@ -3168,6 +3327,29 @@ export declare const MODELS: {
3168
3327
  contextWindow: number;
3169
3328
  maxTokens: number;
3170
3329
  };
3330
+ readonly "gpt-5.5": {
3331
+ id: string;
3332
+ name: string;
3333
+ api: "openai-responses";
3334
+ provider: string;
3335
+ baseUrl: string;
3336
+ headers: {
3337
+ "User-Agent": string;
3338
+ "Editor-Version": string;
3339
+ "Editor-Plugin-Version": string;
3340
+ "Copilot-Integration-Id": string;
3341
+ };
3342
+ reasoning: true;
3343
+ input: ("image" | "text")[];
3344
+ cost: {
3345
+ input: number;
3346
+ output: number;
3347
+ cacheRead: number;
3348
+ cacheWrite: number;
3349
+ };
3350
+ contextWindow: number;
3351
+ maxTokens: number;
3352
+ };
3171
3353
  readonly "grok-code-fast-1": {
3172
3354
  id: string;
3173
3355
  name: string;
@@ -3623,7 +3805,7 @@ export declare const MODELS: {
3623
3805
  contextWindow: number;
3624
3806
  maxTokens: number;
3625
3807
  };
3626
- readonly "gemma-4-26b": {
3808
+ readonly "gemma-4-26b-a4b-it": {
3627
3809
  id: string;
3628
3810
  name: string;
3629
3811
  api: "google-generative-ai";
@@ -3640,7 +3822,7 @@ export declare const MODELS: {
3640
3822
  contextWindow: number;
3641
3823
  maxTokens: number;
3642
3824
  };
3643
- readonly "gemma-4-31b": {
3825
+ readonly "gemma-4-31b-it": {
3644
3826
  id: string;
3645
3827
  name: string;
3646
3828
  api: "google-generative-ai";
@@ -4122,6 +4304,23 @@ export declare const MODELS: {
4122
4304
  contextWindow: number;
4123
4305
  maxTokens: number;
4124
4306
  };
4307
+ readonly "gemini-3.1-pro-preview-customtools": {
4308
+ id: string;
4309
+ name: string;
4310
+ api: "google-vertex";
4311
+ provider: string;
4312
+ baseUrl: string;
4313
+ reasoning: true;
4314
+ input: ("image" | "text")[];
4315
+ cost: {
4316
+ input: number;
4317
+ output: number;
4318
+ cacheRead: number;
4319
+ cacheWrite: number;
4320
+ };
4321
+ contextWindow: number;
4322
+ maxTokens: number;
4323
+ };
4125
4324
  };
4126
4325
  readonly groq: {
4127
4326
  readonly "deepseek-r1-distill-llama-70b": {
@@ -4472,6 +4671,26 @@ export declare const MODELS: {
4472
4671
  contextWindow: number;
4473
4672
  maxTokens: number;
4474
4673
  };
4674
+ readonly "MiniMaxAI/MiniMax-M2.7": {
4675
+ id: string;
4676
+ name: string;
4677
+ api: "openai-completions";
4678
+ provider: string;
4679
+ baseUrl: string;
4680
+ compat: {
4681
+ supportsDeveloperRole: false;
4682
+ };
4683
+ reasoning: true;
4684
+ input: "text"[];
4685
+ cost: {
4686
+ input: number;
4687
+ output: number;
4688
+ cacheRead: number;
4689
+ cacheWrite: number;
4690
+ };
4691
+ contextWindow: number;
4692
+ maxTokens: number;
4693
+ };
4475
4694
  readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4476
4695
  id: string;
4477
4696
  name: string;
@@ -4732,7 +4951,7 @@ export declare const MODELS: {
4732
4951
  contextWindow: number;
4733
4952
  maxTokens: number;
4734
4953
  };
4735
- readonly "zai-org/GLM-4.7": {
4954
+ readonly "moonshotai/Kimi-K2.6": {
4736
4955
  id: string;
4737
4956
  name: string;
4738
4957
  api: "openai-completions";
@@ -4742,7 +4961,7 @@ export declare const MODELS: {
4742
4961
  supportsDeveloperRole: false;
4743
4962
  };
4744
4963
  reasoning: true;
4745
- input: "text"[];
4964
+ input: ("image" | "text")[];
4746
4965
  cost: {
4747
4966
  input: number;
4748
4967
  output: number;
@@ -4752,7 +4971,7 @@ export declare const MODELS: {
4752
4971
  contextWindow: number;
4753
4972
  maxTokens: number;
4754
4973
  };
4755
- readonly "zai-org/GLM-4.7-Flash": {
4974
+ readonly "zai-org/GLM-4.7": {
4756
4975
  id: string;
4757
4976
  name: string;
4758
4977
  api: "openai-completions";
@@ -4772,7 +4991,7 @@ export declare const MODELS: {
4772
4991
  contextWindow: number;
4773
4992
  maxTokens: number;
4774
4993
  };
4775
- readonly "zai-org/GLM-5": {
4994
+ readonly "zai-org/GLM-4.7-Flash": {
4776
4995
  id: string;
4777
4996
  name: string;
4778
4997
  api: "openai-completions";
@@ -4792,16 +5011,17 @@ export declare const MODELS: {
4792
5011
  contextWindow: number;
4793
5012
  maxTokens: number;
4794
5013
  };
4795
- };
4796
- readonly "kimi-coding": {
4797
- readonly k2p5: {
5014
+ readonly "zai-org/GLM-5": {
4798
5015
  id: string;
4799
5016
  name: string;
4800
- api: "anthropic-messages";
5017
+ api: "openai-completions";
4801
5018
  provider: string;
4802
5019
  baseUrl: string;
5020
+ compat: {
5021
+ supportsDeveloperRole: false;
5022
+ };
4803
5023
  reasoning: true;
4804
- input: ("image" | "text")[];
5024
+ input: "text"[];
4805
5025
  cost: {
4806
5026
  input: number;
4807
5027
  output: number;
@@ -4811,12 +5031,15 @@ export declare const MODELS: {
4811
5031
  contextWindow: number;
4812
5032
  maxTokens: number;
4813
5033
  };
4814
- readonly "kimi-k2-thinking": {
5034
+ readonly "zai-org/GLM-5.1": {
4815
5035
  id: string;
4816
5036
  name: string;
4817
- api: "anthropic-messages";
5037
+ api: "openai-completions";
4818
5038
  provider: string;
4819
5039
  baseUrl: string;
5040
+ compat: {
5041
+ supportsDeveloperRole: false;
5042
+ };
4820
5043
  reasoning: true;
4821
5044
  input: "text"[];
4822
5045
  cost: {
@@ -4829,15 +5052,15 @@ export declare const MODELS: {
4829
5052
  maxTokens: number;
4830
5053
  };
4831
5054
  };
4832
- readonly minimax: {
4833
- readonly "MiniMax-M2": {
5055
+ readonly "kimi-coding": {
5056
+ readonly k2p5: {
4834
5057
  id: string;
4835
5058
  name: string;
4836
5059
  api: "anthropic-messages";
4837
5060
  provider: string;
4838
5061
  baseUrl: string;
4839
5062
  reasoning: true;
4840
- input: "text"[];
5063
+ input: ("image" | "text")[];
4841
5064
  cost: {
4842
5065
  input: number;
4843
5066
  output: number;
@@ -4847,48 +5070,14 @@ export declare const MODELS: {
4847
5070
  contextWindow: number;
4848
5071
  maxTokens: number;
4849
5072
  };
4850
- readonly "MiniMax-M2.1": {
5073
+ readonly k2p6: {
4851
5074
  id: string;
4852
5075
  name: string;
4853
5076
  api: "anthropic-messages";
4854
5077
  provider: string;
4855
5078
  baseUrl: string;
4856
5079
  reasoning: true;
4857
- input: "text"[];
4858
- cost: {
4859
- input: number;
4860
- output: number;
4861
- cacheRead: number;
4862
- cacheWrite: number;
4863
- };
4864
- contextWindow: number;
4865
- maxTokens: number;
4866
- };
4867
- readonly "MiniMax-M2.1-highspeed": {
4868
- id: string;
4869
- name: string;
4870
- api: "anthropic-messages";
4871
- provider: string;
4872
- baseUrl: string;
4873
- reasoning: true;
4874
- input: "text"[];
4875
- cost: {
4876
- input: number;
4877
- output: number;
4878
- cacheRead: number;
4879
- cacheWrite: number;
4880
- };
4881
- contextWindow: number;
4882
- maxTokens: number;
4883
- };
4884
- readonly "MiniMax-M2.5": {
4885
- id: string;
4886
- name: string;
4887
- api: "anthropic-messages";
4888
- provider: string;
4889
- baseUrl: string;
4890
- reasoning: true;
4891
- input: "text"[];
5080
+ input: ("image" | "text")[];
4892
5081
  cost: {
4893
5082
  input: number;
4894
5083
  output: number;
@@ -4898,7 +5087,7 @@ export declare const MODELS: {
4898
5087
  contextWindow: number;
4899
5088
  maxTokens: number;
4900
5089
  };
4901
- readonly "MiniMax-M2.5-highspeed": {
5090
+ readonly "kimi-k2-thinking": {
4902
5091
  id: string;
4903
5092
  name: string;
4904
5093
  api: "anthropic-messages";
@@ -4915,6 +5104,8 @@ export declare const MODELS: {
4915
5104
  contextWindow: number;
4916
5105
  maxTokens: number;
4917
5106
  };
5107
+ };
5108
+ readonly minimax: {
4918
5109
  readonly "MiniMax-M2.7": {
4919
5110
  id: string;
4920
5111
  name: string;
@@ -4951,91 +5142,6 @@ export declare const MODELS: {
4951
5142
  };
4952
5143
  };
4953
5144
  readonly "minimax-cn": {
4954
- readonly "MiniMax-M2": {
4955
- id: string;
4956
- name: string;
4957
- api: "anthropic-messages";
4958
- provider: string;
4959
- baseUrl: string;
4960
- reasoning: true;
4961
- input: "text"[];
4962
- cost: {
4963
- input: number;
4964
- output: number;
4965
- cacheRead: number;
4966
- cacheWrite: number;
4967
- };
4968
- contextWindow: number;
4969
- maxTokens: number;
4970
- };
4971
- readonly "MiniMax-M2.1": {
4972
- id: string;
4973
- name: string;
4974
- api: "anthropic-messages";
4975
- provider: string;
4976
- baseUrl: string;
4977
- reasoning: true;
4978
- input: "text"[];
4979
- cost: {
4980
- input: number;
4981
- output: number;
4982
- cacheRead: number;
4983
- cacheWrite: number;
4984
- };
4985
- contextWindow: number;
4986
- maxTokens: number;
4987
- };
4988
- readonly "MiniMax-M2.1-highspeed": {
4989
- id: string;
4990
- name: string;
4991
- api: "anthropic-messages";
4992
- provider: string;
4993
- baseUrl: string;
4994
- reasoning: true;
4995
- input: "text"[];
4996
- cost: {
4997
- input: number;
4998
- output: number;
4999
- cacheRead: number;
5000
- cacheWrite: number;
5001
- };
5002
- contextWindow: number;
5003
- maxTokens: number;
5004
- };
5005
- readonly "MiniMax-M2.5": {
5006
- id: string;
5007
- name: string;
5008
- api: "anthropic-messages";
5009
- provider: string;
5010
- baseUrl: string;
5011
- reasoning: true;
5012
- input: "text"[];
5013
- cost: {
5014
- input: number;
5015
- output: number;
5016
- cacheRead: number;
5017
- cacheWrite: number;
5018
- };
5019
- contextWindow: number;
5020
- maxTokens: number;
5021
- };
5022
- readonly "MiniMax-M2.5-highspeed": {
5023
- id: string;
5024
- name: string;
5025
- api: "anthropic-messages";
5026
- provider: string;
5027
- baseUrl: string;
5028
- reasoning: true;
5029
- input: "text"[];
5030
- cost: {
5031
- input: number;
5032
- output: number;
5033
- cacheRead: number;
5034
- cacheWrite: number;
5035
- };
5036
- contextWindow: number;
5037
- maxTokens: number;
5038
- };
5039
5145
  readonly "MiniMax-M2.7": {
5040
5146
  id: string;
5041
5147
  name: string;
@@ -5075,7 +5181,7 @@ export declare const MODELS: {
5075
5181
  readonly "codestral-latest": {
5076
5182
  id: string;
5077
5183
  name: string;
5078
- api: "mistral-conversations";
5184
+ api: "openai-completions";
5079
5185
  provider: string;
5080
5186
  baseUrl: string;
5081
5187
  reasoning: false;
@@ -5092,7 +5198,7 @@ export declare const MODELS: {
5092
5198
  readonly "devstral-2512": {
5093
5199
  id: string;
5094
5200
  name: string;
5095
- api: "mistral-conversations";
5201
+ api: "openai-completions";
5096
5202
  provider: string;
5097
5203
  baseUrl: string;
5098
5204
  reasoning: false;
@@ -5109,7 +5215,7 @@ export declare const MODELS: {
5109
5215
  readonly "devstral-medium-2507": {
5110
5216
  id: string;
5111
5217
  name: string;
5112
- api: "mistral-conversations";
5218
+ api: "openai-completions";
5113
5219
  provider: string;
5114
5220
  baseUrl: string;
5115
5221
  reasoning: false;
@@ -5126,7 +5232,7 @@ export declare const MODELS: {
5126
5232
  readonly "devstral-medium-latest": {
5127
5233
  id: string;
5128
5234
  name: string;
5129
- api: "mistral-conversations";
5235
+ api: "openai-completions";
5130
5236
  provider: string;
5131
5237
  baseUrl: string;
5132
5238
  reasoning: false;
@@ -5143,7 +5249,7 @@ export declare const MODELS: {
5143
5249
  readonly "devstral-small-2505": {
5144
5250
  id: string;
5145
5251
  name: string;
5146
- api: "mistral-conversations";
5252
+ api: "openai-completions";
5147
5253
  provider: string;
5148
5254
  baseUrl: string;
5149
5255
  reasoning: false;
@@ -5160,7 +5266,7 @@ export declare const MODELS: {
5160
5266
  readonly "devstral-small-2507": {
5161
5267
  id: string;
5162
5268
  name: string;
5163
- api: "mistral-conversations";
5269
+ api: "openai-completions";
5164
5270
  provider: string;
5165
5271
  baseUrl: string;
5166
5272
  reasoning: false;
@@ -5177,7 +5283,7 @@ export declare const MODELS: {
5177
5283
  readonly "labs-devstral-small-2512": {
5178
5284
  id: string;
5179
5285
  name: string;
5180
- api: "mistral-conversations";
5286
+ api: "openai-completions";
5181
5287
  provider: string;
5182
5288
  baseUrl: string;
5183
5289
  reasoning: false;
@@ -5194,7 +5300,7 @@ export declare const MODELS: {
5194
5300
  readonly "magistral-medium-latest": {
5195
5301
  id: string;
5196
5302
  name: string;
5197
- api: "mistral-conversations";
5303
+ api: "openai-completions";
5198
5304
  provider: string;
5199
5305
  baseUrl: string;
5200
5306
  reasoning: true;
@@ -5211,7 +5317,7 @@ export declare const MODELS: {
5211
5317
  readonly "magistral-small": {
5212
5318
  id: string;
5213
5319
  name: string;
5214
- api: "mistral-conversations";
5320
+ api: "openai-completions";
5215
5321
  provider: string;
5216
5322
  baseUrl: string;
5217
5323
  reasoning: true;
@@ -5228,7 +5334,7 @@ export declare const MODELS: {
5228
5334
  readonly "ministral-3b-latest": {
5229
5335
  id: string;
5230
5336
  name: string;
5231
- api: "mistral-conversations";
5337
+ api: "openai-completions";
5232
5338
  provider: string;
5233
5339
  baseUrl: string;
5234
5340
  reasoning: false;
@@ -5245,7 +5351,7 @@ export declare const MODELS: {
5245
5351
  readonly "ministral-8b-latest": {
5246
5352
  id: string;
5247
5353
  name: string;
5248
- api: "mistral-conversations";
5354
+ api: "openai-completions";
5249
5355
  provider: string;
5250
5356
  baseUrl: string;
5251
5357
  reasoning: false;
@@ -5262,7 +5368,7 @@ export declare const MODELS: {
5262
5368
  readonly "mistral-large-2411": {
5263
5369
  id: string;
5264
5370
  name: string;
5265
- api: "mistral-conversations";
5371
+ api: "openai-completions";
5266
5372
  provider: string;
5267
5373
  baseUrl: string;
5268
5374
  reasoning: false;
@@ -5279,7 +5385,7 @@ export declare const MODELS: {
5279
5385
  readonly "mistral-large-2512": {
5280
5386
  id: string;
5281
5387
  name: string;
5282
- api: "mistral-conversations";
5388
+ api: "openai-completions";
5283
5389
  provider: string;
5284
5390
  baseUrl: string;
5285
5391
  reasoning: false;
@@ -5296,7 +5402,7 @@ export declare const MODELS: {
5296
5402
  readonly "mistral-large-latest": {
5297
5403
  id: string;
5298
5404
  name: string;
5299
- api: "mistral-conversations";
5405
+ api: "openai-completions";
5300
5406
  provider: string;
5301
5407
  baseUrl: string;
5302
5408
  reasoning: false;
@@ -5313,7 +5419,7 @@ export declare const MODELS: {
5313
5419
  readonly "mistral-medium-2505": {
5314
5420
  id: string;
5315
5421
  name: string;
5316
- api: "mistral-conversations";
5422
+ api: "openai-completions";
5317
5423
  provider: string;
5318
5424
  baseUrl: string;
5319
5425
  reasoning: false;
@@ -5330,7 +5436,7 @@ export declare const MODELS: {
5330
5436
  readonly "mistral-medium-2508": {
5331
5437
  id: string;
5332
5438
  name: string;
5333
- api: "mistral-conversations";
5439
+ api: "openai-completions";
5334
5440
  provider: string;
5335
5441
  baseUrl: string;
5336
5442
  reasoning: false;
@@ -5347,7 +5453,7 @@ export declare const MODELS: {
5347
5453
  readonly "mistral-medium-latest": {
5348
5454
  id: string;
5349
5455
  name: string;
5350
- api: "mistral-conversations";
5456
+ api: "openai-completions";
5351
5457
  provider: string;
5352
5458
  baseUrl: string;
5353
5459
  reasoning: false;
@@ -5364,7 +5470,7 @@ export declare const MODELS: {
5364
5470
  readonly "mistral-nemo": {
5365
5471
  id: string;
5366
5472
  name: string;
5367
- api: "mistral-conversations";
5473
+ api: "openai-completions";
5368
5474
  provider: string;
5369
5475
  baseUrl: string;
5370
5476
  reasoning: false;
@@ -5381,7 +5487,7 @@ export declare const MODELS: {
5381
5487
  readonly "mistral-small-2506": {
5382
5488
  id: string;
5383
5489
  name: string;
5384
- api: "mistral-conversations";
5490
+ api: "openai-completions";
5385
5491
  provider: string;
5386
5492
  baseUrl: string;
5387
5493
  reasoning: false;
@@ -5398,7 +5504,7 @@ export declare const MODELS: {
5398
5504
  readonly "mistral-small-2603": {
5399
5505
  id: string;
5400
5506
  name: string;
5401
- api: "mistral-conversations";
5507
+ api: "openai-completions";
5402
5508
  provider: string;
5403
5509
  baseUrl: string;
5404
5510
  reasoning: true;
@@ -5415,7 +5521,7 @@ export declare const MODELS: {
5415
5521
  readonly "mistral-small-latest": {
5416
5522
  id: string;
5417
5523
  name: string;
5418
- api: "mistral-conversations";
5524
+ api: "openai-completions";
5419
5525
  provider: string;
5420
5526
  baseUrl: string;
5421
5527
  reasoning: true;
@@ -5432,7 +5538,7 @@ export declare const MODELS: {
5432
5538
  readonly "open-mistral-7b": {
5433
5539
  id: string;
5434
5540
  name: string;
5435
- api: "mistral-conversations";
5541
+ api: "openai-completions";
5436
5542
  provider: string;
5437
5543
  baseUrl: string;
5438
5544
  reasoning: false;
@@ -5449,7 +5555,7 @@ export declare const MODELS: {
5449
5555
  readonly "open-mixtral-8x22b": {
5450
5556
  id: string;
5451
5557
  name: string;
5452
- api: "mistral-conversations";
5558
+ api: "openai-completions";
5453
5559
  provider: string;
5454
5560
  baseUrl: string;
5455
5561
  reasoning: false;
@@ -5466,7 +5572,7 @@ export declare const MODELS: {
5466
5572
  readonly "open-mixtral-8x7b": {
5467
5573
  id: string;
5468
5574
  name: string;
5469
- api: "mistral-conversations";
5575
+ api: "openai-completions";
5470
5576
  provider: string;
5471
5577
  baseUrl: string;
5472
5578
  reasoning: false;
@@ -5483,7 +5589,7 @@ export declare const MODELS: {
5483
5589
  readonly "pixtral-12b": {
5484
5590
  id: string;
5485
5591
  name: string;
5486
- api: "mistral-conversations";
5592
+ api: "openai-completions";
5487
5593
  provider: string;
5488
5594
  baseUrl: string;
5489
5595
  reasoning: false;
@@ -5500,7 +5606,7 @@ export declare const MODELS: {
5500
5606
  readonly "pixtral-large-latest": {
5501
5607
  id: string;
5502
5608
  name: string;
5503
- api: "mistral-conversations";
5609
+ api: "openai-completions";
5504
5610
  provider: string;
5505
5611
  baseUrl: string;
5506
5612
  reasoning: false;
@@ -5516,23 +5622,6 @@ export declare const MODELS: {
5516
5622
  };
5517
5623
  };
5518
5624
  readonly openai: {
5519
- readonly "codex-mini-latest": {
5520
- id: string;
5521
- name: string;
5522
- api: "openai-responses";
5523
- provider: string;
5524
- baseUrl: string;
5525
- reasoning: true;
5526
- input: "text"[];
5527
- cost: {
5528
- input: number;
5529
- output: number;
5530
- cacheRead: number;
5531
- cacheWrite: number;
5532
- };
5533
- contextWindow: number;
5534
- maxTokens: number;
5535
- };
5536
5625
  readonly "gpt-4": {
5537
5626
  id: string;
5538
5627
  name: string;
@@ -6077,6 +6166,23 @@ export declare const MODELS: {
6077
6166
  contextWindow: number;
6078
6167
  maxTokens: number;
6079
6168
  };
6169
+ readonly "gpt-5.5": {
6170
+ id: string;
6171
+ name: string;
6172
+ api: "openai-responses";
6173
+ provider: string;
6174
+ baseUrl: string;
6175
+ reasoning: true;
6176
+ input: ("image" | "text")[];
6177
+ cost: {
6178
+ input: number;
6179
+ output: number;
6180
+ cacheRead: number;
6181
+ cacheWrite: number;
6182
+ };
6183
+ contextWindow: number;
6184
+ maxTokens: number;
6185
+ };
6080
6186
  readonly o1: {
6081
6187
  id: string;
6082
6188
  name: string;
@@ -6472,6 +6578,23 @@ export declare const MODELS: {
6472
6578
  contextWindow: number;
6473
6579
  maxTokens: number;
6474
6580
  };
6581
+ readonly "claude-opus-4-7": {
6582
+ id: string;
6583
+ name: string;
6584
+ api: "anthropic-messages";
6585
+ provider: string;
6586
+ baseUrl: string;
6587
+ reasoning: true;
6588
+ input: ("image" | "text")[];
6589
+ cost: {
6590
+ input: number;
6591
+ output: number;
6592
+ cacheRead: number;
6593
+ cacheWrite: number;
6594
+ };
6595
+ contextWindow: number;
6596
+ maxTokens: number;
6597
+ };
6475
6598
  readonly "claude-sonnet-4": {
6476
6599
  id: string;
6477
6600
  name: string;
@@ -6574,6 +6697,23 @@ export declare const MODELS: {
6574
6697
  contextWindow: number;
6575
6698
  maxTokens: number;
6576
6699
  };
6700
+ readonly "glm-5.1": {
6701
+ id: string;
6702
+ name: string;
6703
+ api: "openai-completions";
6704
+ provider: string;
6705
+ baseUrl: string;
6706
+ reasoning: true;
6707
+ input: "text"[];
6708
+ cost: {
6709
+ input: number;
6710
+ output: number;
6711
+ cacheRead: number;
6712
+ cacheWrite: number;
6713
+ };
6714
+ contextWindow: number;
6715
+ maxTokens: number;
6716
+ };
6577
6717
  readonly "gpt-5": {
6578
6718
  id: string;
6579
6719
  name: string;
@@ -6812,10 +6952,10 @@ export declare const MODELS: {
6812
6952
  contextWindow: number;
6813
6953
  maxTokens: number;
6814
6954
  };
6815
- readonly "kimi-k2.5": {
6955
+ readonly "gpt-5.5": {
6816
6956
  id: string;
6817
6957
  name: string;
6818
- api: "openai-completions";
6958
+ api: "openai-responses";
6819
6959
  provider: string;
6820
6960
  baseUrl: string;
6821
6961
  reasoning: true;
@@ -6829,14 +6969,14 @@ export declare const MODELS: {
6829
6969
  contextWindow: number;
6830
6970
  maxTokens: number;
6831
6971
  };
6832
- readonly "minimax-m2.5": {
6972
+ readonly "gpt-5.5-pro": {
6833
6973
  id: string;
6834
6974
  name: string;
6835
- api: "openai-completions";
6975
+ api: "openai-responses";
6836
6976
  provider: string;
6837
6977
  baseUrl: string;
6838
6978
  reasoning: true;
6839
- input: "text"[];
6979
+ input: ("image" | "text")[];
6840
6980
  cost: {
6841
6981
  input: number;
6842
6982
  output: number;
@@ -6846,10 +6986,10 @@ export declare const MODELS: {
6846
6986
  contextWindow: number;
6847
6987
  maxTokens: number;
6848
6988
  };
6849
- readonly "minimax-m2.5-free": {
6989
+ readonly "hy3-preview-free": {
6850
6990
  id: string;
6851
6991
  name: string;
6852
- api: "anthropic-messages";
6992
+ api: "openai-completions";
6853
6993
  provider: string;
6854
6994
  baseUrl: string;
6855
6995
  reasoning: true;
@@ -6863,14 +7003,14 @@ export declare const MODELS: {
6863
7003
  contextWindow: number;
6864
7004
  maxTokens: number;
6865
7005
  };
6866
- readonly "nemotron-3-super-free": {
7006
+ readonly "kimi-k2.5": {
6867
7007
  id: string;
6868
7008
  name: string;
6869
7009
  api: "openai-completions";
6870
7010
  provider: string;
6871
7011
  baseUrl: string;
6872
7012
  reasoning: true;
6873
- input: "text"[];
7013
+ input: ("image" | "text")[];
6874
7014
  cost: {
6875
7015
  input: number;
6876
7016
  output: number;
@@ -6880,14 +7020,14 @@ export declare const MODELS: {
6880
7020
  contextWindow: number;
6881
7021
  maxTokens: number;
6882
7022
  };
6883
- readonly "qwen3.6-plus-free": {
7023
+ readonly "kimi-k2.6": {
6884
7024
  id: string;
6885
7025
  name: string;
6886
7026
  api: "openai-completions";
6887
7027
  provider: string;
6888
7028
  baseUrl: string;
6889
7029
  reasoning: true;
6890
- input: "text"[];
7030
+ input: ("image" | "text")[];
6891
7031
  cost: {
6892
7032
  input: number;
6893
7033
  output: number;
@@ -6897,15 +7037,13 @@ export declare const MODELS: {
6897
7037
  contextWindow: number;
6898
7038
  maxTokens: number;
6899
7039
  };
6900
- };
6901
- readonly "opencode-go": {
6902
- readonly "glm-5": {
7040
+ readonly "ling-2.6-flash-free": {
6903
7041
  id: string;
6904
7042
  name: string;
6905
7043
  api: "openai-completions";
6906
7044
  provider: string;
6907
7045
  baseUrl: string;
6908
- reasoning: true;
7046
+ reasoning: false;
6909
7047
  input: "text"[];
6910
7048
  cost: {
6911
7049
  input: number;
@@ -6916,14 +7054,14 @@ export declare const MODELS: {
6916
7054
  contextWindow: number;
6917
7055
  maxTokens: number;
6918
7056
  };
6919
- readonly "kimi-k2.5": {
7057
+ readonly "minimax-m2.5": {
6920
7058
  id: string;
6921
7059
  name: string;
6922
7060
  api: "openai-completions";
6923
7061
  provider: string;
6924
7062
  baseUrl: string;
6925
7063
  reasoning: true;
6926
- input: ("image" | "text")[];
7064
+ input: "text"[];
6927
7065
  cost: {
6928
7066
  input: number;
6929
7067
  output: number;
@@ -6933,14 +7071,14 @@ export declare const MODELS: {
6933
7071
  contextWindow: number;
6934
7072
  maxTokens: number;
6935
7073
  };
6936
- readonly "mimo-v2-omni": {
7074
+ readonly "minimax-m2.5-free": {
6937
7075
  id: string;
6938
7076
  name: string;
6939
- api: "openai-completions";
7077
+ api: "anthropic-messages";
6940
7078
  provider: string;
6941
7079
  baseUrl: string;
6942
7080
  reasoning: true;
6943
- input: ("image" | "text")[];
7081
+ input: "text"[];
6944
7082
  cost: {
6945
7083
  input: number;
6946
7084
  output: number;
@@ -6950,7 +7088,7 @@ export declare const MODELS: {
6950
7088
  contextWindow: number;
6951
7089
  maxTokens: number;
6952
7090
  };
6953
- readonly "mimo-v2-pro": {
7091
+ readonly "minimax-m2.7": {
6954
7092
  id: string;
6955
7093
  name: string;
6956
7094
  api: "openai-completions";
@@ -6967,7 +7105,7 @@ export declare const MODELS: {
6967
7105
  contextWindow: number;
6968
7106
  maxTokens: number;
6969
7107
  };
6970
- readonly "minimax-m2.5": {
7108
+ readonly "nemotron-3-super-free": {
6971
7109
  id: string;
6972
7110
  name: string;
6973
7111
  api: "openai-completions";
@@ -6984,14 +7122,14 @@ export declare const MODELS: {
6984
7122
  contextWindow: number;
6985
7123
  maxTokens: number;
6986
7124
  };
6987
- readonly "minimax-m2.7": {
7125
+ readonly "qwen3.5-plus": {
6988
7126
  id: string;
6989
7127
  name: string;
6990
7128
  api: "anthropic-messages";
6991
7129
  provider: string;
6992
7130
  baseUrl: string;
6993
7131
  reasoning: true;
6994
- input: "text"[];
7132
+ input: ("image" | "text")[];
6995
7133
  cost: {
6996
7134
  input: number;
6997
7135
  output: number;
@@ -7001,16 +7139,14 @@ export declare const MODELS: {
7001
7139
  contextWindow: number;
7002
7140
  maxTokens: number;
7003
7141
  };
7004
- };
7005
- readonly openrouter: {
7006
- readonly "ai21/jamba-large-1.7": {
7142
+ readonly "qwen3.6-plus": {
7007
7143
  id: string;
7008
7144
  name: string;
7009
- api: "openai-completions";
7145
+ api: "anthropic-messages";
7010
7146
  provider: string;
7011
7147
  baseUrl: string;
7012
- reasoning: false;
7013
- input: "text"[];
7148
+ reasoning: true;
7149
+ input: ("image" | "text")[];
7014
7150
  cost: {
7015
7151
  input: number;
7016
7152
  output: number;
@@ -7020,8 +7156,267 @@ export declare const MODELS: {
7020
7156
  contextWindow: number;
7021
7157
  maxTokens: number;
7022
7158
  };
7023
- readonly "alibaba/tongyi-deepresearch-30b-a3b": {
7024
- id: string;
7159
+ };
7160
+ readonly "opencode-go": {
7161
+ readonly "deepseek-v4-flash": {
7162
+ id: string;
7163
+ name: string;
7164
+ api: "openai-completions";
7165
+ provider: string;
7166
+ baseUrl: string;
7167
+ reasoning: true;
7168
+ input: "text"[];
7169
+ cost: {
7170
+ input: number;
7171
+ output: number;
7172
+ cacheRead: number;
7173
+ cacheWrite: number;
7174
+ };
7175
+ contextWindow: number;
7176
+ maxTokens: number;
7177
+ };
7178
+ readonly "deepseek-v4-pro": {
7179
+ id: string;
7180
+ name: string;
7181
+ api: "openai-completions";
7182
+ provider: string;
7183
+ baseUrl: string;
7184
+ reasoning: true;
7185
+ input: "text"[];
7186
+ cost: {
7187
+ input: number;
7188
+ output: number;
7189
+ cacheRead: number;
7190
+ cacheWrite: number;
7191
+ };
7192
+ contextWindow: number;
7193
+ maxTokens: number;
7194
+ };
7195
+ readonly "glm-5": {
7196
+ id: string;
7197
+ name: string;
7198
+ api: "openai-completions";
7199
+ provider: string;
7200
+ baseUrl: string;
7201
+ reasoning: true;
7202
+ input: "text"[];
7203
+ cost: {
7204
+ input: number;
7205
+ output: number;
7206
+ cacheRead: number;
7207
+ cacheWrite: number;
7208
+ };
7209
+ contextWindow: number;
7210
+ maxTokens: number;
7211
+ };
7212
+ readonly "glm-5.1": {
7213
+ id: string;
7214
+ name: string;
7215
+ api: "openai-completions";
7216
+ provider: string;
7217
+ baseUrl: string;
7218
+ reasoning: true;
7219
+ input: "text"[];
7220
+ cost: {
7221
+ input: number;
7222
+ output: number;
7223
+ cacheRead: number;
7224
+ cacheWrite: number;
7225
+ };
7226
+ contextWindow: number;
7227
+ maxTokens: number;
7228
+ };
7229
+ readonly "kimi-k2.5": {
7230
+ id: string;
7231
+ name: string;
7232
+ api: "openai-completions";
7233
+ provider: string;
7234
+ baseUrl: string;
7235
+ reasoning: true;
7236
+ input: ("image" | "text")[];
7237
+ cost: {
7238
+ input: number;
7239
+ output: number;
7240
+ cacheRead: number;
7241
+ cacheWrite: number;
7242
+ };
7243
+ contextWindow: number;
7244
+ maxTokens: number;
7245
+ };
7246
+ readonly "kimi-k2.6": {
7247
+ id: string;
7248
+ name: string;
7249
+ api: "openai-completions";
7250
+ provider: string;
7251
+ baseUrl: string;
7252
+ reasoning: true;
7253
+ input: ("image" | "text")[];
7254
+ cost: {
7255
+ input: number;
7256
+ output: number;
7257
+ cacheRead: number;
7258
+ cacheWrite: number;
7259
+ };
7260
+ contextWindow: number;
7261
+ maxTokens: number;
7262
+ };
7263
+ readonly "mimo-v2-omni": {
7264
+ id: string;
7265
+ name: string;
7266
+ api: "openai-completions";
7267
+ provider: string;
7268
+ baseUrl: string;
7269
+ reasoning: true;
7270
+ input: ("image" | "text")[];
7271
+ cost: {
7272
+ input: number;
7273
+ output: number;
7274
+ cacheRead: number;
7275
+ cacheWrite: number;
7276
+ };
7277
+ contextWindow: number;
7278
+ maxTokens: number;
7279
+ };
7280
+ readonly "mimo-v2-pro": {
7281
+ id: string;
7282
+ name: string;
7283
+ api: "openai-completions";
7284
+ provider: string;
7285
+ baseUrl: string;
7286
+ reasoning: true;
7287
+ input: "text"[];
7288
+ cost: {
7289
+ input: number;
7290
+ output: number;
7291
+ cacheRead: number;
7292
+ cacheWrite: number;
7293
+ };
7294
+ contextWindow: number;
7295
+ maxTokens: number;
7296
+ };
7297
+ readonly "mimo-v2.5": {
7298
+ id: string;
7299
+ name: string;
7300
+ api: "openai-completions";
7301
+ provider: string;
7302
+ baseUrl: string;
7303
+ reasoning: true;
7304
+ input: ("image" | "text")[];
7305
+ cost: {
7306
+ input: number;
7307
+ output: number;
7308
+ cacheRead: number;
7309
+ cacheWrite: number;
7310
+ };
7311
+ contextWindow: number;
7312
+ maxTokens: number;
7313
+ };
7314
+ readonly "mimo-v2.5-pro": {
7315
+ id: string;
7316
+ name: string;
7317
+ api: "openai-completions";
7318
+ provider: string;
7319
+ baseUrl: string;
7320
+ reasoning: true;
7321
+ input: "text"[];
7322
+ cost: {
7323
+ input: number;
7324
+ output: number;
7325
+ cacheRead: number;
7326
+ cacheWrite: number;
7327
+ };
7328
+ contextWindow: number;
7329
+ maxTokens: number;
7330
+ };
7331
+ readonly "minimax-m2.5": {
7332
+ id: string;
7333
+ name: string;
7334
+ api: "openai-completions";
7335
+ provider: string;
7336
+ baseUrl: string;
7337
+ reasoning: true;
7338
+ input: "text"[];
7339
+ cost: {
7340
+ input: number;
7341
+ output: number;
7342
+ cacheRead: number;
7343
+ cacheWrite: number;
7344
+ };
7345
+ contextWindow: number;
7346
+ maxTokens: number;
7347
+ };
7348
+ readonly "minimax-m2.7": {
7349
+ id: string;
7350
+ name: string;
7351
+ api: "anthropic-messages";
7352
+ provider: string;
7353
+ baseUrl: string;
7354
+ reasoning: true;
7355
+ input: "text"[];
7356
+ cost: {
7357
+ input: number;
7358
+ output: number;
7359
+ cacheRead: number;
7360
+ cacheWrite: number;
7361
+ };
7362
+ contextWindow: number;
7363
+ maxTokens: number;
7364
+ };
7365
+ readonly "qwen3.5-plus": {
7366
+ id: string;
7367
+ name: string;
7368
+ api: "anthropic-messages";
7369
+ provider: string;
7370
+ baseUrl: string;
7371
+ reasoning: true;
7372
+ input: ("image" | "text")[];
7373
+ cost: {
7374
+ input: number;
7375
+ output: number;
7376
+ cacheRead: number;
7377
+ cacheWrite: number;
7378
+ };
7379
+ contextWindow: number;
7380
+ maxTokens: number;
7381
+ };
7382
+ readonly "qwen3.6-plus": {
7383
+ id: string;
7384
+ name: string;
7385
+ api: "anthropic-messages";
7386
+ provider: string;
7387
+ baseUrl: string;
7388
+ reasoning: true;
7389
+ input: ("image" | "text")[];
7390
+ cost: {
7391
+ input: number;
7392
+ output: number;
7393
+ cacheRead: number;
7394
+ cacheWrite: number;
7395
+ };
7396
+ contextWindow: number;
7397
+ maxTokens: number;
7398
+ };
7399
+ };
7400
+ readonly openrouter: {
7401
+ readonly "ai21/jamba-large-1.7": {
7402
+ id: string;
7403
+ name: string;
7404
+ api: "openai-completions";
7405
+ provider: string;
7406
+ baseUrl: string;
7407
+ reasoning: false;
7408
+ input: "text"[];
7409
+ cost: {
7410
+ input: number;
7411
+ output: number;
7412
+ cacheRead: number;
7413
+ cacheWrite: number;
7414
+ };
7415
+ contextWindow: number;
7416
+ maxTokens: number;
7417
+ };
7418
+ readonly "alibaba/tongyi-deepresearch-30b-a3b": {
7419
+ id: string;
7025
7420
  name: string;
7026
7421
  api: "openai-completions";
7027
7422
  provider: string;
@@ -7292,7 +7687,7 @@ export declare const MODELS: {
7292
7687
  contextWindow: number;
7293
7688
  maxTokens: number;
7294
7689
  };
7295
- readonly "anthropic/claude-sonnet-4": {
7690
+ readonly "anthropic/claude-opus-4.6-fast": {
7296
7691
  id: string;
7297
7692
  name: string;
7298
7693
  api: "openai-completions";
@@ -7309,7 +7704,7 @@ export declare const MODELS: {
7309
7704
  contextWindow: number;
7310
7705
  maxTokens: number;
7311
7706
  };
7312
- readonly "anthropic/claude-sonnet-4.5": {
7707
+ readonly "anthropic/claude-opus-4.7": {
7313
7708
  id: string;
7314
7709
  name: string;
7315
7710
  api: "openai-completions";
@@ -7326,7 +7721,7 @@ export declare const MODELS: {
7326
7721
  contextWindow: number;
7327
7722
  maxTokens: number;
7328
7723
  };
7329
- readonly "anthropic/claude-sonnet-4.6": {
7724
+ readonly "anthropic/claude-sonnet-4": {
7330
7725
  id: string;
7331
7726
  name: string;
7332
7727
  api: "openai-completions";
@@ -7343,14 +7738,14 @@ export declare const MODELS: {
7343
7738
  contextWindow: number;
7344
7739
  maxTokens: number;
7345
7740
  };
7346
- readonly "arcee-ai/trinity-large-preview:free": {
7741
+ readonly "anthropic/claude-sonnet-4.5": {
7347
7742
  id: string;
7348
7743
  name: string;
7349
7744
  api: "openai-completions";
7350
7745
  provider: string;
7351
7746
  baseUrl: string;
7352
- reasoning: false;
7353
- input: "text"[];
7747
+ reasoning: true;
7748
+ input: ("image" | "text")[];
7354
7749
  cost: {
7355
7750
  input: number;
7356
7751
  output: number;
@@ -7360,13 +7755,30 @@ export declare const MODELS: {
7360
7755
  contextWindow: number;
7361
7756
  maxTokens: number;
7362
7757
  };
7363
- readonly "arcee-ai/trinity-large-thinking": {
7758
+ readonly "anthropic/claude-sonnet-4.6": {
7364
7759
  id: string;
7365
7760
  name: string;
7366
7761
  api: "openai-completions";
7367
7762
  provider: string;
7368
7763
  baseUrl: string;
7369
7764
  reasoning: true;
7765
+ input: ("image" | "text")[];
7766
+ cost: {
7767
+ input: number;
7768
+ output: number;
7769
+ cacheRead: number;
7770
+ cacheWrite: number;
7771
+ };
7772
+ contextWindow: number;
7773
+ maxTokens: number;
7774
+ };
7775
+ readonly "arcee-ai/trinity-large-preview": {
7776
+ id: string;
7777
+ name: string;
7778
+ api: "openai-completions";
7779
+ provider: string;
7780
+ baseUrl: string;
7781
+ reasoning: false;
7370
7782
  input: "text"[];
7371
7783
  cost: {
7372
7784
  input: number;
@@ -7377,7 +7789,7 @@ export declare const MODELS: {
7377
7789
  contextWindow: number;
7378
7790
  maxTokens: number;
7379
7791
  };
7380
- readonly "arcee-ai/trinity-mini": {
7792
+ readonly "arcee-ai/trinity-large-thinking": {
7381
7793
  id: string;
7382
7794
  name: string;
7383
7795
  api: "openai-completions";
@@ -7394,7 +7806,7 @@ export declare const MODELS: {
7394
7806
  contextWindow: number;
7395
7807
  maxTokens: number;
7396
7808
  };
7397
- readonly "arcee-ai/trinity-mini:free": {
7809
+ readonly "arcee-ai/trinity-mini": {
7398
7810
  id: string;
7399
7811
  name: string;
7400
7812
  api: "openai-completions";
@@ -7604,7 +8016,7 @@ export declare const MODELS: {
7604
8016
  api: "openai-completions";
7605
8017
  provider: string;
7606
8018
  baseUrl: string;
7607
- reasoning: true;
8019
+ reasoning: false;
7608
8020
  input: "text"[];
7609
8021
  cost: {
7610
8022
  input: number;
@@ -7717,6 +8129,40 @@ export declare const MODELS: {
7717
8129
  contextWindow: number;
7718
8130
  maxTokens: number;
7719
8131
  };
8132
+ readonly "deepseek/deepseek-v4-flash": {
8133
+ id: string;
8134
+ name: string;
8135
+ api: "openai-completions";
8136
+ provider: string;
8137
+ baseUrl: string;
8138
+ reasoning: true;
8139
+ input: "text"[];
8140
+ cost: {
8141
+ input: number;
8142
+ output: number;
8143
+ cacheRead: number;
8144
+ cacheWrite: number;
8145
+ };
8146
+ contextWindow: number;
8147
+ maxTokens: number;
8148
+ };
8149
+ readonly "deepseek/deepseek-v4-pro": {
8150
+ id: string;
8151
+ name: string;
8152
+ api: "openai-completions";
8153
+ provider: string;
8154
+ baseUrl: string;
8155
+ reasoning: true;
8156
+ input: "text"[];
8157
+ cost: {
8158
+ input: number;
8159
+ output: number;
8160
+ cacheRead: number;
8161
+ cacheWrite: number;
8162
+ };
8163
+ contextWindow: number;
8164
+ maxTokens: number;
8165
+ };
7720
8166
  readonly "essentialai/rnj-1-instruct": {
7721
8167
  id: string;
7722
8168
  name: string;
@@ -7785,7 +8231,58 @@ export declare const MODELS: {
7785
8231
  contextWindow: number;
7786
8232
  maxTokens: number;
7787
8233
  };
7788
- readonly "google/gemini-2.5-flash-lite": {
8234
+ readonly "google/gemini-2.5-flash-lite": {
8235
+ id: string;
8236
+ name: string;
8237
+ api: "openai-completions";
8238
+ provider: string;
8239
+ baseUrl: string;
8240
+ reasoning: true;
8241
+ input: ("image" | "text")[];
8242
+ cost: {
8243
+ input: number;
8244
+ output: number;
8245
+ cacheRead: number;
8246
+ cacheWrite: number;
8247
+ };
8248
+ contextWindow: number;
8249
+ maxTokens: number;
8250
+ };
8251
+ readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
8252
+ id: string;
8253
+ name: string;
8254
+ api: "openai-completions";
8255
+ provider: string;
8256
+ baseUrl: string;
8257
+ reasoning: true;
8258
+ input: ("image" | "text")[];
8259
+ cost: {
8260
+ input: number;
8261
+ output: number;
8262
+ cacheRead: number;
8263
+ cacheWrite: number;
8264
+ };
8265
+ contextWindow: number;
8266
+ maxTokens: number;
8267
+ };
8268
+ readonly "google/gemini-2.5-pro": {
8269
+ id: string;
8270
+ name: string;
8271
+ api: "openai-completions";
8272
+ provider: string;
8273
+ baseUrl: string;
8274
+ reasoning: true;
8275
+ input: ("image" | "text")[];
8276
+ cost: {
8277
+ input: number;
8278
+ output: number;
8279
+ cacheRead: number;
8280
+ cacheWrite: number;
8281
+ };
8282
+ contextWindow: number;
8283
+ maxTokens: number;
8284
+ };
8285
+ readonly "google/gemini-2.5-pro-preview": {
7789
8286
  id: string;
7790
8287
  name: string;
7791
8288
  api: "openai-completions";
@@ -7802,7 +8299,7 @@ export declare const MODELS: {
7802
8299
  contextWindow: number;
7803
8300
  maxTokens: number;
7804
8301
  };
7805
- readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
8302
+ readonly "google/gemini-2.5-pro-preview-05-06": {
7806
8303
  id: string;
7807
8304
  name: string;
7808
8305
  api: "openai-completions";
@@ -7819,7 +8316,7 @@ export declare const MODELS: {
7819
8316
  contextWindow: number;
7820
8317
  maxTokens: number;
7821
8318
  };
7822
- readonly "google/gemini-2.5-pro": {
8319
+ readonly "google/gemini-3-flash-preview": {
7823
8320
  id: string;
7824
8321
  name: string;
7825
8322
  api: "openai-completions";
@@ -7836,7 +8333,7 @@ export declare const MODELS: {
7836
8333
  contextWindow: number;
7837
8334
  maxTokens: number;
7838
8335
  };
7839
- readonly "google/gemini-2.5-pro-preview": {
8336
+ readonly "google/gemini-3.1-flash-lite-preview": {
7840
8337
  id: string;
7841
8338
  name: string;
7842
8339
  api: "openai-completions";
@@ -7853,7 +8350,7 @@ export declare const MODELS: {
7853
8350
  contextWindow: number;
7854
8351
  maxTokens: number;
7855
8352
  };
7856
- readonly "google/gemini-2.5-pro-preview-05-06": {
8353
+ readonly "google/gemini-3.1-pro-preview": {
7857
8354
  id: string;
7858
8355
  name: string;
7859
8356
  api: "openai-completions";
@@ -7870,7 +8367,7 @@ export declare const MODELS: {
7870
8367
  contextWindow: number;
7871
8368
  maxTokens: number;
7872
8369
  };
7873
- readonly "google/gemini-3-flash-preview": {
8370
+ readonly "google/gemini-3.1-pro-preview-customtools": {
7874
8371
  id: string;
7875
8372
  name: string;
7876
8373
  api: "openai-completions";
@@ -7887,13 +8384,13 @@ export declare const MODELS: {
7887
8384
  contextWindow: number;
7888
8385
  maxTokens: number;
7889
8386
  };
7890
- readonly "google/gemini-3.1-flash-lite-preview": {
8387
+ readonly "google/gemma-3-12b-it": {
7891
8388
  id: string;
7892
8389
  name: string;
7893
8390
  api: "openai-completions";
7894
8391
  provider: string;
7895
8392
  baseUrl: string;
7896
- reasoning: true;
8393
+ reasoning: false;
7897
8394
  input: ("image" | "text")[];
7898
8395
  cost: {
7899
8396
  input: number;
@@ -7904,13 +8401,13 @@ export declare const MODELS: {
7904
8401
  contextWindow: number;
7905
8402
  maxTokens: number;
7906
8403
  };
7907
- readonly "google/gemini-3.1-pro-preview": {
8404
+ readonly "google/gemma-3-27b-it": {
7908
8405
  id: string;
7909
8406
  name: string;
7910
8407
  api: "openai-completions";
7911
8408
  provider: string;
7912
8409
  baseUrl: string;
7913
- reasoning: true;
8410
+ reasoning: false;
7914
8411
  input: ("image" | "text")[];
7915
8412
  cost: {
7916
8413
  input: number;
@@ -7921,7 +8418,7 @@ export declare const MODELS: {
7921
8418
  contextWindow: number;
7922
8419
  maxTokens: number;
7923
8420
  };
7924
- readonly "google/gemini-3.1-pro-preview-customtools": {
8421
+ readonly "google/gemma-4-26b-a4b-it": {
7925
8422
  id: string;
7926
8423
  name: string;
7927
8424
  api: "openai-completions";
@@ -7938,7 +8435,7 @@ export declare const MODELS: {
7938
8435
  contextWindow: number;
7939
8436
  maxTokens: number;
7940
8437
  };
7941
- readonly "google/gemma-4-26b-a4b-it": {
8438
+ readonly "google/gemma-4-26b-a4b-it:free": {
7942
8439
  id: string;
7943
8440
  name: string;
7944
8441
  api: "openai-completions";
@@ -7972,14 +8469,14 @@ export declare const MODELS: {
7972
8469
  contextWindow: number;
7973
8470
  maxTokens: number;
7974
8471
  };
7975
- readonly "inception/mercury": {
8472
+ readonly "google/gemma-4-31b-it:free": {
7976
8473
  id: string;
7977
8474
  name: string;
7978
8475
  api: "openai-completions";
7979
8476
  provider: string;
7980
8477
  baseUrl: string;
7981
- reasoning: false;
7982
- input: "text"[];
8478
+ reasoning: true;
8479
+ input: ("image" | "text")[];
7983
8480
  cost: {
7984
8481
  input: number;
7985
8482
  output: number;
@@ -8006,7 +8503,7 @@ export declare const MODELS: {
8006
8503
  contextWindow: number;
8007
8504
  maxTokens: number;
8008
8505
  };
8009
- readonly "inception/mercury-coder": {
8506
+ readonly "inclusionai/ling-2.6-1t:free": {
8010
8507
  id: string;
8011
8508
  name: string;
8012
8509
  api: "openai-completions";
@@ -8023,7 +8520,7 @@ export declare const MODELS: {
8023
8520
  contextWindow: number;
8024
8521
  maxTokens: number;
8025
8522
  };
8026
- readonly "kwaipilot/kat-coder-pro-v2": {
8523
+ readonly "inclusionai/ling-2.6-flash:free": {
8027
8524
  id: string;
8028
8525
  name: string;
8029
8526
  api: "openai-completions";
@@ -8040,7 +8537,7 @@ export declare const MODELS: {
8040
8537
  contextWindow: number;
8041
8538
  maxTokens: number;
8042
8539
  };
8043
- readonly "meituan/longcat-flash-chat": {
8540
+ readonly "kwaipilot/kat-coder-pro-v2": {
8044
8541
  id: string;
8045
8542
  name: string;
8046
8543
  api: "openai-completions";
@@ -8142,23 +8639,6 @@ export declare const MODELS: {
8142
8639
  contextWindow: number;
8143
8640
  maxTokens: number;
8144
8641
  };
8145
- readonly "meta-llama/llama-4-maverick": {
8146
- id: string;
8147
- name: string;
8148
- api: "openai-completions";
8149
- provider: string;
8150
- baseUrl: string;
8151
- reasoning: false;
8152
- input: ("image" | "text")[];
8153
- cost: {
8154
- input: number;
8155
- output: number;
8156
- cacheRead: number;
8157
- cacheWrite: number;
8158
- };
8159
- contextWindow: number;
8160
- maxTokens: number;
8161
- };
8162
8642
  readonly "meta-llama/llama-4-scout": {
8163
8643
  id: string;
8164
8644
  name: string;
@@ -8720,6 +9200,23 @@ export declare const MODELS: {
8720
9200
  contextWindow: number;
8721
9201
  maxTokens: number;
8722
9202
  };
9203
+ readonly "moonshotai/kimi-k2.6": {
9204
+ id: string;
9205
+ name: string;
9206
+ api: "openai-completions";
9207
+ provider: string;
9208
+ baseUrl: string;
9209
+ reasoning: true;
9210
+ input: ("image" | "text")[];
9211
+ cost: {
9212
+ input: number;
9213
+ output: number;
9214
+ cacheRead: number;
9215
+ cacheWrite: number;
9216
+ };
9217
+ contextWindow: number;
9218
+ maxTokens: number;
9219
+ };
8723
9220
  readonly "nex-agi/deepseek-v3.1-nex-n1": {
8724
9221
  id: string;
8725
9222
  name: string;
@@ -9196,23 +9693,6 @@ export declare const MODELS: {
9196
9693
  contextWindow: number;
9197
9694
  maxTokens: number;
9198
9695
  };
9199
- readonly "openai/gpt-4o:extended": {
9200
- id: string;
9201
- name: string;
9202
- api: "openai-completions";
9203
- provider: string;
9204
- baseUrl: string;
9205
- reasoning: false;
9206
- input: ("image" | "text")[];
9207
- cost: {
9208
- input: number;
9209
- output: number;
9210
- cacheRead: number;
9211
- cacheWrite: number;
9212
- };
9213
- contextWindow: number;
9214
- maxTokens: number;
9215
- };
9216
9696
  readonly "openai/gpt-5": {
9217
9697
  id: string;
9218
9698
  name: string;
@@ -9247,7 +9727,7 @@ export declare const MODELS: {
9247
9727
  contextWindow: number;
9248
9728
  maxTokens: number;
9249
9729
  };
9250
- readonly "openai/gpt-5-image": {
9730
+ readonly "openai/gpt-5-mini": {
9251
9731
  id: string;
9252
9732
  name: string;
9253
9733
  api: "openai-completions";
@@ -9264,7 +9744,7 @@ export declare const MODELS: {
9264
9744
  contextWindow: number;
9265
9745
  maxTokens: number;
9266
9746
  };
9267
- readonly "openai/gpt-5-image-mini": {
9747
+ readonly "openai/gpt-5-nano": {
9268
9748
  id: string;
9269
9749
  name: string;
9270
9750
  api: "openai-completions";
@@ -9281,7 +9761,7 @@ export declare const MODELS: {
9281
9761
  contextWindow: number;
9282
9762
  maxTokens: number;
9283
9763
  };
9284
- readonly "openai/gpt-5-mini": {
9764
+ readonly "openai/gpt-5-pro": {
9285
9765
  id: string;
9286
9766
  name: string;
9287
9767
  api: "openai-completions";
@@ -9298,7 +9778,7 @@ export declare const MODELS: {
9298
9778
  contextWindow: number;
9299
9779
  maxTokens: number;
9300
9780
  };
9301
- readonly "openai/gpt-5-nano": {
9781
+ readonly "openai/gpt-5.1": {
9302
9782
  id: string;
9303
9783
  name: string;
9304
9784
  api: "openai-completions";
@@ -9315,13 +9795,13 @@ export declare const MODELS: {
9315
9795
  contextWindow: number;
9316
9796
  maxTokens: number;
9317
9797
  };
9318
- readonly "openai/gpt-5-pro": {
9798
+ readonly "openai/gpt-5.1-chat": {
9319
9799
  id: string;
9320
9800
  name: string;
9321
9801
  api: "openai-completions";
9322
9802
  provider: string;
9323
9803
  baseUrl: string;
9324
- reasoning: true;
9804
+ reasoning: false;
9325
9805
  input: ("image" | "text")[];
9326
9806
  cost: {
9327
9807
  input: number;
@@ -9332,7 +9812,7 @@ export declare const MODELS: {
9332
9812
  contextWindow: number;
9333
9813
  maxTokens: number;
9334
9814
  };
9335
- readonly "openai/gpt-5.1": {
9815
+ readonly "openai/gpt-5.1-codex": {
9336
9816
  id: string;
9337
9817
  name: string;
9338
9818
  api: "openai-completions";
@@ -9349,13 +9829,13 @@ export declare const MODELS: {
9349
9829
  contextWindow: number;
9350
9830
  maxTokens: number;
9351
9831
  };
9352
- readonly "openai/gpt-5.1-chat": {
9832
+ readonly "openai/gpt-5.1-codex-max": {
9353
9833
  id: string;
9354
9834
  name: string;
9355
9835
  api: "openai-completions";
9356
9836
  provider: string;
9357
9837
  baseUrl: string;
9358
- reasoning: false;
9838
+ reasoning: true;
9359
9839
  input: ("image" | "text")[];
9360
9840
  cost: {
9361
9841
  input: number;
@@ -9366,7 +9846,7 @@ export declare const MODELS: {
9366
9846
  contextWindow: number;
9367
9847
  maxTokens: number;
9368
9848
  };
9369
- readonly "openai/gpt-5.1-codex": {
9849
+ readonly "openai/gpt-5.1-codex-mini": {
9370
9850
  id: string;
9371
9851
  name: string;
9372
9852
  api: "openai-completions";
@@ -9383,7 +9863,7 @@ export declare const MODELS: {
9383
9863
  contextWindow: number;
9384
9864
  maxTokens: number;
9385
9865
  };
9386
- readonly "openai/gpt-5.1-codex-max": {
9866
+ readonly "openai/gpt-5.2": {
9387
9867
  id: string;
9388
9868
  name: string;
9389
9869
  api: "openai-completions";
@@ -9400,13 +9880,13 @@ export declare const MODELS: {
9400
9880
  contextWindow: number;
9401
9881
  maxTokens: number;
9402
9882
  };
9403
- readonly "openai/gpt-5.1-codex-mini": {
9883
+ readonly "openai/gpt-5.2-chat": {
9404
9884
  id: string;
9405
9885
  name: string;
9406
9886
  api: "openai-completions";
9407
9887
  provider: string;
9408
9888
  baseUrl: string;
9409
- reasoning: true;
9889
+ reasoning: false;
9410
9890
  input: ("image" | "text")[];
9411
9891
  cost: {
9412
9892
  input: number;
@@ -9417,7 +9897,7 @@ export declare const MODELS: {
9417
9897
  contextWindow: number;
9418
9898
  maxTokens: number;
9419
9899
  };
9420
- readonly "openai/gpt-5.2": {
9900
+ readonly "openai/gpt-5.2-codex": {
9421
9901
  id: string;
9422
9902
  name: string;
9423
9903
  api: "openai-completions";
@@ -9434,13 +9914,13 @@ export declare const MODELS: {
9434
9914
  contextWindow: number;
9435
9915
  maxTokens: number;
9436
9916
  };
9437
- readonly "openai/gpt-5.2-chat": {
9917
+ readonly "openai/gpt-5.2-pro": {
9438
9918
  id: string;
9439
9919
  name: string;
9440
9920
  api: "openai-completions";
9441
9921
  provider: string;
9442
9922
  baseUrl: string;
9443
- reasoning: false;
9923
+ reasoning: true;
9444
9924
  input: ("image" | "text")[];
9445
9925
  cost: {
9446
9926
  input: number;
@@ -9451,13 +9931,13 @@ export declare const MODELS: {
9451
9931
  contextWindow: number;
9452
9932
  maxTokens: number;
9453
9933
  };
9454
- readonly "openai/gpt-5.2-codex": {
9934
+ readonly "openai/gpt-5.3-chat": {
9455
9935
  id: string;
9456
9936
  name: string;
9457
9937
  api: "openai-completions";
9458
9938
  provider: string;
9459
9939
  baseUrl: string;
9460
- reasoning: true;
9940
+ reasoning: false;
9461
9941
  input: ("image" | "text")[];
9462
9942
  cost: {
9463
9943
  input: number;
@@ -9468,7 +9948,7 @@ export declare const MODELS: {
9468
9948
  contextWindow: number;
9469
9949
  maxTokens: number;
9470
9950
  };
9471
- readonly "openai/gpt-5.2-pro": {
9951
+ readonly "openai/gpt-5.3-codex": {
9472
9952
  id: string;
9473
9953
  name: string;
9474
9954
  api: "openai-completions";
@@ -9485,13 +9965,13 @@ export declare const MODELS: {
9485
9965
  contextWindow: number;
9486
9966
  maxTokens: number;
9487
9967
  };
9488
- readonly "openai/gpt-5.3-chat": {
9968
+ readonly "openai/gpt-5.4": {
9489
9969
  id: string;
9490
9970
  name: string;
9491
9971
  api: "openai-completions";
9492
9972
  provider: string;
9493
9973
  baseUrl: string;
9494
- reasoning: false;
9974
+ reasoning: true;
9495
9975
  input: ("image" | "text")[];
9496
9976
  cost: {
9497
9977
  input: number;
@@ -9502,7 +9982,7 @@ export declare const MODELS: {
9502
9982
  contextWindow: number;
9503
9983
  maxTokens: number;
9504
9984
  };
9505
- readonly "openai/gpt-5.3-codex": {
9985
+ readonly "openai/gpt-5.4-mini": {
9506
9986
  id: string;
9507
9987
  name: string;
9508
9988
  api: "openai-completions";
@@ -9519,7 +9999,7 @@ export declare const MODELS: {
9519
9999
  contextWindow: number;
9520
10000
  maxTokens: number;
9521
10001
  };
9522
- readonly "openai/gpt-5.4": {
10002
+ readonly "openai/gpt-5.4-nano": {
9523
10003
  id: string;
9524
10004
  name: string;
9525
10005
  api: "openai-completions";
@@ -9536,7 +10016,7 @@ export declare const MODELS: {
9536
10016
  contextWindow: number;
9537
10017
  maxTokens: number;
9538
10018
  };
9539
- readonly "openai/gpt-5.4-mini": {
10019
+ readonly "openai/gpt-5.4-pro": {
9540
10020
  id: string;
9541
10021
  name: string;
9542
10022
  api: "openai-completions";
@@ -9553,7 +10033,7 @@ export declare const MODELS: {
9553
10033
  contextWindow: number;
9554
10034
  maxTokens: number;
9555
10035
  };
9556
- readonly "openai/gpt-5.4-nano": {
10036
+ readonly "openai/gpt-5.5": {
9557
10037
  id: string;
9558
10038
  name: string;
9559
10039
  api: "openai-completions";
@@ -9570,7 +10050,7 @@ export declare const MODELS: {
9570
10050
  contextWindow: number;
9571
10051
  maxTokens: number;
9572
10052
  };
9573
- readonly "openai/gpt-5.4-pro": {
10053
+ readonly "openai/gpt-5.5-pro": {
9574
10054
  id: string;
9575
10055
  name: string;
9576
10056
  api: "openai-completions";
@@ -10086,7 +10566,7 @@ export declare const MODELS: {
10086
10566
  api: "openai-completions";
10087
10567
  provider: string;
10088
10568
  baseUrl: string;
10089
- reasoning: true;
10569
+ reasoning: false;
10090
10570
  input: "text"[];
10091
10571
  cost: {
10092
10572
  input: number;
@@ -10624,7 +11104,7 @@ export declare const MODELS: {
10624
11104
  contextWindow: number;
10625
11105
  maxTokens: number;
10626
11106
  };
10627
- readonly "qwen/qwen3.6-plus:free": {
11107
+ readonly "qwen/qwen3.6-plus": {
10628
11108
  id: string;
10629
11109
  name: string;
10630
11110
  api: "openai-completions";
@@ -10743,7 +11223,7 @@ export declare const MODELS: {
10743
11223
  contextWindow: number;
10744
11224
  maxTokens: number;
10745
11225
  };
10746
- readonly "stepfun/step-3.5-flash:free": {
11226
+ readonly "tencent/hy3-preview:free": {
10747
11227
  id: string;
10748
11228
  name: string;
10749
11229
  api: "openai-completions";
@@ -11032,6 +11512,40 @@ export declare const MODELS: {
11032
11512
  contextWindow: number;
11033
11513
  maxTokens: number;
11034
11514
  };
11515
+ readonly "xiaomi/mimo-v2.5": {
11516
+ id: string;
11517
+ name: string;
11518
+ api: "openai-completions";
11519
+ provider: string;
11520
+ baseUrl: string;
11521
+ reasoning: true;
11522
+ input: ("image" | "text")[];
11523
+ cost: {
11524
+ input: number;
11525
+ output: number;
11526
+ cacheRead: number;
11527
+ cacheWrite: number;
11528
+ };
11529
+ contextWindow: number;
11530
+ maxTokens: number;
11531
+ };
11532
+ readonly "xiaomi/mimo-v2.5-pro": {
11533
+ id: string;
11534
+ name: string;
11535
+ api: "openai-completions";
11536
+ provider: string;
11537
+ baseUrl: string;
11538
+ reasoning: true;
11539
+ input: "text"[];
11540
+ cost: {
11541
+ input: number;
11542
+ output: number;
11543
+ cacheRead: number;
11544
+ cacheWrite: number;
11545
+ };
11546
+ contextWindow: number;
11547
+ maxTokens: number;
11548
+ };
11035
11549
  readonly "z-ai/glm-4-32b": {
11036
11550
  id: string;
11037
11551
  name: string;
@@ -11219,6 +11733,23 @@ export declare const MODELS: {
11219
11733
  contextWindow: number;
11220
11734
  maxTokens: number;
11221
11735
  };
11736
+ readonly "z-ai/glm-5.1": {
11737
+ id: string;
11738
+ name: string;
11739
+ api: "openai-completions";
11740
+ provider: string;
11741
+ baseUrl: string;
11742
+ reasoning: true;
11743
+ input: "text"[];
11744
+ cost: {
11745
+ input: number;
11746
+ output: number;
11747
+ cacheRead: number;
11748
+ cacheWrite: number;
11749
+ };
11750
+ contextWindow: number;
11751
+ maxTokens: number;
11752
+ };
11222
11753
  readonly "z-ai/glm-5v-turbo": {
11223
11754
  id: string;
11224
11755
  name: string;
@@ -11236,6 +11767,23 @@ export declare const MODELS: {
11236
11767
  contextWindow: number;
11237
11768
  maxTokens: number;
11238
11769
  };
11770
+ readonly "~anthropic/claude-opus-latest": {
11771
+ id: string;
11772
+ name: string;
11773
+ api: "openai-completions";
11774
+ provider: string;
11775
+ baseUrl: string;
11776
+ reasoning: true;
11777
+ input: ("image" | "text")[];
11778
+ cost: {
11779
+ input: number;
11780
+ output: number;
11781
+ cacheRead: number;
11782
+ cacheWrite: number;
11783
+ };
11784
+ contextWindow: number;
11785
+ maxTokens: number;
11786
+ };
11239
11787
  };
11240
11788
  readonly "vercel-ai-gateway": {
11241
11789
  readonly "alibaba/qwen-3-14b": {
@@ -11306,6 +11854,23 @@ export declare const MODELS: {
11306
11854
  contextWindow: number;
11307
11855
  maxTokens: number;
11308
11856
  };
11857
+ readonly "alibaba/qwen-3.6-max-preview": {
11858
+ id: string;
11859
+ name: string;
11860
+ api: "anthropic-messages";
11861
+ provider: string;
11862
+ baseUrl: string;
11863
+ reasoning: true;
11864
+ input: ("image" | "text")[];
11865
+ cost: {
11866
+ input: number;
11867
+ output: number;
11868
+ cacheRead: number;
11869
+ cacheWrite: number;
11870
+ };
11871
+ contextWindow: number;
11872
+ maxTokens: number;
11873
+ };
11309
11874
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11310
11875
  id: string;
11311
11876
  name: string;
@@ -11646,6 +12211,23 @@ export declare const MODELS: {
11646
12211
  contextWindow: number;
11647
12212
  maxTokens: number;
11648
12213
  };
12214
+ readonly "anthropic/claude-opus-4.7": {
12215
+ id: string;
12216
+ name: string;
12217
+ api: "anthropic-messages";
12218
+ provider: string;
12219
+ baseUrl: string;
12220
+ reasoning: true;
12221
+ input: ("image" | "text")[];
12222
+ cost: {
12223
+ input: number;
12224
+ output: number;
12225
+ cacheRead: number;
12226
+ cacheWrite: number;
12227
+ };
12228
+ contextWindow: number;
12229
+ maxTokens: number;
12230
+ };
11649
12231
  readonly "anthropic/claude-sonnet-4": {
11650
12232
  id: string;
11651
12233
  name: string;
@@ -11805,7 +12387,41 @@ export declare const MODELS: {
11805
12387
  api: "anthropic-messages";
11806
12388
  provider: string;
11807
12389
  baseUrl: string;
11808
- reasoning: true;
12390
+ reasoning: true;
12391
+ input: "text"[];
12392
+ cost: {
12393
+ input: number;
12394
+ output: number;
12395
+ cacheRead: number;
12396
+ cacheWrite: number;
12397
+ };
12398
+ contextWindow: number;
12399
+ maxTokens: number;
12400
+ };
12401
+ readonly "deepseek/deepseek-v3.1-terminus": {
12402
+ id: string;
12403
+ name: string;
12404
+ api: "anthropic-messages";
12405
+ provider: string;
12406
+ baseUrl: string;
12407
+ reasoning: true;
12408
+ input: "text"[];
12409
+ cost: {
12410
+ input: number;
12411
+ output: number;
12412
+ cacheRead: number;
12413
+ cacheWrite: number;
12414
+ };
12415
+ contextWindow: number;
12416
+ maxTokens: number;
12417
+ };
12418
+ readonly "deepseek/deepseek-v3.2": {
12419
+ id: string;
12420
+ name: string;
12421
+ api: "anthropic-messages";
12422
+ provider: string;
12423
+ baseUrl: string;
12424
+ reasoning: false;
11809
12425
  input: "text"[];
11810
12426
  cost: {
11811
12427
  input: number;
@@ -11816,7 +12432,7 @@ export declare const MODELS: {
11816
12432
  contextWindow: number;
11817
12433
  maxTokens: number;
11818
12434
  };
11819
- readonly "deepseek/deepseek-v3.1-terminus": {
12435
+ readonly "deepseek/deepseek-v3.2-thinking": {
11820
12436
  id: string;
11821
12437
  name: string;
11822
12438
  api: "anthropic-messages";
@@ -11833,13 +12449,13 @@ export declare const MODELS: {
11833
12449
  contextWindow: number;
11834
12450
  maxTokens: number;
11835
12451
  };
11836
- readonly "deepseek/deepseek-v3.2": {
12452
+ readonly "deepseek/deepseek-v4-flash": {
11837
12453
  id: string;
11838
12454
  name: string;
11839
12455
  api: "anthropic-messages";
11840
12456
  provider: string;
11841
12457
  baseUrl: string;
11842
- reasoning: false;
12458
+ reasoning: true;
11843
12459
  input: "text"[];
11844
12460
  cost: {
11845
12461
  input: number;
@@ -11850,7 +12466,7 @@ export declare const MODELS: {
11850
12466
  contextWindow: number;
11851
12467
  maxTokens: number;
11852
12468
  };
11853
- readonly "deepseek/deepseek-v3.2-thinking": {
12469
+ readonly "deepseek/deepseek-v4-pro": {
11854
12470
  id: string;
11855
12471
  name: string;
11856
12472
  api: "anthropic-messages";
@@ -12632,6 +13248,23 @@ export declare const MODELS: {
12632
13248
  contextWindow: number;
12633
13249
  maxTokens: number;
12634
13250
  };
13251
+ readonly "moonshotai/kimi-k2.6": {
13252
+ id: string;
13253
+ name: string;
13254
+ api: "anthropic-messages";
13255
+ provider: string;
13256
+ baseUrl: string;
13257
+ reasoning: true;
13258
+ input: ("image" | "text")[];
13259
+ cost: {
13260
+ input: number;
13261
+ output: number;
13262
+ cacheRead: number;
13263
+ cacheWrite: number;
13264
+ };
13265
+ contextWindow: number;
13266
+ maxTokens: number;
13267
+ };
12635
13268
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12636
13269
  id: string;
12637
13270
  name: string;
@@ -13125,14 +13758,31 @@ export declare const MODELS: {
13125
13758
  contextWindow: number;
13126
13759
  maxTokens: number;
13127
13760
  };
13128
- readonly "openai/gpt-oss-120b": {
13761
+ readonly "openai/gpt-5.5": {
13129
13762
  id: string;
13130
13763
  name: string;
13131
13764
  api: "anthropic-messages";
13132
13765
  provider: string;
13133
13766
  baseUrl: string;
13134
13767
  reasoning: true;
13135
- input: "text"[];
13768
+ input: ("image" | "text")[];
13769
+ cost: {
13770
+ input: number;
13771
+ output: number;
13772
+ cacheRead: number;
13773
+ cacheWrite: number;
13774
+ };
13775
+ contextWindow: number;
13776
+ maxTokens: number;
13777
+ };
13778
+ readonly "openai/gpt-5.5-pro": {
13779
+ id: string;
13780
+ name: string;
13781
+ api: "anthropic-messages";
13782
+ provider: string;
13783
+ baseUrl: string;
13784
+ reasoning: true;
13785
+ input: ("image" | "text")[];
13136
13786
  cost: {
13137
13787
  input: number;
13138
13788
  output: number;
@@ -13329,23 +13979,6 @@ export declare const MODELS: {
13329
13979
  contextWindow: number;
13330
13980
  maxTokens: number;
13331
13981
  };
13332
- readonly "xai/grok-2-vision": {
13333
- id: string;
13334
- name: string;
13335
- api: "anthropic-messages";
13336
- provider: string;
13337
- baseUrl: string;
13338
- reasoning: false;
13339
- input: ("image" | "text")[];
13340
- cost: {
13341
- input: number;
13342
- output: number;
13343
- cacheRead: number;
13344
- cacheWrite: number;
13345
- };
13346
- contextWindow: number;
13347
- maxTokens: number;
13348
- };
13349
13982
  readonly "xai/grok-3": {
13350
13983
  id: string;
13351
13984
  name: string;
@@ -13438,7 +14071,7 @@ export declare const MODELS: {
13438
14071
  provider: string;
13439
14072
  baseUrl: string;
13440
14073
  reasoning: false;
13441
- input: "text"[];
14074
+ input: ("image" | "text")[];
13442
14075
  cost: {
13443
14076
  input: number;
13444
14077
  output: number;
@@ -13455,7 +14088,7 @@ export declare const MODELS: {
13455
14088
  provider: string;
13456
14089
  baseUrl: string;
13457
14090
  reasoning: true;
13458
- input: "text"[];
14091
+ input: ("image" | "text")[];
13459
14092
  cost: {
13460
14093
  input: number;
13461
14094
  output: number;
@@ -13472,7 +14105,7 @@ export declare const MODELS: {
13472
14105
  provider: string;
13473
14106
  baseUrl: string;
13474
14107
  reasoning: false;
13475
- input: "text"[];
14108
+ input: ("image" | "text")[];
13476
14109
  cost: {
13477
14110
  input: number;
13478
14111
  output: number;
@@ -13489,7 +14122,7 @@ export declare const MODELS: {
13489
14122
  provider: string;
13490
14123
  baseUrl: string;
13491
14124
  reasoning: true;
13492
- input: "text"[];
14125
+ input: ("image" | "text")[];
13493
14126
  cost: {
13494
14127
  input: number;
13495
14128
  output: number;
@@ -13506,7 +14139,7 @@ export declare const MODELS: {
13506
14139
  provider: string;
13507
14140
  baseUrl: string;
13508
14141
  reasoning: true;
13509
- input: "text"[];
14142
+ input: ("image" | "text")[];
13510
14143
  cost: {
13511
14144
  input: number;
13512
14145
  output: number;
@@ -13523,7 +14156,7 @@ export declare const MODELS: {
13523
14156
  provider: string;
13524
14157
  baseUrl: string;
13525
14158
  reasoning: true;
13526
- input: "text"[];
14159
+ input: ("image" | "text")[];
13527
14160
  cost: {
13528
14161
  input: number;
13529
14162
  output: number;
@@ -13839,6 +14472,23 @@ export declare const MODELS: {
13839
14472
  contextWindow: number;
13840
14473
  maxTokens: number;
13841
14474
  };
14475
+ readonly "zai/glm-5.1": {
14476
+ id: string;
14477
+ name: string;
14478
+ api: "anthropic-messages";
14479
+ provider: string;
14480
+ baseUrl: string;
14481
+ reasoning: true;
14482
+ input: "text"[];
14483
+ cost: {
14484
+ input: number;
14485
+ output: number;
14486
+ cacheRead: number;
14487
+ cacheWrite: number;
14488
+ };
14489
+ contextWindow: number;
14490
+ maxTokens: number;
14491
+ };
13842
14492
  readonly "zai/glm-5v-turbo": {
13843
14493
  id: string;
13844
14494
  name: string;
@@ -14268,27 +14918,6 @@ export declare const MODELS: {
14268
14918
  };
14269
14919
  };
14270
14920
  readonly zai: {
14271
- readonly "glm-4.5": {
14272
- id: string;
14273
- name: string;
14274
- api: "openai-completions";
14275
- provider: string;
14276
- baseUrl: string;
14277
- compat: {
14278
- supportsDeveloperRole: false;
14279
- thinkingFormat: "zai";
14280
- };
14281
- reasoning: true;
14282
- input: "text"[];
14283
- cost: {
14284
- input: number;
14285
- output: number;
14286
- cacheRead: number;
14287
- cacheWrite: number;
14288
- };
14289
- contextWindow: number;
14290
- maxTokens: number;
14291
- };
14292
14921
  readonly "glm-4.5-air": {
14293
14922
  id: string;
14294
14923
  name: string;
@@ -14310,92 +14939,6 @@ export declare const MODELS: {
14310
14939
  contextWindow: number;
14311
14940
  maxTokens: number;
14312
14941
  };
14313
- readonly "glm-4.5-flash": {
14314
- id: string;
14315
- name: string;
14316
- api: "openai-completions";
14317
- provider: string;
14318
- baseUrl: string;
14319
- compat: {
14320
- supportsDeveloperRole: false;
14321
- thinkingFormat: "zai";
14322
- };
14323
- reasoning: true;
14324
- input: "text"[];
14325
- cost: {
14326
- input: number;
14327
- output: number;
14328
- cacheRead: number;
14329
- cacheWrite: number;
14330
- };
14331
- contextWindow: number;
14332
- maxTokens: number;
14333
- };
14334
- readonly "glm-4.5v": {
14335
- id: string;
14336
- name: string;
14337
- api: "openai-completions";
14338
- provider: string;
14339
- baseUrl: string;
14340
- compat: {
14341
- supportsDeveloperRole: false;
14342
- thinkingFormat: "zai";
14343
- };
14344
- reasoning: true;
14345
- input: ("image" | "text")[];
14346
- cost: {
14347
- input: number;
14348
- output: number;
14349
- cacheRead: number;
14350
- cacheWrite: number;
14351
- };
14352
- contextWindow: number;
14353
- maxTokens: number;
14354
- };
14355
- readonly "glm-4.6": {
14356
- id: string;
14357
- name: string;
14358
- api: "openai-completions";
14359
- provider: string;
14360
- baseUrl: string;
14361
- compat: {
14362
- supportsDeveloperRole: false;
14363
- thinkingFormat: "zai";
14364
- zaiToolStream: true;
14365
- };
14366
- reasoning: true;
14367
- input: "text"[];
14368
- cost: {
14369
- input: number;
14370
- output: number;
14371
- cacheRead: number;
14372
- cacheWrite: number;
14373
- };
14374
- contextWindow: number;
14375
- maxTokens: number;
14376
- };
14377
- readonly "glm-4.6v": {
14378
- id: string;
14379
- name: string;
14380
- api: "openai-completions";
14381
- provider: string;
14382
- baseUrl: string;
14383
- compat: {
14384
- supportsDeveloperRole: false;
14385
- thinkingFormat: "zai";
14386
- zaiToolStream: true;
14387
- };
14388
- reasoning: true;
14389
- input: ("image" | "text")[];
14390
- cost: {
14391
- input: number;
14392
- output: number;
14393
- cacheRead: number;
14394
- cacheWrite: number;
14395
- };
14396
- contextWindow: number;
14397
- maxTokens: number;
14398
- };
14399
14942
  readonly "glm-4.7": {
14400
14943
  id: string;
14401
14944
  name: string;
@@ -14418,72 +14961,6 @@ export declare const MODELS: {
14418
14961
  contextWindow: number;
14419
14962
  maxTokens: number;
14420
14963
  };
14421
- readonly "glm-4.7-flash": {
14422
- id: string;
14423
- name: string;
14424
- api: "openai-completions";
14425
- provider: string;
14426
- baseUrl: string;
14427
- compat: {
14428
- supportsDeveloperRole: false;
14429
- thinkingFormat: "zai";
14430
- zaiToolStream: true;
14431
- };
14432
- reasoning: true;
14433
- input: "text"[];
14434
- cost: {
14435
- input: number;
14436
- output: number;
14437
- cacheRead: number;
14438
- cacheWrite: number;
14439
- };
14440
- contextWindow: number;
14441
- maxTokens: number;
14442
- };
14443
- readonly "glm-4.7-flashx": {
14444
- id: string;
14445
- name: string;
14446
- api: "openai-completions";
14447
- provider: string;
14448
- baseUrl: string;
14449
- compat: {
14450
- supportsDeveloperRole: false;
14451
- thinkingFormat: "zai";
14452
- zaiToolStream: true;
14453
- };
14454
- reasoning: true;
14455
- input: "text"[];
14456
- cost: {
14457
- input: number;
14458
- output: number;
14459
- cacheRead: number;
14460
- cacheWrite: number;
14461
- };
14462
- contextWindow: number;
14463
- maxTokens: number;
14464
- };
14465
- readonly "glm-5": {
14466
- id: string;
14467
- name: string;
14468
- api: "openai-completions";
14469
- provider: string;
14470
- baseUrl: string;
14471
- compat: {
14472
- supportsDeveloperRole: false;
14473
- thinkingFormat: "zai";
14474
- zaiToolStream: true;
14475
- };
14476
- reasoning: true;
14477
- input: "text"[];
14478
- cost: {
14479
- input: number;
14480
- output: number;
14481
- cacheRead: number;
14482
- cacheWrite: number;
14483
- };
14484
- contextWindow: number;
14485
- maxTokens: number;
14486
- };
14487
14964
  readonly "glm-5-turbo": {
14488
14965
  id: string;
14489
14966
  name: string;
@@ -14506,7 +14983,7 @@ export declare const MODELS: {
14506
14983
  contextWindow: number;
14507
14984
  maxTokens: number;
14508
14985
  };
14509
- readonly "glm-5v-turbo": {
14986
+ readonly "glm-5.1": {
14510
14987
  id: string;
14511
14988
  name: string;
14512
14989
  api: "openai-completions";
@@ -14518,7 +14995,7 @@ export declare const MODELS: {
14518
14995
  zaiToolStream: true;
14519
14996
  };
14520
14997
  reasoning: true;
14521
- input: ("image" | "text")[];
14998
+ input: "text"[];
14522
14999
  cost: {
14523
15000
  input: number;
14524
15001
  output: number;