@draht/ai 2026.4.5 → 2026.4.23

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (40) hide show
  1. package/README.md +21 -12
  2. package/dist/cli.d.ts.map +1 -1
  3. package/dist/cli.js +1 -1
  4. package/dist/cli.js.map +1 -1
  5. package/dist/models.generated.d.ts +683 -357
  6. package/dist/models.generated.d.ts.map +1 -1
  7. package/dist/models.generated.js +897 -586
  8. package/dist/models.generated.js.map +1 -1
  9. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  10. package/dist/providers/amazon-bedrock.js +37 -9
  11. package/dist/providers/amazon-bedrock.js.map +1 -1
  12. package/dist/providers/anthropic.d.ts.map +1 -1
  13. package/dist/providers/anthropic.js +38 -16
  14. package/dist/providers/anthropic.js.map +1 -1
  15. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  16. package/dist/providers/google-gemini-cli.js +2 -2
  17. package/dist/providers/google-gemini-cli.js.map +1 -1
  18. package/dist/providers/google-vertex.d.ts.map +1 -1
  19. package/dist/providers/google-vertex.js +1 -1
  20. package/dist/providers/google-vertex.js.map +1 -1
  21. package/dist/providers/google.d.ts.map +1 -1
  22. package/dist/providers/google.js +29 -4
  23. package/dist/providers/google.js.map +1 -1
  24. package/dist/providers/openai-completions.d.ts.map +1 -1
  25. package/dist/providers/openai-completions.js +13 -6
  26. package/dist/providers/openai-completions.js.map +1 -1
  27. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  28. package/dist/providers/openai-responses-shared.js +8 -1
  29. package/dist/providers/openai-responses-shared.js.map +1 -1
  30. package/dist/providers/openai-responses.d.ts.map +1 -1
  31. package/dist/providers/openai-responses.js +2 -2
  32. package/dist/providers/openai-responses.js.map +1 -1
  33. package/dist/types.d.ts +62 -5
  34. package/dist/types.d.ts.map +1 -1
  35. package/dist/types.js.map +1 -1
  36. package/dist/utils/overflow.d.ts +3 -2
  37. package/dist/utils/overflow.d.ts.map +1 -1
  38. package/dist/utils/overflow.js +24 -11
  39. package/dist/utils/overflow.js.map +1 -1
  40. package/package.json +5 -8
@@ -255,6 +255,23 @@ export declare const MODELS: {
255
255
  contextWindow: number;
256
256
  maxTokens: number;
257
257
  };
258
+ readonly "anthropic.claude-opus-4-7": {
259
+ id: string;
260
+ name: string;
261
+ api: "bedrock-converse-stream";
262
+ provider: string;
263
+ baseUrl: string;
264
+ reasoning: true;
265
+ input: ("image" | "text")[];
266
+ cost: {
267
+ input: number;
268
+ output: number;
269
+ cacheRead: number;
270
+ cacheWrite: number;
271
+ };
272
+ contextWindow: number;
273
+ maxTokens: number;
274
+ };
258
275
  readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
259
276
  id: string;
260
277
  name: string;
@@ -408,6 +425,23 @@ export declare const MODELS: {
408
425
  contextWindow: number;
409
426
  maxTokens: number;
410
427
  };
428
+ readonly "eu.anthropic.claude-opus-4-7": {
429
+ id: string;
430
+ name: string;
431
+ api: "bedrock-converse-stream";
432
+ provider: string;
433
+ baseUrl: string;
434
+ reasoning: true;
435
+ input: ("image" | "text")[];
436
+ cost: {
437
+ input: number;
438
+ output: number;
439
+ cacheRead: number;
440
+ cacheWrite: number;
441
+ };
442
+ contextWindow: number;
443
+ maxTokens: number;
444
+ };
411
445
  readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
412
446
  id: string;
413
447
  name: string;
@@ -510,6 +544,23 @@ export declare const MODELS: {
510
544
  contextWindow: number;
511
545
  maxTokens: number;
512
546
  };
547
+ readonly "global.anthropic.claude-opus-4-7": {
548
+ id: string;
549
+ name: string;
550
+ api: "bedrock-converse-stream";
551
+ provider: string;
552
+ baseUrl: string;
553
+ reasoning: true;
554
+ input: ("image" | "text")[];
555
+ cost: {
556
+ input: number;
557
+ output: number;
558
+ cacheRead: number;
559
+ cacheWrite: number;
560
+ };
561
+ contextWindow: number;
562
+ maxTokens: number;
563
+ };
513
564
  readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
514
565
  id: string;
515
566
  name: string;
@@ -1207,6 +1258,23 @@ export declare const MODELS: {
1207
1258
  contextWindow: number;
1208
1259
  maxTokens: number;
1209
1260
  };
1261
+ readonly "qwen.qwen3-coder-next": {
1262
+ id: string;
1263
+ name: string;
1264
+ api: "bedrock-converse-stream";
1265
+ provider: string;
1266
+ baseUrl: string;
1267
+ reasoning: true;
1268
+ input: "text"[];
1269
+ cost: {
1270
+ input: number;
1271
+ output: number;
1272
+ cacheRead: number;
1273
+ cacheWrite: number;
1274
+ };
1275
+ contextWindow: number;
1276
+ maxTokens: number;
1277
+ };
1210
1278
  readonly "qwen.qwen3-next-80b-a3b": {
1211
1279
  id: string;
1212
1280
  name: string;
@@ -1326,6 +1394,23 @@ export declare const MODELS: {
1326
1394
  contextWindow: number;
1327
1395
  maxTokens: number;
1328
1396
  };
1397
+ readonly "us.anthropic.claude-opus-4-7": {
1398
+ id: string;
1399
+ name: string;
1400
+ api: "bedrock-converse-stream";
1401
+ provider: string;
1402
+ baseUrl: string;
1403
+ reasoning: true;
1404
+ input: ("image" | "text")[];
1405
+ cost: {
1406
+ input: number;
1407
+ output: number;
1408
+ cacheRead: number;
1409
+ cacheWrite: number;
1410
+ };
1411
+ contextWindow: number;
1412
+ maxTokens: number;
1413
+ };
1329
1414
  readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
1330
1415
  id: string;
1331
1416
  name: string;
@@ -1753,7 +1838,7 @@ export declare const MODELS: {
1753
1838
  contextWindow: number;
1754
1839
  maxTokens: number;
1755
1840
  };
1756
- readonly "claude-sonnet-4-0": {
1841
+ readonly "claude-opus-4-7": {
1757
1842
  id: string;
1758
1843
  name: string;
1759
1844
  api: "anthropic-messages";
@@ -1770,7 +1855,7 @@ export declare const MODELS: {
1770
1855
  contextWindow: number;
1771
1856
  maxTokens: number;
1772
1857
  };
1773
- readonly "claude-sonnet-4-20250514": {
1858
+ readonly "claude-sonnet-4-0": {
1774
1859
  id: string;
1775
1860
  name: string;
1776
1861
  api: "anthropic-messages";
@@ -1787,7 +1872,7 @@ export declare const MODELS: {
1787
1872
  contextWindow: number;
1788
1873
  maxTokens: number;
1789
1874
  };
1790
- readonly "claude-sonnet-4-5": {
1875
+ readonly "claude-sonnet-4-20250514": {
1791
1876
  id: string;
1792
1877
  name: string;
1793
1878
  api: "anthropic-messages";
@@ -1804,7 +1889,7 @@ export declare const MODELS: {
1804
1889
  contextWindow: number;
1805
1890
  maxTokens: number;
1806
1891
  };
1807
- readonly "claude-sonnet-4-5-20250929": {
1892
+ readonly "claude-sonnet-4-5": {
1808
1893
  id: string;
1809
1894
  name: string;
1810
1895
  api: "anthropic-messages";
@@ -1821,7 +1906,7 @@ export declare const MODELS: {
1821
1906
  contextWindow: number;
1822
1907
  maxTokens: number;
1823
1908
  };
1824
- readonly "claude-sonnet-4-6": {
1909
+ readonly "claude-sonnet-4-5-20250929": {
1825
1910
  id: string;
1826
1911
  name: string;
1827
1912
  api: "anthropic-messages";
@@ -1838,16 +1923,14 @@ export declare const MODELS: {
1838
1923
  contextWindow: number;
1839
1924
  maxTokens: number;
1840
1925
  };
1841
- };
1842
- readonly "azure-openai-responses": {
1843
- readonly "codex-mini-latest": {
1926
+ readonly "claude-sonnet-4-6": {
1844
1927
  id: string;
1845
1928
  name: string;
1846
- api: "azure-openai-responses";
1929
+ api: "anthropic-messages";
1847
1930
  provider: string;
1848
1931
  baseUrl: string;
1849
1932
  reasoning: true;
1850
- input: "text"[];
1933
+ input: ("image" | "text")[];
1851
1934
  cost: {
1852
1935
  input: number;
1853
1936
  output: number;
@@ -1857,6 +1940,8 @@ export declare const MODELS: {
1857
1940
  contextWindow: number;
1858
1941
  maxTokens: number;
1859
1942
  };
1943
+ };
1944
+ readonly "azure-openai-responses": {
1860
1945
  readonly "gpt-4": {
1861
1946
  id: string;
1862
1947
  name: string;
@@ -2678,6 +2763,29 @@ export declare const MODELS: {
2678
2763
  contextWindow: number;
2679
2764
  maxTokens: number;
2680
2765
  };
2766
+ readonly "claude-opus-4.7": {
2767
+ id: string;
2768
+ name: string;
2769
+ api: "anthropic-messages";
2770
+ provider: string;
2771
+ baseUrl: string;
2772
+ headers: {
2773
+ "User-Agent": string;
2774
+ "Editor-Version": string;
2775
+ "Editor-Plugin-Version": string;
2776
+ "Copilot-Integration-Id": string;
2777
+ };
2778
+ reasoning: true;
2779
+ input: ("image" | "text")[];
2780
+ cost: {
2781
+ input: number;
2782
+ output: number;
2783
+ cacheRead: number;
2784
+ cacheWrite: number;
2785
+ };
2786
+ contextWindow: number;
2787
+ maxTokens: number;
2788
+ };
2681
2789
  readonly "claude-sonnet-4": {
2682
2790
  id: string;
2683
2791
  name: string;
@@ -3623,7 +3731,7 @@ export declare const MODELS: {
3623
3731
  contextWindow: number;
3624
3732
  maxTokens: number;
3625
3733
  };
3626
- readonly "gemma-4-26b": {
3734
+ readonly "gemma-4-26b-a4b-it": {
3627
3735
  id: string;
3628
3736
  name: string;
3629
3737
  api: "google-generative-ai";
@@ -3640,7 +3748,7 @@ export declare const MODELS: {
3640
3748
  contextWindow: number;
3641
3749
  maxTokens: number;
3642
3750
  };
3643
- readonly "gemma-4-31b": {
3751
+ readonly "gemma-4-31b-it": {
3644
3752
  id: string;
3645
3753
  name: string;
3646
3754
  api: "google-generative-ai";
@@ -4122,6 +4230,23 @@ export declare const MODELS: {
4122
4230
  contextWindow: number;
4123
4231
  maxTokens: number;
4124
4232
  };
4233
+ readonly "gemini-3.1-pro-preview-customtools": {
4234
+ id: string;
4235
+ name: string;
4236
+ api: "google-vertex";
4237
+ provider: string;
4238
+ baseUrl: string;
4239
+ reasoning: true;
4240
+ input: ("image" | "text")[];
4241
+ cost: {
4242
+ input: number;
4243
+ output: number;
4244
+ cacheRead: number;
4245
+ cacheWrite: number;
4246
+ };
4247
+ contextWindow: number;
4248
+ maxTokens: number;
4249
+ };
4125
4250
  };
4126
4251
  readonly groq: {
4127
4252
  readonly "deepseek-r1-distill-llama-70b": {
@@ -4472,6 +4597,26 @@ export declare const MODELS: {
4472
4597
  contextWindow: number;
4473
4598
  maxTokens: number;
4474
4599
  };
4600
+ readonly "MiniMaxAI/MiniMax-M2.7": {
4601
+ id: string;
4602
+ name: string;
4603
+ api: "openai-completions";
4604
+ provider: string;
4605
+ baseUrl: string;
4606
+ compat: {
4607
+ supportsDeveloperRole: false;
4608
+ };
4609
+ reasoning: true;
4610
+ input: "text"[];
4611
+ cost: {
4612
+ input: number;
4613
+ output: number;
4614
+ cacheRead: number;
4615
+ cacheWrite: number;
4616
+ };
4617
+ contextWindow: number;
4618
+ maxTokens: number;
4619
+ };
4475
4620
  readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4476
4621
  id: string;
4477
4622
  name: string;
@@ -4732,7 +4877,7 @@ export declare const MODELS: {
4732
4877
  contextWindow: number;
4733
4878
  maxTokens: number;
4734
4879
  };
4735
- readonly "zai-org/GLM-4.7": {
4880
+ readonly "moonshotai/Kimi-K2.6": {
4736
4881
  id: string;
4737
4882
  name: string;
4738
4883
  api: "openai-completions";
@@ -4742,7 +4887,7 @@ export declare const MODELS: {
4742
4887
  supportsDeveloperRole: false;
4743
4888
  };
4744
4889
  reasoning: true;
4745
- input: "text"[];
4890
+ input: ("image" | "text")[];
4746
4891
  cost: {
4747
4892
  input: number;
4748
4893
  output: number;
@@ -4752,7 +4897,7 @@ export declare const MODELS: {
4752
4897
  contextWindow: number;
4753
4898
  maxTokens: number;
4754
4899
  };
4755
- readonly "zai-org/GLM-4.7-Flash": {
4900
+ readonly "zai-org/GLM-4.7": {
4756
4901
  id: string;
4757
4902
  name: string;
4758
4903
  api: "openai-completions";
@@ -4772,7 +4917,7 @@ export declare const MODELS: {
4772
4917
  contextWindow: number;
4773
4918
  maxTokens: number;
4774
4919
  };
4775
- readonly "zai-org/GLM-5": {
4920
+ readonly "zai-org/GLM-4.7-Flash": {
4776
4921
  id: string;
4777
4922
  name: string;
4778
4923
  api: "openai-completions";
@@ -4792,16 +4937,17 @@ export declare const MODELS: {
4792
4937
  contextWindow: number;
4793
4938
  maxTokens: number;
4794
4939
  };
4795
- };
4796
- readonly "kimi-coding": {
4797
- readonly k2p5: {
4940
+ readonly "zai-org/GLM-5": {
4798
4941
  id: string;
4799
4942
  name: string;
4800
- api: "anthropic-messages";
4943
+ api: "openai-completions";
4801
4944
  provider: string;
4802
4945
  baseUrl: string;
4946
+ compat: {
4947
+ supportsDeveloperRole: false;
4948
+ };
4803
4949
  reasoning: true;
4804
- input: ("image" | "text")[];
4950
+ input: "text"[];
4805
4951
  cost: {
4806
4952
  input: number;
4807
4953
  output: number;
@@ -4811,12 +4957,15 @@ export declare const MODELS: {
4811
4957
  contextWindow: number;
4812
4958
  maxTokens: number;
4813
4959
  };
4814
- readonly "kimi-k2-thinking": {
4960
+ readonly "zai-org/GLM-5.1": {
4815
4961
  id: string;
4816
4962
  name: string;
4817
- api: "anthropic-messages";
4963
+ api: "openai-completions";
4818
4964
  provider: string;
4819
4965
  baseUrl: string;
4966
+ compat: {
4967
+ supportsDeveloperRole: false;
4968
+ };
4820
4969
  reasoning: true;
4821
4970
  input: "text"[];
4822
4971
  cost: {
@@ -4829,15 +4978,15 @@ export declare const MODELS: {
4829
4978
  maxTokens: number;
4830
4979
  };
4831
4980
  };
4832
- readonly minimax: {
4833
- readonly "MiniMax-M2": {
4981
+ readonly "kimi-coding": {
4982
+ readonly k2p5: {
4834
4983
  id: string;
4835
4984
  name: string;
4836
4985
  api: "anthropic-messages";
4837
4986
  provider: string;
4838
4987
  baseUrl: string;
4839
4988
  reasoning: true;
4840
- input: "text"[];
4989
+ input: ("image" | "text")[];
4841
4990
  cost: {
4842
4991
  input: number;
4843
4992
  output: number;
@@ -4847,14 +4996,14 @@ export declare const MODELS: {
4847
4996
  contextWindow: number;
4848
4997
  maxTokens: number;
4849
4998
  };
4850
- readonly "MiniMax-M2.1": {
4999
+ readonly k2p6: {
4851
5000
  id: string;
4852
5001
  name: string;
4853
5002
  api: "anthropic-messages";
4854
5003
  provider: string;
4855
5004
  baseUrl: string;
4856
5005
  reasoning: true;
4857
- input: "text"[];
5006
+ input: ("image" | "text")[];
4858
5007
  cost: {
4859
5008
  input: number;
4860
5009
  output: number;
@@ -4864,7 +5013,7 @@ export declare const MODELS: {
4864
5013
  contextWindow: number;
4865
5014
  maxTokens: number;
4866
5015
  };
4867
- readonly "MiniMax-M2.1-highspeed": {
5016
+ readonly "kimi-k2-thinking": {
4868
5017
  id: string;
4869
5018
  name: string;
4870
5019
  api: "anthropic-messages";
@@ -4881,7 +5030,9 @@ export declare const MODELS: {
4881
5030
  contextWindow: number;
4882
5031
  maxTokens: number;
4883
5032
  };
4884
- readonly "MiniMax-M2.5": {
5033
+ };
5034
+ readonly minimax: {
5035
+ readonly "MiniMax-M2.7": {
4885
5036
  id: string;
4886
5037
  name: string;
4887
5038
  api: "anthropic-messages";
@@ -4898,7 +5049,7 @@ export declare const MODELS: {
4898
5049
  contextWindow: number;
4899
5050
  maxTokens: number;
4900
5051
  };
4901
- readonly "MiniMax-M2.5-highspeed": {
5052
+ readonly "MiniMax-M2.7-highspeed": {
4902
5053
  id: string;
4903
5054
  name: string;
4904
5055
  api: "anthropic-messages";
@@ -4915,6 +5066,8 @@ export declare const MODELS: {
4915
5066
  contextWindow: number;
4916
5067
  maxTokens: number;
4917
5068
  };
5069
+ };
5070
+ readonly "minimax-cn": {
4918
5071
  readonly "MiniMax-M2.7": {
4919
5072
  id: string;
4920
5073
  name: string;
@@ -4950,14 +5103,14 @@ export declare const MODELS: {
4950
5103
  maxTokens: number;
4951
5104
  };
4952
5105
  };
4953
- readonly "minimax-cn": {
4954
- readonly "MiniMax-M2": {
5106
+ readonly mistral: {
5107
+ readonly "codestral-latest": {
4955
5108
  id: string;
4956
5109
  name: string;
4957
- api: "anthropic-messages";
5110
+ api: "openai-completions";
4958
5111
  provider: string;
4959
5112
  baseUrl: string;
4960
- reasoning: true;
5113
+ reasoning: false;
4961
5114
  input: "text"[];
4962
5115
  cost: {
4963
5116
  input: number;
@@ -4968,13 +5121,13 @@ export declare const MODELS: {
4968
5121
  contextWindow: number;
4969
5122
  maxTokens: number;
4970
5123
  };
4971
- readonly "MiniMax-M2.1": {
5124
+ readonly "devstral-2512": {
4972
5125
  id: string;
4973
5126
  name: string;
4974
- api: "anthropic-messages";
5127
+ api: "openai-completions";
4975
5128
  provider: string;
4976
5129
  baseUrl: string;
4977
- reasoning: true;
5130
+ reasoning: false;
4978
5131
  input: "text"[];
4979
5132
  cost: {
4980
5133
  input: number;
@@ -4985,13 +5138,13 @@ export declare const MODELS: {
4985
5138
  contextWindow: number;
4986
5139
  maxTokens: number;
4987
5140
  };
4988
- readonly "MiniMax-M2.1-highspeed": {
5141
+ readonly "devstral-medium-2507": {
4989
5142
  id: string;
4990
5143
  name: string;
4991
- api: "anthropic-messages";
5144
+ api: "openai-completions";
4992
5145
  provider: string;
4993
5146
  baseUrl: string;
4994
- reasoning: true;
5147
+ reasoning: false;
4995
5148
  input: "text"[];
4996
5149
  cost: {
4997
5150
  input: number;
@@ -5002,13 +5155,13 @@ export declare const MODELS: {
5002
5155
  contextWindow: number;
5003
5156
  maxTokens: number;
5004
5157
  };
5005
- readonly "MiniMax-M2.5": {
5158
+ readonly "devstral-medium-latest": {
5006
5159
  id: string;
5007
5160
  name: string;
5008
- api: "anthropic-messages";
5161
+ api: "openai-completions";
5009
5162
  provider: string;
5010
5163
  baseUrl: string;
5011
- reasoning: true;
5164
+ reasoning: false;
5012
5165
  input: "text"[];
5013
5166
  cost: {
5014
5167
  input: number;
@@ -5019,13 +5172,13 @@ export declare const MODELS: {
5019
5172
  contextWindow: number;
5020
5173
  maxTokens: number;
5021
5174
  };
5022
- readonly "MiniMax-M2.5-highspeed": {
5175
+ readonly "devstral-small-2505": {
5023
5176
  id: string;
5024
5177
  name: string;
5025
- api: "anthropic-messages";
5178
+ api: "openai-completions";
5026
5179
  provider: string;
5027
5180
  baseUrl: string;
5028
- reasoning: true;
5181
+ reasoning: false;
5029
5182
  input: "text"[];
5030
5183
  cost: {
5031
5184
  input: number;
@@ -5036,13 +5189,13 @@ export declare const MODELS: {
5036
5189
  contextWindow: number;
5037
5190
  maxTokens: number;
5038
5191
  };
5039
- readonly "MiniMax-M2.7": {
5192
+ readonly "devstral-small-2507": {
5040
5193
  id: string;
5041
5194
  name: string;
5042
- api: "anthropic-messages";
5195
+ api: "openai-completions";
5043
5196
  provider: string;
5044
5197
  baseUrl: string;
5045
- reasoning: true;
5198
+ reasoning: false;
5046
5199
  input: "text"[];
5047
5200
  cost: {
5048
5201
  input: number;
@@ -5053,14 +5206,14 @@ export declare const MODELS: {
5053
5206
  contextWindow: number;
5054
5207
  maxTokens: number;
5055
5208
  };
5056
- readonly "MiniMax-M2.7-highspeed": {
5209
+ readonly "labs-devstral-small-2512": {
5057
5210
  id: string;
5058
5211
  name: string;
5059
- api: "anthropic-messages";
5212
+ api: "openai-completions";
5060
5213
  provider: string;
5061
5214
  baseUrl: string;
5062
- reasoning: true;
5063
- input: "text"[];
5215
+ reasoning: false;
5216
+ input: ("image" | "text")[];
5064
5217
  cost: {
5065
5218
  input: number;
5066
5219
  output: number;
@@ -5070,15 +5223,13 @@ export declare const MODELS: {
5070
5223
  contextWindow: number;
5071
5224
  maxTokens: number;
5072
5225
  };
5073
- };
5074
- readonly mistral: {
5075
- readonly "codestral-latest": {
5226
+ readonly "magistral-medium-latest": {
5076
5227
  id: string;
5077
5228
  name: string;
5078
- api: "mistral-conversations";
5229
+ api: "openai-completions";
5079
5230
  provider: string;
5080
5231
  baseUrl: string;
5081
- reasoning: false;
5232
+ reasoning: true;
5082
5233
  input: "text"[];
5083
5234
  cost: {
5084
5235
  input: number;
@@ -5089,13 +5240,13 @@ export declare const MODELS: {
5089
5240
  contextWindow: number;
5090
5241
  maxTokens: number;
5091
5242
  };
5092
- readonly "devstral-2512": {
5243
+ readonly "magistral-small": {
5093
5244
  id: string;
5094
5245
  name: string;
5095
- api: "mistral-conversations";
5246
+ api: "openai-completions";
5096
5247
  provider: string;
5097
5248
  baseUrl: string;
5098
- reasoning: false;
5249
+ reasoning: true;
5099
5250
  input: "text"[];
5100
5251
  cost: {
5101
5252
  input: number;
@@ -5106,10 +5257,10 @@ export declare const MODELS: {
5106
5257
  contextWindow: number;
5107
5258
  maxTokens: number;
5108
5259
  };
5109
- readonly "devstral-medium-2507": {
5260
+ readonly "ministral-3b-latest": {
5110
5261
  id: string;
5111
5262
  name: string;
5112
- api: "mistral-conversations";
5263
+ api: "openai-completions";
5113
5264
  provider: string;
5114
5265
  baseUrl: string;
5115
5266
  reasoning: false;
@@ -5123,10 +5274,10 @@ export declare const MODELS: {
5123
5274
  contextWindow: number;
5124
5275
  maxTokens: number;
5125
5276
  };
5126
- readonly "devstral-medium-latest": {
5277
+ readonly "ministral-8b-latest": {
5127
5278
  id: string;
5128
5279
  name: string;
5129
- api: "mistral-conversations";
5280
+ api: "openai-completions";
5130
5281
  provider: string;
5131
5282
  baseUrl: string;
5132
5283
  reasoning: false;
@@ -5140,129 +5291,10 @@ export declare const MODELS: {
5140
5291
  contextWindow: number;
5141
5292
  maxTokens: number;
5142
5293
  };
5143
- readonly "devstral-small-2505": {
5294
+ readonly "mistral-large-2411": {
5144
5295
  id: string;
5145
5296
  name: string;
5146
- api: "mistral-conversations";
5147
- provider: string;
5148
- baseUrl: string;
5149
- reasoning: false;
5150
- input: "text"[];
5151
- cost: {
5152
- input: number;
5153
- output: number;
5154
- cacheRead: number;
5155
- cacheWrite: number;
5156
- };
5157
- contextWindow: number;
5158
- maxTokens: number;
5159
- };
5160
- readonly "devstral-small-2507": {
5161
- id: string;
5162
- name: string;
5163
- api: "mistral-conversations";
5164
- provider: string;
5165
- baseUrl: string;
5166
- reasoning: false;
5167
- input: "text"[];
5168
- cost: {
5169
- input: number;
5170
- output: number;
5171
- cacheRead: number;
5172
- cacheWrite: number;
5173
- };
5174
- contextWindow: number;
5175
- maxTokens: number;
5176
- };
5177
- readonly "labs-devstral-small-2512": {
5178
- id: string;
5179
- name: string;
5180
- api: "mistral-conversations";
5181
- provider: string;
5182
- baseUrl: string;
5183
- reasoning: false;
5184
- input: ("image" | "text")[];
5185
- cost: {
5186
- input: number;
5187
- output: number;
5188
- cacheRead: number;
5189
- cacheWrite: number;
5190
- };
5191
- contextWindow: number;
5192
- maxTokens: number;
5193
- };
5194
- readonly "magistral-medium-latest": {
5195
- id: string;
5196
- name: string;
5197
- api: "mistral-conversations";
5198
- provider: string;
5199
- baseUrl: string;
5200
- reasoning: true;
5201
- input: "text"[];
5202
- cost: {
5203
- input: number;
5204
- output: number;
5205
- cacheRead: number;
5206
- cacheWrite: number;
5207
- };
5208
- contextWindow: number;
5209
- maxTokens: number;
5210
- };
5211
- readonly "magistral-small": {
5212
- id: string;
5213
- name: string;
5214
- api: "mistral-conversations";
5215
- provider: string;
5216
- baseUrl: string;
5217
- reasoning: true;
5218
- input: "text"[];
5219
- cost: {
5220
- input: number;
5221
- output: number;
5222
- cacheRead: number;
5223
- cacheWrite: number;
5224
- };
5225
- contextWindow: number;
5226
- maxTokens: number;
5227
- };
5228
- readonly "ministral-3b-latest": {
5229
- id: string;
5230
- name: string;
5231
- api: "mistral-conversations";
5232
- provider: string;
5233
- baseUrl: string;
5234
- reasoning: false;
5235
- input: "text"[];
5236
- cost: {
5237
- input: number;
5238
- output: number;
5239
- cacheRead: number;
5240
- cacheWrite: number;
5241
- };
5242
- contextWindow: number;
5243
- maxTokens: number;
5244
- };
5245
- readonly "ministral-8b-latest": {
5246
- id: string;
5247
- name: string;
5248
- api: "mistral-conversations";
5249
- provider: string;
5250
- baseUrl: string;
5251
- reasoning: false;
5252
- input: "text"[];
5253
- cost: {
5254
- input: number;
5255
- output: number;
5256
- cacheRead: number;
5257
- cacheWrite: number;
5258
- };
5259
- contextWindow: number;
5260
- maxTokens: number;
5261
- };
5262
- readonly "mistral-large-2411": {
5263
- id: string;
5264
- name: string;
5265
- api: "mistral-conversations";
5297
+ api: "openai-completions";
5266
5298
  provider: string;
5267
5299
  baseUrl: string;
5268
5300
  reasoning: false;
@@ -5279,7 +5311,7 @@ export declare const MODELS: {
5279
5311
  readonly "mistral-large-2512": {
5280
5312
  id: string;
5281
5313
  name: string;
5282
- api: "mistral-conversations";
5314
+ api: "openai-completions";
5283
5315
  provider: string;
5284
5316
  baseUrl: string;
5285
5317
  reasoning: false;
@@ -5296,7 +5328,7 @@ export declare const MODELS: {
5296
5328
  readonly "mistral-large-latest": {
5297
5329
  id: string;
5298
5330
  name: string;
5299
- api: "mistral-conversations";
5331
+ api: "openai-completions";
5300
5332
  provider: string;
5301
5333
  baseUrl: string;
5302
5334
  reasoning: false;
@@ -5313,7 +5345,7 @@ export declare const MODELS: {
5313
5345
  readonly "mistral-medium-2505": {
5314
5346
  id: string;
5315
5347
  name: string;
5316
- api: "mistral-conversations";
5348
+ api: "openai-completions";
5317
5349
  provider: string;
5318
5350
  baseUrl: string;
5319
5351
  reasoning: false;
@@ -5330,7 +5362,7 @@ export declare const MODELS: {
5330
5362
  readonly "mistral-medium-2508": {
5331
5363
  id: string;
5332
5364
  name: string;
5333
- api: "mistral-conversations";
5365
+ api: "openai-completions";
5334
5366
  provider: string;
5335
5367
  baseUrl: string;
5336
5368
  reasoning: false;
@@ -5347,7 +5379,7 @@ export declare const MODELS: {
5347
5379
  readonly "mistral-medium-latest": {
5348
5380
  id: string;
5349
5381
  name: string;
5350
- api: "mistral-conversations";
5382
+ api: "openai-completions";
5351
5383
  provider: string;
5352
5384
  baseUrl: string;
5353
5385
  reasoning: false;
@@ -5364,7 +5396,7 @@ export declare const MODELS: {
5364
5396
  readonly "mistral-nemo": {
5365
5397
  id: string;
5366
5398
  name: string;
5367
- api: "mistral-conversations";
5399
+ api: "openai-completions";
5368
5400
  provider: string;
5369
5401
  baseUrl: string;
5370
5402
  reasoning: false;
@@ -5381,7 +5413,7 @@ export declare const MODELS: {
5381
5413
  readonly "mistral-small-2506": {
5382
5414
  id: string;
5383
5415
  name: string;
5384
- api: "mistral-conversations";
5416
+ api: "openai-completions";
5385
5417
  provider: string;
5386
5418
  baseUrl: string;
5387
5419
  reasoning: false;
@@ -5398,7 +5430,7 @@ export declare const MODELS: {
5398
5430
  readonly "mistral-small-2603": {
5399
5431
  id: string;
5400
5432
  name: string;
5401
- api: "mistral-conversations";
5433
+ api: "openai-completions";
5402
5434
  provider: string;
5403
5435
  baseUrl: string;
5404
5436
  reasoning: true;
@@ -5415,7 +5447,7 @@ export declare const MODELS: {
5415
5447
  readonly "mistral-small-latest": {
5416
5448
  id: string;
5417
5449
  name: string;
5418
- api: "mistral-conversations";
5450
+ api: "openai-completions";
5419
5451
  provider: string;
5420
5452
  baseUrl: string;
5421
5453
  reasoning: true;
@@ -5432,7 +5464,7 @@ export declare const MODELS: {
5432
5464
  readonly "open-mistral-7b": {
5433
5465
  id: string;
5434
5466
  name: string;
5435
- api: "mistral-conversations";
5467
+ api: "openai-completions";
5436
5468
  provider: string;
5437
5469
  baseUrl: string;
5438
5470
  reasoning: false;
@@ -5449,7 +5481,7 @@ export declare const MODELS: {
5449
5481
  readonly "open-mixtral-8x22b": {
5450
5482
  id: string;
5451
5483
  name: string;
5452
- api: "mistral-conversations";
5484
+ api: "openai-completions";
5453
5485
  provider: string;
5454
5486
  baseUrl: string;
5455
5487
  reasoning: false;
@@ -5466,7 +5498,7 @@ export declare const MODELS: {
5466
5498
  readonly "open-mixtral-8x7b": {
5467
5499
  id: string;
5468
5500
  name: string;
5469
- api: "mistral-conversations";
5501
+ api: "openai-completions";
5470
5502
  provider: string;
5471
5503
  baseUrl: string;
5472
5504
  reasoning: false;
@@ -5483,7 +5515,7 @@ export declare const MODELS: {
5483
5515
  readonly "pixtral-12b": {
5484
5516
  id: string;
5485
5517
  name: string;
5486
- api: "mistral-conversations";
5518
+ api: "openai-completions";
5487
5519
  provider: string;
5488
5520
  baseUrl: string;
5489
5521
  reasoning: false;
@@ -5500,7 +5532,7 @@ export declare const MODELS: {
5500
5532
  readonly "pixtral-large-latest": {
5501
5533
  id: string;
5502
5534
  name: string;
5503
- api: "mistral-conversations";
5535
+ api: "openai-completions";
5504
5536
  provider: string;
5505
5537
  baseUrl: string;
5506
5538
  reasoning: false;
@@ -5516,23 +5548,6 @@ export declare const MODELS: {
5516
5548
  };
5517
5549
  };
5518
5550
  readonly openai: {
5519
- readonly "codex-mini-latest": {
5520
- id: string;
5521
- name: string;
5522
- api: "openai-responses";
5523
- provider: string;
5524
- baseUrl: string;
5525
- reasoning: true;
5526
- input: "text"[];
5527
- cost: {
5528
- input: number;
5529
- output: number;
5530
- cacheRead: number;
5531
- cacheWrite: number;
5532
- };
5533
- contextWindow: number;
5534
- maxTokens: number;
5535
- };
5536
5551
  readonly "gpt-4": {
5537
5552
  id: string;
5538
5553
  name: string;
@@ -6472,6 +6487,23 @@ export declare const MODELS: {
6472
6487
  contextWindow: number;
6473
6488
  maxTokens: number;
6474
6489
  };
6490
+ readonly "claude-opus-4-7": {
6491
+ id: string;
6492
+ name: string;
6493
+ api: "anthropic-messages";
6494
+ provider: string;
6495
+ baseUrl: string;
6496
+ reasoning: true;
6497
+ input: ("image" | "text")[];
6498
+ cost: {
6499
+ input: number;
6500
+ output: number;
6501
+ cacheRead: number;
6502
+ cacheWrite: number;
6503
+ };
6504
+ contextWindow: number;
6505
+ maxTokens: number;
6506
+ };
6475
6507
  readonly "claude-sonnet-4": {
6476
6508
  id: string;
6477
6509
  name: string;
@@ -6574,6 +6606,23 @@ export declare const MODELS: {
6574
6606
  contextWindow: number;
6575
6607
  maxTokens: number;
6576
6608
  };
6609
+ readonly "glm-5.1": {
6610
+ id: string;
6611
+ name: string;
6612
+ api: "openai-completions";
6613
+ provider: string;
6614
+ baseUrl: string;
6615
+ reasoning: true;
6616
+ input: "text"[];
6617
+ cost: {
6618
+ input: number;
6619
+ output: number;
6620
+ cacheRead: number;
6621
+ cacheWrite: number;
6622
+ };
6623
+ contextWindow: number;
6624
+ maxTokens: number;
6625
+ };
6577
6626
  readonly "gpt-5": {
6578
6627
  id: string;
6579
6628
  name: string;
@@ -6812,6 +6861,23 @@ export declare const MODELS: {
6812
6861
  contextWindow: number;
6813
6862
  maxTokens: number;
6814
6863
  };
6864
+ readonly "hy3-preview-free": {
6865
+ id: string;
6866
+ name: string;
6867
+ api: "openai-completions";
6868
+ provider: string;
6869
+ baseUrl: string;
6870
+ reasoning: true;
6871
+ input: "text"[];
6872
+ cost: {
6873
+ input: number;
6874
+ output: number;
6875
+ cacheRead: number;
6876
+ cacheWrite: number;
6877
+ };
6878
+ contextWindow: number;
6879
+ maxTokens: number;
6880
+ };
6815
6881
  readonly "kimi-k2.5": {
6816
6882
  id: string;
6817
6883
  name: string;
@@ -6829,6 +6895,40 @@ export declare const MODELS: {
6829
6895
  contextWindow: number;
6830
6896
  maxTokens: number;
6831
6897
  };
6898
+ readonly "kimi-k2.6": {
6899
+ id: string;
6900
+ name: string;
6901
+ api: "openai-completions";
6902
+ provider: string;
6903
+ baseUrl: string;
6904
+ reasoning: true;
6905
+ input: ("image" | "text")[];
6906
+ cost: {
6907
+ input: number;
6908
+ output: number;
6909
+ cacheRead: number;
6910
+ cacheWrite: number;
6911
+ };
6912
+ contextWindow: number;
6913
+ maxTokens: number;
6914
+ };
6915
+ readonly "ling-2.6-flash-free": {
6916
+ id: string;
6917
+ name: string;
6918
+ api: "openai-completions";
6919
+ provider: string;
6920
+ baseUrl: string;
6921
+ reasoning: false;
6922
+ input: "text"[];
6923
+ cost: {
6924
+ input: number;
6925
+ output: number;
6926
+ cacheRead: number;
6927
+ cacheWrite: number;
6928
+ };
6929
+ contextWindow: number;
6930
+ maxTokens: number;
6931
+ };
6832
6932
  readonly "minimax-m2.5": {
6833
6933
  id: string;
6834
6934
  name: string;
@@ -6863,7 +6963,7 @@ export declare const MODELS: {
6863
6963
  contextWindow: number;
6864
6964
  maxTokens: number;
6865
6965
  };
6866
- readonly "nemotron-3-super-free": {
6966
+ readonly "minimax-m2.7": {
6867
6967
  id: string;
6868
6968
  name: string;
6869
6969
  api: "openai-completions";
@@ -6880,7 +6980,7 @@ export declare const MODELS: {
6880
6980
  contextWindow: number;
6881
6981
  maxTokens: number;
6882
6982
  };
6883
- readonly "qwen3.6-plus-free": {
6983
+ readonly "nemotron-3-super-free": {
6884
6984
  id: string;
6885
6985
  name: string;
6886
6986
  api: "openai-completions";
@@ -6897,6 +6997,40 @@ export declare const MODELS: {
6897
6997
  contextWindow: number;
6898
6998
  maxTokens: number;
6899
6999
  };
7000
+ readonly "qwen3.5-plus": {
7001
+ id: string;
7002
+ name: string;
7003
+ api: "anthropic-messages";
7004
+ provider: string;
7005
+ baseUrl: string;
7006
+ reasoning: true;
7007
+ input: ("image" | "text")[];
7008
+ cost: {
7009
+ input: number;
7010
+ output: number;
7011
+ cacheRead: number;
7012
+ cacheWrite: number;
7013
+ };
7014
+ contextWindow: number;
7015
+ maxTokens: number;
7016
+ };
7017
+ readonly "qwen3.6-plus": {
7018
+ id: string;
7019
+ name: string;
7020
+ api: "anthropic-messages";
7021
+ provider: string;
7022
+ baseUrl: string;
7023
+ reasoning: true;
7024
+ input: ("image" | "text")[];
7025
+ cost: {
7026
+ input: number;
7027
+ output: number;
7028
+ cacheRead: number;
7029
+ cacheWrite: number;
7030
+ };
7031
+ contextWindow: number;
7032
+ maxTokens: number;
7033
+ };
6900
7034
  };
6901
7035
  readonly "opencode-go": {
6902
7036
  readonly "glm-5": {
@@ -6916,6 +7050,23 @@ export declare const MODELS: {
6916
7050
  contextWindow: number;
6917
7051
  maxTokens: number;
6918
7052
  };
7053
+ readonly "glm-5.1": {
7054
+ id: string;
7055
+ name: string;
7056
+ api: "openai-completions";
7057
+ provider: string;
7058
+ baseUrl: string;
7059
+ reasoning: true;
7060
+ input: "text"[];
7061
+ cost: {
7062
+ input: number;
7063
+ output: number;
7064
+ cacheRead: number;
7065
+ cacheWrite: number;
7066
+ };
7067
+ contextWindow: number;
7068
+ maxTokens: number;
7069
+ };
6919
7070
  readonly "kimi-k2.5": {
6920
7071
  id: string;
6921
7072
  name: string;
@@ -6933,6 +7084,23 @@ export declare const MODELS: {
6933
7084
  contextWindow: number;
6934
7085
  maxTokens: number;
6935
7086
  };
7087
+ readonly "kimi-k2.6": {
7088
+ id: string;
7089
+ name: string;
7090
+ api: "openai-completions";
7091
+ provider: string;
7092
+ baseUrl: string;
7093
+ reasoning: true;
7094
+ input: ("image" | "text")[];
7095
+ cost: {
7096
+ input: number;
7097
+ output: number;
7098
+ cacheRead: number;
7099
+ cacheWrite: number;
7100
+ };
7101
+ contextWindow: number;
7102
+ maxTokens: number;
7103
+ };
6936
7104
  readonly "mimo-v2-omni": {
6937
7105
  id: string;
6938
7106
  name: string;
@@ -6967,6 +7135,40 @@ export declare const MODELS: {
6967
7135
  contextWindow: number;
6968
7136
  maxTokens: number;
6969
7137
  };
7138
+ readonly "mimo-v2.5": {
7139
+ id: string;
7140
+ name: string;
7141
+ api: "openai-completions";
7142
+ provider: string;
7143
+ baseUrl: string;
7144
+ reasoning: true;
7145
+ input: ("image" | "text")[];
7146
+ cost: {
7147
+ input: number;
7148
+ output: number;
7149
+ cacheRead: number;
7150
+ cacheWrite: number;
7151
+ };
7152
+ contextWindow: number;
7153
+ maxTokens: number;
7154
+ };
7155
+ readonly "mimo-v2.5-pro": {
7156
+ id: string;
7157
+ name: string;
7158
+ api: "openai-completions";
7159
+ provider: string;
7160
+ baseUrl: string;
7161
+ reasoning: true;
7162
+ input: "text"[];
7163
+ cost: {
7164
+ input: number;
7165
+ output: number;
7166
+ cacheRead: number;
7167
+ cacheWrite: number;
7168
+ };
7169
+ contextWindow: number;
7170
+ maxTokens: number;
7171
+ };
6970
7172
  readonly "minimax-m2.5": {
6971
7173
  id: string;
6972
7174
  name: string;
@@ -7001,6 +7203,40 @@ export declare const MODELS: {
7001
7203
  contextWindow: number;
7002
7204
  maxTokens: number;
7003
7205
  };
7206
+ readonly "qwen3.5-plus": {
7207
+ id: string;
7208
+ name: string;
7209
+ api: "anthropic-messages";
7210
+ provider: string;
7211
+ baseUrl: string;
7212
+ reasoning: true;
7213
+ input: ("image" | "text")[];
7214
+ cost: {
7215
+ input: number;
7216
+ output: number;
7217
+ cacheRead: number;
7218
+ cacheWrite: number;
7219
+ };
7220
+ contextWindow: number;
7221
+ maxTokens: number;
7222
+ };
7223
+ readonly "qwen3.6-plus": {
7224
+ id: string;
7225
+ name: string;
7226
+ api: "anthropic-messages";
7227
+ provider: string;
7228
+ baseUrl: string;
7229
+ reasoning: true;
7230
+ input: ("image" | "text")[];
7231
+ cost: {
7232
+ input: number;
7233
+ output: number;
7234
+ cacheRead: number;
7235
+ cacheWrite: number;
7236
+ };
7237
+ contextWindow: number;
7238
+ maxTokens: number;
7239
+ };
7004
7240
  };
7005
7241
  readonly openrouter: {
7006
7242
  readonly "ai21/jamba-large-1.7": {
@@ -7292,7 +7528,7 @@ export declare const MODELS: {
7292
7528
  contextWindow: number;
7293
7529
  maxTokens: number;
7294
7530
  };
7295
- readonly "anthropic/claude-sonnet-4": {
7531
+ readonly "anthropic/claude-opus-4.6-fast": {
7296
7532
  id: string;
7297
7533
  name: string;
7298
7534
  api: "openai-completions";
@@ -7309,7 +7545,7 @@ export declare const MODELS: {
7309
7545
  contextWindow: number;
7310
7546
  maxTokens: number;
7311
7547
  };
7312
- readonly "anthropic/claude-sonnet-4.5": {
7548
+ readonly "anthropic/claude-opus-4.7": {
7313
7549
  id: string;
7314
7550
  name: string;
7315
7551
  api: "openai-completions";
@@ -7326,7 +7562,7 @@ export declare const MODELS: {
7326
7562
  contextWindow: number;
7327
7563
  maxTokens: number;
7328
7564
  };
7329
- readonly "anthropic/claude-sonnet-4.6": {
7565
+ readonly "anthropic/claude-sonnet-4": {
7330
7566
  id: string;
7331
7567
  name: string;
7332
7568
  api: "openai-completions";
@@ -7343,14 +7579,14 @@ export declare const MODELS: {
7343
7579
  contextWindow: number;
7344
7580
  maxTokens: number;
7345
7581
  };
7346
- readonly "arcee-ai/trinity-large-preview:free": {
7582
+ readonly "anthropic/claude-sonnet-4.5": {
7347
7583
  id: string;
7348
7584
  name: string;
7349
7585
  api: "openai-completions";
7350
7586
  provider: string;
7351
7587
  baseUrl: string;
7352
- reasoning: false;
7353
- input: "text"[];
7588
+ reasoning: true;
7589
+ input: ("image" | "text")[];
7354
7590
  cost: {
7355
7591
  input: number;
7356
7592
  output: number;
@@ -7360,13 +7596,30 @@ export declare const MODELS: {
7360
7596
  contextWindow: number;
7361
7597
  maxTokens: number;
7362
7598
  };
7363
- readonly "arcee-ai/trinity-large-thinking": {
7599
+ readonly "anthropic/claude-sonnet-4.6": {
7364
7600
  id: string;
7365
7601
  name: string;
7366
7602
  api: "openai-completions";
7367
7603
  provider: string;
7368
7604
  baseUrl: string;
7369
7605
  reasoning: true;
7606
+ input: ("image" | "text")[];
7607
+ cost: {
7608
+ input: number;
7609
+ output: number;
7610
+ cacheRead: number;
7611
+ cacheWrite: number;
7612
+ };
7613
+ contextWindow: number;
7614
+ maxTokens: number;
7615
+ };
7616
+ readonly "arcee-ai/trinity-large-preview": {
7617
+ id: string;
7618
+ name: string;
7619
+ api: "openai-completions";
7620
+ provider: string;
7621
+ baseUrl: string;
7622
+ reasoning: false;
7370
7623
  input: "text"[];
7371
7624
  cost: {
7372
7625
  input: number;
@@ -7377,7 +7630,7 @@ export declare const MODELS: {
7377
7630
  contextWindow: number;
7378
7631
  maxTokens: number;
7379
7632
  };
7380
- readonly "arcee-ai/trinity-mini": {
7633
+ readonly "arcee-ai/trinity-large-thinking": {
7381
7634
  id: string;
7382
7635
  name: string;
7383
7636
  api: "openai-completions";
@@ -7394,7 +7647,7 @@ export declare const MODELS: {
7394
7647
  contextWindow: number;
7395
7648
  maxTokens: number;
7396
7649
  };
7397
- readonly "arcee-ai/trinity-mini:free": {
7650
+ readonly "arcee-ai/trinity-mini": {
7398
7651
  id: string;
7399
7652
  name: string;
7400
7653
  api: "openai-completions";
@@ -7604,7 +7857,7 @@ export declare const MODELS: {
7604
7857
  api: "openai-completions";
7605
7858
  provider: string;
7606
7859
  baseUrl: string;
7607
- reasoning: true;
7860
+ reasoning: false;
7608
7861
  input: "text"[];
7609
7862
  cost: {
7610
7863
  input: number;
@@ -7955,7 +8208,7 @@ export declare const MODELS: {
7955
8208
  contextWindow: number;
7956
8209
  maxTokens: number;
7957
8210
  };
7958
- readonly "google/gemma-4-31b-it": {
8211
+ readonly "google/gemma-4-26b-a4b-it:free": {
7959
8212
  id: string;
7960
8213
  name: string;
7961
8214
  api: "openai-completions";
@@ -7972,14 +8225,14 @@ export declare const MODELS: {
7972
8225
  contextWindow: number;
7973
8226
  maxTokens: number;
7974
8227
  };
7975
- readonly "inception/mercury": {
8228
+ readonly "google/gemma-4-31b-it": {
7976
8229
  id: string;
7977
8230
  name: string;
7978
8231
  api: "openai-completions";
7979
8232
  provider: string;
7980
8233
  baseUrl: string;
7981
- reasoning: false;
7982
- input: "text"[];
8234
+ reasoning: true;
8235
+ input: ("image" | "text")[];
7983
8236
  cost: {
7984
8237
  input: number;
7985
8238
  output: number;
@@ -7989,14 +8242,14 @@ export declare const MODELS: {
7989
8242
  contextWindow: number;
7990
8243
  maxTokens: number;
7991
8244
  };
7992
- readonly "inception/mercury-2": {
8245
+ readonly "google/gemma-4-31b-it:free": {
7993
8246
  id: string;
7994
8247
  name: string;
7995
8248
  api: "openai-completions";
7996
8249
  provider: string;
7997
8250
  baseUrl: string;
7998
8251
  reasoning: true;
7999
- input: "text"[];
8252
+ input: ("image" | "text")[];
8000
8253
  cost: {
8001
8254
  input: number;
8002
8255
  output: number;
@@ -8006,13 +8259,13 @@ export declare const MODELS: {
8006
8259
  contextWindow: number;
8007
8260
  maxTokens: number;
8008
8261
  };
8009
- readonly "inception/mercury-coder": {
8262
+ readonly "inception/mercury-2": {
8010
8263
  id: string;
8011
8264
  name: string;
8012
8265
  api: "openai-completions";
8013
8266
  provider: string;
8014
8267
  baseUrl: string;
8015
- reasoning: false;
8268
+ reasoning: true;
8016
8269
  input: "text"[];
8017
8270
  cost: {
8018
8271
  input: number;
@@ -8023,7 +8276,7 @@ export declare const MODELS: {
8023
8276
  contextWindow: number;
8024
8277
  maxTokens: number;
8025
8278
  };
8026
- readonly "kwaipilot/kat-coder-pro-v2": {
8279
+ readonly "inclusionai/ling-2.6-flash:free": {
8027
8280
  id: string;
8028
8281
  name: string;
8029
8282
  api: "openai-completions";
@@ -8040,7 +8293,7 @@ export declare const MODELS: {
8040
8293
  contextWindow: number;
8041
8294
  maxTokens: number;
8042
8295
  };
8043
- readonly "meituan/longcat-flash-chat": {
8296
+ readonly "kwaipilot/kat-coder-pro-v2": {
8044
8297
  id: string;
8045
8298
  name: string;
8046
8299
  api: "openai-completions";
@@ -8142,23 +8395,6 @@ export declare const MODELS: {
8142
8395
  contextWindow: number;
8143
8396
  maxTokens: number;
8144
8397
  };
8145
- readonly "meta-llama/llama-4-maverick": {
8146
- id: string;
8147
- name: string;
8148
- api: "openai-completions";
8149
- provider: string;
8150
- baseUrl: string;
8151
- reasoning: false;
8152
- input: ("image" | "text")[];
8153
- cost: {
8154
- input: number;
8155
- output: number;
8156
- cacheRead: number;
8157
- cacheWrite: number;
8158
- };
8159
- contextWindow: number;
8160
- maxTokens: number;
8161
- };
8162
8398
  readonly "meta-llama/llama-4-scout": {
8163
8399
  id: string;
8164
8400
  name: string;
@@ -8652,7 +8888,24 @@ export declare const MODELS: {
8652
8888
  contextWindow: number;
8653
8889
  maxTokens: number;
8654
8890
  };
8655
- readonly "moonshotai/kimi-k2": {
8891
+ readonly "moonshotai/kimi-k2": {
8892
+ id: string;
8893
+ name: string;
8894
+ api: "openai-completions";
8895
+ provider: string;
8896
+ baseUrl: string;
8897
+ reasoning: false;
8898
+ input: "text"[];
8899
+ cost: {
8900
+ input: number;
8901
+ output: number;
8902
+ cacheRead: number;
8903
+ cacheWrite: number;
8904
+ };
8905
+ contextWindow: number;
8906
+ maxTokens: number;
8907
+ };
8908
+ readonly "moonshotai/kimi-k2-0905": {
8656
8909
  id: string;
8657
8910
  name: string;
8658
8911
  api: "openai-completions";
@@ -8669,13 +8922,13 @@ export declare const MODELS: {
8669
8922
  contextWindow: number;
8670
8923
  maxTokens: number;
8671
8924
  };
8672
- readonly "moonshotai/kimi-k2-0905": {
8925
+ readonly "moonshotai/kimi-k2-thinking": {
8673
8926
  id: string;
8674
8927
  name: string;
8675
8928
  api: "openai-completions";
8676
8929
  provider: string;
8677
8930
  baseUrl: string;
8678
- reasoning: false;
8931
+ reasoning: true;
8679
8932
  input: "text"[];
8680
8933
  cost: {
8681
8934
  input: number;
@@ -8686,14 +8939,14 @@ export declare const MODELS: {
8686
8939
  contextWindow: number;
8687
8940
  maxTokens: number;
8688
8941
  };
8689
- readonly "moonshotai/kimi-k2-thinking": {
8942
+ readonly "moonshotai/kimi-k2.5": {
8690
8943
  id: string;
8691
8944
  name: string;
8692
8945
  api: "openai-completions";
8693
8946
  provider: string;
8694
8947
  baseUrl: string;
8695
8948
  reasoning: true;
8696
- input: "text"[];
8949
+ input: ("image" | "text")[];
8697
8950
  cost: {
8698
8951
  input: number;
8699
8952
  output: number;
@@ -8703,7 +8956,7 @@ export declare const MODELS: {
8703
8956
  contextWindow: number;
8704
8957
  maxTokens: number;
8705
8958
  };
8706
- readonly "moonshotai/kimi-k2.5": {
8959
+ readonly "moonshotai/kimi-k2.6": {
8707
8960
  id: string;
8708
8961
  name: string;
8709
8962
  api: "openai-completions";
@@ -9196,23 +9449,6 @@ export declare const MODELS: {
9196
9449
  contextWindow: number;
9197
9450
  maxTokens: number;
9198
9451
  };
9199
- readonly "openai/gpt-4o:extended": {
9200
- id: string;
9201
- name: string;
9202
- api: "openai-completions";
9203
- provider: string;
9204
- baseUrl: string;
9205
- reasoning: false;
9206
- input: ("image" | "text")[];
9207
- cost: {
9208
- input: number;
9209
- output: number;
9210
- cacheRead: number;
9211
- cacheWrite: number;
9212
- };
9213
- contextWindow: number;
9214
- maxTokens: number;
9215
- };
9216
9452
  readonly "openai/gpt-5": {
9217
9453
  id: string;
9218
9454
  name: string;
@@ -9247,40 +9483,6 @@ export declare const MODELS: {
9247
9483
  contextWindow: number;
9248
9484
  maxTokens: number;
9249
9485
  };
9250
- readonly "openai/gpt-5-image": {
9251
- id: string;
9252
- name: string;
9253
- api: "openai-completions";
9254
- provider: string;
9255
- baseUrl: string;
9256
- reasoning: true;
9257
- input: ("image" | "text")[];
9258
- cost: {
9259
- input: number;
9260
- output: number;
9261
- cacheRead: number;
9262
- cacheWrite: number;
9263
- };
9264
- contextWindow: number;
9265
- maxTokens: number;
9266
- };
9267
- readonly "openai/gpt-5-image-mini": {
9268
- id: string;
9269
- name: string;
9270
- api: "openai-completions";
9271
- provider: string;
9272
- baseUrl: string;
9273
- reasoning: true;
9274
- input: ("image" | "text")[];
9275
- cost: {
9276
- input: number;
9277
- output: number;
9278
- cacheRead: number;
9279
- cacheWrite: number;
9280
- };
9281
- contextWindow: number;
9282
- maxTokens: number;
9283
- };
9284
9486
  readonly "openai/gpt-5-mini": {
9285
9487
  id: string;
9286
9488
  name: string;
@@ -10086,7 +10288,7 @@ export declare const MODELS: {
10086
10288
  api: "openai-completions";
10087
10289
  provider: string;
10088
10290
  baseUrl: string;
10089
- reasoning: true;
10291
+ reasoning: false;
10090
10292
  input: "text"[];
10091
10293
  cost: {
10092
10294
  input: number;
@@ -10624,7 +10826,7 @@ export declare const MODELS: {
10624
10826
  contextWindow: number;
10625
10827
  maxTokens: number;
10626
10828
  };
10627
- readonly "qwen/qwen3.6-plus:free": {
10829
+ readonly "qwen/qwen3.6-plus": {
10628
10830
  id: string;
10629
10831
  name: string;
10630
10832
  api: "openai-completions";
@@ -10743,7 +10945,7 @@ export declare const MODELS: {
10743
10945
  contextWindow: number;
10744
10946
  maxTokens: number;
10745
10947
  };
10746
- readonly "stepfun/step-3.5-flash:free": {
10948
+ readonly "tencent/hy3-preview:free": {
10747
10949
  id: string;
10748
10950
  name: string;
10749
10951
  api: "openai-completions";
@@ -11032,6 +11234,40 @@ export declare const MODELS: {
11032
11234
  contextWindow: number;
11033
11235
  maxTokens: number;
11034
11236
  };
11237
+ readonly "xiaomi/mimo-v2.5": {
11238
+ id: string;
11239
+ name: string;
11240
+ api: "openai-completions";
11241
+ provider: string;
11242
+ baseUrl: string;
11243
+ reasoning: true;
11244
+ input: ("image" | "text")[];
11245
+ cost: {
11246
+ input: number;
11247
+ output: number;
11248
+ cacheRead: number;
11249
+ cacheWrite: number;
11250
+ };
11251
+ contextWindow: number;
11252
+ maxTokens: number;
11253
+ };
11254
+ readonly "xiaomi/mimo-v2.5-pro": {
11255
+ id: string;
11256
+ name: string;
11257
+ api: "openai-completions";
11258
+ provider: string;
11259
+ baseUrl: string;
11260
+ reasoning: true;
11261
+ input: "text"[];
11262
+ cost: {
11263
+ input: number;
11264
+ output: number;
11265
+ cacheRead: number;
11266
+ cacheWrite: number;
11267
+ };
11268
+ contextWindow: number;
11269
+ maxTokens: number;
11270
+ };
11035
11271
  readonly "z-ai/glm-4-32b": {
11036
11272
  id: string;
11037
11273
  name: string;
@@ -11219,6 +11455,23 @@ export declare const MODELS: {
11219
11455
  contextWindow: number;
11220
11456
  maxTokens: number;
11221
11457
  };
11458
+ readonly "z-ai/glm-5.1": {
11459
+ id: string;
11460
+ name: string;
11461
+ api: "openai-completions";
11462
+ provider: string;
11463
+ baseUrl: string;
11464
+ reasoning: true;
11465
+ input: "text"[];
11466
+ cost: {
11467
+ input: number;
11468
+ output: number;
11469
+ cacheRead: number;
11470
+ cacheWrite: number;
11471
+ };
11472
+ contextWindow: number;
11473
+ maxTokens: number;
11474
+ };
11222
11475
  readonly "z-ai/glm-5v-turbo": {
11223
11476
  id: string;
11224
11477
  name: string;
@@ -11236,6 +11489,23 @@ export declare const MODELS: {
11236
11489
  contextWindow: number;
11237
11490
  maxTokens: number;
11238
11491
  };
11492
+ readonly "~anthropic/claude-opus-latest": {
11493
+ id: string;
11494
+ name: string;
11495
+ api: "openai-completions";
11496
+ provider: string;
11497
+ baseUrl: string;
11498
+ reasoning: true;
11499
+ input: ("image" | "text")[];
11500
+ cost: {
11501
+ input: number;
11502
+ output: number;
11503
+ cacheRead: number;
11504
+ cacheWrite: number;
11505
+ };
11506
+ contextWindow: number;
11507
+ maxTokens: number;
11508
+ };
11239
11509
  };
11240
11510
  readonly "vercel-ai-gateway": {
11241
11511
  readonly "alibaba/qwen-3-14b": {
@@ -11306,6 +11576,23 @@ export declare const MODELS: {
11306
11576
  contextWindow: number;
11307
11577
  maxTokens: number;
11308
11578
  };
11579
+ readonly "alibaba/qwen-3.6-max-preview": {
11580
+ id: string;
11581
+ name: string;
11582
+ api: "anthropic-messages";
11583
+ provider: string;
11584
+ baseUrl: string;
11585
+ reasoning: true;
11586
+ input: ("image" | "text")[];
11587
+ cost: {
11588
+ input: number;
11589
+ output: number;
11590
+ cacheRead: number;
11591
+ cacheWrite: number;
11592
+ };
11593
+ contextWindow: number;
11594
+ maxTokens: number;
11595
+ };
11309
11596
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11310
11597
  id: string;
11311
11598
  name: string;
@@ -11646,6 +11933,23 @@ export declare const MODELS: {
11646
11933
  contextWindow: number;
11647
11934
  maxTokens: number;
11648
11935
  };
11936
+ readonly "anthropic/claude-opus-4.7": {
11937
+ id: string;
11938
+ name: string;
11939
+ api: "anthropic-messages";
11940
+ provider: string;
11941
+ baseUrl: string;
11942
+ reasoning: true;
11943
+ input: ("image" | "text")[];
11944
+ cost: {
11945
+ input: number;
11946
+ output: number;
11947
+ cacheRead: number;
11948
+ cacheWrite: number;
11949
+ };
11950
+ contextWindow: number;
11951
+ maxTokens: number;
11952
+ };
11649
11953
  readonly "anthropic/claude-sonnet-4": {
11650
11954
  id: string;
11651
11955
  name: string;
@@ -12632,6 +12936,23 @@ export declare const MODELS: {
12632
12936
  contextWindow: number;
12633
12937
  maxTokens: number;
12634
12938
  };
12939
+ readonly "moonshotai/kimi-k2.6": {
12940
+ id: string;
12941
+ name: string;
12942
+ api: "anthropic-messages";
12943
+ provider: string;
12944
+ baseUrl: string;
12945
+ reasoning: true;
12946
+ input: ("image" | "text")[];
12947
+ cost: {
12948
+ input: number;
12949
+ output: number;
12950
+ cacheRead: number;
12951
+ cacheWrite: number;
12952
+ };
12953
+ contextWindow: number;
12954
+ maxTokens: number;
12955
+ };
12635
12956
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12636
12957
  id: string;
12637
12958
  name: string;
@@ -13125,23 +13446,6 @@ export declare const MODELS: {
13125
13446
  contextWindow: number;
13126
13447
  maxTokens: number;
13127
13448
  };
13128
- readonly "openai/gpt-oss-120b": {
13129
- id: string;
13130
- name: string;
13131
- api: "anthropic-messages";
13132
- provider: string;
13133
- baseUrl: string;
13134
- reasoning: true;
13135
- input: "text"[];
13136
- cost: {
13137
- input: number;
13138
- output: number;
13139
- cacheRead: number;
13140
- cacheWrite: number;
13141
- };
13142
- contextWindow: number;
13143
- maxTokens: number;
13144
- };
13145
13449
  readonly "openai/gpt-oss-20b": {
13146
13450
  id: string;
13147
13451
  name: string;
@@ -13329,23 +13633,6 @@ export declare const MODELS: {
13329
13633
  contextWindow: number;
13330
13634
  maxTokens: number;
13331
13635
  };
13332
- readonly "xai/grok-2-vision": {
13333
- id: string;
13334
- name: string;
13335
- api: "anthropic-messages";
13336
- provider: string;
13337
- baseUrl: string;
13338
- reasoning: false;
13339
- input: ("image" | "text")[];
13340
- cost: {
13341
- input: number;
13342
- output: number;
13343
- cacheRead: number;
13344
- cacheWrite: number;
13345
- };
13346
- contextWindow: number;
13347
- maxTokens: number;
13348
- };
13349
13636
  readonly "xai/grok-3": {
13350
13637
  id: string;
13351
13638
  name: string;
@@ -13438,7 +13725,7 @@ export declare const MODELS: {
13438
13725
  provider: string;
13439
13726
  baseUrl: string;
13440
13727
  reasoning: false;
13441
- input: "text"[];
13728
+ input: ("image" | "text")[];
13442
13729
  cost: {
13443
13730
  input: number;
13444
13731
  output: number;
@@ -13455,7 +13742,7 @@ export declare const MODELS: {
13455
13742
  provider: string;
13456
13743
  baseUrl: string;
13457
13744
  reasoning: true;
13458
- input: "text"[];
13745
+ input: ("image" | "text")[];
13459
13746
  cost: {
13460
13747
  input: number;
13461
13748
  output: number;
@@ -13472,7 +13759,7 @@ export declare const MODELS: {
13472
13759
  provider: string;
13473
13760
  baseUrl: string;
13474
13761
  reasoning: false;
13475
- input: "text"[];
13762
+ input: ("image" | "text")[];
13476
13763
  cost: {
13477
13764
  input: number;
13478
13765
  output: number;
@@ -13489,7 +13776,7 @@ export declare const MODELS: {
13489
13776
  provider: string;
13490
13777
  baseUrl: string;
13491
13778
  reasoning: true;
13492
- input: "text"[];
13779
+ input: ("image" | "text")[];
13493
13780
  cost: {
13494
13781
  input: number;
13495
13782
  output: number;
@@ -13506,7 +13793,7 @@ export declare const MODELS: {
13506
13793
  provider: string;
13507
13794
  baseUrl: string;
13508
13795
  reasoning: true;
13509
- input: "text"[];
13796
+ input: ("image" | "text")[];
13510
13797
  cost: {
13511
13798
  input: number;
13512
13799
  output: number;
@@ -13523,7 +13810,7 @@ export declare const MODELS: {
13523
13810
  provider: string;
13524
13811
  baseUrl: string;
13525
13812
  reasoning: true;
13526
- input: "text"[];
13813
+ input: ("image" | "text")[];
13527
13814
  cost: {
13528
13815
  input: number;
13529
13816
  output: number;
@@ -13839,6 +14126,23 @@ export declare const MODELS: {
13839
14126
  contextWindow: number;
13840
14127
  maxTokens: number;
13841
14128
  };
14129
+ readonly "zai/glm-5.1": {
14130
+ id: string;
14131
+ name: string;
14132
+ api: "anthropic-messages";
14133
+ provider: string;
14134
+ baseUrl: string;
14135
+ reasoning: true;
14136
+ input: "text"[];
14137
+ cost: {
14138
+ input: number;
14139
+ output: number;
14140
+ cacheRead: number;
14141
+ cacheWrite: number;
14142
+ };
14143
+ contextWindow: number;
14144
+ maxTokens: number;
14145
+ };
13842
14146
  readonly "zai/glm-5v-turbo": {
13843
14147
  id: string;
13844
14148
  name: string;
@@ -14506,6 +14810,28 @@ export declare const MODELS: {
14506
14810
  contextWindow: number;
14507
14811
  maxTokens: number;
14508
14812
  };
14813
+ readonly "glm-5.1": {
14814
+ id: string;
14815
+ name: string;
14816
+ api: "openai-completions";
14817
+ provider: string;
14818
+ baseUrl: string;
14819
+ compat: {
14820
+ supportsDeveloperRole: false;
14821
+ thinkingFormat: "zai";
14822
+ zaiToolStream: true;
14823
+ };
14824
+ reasoning: true;
14825
+ input: "text"[];
14826
+ cost: {
14827
+ input: number;
14828
+ output: number;
14829
+ cacheRead: number;
14830
+ cacheWrite: number;
14831
+ };
14832
+ contextWindow: number;
14833
+ maxTokens: number;
14834
+ };
14509
14835
  readonly "glm-5v-turbo": {
14510
14836
  id: string;
14511
14837
  name: string;