@draht/ai 2026.4.5 → 2026.4.23
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +21 -12
- package/dist/cli.d.ts.map +1 -1
- package/dist/cli.js +1 -1
- package/dist/cli.js.map +1 -1
- package/dist/models.generated.d.ts +683 -357
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +897 -586
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +37 -9
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +38 -16
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +2 -2
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +1 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +29 -4
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +13 -6
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +8 -1
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +2 -2
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/types.d.ts +62 -5
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts +3 -2
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +24 -11
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +5 -8
|
@@ -255,6 +255,23 @@ export declare const MODELS: {
|
|
|
255
255
|
contextWindow: number;
|
|
256
256
|
maxTokens: number;
|
|
257
257
|
};
|
|
258
|
+
readonly "anthropic.claude-opus-4-7": {
|
|
259
|
+
id: string;
|
|
260
|
+
name: string;
|
|
261
|
+
api: "bedrock-converse-stream";
|
|
262
|
+
provider: string;
|
|
263
|
+
baseUrl: string;
|
|
264
|
+
reasoning: true;
|
|
265
|
+
input: ("image" | "text")[];
|
|
266
|
+
cost: {
|
|
267
|
+
input: number;
|
|
268
|
+
output: number;
|
|
269
|
+
cacheRead: number;
|
|
270
|
+
cacheWrite: number;
|
|
271
|
+
};
|
|
272
|
+
contextWindow: number;
|
|
273
|
+
maxTokens: number;
|
|
274
|
+
};
|
|
258
275
|
readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
259
276
|
id: string;
|
|
260
277
|
name: string;
|
|
@@ -408,6 +425,23 @@ export declare const MODELS: {
|
|
|
408
425
|
contextWindow: number;
|
|
409
426
|
maxTokens: number;
|
|
410
427
|
};
|
|
428
|
+
readonly "eu.anthropic.claude-opus-4-7": {
|
|
429
|
+
id: string;
|
|
430
|
+
name: string;
|
|
431
|
+
api: "bedrock-converse-stream";
|
|
432
|
+
provider: string;
|
|
433
|
+
baseUrl: string;
|
|
434
|
+
reasoning: true;
|
|
435
|
+
input: ("image" | "text")[];
|
|
436
|
+
cost: {
|
|
437
|
+
input: number;
|
|
438
|
+
output: number;
|
|
439
|
+
cacheRead: number;
|
|
440
|
+
cacheWrite: number;
|
|
441
|
+
};
|
|
442
|
+
contextWindow: number;
|
|
443
|
+
maxTokens: number;
|
|
444
|
+
};
|
|
411
445
|
readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
412
446
|
id: string;
|
|
413
447
|
name: string;
|
|
@@ -510,6 +544,23 @@ export declare const MODELS: {
|
|
|
510
544
|
contextWindow: number;
|
|
511
545
|
maxTokens: number;
|
|
512
546
|
};
|
|
547
|
+
readonly "global.anthropic.claude-opus-4-7": {
|
|
548
|
+
id: string;
|
|
549
|
+
name: string;
|
|
550
|
+
api: "bedrock-converse-stream";
|
|
551
|
+
provider: string;
|
|
552
|
+
baseUrl: string;
|
|
553
|
+
reasoning: true;
|
|
554
|
+
input: ("image" | "text")[];
|
|
555
|
+
cost: {
|
|
556
|
+
input: number;
|
|
557
|
+
output: number;
|
|
558
|
+
cacheRead: number;
|
|
559
|
+
cacheWrite: number;
|
|
560
|
+
};
|
|
561
|
+
contextWindow: number;
|
|
562
|
+
maxTokens: number;
|
|
563
|
+
};
|
|
513
564
|
readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
514
565
|
id: string;
|
|
515
566
|
name: string;
|
|
@@ -1207,6 +1258,23 @@ export declare const MODELS: {
|
|
|
1207
1258
|
contextWindow: number;
|
|
1208
1259
|
maxTokens: number;
|
|
1209
1260
|
};
|
|
1261
|
+
readonly "qwen.qwen3-coder-next": {
|
|
1262
|
+
id: string;
|
|
1263
|
+
name: string;
|
|
1264
|
+
api: "bedrock-converse-stream";
|
|
1265
|
+
provider: string;
|
|
1266
|
+
baseUrl: string;
|
|
1267
|
+
reasoning: true;
|
|
1268
|
+
input: "text"[];
|
|
1269
|
+
cost: {
|
|
1270
|
+
input: number;
|
|
1271
|
+
output: number;
|
|
1272
|
+
cacheRead: number;
|
|
1273
|
+
cacheWrite: number;
|
|
1274
|
+
};
|
|
1275
|
+
contextWindow: number;
|
|
1276
|
+
maxTokens: number;
|
|
1277
|
+
};
|
|
1210
1278
|
readonly "qwen.qwen3-next-80b-a3b": {
|
|
1211
1279
|
id: string;
|
|
1212
1280
|
name: string;
|
|
@@ -1326,6 +1394,23 @@ export declare const MODELS: {
|
|
|
1326
1394
|
contextWindow: number;
|
|
1327
1395
|
maxTokens: number;
|
|
1328
1396
|
};
|
|
1397
|
+
readonly "us.anthropic.claude-opus-4-7": {
|
|
1398
|
+
id: string;
|
|
1399
|
+
name: string;
|
|
1400
|
+
api: "bedrock-converse-stream";
|
|
1401
|
+
provider: string;
|
|
1402
|
+
baseUrl: string;
|
|
1403
|
+
reasoning: true;
|
|
1404
|
+
input: ("image" | "text")[];
|
|
1405
|
+
cost: {
|
|
1406
|
+
input: number;
|
|
1407
|
+
output: number;
|
|
1408
|
+
cacheRead: number;
|
|
1409
|
+
cacheWrite: number;
|
|
1410
|
+
};
|
|
1411
|
+
contextWindow: number;
|
|
1412
|
+
maxTokens: number;
|
|
1413
|
+
};
|
|
1329
1414
|
readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
1330
1415
|
id: string;
|
|
1331
1416
|
name: string;
|
|
@@ -1753,7 +1838,7 @@ export declare const MODELS: {
|
|
|
1753
1838
|
contextWindow: number;
|
|
1754
1839
|
maxTokens: number;
|
|
1755
1840
|
};
|
|
1756
|
-
readonly "claude-
|
|
1841
|
+
readonly "claude-opus-4-7": {
|
|
1757
1842
|
id: string;
|
|
1758
1843
|
name: string;
|
|
1759
1844
|
api: "anthropic-messages";
|
|
@@ -1770,7 +1855,7 @@ export declare const MODELS: {
|
|
|
1770
1855
|
contextWindow: number;
|
|
1771
1856
|
maxTokens: number;
|
|
1772
1857
|
};
|
|
1773
|
-
readonly "claude-sonnet-4-
|
|
1858
|
+
readonly "claude-sonnet-4-0": {
|
|
1774
1859
|
id: string;
|
|
1775
1860
|
name: string;
|
|
1776
1861
|
api: "anthropic-messages";
|
|
@@ -1787,7 +1872,7 @@ export declare const MODELS: {
|
|
|
1787
1872
|
contextWindow: number;
|
|
1788
1873
|
maxTokens: number;
|
|
1789
1874
|
};
|
|
1790
|
-
readonly "claude-sonnet-4-
|
|
1875
|
+
readonly "claude-sonnet-4-20250514": {
|
|
1791
1876
|
id: string;
|
|
1792
1877
|
name: string;
|
|
1793
1878
|
api: "anthropic-messages";
|
|
@@ -1804,7 +1889,7 @@ export declare const MODELS: {
|
|
|
1804
1889
|
contextWindow: number;
|
|
1805
1890
|
maxTokens: number;
|
|
1806
1891
|
};
|
|
1807
|
-
readonly "claude-sonnet-4-5
|
|
1892
|
+
readonly "claude-sonnet-4-5": {
|
|
1808
1893
|
id: string;
|
|
1809
1894
|
name: string;
|
|
1810
1895
|
api: "anthropic-messages";
|
|
@@ -1821,7 +1906,7 @@ export declare const MODELS: {
|
|
|
1821
1906
|
contextWindow: number;
|
|
1822
1907
|
maxTokens: number;
|
|
1823
1908
|
};
|
|
1824
|
-
readonly "claude-sonnet-4-
|
|
1909
|
+
readonly "claude-sonnet-4-5-20250929": {
|
|
1825
1910
|
id: string;
|
|
1826
1911
|
name: string;
|
|
1827
1912
|
api: "anthropic-messages";
|
|
@@ -1838,16 +1923,14 @@ export declare const MODELS: {
|
|
|
1838
1923
|
contextWindow: number;
|
|
1839
1924
|
maxTokens: number;
|
|
1840
1925
|
};
|
|
1841
|
-
|
|
1842
|
-
readonly "azure-openai-responses": {
|
|
1843
|
-
readonly "codex-mini-latest": {
|
|
1926
|
+
readonly "claude-sonnet-4-6": {
|
|
1844
1927
|
id: string;
|
|
1845
1928
|
name: string;
|
|
1846
|
-
api: "
|
|
1929
|
+
api: "anthropic-messages";
|
|
1847
1930
|
provider: string;
|
|
1848
1931
|
baseUrl: string;
|
|
1849
1932
|
reasoning: true;
|
|
1850
|
-
input: "text"[];
|
|
1933
|
+
input: ("image" | "text")[];
|
|
1851
1934
|
cost: {
|
|
1852
1935
|
input: number;
|
|
1853
1936
|
output: number;
|
|
@@ -1857,6 +1940,8 @@ export declare const MODELS: {
|
|
|
1857
1940
|
contextWindow: number;
|
|
1858
1941
|
maxTokens: number;
|
|
1859
1942
|
};
|
|
1943
|
+
};
|
|
1944
|
+
readonly "azure-openai-responses": {
|
|
1860
1945
|
readonly "gpt-4": {
|
|
1861
1946
|
id: string;
|
|
1862
1947
|
name: string;
|
|
@@ -2678,6 +2763,29 @@ export declare const MODELS: {
|
|
|
2678
2763
|
contextWindow: number;
|
|
2679
2764
|
maxTokens: number;
|
|
2680
2765
|
};
|
|
2766
|
+
readonly "claude-opus-4.7": {
|
|
2767
|
+
id: string;
|
|
2768
|
+
name: string;
|
|
2769
|
+
api: "anthropic-messages";
|
|
2770
|
+
provider: string;
|
|
2771
|
+
baseUrl: string;
|
|
2772
|
+
headers: {
|
|
2773
|
+
"User-Agent": string;
|
|
2774
|
+
"Editor-Version": string;
|
|
2775
|
+
"Editor-Plugin-Version": string;
|
|
2776
|
+
"Copilot-Integration-Id": string;
|
|
2777
|
+
};
|
|
2778
|
+
reasoning: true;
|
|
2779
|
+
input: ("image" | "text")[];
|
|
2780
|
+
cost: {
|
|
2781
|
+
input: number;
|
|
2782
|
+
output: number;
|
|
2783
|
+
cacheRead: number;
|
|
2784
|
+
cacheWrite: number;
|
|
2785
|
+
};
|
|
2786
|
+
contextWindow: number;
|
|
2787
|
+
maxTokens: number;
|
|
2788
|
+
};
|
|
2681
2789
|
readonly "claude-sonnet-4": {
|
|
2682
2790
|
id: string;
|
|
2683
2791
|
name: string;
|
|
@@ -3623,7 +3731,7 @@ export declare const MODELS: {
|
|
|
3623
3731
|
contextWindow: number;
|
|
3624
3732
|
maxTokens: number;
|
|
3625
3733
|
};
|
|
3626
|
-
readonly "gemma-4-26b": {
|
|
3734
|
+
readonly "gemma-4-26b-a4b-it": {
|
|
3627
3735
|
id: string;
|
|
3628
3736
|
name: string;
|
|
3629
3737
|
api: "google-generative-ai";
|
|
@@ -3640,7 +3748,7 @@ export declare const MODELS: {
|
|
|
3640
3748
|
contextWindow: number;
|
|
3641
3749
|
maxTokens: number;
|
|
3642
3750
|
};
|
|
3643
|
-
readonly "gemma-4-31b": {
|
|
3751
|
+
readonly "gemma-4-31b-it": {
|
|
3644
3752
|
id: string;
|
|
3645
3753
|
name: string;
|
|
3646
3754
|
api: "google-generative-ai";
|
|
@@ -4122,6 +4230,23 @@ export declare const MODELS: {
|
|
|
4122
4230
|
contextWindow: number;
|
|
4123
4231
|
maxTokens: number;
|
|
4124
4232
|
};
|
|
4233
|
+
readonly "gemini-3.1-pro-preview-customtools": {
|
|
4234
|
+
id: string;
|
|
4235
|
+
name: string;
|
|
4236
|
+
api: "google-vertex";
|
|
4237
|
+
provider: string;
|
|
4238
|
+
baseUrl: string;
|
|
4239
|
+
reasoning: true;
|
|
4240
|
+
input: ("image" | "text")[];
|
|
4241
|
+
cost: {
|
|
4242
|
+
input: number;
|
|
4243
|
+
output: number;
|
|
4244
|
+
cacheRead: number;
|
|
4245
|
+
cacheWrite: number;
|
|
4246
|
+
};
|
|
4247
|
+
contextWindow: number;
|
|
4248
|
+
maxTokens: number;
|
|
4249
|
+
};
|
|
4125
4250
|
};
|
|
4126
4251
|
readonly groq: {
|
|
4127
4252
|
readonly "deepseek-r1-distill-llama-70b": {
|
|
@@ -4472,6 +4597,26 @@ export declare const MODELS: {
|
|
|
4472
4597
|
contextWindow: number;
|
|
4473
4598
|
maxTokens: number;
|
|
4474
4599
|
};
|
|
4600
|
+
readonly "MiniMaxAI/MiniMax-M2.7": {
|
|
4601
|
+
id: string;
|
|
4602
|
+
name: string;
|
|
4603
|
+
api: "openai-completions";
|
|
4604
|
+
provider: string;
|
|
4605
|
+
baseUrl: string;
|
|
4606
|
+
compat: {
|
|
4607
|
+
supportsDeveloperRole: false;
|
|
4608
|
+
};
|
|
4609
|
+
reasoning: true;
|
|
4610
|
+
input: "text"[];
|
|
4611
|
+
cost: {
|
|
4612
|
+
input: number;
|
|
4613
|
+
output: number;
|
|
4614
|
+
cacheRead: number;
|
|
4615
|
+
cacheWrite: number;
|
|
4616
|
+
};
|
|
4617
|
+
contextWindow: number;
|
|
4618
|
+
maxTokens: number;
|
|
4619
|
+
};
|
|
4475
4620
|
readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4476
4621
|
id: string;
|
|
4477
4622
|
name: string;
|
|
@@ -4732,7 +4877,7 @@ export declare const MODELS: {
|
|
|
4732
4877
|
contextWindow: number;
|
|
4733
4878
|
maxTokens: number;
|
|
4734
4879
|
};
|
|
4735
|
-
readonly "
|
|
4880
|
+
readonly "moonshotai/Kimi-K2.6": {
|
|
4736
4881
|
id: string;
|
|
4737
4882
|
name: string;
|
|
4738
4883
|
api: "openai-completions";
|
|
@@ -4742,7 +4887,7 @@ export declare const MODELS: {
|
|
|
4742
4887
|
supportsDeveloperRole: false;
|
|
4743
4888
|
};
|
|
4744
4889
|
reasoning: true;
|
|
4745
|
-
input: "text"[];
|
|
4890
|
+
input: ("image" | "text")[];
|
|
4746
4891
|
cost: {
|
|
4747
4892
|
input: number;
|
|
4748
4893
|
output: number;
|
|
@@ -4752,7 +4897,7 @@ export declare const MODELS: {
|
|
|
4752
4897
|
contextWindow: number;
|
|
4753
4898
|
maxTokens: number;
|
|
4754
4899
|
};
|
|
4755
|
-
readonly "zai-org/GLM-4.7
|
|
4900
|
+
readonly "zai-org/GLM-4.7": {
|
|
4756
4901
|
id: string;
|
|
4757
4902
|
name: string;
|
|
4758
4903
|
api: "openai-completions";
|
|
@@ -4772,7 +4917,7 @@ export declare const MODELS: {
|
|
|
4772
4917
|
contextWindow: number;
|
|
4773
4918
|
maxTokens: number;
|
|
4774
4919
|
};
|
|
4775
|
-
readonly "zai-org/GLM-
|
|
4920
|
+
readonly "zai-org/GLM-4.7-Flash": {
|
|
4776
4921
|
id: string;
|
|
4777
4922
|
name: string;
|
|
4778
4923
|
api: "openai-completions";
|
|
@@ -4792,16 +4937,17 @@ export declare const MODELS: {
|
|
|
4792
4937
|
contextWindow: number;
|
|
4793
4938
|
maxTokens: number;
|
|
4794
4939
|
};
|
|
4795
|
-
|
|
4796
|
-
readonly "kimi-coding": {
|
|
4797
|
-
readonly k2p5: {
|
|
4940
|
+
readonly "zai-org/GLM-5": {
|
|
4798
4941
|
id: string;
|
|
4799
4942
|
name: string;
|
|
4800
|
-
api: "
|
|
4943
|
+
api: "openai-completions";
|
|
4801
4944
|
provider: string;
|
|
4802
4945
|
baseUrl: string;
|
|
4946
|
+
compat: {
|
|
4947
|
+
supportsDeveloperRole: false;
|
|
4948
|
+
};
|
|
4803
4949
|
reasoning: true;
|
|
4804
|
-
input:
|
|
4950
|
+
input: "text"[];
|
|
4805
4951
|
cost: {
|
|
4806
4952
|
input: number;
|
|
4807
4953
|
output: number;
|
|
@@ -4811,12 +4957,15 @@ export declare const MODELS: {
|
|
|
4811
4957
|
contextWindow: number;
|
|
4812
4958
|
maxTokens: number;
|
|
4813
4959
|
};
|
|
4814
|
-
readonly "
|
|
4960
|
+
readonly "zai-org/GLM-5.1": {
|
|
4815
4961
|
id: string;
|
|
4816
4962
|
name: string;
|
|
4817
|
-
api: "
|
|
4963
|
+
api: "openai-completions";
|
|
4818
4964
|
provider: string;
|
|
4819
4965
|
baseUrl: string;
|
|
4966
|
+
compat: {
|
|
4967
|
+
supportsDeveloperRole: false;
|
|
4968
|
+
};
|
|
4820
4969
|
reasoning: true;
|
|
4821
4970
|
input: "text"[];
|
|
4822
4971
|
cost: {
|
|
@@ -4829,15 +4978,15 @@ export declare const MODELS: {
|
|
|
4829
4978
|
maxTokens: number;
|
|
4830
4979
|
};
|
|
4831
4980
|
};
|
|
4832
|
-
readonly
|
|
4833
|
-
readonly
|
|
4981
|
+
readonly "kimi-coding": {
|
|
4982
|
+
readonly k2p5: {
|
|
4834
4983
|
id: string;
|
|
4835
4984
|
name: string;
|
|
4836
4985
|
api: "anthropic-messages";
|
|
4837
4986
|
provider: string;
|
|
4838
4987
|
baseUrl: string;
|
|
4839
4988
|
reasoning: true;
|
|
4840
|
-
input: "text"[];
|
|
4989
|
+
input: ("image" | "text")[];
|
|
4841
4990
|
cost: {
|
|
4842
4991
|
input: number;
|
|
4843
4992
|
output: number;
|
|
@@ -4847,14 +4996,14 @@ export declare const MODELS: {
|
|
|
4847
4996
|
contextWindow: number;
|
|
4848
4997
|
maxTokens: number;
|
|
4849
4998
|
};
|
|
4850
|
-
readonly
|
|
4999
|
+
readonly k2p6: {
|
|
4851
5000
|
id: string;
|
|
4852
5001
|
name: string;
|
|
4853
5002
|
api: "anthropic-messages";
|
|
4854
5003
|
provider: string;
|
|
4855
5004
|
baseUrl: string;
|
|
4856
5005
|
reasoning: true;
|
|
4857
|
-
input: "text"[];
|
|
5006
|
+
input: ("image" | "text")[];
|
|
4858
5007
|
cost: {
|
|
4859
5008
|
input: number;
|
|
4860
5009
|
output: number;
|
|
@@ -4864,7 +5013,7 @@ export declare const MODELS: {
|
|
|
4864
5013
|
contextWindow: number;
|
|
4865
5014
|
maxTokens: number;
|
|
4866
5015
|
};
|
|
4867
|
-
readonly "
|
|
5016
|
+
readonly "kimi-k2-thinking": {
|
|
4868
5017
|
id: string;
|
|
4869
5018
|
name: string;
|
|
4870
5019
|
api: "anthropic-messages";
|
|
@@ -4881,7 +5030,9 @@ export declare const MODELS: {
|
|
|
4881
5030
|
contextWindow: number;
|
|
4882
5031
|
maxTokens: number;
|
|
4883
5032
|
};
|
|
4884
|
-
|
|
5033
|
+
};
|
|
5034
|
+
readonly minimax: {
|
|
5035
|
+
readonly "MiniMax-M2.7": {
|
|
4885
5036
|
id: string;
|
|
4886
5037
|
name: string;
|
|
4887
5038
|
api: "anthropic-messages";
|
|
@@ -4898,7 +5049,7 @@ export declare const MODELS: {
|
|
|
4898
5049
|
contextWindow: number;
|
|
4899
5050
|
maxTokens: number;
|
|
4900
5051
|
};
|
|
4901
|
-
readonly "MiniMax-M2.
|
|
5052
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4902
5053
|
id: string;
|
|
4903
5054
|
name: string;
|
|
4904
5055
|
api: "anthropic-messages";
|
|
@@ -4915,6 +5066,8 @@ export declare const MODELS: {
|
|
|
4915
5066
|
contextWindow: number;
|
|
4916
5067
|
maxTokens: number;
|
|
4917
5068
|
};
|
|
5069
|
+
};
|
|
5070
|
+
readonly "minimax-cn": {
|
|
4918
5071
|
readonly "MiniMax-M2.7": {
|
|
4919
5072
|
id: string;
|
|
4920
5073
|
name: string;
|
|
@@ -4950,14 +5103,14 @@ export declare const MODELS: {
|
|
|
4950
5103
|
maxTokens: number;
|
|
4951
5104
|
};
|
|
4952
5105
|
};
|
|
4953
|
-
readonly
|
|
4954
|
-
readonly "
|
|
5106
|
+
readonly mistral: {
|
|
5107
|
+
readonly "codestral-latest": {
|
|
4955
5108
|
id: string;
|
|
4956
5109
|
name: string;
|
|
4957
|
-
api: "
|
|
5110
|
+
api: "openai-completions";
|
|
4958
5111
|
provider: string;
|
|
4959
5112
|
baseUrl: string;
|
|
4960
|
-
reasoning:
|
|
5113
|
+
reasoning: false;
|
|
4961
5114
|
input: "text"[];
|
|
4962
5115
|
cost: {
|
|
4963
5116
|
input: number;
|
|
@@ -4968,13 +5121,13 @@ export declare const MODELS: {
|
|
|
4968
5121
|
contextWindow: number;
|
|
4969
5122
|
maxTokens: number;
|
|
4970
5123
|
};
|
|
4971
|
-
readonly "
|
|
5124
|
+
readonly "devstral-2512": {
|
|
4972
5125
|
id: string;
|
|
4973
5126
|
name: string;
|
|
4974
|
-
api: "
|
|
5127
|
+
api: "openai-completions";
|
|
4975
5128
|
provider: string;
|
|
4976
5129
|
baseUrl: string;
|
|
4977
|
-
reasoning:
|
|
5130
|
+
reasoning: false;
|
|
4978
5131
|
input: "text"[];
|
|
4979
5132
|
cost: {
|
|
4980
5133
|
input: number;
|
|
@@ -4985,13 +5138,13 @@ export declare const MODELS: {
|
|
|
4985
5138
|
contextWindow: number;
|
|
4986
5139
|
maxTokens: number;
|
|
4987
5140
|
};
|
|
4988
|
-
readonly "
|
|
5141
|
+
readonly "devstral-medium-2507": {
|
|
4989
5142
|
id: string;
|
|
4990
5143
|
name: string;
|
|
4991
|
-
api: "
|
|
5144
|
+
api: "openai-completions";
|
|
4992
5145
|
provider: string;
|
|
4993
5146
|
baseUrl: string;
|
|
4994
|
-
reasoning:
|
|
5147
|
+
reasoning: false;
|
|
4995
5148
|
input: "text"[];
|
|
4996
5149
|
cost: {
|
|
4997
5150
|
input: number;
|
|
@@ -5002,13 +5155,13 @@ export declare const MODELS: {
|
|
|
5002
5155
|
contextWindow: number;
|
|
5003
5156
|
maxTokens: number;
|
|
5004
5157
|
};
|
|
5005
|
-
readonly "
|
|
5158
|
+
readonly "devstral-medium-latest": {
|
|
5006
5159
|
id: string;
|
|
5007
5160
|
name: string;
|
|
5008
|
-
api: "
|
|
5161
|
+
api: "openai-completions";
|
|
5009
5162
|
provider: string;
|
|
5010
5163
|
baseUrl: string;
|
|
5011
|
-
reasoning:
|
|
5164
|
+
reasoning: false;
|
|
5012
5165
|
input: "text"[];
|
|
5013
5166
|
cost: {
|
|
5014
5167
|
input: number;
|
|
@@ -5019,13 +5172,13 @@ export declare const MODELS: {
|
|
|
5019
5172
|
contextWindow: number;
|
|
5020
5173
|
maxTokens: number;
|
|
5021
5174
|
};
|
|
5022
|
-
readonly "
|
|
5175
|
+
readonly "devstral-small-2505": {
|
|
5023
5176
|
id: string;
|
|
5024
5177
|
name: string;
|
|
5025
|
-
api: "
|
|
5178
|
+
api: "openai-completions";
|
|
5026
5179
|
provider: string;
|
|
5027
5180
|
baseUrl: string;
|
|
5028
|
-
reasoning:
|
|
5181
|
+
reasoning: false;
|
|
5029
5182
|
input: "text"[];
|
|
5030
5183
|
cost: {
|
|
5031
5184
|
input: number;
|
|
@@ -5036,13 +5189,13 @@ export declare const MODELS: {
|
|
|
5036
5189
|
contextWindow: number;
|
|
5037
5190
|
maxTokens: number;
|
|
5038
5191
|
};
|
|
5039
|
-
readonly "
|
|
5192
|
+
readonly "devstral-small-2507": {
|
|
5040
5193
|
id: string;
|
|
5041
5194
|
name: string;
|
|
5042
|
-
api: "
|
|
5195
|
+
api: "openai-completions";
|
|
5043
5196
|
provider: string;
|
|
5044
5197
|
baseUrl: string;
|
|
5045
|
-
reasoning:
|
|
5198
|
+
reasoning: false;
|
|
5046
5199
|
input: "text"[];
|
|
5047
5200
|
cost: {
|
|
5048
5201
|
input: number;
|
|
@@ -5053,14 +5206,14 @@ export declare const MODELS: {
|
|
|
5053
5206
|
contextWindow: number;
|
|
5054
5207
|
maxTokens: number;
|
|
5055
5208
|
};
|
|
5056
|
-
readonly "
|
|
5209
|
+
readonly "labs-devstral-small-2512": {
|
|
5057
5210
|
id: string;
|
|
5058
5211
|
name: string;
|
|
5059
|
-
api: "
|
|
5212
|
+
api: "openai-completions";
|
|
5060
5213
|
provider: string;
|
|
5061
5214
|
baseUrl: string;
|
|
5062
|
-
reasoning:
|
|
5063
|
-
input: "text"[];
|
|
5215
|
+
reasoning: false;
|
|
5216
|
+
input: ("image" | "text")[];
|
|
5064
5217
|
cost: {
|
|
5065
5218
|
input: number;
|
|
5066
5219
|
output: number;
|
|
@@ -5070,15 +5223,13 @@ export declare const MODELS: {
|
|
|
5070
5223
|
contextWindow: number;
|
|
5071
5224
|
maxTokens: number;
|
|
5072
5225
|
};
|
|
5073
|
-
|
|
5074
|
-
readonly mistral: {
|
|
5075
|
-
readonly "codestral-latest": {
|
|
5226
|
+
readonly "magistral-medium-latest": {
|
|
5076
5227
|
id: string;
|
|
5077
5228
|
name: string;
|
|
5078
|
-
api: "
|
|
5229
|
+
api: "openai-completions";
|
|
5079
5230
|
provider: string;
|
|
5080
5231
|
baseUrl: string;
|
|
5081
|
-
reasoning:
|
|
5232
|
+
reasoning: true;
|
|
5082
5233
|
input: "text"[];
|
|
5083
5234
|
cost: {
|
|
5084
5235
|
input: number;
|
|
@@ -5089,13 +5240,13 @@ export declare const MODELS: {
|
|
|
5089
5240
|
contextWindow: number;
|
|
5090
5241
|
maxTokens: number;
|
|
5091
5242
|
};
|
|
5092
|
-
readonly "
|
|
5243
|
+
readonly "magistral-small": {
|
|
5093
5244
|
id: string;
|
|
5094
5245
|
name: string;
|
|
5095
|
-
api: "
|
|
5246
|
+
api: "openai-completions";
|
|
5096
5247
|
provider: string;
|
|
5097
5248
|
baseUrl: string;
|
|
5098
|
-
reasoning:
|
|
5249
|
+
reasoning: true;
|
|
5099
5250
|
input: "text"[];
|
|
5100
5251
|
cost: {
|
|
5101
5252
|
input: number;
|
|
@@ -5106,10 +5257,10 @@ export declare const MODELS: {
|
|
|
5106
5257
|
contextWindow: number;
|
|
5107
5258
|
maxTokens: number;
|
|
5108
5259
|
};
|
|
5109
|
-
readonly "
|
|
5260
|
+
readonly "ministral-3b-latest": {
|
|
5110
5261
|
id: string;
|
|
5111
5262
|
name: string;
|
|
5112
|
-
api: "
|
|
5263
|
+
api: "openai-completions";
|
|
5113
5264
|
provider: string;
|
|
5114
5265
|
baseUrl: string;
|
|
5115
5266
|
reasoning: false;
|
|
@@ -5123,10 +5274,10 @@ export declare const MODELS: {
|
|
|
5123
5274
|
contextWindow: number;
|
|
5124
5275
|
maxTokens: number;
|
|
5125
5276
|
};
|
|
5126
|
-
readonly "
|
|
5277
|
+
readonly "ministral-8b-latest": {
|
|
5127
5278
|
id: string;
|
|
5128
5279
|
name: string;
|
|
5129
|
-
api: "
|
|
5280
|
+
api: "openai-completions";
|
|
5130
5281
|
provider: string;
|
|
5131
5282
|
baseUrl: string;
|
|
5132
5283
|
reasoning: false;
|
|
@@ -5140,129 +5291,10 @@ export declare const MODELS: {
|
|
|
5140
5291
|
contextWindow: number;
|
|
5141
5292
|
maxTokens: number;
|
|
5142
5293
|
};
|
|
5143
|
-
readonly "
|
|
5294
|
+
readonly "mistral-large-2411": {
|
|
5144
5295
|
id: string;
|
|
5145
5296
|
name: string;
|
|
5146
|
-
api: "
|
|
5147
|
-
provider: string;
|
|
5148
|
-
baseUrl: string;
|
|
5149
|
-
reasoning: false;
|
|
5150
|
-
input: "text"[];
|
|
5151
|
-
cost: {
|
|
5152
|
-
input: number;
|
|
5153
|
-
output: number;
|
|
5154
|
-
cacheRead: number;
|
|
5155
|
-
cacheWrite: number;
|
|
5156
|
-
};
|
|
5157
|
-
contextWindow: number;
|
|
5158
|
-
maxTokens: number;
|
|
5159
|
-
};
|
|
5160
|
-
readonly "devstral-small-2507": {
|
|
5161
|
-
id: string;
|
|
5162
|
-
name: string;
|
|
5163
|
-
api: "mistral-conversations";
|
|
5164
|
-
provider: string;
|
|
5165
|
-
baseUrl: string;
|
|
5166
|
-
reasoning: false;
|
|
5167
|
-
input: "text"[];
|
|
5168
|
-
cost: {
|
|
5169
|
-
input: number;
|
|
5170
|
-
output: number;
|
|
5171
|
-
cacheRead: number;
|
|
5172
|
-
cacheWrite: number;
|
|
5173
|
-
};
|
|
5174
|
-
contextWindow: number;
|
|
5175
|
-
maxTokens: number;
|
|
5176
|
-
};
|
|
5177
|
-
readonly "labs-devstral-small-2512": {
|
|
5178
|
-
id: string;
|
|
5179
|
-
name: string;
|
|
5180
|
-
api: "mistral-conversations";
|
|
5181
|
-
provider: string;
|
|
5182
|
-
baseUrl: string;
|
|
5183
|
-
reasoning: false;
|
|
5184
|
-
input: ("image" | "text")[];
|
|
5185
|
-
cost: {
|
|
5186
|
-
input: number;
|
|
5187
|
-
output: number;
|
|
5188
|
-
cacheRead: number;
|
|
5189
|
-
cacheWrite: number;
|
|
5190
|
-
};
|
|
5191
|
-
contextWindow: number;
|
|
5192
|
-
maxTokens: number;
|
|
5193
|
-
};
|
|
5194
|
-
readonly "magistral-medium-latest": {
|
|
5195
|
-
id: string;
|
|
5196
|
-
name: string;
|
|
5197
|
-
api: "mistral-conversations";
|
|
5198
|
-
provider: string;
|
|
5199
|
-
baseUrl: string;
|
|
5200
|
-
reasoning: true;
|
|
5201
|
-
input: "text"[];
|
|
5202
|
-
cost: {
|
|
5203
|
-
input: number;
|
|
5204
|
-
output: number;
|
|
5205
|
-
cacheRead: number;
|
|
5206
|
-
cacheWrite: number;
|
|
5207
|
-
};
|
|
5208
|
-
contextWindow: number;
|
|
5209
|
-
maxTokens: number;
|
|
5210
|
-
};
|
|
5211
|
-
readonly "magistral-small": {
|
|
5212
|
-
id: string;
|
|
5213
|
-
name: string;
|
|
5214
|
-
api: "mistral-conversations";
|
|
5215
|
-
provider: string;
|
|
5216
|
-
baseUrl: string;
|
|
5217
|
-
reasoning: true;
|
|
5218
|
-
input: "text"[];
|
|
5219
|
-
cost: {
|
|
5220
|
-
input: number;
|
|
5221
|
-
output: number;
|
|
5222
|
-
cacheRead: number;
|
|
5223
|
-
cacheWrite: number;
|
|
5224
|
-
};
|
|
5225
|
-
contextWindow: number;
|
|
5226
|
-
maxTokens: number;
|
|
5227
|
-
};
|
|
5228
|
-
readonly "ministral-3b-latest": {
|
|
5229
|
-
id: string;
|
|
5230
|
-
name: string;
|
|
5231
|
-
api: "mistral-conversations";
|
|
5232
|
-
provider: string;
|
|
5233
|
-
baseUrl: string;
|
|
5234
|
-
reasoning: false;
|
|
5235
|
-
input: "text"[];
|
|
5236
|
-
cost: {
|
|
5237
|
-
input: number;
|
|
5238
|
-
output: number;
|
|
5239
|
-
cacheRead: number;
|
|
5240
|
-
cacheWrite: number;
|
|
5241
|
-
};
|
|
5242
|
-
contextWindow: number;
|
|
5243
|
-
maxTokens: number;
|
|
5244
|
-
};
|
|
5245
|
-
readonly "ministral-8b-latest": {
|
|
5246
|
-
id: string;
|
|
5247
|
-
name: string;
|
|
5248
|
-
api: "mistral-conversations";
|
|
5249
|
-
provider: string;
|
|
5250
|
-
baseUrl: string;
|
|
5251
|
-
reasoning: false;
|
|
5252
|
-
input: "text"[];
|
|
5253
|
-
cost: {
|
|
5254
|
-
input: number;
|
|
5255
|
-
output: number;
|
|
5256
|
-
cacheRead: number;
|
|
5257
|
-
cacheWrite: number;
|
|
5258
|
-
};
|
|
5259
|
-
contextWindow: number;
|
|
5260
|
-
maxTokens: number;
|
|
5261
|
-
};
|
|
5262
|
-
readonly "mistral-large-2411": {
|
|
5263
|
-
id: string;
|
|
5264
|
-
name: string;
|
|
5265
|
-
api: "mistral-conversations";
|
|
5297
|
+
api: "openai-completions";
|
|
5266
5298
|
provider: string;
|
|
5267
5299
|
baseUrl: string;
|
|
5268
5300
|
reasoning: false;
|
|
@@ -5279,7 +5311,7 @@ export declare const MODELS: {
|
|
|
5279
5311
|
readonly "mistral-large-2512": {
|
|
5280
5312
|
id: string;
|
|
5281
5313
|
name: string;
|
|
5282
|
-
api: "
|
|
5314
|
+
api: "openai-completions";
|
|
5283
5315
|
provider: string;
|
|
5284
5316
|
baseUrl: string;
|
|
5285
5317
|
reasoning: false;
|
|
@@ -5296,7 +5328,7 @@ export declare const MODELS: {
|
|
|
5296
5328
|
readonly "mistral-large-latest": {
|
|
5297
5329
|
id: string;
|
|
5298
5330
|
name: string;
|
|
5299
|
-
api: "
|
|
5331
|
+
api: "openai-completions";
|
|
5300
5332
|
provider: string;
|
|
5301
5333
|
baseUrl: string;
|
|
5302
5334
|
reasoning: false;
|
|
@@ -5313,7 +5345,7 @@ export declare const MODELS: {
|
|
|
5313
5345
|
readonly "mistral-medium-2505": {
|
|
5314
5346
|
id: string;
|
|
5315
5347
|
name: string;
|
|
5316
|
-
api: "
|
|
5348
|
+
api: "openai-completions";
|
|
5317
5349
|
provider: string;
|
|
5318
5350
|
baseUrl: string;
|
|
5319
5351
|
reasoning: false;
|
|
@@ -5330,7 +5362,7 @@ export declare const MODELS: {
|
|
|
5330
5362
|
readonly "mistral-medium-2508": {
|
|
5331
5363
|
id: string;
|
|
5332
5364
|
name: string;
|
|
5333
|
-
api: "
|
|
5365
|
+
api: "openai-completions";
|
|
5334
5366
|
provider: string;
|
|
5335
5367
|
baseUrl: string;
|
|
5336
5368
|
reasoning: false;
|
|
@@ -5347,7 +5379,7 @@ export declare const MODELS: {
|
|
|
5347
5379
|
readonly "mistral-medium-latest": {
|
|
5348
5380
|
id: string;
|
|
5349
5381
|
name: string;
|
|
5350
|
-
api: "
|
|
5382
|
+
api: "openai-completions";
|
|
5351
5383
|
provider: string;
|
|
5352
5384
|
baseUrl: string;
|
|
5353
5385
|
reasoning: false;
|
|
@@ -5364,7 +5396,7 @@ export declare const MODELS: {
|
|
|
5364
5396
|
readonly "mistral-nemo": {
|
|
5365
5397
|
id: string;
|
|
5366
5398
|
name: string;
|
|
5367
|
-
api: "
|
|
5399
|
+
api: "openai-completions";
|
|
5368
5400
|
provider: string;
|
|
5369
5401
|
baseUrl: string;
|
|
5370
5402
|
reasoning: false;
|
|
@@ -5381,7 +5413,7 @@ export declare const MODELS: {
|
|
|
5381
5413
|
readonly "mistral-small-2506": {
|
|
5382
5414
|
id: string;
|
|
5383
5415
|
name: string;
|
|
5384
|
-
api: "
|
|
5416
|
+
api: "openai-completions";
|
|
5385
5417
|
provider: string;
|
|
5386
5418
|
baseUrl: string;
|
|
5387
5419
|
reasoning: false;
|
|
@@ -5398,7 +5430,7 @@ export declare const MODELS: {
|
|
|
5398
5430
|
readonly "mistral-small-2603": {
|
|
5399
5431
|
id: string;
|
|
5400
5432
|
name: string;
|
|
5401
|
-
api: "
|
|
5433
|
+
api: "openai-completions";
|
|
5402
5434
|
provider: string;
|
|
5403
5435
|
baseUrl: string;
|
|
5404
5436
|
reasoning: true;
|
|
@@ -5415,7 +5447,7 @@ export declare const MODELS: {
|
|
|
5415
5447
|
readonly "mistral-small-latest": {
|
|
5416
5448
|
id: string;
|
|
5417
5449
|
name: string;
|
|
5418
|
-
api: "
|
|
5450
|
+
api: "openai-completions";
|
|
5419
5451
|
provider: string;
|
|
5420
5452
|
baseUrl: string;
|
|
5421
5453
|
reasoning: true;
|
|
@@ -5432,7 +5464,7 @@ export declare const MODELS: {
|
|
|
5432
5464
|
readonly "open-mistral-7b": {
|
|
5433
5465
|
id: string;
|
|
5434
5466
|
name: string;
|
|
5435
|
-
api: "
|
|
5467
|
+
api: "openai-completions";
|
|
5436
5468
|
provider: string;
|
|
5437
5469
|
baseUrl: string;
|
|
5438
5470
|
reasoning: false;
|
|
@@ -5449,7 +5481,7 @@ export declare const MODELS: {
|
|
|
5449
5481
|
readonly "open-mixtral-8x22b": {
|
|
5450
5482
|
id: string;
|
|
5451
5483
|
name: string;
|
|
5452
|
-
api: "
|
|
5484
|
+
api: "openai-completions";
|
|
5453
5485
|
provider: string;
|
|
5454
5486
|
baseUrl: string;
|
|
5455
5487
|
reasoning: false;
|
|
@@ -5466,7 +5498,7 @@ export declare const MODELS: {
|
|
|
5466
5498
|
readonly "open-mixtral-8x7b": {
|
|
5467
5499
|
id: string;
|
|
5468
5500
|
name: string;
|
|
5469
|
-
api: "
|
|
5501
|
+
api: "openai-completions";
|
|
5470
5502
|
provider: string;
|
|
5471
5503
|
baseUrl: string;
|
|
5472
5504
|
reasoning: false;
|
|
@@ -5483,7 +5515,7 @@ export declare const MODELS: {
|
|
|
5483
5515
|
readonly "pixtral-12b": {
|
|
5484
5516
|
id: string;
|
|
5485
5517
|
name: string;
|
|
5486
|
-
api: "
|
|
5518
|
+
api: "openai-completions";
|
|
5487
5519
|
provider: string;
|
|
5488
5520
|
baseUrl: string;
|
|
5489
5521
|
reasoning: false;
|
|
@@ -5500,7 +5532,7 @@ export declare const MODELS: {
|
|
|
5500
5532
|
readonly "pixtral-large-latest": {
|
|
5501
5533
|
id: string;
|
|
5502
5534
|
name: string;
|
|
5503
|
-
api: "
|
|
5535
|
+
api: "openai-completions";
|
|
5504
5536
|
provider: string;
|
|
5505
5537
|
baseUrl: string;
|
|
5506
5538
|
reasoning: false;
|
|
@@ -5516,23 +5548,6 @@ export declare const MODELS: {
|
|
|
5516
5548
|
};
|
|
5517
5549
|
};
|
|
5518
5550
|
readonly openai: {
|
|
5519
|
-
readonly "codex-mini-latest": {
|
|
5520
|
-
id: string;
|
|
5521
|
-
name: string;
|
|
5522
|
-
api: "openai-responses";
|
|
5523
|
-
provider: string;
|
|
5524
|
-
baseUrl: string;
|
|
5525
|
-
reasoning: true;
|
|
5526
|
-
input: "text"[];
|
|
5527
|
-
cost: {
|
|
5528
|
-
input: number;
|
|
5529
|
-
output: number;
|
|
5530
|
-
cacheRead: number;
|
|
5531
|
-
cacheWrite: number;
|
|
5532
|
-
};
|
|
5533
|
-
contextWindow: number;
|
|
5534
|
-
maxTokens: number;
|
|
5535
|
-
};
|
|
5536
5551
|
readonly "gpt-4": {
|
|
5537
5552
|
id: string;
|
|
5538
5553
|
name: string;
|
|
@@ -6472,6 +6487,23 @@ export declare const MODELS: {
|
|
|
6472
6487
|
contextWindow: number;
|
|
6473
6488
|
maxTokens: number;
|
|
6474
6489
|
};
|
|
6490
|
+
readonly "claude-opus-4-7": {
|
|
6491
|
+
id: string;
|
|
6492
|
+
name: string;
|
|
6493
|
+
api: "anthropic-messages";
|
|
6494
|
+
provider: string;
|
|
6495
|
+
baseUrl: string;
|
|
6496
|
+
reasoning: true;
|
|
6497
|
+
input: ("image" | "text")[];
|
|
6498
|
+
cost: {
|
|
6499
|
+
input: number;
|
|
6500
|
+
output: number;
|
|
6501
|
+
cacheRead: number;
|
|
6502
|
+
cacheWrite: number;
|
|
6503
|
+
};
|
|
6504
|
+
contextWindow: number;
|
|
6505
|
+
maxTokens: number;
|
|
6506
|
+
};
|
|
6475
6507
|
readonly "claude-sonnet-4": {
|
|
6476
6508
|
id: string;
|
|
6477
6509
|
name: string;
|
|
@@ -6574,6 +6606,23 @@ export declare const MODELS: {
|
|
|
6574
6606
|
contextWindow: number;
|
|
6575
6607
|
maxTokens: number;
|
|
6576
6608
|
};
|
|
6609
|
+
readonly "glm-5.1": {
|
|
6610
|
+
id: string;
|
|
6611
|
+
name: string;
|
|
6612
|
+
api: "openai-completions";
|
|
6613
|
+
provider: string;
|
|
6614
|
+
baseUrl: string;
|
|
6615
|
+
reasoning: true;
|
|
6616
|
+
input: "text"[];
|
|
6617
|
+
cost: {
|
|
6618
|
+
input: number;
|
|
6619
|
+
output: number;
|
|
6620
|
+
cacheRead: number;
|
|
6621
|
+
cacheWrite: number;
|
|
6622
|
+
};
|
|
6623
|
+
contextWindow: number;
|
|
6624
|
+
maxTokens: number;
|
|
6625
|
+
};
|
|
6577
6626
|
readonly "gpt-5": {
|
|
6578
6627
|
id: string;
|
|
6579
6628
|
name: string;
|
|
@@ -6812,6 +6861,23 @@ export declare const MODELS: {
|
|
|
6812
6861
|
contextWindow: number;
|
|
6813
6862
|
maxTokens: number;
|
|
6814
6863
|
};
|
|
6864
|
+
readonly "hy3-preview-free": {
|
|
6865
|
+
id: string;
|
|
6866
|
+
name: string;
|
|
6867
|
+
api: "openai-completions";
|
|
6868
|
+
provider: string;
|
|
6869
|
+
baseUrl: string;
|
|
6870
|
+
reasoning: true;
|
|
6871
|
+
input: "text"[];
|
|
6872
|
+
cost: {
|
|
6873
|
+
input: number;
|
|
6874
|
+
output: number;
|
|
6875
|
+
cacheRead: number;
|
|
6876
|
+
cacheWrite: number;
|
|
6877
|
+
};
|
|
6878
|
+
contextWindow: number;
|
|
6879
|
+
maxTokens: number;
|
|
6880
|
+
};
|
|
6815
6881
|
readonly "kimi-k2.5": {
|
|
6816
6882
|
id: string;
|
|
6817
6883
|
name: string;
|
|
@@ -6829,6 +6895,40 @@ export declare const MODELS: {
|
|
|
6829
6895
|
contextWindow: number;
|
|
6830
6896
|
maxTokens: number;
|
|
6831
6897
|
};
|
|
6898
|
+
readonly "kimi-k2.6": {
|
|
6899
|
+
id: string;
|
|
6900
|
+
name: string;
|
|
6901
|
+
api: "openai-completions";
|
|
6902
|
+
provider: string;
|
|
6903
|
+
baseUrl: string;
|
|
6904
|
+
reasoning: true;
|
|
6905
|
+
input: ("image" | "text")[];
|
|
6906
|
+
cost: {
|
|
6907
|
+
input: number;
|
|
6908
|
+
output: number;
|
|
6909
|
+
cacheRead: number;
|
|
6910
|
+
cacheWrite: number;
|
|
6911
|
+
};
|
|
6912
|
+
contextWindow: number;
|
|
6913
|
+
maxTokens: number;
|
|
6914
|
+
};
|
|
6915
|
+
readonly "ling-2.6-flash-free": {
|
|
6916
|
+
id: string;
|
|
6917
|
+
name: string;
|
|
6918
|
+
api: "openai-completions";
|
|
6919
|
+
provider: string;
|
|
6920
|
+
baseUrl: string;
|
|
6921
|
+
reasoning: false;
|
|
6922
|
+
input: "text"[];
|
|
6923
|
+
cost: {
|
|
6924
|
+
input: number;
|
|
6925
|
+
output: number;
|
|
6926
|
+
cacheRead: number;
|
|
6927
|
+
cacheWrite: number;
|
|
6928
|
+
};
|
|
6929
|
+
contextWindow: number;
|
|
6930
|
+
maxTokens: number;
|
|
6931
|
+
};
|
|
6832
6932
|
readonly "minimax-m2.5": {
|
|
6833
6933
|
id: string;
|
|
6834
6934
|
name: string;
|
|
@@ -6863,7 +6963,7 @@ export declare const MODELS: {
|
|
|
6863
6963
|
contextWindow: number;
|
|
6864
6964
|
maxTokens: number;
|
|
6865
6965
|
};
|
|
6866
|
-
readonly "
|
|
6966
|
+
readonly "minimax-m2.7": {
|
|
6867
6967
|
id: string;
|
|
6868
6968
|
name: string;
|
|
6869
6969
|
api: "openai-completions";
|
|
@@ -6880,7 +6980,7 @@ export declare const MODELS: {
|
|
|
6880
6980
|
contextWindow: number;
|
|
6881
6981
|
maxTokens: number;
|
|
6882
6982
|
};
|
|
6883
|
-
readonly "
|
|
6983
|
+
readonly "nemotron-3-super-free": {
|
|
6884
6984
|
id: string;
|
|
6885
6985
|
name: string;
|
|
6886
6986
|
api: "openai-completions";
|
|
@@ -6897,6 +6997,40 @@ export declare const MODELS: {
|
|
|
6897
6997
|
contextWindow: number;
|
|
6898
6998
|
maxTokens: number;
|
|
6899
6999
|
};
|
|
7000
|
+
readonly "qwen3.5-plus": {
|
|
7001
|
+
id: string;
|
|
7002
|
+
name: string;
|
|
7003
|
+
api: "anthropic-messages";
|
|
7004
|
+
provider: string;
|
|
7005
|
+
baseUrl: string;
|
|
7006
|
+
reasoning: true;
|
|
7007
|
+
input: ("image" | "text")[];
|
|
7008
|
+
cost: {
|
|
7009
|
+
input: number;
|
|
7010
|
+
output: number;
|
|
7011
|
+
cacheRead: number;
|
|
7012
|
+
cacheWrite: number;
|
|
7013
|
+
};
|
|
7014
|
+
contextWindow: number;
|
|
7015
|
+
maxTokens: number;
|
|
7016
|
+
};
|
|
7017
|
+
readonly "qwen3.6-plus": {
|
|
7018
|
+
id: string;
|
|
7019
|
+
name: string;
|
|
7020
|
+
api: "anthropic-messages";
|
|
7021
|
+
provider: string;
|
|
7022
|
+
baseUrl: string;
|
|
7023
|
+
reasoning: true;
|
|
7024
|
+
input: ("image" | "text")[];
|
|
7025
|
+
cost: {
|
|
7026
|
+
input: number;
|
|
7027
|
+
output: number;
|
|
7028
|
+
cacheRead: number;
|
|
7029
|
+
cacheWrite: number;
|
|
7030
|
+
};
|
|
7031
|
+
contextWindow: number;
|
|
7032
|
+
maxTokens: number;
|
|
7033
|
+
};
|
|
6900
7034
|
};
|
|
6901
7035
|
readonly "opencode-go": {
|
|
6902
7036
|
readonly "glm-5": {
|
|
@@ -6916,6 +7050,23 @@ export declare const MODELS: {
|
|
|
6916
7050
|
contextWindow: number;
|
|
6917
7051
|
maxTokens: number;
|
|
6918
7052
|
};
|
|
7053
|
+
readonly "glm-5.1": {
|
|
7054
|
+
id: string;
|
|
7055
|
+
name: string;
|
|
7056
|
+
api: "openai-completions";
|
|
7057
|
+
provider: string;
|
|
7058
|
+
baseUrl: string;
|
|
7059
|
+
reasoning: true;
|
|
7060
|
+
input: "text"[];
|
|
7061
|
+
cost: {
|
|
7062
|
+
input: number;
|
|
7063
|
+
output: number;
|
|
7064
|
+
cacheRead: number;
|
|
7065
|
+
cacheWrite: number;
|
|
7066
|
+
};
|
|
7067
|
+
contextWindow: number;
|
|
7068
|
+
maxTokens: number;
|
|
7069
|
+
};
|
|
6919
7070
|
readonly "kimi-k2.5": {
|
|
6920
7071
|
id: string;
|
|
6921
7072
|
name: string;
|
|
@@ -6933,6 +7084,23 @@ export declare const MODELS: {
|
|
|
6933
7084
|
contextWindow: number;
|
|
6934
7085
|
maxTokens: number;
|
|
6935
7086
|
};
|
|
7087
|
+
readonly "kimi-k2.6": {
|
|
7088
|
+
id: string;
|
|
7089
|
+
name: string;
|
|
7090
|
+
api: "openai-completions";
|
|
7091
|
+
provider: string;
|
|
7092
|
+
baseUrl: string;
|
|
7093
|
+
reasoning: true;
|
|
7094
|
+
input: ("image" | "text")[];
|
|
7095
|
+
cost: {
|
|
7096
|
+
input: number;
|
|
7097
|
+
output: number;
|
|
7098
|
+
cacheRead: number;
|
|
7099
|
+
cacheWrite: number;
|
|
7100
|
+
};
|
|
7101
|
+
contextWindow: number;
|
|
7102
|
+
maxTokens: number;
|
|
7103
|
+
};
|
|
6936
7104
|
readonly "mimo-v2-omni": {
|
|
6937
7105
|
id: string;
|
|
6938
7106
|
name: string;
|
|
@@ -6967,6 +7135,40 @@ export declare const MODELS: {
|
|
|
6967
7135
|
contextWindow: number;
|
|
6968
7136
|
maxTokens: number;
|
|
6969
7137
|
};
|
|
7138
|
+
readonly "mimo-v2.5": {
|
|
7139
|
+
id: string;
|
|
7140
|
+
name: string;
|
|
7141
|
+
api: "openai-completions";
|
|
7142
|
+
provider: string;
|
|
7143
|
+
baseUrl: string;
|
|
7144
|
+
reasoning: true;
|
|
7145
|
+
input: ("image" | "text")[];
|
|
7146
|
+
cost: {
|
|
7147
|
+
input: number;
|
|
7148
|
+
output: number;
|
|
7149
|
+
cacheRead: number;
|
|
7150
|
+
cacheWrite: number;
|
|
7151
|
+
};
|
|
7152
|
+
contextWindow: number;
|
|
7153
|
+
maxTokens: number;
|
|
7154
|
+
};
|
|
7155
|
+
readonly "mimo-v2.5-pro": {
|
|
7156
|
+
id: string;
|
|
7157
|
+
name: string;
|
|
7158
|
+
api: "openai-completions";
|
|
7159
|
+
provider: string;
|
|
7160
|
+
baseUrl: string;
|
|
7161
|
+
reasoning: true;
|
|
7162
|
+
input: "text"[];
|
|
7163
|
+
cost: {
|
|
7164
|
+
input: number;
|
|
7165
|
+
output: number;
|
|
7166
|
+
cacheRead: number;
|
|
7167
|
+
cacheWrite: number;
|
|
7168
|
+
};
|
|
7169
|
+
contextWindow: number;
|
|
7170
|
+
maxTokens: number;
|
|
7171
|
+
};
|
|
6970
7172
|
readonly "minimax-m2.5": {
|
|
6971
7173
|
id: string;
|
|
6972
7174
|
name: string;
|
|
@@ -7001,6 +7203,40 @@ export declare const MODELS: {
|
|
|
7001
7203
|
contextWindow: number;
|
|
7002
7204
|
maxTokens: number;
|
|
7003
7205
|
};
|
|
7206
|
+
readonly "qwen3.5-plus": {
|
|
7207
|
+
id: string;
|
|
7208
|
+
name: string;
|
|
7209
|
+
api: "anthropic-messages";
|
|
7210
|
+
provider: string;
|
|
7211
|
+
baseUrl: string;
|
|
7212
|
+
reasoning: true;
|
|
7213
|
+
input: ("image" | "text")[];
|
|
7214
|
+
cost: {
|
|
7215
|
+
input: number;
|
|
7216
|
+
output: number;
|
|
7217
|
+
cacheRead: number;
|
|
7218
|
+
cacheWrite: number;
|
|
7219
|
+
};
|
|
7220
|
+
contextWindow: number;
|
|
7221
|
+
maxTokens: number;
|
|
7222
|
+
};
|
|
7223
|
+
readonly "qwen3.6-plus": {
|
|
7224
|
+
id: string;
|
|
7225
|
+
name: string;
|
|
7226
|
+
api: "anthropic-messages";
|
|
7227
|
+
provider: string;
|
|
7228
|
+
baseUrl: string;
|
|
7229
|
+
reasoning: true;
|
|
7230
|
+
input: ("image" | "text")[];
|
|
7231
|
+
cost: {
|
|
7232
|
+
input: number;
|
|
7233
|
+
output: number;
|
|
7234
|
+
cacheRead: number;
|
|
7235
|
+
cacheWrite: number;
|
|
7236
|
+
};
|
|
7237
|
+
contextWindow: number;
|
|
7238
|
+
maxTokens: number;
|
|
7239
|
+
};
|
|
7004
7240
|
};
|
|
7005
7241
|
readonly openrouter: {
|
|
7006
7242
|
readonly "ai21/jamba-large-1.7": {
|
|
@@ -7292,7 +7528,7 @@ export declare const MODELS: {
|
|
|
7292
7528
|
contextWindow: number;
|
|
7293
7529
|
maxTokens: number;
|
|
7294
7530
|
};
|
|
7295
|
-
readonly "anthropic/claude-
|
|
7531
|
+
readonly "anthropic/claude-opus-4.6-fast": {
|
|
7296
7532
|
id: string;
|
|
7297
7533
|
name: string;
|
|
7298
7534
|
api: "openai-completions";
|
|
@@ -7309,7 +7545,7 @@ export declare const MODELS: {
|
|
|
7309
7545
|
contextWindow: number;
|
|
7310
7546
|
maxTokens: number;
|
|
7311
7547
|
};
|
|
7312
|
-
readonly "anthropic/claude-
|
|
7548
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
7313
7549
|
id: string;
|
|
7314
7550
|
name: string;
|
|
7315
7551
|
api: "openai-completions";
|
|
@@ -7326,7 +7562,7 @@ export declare const MODELS: {
|
|
|
7326
7562
|
contextWindow: number;
|
|
7327
7563
|
maxTokens: number;
|
|
7328
7564
|
};
|
|
7329
|
-
readonly "anthropic/claude-sonnet-4
|
|
7565
|
+
readonly "anthropic/claude-sonnet-4": {
|
|
7330
7566
|
id: string;
|
|
7331
7567
|
name: string;
|
|
7332
7568
|
api: "openai-completions";
|
|
@@ -7343,14 +7579,14 @@ export declare const MODELS: {
|
|
|
7343
7579
|
contextWindow: number;
|
|
7344
7580
|
maxTokens: number;
|
|
7345
7581
|
};
|
|
7346
|
-
readonly "
|
|
7582
|
+
readonly "anthropic/claude-sonnet-4.5": {
|
|
7347
7583
|
id: string;
|
|
7348
7584
|
name: string;
|
|
7349
7585
|
api: "openai-completions";
|
|
7350
7586
|
provider: string;
|
|
7351
7587
|
baseUrl: string;
|
|
7352
|
-
reasoning:
|
|
7353
|
-
input: "text"[];
|
|
7588
|
+
reasoning: true;
|
|
7589
|
+
input: ("image" | "text")[];
|
|
7354
7590
|
cost: {
|
|
7355
7591
|
input: number;
|
|
7356
7592
|
output: number;
|
|
@@ -7360,13 +7596,30 @@ export declare const MODELS: {
|
|
|
7360
7596
|
contextWindow: number;
|
|
7361
7597
|
maxTokens: number;
|
|
7362
7598
|
};
|
|
7363
|
-
readonly "
|
|
7599
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
7364
7600
|
id: string;
|
|
7365
7601
|
name: string;
|
|
7366
7602
|
api: "openai-completions";
|
|
7367
7603
|
provider: string;
|
|
7368
7604
|
baseUrl: string;
|
|
7369
7605
|
reasoning: true;
|
|
7606
|
+
input: ("image" | "text")[];
|
|
7607
|
+
cost: {
|
|
7608
|
+
input: number;
|
|
7609
|
+
output: number;
|
|
7610
|
+
cacheRead: number;
|
|
7611
|
+
cacheWrite: number;
|
|
7612
|
+
};
|
|
7613
|
+
contextWindow: number;
|
|
7614
|
+
maxTokens: number;
|
|
7615
|
+
};
|
|
7616
|
+
readonly "arcee-ai/trinity-large-preview": {
|
|
7617
|
+
id: string;
|
|
7618
|
+
name: string;
|
|
7619
|
+
api: "openai-completions";
|
|
7620
|
+
provider: string;
|
|
7621
|
+
baseUrl: string;
|
|
7622
|
+
reasoning: false;
|
|
7370
7623
|
input: "text"[];
|
|
7371
7624
|
cost: {
|
|
7372
7625
|
input: number;
|
|
@@ -7377,7 +7630,7 @@ export declare const MODELS: {
|
|
|
7377
7630
|
contextWindow: number;
|
|
7378
7631
|
maxTokens: number;
|
|
7379
7632
|
};
|
|
7380
|
-
readonly "arcee-ai/trinity-
|
|
7633
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
7381
7634
|
id: string;
|
|
7382
7635
|
name: string;
|
|
7383
7636
|
api: "openai-completions";
|
|
@@ -7394,7 +7647,7 @@ export declare const MODELS: {
|
|
|
7394
7647
|
contextWindow: number;
|
|
7395
7648
|
maxTokens: number;
|
|
7396
7649
|
};
|
|
7397
|
-
readonly "arcee-ai/trinity-mini
|
|
7650
|
+
readonly "arcee-ai/trinity-mini": {
|
|
7398
7651
|
id: string;
|
|
7399
7652
|
name: string;
|
|
7400
7653
|
api: "openai-completions";
|
|
@@ -7604,7 +7857,7 @@ export declare const MODELS: {
|
|
|
7604
7857
|
api: "openai-completions";
|
|
7605
7858
|
provider: string;
|
|
7606
7859
|
baseUrl: string;
|
|
7607
|
-
reasoning:
|
|
7860
|
+
reasoning: false;
|
|
7608
7861
|
input: "text"[];
|
|
7609
7862
|
cost: {
|
|
7610
7863
|
input: number;
|
|
@@ -7955,7 +8208,7 @@ export declare const MODELS: {
|
|
|
7955
8208
|
contextWindow: number;
|
|
7956
8209
|
maxTokens: number;
|
|
7957
8210
|
};
|
|
7958
|
-
readonly "google/gemma-4-
|
|
8211
|
+
readonly "google/gemma-4-26b-a4b-it:free": {
|
|
7959
8212
|
id: string;
|
|
7960
8213
|
name: string;
|
|
7961
8214
|
api: "openai-completions";
|
|
@@ -7972,14 +8225,14 @@ export declare const MODELS: {
|
|
|
7972
8225
|
contextWindow: number;
|
|
7973
8226
|
maxTokens: number;
|
|
7974
8227
|
};
|
|
7975
|
-
readonly "
|
|
8228
|
+
readonly "google/gemma-4-31b-it": {
|
|
7976
8229
|
id: string;
|
|
7977
8230
|
name: string;
|
|
7978
8231
|
api: "openai-completions";
|
|
7979
8232
|
provider: string;
|
|
7980
8233
|
baseUrl: string;
|
|
7981
|
-
reasoning:
|
|
7982
|
-
input: "text"[];
|
|
8234
|
+
reasoning: true;
|
|
8235
|
+
input: ("image" | "text")[];
|
|
7983
8236
|
cost: {
|
|
7984
8237
|
input: number;
|
|
7985
8238
|
output: number;
|
|
@@ -7989,14 +8242,14 @@ export declare const MODELS: {
|
|
|
7989
8242
|
contextWindow: number;
|
|
7990
8243
|
maxTokens: number;
|
|
7991
8244
|
};
|
|
7992
|
-
readonly "
|
|
8245
|
+
readonly "google/gemma-4-31b-it:free": {
|
|
7993
8246
|
id: string;
|
|
7994
8247
|
name: string;
|
|
7995
8248
|
api: "openai-completions";
|
|
7996
8249
|
provider: string;
|
|
7997
8250
|
baseUrl: string;
|
|
7998
8251
|
reasoning: true;
|
|
7999
|
-
input: "text"[];
|
|
8252
|
+
input: ("image" | "text")[];
|
|
8000
8253
|
cost: {
|
|
8001
8254
|
input: number;
|
|
8002
8255
|
output: number;
|
|
@@ -8006,13 +8259,13 @@ export declare const MODELS: {
|
|
|
8006
8259
|
contextWindow: number;
|
|
8007
8260
|
maxTokens: number;
|
|
8008
8261
|
};
|
|
8009
|
-
readonly "inception/mercury-
|
|
8262
|
+
readonly "inception/mercury-2": {
|
|
8010
8263
|
id: string;
|
|
8011
8264
|
name: string;
|
|
8012
8265
|
api: "openai-completions";
|
|
8013
8266
|
provider: string;
|
|
8014
8267
|
baseUrl: string;
|
|
8015
|
-
reasoning:
|
|
8268
|
+
reasoning: true;
|
|
8016
8269
|
input: "text"[];
|
|
8017
8270
|
cost: {
|
|
8018
8271
|
input: number;
|
|
@@ -8023,7 +8276,7 @@ export declare const MODELS: {
|
|
|
8023
8276
|
contextWindow: number;
|
|
8024
8277
|
maxTokens: number;
|
|
8025
8278
|
};
|
|
8026
|
-
readonly "
|
|
8279
|
+
readonly "inclusionai/ling-2.6-flash:free": {
|
|
8027
8280
|
id: string;
|
|
8028
8281
|
name: string;
|
|
8029
8282
|
api: "openai-completions";
|
|
@@ -8040,7 +8293,7 @@ export declare const MODELS: {
|
|
|
8040
8293
|
contextWindow: number;
|
|
8041
8294
|
maxTokens: number;
|
|
8042
8295
|
};
|
|
8043
|
-
readonly "
|
|
8296
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
8044
8297
|
id: string;
|
|
8045
8298
|
name: string;
|
|
8046
8299
|
api: "openai-completions";
|
|
@@ -8142,23 +8395,6 @@ export declare const MODELS: {
|
|
|
8142
8395
|
contextWindow: number;
|
|
8143
8396
|
maxTokens: number;
|
|
8144
8397
|
};
|
|
8145
|
-
readonly "meta-llama/llama-4-maverick": {
|
|
8146
|
-
id: string;
|
|
8147
|
-
name: string;
|
|
8148
|
-
api: "openai-completions";
|
|
8149
|
-
provider: string;
|
|
8150
|
-
baseUrl: string;
|
|
8151
|
-
reasoning: false;
|
|
8152
|
-
input: ("image" | "text")[];
|
|
8153
|
-
cost: {
|
|
8154
|
-
input: number;
|
|
8155
|
-
output: number;
|
|
8156
|
-
cacheRead: number;
|
|
8157
|
-
cacheWrite: number;
|
|
8158
|
-
};
|
|
8159
|
-
contextWindow: number;
|
|
8160
|
-
maxTokens: number;
|
|
8161
|
-
};
|
|
8162
8398
|
readonly "meta-llama/llama-4-scout": {
|
|
8163
8399
|
id: string;
|
|
8164
8400
|
name: string;
|
|
@@ -8652,7 +8888,24 @@ export declare const MODELS: {
|
|
|
8652
8888
|
contextWindow: number;
|
|
8653
8889
|
maxTokens: number;
|
|
8654
8890
|
};
|
|
8655
|
-
readonly "moonshotai/kimi-k2": {
|
|
8891
|
+
readonly "moonshotai/kimi-k2": {
|
|
8892
|
+
id: string;
|
|
8893
|
+
name: string;
|
|
8894
|
+
api: "openai-completions";
|
|
8895
|
+
provider: string;
|
|
8896
|
+
baseUrl: string;
|
|
8897
|
+
reasoning: false;
|
|
8898
|
+
input: "text"[];
|
|
8899
|
+
cost: {
|
|
8900
|
+
input: number;
|
|
8901
|
+
output: number;
|
|
8902
|
+
cacheRead: number;
|
|
8903
|
+
cacheWrite: number;
|
|
8904
|
+
};
|
|
8905
|
+
contextWindow: number;
|
|
8906
|
+
maxTokens: number;
|
|
8907
|
+
};
|
|
8908
|
+
readonly "moonshotai/kimi-k2-0905": {
|
|
8656
8909
|
id: string;
|
|
8657
8910
|
name: string;
|
|
8658
8911
|
api: "openai-completions";
|
|
@@ -8669,13 +8922,13 @@ export declare const MODELS: {
|
|
|
8669
8922
|
contextWindow: number;
|
|
8670
8923
|
maxTokens: number;
|
|
8671
8924
|
};
|
|
8672
|
-
readonly "moonshotai/kimi-k2-
|
|
8925
|
+
readonly "moonshotai/kimi-k2-thinking": {
|
|
8673
8926
|
id: string;
|
|
8674
8927
|
name: string;
|
|
8675
8928
|
api: "openai-completions";
|
|
8676
8929
|
provider: string;
|
|
8677
8930
|
baseUrl: string;
|
|
8678
|
-
reasoning:
|
|
8931
|
+
reasoning: true;
|
|
8679
8932
|
input: "text"[];
|
|
8680
8933
|
cost: {
|
|
8681
8934
|
input: number;
|
|
@@ -8686,14 +8939,14 @@ export declare const MODELS: {
|
|
|
8686
8939
|
contextWindow: number;
|
|
8687
8940
|
maxTokens: number;
|
|
8688
8941
|
};
|
|
8689
|
-
readonly "moonshotai/kimi-k2
|
|
8942
|
+
readonly "moonshotai/kimi-k2.5": {
|
|
8690
8943
|
id: string;
|
|
8691
8944
|
name: string;
|
|
8692
8945
|
api: "openai-completions";
|
|
8693
8946
|
provider: string;
|
|
8694
8947
|
baseUrl: string;
|
|
8695
8948
|
reasoning: true;
|
|
8696
|
-
input: "text"[];
|
|
8949
|
+
input: ("image" | "text")[];
|
|
8697
8950
|
cost: {
|
|
8698
8951
|
input: number;
|
|
8699
8952
|
output: number;
|
|
@@ -8703,7 +8956,7 @@ export declare const MODELS: {
|
|
|
8703
8956
|
contextWindow: number;
|
|
8704
8957
|
maxTokens: number;
|
|
8705
8958
|
};
|
|
8706
|
-
readonly "moonshotai/kimi-k2.
|
|
8959
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
8707
8960
|
id: string;
|
|
8708
8961
|
name: string;
|
|
8709
8962
|
api: "openai-completions";
|
|
@@ -9196,23 +9449,6 @@ export declare const MODELS: {
|
|
|
9196
9449
|
contextWindow: number;
|
|
9197
9450
|
maxTokens: number;
|
|
9198
9451
|
};
|
|
9199
|
-
readonly "openai/gpt-4o:extended": {
|
|
9200
|
-
id: string;
|
|
9201
|
-
name: string;
|
|
9202
|
-
api: "openai-completions";
|
|
9203
|
-
provider: string;
|
|
9204
|
-
baseUrl: string;
|
|
9205
|
-
reasoning: false;
|
|
9206
|
-
input: ("image" | "text")[];
|
|
9207
|
-
cost: {
|
|
9208
|
-
input: number;
|
|
9209
|
-
output: number;
|
|
9210
|
-
cacheRead: number;
|
|
9211
|
-
cacheWrite: number;
|
|
9212
|
-
};
|
|
9213
|
-
contextWindow: number;
|
|
9214
|
-
maxTokens: number;
|
|
9215
|
-
};
|
|
9216
9452
|
readonly "openai/gpt-5": {
|
|
9217
9453
|
id: string;
|
|
9218
9454
|
name: string;
|
|
@@ -9247,40 +9483,6 @@ export declare const MODELS: {
|
|
|
9247
9483
|
contextWindow: number;
|
|
9248
9484
|
maxTokens: number;
|
|
9249
9485
|
};
|
|
9250
|
-
readonly "openai/gpt-5-image": {
|
|
9251
|
-
id: string;
|
|
9252
|
-
name: string;
|
|
9253
|
-
api: "openai-completions";
|
|
9254
|
-
provider: string;
|
|
9255
|
-
baseUrl: string;
|
|
9256
|
-
reasoning: true;
|
|
9257
|
-
input: ("image" | "text")[];
|
|
9258
|
-
cost: {
|
|
9259
|
-
input: number;
|
|
9260
|
-
output: number;
|
|
9261
|
-
cacheRead: number;
|
|
9262
|
-
cacheWrite: number;
|
|
9263
|
-
};
|
|
9264
|
-
contextWindow: number;
|
|
9265
|
-
maxTokens: number;
|
|
9266
|
-
};
|
|
9267
|
-
readonly "openai/gpt-5-image-mini": {
|
|
9268
|
-
id: string;
|
|
9269
|
-
name: string;
|
|
9270
|
-
api: "openai-completions";
|
|
9271
|
-
provider: string;
|
|
9272
|
-
baseUrl: string;
|
|
9273
|
-
reasoning: true;
|
|
9274
|
-
input: ("image" | "text")[];
|
|
9275
|
-
cost: {
|
|
9276
|
-
input: number;
|
|
9277
|
-
output: number;
|
|
9278
|
-
cacheRead: number;
|
|
9279
|
-
cacheWrite: number;
|
|
9280
|
-
};
|
|
9281
|
-
contextWindow: number;
|
|
9282
|
-
maxTokens: number;
|
|
9283
|
-
};
|
|
9284
9486
|
readonly "openai/gpt-5-mini": {
|
|
9285
9487
|
id: string;
|
|
9286
9488
|
name: string;
|
|
@@ -10086,7 +10288,7 @@ export declare const MODELS: {
|
|
|
10086
10288
|
api: "openai-completions";
|
|
10087
10289
|
provider: string;
|
|
10088
10290
|
baseUrl: string;
|
|
10089
|
-
reasoning:
|
|
10291
|
+
reasoning: false;
|
|
10090
10292
|
input: "text"[];
|
|
10091
10293
|
cost: {
|
|
10092
10294
|
input: number;
|
|
@@ -10624,7 +10826,7 @@ export declare const MODELS: {
|
|
|
10624
10826
|
contextWindow: number;
|
|
10625
10827
|
maxTokens: number;
|
|
10626
10828
|
};
|
|
10627
|
-
readonly "qwen/qwen3.6-plus
|
|
10829
|
+
readonly "qwen/qwen3.6-plus": {
|
|
10628
10830
|
id: string;
|
|
10629
10831
|
name: string;
|
|
10630
10832
|
api: "openai-completions";
|
|
@@ -10743,7 +10945,7 @@ export declare const MODELS: {
|
|
|
10743
10945
|
contextWindow: number;
|
|
10744
10946
|
maxTokens: number;
|
|
10745
10947
|
};
|
|
10746
|
-
readonly "
|
|
10948
|
+
readonly "tencent/hy3-preview:free": {
|
|
10747
10949
|
id: string;
|
|
10748
10950
|
name: string;
|
|
10749
10951
|
api: "openai-completions";
|
|
@@ -11032,6 +11234,40 @@ export declare const MODELS: {
|
|
|
11032
11234
|
contextWindow: number;
|
|
11033
11235
|
maxTokens: number;
|
|
11034
11236
|
};
|
|
11237
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
11238
|
+
id: string;
|
|
11239
|
+
name: string;
|
|
11240
|
+
api: "openai-completions";
|
|
11241
|
+
provider: string;
|
|
11242
|
+
baseUrl: string;
|
|
11243
|
+
reasoning: true;
|
|
11244
|
+
input: ("image" | "text")[];
|
|
11245
|
+
cost: {
|
|
11246
|
+
input: number;
|
|
11247
|
+
output: number;
|
|
11248
|
+
cacheRead: number;
|
|
11249
|
+
cacheWrite: number;
|
|
11250
|
+
};
|
|
11251
|
+
contextWindow: number;
|
|
11252
|
+
maxTokens: number;
|
|
11253
|
+
};
|
|
11254
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
11255
|
+
id: string;
|
|
11256
|
+
name: string;
|
|
11257
|
+
api: "openai-completions";
|
|
11258
|
+
provider: string;
|
|
11259
|
+
baseUrl: string;
|
|
11260
|
+
reasoning: true;
|
|
11261
|
+
input: "text"[];
|
|
11262
|
+
cost: {
|
|
11263
|
+
input: number;
|
|
11264
|
+
output: number;
|
|
11265
|
+
cacheRead: number;
|
|
11266
|
+
cacheWrite: number;
|
|
11267
|
+
};
|
|
11268
|
+
contextWindow: number;
|
|
11269
|
+
maxTokens: number;
|
|
11270
|
+
};
|
|
11035
11271
|
readonly "z-ai/glm-4-32b": {
|
|
11036
11272
|
id: string;
|
|
11037
11273
|
name: string;
|
|
@@ -11219,6 +11455,23 @@ export declare const MODELS: {
|
|
|
11219
11455
|
contextWindow: number;
|
|
11220
11456
|
maxTokens: number;
|
|
11221
11457
|
};
|
|
11458
|
+
readonly "z-ai/glm-5.1": {
|
|
11459
|
+
id: string;
|
|
11460
|
+
name: string;
|
|
11461
|
+
api: "openai-completions";
|
|
11462
|
+
provider: string;
|
|
11463
|
+
baseUrl: string;
|
|
11464
|
+
reasoning: true;
|
|
11465
|
+
input: "text"[];
|
|
11466
|
+
cost: {
|
|
11467
|
+
input: number;
|
|
11468
|
+
output: number;
|
|
11469
|
+
cacheRead: number;
|
|
11470
|
+
cacheWrite: number;
|
|
11471
|
+
};
|
|
11472
|
+
contextWindow: number;
|
|
11473
|
+
maxTokens: number;
|
|
11474
|
+
};
|
|
11222
11475
|
readonly "z-ai/glm-5v-turbo": {
|
|
11223
11476
|
id: string;
|
|
11224
11477
|
name: string;
|
|
@@ -11236,6 +11489,23 @@ export declare const MODELS: {
|
|
|
11236
11489
|
contextWindow: number;
|
|
11237
11490
|
maxTokens: number;
|
|
11238
11491
|
};
|
|
11492
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
11493
|
+
id: string;
|
|
11494
|
+
name: string;
|
|
11495
|
+
api: "openai-completions";
|
|
11496
|
+
provider: string;
|
|
11497
|
+
baseUrl: string;
|
|
11498
|
+
reasoning: true;
|
|
11499
|
+
input: ("image" | "text")[];
|
|
11500
|
+
cost: {
|
|
11501
|
+
input: number;
|
|
11502
|
+
output: number;
|
|
11503
|
+
cacheRead: number;
|
|
11504
|
+
cacheWrite: number;
|
|
11505
|
+
};
|
|
11506
|
+
contextWindow: number;
|
|
11507
|
+
maxTokens: number;
|
|
11508
|
+
};
|
|
11239
11509
|
};
|
|
11240
11510
|
readonly "vercel-ai-gateway": {
|
|
11241
11511
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -11306,6 +11576,23 @@ export declare const MODELS: {
|
|
|
11306
11576
|
contextWindow: number;
|
|
11307
11577
|
maxTokens: number;
|
|
11308
11578
|
};
|
|
11579
|
+
readonly "alibaba/qwen-3.6-max-preview": {
|
|
11580
|
+
id: string;
|
|
11581
|
+
name: string;
|
|
11582
|
+
api: "anthropic-messages";
|
|
11583
|
+
provider: string;
|
|
11584
|
+
baseUrl: string;
|
|
11585
|
+
reasoning: true;
|
|
11586
|
+
input: ("image" | "text")[];
|
|
11587
|
+
cost: {
|
|
11588
|
+
input: number;
|
|
11589
|
+
output: number;
|
|
11590
|
+
cacheRead: number;
|
|
11591
|
+
cacheWrite: number;
|
|
11592
|
+
};
|
|
11593
|
+
contextWindow: number;
|
|
11594
|
+
maxTokens: number;
|
|
11595
|
+
};
|
|
11309
11596
|
readonly "alibaba/qwen3-235b-a22b-thinking": {
|
|
11310
11597
|
id: string;
|
|
11311
11598
|
name: string;
|
|
@@ -11646,6 +11933,23 @@ export declare const MODELS: {
|
|
|
11646
11933
|
contextWindow: number;
|
|
11647
11934
|
maxTokens: number;
|
|
11648
11935
|
};
|
|
11936
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
11937
|
+
id: string;
|
|
11938
|
+
name: string;
|
|
11939
|
+
api: "anthropic-messages";
|
|
11940
|
+
provider: string;
|
|
11941
|
+
baseUrl: string;
|
|
11942
|
+
reasoning: true;
|
|
11943
|
+
input: ("image" | "text")[];
|
|
11944
|
+
cost: {
|
|
11945
|
+
input: number;
|
|
11946
|
+
output: number;
|
|
11947
|
+
cacheRead: number;
|
|
11948
|
+
cacheWrite: number;
|
|
11949
|
+
};
|
|
11950
|
+
contextWindow: number;
|
|
11951
|
+
maxTokens: number;
|
|
11952
|
+
};
|
|
11649
11953
|
readonly "anthropic/claude-sonnet-4": {
|
|
11650
11954
|
id: string;
|
|
11651
11955
|
name: string;
|
|
@@ -12632,6 +12936,23 @@ export declare const MODELS: {
|
|
|
12632
12936
|
contextWindow: number;
|
|
12633
12937
|
maxTokens: number;
|
|
12634
12938
|
};
|
|
12939
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
12940
|
+
id: string;
|
|
12941
|
+
name: string;
|
|
12942
|
+
api: "anthropic-messages";
|
|
12943
|
+
provider: string;
|
|
12944
|
+
baseUrl: string;
|
|
12945
|
+
reasoning: true;
|
|
12946
|
+
input: ("image" | "text")[];
|
|
12947
|
+
cost: {
|
|
12948
|
+
input: number;
|
|
12949
|
+
output: number;
|
|
12950
|
+
cacheRead: number;
|
|
12951
|
+
cacheWrite: number;
|
|
12952
|
+
};
|
|
12953
|
+
contextWindow: number;
|
|
12954
|
+
maxTokens: number;
|
|
12955
|
+
};
|
|
12635
12956
|
readonly "nvidia/nemotron-nano-12b-v2-vl": {
|
|
12636
12957
|
id: string;
|
|
12637
12958
|
name: string;
|
|
@@ -13125,23 +13446,6 @@ export declare const MODELS: {
|
|
|
13125
13446
|
contextWindow: number;
|
|
13126
13447
|
maxTokens: number;
|
|
13127
13448
|
};
|
|
13128
|
-
readonly "openai/gpt-oss-120b": {
|
|
13129
|
-
id: string;
|
|
13130
|
-
name: string;
|
|
13131
|
-
api: "anthropic-messages";
|
|
13132
|
-
provider: string;
|
|
13133
|
-
baseUrl: string;
|
|
13134
|
-
reasoning: true;
|
|
13135
|
-
input: "text"[];
|
|
13136
|
-
cost: {
|
|
13137
|
-
input: number;
|
|
13138
|
-
output: number;
|
|
13139
|
-
cacheRead: number;
|
|
13140
|
-
cacheWrite: number;
|
|
13141
|
-
};
|
|
13142
|
-
contextWindow: number;
|
|
13143
|
-
maxTokens: number;
|
|
13144
|
-
};
|
|
13145
13449
|
readonly "openai/gpt-oss-20b": {
|
|
13146
13450
|
id: string;
|
|
13147
13451
|
name: string;
|
|
@@ -13329,23 +13633,6 @@ export declare const MODELS: {
|
|
|
13329
13633
|
contextWindow: number;
|
|
13330
13634
|
maxTokens: number;
|
|
13331
13635
|
};
|
|
13332
|
-
readonly "xai/grok-2-vision": {
|
|
13333
|
-
id: string;
|
|
13334
|
-
name: string;
|
|
13335
|
-
api: "anthropic-messages";
|
|
13336
|
-
provider: string;
|
|
13337
|
-
baseUrl: string;
|
|
13338
|
-
reasoning: false;
|
|
13339
|
-
input: ("image" | "text")[];
|
|
13340
|
-
cost: {
|
|
13341
|
-
input: number;
|
|
13342
|
-
output: number;
|
|
13343
|
-
cacheRead: number;
|
|
13344
|
-
cacheWrite: number;
|
|
13345
|
-
};
|
|
13346
|
-
contextWindow: number;
|
|
13347
|
-
maxTokens: number;
|
|
13348
|
-
};
|
|
13349
13636
|
readonly "xai/grok-3": {
|
|
13350
13637
|
id: string;
|
|
13351
13638
|
name: string;
|
|
@@ -13438,7 +13725,7 @@ export declare const MODELS: {
|
|
|
13438
13725
|
provider: string;
|
|
13439
13726
|
baseUrl: string;
|
|
13440
13727
|
reasoning: false;
|
|
13441
|
-
input: "text"[];
|
|
13728
|
+
input: ("image" | "text")[];
|
|
13442
13729
|
cost: {
|
|
13443
13730
|
input: number;
|
|
13444
13731
|
output: number;
|
|
@@ -13455,7 +13742,7 @@ export declare const MODELS: {
|
|
|
13455
13742
|
provider: string;
|
|
13456
13743
|
baseUrl: string;
|
|
13457
13744
|
reasoning: true;
|
|
13458
|
-
input: "text"[];
|
|
13745
|
+
input: ("image" | "text")[];
|
|
13459
13746
|
cost: {
|
|
13460
13747
|
input: number;
|
|
13461
13748
|
output: number;
|
|
@@ -13472,7 +13759,7 @@ export declare const MODELS: {
|
|
|
13472
13759
|
provider: string;
|
|
13473
13760
|
baseUrl: string;
|
|
13474
13761
|
reasoning: false;
|
|
13475
|
-
input: "text"[];
|
|
13762
|
+
input: ("image" | "text")[];
|
|
13476
13763
|
cost: {
|
|
13477
13764
|
input: number;
|
|
13478
13765
|
output: number;
|
|
@@ -13489,7 +13776,7 @@ export declare const MODELS: {
|
|
|
13489
13776
|
provider: string;
|
|
13490
13777
|
baseUrl: string;
|
|
13491
13778
|
reasoning: true;
|
|
13492
|
-
input: "text"[];
|
|
13779
|
+
input: ("image" | "text")[];
|
|
13493
13780
|
cost: {
|
|
13494
13781
|
input: number;
|
|
13495
13782
|
output: number;
|
|
@@ -13506,7 +13793,7 @@ export declare const MODELS: {
|
|
|
13506
13793
|
provider: string;
|
|
13507
13794
|
baseUrl: string;
|
|
13508
13795
|
reasoning: true;
|
|
13509
|
-
input: "text"[];
|
|
13796
|
+
input: ("image" | "text")[];
|
|
13510
13797
|
cost: {
|
|
13511
13798
|
input: number;
|
|
13512
13799
|
output: number;
|
|
@@ -13523,7 +13810,7 @@ export declare const MODELS: {
|
|
|
13523
13810
|
provider: string;
|
|
13524
13811
|
baseUrl: string;
|
|
13525
13812
|
reasoning: true;
|
|
13526
|
-
input: "text"[];
|
|
13813
|
+
input: ("image" | "text")[];
|
|
13527
13814
|
cost: {
|
|
13528
13815
|
input: number;
|
|
13529
13816
|
output: number;
|
|
@@ -13839,6 +14126,23 @@ export declare const MODELS: {
|
|
|
13839
14126
|
contextWindow: number;
|
|
13840
14127
|
maxTokens: number;
|
|
13841
14128
|
};
|
|
14129
|
+
readonly "zai/glm-5.1": {
|
|
14130
|
+
id: string;
|
|
14131
|
+
name: string;
|
|
14132
|
+
api: "anthropic-messages";
|
|
14133
|
+
provider: string;
|
|
14134
|
+
baseUrl: string;
|
|
14135
|
+
reasoning: true;
|
|
14136
|
+
input: "text"[];
|
|
14137
|
+
cost: {
|
|
14138
|
+
input: number;
|
|
14139
|
+
output: number;
|
|
14140
|
+
cacheRead: number;
|
|
14141
|
+
cacheWrite: number;
|
|
14142
|
+
};
|
|
14143
|
+
contextWindow: number;
|
|
14144
|
+
maxTokens: number;
|
|
14145
|
+
};
|
|
13842
14146
|
readonly "zai/glm-5v-turbo": {
|
|
13843
14147
|
id: string;
|
|
13844
14148
|
name: string;
|
|
@@ -14506,6 +14810,28 @@ export declare const MODELS: {
|
|
|
14506
14810
|
contextWindow: number;
|
|
14507
14811
|
maxTokens: number;
|
|
14508
14812
|
};
|
|
14813
|
+
readonly "glm-5.1": {
|
|
14814
|
+
id: string;
|
|
14815
|
+
name: string;
|
|
14816
|
+
api: "openai-completions";
|
|
14817
|
+
provider: string;
|
|
14818
|
+
baseUrl: string;
|
|
14819
|
+
compat: {
|
|
14820
|
+
supportsDeveloperRole: false;
|
|
14821
|
+
thinkingFormat: "zai";
|
|
14822
|
+
zaiToolStream: true;
|
|
14823
|
+
};
|
|
14824
|
+
reasoning: true;
|
|
14825
|
+
input: "text"[];
|
|
14826
|
+
cost: {
|
|
14827
|
+
input: number;
|
|
14828
|
+
output: number;
|
|
14829
|
+
cacheRead: number;
|
|
14830
|
+
cacheWrite: number;
|
|
14831
|
+
};
|
|
14832
|
+
contextWindow: number;
|
|
14833
|
+
maxTokens: number;
|
|
14834
|
+
};
|
|
14509
14835
|
readonly "glm-5v-turbo": {
|
|
14510
14836
|
id: string;
|
|
14511
14837
|
name: string;
|