@draht/ai 2026.3.25 → 2026.4.23
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +107 -12
- package/dist/cli.d.ts.map +1 -1
- package/dist/cli.js +1 -1
- package/dist/cli.js.map +1 -1
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -0
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +1163 -498
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +1392 -756
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +37 -9
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +38 -16
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/faux.d.ts +56 -0
- package/dist/providers/faux.d.ts.map +1 -0
- package/dist/providers/faux.js +367 -0
- package/dist/providers/faux.js.map +1 -0
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +2 -2
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +1 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +29 -4
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +18 -6
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +20 -1
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +2 -2
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/types.d.ts +64 -5
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts +4 -3
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +28 -13
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +5 -8
|
@@ -255,6 +255,23 @@ export declare const MODELS: {
|
|
|
255
255
|
contextWindow: number;
|
|
256
256
|
maxTokens: number;
|
|
257
257
|
};
|
|
258
|
+
readonly "anthropic.claude-opus-4-7": {
|
|
259
|
+
id: string;
|
|
260
|
+
name: string;
|
|
261
|
+
api: "bedrock-converse-stream";
|
|
262
|
+
provider: string;
|
|
263
|
+
baseUrl: string;
|
|
264
|
+
reasoning: true;
|
|
265
|
+
input: ("image" | "text")[];
|
|
266
|
+
cost: {
|
|
267
|
+
input: number;
|
|
268
|
+
output: number;
|
|
269
|
+
cacheRead: number;
|
|
270
|
+
cacheWrite: number;
|
|
271
|
+
};
|
|
272
|
+
contextWindow: number;
|
|
273
|
+
maxTokens: number;
|
|
274
|
+
};
|
|
258
275
|
readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
259
276
|
id: string;
|
|
260
277
|
name: string;
|
|
@@ -408,6 +425,23 @@ export declare const MODELS: {
|
|
|
408
425
|
contextWindow: number;
|
|
409
426
|
maxTokens: number;
|
|
410
427
|
};
|
|
428
|
+
readonly "eu.anthropic.claude-opus-4-7": {
|
|
429
|
+
id: string;
|
|
430
|
+
name: string;
|
|
431
|
+
api: "bedrock-converse-stream";
|
|
432
|
+
provider: string;
|
|
433
|
+
baseUrl: string;
|
|
434
|
+
reasoning: true;
|
|
435
|
+
input: ("image" | "text")[];
|
|
436
|
+
cost: {
|
|
437
|
+
input: number;
|
|
438
|
+
output: number;
|
|
439
|
+
cacheRead: number;
|
|
440
|
+
cacheWrite: number;
|
|
441
|
+
};
|
|
442
|
+
contextWindow: number;
|
|
443
|
+
maxTokens: number;
|
|
444
|
+
};
|
|
411
445
|
readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
412
446
|
id: string;
|
|
413
447
|
name: string;
|
|
@@ -510,6 +544,23 @@ export declare const MODELS: {
|
|
|
510
544
|
contextWindow: number;
|
|
511
545
|
maxTokens: number;
|
|
512
546
|
};
|
|
547
|
+
readonly "global.anthropic.claude-opus-4-7": {
|
|
548
|
+
id: string;
|
|
549
|
+
name: string;
|
|
550
|
+
api: "bedrock-converse-stream";
|
|
551
|
+
provider: string;
|
|
552
|
+
baseUrl: string;
|
|
553
|
+
reasoning: true;
|
|
554
|
+
input: ("image" | "text")[];
|
|
555
|
+
cost: {
|
|
556
|
+
input: number;
|
|
557
|
+
output: number;
|
|
558
|
+
cacheRead: number;
|
|
559
|
+
cacheWrite: number;
|
|
560
|
+
};
|
|
561
|
+
contextWindow: number;
|
|
562
|
+
maxTokens: number;
|
|
563
|
+
};
|
|
513
564
|
readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
514
565
|
id: string;
|
|
515
566
|
name: string;
|
|
@@ -1054,6 +1105,23 @@ export declare const MODELS: {
|
|
|
1054
1105
|
contextWindow: number;
|
|
1055
1106
|
maxTokens: number;
|
|
1056
1107
|
};
|
|
1108
|
+
readonly "nvidia.nemotron-super-3-120b": {
|
|
1109
|
+
id: string;
|
|
1110
|
+
name: string;
|
|
1111
|
+
api: "bedrock-converse-stream";
|
|
1112
|
+
provider: string;
|
|
1113
|
+
baseUrl: string;
|
|
1114
|
+
reasoning: true;
|
|
1115
|
+
input: "text"[];
|
|
1116
|
+
cost: {
|
|
1117
|
+
input: number;
|
|
1118
|
+
output: number;
|
|
1119
|
+
cacheRead: number;
|
|
1120
|
+
cacheWrite: number;
|
|
1121
|
+
};
|
|
1122
|
+
contextWindow: number;
|
|
1123
|
+
maxTokens: number;
|
|
1124
|
+
};
|
|
1057
1125
|
readonly "openai.gpt-oss-120b-1:0": {
|
|
1058
1126
|
id: string;
|
|
1059
1127
|
name: string;
|
|
@@ -1190,6 +1258,23 @@ export declare const MODELS: {
|
|
|
1190
1258
|
contextWindow: number;
|
|
1191
1259
|
maxTokens: number;
|
|
1192
1260
|
};
|
|
1261
|
+
readonly "qwen.qwen3-coder-next": {
|
|
1262
|
+
id: string;
|
|
1263
|
+
name: string;
|
|
1264
|
+
api: "bedrock-converse-stream";
|
|
1265
|
+
provider: string;
|
|
1266
|
+
baseUrl: string;
|
|
1267
|
+
reasoning: true;
|
|
1268
|
+
input: "text"[];
|
|
1269
|
+
cost: {
|
|
1270
|
+
input: number;
|
|
1271
|
+
output: number;
|
|
1272
|
+
cacheRead: number;
|
|
1273
|
+
cacheWrite: number;
|
|
1274
|
+
};
|
|
1275
|
+
contextWindow: number;
|
|
1276
|
+
maxTokens: number;
|
|
1277
|
+
};
|
|
1193
1278
|
readonly "qwen.qwen3-next-80b-a3b": {
|
|
1194
1279
|
id: string;
|
|
1195
1280
|
name: string;
|
|
@@ -1309,6 +1394,23 @@ export declare const MODELS: {
|
|
|
1309
1394
|
contextWindow: number;
|
|
1310
1395
|
maxTokens: number;
|
|
1311
1396
|
};
|
|
1397
|
+
readonly "us.anthropic.claude-opus-4-7": {
|
|
1398
|
+
id: string;
|
|
1399
|
+
name: string;
|
|
1400
|
+
api: "bedrock-converse-stream";
|
|
1401
|
+
provider: string;
|
|
1402
|
+
baseUrl: string;
|
|
1403
|
+
reasoning: true;
|
|
1404
|
+
input: ("image" | "text")[];
|
|
1405
|
+
cost: {
|
|
1406
|
+
input: number;
|
|
1407
|
+
output: number;
|
|
1408
|
+
cacheRead: number;
|
|
1409
|
+
cacheWrite: number;
|
|
1410
|
+
};
|
|
1411
|
+
contextWindow: number;
|
|
1412
|
+
maxTokens: number;
|
|
1413
|
+
};
|
|
1312
1414
|
readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
1313
1415
|
id: string;
|
|
1314
1416
|
name: string;
|
|
@@ -1532,23 +1634,6 @@ export declare const MODELS: {
|
|
|
1532
1634
|
contextWindow: number;
|
|
1533
1635
|
maxTokens: number;
|
|
1534
1636
|
};
|
|
1535
|
-
readonly "claude-3-7-sonnet-latest": {
|
|
1536
|
-
id: string;
|
|
1537
|
-
name: string;
|
|
1538
|
-
api: "anthropic-messages";
|
|
1539
|
-
provider: string;
|
|
1540
|
-
baseUrl: string;
|
|
1541
|
-
reasoning: true;
|
|
1542
|
-
input: ("image" | "text")[];
|
|
1543
|
-
cost: {
|
|
1544
|
-
input: number;
|
|
1545
|
-
output: number;
|
|
1546
|
-
cacheRead: number;
|
|
1547
|
-
cacheWrite: number;
|
|
1548
|
-
};
|
|
1549
|
-
contextWindow: number;
|
|
1550
|
-
maxTokens: number;
|
|
1551
|
-
};
|
|
1552
1637
|
readonly "claude-3-haiku-20240307": {
|
|
1553
1638
|
id: string;
|
|
1554
1639
|
name: string;
|
|
@@ -1753,7 +1838,7 @@ export declare const MODELS: {
|
|
|
1753
1838
|
contextWindow: number;
|
|
1754
1839
|
maxTokens: number;
|
|
1755
1840
|
};
|
|
1756
|
-
readonly "claude-
|
|
1841
|
+
readonly "claude-opus-4-7": {
|
|
1757
1842
|
id: string;
|
|
1758
1843
|
name: string;
|
|
1759
1844
|
api: "anthropic-messages";
|
|
@@ -1770,7 +1855,7 @@ export declare const MODELS: {
|
|
|
1770
1855
|
contextWindow: number;
|
|
1771
1856
|
maxTokens: number;
|
|
1772
1857
|
};
|
|
1773
|
-
readonly "claude-sonnet-4-
|
|
1858
|
+
readonly "claude-sonnet-4-0": {
|
|
1774
1859
|
id: string;
|
|
1775
1860
|
name: string;
|
|
1776
1861
|
api: "anthropic-messages";
|
|
@@ -1787,7 +1872,7 @@ export declare const MODELS: {
|
|
|
1787
1872
|
contextWindow: number;
|
|
1788
1873
|
maxTokens: number;
|
|
1789
1874
|
};
|
|
1790
|
-
readonly "claude-sonnet-4-
|
|
1875
|
+
readonly "claude-sonnet-4-20250514": {
|
|
1791
1876
|
id: string;
|
|
1792
1877
|
name: string;
|
|
1793
1878
|
api: "anthropic-messages";
|
|
@@ -1804,7 +1889,7 @@ export declare const MODELS: {
|
|
|
1804
1889
|
contextWindow: number;
|
|
1805
1890
|
maxTokens: number;
|
|
1806
1891
|
};
|
|
1807
|
-
readonly "claude-sonnet-4-5
|
|
1892
|
+
readonly "claude-sonnet-4-5": {
|
|
1808
1893
|
id: string;
|
|
1809
1894
|
name: string;
|
|
1810
1895
|
api: "anthropic-messages";
|
|
@@ -1821,7 +1906,7 @@ export declare const MODELS: {
|
|
|
1821
1906
|
contextWindow: number;
|
|
1822
1907
|
maxTokens: number;
|
|
1823
1908
|
};
|
|
1824
|
-
readonly "claude-sonnet-4-
|
|
1909
|
+
readonly "claude-sonnet-4-5-20250929": {
|
|
1825
1910
|
id: string;
|
|
1826
1911
|
name: string;
|
|
1827
1912
|
api: "anthropic-messages";
|
|
@@ -1838,16 +1923,14 @@ export declare const MODELS: {
|
|
|
1838
1923
|
contextWindow: number;
|
|
1839
1924
|
maxTokens: number;
|
|
1840
1925
|
};
|
|
1841
|
-
|
|
1842
|
-
readonly "azure-openai-responses": {
|
|
1843
|
-
readonly "codex-mini-latest": {
|
|
1926
|
+
readonly "claude-sonnet-4-6": {
|
|
1844
1927
|
id: string;
|
|
1845
1928
|
name: string;
|
|
1846
|
-
api: "
|
|
1929
|
+
api: "anthropic-messages";
|
|
1847
1930
|
provider: string;
|
|
1848
1931
|
baseUrl: string;
|
|
1849
1932
|
reasoning: true;
|
|
1850
|
-
input: "text"[];
|
|
1933
|
+
input: ("image" | "text")[];
|
|
1851
1934
|
cost: {
|
|
1852
1935
|
input: number;
|
|
1853
1936
|
output: number;
|
|
@@ -1857,6 +1940,8 @@ export declare const MODELS: {
|
|
|
1857
1940
|
contextWindow: number;
|
|
1858
1941
|
maxTokens: number;
|
|
1859
1942
|
};
|
|
1943
|
+
};
|
|
1944
|
+
readonly "azure-openai-responses": {
|
|
1860
1945
|
readonly "gpt-4": {
|
|
1861
1946
|
id: string;
|
|
1862
1947
|
name: string;
|
|
@@ -2282,6 +2367,23 @@ export declare const MODELS: {
|
|
|
2282
2367
|
contextWindow: number;
|
|
2283
2368
|
maxTokens: number;
|
|
2284
2369
|
};
|
|
2370
|
+
readonly "gpt-5.3-chat-latest": {
|
|
2371
|
+
id: string;
|
|
2372
|
+
name: string;
|
|
2373
|
+
api: "azure-openai-responses";
|
|
2374
|
+
provider: string;
|
|
2375
|
+
baseUrl: string;
|
|
2376
|
+
reasoning: false;
|
|
2377
|
+
input: ("image" | "text")[];
|
|
2378
|
+
cost: {
|
|
2379
|
+
input: number;
|
|
2380
|
+
output: number;
|
|
2381
|
+
cacheRead: number;
|
|
2382
|
+
cacheWrite: number;
|
|
2383
|
+
};
|
|
2384
|
+
contextWindow: number;
|
|
2385
|
+
maxTokens: number;
|
|
2386
|
+
};
|
|
2285
2387
|
readonly "gpt-5.3-codex": {
|
|
2286
2388
|
id: string;
|
|
2287
2389
|
name: string;
|
|
@@ -2661,6 +2763,29 @@ export declare const MODELS: {
|
|
|
2661
2763
|
contextWindow: number;
|
|
2662
2764
|
maxTokens: number;
|
|
2663
2765
|
};
|
|
2766
|
+
readonly "claude-opus-4.7": {
|
|
2767
|
+
id: string;
|
|
2768
|
+
name: string;
|
|
2769
|
+
api: "anthropic-messages";
|
|
2770
|
+
provider: string;
|
|
2771
|
+
baseUrl: string;
|
|
2772
|
+
headers: {
|
|
2773
|
+
"User-Agent": string;
|
|
2774
|
+
"Editor-Version": string;
|
|
2775
|
+
"Editor-Plugin-Version": string;
|
|
2776
|
+
"Copilot-Integration-Id": string;
|
|
2777
|
+
};
|
|
2778
|
+
reasoning: true;
|
|
2779
|
+
input: ("image" | "text")[];
|
|
2780
|
+
cost: {
|
|
2781
|
+
input: number;
|
|
2782
|
+
output: number;
|
|
2783
|
+
cacheRead: number;
|
|
2784
|
+
cacheWrite: number;
|
|
2785
|
+
};
|
|
2786
|
+
contextWindow: number;
|
|
2787
|
+
maxTokens: number;
|
|
2788
|
+
};
|
|
2664
2789
|
readonly "claude-sonnet-4": {
|
|
2665
2790
|
id: string;
|
|
2666
2791
|
name: string;
|
|
@@ -3589,6 +3714,57 @@ export declare const MODELS: {
|
|
|
3589
3714
|
contextWindow: number;
|
|
3590
3715
|
maxTokens: number;
|
|
3591
3716
|
};
|
|
3717
|
+
readonly "gemma-3-27b-it": {
|
|
3718
|
+
id: string;
|
|
3719
|
+
name: string;
|
|
3720
|
+
api: "google-generative-ai";
|
|
3721
|
+
provider: string;
|
|
3722
|
+
baseUrl: string;
|
|
3723
|
+
reasoning: false;
|
|
3724
|
+
input: ("image" | "text")[];
|
|
3725
|
+
cost: {
|
|
3726
|
+
input: number;
|
|
3727
|
+
output: number;
|
|
3728
|
+
cacheRead: number;
|
|
3729
|
+
cacheWrite: number;
|
|
3730
|
+
};
|
|
3731
|
+
contextWindow: number;
|
|
3732
|
+
maxTokens: number;
|
|
3733
|
+
};
|
|
3734
|
+
readonly "gemma-4-26b-a4b-it": {
|
|
3735
|
+
id: string;
|
|
3736
|
+
name: string;
|
|
3737
|
+
api: "google-generative-ai";
|
|
3738
|
+
provider: string;
|
|
3739
|
+
baseUrl: string;
|
|
3740
|
+
reasoning: true;
|
|
3741
|
+
input: ("image" | "text")[];
|
|
3742
|
+
cost: {
|
|
3743
|
+
input: number;
|
|
3744
|
+
output: number;
|
|
3745
|
+
cacheRead: number;
|
|
3746
|
+
cacheWrite: number;
|
|
3747
|
+
};
|
|
3748
|
+
contextWindow: number;
|
|
3749
|
+
maxTokens: number;
|
|
3750
|
+
};
|
|
3751
|
+
readonly "gemma-4-31b-it": {
|
|
3752
|
+
id: string;
|
|
3753
|
+
name: string;
|
|
3754
|
+
api: "google-generative-ai";
|
|
3755
|
+
provider: string;
|
|
3756
|
+
baseUrl: string;
|
|
3757
|
+
reasoning: true;
|
|
3758
|
+
input: ("image" | "text")[];
|
|
3759
|
+
cost: {
|
|
3760
|
+
input: number;
|
|
3761
|
+
output: number;
|
|
3762
|
+
cacheRead: number;
|
|
3763
|
+
cacheWrite: number;
|
|
3764
|
+
};
|
|
3765
|
+
contextWindow: number;
|
|
3766
|
+
maxTokens: number;
|
|
3767
|
+
};
|
|
3592
3768
|
};
|
|
3593
3769
|
readonly "google-antigravity": {
|
|
3594
3770
|
readonly "claude-opus-4-5-thinking": {
|
|
@@ -4054,16 +4230,14 @@ export declare const MODELS: {
|
|
|
4054
4230
|
contextWindow: number;
|
|
4055
4231
|
maxTokens: number;
|
|
4056
4232
|
};
|
|
4057
|
-
|
|
4058
|
-
readonly groq: {
|
|
4059
|
-
readonly "deepseek-r1-distill-llama-70b": {
|
|
4233
|
+
readonly "gemini-3.1-pro-preview-customtools": {
|
|
4060
4234
|
id: string;
|
|
4061
4235
|
name: string;
|
|
4062
|
-
api: "
|
|
4236
|
+
api: "google-vertex";
|
|
4063
4237
|
provider: string;
|
|
4064
4238
|
baseUrl: string;
|
|
4065
4239
|
reasoning: true;
|
|
4066
|
-
input: "text"[];
|
|
4240
|
+
input: ("image" | "text")[];
|
|
4067
4241
|
cost: {
|
|
4068
4242
|
input: number;
|
|
4069
4243
|
output: number;
|
|
@@ -4073,13 +4247,15 @@ export declare const MODELS: {
|
|
|
4073
4247
|
contextWindow: number;
|
|
4074
4248
|
maxTokens: number;
|
|
4075
4249
|
};
|
|
4076
|
-
|
|
4250
|
+
};
|
|
4251
|
+
readonly groq: {
|
|
4252
|
+
readonly "deepseek-r1-distill-llama-70b": {
|
|
4077
4253
|
id: string;
|
|
4078
4254
|
name: string;
|
|
4079
4255
|
api: "openai-completions";
|
|
4080
4256
|
provider: string;
|
|
4081
4257
|
baseUrl: string;
|
|
4082
|
-
reasoning:
|
|
4258
|
+
reasoning: true;
|
|
4083
4259
|
input: "text"[];
|
|
4084
4260
|
cost: {
|
|
4085
4261
|
input: number;
|
|
@@ -4090,7 +4266,7 @@ export declare const MODELS: {
|
|
|
4090
4266
|
contextWindow: number;
|
|
4091
4267
|
maxTokens: number;
|
|
4092
4268
|
};
|
|
4093
|
-
readonly "
|
|
4269
|
+
readonly "gemma2-9b-it": {
|
|
4094
4270
|
id: string;
|
|
4095
4271
|
name: string;
|
|
4096
4272
|
api: "openai-completions";
|
|
@@ -4107,13 +4283,13 @@ export declare const MODELS: {
|
|
|
4107
4283
|
contextWindow: number;
|
|
4108
4284
|
maxTokens: number;
|
|
4109
4285
|
};
|
|
4110
|
-
readonly "
|
|
4286
|
+
readonly "groq/compound": {
|
|
4111
4287
|
id: string;
|
|
4112
4288
|
name: string;
|
|
4113
4289
|
api: "openai-completions";
|
|
4114
4290
|
provider: string;
|
|
4115
4291
|
baseUrl: string;
|
|
4116
|
-
reasoning:
|
|
4292
|
+
reasoning: true;
|
|
4117
4293
|
input: "text"[];
|
|
4118
4294
|
cost: {
|
|
4119
4295
|
input: number;
|
|
@@ -4124,13 +4300,13 @@ export declare const MODELS: {
|
|
|
4124
4300
|
contextWindow: number;
|
|
4125
4301
|
maxTokens: number;
|
|
4126
4302
|
};
|
|
4127
|
-
readonly "
|
|
4303
|
+
readonly "groq/compound-mini": {
|
|
4128
4304
|
id: string;
|
|
4129
4305
|
name: string;
|
|
4130
4306
|
api: "openai-completions";
|
|
4131
4307
|
provider: string;
|
|
4132
4308
|
baseUrl: string;
|
|
4133
|
-
reasoning:
|
|
4309
|
+
reasoning: true;
|
|
4134
4310
|
input: "text"[];
|
|
4135
4311
|
cost: {
|
|
4136
4312
|
input: number;
|
|
@@ -4141,7 +4317,7 @@ export declare const MODELS: {
|
|
|
4141
4317
|
contextWindow: number;
|
|
4142
4318
|
maxTokens: number;
|
|
4143
4319
|
};
|
|
4144
|
-
readonly "
|
|
4320
|
+
readonly "llama-3.1-8b-instant": {
|
|
4145
4321
|
id: string;
|
|
4146
4322
|
name: string;
|
|
4147
4323
|
api: "openai-completions";
|
|
@@ -4158,14 +4334,14 @@ export declare const MODELS: {
|
|
|
4158
4334
|
contextWindow: number;
|
|
4159
4335
|
maxTokens: number;
|
|
4160
4336
|
};
|
|
4161
|
-
readonly "
|
|
4337
|
+
readonly "llama-3.3-70b-versatile": {
|
|
4162
4338
|
id: string;
|
|
4163
4339
|
name: string;
|
|
4164
4340
|
api: "openai-completions";
|
|
4165
4341
|
provider: string;
|
|
4166
4342
|
baseUrl: string;
|
|
4167
4343
|
reasoning: false;
|
|
4168
|
-
input:
|
|
4344
|
+
input: "text"[];
|
|
4169
4345
|
cost: {
|
|
4170
4346
|
input: number;
|
|
4171
4347
|
output: number;
|
|
@@ -4175,14 +4351,14 @@ export declare const MODELS: {
|
|
|
4175
4351
|
contextWindow: number;
|
|
4176
4352
|
maxTokens: number;
|
|
4177
4353
|
};
|
|
4178
|
-
readonly "
|
|
4354
|
+
readonly "llama3-70b-8192": {
|
|
4179
4355
|
id: string;
|
|
4180
4356
|
name: string;
|
|
4181
4357
|
api: "openai-completions";
|
|
4182
4358
|
provider: string;
|
|
4183
4359
|
baseUrl: string;
|
|
4184
4360
|
reasoning: false;
|
|
4185
|
-
input:
|
|
4361
|
+
input: "text"[];
|
|
4186
4362
|
cost: {
|
|
4187
4363
|
input: number;
|
|
4188
4364
|
output: number;
|
|
@@ -4192,7 +4368,7 @@ export declare const MODELS: {
|
|
|
4192
4368
|
contextWindow: number;
|
|
4193
4369
|
maxTokens: number;
|
|
4194
4370
|
};
|
|
4195
|
-
readonly "
|
|
4371
|
+
readonly "llama3-8b-8192": {
|
|
4196
4372
|
id: string;
|
|
4197
4373
|
name: string;
|
|
4198
4374
|
api: "openai-completions";
|
|
@@ -4209,14 +4385,14 @@ export declare const MODELS: {
|
|
|
4209
4385
|
contextWindow: number;
|
|
4210
4386
|
maxTokens: number;
|
|
4211
4387
|
};
|
|
4212
|
-
readonly "
|
|
4388
|
+
readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
|
|
4213
4389
|
id: string;
|
|
4214
4390
|
name: string;
|
|
4215
4391
|
api: "openai-completions";
|
|
4216
4392
|
provider: string;
|
|
4217
4393
|
baseUrl: string;
|
|
4218
4394
|
reasoning: false;
|
|
4219
|
-
input: "text"[];
|
|
4395
|
+
input: ("image" | "text")[];
|
|
4220
4396
|
cost: {
|
|
4221
4397
|
input: number;
|
|
4222
4398
|
output: number;
|
|
@@ -4226,14 +4402,14 @@ export declare const MODELS: {
|
|
|
4226
4402
|
contextWindow: number;
|
|
4227
4403
|
maxTokens: number;
|
|
4228
4404
|
};
|
|
4229
|
-
readonly "
|
|
4405
|
+
readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
4230
4406
|
id: string;
|
|
4231
4407
|
name: string;
|
|
4232
4408
|
api: "openai-completions";
|
|
4233
4409
|
provider: string;
|
|
4234
4410
|
baseUrl: string;
|
|
4235
4411
|
reasoning: false;
|
|
4236
|
-
input: "text"[];
|
|
4412
|
+
input: ("image" | "text")[];
|
|
4237
4413
|
cost: {
|
|
4238
4414
|
input: number;
|
|
4239
4415
|
output: number;
|
|
@@ -4243,13 +4419,13 @@ export declare const MODELS: {
|
|
|
4243
4419
|
contextWindow: number;
|
|
4244
4420
|
maxTokens: number;
|
|
4245
4421
|
};
|
|
4246
|
-
readonly "
|
|
4422
|
+
readonly "mistral-saba-24b": {
|
|
4247
4423
|
id: string;
|
|
4248
4424
|
name: string;
|
|
4249
4425
|
api: "openai-completions";
|
|
4250
4426
|
provider: string;
|
|
4251
4427
|
baseUrl: string;
|
|
4252
|
-
reasoning:
|
|
4428
|
+
reasoning: false;
|
|
4253
4429
|
input: "text"[];
|
|
4254
4430
|
cost: {
|
|
4255
4431
|
input: number;
|
|
@@ -4260,13 +4436,13 @@ export declare const MODELS: {
|
|
|
4260
4436
|
contextWindow: number;
|
|
4261
4437
|
maxTokens: number;
|
|
4262
4438
|
};
|
|
4263
|
-
readonly "
|
|
4439
|
+
readonly "moonshotai/kimi-k2-instruct": {
|
|
4264
4440
|
id: string;
|
|
4265
4441
|
name: string;
|
|
4266
4442
|
api: "openai-completions";
|
|
4267
4443
|
provider: string;
|
|
4268
4444
|
baseUrl: string;
|
|
4269
|
-
reasoning:
|
|
4445
|
+
reasoning: false;
|
|
4270
4446
|
input: "text"[];
|
|
4271
4447
|
cost: {
|
|
4272
4448
|
input: number;
|
|
@@ -4277,13 +4453,13 @@ export declare const MODELS: {
|
|
|
4277
4453
|
contextWindow: number;
|
|
4278
4454
|
maxTokens: number;
|
|
4279
4455
|
};
|
|
4280
|
-
readonly "
|
|
4456
|
+
readonly "moonshotai/kimi-k2-instruct-0905": {
|
|
4281
4457
|
id: string;
|
|
4282
4458
|
name: string;
|
|
4283
4459
|
api: "openai-completions";
|
|
4284
4460
|
provider: string;
|
|
4285
4461
|
baseUrl: string;
|
|
4286
|
-
reasoning:
|
|
4462
|
+
reasoning: false;
|
|
4287
4463
|
input: "text"[];
|
|
4288
4464
|
cost: {
|
|
4289
4465
|
input: number;
|
|
@@ -4294,7 +4470,7 @@ export declare const MODELS: {
|
|
|
4294
4470
|
contextWindow: number;
|
|
4295
4471
|
maxTokens: number;
|
|
4296
4472
|
};
|
|
4297
|
-
readonly "
|
|
4473
|
+
readonly "openai/gpt-oss-120b": {
|
|
4298
4474
|
id: string;
|
|
4299
4475
|
name: string;
|
|
4300
4476
|
api: "openai-completions";
|
|
@@ -4311,17 +4487,12 @@ export declare const MODELS: {
|
|
|
4311
4487
|
contextWindow: number;
|
|
4312
4488
|
maxTokens: number;
|
|
4313
4489
|
};
|
|
4314
|
-
|
|
4315
|
-
readonly huggingface: {
|
|
4316
|
-
readonly "MiniMaxAI/MiniMax-M2.1": {
|
|
4490
|
+
readonly "openai/gpt-oss-20b": {
|
|
4317
4491
|
id: string;
|
|
4318
4492
|
name: string;
|
|
4319
4493
|
api: "openai-completions";
|
|
4320
4494
|
provider: string;
|
|
4321
4495
|
baseUrl: string;
|
|
4322
|
-
compat: {
|
|
4323
|
-
supportsDeveloperRole: false;
|
|
4324
|
-
};
|
|
4325
4496
|
reasoning: true;
|
|
4326
4497
|
input: "text"[];
|
|
4327
4498
|
cost: {
|
|
@@ -4333,15 +4504,12 @@ export declare const MODELS: {
|
|
|
4333
4504
|
contextWindow: number;
|
|
4334
4505
|
maxTokens: number;
|
|
4335
4506
|
};
|
|
4336
|
-
readonly "
|
|
4507
|
+
readonly "openai/gpt-oss-safeguard-20b": {
|
|
4337
4508
|
id: string;
|
|
4338
4509
|
name: string;
|
|
4339
4510
|
api: "openai-completions";
|
|
4340
4511
|
provider: string;
|
|
4341
4512
|
baseUrl: string;
|
|
4342
|
-
compat: {
|
|
4343
|
-
supportsDeveloperRole: false;
|
|
4344
|
-
};
|
|
4345
4513
|
reasoning: true;
|
|
4346
4514
|
input: "text"[];
|
|
4347
4515
|
cost: {
|
|
@@ -4353,15 +4521,12 @@ export declare const MODELS: {
|
|
|
4353
4521
|
contextWindow: number;
|
|
4354
4522
|
maxTokens: number;
|
|
4355
4523
|
};
|
|
4356
|
-
readonly "
|
|
4524
|
+
readonly "qwen-qwq-32b": {
|
|
4357
4525
|
id: string;
|
|
4358
4526
|
name: string;
|
|
4359
4527
|
api: "openai-completions";
|
|
4360
4528
|
provider: string;
|
|
4361
4529
|
baseUrl: string;
|
|
4362
|
-
compat: {
|
|
4363
|
-
supportsDeveloperRole: false;
|
|
4364
|
-
};
|
|
4365
4530
|
reasoning: true;
|
|
4366
4531
|
input: "text"[];
|
|
4367
4532
|
cost: {
|
|
@@ -4373,16 +4538,13 @@ export declare const MODELS: {
|
|
|
4373
4538
|
contextWindow: number;
|
|
4374
4539
|
maxTokens: number;
|
|
4375
4540
|
};
|
|
4376
|
-
readonly "
|
|
4541
|
+
readonly "qwen/qwen3-32b": {
|
|
4377
4542
|
id: string;
|
|
4378
4543
|
name: string;
|
|
4379
4544
|
api: "openai-completions";
|
|
4380
4545
|
provider: string;
|
|
4381
4546
|
baseUrl: string;
|
|
4382
|
-
|
|
4383
|
-
supportsDeveloperRole: false;
|
|
4384
|
-
};
|
|
4385
|
-
reasoning: false;
|
|
4547
|
+
reasoning: true;
|
|
4386
4548
|
input: "text"[];
|
|
4387
4549
|
cost: {
|
|
4388
4550
|
input: number;
|
|
@@ -4393,7 +4555,9 @@ export declare const MODELS: {
|
|
|
4393
4555
|
contextWindow: number;
|
|
4394
4556
|
maxTokens: number;
|
|
4395
4557
|
};
|
|
4396
|
-
|
|
4558
|
+
};
|
|
4559
|
+
readonly huggingface: {
|
|
4560
|
+
readonly "MiniMaxAI/MiniMax-M2.1": {
|
|
4397
4561
|
id: string;
|
|
4398
4562
|
name: string;
|
|
4399
4563
|
api: "openai-completions";
|
|
@@ -4402,7 +4566,7 @@ export declare const MODELS: {
|
|
|
4402
4566
|
compat: {
|
|
4403
4567
|
supportsDeveloperRole: false;
|
|
4404
4568
|
};
|
|
4405
|
-
reasoning:
|
|
4569
|
+
reasoning: true;
|
|
4406
4570
|
input: "text"[];
|
|
4407
4571
|
cost: {
|
|
4408
4572
|
input: number;
|
|
@@ -4413,7 +4577,7 @@ export declare const MODELS: {
|
|
|
4413
4577
|
contextWindow: number;
|
|
4414
4578
|
maxTokens: number;
|
|
4415
4579
|
};
|
|
4416
|
-
readonly "
|
|
4580
|
+
readonly "MiniMaxAI/MiniMax-M2.5": {
|
|
4417
4581
|
id: string;
|
|
4418
4582
|
name: string;
|
|
4419
4583
|
api: "openai-completions";
|
|
@@ -4422,7 +4586,7 @@ export declare const MODELS: {
|
|
|
4422
4586
|
compat: {
|
|
4423
4587
|
supportsDeveloperRole: false;
|
|
4424
4588
|
};
|
|
4425
|
-
reasoning:
|
|
4589
|
+
reasoning: true;
|
|
4426
4590
|
input: "text"[];
|
|
4427
4591
|
cost: {
|
|
4428
4592
|
input: number;
|
|
@@ -4433,7 +4597,7 @@ export declare const MODELS: {
|
|
|
4433
4597
|
contextWindow: number;
|
|
4434
4598
|
maxTokens: number;
|
|
4435
4599
|
};
|
|
4436
|
-
readonly "
|
|
4600
|
+
readonly "MiniMaxAI/MiniMax-M2.7": {
|
|
4437
4601
|
id: string;
|
|
4438
4602
|
name: string;
|
|
4439
4603
|
api: "openai-completions";
|
|
@@ -4442,7 +4606,7 @@ export declare const MODELS: {
|
|
|
4442
4606
|
compat: {
|
|
4443
4607
|
supportsDeveloperRole: false;
|
|
4444
4608
|
};
|
|
4445
|
-
reasoning:
|
|
4609
|
+
reasoning: true;
|
|
4446
4610
|
input: "text"[];
|
|
4447
4611
|
cost: {
|
|
4448
4612
|
input: number;
|
|
@@ -4453,7 +4617,7 @@ export declare const MODELS: {
|
|
|
4453
4617
|
contextWindow: number;
|
|
4454
4618
|
maxTokens: number;
|
|
4455
4619
|
};
|
|
4456
|
-
readonly "Qwen/Qwen3
|
|
4620
|
+
readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4457
4621
|
id: string;
|
|
4458
4622
|
name: string;
|
|
4459
4623
|
api: "openai-completions";
|
|
@@ -4463,7 +4627,7 @@ export declare const MODELS: {
|
|
|
4463
4627
|
supportsDeveloperRole: false;
|
|
4464
4628
|
};
|
|
4465
4629
|
reasoning: true;
|
|
4466
|
-
input:
|
|
4630
|
+
input: "text"[];
|
|
4467
4631
|
cost: {
|
|
4468
4632
|
input: number;
|
|
4469
4633
|
output: number;
|
|
@@ -4473,7 +4637,7 @@ export declare const MODELS: {
|
|
|
4473
4637
|
contextWindow: number;
|
|
4474
4638
|
maxTokens: number;
|
|
4475
4639
|
};
|
|
4476
|
-
readonly "
|
|
4640
|
+
readonly "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
|
|
4477
4641
|
id: string;
|
|
4478
4642
|
name: string;
|
|
4479
4643
|
api: "openai-completions";
|
|
@@ -4482,7 +4646,7 @@ export declare const MODELS: {
|
|
|
4482
4646
|
compat: {
|
|
4483
4647
|
supportsDeveloperRole: false;
|
|
4484
4648
|
};
|
|
4485
|
-
reasoning:
|
|
4649
|
+
reasoning: false;
|
|
4486
4650
|
input: "text"[];
|
|
4487
4651
|
cost: {
|
|
4488
4652
|
input: number;
|
|
@@ -4493,7 +4657,7 @@ export declare const MODELS: {
|
|
|
4493
4657
|
contextWindow: number;
|
|
4494
4658
|
maxTokens: number;
|
|
4495
4659
|
};
|
|
4496
|
-
readonly "
|
|
4660
|
+
readonly "Qwen/Qwen3-Coder-Next": {
|
|
4497
4661
|
id: string;
|
|
4498
4662
|
name: string;
|
|
4499
4663
|
api: "openai-completions";
|
|
@@ -4502,7 +4666,7 @@ export declare const MODELS: {
|
|
|
4502
4666
|
compat: {
|
|
4503
4667
|
supportsDeveloperRole: false;
|
|
4504
4668
|
};
|
|
4505
|
-
reasoning:
|
|
4669
|
+
reasoning: false;
|
|
4506
4670
|
input: "text"[];
|
|
4507
4671
|
cost: {
|
|
4508
4672
|
input: number;
|
|
@@ -4513,7 +4677,7 @@ export declare const MODELS: {
|
|
|
4513
4677
|
contextWindow: number;
|
|
4514
4678
|
maxTokens: number;
|
|
4515
4679
|
};
|
|
4516
|
-
readonly "
|
|
4680
|
+
readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
|
|
4517
4681
|
id: string;
|
|
4518
4682
|
name: string;
|
|
4519
4683
|
api: "openai-completions";
|
|
@@ -4522,7 +4686,7 @@ export declare const MODELS: {
|
|
|
4522
4686
|
compat: {
|
|
4523
4687
|
supportsDeveloperRole: false;
|
|
4524
4688
|
};
|
|
4525
|
-
reasoning:
|
|
4689
|
+
reasoning: false;
|
|
4526
4690
|
input: "text"[];
|
|
4527
4691
|
cost: {
|
|
4528
4692
|
input: number;
|
|
@@ -4533,7 +4697,7 @@ export declare const MODELS: {
|
|
|
4533
4697
|
contextWindow: number;
|
|
4534
4698
|
maxTokens: number;
|
|
4535
4699
|
};
|
|
4536
|
-
readonly "
|
|
4700
|
+
readonly "Qwen/Qwen3-Next-80B-A3B-Thinking": {
|
|
4537
4701
|
id: string;
|
|
4538
4702
|
name: string;
|
|
4539
4703
|
api: "openai-completions";
|
|
@@ -4553,7 +4717,7 @@ export declare const MODELS: {
|
|
|
4553
4717
|
contextWindow: number;
|
|
4554
4718
|
maxTokens: number;
|
|
4555
4719
|
};
|
|
4556
|
-
readonly "
|
|
4720
|
+
readonly "Qwen/Qwen3.5-397B-A17B": {
|
|
4557
4721
|
id: string;
|
|
4558
4722
|
name: string;
|
|
4559
4723
|
api: "openai-completions";
|
|
@@ -4562,8 +4726,8 @@ export declare const MODELS: {
|
|
|
4562
4726
|
compat: {
|
|
4563
4727
|
supportsDeveloperRole: false;
|
|
4564
4728
|
};
|
|
4565
|
-
reasoning:
|
|
4566
|
-
input: "text"[];
|
|
4729
|
+
reasoning: true;
|
|
4730
|
+
input: ("image" | "text")[];
|
|
4567
4731
|
cost: {
|
|
4568
4732
|
input: number;
|
|
4569
4733
|
output: number;
|
|
@@ -4573,7 +4737,7 @@ export declare const MODELS: {
|
|
|
4573
4737
|
contextWindow: number;
|
|
4574
4738
|
maxTokens: number;
|
|
4575
4739
|
};
|
|
4576
|
-
readonly "
|
|
4740
|
+
readonly "XiaomiMiMo/MiMo-V2-Flash": {
|
|
4577
4741
|
id: string;
|
|
4578
4742
|
name: string;
|
|
4579
4743
|
api: "openai-completions";
|
|
@@ -4593,7 +4757,7 @@ export declare const MODELS: {
|
|
|
4593
4757
|
contextWindow: number;
|
|
4594
4758
|
maxTokens: number;
|
|
4595
4759
|
};
|
|
4596
|
-
readonly "
|
|
4760
|
+
readonly "deepseek-ai/DeepSeek-R1-0528": {
|
|
4597
4761
|
id: string;
|
|
4598
4762
|
name: string;
|
|
4599
4763
|
api: "openai-completions";
|
|
@@ -4603,7 +4767,7 @@ export declare const MODELS: {
|
|
|
4603
4767
|
supportsDeveloperRole: false;
|
|
4604
4768
|
};
|
|
4605
4769
|
reasoning: true;
|
|
4606
|
-
input:
|
|
4770
|
+
input: "text"[];
|
|
4607
4771
|
cost: {
|
|
4608
4772
|
input: number;
|
|
4609
4773
|
output: number;
|
|
@@ -4613,7 +4777,7 @@ export declare const MODELS: {
|
|
|
4613
4777
|
contextWindow: number;
|
|
4614
4778
|
maxTokens: number;
|
|
4615
4779
|
};
|
|
4616
|
-
readonly "
|
|
4780
|
+
readonly "deepseek-ai/DeepSeek-V3.2": {
|
|
4617
4781
|
id: string;
|
|
4618
4782
|
name: string;
|
|
4619
4783
|
api: "openai-completions";
|
|
@@ -4633,7 +4797,7 @@ export declare const MODELS: {
|
|
|
4633
4797
|
contextWindow: number;
|
|
4634
4798
|
maxTokens: number;
|
|
4635
4799
|
};
|
|
4636
|
-
readonly "
|
|
4800
|
+
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
4637
4801
|
id: string;
|
|
4638
4802
|
name: string;
|
|
4639
4803
|
api: "openai-completions";
|
|
@@ -4642,7 +4806,7 @@ export declare const MODELS: {
|
|
|
4642
4806
|
compat: {
|
|
4643
4807
|
supportsDeveloperRole: false;
|
|
4644
4808
|
};
|
|
4645
|
-
reasoning:
|
|
4809
|
+
reasoning: false;
|
|
4646
4810
|
input: "text"[];
|
|
4647
4811
|
cost: {
|
|
4648
4812
|
input: number;
|
|
@@ -4653,7 +4817,7 @@ export declare const MODELS: {
|
|
|
4653
4817
|
contextWindow: number;
|
|
4654
4818
|
maxTokens: number;
|
|
4655
4819
|
};
|
|
4656
|
-
readonly "
|
|
4820
|
+
readonly "moonshotai/Kimi-K2-Instruct-0905": {
|
|
4657
4821
|
id: string;
|
|
4658
4822
|
name: string;
|
|
4659
4823
|
api: "openai-completions";
|
|
@@ -4662,7 +4826,7 @@ export declare const MODELS: {
|
|
|
4662
4826
|
compat: {
|
|
4663
4827
|
supportsDeveloperRole: false;
|
|
4664
4828
|
};
|
|
4665
|
-
reasoning:
|
|
4829
|
+
reasoning: false;
|
|
4666
4830
|
input: "text"[];
|
|
4667
4831
|
cost: {
|
|
4668
4832
|
input: number;
|
|
@@ -4673,31 +4837,15 @@ export declare const MODELS: {
|
|
|
4673
4837
|
contextWindow: number;
|
|
4674
4838
|
maxTokens: number;
|
|
4675
4839
|
};
|
|
4676
|
-
|
|
4677
|
-
readonly "kimi-coding": {
|
|
4678
|
-
readonly k2p5: {
|
|
4840
|
+
readonly "moonshotai/Kimi-K2-Thinking": {
|
|
4679
4841
|
id: string;
|
|
4680
4842
|
name: string;
|
|
4681
|
-
api: "
|
|
4843
|
+
api: "openai-completions";
|
|
4682
4844
|
provider: string;
|
|
4683
4845
|
baseUrl: string;
|
|
4684
|
-
|
|
4685
|
-
|
|
4686
|
-
cost: {
|
|
4687
|
-
input: number;
|
|
4688
|
-
output: number;
|
|
4689
|
-
cacheRead: number;
|
|
4690
|
-
cacheWrite: number;
|
|
4846
|
+
compat: {
|
|
4847
|
+
supportsDeveloperRole: false;
|
|
4691
4848
|
};
|
|
4692
|
-
contextWindow: number;
|
|
4693
|
-
maxTokens: number;
|
|
4694
|
-
};
|
|
4695
|
-
readonly "kimi-k2-thinking": {
|
|
4696
|
-
id: string;
|
|
4697
|
-
name: string;
|
|
4698
|
-
api: "anthropic-messages";
|
|
4699
|
-
provider: string;
|
|
4700
|
-
baseUrl: string;
|
|
4701
4849
|
reasoning: true;
|
|
4702
4850
|
input: "text"[];
|
|
4703
4851
|
cost: {
|
|
@@ -4709,33 +4857,17 @@ export declare const MODELS: {
|
|
|
4709
4857
|
contextWindow: number;
|
|
4710
4858
|
maxTokens: number;
|
|
4711
4859
|
};
|
|
4712
|
-
|
|
4713
|
-
readonly minimax: {
|
|
4714
|
-
readonly "MiniMax-M2": {
|
|
4860
|
+
readonly "moonshotai/Kimi-K2.5": {
|
|
4715
4861
|
id: string;
|
|
4716
4862
|
name: string;
|
|
4717
|
-
api: "
|
|
4863
|
+
api: "openai-completions";
|
|
4718
4864
|
provider: string;
|
|
4719
4865
|
baseUrl: string;
|
|
4720
|
-
|
|
4721
|
-
|
|
4722
|
-
cost: {
|
|
4723
|
-
input: number;
|
|
4724
|
-
output: number;
|
|
4725
|
-
cacheRead: number;
|
|
4726
|
-
cacheWrite: number;
|
|
4866
|
+
compat: {
|
|
4867
|
+
supportsDeveloperRole: false;
|
|
4727
4868
|
};
|
|
4728
|
-
contextWindow: number;
|
|
4729
|
-
maxTokens: number;
|
|
4730
|
-
};
|
|
4731
|
-
readonly "MiniMax-M2.1": {
|
|
4732
|
-
id: string;
|
|
4733
|
-
name: string;
|
|
4734
|
-
api: "anthropic-messages";
|
|
4735
|
-
provider: string;
|
|
4736
|
-
baseUrl: string;
|
|
4737
4869
|
reasoning: true;
|
|
4738
|
-
input: "text"[];
|
|
4870
|
+
input: ("image" | "text")[];
|
|
4739
4871
|
cost: {
|
|
4740
4872
|
input: number;
|
|
4741
4873
|
output: number;
|
|
@@ -4745,14 +4877,17 @@ export declare const MODELS: {
|
|
|
4745
4877
|
contextWindow: number;
|
|
4746
4878
|
maxTokens: number;
|
|
4747
4879
|
};
|
|
4748
|
-
readonly "
|
|
4880
|
+
readonly "moonshotai/Kimi-K2.6": {
|
|
4749
4881
|
id: string;
|
|
4750
4882
|
name: string;
|
|
4751
|
-
api: "
|
|
4883
|
+
api: "openai-completions";
|
|
4752
4884
|
provider: string;
|
|
4753
4885
|
baseUrl: string;
|
|
4886
|
+
compat: {
|
|
4887
|
+
supportsDeveloperRole: false;
|
|
4888
|
+
};
|
|
4754
4889
|
reasoning: true;
|
|
4755
|
-
input: "text"[];
|
|
4890
|
+
input: ("image" | "text")[];
|
|
4756
4891
|
cost: {
|
|
4757
4892
|
input: number;
|
|
4758
4893
|
output: number;
|
|
@@ -4762,12 +4897,15 @@ export declare const MODELS: {
|
|
|
4762
4897
|
contextWindow: number;
|
|
4763
4898
|
maxTokens: number;
|
|
4764
4899
|
};
|
|
4765
|
-
readonly "
|
|
4900
|
+
readonly "zai-org/GLM-4.7": {
|
|
4766
4901
|
id: string;
|
|
4767
4902
|
name: string;
|
|
4768
|
-
api: "
|
|
4903
|
+
api: "openai-completions";
|
|
4769
4904
|
provider: string;
|
|
4770
4905
|
baseUrl: string;
|
|
4906
|
+
compat: {
|
|
4907
|
+
supportsDeveloperRole: false;
|
|
4908
|
+
};
|
|
4771
4909
|
reasoning: true;
|
|
4772
4910
|
input: "text"[];
|
|
4773
4911
|
cost: {
|
|
@@ -4779,12 +4917,15 @@ export declare const MODELS: {
|
|
|
4779
4917
|
contextWindow: number;
|
|
4780
4918
|
maxTokens: number;
|
|
4781
4919
|
};
|
|
4782
|
-
readonly "
|
|
4920
|
+
readonly "zai-org/GLM-4.7-Flash": {
|
|
4783
4921
|
id: string;
|
|
4784
4922
|
name: string;
|
|
4785
|
-
api: "
|
|
4923
|
+
api: "openai-completions";
|
|
4786
4924
|
provider: string;
|
|
4787
4925
|
baseUrl: string;
|
|
4926
|
+
compat: {
|
|
4927
|
+
supportsDeveloperRole: false;
|
|
4928
|
+
};
|
|
4788
4929
|
reasoning: true;
|
|
4789
4930
|
input: "text"[];
|
|
4790
4931
|
cost: {
|
|
@@ -4796,12 +4937,15 @@ export declare const MODELS: {
|
|
|
4796
4937
|
contextWindow: number;
|
|
4797
4938
|
maxTokens: number;
|
|
4798
4939
|
};
|
|
4799
|
-
readonly "
|
|
4940
|
+
readonly "zai-org/GLM-5": {
|
|
4800
4941
|
id: string;
|
|
4801
4942
|
name: string;
|
|
4802
|
-
api: "
|
|
4943
|
+
api: "openai-completions";
|
|
4803
4944
|
provider: string;
|
|
4804
4945
|
baseUrl: string;
|
|
4946
|
+
compat: {
|
|
4947
|
+
supportsDeveloperRole: false;
|
|
4948
|
+
};
|
|
4805
4949
|
reasoning: true;
|
|
4806
4950
|
input: "text"[];
|
|
4807
4951
|
cost: {
|
|
@@ -4813,12 +4957,15 @@ export declare const MODELS: {
|
|
|
4813
4957
|
contextWindow: number;
|
|
4814
4958
|
maxTokens: number;
|
|
4815
4959
|
};
|
|
4816
|
-
readonly "
|
|
4960
|
+
readonly "zai-org/GLM-5.1": {
|
|
4817
4961
|
id: string;
|
|
4818
4962
|
name: string;
|
|
4819
|
-
api: "
|
|
4963
|
+
api: "openai-completions";
|
|
4820
4964
|
provider: string;
|
|
4821
4965
|
baseUrl: string;
|
|
4966
|
+
compat: {
|
|
4967
|
+
supportsDeveloperRole: false;
|
|
4968
|
+
};
|
|
4822
4969
|
reasoning: true;
|
|
4823
4970
|
input: "text"[];
|
|
4824
4971
|
cost: {
|
|
@@ -4831,15 +4978,15 @@ export declare const MODELS: {
|
|
|
4831
4978
|
maxTokens: number;
|
|
4832
4979
|
};
|
|
4833
4980
|
};
|
|
4834
|
-
readonly "
|
|
4835
|
-
readonly
|
|
4981
|
+
readonly "kimi-coding": {
|
|
4982
|
+
readonly k2p5: {
|
|
4836
4983
|
id: string;
|
|
4837
4984
|
name: string;
|
|
4838
4985
|
api: "anthropic-messages";
|
|
4839
4986
|
provider: string;
|
|
4840
4987
|
baseUrl: string;
|
|
4841
4988
|
reasoning: true;
|
|
4842
|
-
input: "text"[];
|
|
4989
|
+
input: ("image" | "text")[];
|
|
4843
4990
|
cost: {
|
|
4844
4991
|
input: number;
|
|
4845
4992
|
output: number;
|
|
@@ -4849,14 +4996,14 @@ export declare const MODELS: {
|
|
|
4849
4996
|
contextWindow: number;
|
|
4850
4997
|
maxTokens: number;
|
|
4851
4998
|
};
|
|
4852
|
-
readonly
|
|
4999
|
+
readonly k2p6: {
|
|
4853
5000
|
id: string;
|
|
4854
5001
|
name: string;
|
|
4855
5002
|
api: "anthropic-messages";
|
|
4856
5003
|
provider: string;
|
|
4857
5004
|
baseUrl: string;
|
|
4858
5005
|
reasoning: true;
|
|
4859
|
-
input: "text"[];
|
|
5006
|
+
input: ("image" | "text")[];
|
|
4860
5007
|
cost: {
|
|
4861
5008
|
input: number;
|
|
4862
5009
|
output: number;
|
|
@@ -4866,7 +5013,7 @@ export declare const MODELS: {
|
|
|
4866
5013
|
contextWindow: number;
|
|
4867
5014
|
maxTokens: number;
|
|
4868
5015
|
};
|
|
4869
|
-
readonly "
|
|
5016
|
+
readonly "kimi-k2-thinking": {
|
|
4870
5017
|
id: string;
|
|
4871
5018
|
name: string;
|
|
4872
5019
|
api: "anthropic-messages";
|
|
@@ -4883,7 +5030,9 @@ export declare const MODELS: {
|
|
|
4883
5030
|
contextWindow: number;
|
|
4884
5031
|
maxTokens: number;
|
|
4885
5032
|
};
|
|
4886
|
-
|
|
5033
|
+
};
|
|
5034
|
+
readonly minimax: {
|
|
5035
|
+
readonly "MiniMax-M2.7": {
|
|
4887
5036
|
id: string;
|
|
4888
5037
|
name: string;
|
|
4889
5038
|
api: "anthropic-messages";
|
|
@@ -4900,7 +5049,7 @@ export declare const MODELS: {
|
|
|
4900
5049
|
contextWindow: number;
|
|
4901
5050
|
maxTokens: number;
|
|
4902
5051
|
};
|
|
4903
|
-
readonly "MiniMax-M2.
|
|
5052
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4904
5053
|
id: string;
|
|
4905
5054
|
name: string;
|
|
4906
5055
|
api: "anthropic-messages";
|
|
@@ -4917,6 +5066,8 @@ export declare const MODELS: {
|
|
|
4917
5066
|
contextWindow: number;
|
|
4918
5067
|
maxTokens: number;
|
|
4919
5068
|
};
|
|
5069
|
+
};
|
|
5070
|
+
readonly "minimax-cn": {
|
|
4920
5071
|
readonly "MiniMax-M2.7": {
|
|
4921
5072
|
id: string;
|
|
4922
5073
|
name: string;
|
|
@@ -4956,7 +5107,7 @@ export declare const MODELS: {
|
|
|
4956
5107
|
readonly "codestral-latest": {
|
|
4957
5108
|
id: string;
|
|
4958
5109
|
name: string;
|
|
4959
|
-
api: "
|
|
5110
|
+
api: "openai-completions";
|
|
4960
5111
|
provider: string;
|
|
4961
5112
|
baseUrl: string;
|
|
4962
5113
|
reasoning: false;
|
|
@@ -4973,7 +5124,7 @@ export declare const MODELS: {
|
|
|
4973
5124
|
readonly "devstral-2512": {
|
|
4974
5125
|
id: string;
|
|
4975
5126
|
name: string;
|
|
4976
|
-
api: "
|
|
5127
|
+
api: "openai-completions";
|
|
4977
5128
|
provider: string;
|
|
4978
5129
|
baseUrl: string;
|
|
4979
5130
|
reasoning: false;
|
|
@@ -4990,7 +5141,7 @@ export declare const MODELS: {
|
|
|
4990
5141
|
readonly "devstral-medium-2507": {
|
|
4991
5142
|
id: string;
|
|
4992
5143
|
name: string;
|
|
4993
|
-
api: "
|
|
5144
|
+
api: "openai-completions";
|
|
4994
5145
|
provider: string;
|
|
4995
5146
|
baseUrl: string;
|
|
4996
5147
|
reasoning: false;
|
|
@@ -5007,7 +5158,7 @@ export declare const MODELS: {
|
|
|
5007
5158
|
readonly "devstral-medium-latest": {
|
|
5008
5159
|
id: string;
|
|
5009
5160
|
name: string;
|
|
5010
|
-
api: "
|
|
5161
|
+
api: "openai-completions";
|
|
5011
5162
|
provider: string;
|
|
5012
5163
|
baseUrl: string;
|
|
5013
5164
|
reasoning: false;
|
|
@@ -5024,7 +5175,7 @@ export declare const MODELS: {
|
|
|
5024
5175
|
readonly "devstral-small-2505": {
|
|
5025
5176
|
id: string;
|
|
5026
5177
|
name: string;
|
|
5027
|
-
api: "
|
|
5178
|
+
api: "openai-completions";
|
|
5028
5179
|
provider: string;
|
|
5029
5180
|
baseUrl: string;
|
|
5030
5181
|
reasoning: false;
|
|
@@ -5041,7 +5192,7 @@ export declare const MODELS: {
|
|
|
5041
5192
|
readonly "devstral-small-2507": {
|
|
5042
5193
|
id: string;
|
|
5043
5194
|
name: string;
|
|
5044
|
-
api: "
|
|
5195
|
+
api: "openai-completions";
|
|
5045
5196
|
provider: string;
|
|
5046
5197
|
baseUrl: string;
|
|
5047
5198
|
reasoning: false;
|
|
@@ -5058,7 +5209,7 @@ export declare const MODELS: {
|
|
|
5058
5209
|
readonly "labs-devstral-small-2512": {
|
|
5059
5210
|
id: string;
|
|
5060
5211
|
name: string;
|
|
5061
|
-
api: "
|
|
5212
|
+
api: "openai-completions";
|
|
5062
5213
|
provider: string;
|
|
5063
5214
|
baseUrl: string;
|
|
5064
5215
|
reasoning: false;
|
|
@@ -5075,7 +5226,7 @@ export declare const MODELS: {
|
|
|
5075
5226
|
readonly "magistral-medium-latest": {
|
|
5076
5227
|
id: string;
|
|
5077
5228
|
name: string;
|
|
5078
|
-
api: "
|
|
5229
|
+
api: "openai-completions";
|
|
5079
5230
|
provider: string;
|
|
5080
5231
|
baseUrl: string;
|
|
5081
5232
|
reasoning: true;
|
|
@@ -5092,7 +5243,7 @@ export declare const MODELS: {
|
|
|
5092
5243
|
readonly "magistral-small": {
|
|
5093
5244
|
id: string;
|
|
5094
5245
|
name: string;
|
|
5095
|
-
api: "
|
|
5246
|
+
api: "openai-completions";
|
|
5096
5247
|
provider: string;
|
|
5097
5248
|
baseUrl: string;
|
|
5098
5249
|
reasoning: true;
|
|
@@ -5109,7 +5260,7 @@ export declare const MODELS: {
|
|
|
5109
5260
|
readonly "ministral-3b-latest": {
|
|
5110
5261
|
id: string;
|
|
5111
5262
|
name: string;
|
|
5112
|
-
api: "
|
|
5263
|
+
api: "openai-completions";
|
|
5113
5264
|
provider: string;
|
|
5114
5265
|
baseUrl: string;
|
|
5115
5266
|
reasoning: false;
|
|
@@ -5126,7 +5277,7 @@ export declare const MODELS: {
|
|
|
5126
5277
|
readonly "ministral-8b-latest": {
|
|
5127
5278
|
id: string;
|
|
5128
5279
|
name: string;
|
|
5129
|
-
api: "
|
|
5280
|
+
api: "openai-completions";
|
|
5130
5281
|
provider: string;
|
|
5131
5282
|
baseUrl: string;
|
|
5132
5283
|
reasoning: false;
|
|
@@ -5143,7 +5294,7 @@ export declare const MODELS: {
|
|
|
5143
5294
|
readonly "mistral-large-2411": {
|
|
5144
5295
|
id: string;
|
|
5145
5296
|
name: string;
|
|
5146
|
-
api: "
|
|
5297
|
+
api: "openai-completions";
|
|
5147
5298
|
provider: string;
|
|
5148
5299
|
baseUrl: string;
|
|
5149
5300
|
reasoning: false;
|
|
@@ -5160,7 +5311,7 @@ export declare const MODELS: {
|
|
|
5160
5311
|
readonly "mistral-large-2512": {
|
|
5161
5312
|
id: string;
|
|
5162
5313
|
name: string;
|
|
5163
|
-
api: "
|
|
5314
|
+
api: "openai-completions";
|
|
5164
5315
|
provider: string;
|
|
5165
5316
|
baseUrl: string;
|
|
5166
5317
|
reasoning: false;
|
|
@@ -5177,7 +5328,7 @@ export declare const MODELS: {
|
|
|
5177
5328
|
readonly "mistral-large-latest": {
|
|
5178
5329
|
id: string;
|
|
5179
5330
|
name: string;
|
|
5180
|
-
api: "
|
|
5331
|
+
api: "openai-completions";
|
|
5181
5332
|
provider: string;
|
|
5182
5333
|
baseUrl: string;
|
|
5183
5334
|
reasoning: false;
|
|
@@ -5194,7 +5345,7 @@ export declare const MODELS: {
|
|
|
5194
5345
|
readonly "mistral-medium-2505": {
|
|
5195
5346
|
id: string;
|
|
5196
5347
|
name: string;
|
|
5197
|
-
api: "
|
|
5348
|
+
api: "openai-completions";
|
|
5198
5349
|
provider: string;
|
|
5199
5350
|
baseUrl: string;
|
|
5200
5351
|
reasoning: false;
|
|
@@ -5211,7 +5362,7 @@ export declare const MODELS: {
|
|
|
5211
5362
|
readonly "mistral-medium-2508": {
|
|
5212
5363
|
id: string;
|
|
5213
5364
|
name: string;
|
|
5214
|
-
api: "
|
|
5365
|
+
api: "openai-completions";
|
|
5215
5366
|
provider: string;
|
|
5216
5367
|
baseUrl: string;
|
|
5217
5368
|
reasoning: false;
|
|
@@ -5228,7 +5379,7 @@ export declare const MODELS: {
|
|
|
5228
5379
|
readonly "mistral-medium-latest": {
|
|
5229
5380
|
id: string;
|
|
5230
5381
|
name: string;
|
|
5231
|
-
api: "
|
|
5382
|
+
api: "openai-completions";
|
|
5232
5383
|
provider: string;
|
|
5233
5384
|
baseUrl: string;
|
|
5234
5385
|
reasoning: false;
|
|
@@ -5245,7 +5396,7 @@ export declare const MODELS: {
|
|
|
5245
5396
|
readonly "mistral-nemo": {
|
|
5246
5397
|
id: string;
|
|
5247
5398
|
name: string;
|
|
5248
|
-
api: "
|
|
5399
|
+
api: "openai-completions";
|
|
5249
5400
|
provider: string;
|
|
5250
5401
|
baseUrl: string;
|
|
5251
5402
|
reasoning: false;
|
|
@@ -5262,7 +5413,7 @@ export declare const MODELS: {
|
|
|
5262
5413
|
readonly "mistral-small-2506": {
|
|
5263
5414
|
id: string;
|
|
5264
5415
|
name: string;
|
|
5265
|
-
api: "
|
|
5416
|
+
api: "openai-completions";
|
|
5266
5417
|
provider: string;
|
|
5267
5418
|
baseUrl: string;
|
|
5268
5419
|
reasoning: false;
|
|
@@ -5276,13 +5427,13 @@ export declare const MODELS: {
|
|
|
5276
5427
|
contextWindow: number;
|
|
5277
5428
|
maxTokens: number;
|
|
5278
5429
|
};
|
|
5279
|
-
readonly "mistral-small-
|
|
5430
|
+
readonly "mistral-small-2603": {
|
|
5280
5431
|
id: string;
|
|
5281
5432
|
name: string;
|
|
5282
|
-
api: "
|
|
5433
|
+
api: "openai-completions";
|
|
5283
5434
|
provider: string;
|
|
5284
5435
|
baseUrl: string;
|
|
5285
|
-
reasoning:
|
|
5436
|
+
reasoning: true;
|
|
5286
5437
|
input: ("image" | "text")[];
|
|
5287
5438
|
cost: {
|
|
5288
5439
|
input: number;
|
|
@@ -5293,14 +5444,14 @@ export declare const MODELS: {
|
|
|
5293
5444
|
contextWindow: number;
|
|
5294
5445
|
maxTokens: number;
|
|
5295
5446
|
};
|
|
5296
|
-
readonly "
|
|
5447
|
+
readonly "mistral-small-latest": {
|
|
5297
5448
|
id: string;
|
|
5298
5449
|
name: string;
|
|
5299
|
-
api: "
|
|
5450
|
+
api: "openai-completions";
|
|
5300
5451
|
provider: string;
|
|
5301
5452
|
baseUrl: string;
|
|
5302
|
-
reasoning:
|
|
5303
|
-
input: "text"[];
|
|
5453
|
+
reasoning: true;
|
|
5454
|
+
input: ("image" | "text")[];
|
|
5304
5455
|
cost: {
|
|
5305
5456
|
input: number;
|
|
5306
5457
|
output: number;
|
|
@@ -5310,10 +5461,10 @@ export declare const MODELS: {
|
|
|
5310
5461
|
contextWindow: number;
|
|
5311
5462
|
maxTokens: number;
|
|
5312
5463
|
};
|
|
5313
|
-
readonly "open-
|
|
5464
|
+
readonly "open-mistral-7b": {
|
|
5314
5465
|
id: string;
|
|
5315
5466
|
name: string;
|
|
5316
|
-
api: "
|
|
5467
|
+
api: "openai-completions";
|
|
5317
5468
|
provider: string;
|
|
5318
5469
|
baseUrl: string;
|
|
5319
5470
|
reasoning: false;
|
|
@@ -5327,10 +5478,10 @@ export declare const MODELS: {
|
|
|
5327
5478
|
contextWindow: number;
|
|
5328
5479
|
maxTokens: number;
|
|
5329
5480
|
};
|
|
5330
|
-
readonly "open-mixtral-
|
|
5481
|
+
readonly "open-mixtral-8x22b": {
|
|
5331
5482
|
id: string;
|
|
5332
5483
|
name: string;
|
|
5333
|
-
api: "
|
|
5484
|
+
api: "openai-completions";
|
|
5334
5485
|
provider: string;
|
|
5335
5486
|
baseUrl: string;
|
|
5336
5487
|
reasoning: false;
|
|
@@ -5344,14 +5495,14 @@ export declare const MODELS: {
|
|
|
5344
5495
|
contextWindow: number;
|
|
5345
5496
|
maxTokens: number;
|
|
5346
5497
|
};
|
|
5347
|
-
readonly "
|
|
5498
|
+
readonly "open-mixtral-8x7b": {
|
|
5348
5499
|
id: string;
|
|
5349
5500
|
name: string;
|
|
5350
|
-
api: "
|
|
5501
|
+
api: "openai-completions";
|
|
5351
5502
|
provider: string;
|
|
5352
5503
|
baseUrl: string;
|
|
5353
5504
|
reasoning: false;
|
|
5354
|
-
input:
|
|
5505
|
+
input: "text"[];
|
|
5355
5506
|
cost: {
|
|
5356
5507
|
input: number;
|
|
5357
5508
|
output: number;
|
|
@@ -5361,10 +5512,10 @@ export declare const MODELS: {
|
|
|
5361
5512
|
contextWindow: number;
|
|
5362
5513
|
maxTokens: number;
|
|
5363
5514
|
};
|
|
5364
|
-
readonly "pixtral-
|
|
5515
|
+
readonly "pixtral-12b": {
|
|
5365
5516
|
id: string;
|
|
5366
5517
|
name: string;
|
|
5367
|
-
api: "
|
|
5518
|
+
api: "openai-completions";
|
|
5368
5519
|
provider: string;
|
|
5369
5520
|
baseUrl: string;
|
|
5370
5521
|
reasoning: false;
|
|
@@ -5378,16 +5529,14 @@ export declare const MODELS: {
|
|
|
5378
5529
|
contextWindow: number;
|
|
5379
5530
|
maxTokens: number;
|
|
5380
5531
|
};
|
|
5381
|
-
|
|
5382
|
-
readonly openai: {
|
|
5383
|
-
readonly "codex-mini-latest": {
|
|
5532
|
+
readonly "pixtral-large-latest": {
|
|
5384
5533
|
id: string;
|
|
5385
5534
|
name: string;
|
|
5386
|
-
api: "openai-
|
|
5535
|
+
api: "openai-completions";
|
|
5387
5536
|
provider: string;
|
|
5388
5537
|
baseUrl: string;
|
|
5389
|
-
reasoning:
|
|
5390
|
-
input: "text"[];
|
|
5538
|
+
reasoning: false;
|
|
5539
|
+
input: ("image" | "text")[];
|
|
5391
5540
|
cost: {
|
|
5392
5541
|
input: number;
|
|
5393
5542
|
output: number;
|
|
@@ -5397,6 +5546,8 @@ export declare const MODELS: {
|
|
|
5397
5546
|
contextWindow: number;
|
|
5398
5547
|
maxTokens: number;
|
|
5399
5548
|
};
|
|
5549
|
+
};
|
|
5550
|
+
readonly openai: {
|
|
5400
5551
|
readonly "gpt-4": {
|
|
5401
5552
|
id: string;
|
|
5402
5553
|
name: string;
|
|
@@ -5822,6 +5973,23 @@ export declare const MODELS: {
|
|
|
5822
5973
|
contextWindow: number;
|
|
5823
5974
|
maxTokens: number;
|
|
5824
5975
|
};
|
|
5976
|
+
readonly "gpt-5.3-chat-latest": {
|
|
5977
|
+
id: string;
|
|
5978
|
+
name: string;
|
|
5979
|
+
api: "openai-responses";
|
|
5980
|
+
provider: string;
|
|
5981
|
+
baseUrl: string;
|
|
5982
|
+
reasoning: false;
|
|
5983
|
+
input: ("image" | "text")[];
|
|
5984
|
+
cost: {
|
|
5985
|
+
input: number;
|
|
5986
|
+
output: number;
|
|
5987
|
+
cacheRead: number;
|
|
5988
|
+
cacheWrite: number;
|
|
5989
|
+
};
|
|
5990
|
+
contextWindow: number;
|
|
5991
|
+
maxTokens: number;
|
|
5992
|
+
};
|
|
5825
5993
|
readonly "gpt-5.3-codex": {
|
|
5826
5994
|
id: string;
|
|
5827
5995
|
name: string;
|
|
@@ -6046,7 +6214,249 @@ export declare const MODELS: {
|
|
|
6046
6214
|
readonly "o4-mini-deep-research": {
|
|
6047
6215
|
id: string;
|
|
6048
6216
|
name: string;
|
|
6049
|
-
api: "openai-responses";
|
|
6217
|
+
api: "openai-responses";
|
|
6218
|
+
provider: string;
|
|
6219
|
+
baseUrl: string;
|
|
6220
|
+
reasoning: true;
|
|
6221
|
+
input: ("image" | "text")[];
|
|
6222
|
+
cost: {
|
|
6223
|
+
input: number;
|
|
6224
|
+
output: number;
|
|
6225
|
+
cacheRead: number;
|
|
6226
|
+
cacheWrite: number;
|
|
6227
|
+
};
|
|
6228
|
+
contextWindow: number;
|
|
6229
|
+
maxTokens: number;
|
|
6230
|
+
};
|
|
6231
|
+
};
|
|
6232
|
+
readonly "openai-codex": {
|
|
6233
|
+
readonly "gpt-5.1": {
|
|
6234
|
+
id: string;
|
|
6235
|
+
name: string;
|
|
6236
|
+
api: "openai-codex-responses";
|
|
6237
|
+
provider: string;
|
|
6238
|
+
baseUrl: string;
|
|
6239
|
+
reasoning: true;
|
|
6240
|
+
input: ("image" | "text")[];
|
|
6241
|
+
cost: {
|
|
6242
|
+
input: number;
|
|
6243
|
+
output: number;
|
|
6244
|
+
cacheRead: number;
|
|
6245
|
+
cacheWrite: number;
|
|
6246
|
+
};
|
|
6247
|
+
contextWindow: number;
|
|
6248
|
+
maxTokens: number;
|
|
6249
|
+
};
|
|
6250
|
+
readonly "gpt-5.1-codex-max": {
|
|
6251
|
+
id: string;
|
|
6252
|
+
name: string;
|
|
6253
|
+
api: "openai-codex-responses";
|
|
6254
|
+
provider: string;
|
|
6255
|
+
baseUrl: string;
|
|
6256
|
+
reasoning: true;
|
|
6257
|
+
input: ("image" | "text")[];
|
|
6258
|
+
cost: {
|
|
6259
|
+
input: number;
|
|
6260
|
+
output: number;
|
|
6261
|
+
cacheRead: number;
|
|
6262
|
+
cacheWrite: number;
|
|
6263
|
+
};
|
|
6264
|
+
contextWindow: number;
|
|
6265
|
+
maxTokens: number;
|
|
6266
|
+
};
|
|
6267
|
+
readonly "gpt-5.1-codex-mini": {
|
|
6268
|
+
id: string;
|
|
6269
|
+
name: string;
|
|
6270
|
+
api: "openai-codex-responses";
|
|
6271
|
+
provider: string;
|
|
6272
|
+
baseUrl: string;
|
|
6273
|
+
reasoning: true;
|
|
6274
|
+
input: ("image" | "text")[];
|
|
6275
|
+
cost: {
|
|
6276
|
+
input: number;
|
|
6277
|
+
output: number;
|
|
6278
|
+
cacheRead: number;
|
|
6279
|
+
cacheWrite: number;
|
|
6280
|
+
};
|
|
6281
|
+
contextWindow: number;
|
|
6282
|
+
maxTokens: number;
|
|
6283
|
+
};
|
|
6284
|
+
readonly "gpt-5.2": {
|
|
6285
|
+
id: string;
|
|
6286
|
+
name: string;
|
|
6287
|
+
api: "openai-codex-responses";
|
|
6288
|
+
provider: string;
|
|
6289
|
+
baseUrl: string;
|
|
6290
|
+
reasoning: true;
|
|
6291
|
+
input: ("image" | "text")[];
|
|
6292
|
+
cost: {
|
|
6293
|
+
input: number;
|
|
6294
|
+
output: number;
|
|
6295
|
+
cacheRead: number;
|
|
6296
|
+
cacheWrite: number;
|
|
6297
|
+
};
|
|
6298
|
+
contextWindow: number;
|
|
6299
|
+
maxTokens: number;
|
|
6300
|
+
};
|
|
6301
|
+
readonly "gpt-5.2-codex": {
|
|
6302
|
+
id: string;
|
|
6303
|
+
name: string;
|
|
6304
|
+
api: "openai-codex-responses";
|
|
6305
|
+
provider: string;
|
|
6306
|
+
baseUrl: string;
|
|
6307
|
+
reasoning: true;
|
|
6308
|
+
input: ("image" | "text")[];
|
|
6309
|
+
cost: {
|
|
6310
|
+
input: number;
|
|
6311
|
+
output: number;
|
|
6312
|
+
cacheRead: number;
|
|
6313
|
+
cacheWrite: number;
|
|
6314
|
+
};
|
|
6315
|
+
contextWindow: number;
|
|
6316
|
+
maxTokens: number;
|
|
6317
|
+
};
|
|
6318
|
+
readonly "gpt-5.3-codex": {
|
|
6319
|
+
id: string;
|
|
6320
|
+
name: string;
|
|
6321
|
+
api: "openai-codex-responses";
|
|
6322
|
+
provider: string;
|
|
6323
|
+
baseUrl: string;
|
|
6324
|
+
reasoning: true;
|
|
6325
|
+
input: ("image" | "text")[];
|
|
6326
|
+
cost: {
|
|
6327
|
+
input: number;
|
|
6328
|
+
output: number;
|
|
6329
|
+
cacheRead: number;
|
|
6330
|
+
cacheWrite: number;
|
|
6331
|
+
};
|
|
6332
|
+
contextWindow: number;
|
|
6333
|
+
maxTokens: number;
|
|
6334
|
+
};
|
|
6335
|
+
readonly "gpt-5.3-codex-spark": {
|
|
6336
|
+
id: string;
|
|
6337
|
+
name: string;
|
|
6338
|
+
api: "openai-codex-responses";
|
|
6339
|
+
provider: string;
|
|
6340
|
+
baseUrl: string;
|
|
6341
|
+
reasoning: true;
|
|
6342
|
+
input: "text"[];
|
|
6343
|
+
cost: {
|
|
6344
|
+
input: number;
|
|
6345
|
+
output: number;
|
|
6346
|
+
cacheRead: number;
|
|
6347
|
+
cacheWrite: number;
|
|
6348
|
+
};
|
|
6349
|
+
contextWindow: number;
|
|
6350
|
+
maxTokens: number;
|
|
6351
|
+
};
|
|
6352
|
+
readonly "gpt-5.4": {
|
|
6353
|
+
id: string;
|
|
6354
|
+
name: string;
|
|
6355
|
+
api: "openai-codex-responses";
|
|
6356
|
+
provider: string;
|
|
6357
|
+
baseUrl: string;
|
|
6358
|
+
reasoning: true;
|
|
6359
|
+
input: ("image" | "text")[];
|
|
6360
|
+
cost: {
|
|
6361
|
+
input: number;
|
|
6362
|
+
output: number;
|
|
6363
|
+
cacheRead: number;
|
|
6364
|
+
cacheWrite: number;
|
|
6365
|
+
};
|
|
6366
|
+
contextWindow: number;
|
|
6367
|
+
maxTokens: number;
|
|
6368
|
+
};
|
|
6369
|
+
readonly "gpt-5.4-mini": {
|
|
6370
|
+
id: string;
|
|
6371
|
+
name: string;
|
|
6372
|
+
api: "openai-codex-responses";
|
|
6373
|
+
provider: string;
|
|
6374
|
+
baseUrl: string;
|
|
6375
|
+
reasoning: true;
|
|
6376
|
+
input: ("image" | "text")[];
|
|
6377
|
+
cost: {
|
|
6378
|
+
input: number;
|
|
6379
|
+
output: number;
|
|
6380
|
+
cacheRead: number;
|
|
6381
|
+
cacheWrite: number;
|
|
6382
|
+
};
|
|
6383
|
+
contextWindow: number;
|
|
6384
|
+
maxTokens: number;
|
|
6385
|
+
};
|
|
6386
|
+
};
|
|
6387
|
+
readonly opencode: {
|
|
6388
|
+
readonly "big-pickle": {
|
|
6389
|
+
id: string;
|
|
6390
|
+
name: string;
|
|
6391
|
+
api: "anthropic-messages";
|
|
6392
|
+
provider: string;
|
|
6393
|
+
baseUrl: string;
|
|
6394
|
+
reasoning: true;
|
|
6395
|
+
input: "text"[];
|
|
6396
|
+
cost: {
|
|
6397
|
+
input: number;
|
|
6398
|
+
output: number;
|
|
6399
|
+
cacheRead: number;
|
|
6400
|
+
cacheWrite: number;
|
|
6401
|
+
};
|
|
6402
|
+
contextWindow: number;
|
|
6403
|
+
maxTokens: number;
|
|
6404
|
+
};
|
|
6405
|
+
readonly "claude-3-5-haiku": {
|
|
6406
|
+
id: string;
|
|
6407
|
+
name: string;
|
|
6408
|
+
api: "anthropic-messages";
|
|
6409
|
+
provider: string;
|
|
6410
|
+
baseUrl: string;
|
|
6411
|
+
reasoning: false;
|
|
6412
|
+
input: ("image" | "text")[];
|
|
6413
|
+
cost: {
|
|
6414
|
+
input: number;
|
|
6415
|
+
output: number;
|
|
6416
|
+
cacheRead: number;
|
|
6417
|
+
cacheWrite: number;
|
|
6418
|
+
};
|
|
6419
|
+
contextWindow: number;
|
|
6420
|
+
maxTokens: number;
|
|
6421
|
+
};
|
|
6422
|
+
readonly "claude-haiku-4-5": {
|
|
6423
|
+
id: string;
|
|
6424
|
+
name: string;
|
|
6425
|
+
api: "anthropic-messages";
|
|
6426
|
+
provider: string;
|
|
6427
|
+
baseUrl: string;
|
|
6428
|
+
reasoning: true;
|
|
6429
|
+
input: ("image" | "text")[];
|
|
6430
|
+
cost: {
|
|
6431
|
+
input: number;
|
|
6432
|
+
output: number;
|
|
6433
|
+
cacheRead: number;
|
|
6434
|
+
cacheWrite: number;
|
|
6435
|
+
};
|
|
6436
|
+
contextWindow: number;
|
|
6437
|
+
maxTokens: number;
|
|
6438
|
+
};
|
|
6439
|
+
readonly "claude-opus-4-1": {
|
|
6440
|
+
id: string;
|
|
6441
|
+
name: string;
|
|
6442
|
+
api: "anthropic-messages";
|
|
6443
|
+
provider: string;
|
|
6444
|
+
baseUrl: string;
|
|
6445
|
+
reasoning: true;
|
|
6446
|
+
input: ("image" | "text")[];
|
|
6447
|
+
cost: {
|
|
6448
|
+
input: number;
|
|
6449
|
+
output: number;
|
|
6450
|
+
cacheRead: number;
|
|
6451
|
+
cacheWrite: number;
|
|
6452
|
+
};
|
|
6453
|
+
contextWindow: number;
|
|
6454
|
+
maxTokens: number;
|
|
6455
|
+
};
|
|
6456
|
+
readonly "claude-opus-4-5": {
|
|
6457
|
+
id: string;
|
|
6458
|
+
name: string;
|
|
6459
|
+
api: "anthropic-messages";
|
|
6050
6460
|
provider: string;
|
|
6051
6461
|
baseUrl: string;
|
|
6052
6462
|
reasoning: true;
|
|
@@ -6060,12 +6470,10 @@ export declare const MODELS: {
|
|
|
6060
6470
|
contextWindow: number;
|
|
6061
6471
|
maxTokens: number;
|
|
6062
6472
|
};
|
|
6063
|
-
|
|
6064
|
-
readonly "openai-codex": {
|
|
6065
|
-
readonly "gpt-5.1": {
|
|
6473
|
+
readonly "claude-opus-4-6": {
|
|
6066
6474
|
id: string;
|
|
6067
6475
|
name: string;
|
|
6068
|
-
api: "
|
|
6476
|
+
api: "anthropic-messages";
|
|
6069
6477
|
provider: string;
|
|
6070
6478
|
baseUrl: string;
|
|
6071
6479
|
reasoning: true;
|
|
@@ -6079,10 +6487,10 @@ export declare const MODELS: {
|
|
|
6079
6487
|
contextWindow: number;
|
|
6080
6488
|
maxTokens: number;
|
|
6081
6489
|
};
|
|
6082
|
-
readonly "
|
|
6490
|
+
readonly "claude-opus-4-7": {
|
|
6083
6491
|
id: string;
|
|
6084
6492
|
name: string;
|
|
6085
|
-
api: "
|
|
6493
|
+
api: "anthropic-messages";
|
|
6086
6494
|
provider: string;
|
|
6087
6495
|
baseUrl: string;
|
|
6088
6496
|
reasoning: true;
|
|
@@ -6096,10 +6504,10 @@ export declare const MODELS: {
|
|
|
6096
6504
|
contextWindow: number;
|
|
6097
6505
|
maxTokens: number;
|
|
6098
6506
|
};
|
|
6099
|
-
readonly "
|
|
6507
|
+
readonly "claude-sonnet-4": {
|
|
6100
6508
|
id: string;
|
|
6101
6509
|
name: string;
|
|
6102
|
-
api: "
|
|
6510
|
+
api: "anthropic-messages";
|
|
6103
6511
|
provider: string;
|
|
6104
6512
|
baseUrl: string;
|
|
6105
6513
|
reasoning: true;
|
|
@@ -6113,10 +6521,10 @@ export declare const MODELS: {
|
|
|
6113
6521
|
contextWindow: number;
|
|
6114
6522
|
maxTokens: number;
|
|
6115
6523
|
};
|
|
6116
|
-
readonly "
|
|
6524
|
+
readonly "claude-sonnet-4-5": {
|
|
6117
6525
|
id: string;
|
|
6118
6526
|
name: string;
|
|
6119
|
-
api: "
|
|
6527
|
+
api: "anthropic-messages";
|
|
6120
6528
|
provider: string;
|
|
6121
6529
|
baseUrl: string;
|
|
6122
6530
|
reasoning: true;
|
|
@@ -6130,10 +6538,10 @@ export declare const MODELS: {
|
|
|
6130
6538
|
contextWindow: number;
|
|
6131
6539
|
maxTokens: number;
|
|
6132
6540
|
};
|
|
6133
|
-
readonly "
|
|
6541
|
+
readonly "claude-sonnet-4-6": {
|
|
6134
6542
|
id: string;
|
|
6135
6543
|
name: string;
|
|
6136
|
-
api: "
|
|
6544
|
+
api: "anthropic-messages";
|
|
6137
6545
|
provider: string;
|
|
6138
6546
|
baseUrl: string;
|
|
6139
6547
|
reasoning: true;
|
|
@@ -6147,10 +6555,10 @@ export declare const MODELS: {
|
|
|
6147
6555
|
contextWindow: number;
|
|
6148
6556
|
maxTokens: number;
|
|
6149
6557
|
};
|
|
6150
|
-
readonly "
|
|
6558
|
+
readonly "gemini-3-flash": {
|
|
6151
6559
|
id: string;
|
|
6152
6560
|
name: string;
|
|
6153
|
-
api: "
|
|
6561
|
+
api: "google-generative-ai";
|
|
6154
6562
|
provider: string;
|
|
6155
6563
|
baseUrl: string;
|
|
6156
6564
|
reasoning: true;
|
|
@@ -6164,14 +6572,14 @@ export declare const MODELS: {
|
|
|
6164
6572
|
contextWindow: number;
|
|
6165
6573
|
maxTokens: number;
|
|
6166
6574
|
};
|
|
6167
|
-
readonly "
|
|
6575
|
+
readonly "gemini-3.1-pro": {
|
|
6168
6576
|
id: string;
|
|
6169
6577
|
name: string;
|
|
6170
|
-
api: "
|
|
6578
|
+
api: "google-generative-ai";
|
|
6171
6579
|
provider: string;
|
|
6172
6580
|
baseUrl: string;
|
|
6173
6581
|
reasoning: true;
|
|
6174
|
-
input: "text"[];
|
|
6582
|
+
input: ("image" | "text")[];
|
|
6175
6583
|
cost: {
|
|
6176
6584
|
input: number;
|
|
6177
6585
|
output: number;
|
|
@@ -6181,14 +6589,14 @@ export declare const MODELS: {
|
|
|
6181
6589
|
contextWindow: number;
|
|
6182
6590
|
maxTokens: number;
|
|
6183
6591
|
};
|
|
6184
|
-
readonly "
|
|
6592
|
+
readonly "glm-5": {
|
|
6185
6593
|
id: string;
|
|
6186
6594
|
name: string;
|
|
6187
|
-
api: "openai-
|
|
6595
|
+
api: "openai-completions";
|
|
6188
6596
|
provider: string;
|
|
6189
6597
|
baseUrl: string;
|
|
6190
6598
|
reasoning: true;
|
|
6191
|
-
input:
|
|
6599
|
+
input: "text"[];
|
|
6192
6600
|
cost: {
|
|
6193
6601
|
input: number;
|
|
6194
6602
|
output: number;
|
|
@@ -6198,14 +6606,14 @@ export declare const MODELS: {
|
|
|
6198
6606
|
contextWindow: number;
|
|
6199
6607
|
maxTokens: number;
|
|
6200
6608
|
};
|
|
6201
|
-
readonly "
|
|
6609
|
+
readonly "glm-5.1": {
|
|
6202
6610
|
id: string;
|
|
6203
6611
|
name: string;
|
|
6204
|
-
api: "openai-
|
|
6612
|
+
api: "openai-completions";
|
|
6205
6613
|
provider: string;
|
|
6206
6614
|
baseUrl: string;
|
|
6207
6615
|
reasoning: true;
|
|
6208
|
-
input:
|
|
6616
|
+
input: "text"[];
|
|
6209
6617
|
cost: {
|
|
6210
6618
|
input: number;
|
|
6211
6619
|
output: number;
|
|
@@ -6215,16 +6623,14 @@ export declare const MODELS: {
|
|
|
6215
6623
|
contextWindow: number;
|
|
6216
6624
|
maxTokens: number;
|
|
6217
6625
|
};
|
|
6218
|
-
|
|
6219
|
-
readonly opencode: {
|
|
6220
|
-
readonly "big-pickle": {
|
|
6626
|
+
readonly "gpt-5": {
|
|
6221
6627
|
id: string;
|
|
6222
6628
|
name: string;
|
|
6223
|
-
api: "
|
|
6629
|
+
api: "openai-responses";
|
|
6224
6630
|
provider: string;
|
|
6225
6631
|
baseUrl: string;
|
|
6226
6632
|
reasoning: true;
|
|
6227
|
-
input: "text"[];
|
|
6633
|
+
input: ("image" | "text")[];
|
|
6228
6634
|
cost: {
|
|
6229
6635
|
input: number;
|
|
6230
6636
|
output: number;
|
|
@@ -6234,13 +6640,13 @@ export declare const MODELS: {
|
|
|
6234
6640
|
contextWindow: number;
|
|
6235
6641
|
maxTokens: number;
|
|
6236
6642
|
};
|
|
6237
|
-
readonly "
|
|
6643
|
+
readonly "gpt-5-codex": {
|
|
6238
6644
|
id: string;
|
|
6239
6645
|
name: string;
|
|
6240
|
-
api: "
|
|
6646
|
+
api: "openai-responses";
|
|
6241
6647
|
provider: string;
|
|
6242
6648
|
baseUrl: string;
|
|
6243
|
-
reasoning:
|
|
6649
|
+
reasoning: true;
|
|
6244
6650
|
input: ("image" | "text")[];
|
|
6245
6651
|
cost: {
|
|
6246
6652
|
input: number;
|
|
@@ -6251,10 +6657,10 @@ export declare const MODELS: {
|
|
|
6251
6657
|
contextWindow: number;
|
|
6252
6658
|
maxTokens: number;
|
|
6253
6659
|
};
|
|
6254
|
-
readonly "
|
|
6660
|
+
readonly "gpt-5-nano": {
|
|
6255
6661
|
id: string;
|
|
6256
6662
|
name: string;
|
|
6257
|
-
api: "
|
|
6663
|
+
api: "openai-responses";
|
|
6258
6664
|
provider: string;
|
|
6259
6665
|
baseUrl: string;
|
|
6260
6666
|
reasoning: true;
|
|
@@ -6268,10 +6674,10 @@ export declare const MODELS: {
|
|
|
6268
6674
|
contextWindow: number;
|
|
6269
6675
|
maxTokens: number;
|
|
6270
6676
|
};
|
|
6271
|
-
readonly "
|
|
6677
|
+
readonly "gpt-5.1": {
|
|
6272
6678
|
id: string;
|
|
6273
6679
|
name: string;
|
|
6274
|
-
api: "
|
|
6680
|
+
api: "openai-responses";
|
|
6275
6681
|
provider: string;
|
|
6276
6682
|
baseUrl: string;
|
|
6277
6683
|
reasoning: true;
|
|
@@ -6285,10 +6691,10 @@ export declare const MODELS: {
|
|
|
6285
6691
|
contextWindow: number;
|
|
6286
6692
|
maxTokens: number;
|
|
6287
6693
|
};
|
|
6288
|
-
readonly "
|
|
6694
|
+
readonly "gpt-5.1-codex": {
|
|
6289
6695
|
id: string;
|
|
6290
6696
|
name: string;
|
|
6291
|
-
api: "
|
|
6697
|
+
api: "openai-responses";
|
|
6292
6698
|
provider: string;
|
|
6293
6699
|
baseUrl: string;
|
|
6294
6700
|
reasoning: true;
|
|
@@ -6302,10 +6708,10 @@ export declare const MODELS: {
|
|
|
6302
6708
|
contextWindow: number;
|
|
6303
6709
|
maxTokens: number;
|
|
6304
6710
|
};
|
|
6305
|
-
readonly "
|
|
6711
|
+
readonly "gpt-5.1-codex-max": {
|
|
6306
6712
|
id: string;
|
|
6307
6713
|
name: string;
|
|
6308
|
-
api: "
|
|
6714
|
+
api: "openai-responses";
|
|
6309
6715
|
provider: string;
|
|
6310
6716
|
baseUrl: string;
|
|
6311
6717
|
reasoning: true;
|
|
@@ -6319,10 +6725,10 @@ export declare const MODELS: {
|
|
|
6319
6725
|
contextWindow: number;
|
|
6320
6726
|
maxTokens: number;
|
|
6321
6727
|
};
|
|
6322
|
-
readonly "
|
|
6728
|
+
readonly "gpt-5.1-codex-mini": {
|
|
6323
6729
|
id: string;
|
|
6324
6730
|
name: string;
|
|
6325
|
-
api: "
|
|
6731
|
+
api: "openai-responses";
|
|
6326
6732
|
provider: string;
|
|
6327
6733
|
baseUrl: string;
|
|
6328
6734
|
reasoning: true;
|
|
@@ -6336,10 +6742,10 @@ export declare const MODELS: {
|
|
|
6336
6742
|
contextWindow: number;
|
|
6337
6743
|
maxTokens: number;
|
|
6338
6744
|
};
|
|
6339
|
-
readonly "
|
|
6745
|
+
readonly "gpt-5.2": {
|
|
6340
6746
|
id: string;
|
|
6341
6747
|
name: string;
|
|
6342
|
-
api: "
|
|
6748
|
+
api: "openai-responses";
|
|
6343
6749
|
provider: string;
|
|
6344
6750
|
baseUrl: string;
|
|
6345
6751
|
reasoning: true;
|
|
@@ -6353,10 +6759,10 @@ export declare const MODELS: {
|
|
|
6353
6759
|
contextWindow: number;
|
|
6354
6760
|
maxTokens: number;
|
|
6355
6761
|
};
|
|
6356
|
-
readonly "
|
|
6762
|
+
readonly "gpt-5.2-codex": {
|
|
6357
6763
|
id: string;
|
|
6358
6764
|
name: string;
|
|
6359
|
-
api: "
|
|
6765
|
+
api: "openai-responses";
|
|
6360
6766
|
provider: string;
|
|
6361
6767
|
baseUrl: string;
|
|
6362
6768
|
reasoning: true;
|
|
@@ -6370,10 +6776,10 @@ export declare const MODELS: {
|
|
|
6370
6776
|
contextWindow: number;
|
|
6371
6777
|
maxTokens: number;
|
|
6372
6778
|
};
|
|
6373
|
-
readonly "
|
|
6779
|
+
readonly "gpt-5.3-codex": {
|
|
6374
6780
|
id: string;
|
|
6375
6781
|
name: string;
|
|
6376
|
-
api: "
|
|
6782
|
+
api: "openai-responses";
|
|
6377
6783
|
provider: string;
|
|
6378
6784
|
baseUrl: string;
|
|
6379
6785
|
reasoning: true;
|
|
@@ -6387,10 +6793,10 @@ export declare const MODELS: {
|
|
|
6387
6793
|
contextWindow: number;
|
|
6388
6794
|
maxTokens: number;
|
|
6389
6795
|
};
|
|
6390
|
-
readonly "
|
|
6796
|
+
readonly "gpt-5.4": {
|
|
6391
6797
|
id: string;
|
|
6392
6798
|
name: string;
|
|
6393
|
-
api: "
|
|
6799
|
+
api: "openai-responses";
|
|
6394
6800
|
provider: string;
|
|
6395
6801
|
baseUrl: string;
|
|
6396
6802
|
reasoning: true;
|
|
@@ -6404,14 +6810,14 @@ export declare const MODELS: {
|
|
|
6404
6810
|
contextWindow: number;
|
|
6405
6811
|
maxTokens: number;
|
|
6406
6812
|
};
|
|
6407
|
-
readonly "
|
|
6813
|
+
readonly "gpt-5.4-mini": {
|
|
6408
6814
|
id: string;
|
|
6409
6815
|
name: string;
|
|
6410
|
-
api: "openai-
|
|
6816
|
+
api: "openai-responses";
|
|
6411
6817
|
provider: string;
|
|
6412
6818
|
baseUrl: string;
|
|
6413
6819
|
reasoning: true;
|
|
6414
|
-
input: "text"[];
|
|
6820
|
+
input: ("image" | "text")[];
|
|
6415
6821
|
cost: {
|
|
6416
6822
|
input: number;
|
|
6417
6823
|
output: number;
|
|
@@ -6421,7 +6827,7 @@ export declare const MODELS: {
|
|
|
6421
6827
|
contextWindow: number;
|
|
6422
6828
|
maxTokens: number;
|
|
6423
6829
|
};
|
|
6424
|
-
readonly "gpt-5": {
|
|
6830
|
+
readonly "gpt-5.4-nano": {
|
|
6425
6831
|
id: string;
|
|
6426
6832
|
name: string;
|
|
6427
6833
|
api: "openai-responses";
|
|
@@ -6438,7 +6844,7 @@ export declare const MODELS: {
|
|
|
6438
6844
|
contextWindow: number;
|
|
6439
6845
|
maxTokens: number;
|
|
6440
6846
|
};
|
|
6441
|
-
readonly "gpt-5-
|
|
6847
|
+
readonly "gpt-5.4-pro": {
|
|
6442
6848
|
id: string;
|
|
6443
6849
|
name: string;
|
|
6444
6850
|
api: "openai-responses";
|
|
@@ -6455,14 +6861,14 @@ export declare const MODELS: {
|
|
|
6455
6861
|
contextWindow: number;
|
|
6456
6862
|
maxTokens: number;
|
|
6457
6863
|
};
|
|
6458
|
-
readonly "
|
|
6864
|
+
readonly "hy3-preview-free": {
|
|
6459
6865
|
id: string;
|
|
6460
6866
|
name: string;
|
|
6461
|
-
api: "openai-
|
|
6867
|
+
api: "openai-completions";
|
|
6462
6868
|
provider: string;
|
|
6463
6869
|
baseUrl: string;
|
|
6464
6870
|
reasoning: true;
|
|
6465
|
-
input:
|
|
6871
|
+
input: "text"[];
|
|
6466
6872
|
cost: {
|
|
6467
6873
|
input: number;
|
|
6468
6874
|
output: number;
|
|
@@ -6472,10 +6878,10 @@ export declare const MODELS: {
|
|
|
6472
6878
|
contextWindow: number;
|
|
6473
6879
|
maxTokens: number;
|
|
6474
6880
|
};
|
|
6475
|
-
readonly "
|
|
6881
|
+
readonly "kimi-k2.5": {
|
|
6476
6882
|
id: string;
|
|
6477
6883
|
name: string;
|
|
6478
|
-
api: "openai-
|
|
6884
|
+
api: "openai-completions";
|
|
6479
6885
|
provider: string;
|
|
6480
6886
|
baseUrl: string;
|
|
6481
6887
|
reasoning: true;
|
|
@@ -6489,10 +6895,10 @@ export declare const MODELS: {
|
|
|
6489
6895
|
contextWindow: number;
|
|
6490
6896
|
maxTokens: number;
|
|
6491
6897
|
};
|
|
6492
|
-
readonly "
|
|
6898
|
+
readonly "kimi-k2.6": {
|
|
6493
6899
|
id: string;
|
|
6494
6900
|
name: string;
|
|
6495
|
-
api: "openai-
|
|
6901
|
+
api: "openai-completions";
|
|
6496
6902
|
provider: string;
|
|
6497
6903
|
baseUrl: string;
|
|
6498
6904
|
reasoning: true;
|
|
@@ -6506,14 +6912,14 @@ export declare const MODELS: {
|
|
|
6506
6912
|
contextWindow: number;
|
|
6507
6913
|
maxTokens: number;
|
|
6508
6914
|
};
|
|
6509
|
-
readonly "
|
|
6915
|
+
readonly "ling-2.6-flash-free": {
|
|
6510
6916
|
id: string;
|
|
6511
6917
|
name: string;
|
|
6512
|
-
api: "openai-
|
|
6918
|
+
api: "openai-completions";
|
|
6513
6919
|
provider: string;
|
|
6514
6920
|
baseUrl: string;
|
|
6515
|
-
reasoning:
|
|
6516
|
-
input:
|
|
6921
|
+
reasoning: false;
|
|
6922
|
+
input: "text"[];
|
|
6517
6923
|
cost: {
|
|
6518
6924
|
input: number;
|
|
6519
6925
|
output: number;
|
|
@@ -6523,14 +6929,14 @@ export declare const MODELS: {
|
|
|
6523
6929
|
contextWindow: number;
|
|
6524
6930
|
maxTokens: number;
|
|
6525
6931
|
};
|
|
6526
|
-
readonly "
|
|
6932
|
+
readonly "minimax-m2.5": {
|
|
6527
6933
|
id: string;
|
|
6528
6934
|
name: string;
|
|
6529
|
-
api: "openai-
|
|
6935
|
+
api: "openai-completions";
|
|
6530
6936
|
provider: string;
|
|
6531
6937
|
baseUrl: string;
|
|
6532
6938
|
reasoning: true;
|
|
6533
|
-
input:
|
|
6939
|
+
input: "text"[];
|
|
6534
6940
|
cost: {
|
|
6535
6941
|
input: number;
|
|
6536
6942
|
output: number;
|
|
@@ -6540,14 +6946,14 @@ export declare const MODELS: {
|
|
|
6540
6946
|
contextWindow: number;
|
|
6541
6947
|
maxTokens: number;
|
|
6542
6948
|
};
|
|
6543
|
-
readonly "
|
|
6949
|
+
readonly "minimax-m2.5-free": {
|
|
6544
6950
|
id: string;
|
|
6545
6951
|
name: string;
|
|
6546
|
-
api: "
|
|
6952
|
+
api: "anthropic-messages";
|
|
6547
6953
|
provider: string;
|
|
6548
6954
|
baseUrl: string;
|
|
6549
6955
|
reasoning: true;
|
|
6550
|
-
input:
|
|
6956
|
+
input: "text"[];
|
|
6551
6957
|
cost: {
|
|
6552
6958
|
input: number;
|
|
6553
6959
|
output: number;
|
|
@@ -6557,14 +6963,14 @@ export declare const MODELS: {
|
|
|
6557
6963
|
contextWindow: number;
|
|
6558
6964
|
maxTokens: number;
|
|
6559
6965
|
};
|
|
6560
|
-
readonly "
|
|
6966
|
+
readonly "minimax-m2.7": {
|
|
6561
6967
|
id: string;
|
|
6562
6968
|
name: string;
|
|
6563
|
-
api: "openai-
|
|
6969
|
+
api: "openai-completions";
|
|
6564
6970
|
provider: string;
|
|
6565
6971
|
baseUrl: string;
|
|
6566
6972
|
reasoning: true;
|
|
6567
|
-
input:
|
|
6973
|
+
input: "text"[];
|
|
6568
6974
|
cost: {
|
|
6569
6975
|
input: number;
|
|
6570
6976
|
output: number;
|
|
@@ -6574,14 +6980,14 @@ export declare const MODELS: {
|
|
|
6574
6980
|
contextWindow: number;
|
|
6575
6981
|
maxTokens: number;
|
|
6576
6982
|
};
|
|
6577
|
-
readonly "
|
|
6983
|
+
readonly "nemotron-3-super-free": {
|
|
6578
6984
|
id: string;
|
|
6579
6985
|
name: string;
|
|
6580
|
-
api: "openai-
|
|
6986
|
+
api: "openai-completions";
|
|
6581
6987
|
provider: string;
|
|
6582
6988
|
baseUrl: string;
|
|
6583
6989
|
reasoning: true;
|
|
6584
|
-
input:
|
|
6990
|
+
input: "text"[];
|
|
6585
6991
|
cost: {
|
|
6586
6992
|
input: number;
|
|
6587
6993
|
output: number;
|
|
@@ -6591,10 +6997,10 @@ export declare const MODELS: {
|
|
|
6591
6997
|
contextWindow: number;
|
|
6592
6998
|
maxTokens: number;
|
|
6593
6999
|
};
|
|
6594
|
-
readonly "
|
|
7000
|
+
readonly "qwen3.5-plus": {
|
|
6595
7001
|
id: string;
|
|
6596
7002
|
name: string;
|
|
6597
|
-
api: "
|
|
7003
|
+
api: "anthropic-messages";
|
|
6598
7004
|
provider: string;
|
|
6599
7005
|
baseUrl: string;
|
|
6600
7006
|
reasoning: true;
|
|
@@ -6608,10 +7014,10 @@ export declare const MODELS: {
|
|
|
6608
7014
|
contextWindow: number;
|
|
6609
7015
|
maxTokens: number;
|
|
6610
7016
|
};
|
|
6611
|
-
readonly "
|
|
7017
|
+
readonly "qwen3.6-plus": {
|
|
6612
7018
|
id: string;
|
|
6613
7019
|
name: string;
|
|
6614
|
-
api: "
|
|
7020
|
+
api: "anthropic-messages";
|
|
6615
7021
|
provider: string;
|
|
6616
7022
|
baseUrl: string;
|
|
6617
7023
|
reasoning: true;
|
|
@@ -6625,14 +7031,16 @@ export declare const MODELS: {
|
|
|
6625
7031
|
contextWindow: number;
|
|
6626
7032
|
maxTokens: number;
|
|
6627
7033
|
};
|
|
6628
|
-
|
|
7034
|
+
};
|
|
7035
|
+
readonly "opencode-go": {
|
|
7036
|
+
readonly "glm-5": {
|
|
6629
7037
|
id: string;
|
|
6630
7038
|
name: string;
|
|
6631
|
-
api: "openai-
|
|
7039
|
+
api: "openai-completions";
|
|
6632
7040
|
provider: string;
|
|
6633
7041
|
baseUrl: string;
|
|
6634
7042
|
reasoning: true;
|
|
6635
|
-
input:
|
|
7043
|
+
input: "text"[];
|
|
6636
7044
|
cost: {
|
|
6637
7045
|
input: number;
|
|
6638
7046
|
output: number;
|
|
@@ -6642,14 +7050,14 @@ export declare const MODELS: {
|
|
|
6642
7050
|
contextWindow: number;
|
|
6643
7051
|
maxTokens: number;
|
|
6644
7052
|
};
|
|
6645
|
-
readonly "
|
|
7053
|
+
readonly "glm-5.1": {
|
|
6646
7054
|
id: string;
|
|
6647
7055
|
name: string;
|
|
6648
|
-
api: "openai-
|
|
7056
|
+
api: "openai-completions";
|
|
6649
7057
|
provider: string;
|
|
6650
7058
|
baseUrl: string;
|
|
6651
7059
|
reasoning: true;
|
|
6652
|
-
input:
|
|
7060
|
+
input: "text"[];
|
|
6653
7061
|
cost: {
|
|
6654
7062
|
input: number;
|
|
6655
7063
|
output: number;
|
|
@@ -6676,7 +7084,7 @@ export declare const MODELS: {
|
|
|
6676
7084
|
contextWindow: number;
|
|
6677
7085
|
maxTokens: number;
|
|
6678
7086
|
};
|
|
6679
|
-
readonly "
|
|
7087
|
+
readonly "kimi-k2.6": {
|
|
6680
7088
|
id: string;
|
|
6681
7089
|
name: string;
|
|
6682
7090
|
api: "openai-completions";
|
|
@@ -6693,14 +7101,14 @@ export declare const MODELS: {
|
|
|
6693
7101
|
contextWindow: number;
|
|
6694
7102
|
maxTokens: number;
|
|
6695
7103
|
};
|
|
6696
|
-
readonly "mimo-v2-
|
|
7104
|
+
readonly "mimo-v2-omni": {
|
|
6697
7105
|
id: string;
|
|
6698
7106
|
name: string;
|
|
6699
7107
|
api: "openai-completions";
|
|
6700
7108
|
provider: string;
|
|
6701
7109
|
baseUrl: string;
|
|
6702
7110
|
reasoning: true;
|
|
6703
|
-
input: "text"[];
|
|
7111
|
+
input: ("image" | "text")[];
|
|
6704
7112
|
cost: {
|
|
6705
7113
|
input: number;
|
|
6706
7114
|
output: number;
|
|
@@ -6710,7 +7118,7 @@ export declare const MODELS: {
|
|
|
6710
7118
|
contextWindow: number;
|
|
6711
7119
|
maxTokens: number;
|
|
6712
7120
|
};
|
|
6713
|
-
readonly "
|
|
7121
|
+
readonly "mimo-v2-pro": {
|
|
6714
7122
|
id: string;
|
|
6715
7123
|
name: string;
|
|
6716
7124
|
api: "openai-completions";
|
|
@@ -6727,14 +7135,14 @@ export declare const MODELS: {
|
|
|
6727
7135
|
contextWindow: number;
|
|
6728
7136
|
maxTokens: number;
|
|
6729
7137
|
};
|
|
6730
|
-
readonly "
|
|
7138
|
+
readonly "mimo-v2.5": {
|
|
6731
7139
|
id: string;
|
|
6732
7140
|
name: string;
|
|
6733
|
-
api: "
|
|
7141
|
+
api: "openai-completions";
|
|
6734
7142
|
provider: string;
|
|
6735
7143
|
baseUrl: string;
|
|
6736
7144
|
reasoning: true;
|
|
6737
|
-
input: "text"[];
|
|
7145
|
+
input: ("image" | "text")[];
|
|
6738
7146
|
cost: {
|
|
6739
7147
|
input: number;
|
|
6740
7148
|
output: number;
|
|
@@ -6744,7 +7152,7 @@ export declare const MODELS: {
|
|
|
6744
7152
|
contextWindow: number;
|
|
6745
7153
|
maxTokens: number;
|
|
6746
7154
|
};
|
|
6747
|
-
readonly "
|
|
7155
|
+
readonly "mimo-v2.5-pro": {
|
|
6748
7156
|
id: string;
|
|
6749
7157
|
name: string;
|
|
6750
7158
|
api: "openai-completions";
|
|
@@ -6761,9 +7169,7 @@ export declare const MODELS: {
|
|
|
6761
7169
|
contextWindow: number;
|
|
6762
7170
|
maxTokens: number;
|
|
6763
7171
|
};
|
|
6764
|
-
|
|
6765
|
-
readonly "opencode-go": {
|
|
6766
|
-
readonly "glm-5": {
|
|
7172
|
+
readonly "minimax-m2.5": {
|
|
6767
7173
|
id: string;
|
|
6768
7174
|
name: string;
|
|
6769
7175
|
api: "openai-completions";
|
|
@@ -6780,14 +7186,14 @@ export declare const MODELS: {
|
|
|
6780
7186
|
contextWindow: number;
|
|
6781
7187
|
maxTokens: number;
|
|
6782
7188
|
};
|
|
6783
|
-
readonly "
|
|
7189
|
+
readonly "minimax-m2.7": {
|
|
6784
7190
|
id: string;
|
|
6785
7191
|
name: string;
|
|
6786
|
-
api: "
|
|
7192
|
+
api: "anthropic-messages";
|
|
6787
7193
|
provider: string;
|
|
6788
7194
|
baseUrl: string;
|
|
6789
7195
|
reasoning: true;
|
|
6790
|
-
input:
|
|
7196
|
+
input: "text"[];
|
|
6791
7197
|
cost: {
|
|
6792
7198
|
input: number;
|
|
6793
7199
|
output: number;
|
|
@@ -6797,14 +7203,14 @@ export declare const MODELS: {
|
|
|
6797
7203
|
contextWindow: number;
|
|
6798
7204
|
maxTokens: number;
|
|
6799
7205
|
};
|
|
6800
|
-
readonly "
|
|
7206
|
+
readonly "qwen3.5-plus": {
|
|
6801
7207
|
id: string;
|
|
6802
7208
|
name: string;
|
|
6803
7209
|
api: "anthropic-messages";
|
|
6804
7210
|
provider: string;
|
|
6805
7211
|
baseUrl: string;
|
|
6806
7212
|
reasoning: true;
|
|
6807
|
-
input: "text"[];
|
|
7213
|
+
input: ("image" | "text")[];
|
|
6808
7214
|
cost: {
|
|
6809
7215
|
input: number;
|
|
6810
7216
|
output: number;
|
|
@@ -6814,14 +7220,14 @@ export declare const MODELS: {
|
|
|
6814
7220
|
contextWindow: number;
|
|
6815
7221
|
maxTokens: number;
|
|
6816
7222
|
};
|
|
6817
|
-
readonly "
|
|
7223
|
+
readonly "qwen3.6-plus": {
|
|
6818
7224
|
id: string;
|
|
6819
7225
|
name: string;
|
|
6820
7226
|
api: "anthropic-messages";
|
|
6821
7227
|
provider: string;
|
|
6822
7228
|
baseUrl: string;
|
|
6823
7229
|
reasoning: true;
|
|
6824
|
-
input: "text"[];
|
|
7230
|
+
input: ("image" | "text")[];
|
|
6825
7231
|
cost: {
|
|
6826
7232
|
input: number;
|
|
6827
7233
|
output: number;
|
|
@@ -7003,13 +7409,30 @@ export declare const MODELS: {
|
|
|
7003
7409
|
contextWindow: number;
|
|
7004
7410
|
maxTokens: number;
|
|
7005
7411
|
};
|
|
7006
|
-
readonly "anthropic/claude-3.
|
|
7412
|
+
readonly "anthropic/claude-3.7-sonnet": {
|
|
7413
|
+
id: string;
|
|
7414
|
+
name: string;
|
|
7415
|
+
api: "openai-completions";
|
|
7416
|
+
provider: string;
|
|
7417
|
+
baseUrl: string;
|
|
7418
|
+
reasoning: true;
|
|
7419
|
+
input: ("image" | "text")[];
|
|
7420
|
+
cost: {
|
|
7421
|
+
input: number;
|
|
7422
|
+
output: number;
|
|
7423
|
+
cacheRead: number;
|
|
7424
|
+
cacheWrite: number;
|
|
7425
|
+
};
|
|
7426
|
+
contextWindow: number;
|
|
7427
|
+
maxTokens: number;
|
|
7428
|
+
};
|
|
7429
|
+
readonly "anthropic/claude-3.7-sonnet:thinking": {
|
|
7007
7430
|
id: string;
|
|
7008
7431
|
name: string;
|
|
7009
7432
|
api: "openai-completions";
|
|
7010
7433
|
provider: string;
|
|
7011
7434
|
baseUrl: string;
|
|
7012
|
-
reasoning:
|
|
7435
|
+
reasoning: true;
|
|
7013
7436
|
input: ("image" | "text")[];
|
|
7014
7437
|
cost: {
|
|
7015
7438
|
input: number;
|
|
@@ -7020,7 +7443,7 @@ export declare const MODELS: {
|
|
|
7020
7443
|
contextWindow: number;
|
|
7021
7444
|
maxTokens: number;
|
|
7022
7445
|
};
|
|
7023
|
-
readonly "anthropic/claude-
|
|
7446
|
+
readonly "anthropic/claude-haiku-4.5": {
|
|
7024
7447
|
id: string;
|
|
7025
7448
|
name: string;
|
|
7026
7449
|
api: "openai-completions";
|
|
@@ -7037,7 +7460,7 @@ export declare const MODELS: {
|
|
|
7037
7460
|
contextWindow: number;
|
|
7038
7461
|
maxTokens: number;
|
|
7039
7462
|
};
|
|
7040
|
-
readonly "anthropic/claude-
|
|
7463
|
+
readonly "anthropic/claude-opus-4": {
|
|
7041
7464
|
id: string;
|
|
7042
7465
|
name: string;
|
|
7043
7466
|
api: "openai-completions";
|
|
@@ -7054,7 +7477,7 @@ export declare const MODELS: {
|
|
|
7054
7477
|
contextWindow: number;
|
|
7055
7478
|
maxTokens: number;
|
|
7056
7479
|
};
|
|
7057
|
-
readonly "anthropic/claude-
|
|
7480
|
+
readonly "anthropic/claude-opus-4.1": {
|
|
7058
7481
|
id: string;
|
|
7059
7482
|
name: string;
|
|
7060
7483
|
api: "openai-completions";
|
|
@@ -7071,7 +7494,7 @@ export declare const MODELS: {
|
|
|
7071
7494
|
contextWindow: number;
|
|
7072
7495
|
maxTokens: number;
|
|
7073
7496
|
};
|
|
7074
|
-
readonly "anthropic/claude-opus-4": {
|
|
7497
|
+
readonly "anthropic/claude-opus-4.5": {
|
|
7075
7498
|
id: string;
|
|
7076
7499
|
name: string;
|
|
7077
7500
|
api: "openai-completions";
|
|
@@ -7088,7 +7511,7 @@ export declare const MODELS: {
|
|
|
7088
7511
|
contextWindow: number;
|
|
7089
7512
|
maxTokens: number;
|
|
7090
7513
|
};
|
|
7091
|
-
readonly "anthropic/claude-opus-4.
|
|
7514
|
+
readonly "anthropic/claude-opus-4.6": {
|
|
7092
7515
|
id: string;
|
|
7093
7516
|
name: string;
|
|
7094
7517
|
api: "openai-completions";
|
|
@@ -7105,7 +7528,7 @@ export declare const MODELS: {
|
|
|
7105
7528
|
contextWindow: number;
|
|
7106
7529
|
maxTokens: number;
|
|
7107
7530
|
};
|
|
7108
|
-
readonly "anthropic/claude-opus-4.
|
|
7531
|
+
readonly "anthropic/claude-opus-4.6-fast": {
|
|
7109
7532
|
id: string;
|
|
7110
7533
|
name: string;
|
|
7111
7534
|
api: "openai-completions";
|
|
@@ -7122,7 +7545,7 @@ export declare const MODELS: {
|
|
|
7122
7545
|
contextWindow: number;
|
|
7123
7546
|
maxTokens: number;
|
|
7124
7547
|
};
|
|
7125
|
-
readonly "anthropic/claude-opus-4.
|
|
7548
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
7126
7549
|
id: string;
|
|
7127
7550
|
name: string;
|
|
7128
7551
|
api: "openai-completions";
|
|
@@ -7190,7 +7613,7 @@ export declare const MODELS: {
|
|
|
7190
7613
|
contextWindow: number;
|
|
7191
7614
|
maxTokens: number;
|
|
7192
7615
|
};
|
|
7193
|
-
readonly "arcee-ai/trinity-large-preview
|
|
7616
|
+
readonly "arcee-ai/trinity-large-preview": {
|
|
7194
7617
|
id: string;
|
|
7195
7618
|
name: string;
|
|
7196
7619
|
api: "openai-completions";
|
|
@@ -7207,7 +7630,7 @@ export declare const MODELS: {
|
|
|
7207
7630
|
contextWindow: number;
|
|
7208
7631
|
maxTokens: number;
|
|
7209
7632
|
};
|
|
7210
|
-
readonly "arcee-ai/trinity-
|
|
7633
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
7211
7634
|
id: string;
|
|
7212
7635
|
name: string;
|
|
7213
7636
|
api: "openai-completions";
|
|
@@ -7224,7 +7647,7 @@ export declare const MODELS: {
|
|
|
7224
7647
|
contextWindow: number;
|
|
7225
7648
|
maxTokens: number;
|
|
7226
7649
|
};
|
|
7227
|
-
readonly "arcee-ai/trinity-mini
|
|
7650
|
+
readonly "arcee-ai/trinity-mini": {
|
|
7228
7651
|
id: string;
|
|
7229
7652
|
name: string;
|
|
7230
7653
|
api: "openai-completions";
|
|
@@ -7434,7 +7857,7 @@ export declare const MODELS: {
|
|
|
7434
7857
|
api: "openai-completions";
|
|
7435
7858
|
provider: string;
|
|
7436
7859
|
baseUrl: string;
|
|
7437
|
-
reasoning:
|
|
7860
|
+
reasoning: false;
|
|
7438
7861
|
input: "text"[];
|
|
7439
7862
|
cost: {
|
|
7440
7863
|
input: number;
|
|
@@ -7717,7 +8140,7 @@ export declare const MODELS: {
|
|
|
7717
8140
|
contextWindow: number;
|
|
7718
8141
|
maxTokens: number;
|
|
7719
8142
|
};
|
|
7720
|
-
readonly "google/gemini-3-
|
|
8143
|
+
readonly "google/gemini-3.1-flash-lite-preview": {
|
|
7721
8144
|
id: string;
|
|
7722
8145
|
name: string;
|
|
7723
8146
|
api: "openai-completions";
|
|
@@ -7734,7 +8157,7 @@ export declare const MODELS: {
|
|
|
7734
8157
|
contextWindow: number;
|
|
7735
8158
|
maxTokens: number;
|
|
7736
8159
|
};
|
|
7737
|
-
readonly "google/gemini-3.1-
|
|
8160
|
+
readonly "google/gemini-3.1-pro-preview": {
|
|
7738
8161
|
id: string;
|
|
7739
8162
|
name: string;
|
|
7740
8163
|
api: "openai-completions";
|
|
@@ -7751,7 +8174,7 @@ export declare const MODELS: {
|
|
|
7751
8174
|
contextWindow: number;
|
|
7752
8175
|
maxTokens: number;
|
|
7753
8176
|
};
|
|
7754
|
-
readonly "google/gemini-3.1-pro-preview": {
|
|
8177
|
+
readonly "google/gemini-3.1-pro-preview-customtools": {
|
|
7755
8178
|
id: string;
|
|
7756
8179
|
name: string;
|
|
7757
8180
|
api: "openai-completions";
|
|
@@ -7768,7 +8191,7 @@ export declare const MODELS: {
|
|
|
7768
8191
|
contextWindow: number;
|
|
7769
8192
|
maxTokens: number;
|
|
7770
8193
|
};
|
|
7771
|
-
readonly "google/
|
|
8194
|
+
readonly "google/gemma-4-26b-a4b-it": {
|
|
7772
8195
|
id: string;
|
|
7773
8196
|
name: string;
|
|
7774
8197
|
api: "openai-completions";
|
|
@@ -7785,14 +8208,14 @@ export declare const MODELS: {
|
|
|
7785
8208
|
contextWindow: number;
|
|
7786
8209
|
maxTokens: number;
|
|
7787
8210
|
};
|
|
7788
|
-
readonly "
|
|
8211
|
+
readonly "google/gemma-4-26b-a4b-it:free": {
|
|
7789
8212
|
id: string;
|
|
7790
8213
|
name: string;
|
|
7791
8214
|
api: "openai-completions";
|
|
7792
8215
|
provider: string;
|
|
7793
8216
|
baseUrl: string;
|
|
7794
|
-
reasoning:
|
|
7795
|
-
input: "text"[];
|
|
8217
|
+
reasoning: true;
|
|
8218
|
+
input: ("image" | "text")[];
|
|
7796
8219
|
cost: {
|
|
7797
8220
|
input: number;
|
|
7798
8221
|
output: number;
|
|
@@ -7802,14 +8225,14 @@ export declare const MODELS: {
|
|
|
7802
8225
|
contextWindow: number;
|
|
7803
8226
|
maxTokens: number;
|
|
7804
8227
|
};
|
|
7805
|
-
readonly "
|
|
8228
|
+
readonly "google/gemma-4-31b-it": {
|
|
7806
8229
|
id: string;
|
|
7807
8230
|
name: string;
|
|
7808
8231
|
api: "openai-completions";
|
|
7809
8232
|
provider: string;
|
|
7810
8233
|
baseUrl: string;
|
|
7811
8234
|
reasoning: true;
|
|
7812
|
-
input: "text"[];
|
|
8235
|
+
input: ("image" | "text")[];
|
|
7813
8236
|
cost: {
|
|
7814
8237
|
input: number;
|
|
7815
8238
|
output: number;
|
|
@@ -7819,14 +8242,14 @@ export declare const MODELS: {
|
|
|
7819
8242
|
contextWindow: number;
|
|
7820
8243
|
maxTokens: number;
|
|
7821
8244
|
};
|
|
7822
|
-
readonly "
|
|
8245
|
+
readonly "google/gemma-4-31b-it:free": {
|
|
7823
8246
|
id: string;
|
|
7824
8247
|
name: string;
|
|
7825
8248
|
api: "openai-completions";
|
|
7826
8249
|
provider: string;
|
|
7827
8250
|
baseUrl: string;
|
|
7828
|
-
reasoning:
|
|
7829
|
-
input: "text"[];
|
|
8251
|
+
reasoning: true;
|
|
8252
|
+
input: ("image" | "text")[];
|
|
7830
8253
|
cost: {
|
|
7831
8254
|
input: number;
|
|
7832
8255
|
output: number;
|
|
@@ -7836,13 +8259,13 @@ export declare const MODELS: {
|
|
|
7836
8259
|
contextWindow: number;
|
|
7837
8260
|
maxTokens: number;
|
|
7838
8261
|
};
|
|
7839
|
-
readonly "
|
|
8262
|
+
readonly "inception/mercury-2": {
|
|
7840
8263
|
id: string;
|
|
7841
8264
|
name: string;
|
|
7842
8265
|
api: "openai-completions";
|
|
7843
8266
|
provider: string;
|
|
7844
8267
|
baseUrl: string;
|
|
7845
|
-
reasoning:
|
|
8268
|
+
reasoning: true;
|
|
7846
8269
|
input: "text"[];
|
|
7847
8270
|
cost: {
|
|
7848
8271
|
input: number;
|
|
@@ -7853,7 +8276,7 @@ export declare const MODELS: {
|
|
|
7853
8276
|
contextWindow: number;
|
|
7854
8277
|
maxTokens: number;
|
|
7855
8278
|
};
|
|
7856
|
-
readonly "
|
|
8279
|
+
readonly "inclusionai/ling-2.6-flash:free": {
|
|
7857
8280
|
id: string;
|
|
7858
8281
|
name: string;
|
|
7859
8282
|
api: "openai-completions";
|
|
@@ -7870,7 +8293,7 @@ export declare const MODELS: {
|
|
|
7870
8293
|
contextWindow: number;
|
|
7871
8294
|
maxTokens: number;
|
|
7872
8295
|
};
|
|
7873
|
-
readonly "
|
|
8296
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
7874
8297
|
id: string;
|
|
7875
8298
|
name: string;
|
|
7876
8299
|
api: "openai-completions";
|
|
@@ -7887,7 +8310,7 @@ export declare const MODELS: {
|
|
|
7887
8310
|
contextWindow: number;
|
|
7888
8311
|
maxTokens: number;
|
|
7889
8312
|
};
|
|
7890
|
-
readonly "meta-llama/llama-3
|
|
8313
|
+
readonly "meta-llama/llama-3-8b-instruct": {
|
|
7891
8314
|
id: string;
|
|
7892
8315
|
name: string;
|
|
7893
8316
|
api: "openai-completions";
|
|
@@ -7904,7 +8327,7 @@ export declare const MODELS: {
|
|
|
7904
8327
|
contextWindow: number;
|
|
7905
8328
|
maxTokens: number;
|
|
7906
8329
|
};
|
|
7907
|
-
readonly "meta-llama/llama-3.1-
|
|
8330
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
7908
8331
|
id: string;
|
|
7909
8332
|
name: string;
|
|
7910
8333
|
api: "openai-completions";
|
|
@@ -7921,7 +8344,7 @@ export declare const MODELS: {
|
|
|
7921
8344
|
contextWindow: number;
|
|
7922
8345
|
maxTokens: number;
|
|
7923
8346
|
};
|
|
7924
|
-
readonly "meta-llama/llama-3.
|
|
8347
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
7925
8348
|
id: string;
|
|
7926
8349
|
name: string;
|
|
7927
8350
|
api: "openai-completions";
|
|
@@ -7938,7 +8361,7 @@ export declare const MODELS: {
|
|
|
7938
8361
|
contextWindow: number;
|
|
7939
8362
|
maxTokens: number;
|
|
7940
8363
|
};
|
|
7941
|
-
readonly "meta-llama/llama-3.3-70b-instruct
|
|
8364
|
+
readonly "meta-llama/llama-3.3-70b-instruct": {
|
|
7942
8365
|
id: string;
|
|
7943
8366
|
name: string;
|
|
7944
8367
|
api: "openai-completions";
|
|
@@ -7955,14 +8378,14 @@ export declare const MODELS: {
|
|
|
7955
8378
|
contextWindow: number;
|
|
7956
8379
|
maxTokens: number;
|
|
7957
8380
|
};
|
|
7958
|
-
readonly "meta-llama/llama-
|
|
8381
|
+
readonly "meta-llama/llama-3.3-70b-instruct:free": {
|
|
7959
8382
|
id: string;
|
|
7960
8383
|
name: string;
|
|
7961
8384
|
api: "openai-completions";
|
|
7962
8385
|
provider: string;
|
|
7963
8386
|
baseUrl: string;
|
|
7964
8387
|
reasoning: false;
|
|
7965
|
-
input:
|
|
8388
|
+
input: "text"[];
|
|
7966
8389
|
cost: {
|
|
7967
8390
|
input: number;
|
|
7968
8391
|
output: number;
|
|
@@ -8346,23 +8769,6 @@ export declare const MODELS: {
|
|
|
8346
8769
|
contextWindow: number;
|
|
8347
8770
|
maxTokens: number;
|
|
8348
8771
|
};
|
|
8349
|
-
readonly "mistralai/mistral-small-24b-instruct-2501": {
|
|
8350
|
-
id: string;
|
|
8351
|
-
name: string;
|
|
8352
|
-
api: "openai-completions";
|
|
8353
|
-
provider: string;
|
|
8354
|
-
baseUrl: string;
|
|
8355
|
-
reasoning: false;
|
|
8356
|
-
input: "text"[];
|
|
8357
|
-
cost: {
|
|
8358
|
-
input: number;
|
|
8359
|
-
output: number;
|
|
8360
|
-
cacheRead: number;
|
|
8361
|
-
cacheWrite: number;
|
|
8362
|
-
};
|
|
8363
|
-
contextWindow: number;
|
|
8364
|
-
maxTokens: number;
|
|
8365
|
-
};
|
|
8366
8772
|
readonly "mistralai/mistral-small-2603": {
|
|
8367
8773
|
id: string;
|
|
8368
8774
|
name: string;
|
|
@@ -8380,7 +8786,7 @@ export declare const MODELS: {
|
|
|
8380
8786
|
contextWindow: number;
|
|
8381
8787
|
maxTokens: number;
|
|
8382
8788
|
};
|
|
8383
|
-
readonly "mistralai/mistral-small-3.
|
|
8789
|
+
readonly "mistralai/mistral-small-3.2-24b-instruct": {
|
|
8384
8790
|
id: string;
|
|
8385
8791
|
name: string;
|
|
8386
8792
|
api: "openai-completions";
|
|
@@ -8397,14 +8803,14 @@ export declare const MODELS: {
|
|
|
8397
8803
|
contextWindow: number;
|
|
8398
8804
|
maxTokens: number;
|
|
8399
8805
|
};
|
|
8400
|
-
readonly "mistralai/mistral-small-
|
|
8806
|
+
readonly "mistralai/mistral-small-creative": {
|
|
8401
8807
|
id: string;
|
|
8402
8808
|
name: string;
|
|
8403
8809
|
api: "openai-completions";
|
|
8404
8810
|
provider: string;
|
|
8405
8811
|
baseUrl: string;
|
|
8406
8812
|
reasoning: false;
|
|
8407
|
-
input:
|
|
8813
|
+
input: "text"[];
|
|
8408
8814
|
cost: {
|
|
8409
8815
|
input: number;
|
|
8410
8816
|
output: number;
|
|
@@ -8414,7 +8820,7 @@ export declare const MODELS: {
|
|
|
8414
8820
|
contextWindow: number;
|
|
8415
8821
|
maxTokens: number;
|
|
8416
8822
|
};
|
|
8417
|
-
readonly "mistralai/
|
|
8823
|
+
readonly "mistralai/mixtral-8x22b-instruct": {
|
|
8418
8824
|
id: string;
|
|
8419
8825
|
name: string;
|
|
8420
8826
|
api: "openai-completions";
|
|
@@ -8431,7 +8837,7 @@ export declare const MODELS: {
|
|
|
8431
8837
|
contextWindow: number;
|
|
8432
8838
|
maxTokens: number;
|
|
8433
8839
|
};
|
|
8434
|
-
readonly "mistralai/mixtral-
|
|
8840
|
+
readonly "mistralai/mixtral-8x7b-instruct": {
|
|
8435
8841
|
id: string;
|
|
8436
8842
|
name: string;
|
|
8437
8843
|
api: "openai-completions";
|
|
@@ -8448,14 +8854,14 @@ export declare const MODELS: {
|
|
|
8448
8854
|
contextWindow: number;
|
|
8449
8855
|
maxTokens: number;
|
|
8450
8856
|
};
|
|
8451
|
-
readonly "mistralai/
|
|
8857
|
+
readonly "mistralai/pixtral-large-2411": {
|
|
8452
8858
|
id: string;
|
|
8453
8859
|
name: string;
|
|
8454
8860
|
api: "openai-completions";
|
|
8455
8861
|
provider: string;
|
|
8456
8862
|
baseUrl: string;
|
|
8457
8863
|
reasoning: false;
|
|
8458
|
-
input: "text"[];
|
|
8864
|
+
input: ("image" | "text")[];
|
|
8459
8865
|
cost: {
|
|
8460
8866
|
input: number;
|
|
8461
8867
|
output: number;
|
|
@@ -8465,14 +8871,14 @@ export declare const MODELS: {
|
|
|
8465
8871
|
contextWindow: number;
|
|
8466
8872
|
maxTokens: number;
|
|
8467
8873
|
};
|
|
8468
|
-
readonly "mistralai/
|
|
8874
|
+
readonly "mistralai/voxtral-small-24b-2507": {
|
|
8469
8875
|
id: string;
|
|
8470
8876
|
name: string;
|
|
8471
8877
|
api: "openai-completions";
|
|
8472
8878
|
provider: string;
|
|
8473
8879
|
baseUrl: string;
|
|
8474
8880
|
reasoning: false;
|
|
8475
|
-
input:
|
|
8881
|
+
input: "text"[];
|
|
8476
8882
|
cost: {
|
|
8477
8883
|
input: number;
|
|
8478
8884
|
output: number;
|
|
@@ -8482,7 +8888,7 @@ export declare const MODELS: {
|
|
|
8482
8888
|
contextWindow: number;
|
|
8483
8889
|
maxTokens: number;
|
|
8484
8890
|
};
|
|
8485
|
-
readonly "
|
|
8891
|
+
readonly "moonshotai/kimi-k2": {
|
|
8486
8892
|
id: string;
|
|
8487
8893
|
name: string;
|
|
8488
8894
|
api: "openai-completions";
|
|
@@ -8499,7 +8905,7 @@ export declare const MODELS: {
|
|
|
8499
8905
|
contextWindow: number;
|
|
8500
8906
|
maxTokens: number;
|
|
8501
8907
|
};
|
|
8502
|
-
readonly "moonshotai/kimi-k2": {
|
|
8908
|
+
readonly "moonshotai/kimi-k2-0905": {
|
|
8503
8909
|
id: string;
|
|
8504
8910
|
name: string;
|
|
8505
8911
|
api: "openai-completions";
|
|
@@ -8516,13 +8922,13 @@ export declare const MODELS: {
|
|
|
8516
8922
|
contextWindow: number;
|
|
8517
8923
|
maxTokens: number;
|
|
8518
8924
|
};
|
|
8519
|
-
readonly "moonshotai/kimi-k2-
|
|
8925
|
+
readonly "moonshotai/kimi-k2-thinking": {
|
|
8520
8926
|
id: string;
|
|
8521
8927
|
name: string;
|
|
8522
8928
|
api: "openai-completions";
|
|
8523
8929
|
provider: string;
|
|
8524
8930
|
baseUrl: string;
|
|
8525
|
-
reasoning:
|
|
8931
|
+
reasoning: true;
|
|
8526
8932
|
input: "text"[];
|
|
8527
8933
|
cost: {
|
|
8528
8934
|
input: number;
|
|
@@ -8533,14 +8939,14 @@ export declare const MODELS: {
|
|
|
8533
8939
|
contextWindow: number;
|
|
8534
8940
|
maxTokens: number;
|
|
8535
8941
|
};
|
|
8536
|
-
readonly "moonshotai/kimi-k2
|
|
8942
|
+
readonly "moonshotai/kimi-k2.5": {
|
|
8537
8943
|
id: string;
|
|
8538
8944
|
name: string;
|
|
8539
8945
|
api: "openai-completions";
|
|
8540
8946
|
provider: string;
|
|
8541
8947
|
baseUrl: string;
|
|
8542
8948
|
reasoning: true;
|
|
8543
|
-
input: "text"[];
|
|
8949
|
+
input: ("image" | "text")[];
|
|
8544
8950
|
cost: {
|
|
8545
8951
|
input: number;
|
|
8546
8952
|
output: number;
|
|
@@ -8550,7 +8956,7 @@ export declare const MODELS: {
|
|
|
8550
8956
|
contextWindow: number;
|
|
8551
8957
|
maxTokens: number;
|
|
8552
8958
|
};
|
|
8553
|
-
readonly "moonshotai/kimi-k2.
|
|
8959
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
8554
8960
|
id: string;
|
|
8555
8961
|
name: string;
|
|
8556
8962
|
api: "openai-completions";
|
|
@@ -9043,23 +9449,6 @@ export declare const MODELS: {
|
|
|
9043
9449
|
contextWindow: number;
|
|
9044
9450
|
maxTokens: number;
|
|
9045
9451
|
};
|
|
9046
|
-
readonly "openai/gpt-4o:extended": {
|
|
9047
|
-
id: string;
|
|
9048
|
-
name: string;
|
|
9049
|
-
api: "openai-completions";
|
|
9050
|
-
provider: string;
|
|
9051
|
-
baseUrl: string;
|
|
9052
|
-
reasoning: false;
|
|
9053
|
-
input: ("image" | "text")[];
|
|
9054
|
-
cost: {
|
|
9055
|
-
input: number;
|
|
9056
|
-
output: number;
|
|
9057
|
-
cacheRead: number;
|
|
9058
|
-
cacheWrite: number;
|
|
9059
|
-
};
|
|
9060
|
-
contextWindow: number;
|
|
9061
|
-
maxTokens: number;
|
|
9062
|
-
};
|
|
9063
9452
|
readonly "openai/gpt-5": {
|
|
9064
9453
|
id: string;
|
|
9065
9454
|
name: string;
|
|
@@ -9094,7 +9483,7 @@ export declare const MODELS: {
|
|
|
9094
9483
|
contextWindow: number;
|
|
9095
9484
|
maxTokens: number;
|
|
9096
9485
|
};
|
|
9097
|
-
readonly "openai/gpt-5-
|
|
9486
|
+
readonly "openai/gpt-5-mini": {
|
|
9098
9487
|
id: string;
|
|
9099
9488
|
name: string;
|
|
9100
9489
|
api: "openai-completions";
|
|
@@ -9111,7 +9500,7 @@ export declare const MODELS: {
|
|
|
9111
9500
|
contextWindow: number;
|
|
9112
9501
|
maxTokens: number;
|
|
9113
9502
|
};
|
|
9114
|
-
readonly "openai/gpt-5-
|
|
9503
|
+
readonly "openai/gpt-5-nano": {
|
|
9115
9504
|
id: string;
|
|
9116
9505
|
name: string;
|
|
9117
9506
|
api: "openai-completions";
|
|
@@ -9128,7 +9517,7 @@ export declare const MODELS: {
|
|
|
9128
9517
|
contextWindow: number;
|
|
9129
9518
|
maxTokens: number;
|
|
9130
9519
|
};
|
|
9131
|
-
readonly "openai/gpt-5-
|
|
9520
|
+
readonly "openai/gpt-5-pro": {
|
|
9132
9521
|
id: string;
|
|
9133
9522
|
name: string;
|
|
9134
9523
|
api: "openai-completions";
|
|
@@ -9145,7 +9534,7 @@ export declare const MODELS: {
|
|
|
9145
9534
|
contextWindow: number;
|
|
9146
9535
|
maxTokens: number;
|
|
9147
9536
|
};
|
|
9148
|
-
readonly "openai/gpt-5
|
|
9537
|
+
readonly "openai/gpt-5.1": {
|
|
9149
9538
|
id: string;
|
|
9150
9539
|
name: string;
|
|
9151
9540
|
api: "openai-completions";
|
|
@@ -9162,13 +9551,13 @@ export declare const MODELS: {
|
|
|
9162
9551
|
contextWindow: number;
|
|
9163
9552
|
maxTokens: number;
|
|
9164
9553
|
};
|
|
9165
|
-
readonly "openai/gpt-5-
|
|
9554
|
+
readonly "openai/gpt-5.1-chat": {
|
|
9166
9555
|
id: string;
|
|
9167
9556
|
name: string;
|
|
9168
9557
|
api: "openai-completions";
|
|
9169
9558
|
provider: string;
|
|
9170
9559
|
baseUrl: string;
|
|
9171
|
-
reasoning:
|
|
9560
|
+
reasoning: false;
|
|
9172
9561
|
input: ("image" | "text")[];
|
|
9173
9562
|
cost: {
|
|
9174
9563
|
input: number;
|
|
@@ -9179,7 +9568,7 @@ export declare const MODELS: {
|
|
|
9179
9568
|
contextWindow: number;
|
|
9180
9569
|
maxTokens: number;
|
|
9181
9570
|
};
|
|
9182
|
-
readonly "openai/gpt-5.1": {
|
|
9571
|
+
readonly "openai/gpt-5.1-codex": {
|
|
9183
9572
|
id: string;
|
|
9184
9573
|
name: string;
|
|
9185
9574
|
api: "openai-completions";
|
|
@@ -9196,13 +9585,13 @@ export declare const MODELS: {
|
|
|
9196
9585
|
contextWindow: number;
|
|
9197
9586
|
maxTokens: number;
|
|
9198
9587
|
};
|
|
9199
|
-
readonly "openai/gpt-5.1-
|
|
9588
|
+
readonly "openai/gpt-5.1-codex-max": {
|
|
9200
9589
|
id: string;
|
|
9201
9590
|
name: string;
|
|
9202
9591
|
api: "openai-completions";
|
|
9203
9592
|
provider: string;
|
|
9204
9593
|
baseUrl: string;
|
|
9205
|
-
reasoning:
|
|
9594
|
+
reasoning: true;
|
|
9206
9595
|
input: ("image" | "text")[];
|
|
9207
9596
|
cost: {
|
|
9208
9597
|
input: number;
|
|
@@ -9213,7 +9602,7 @@ export declare const MODELS: {
|
|
|
9213
9602
|
contextWindow: number;
|
|
9214
9603
|
maxTokens: number;
|
|
9215
9604
|
};
|
|
9216
|
-
readonly "openai/gpt-5.1-codex": {
|
|
9605
|
+
readonly "openai/gpt-5.1-codex-mini": {
|
|
9217
9606
|
id: string;
|
|
9218
9607
|
name: string;
|
|
9219
9608
|
api: "openai-completions";
|
|
@@ -9230,7 +9619,7 @@ export declare const MODELS: {
|
|
|
9230
9619
|
contextWindow: number;
|
|
9231
9620
|
maxTokens: number;
|
|
9232
9621
|
};
|
|
9233
|
-
readonly "openai/gpt-5.
|
|
9622
|
+
readonly "openai/gpt-5.2": {
|
|
9234
9623
|
id: string;
|
|
9235
9624
|
name: string;
|
|
9236
9625
|
api: "openai-completions";
|
|
@@ -9247,13 +9636,13 @@ export declare const MODELS: {
|
|
|
9247
9636
|
contextWindow: number;
|
|
9248
9637
|
maxTokens: number;
|
|
9249
9638
|
};
|
|
9250
|
-
readonly "openai/gpt-5.
|
|
9639
|
+
readonly "openai/gpt-5.2-chat": {
|
|
9251
9640
|
id: string;
|
|
9252
9641
|
name: string;
|
|
9253
9642
|
api: "openai-completions";
|
|
9254
9643
|
provider: string;
|
|
9255
9644
|
baseUrl: string;
|
|
9256
|
-
reasoning:
|
|
9645
|
+
reasoning: false;
|
|
9257
9646
|
input: ("image" | "text")[];
|
|
9258
9647
|
cost: {
|
|
9259
9648
|
input: number;
|
|
@@ -9264,7 +9653,7 @@ export declare const MODELS: {
|
|
|
9264
9653
|
contextWindow: number;
|
|
9265
9654
|
maxTokens: number;
|
|
9266
9655
|
};
|
|
9267
|
-
readonly "openai/gpt-5.2": {
|
|
9656
|
+
readonly "openai/gpt-5.2-codex": {
|
|
9268
9657
|
id: string;
|
|
9269
9658
|
name: string;
|
|
9270
9659
|
api: "openai-completions";
|
|
@@ -9281,13 +9670,13 @@ export declare const MODELS: {
|
|
|
9281
9670
|
contextWindow: number;
|
|
9282
9671
|
maxTokens: number;
|
|
9283
9672
|
};
|
|
9284
|
-
readonly "openai/gpt-5.2-
|
|
9673
|
+
readonly "openai/gpt-5.2-pro": {
|
|
9285
9674
|
id: string;
|
|
9286
9675
|
name: string;
|
|
9287
9676
|
api: "openai-completions";
|
|
9288
9677
|
provider: string;
|
|
9289
9678
|
baseUrl: string;
|
|
9290
|
-
reasoning:
|
|
9679
|
+
reasoning: true;
|
|
9291
9680
|
input: ("image" | "text")[];
|
|
9292
9681
|
cost: {
|
|
9293
9682
|
input: number;
|
|
@@ -9298,13 +9687,13 @@ export declare const MODELS: {
|
|
|
9298
9687
|
contextWindow: number;
|
|
9299
9688
|
maxTokens: number;
|
|
9300
9689
|
};
|
|
9301
|
-
readonly "openai/gpt-5.
|
|
9690
|
+
readonly "openai/gpt-5.3-chat": {
|
|
9302
9691
|
id: string;
|
|
9303
9692
|
name: string;
|
|
9304
9693
|
api: "openai-completions";
|
|
9305
9694
|
provider: string;
|
|
9306
9695
|
baseUrl: string;
|
|
9307
|
-
reasoning:
|
|
9696
|
+
reasoning: false;
|
|
9308
9697
|
input: ("image" | "text")[];
|
|
9309
9698
|
cost: {
|
|
9310
9699
|
input: number;
|
|
@@ -9315,7 +9704,7 @@ export declare const MODELS: {
|
|
|
9315
9704
|
contextWindow: number;
|
|
9316
9705
|
maxTokens: number;
|
|
9317
9706
|
};
|
|
9318
|
-
readonly "openai/gpt-5.
|
|
9707
|
+
readonly "openai/gpt-5.3-codex": {
|
|
9319
9708
|
id: string;
|
|
9320
9709
|
name: string;
|
|
9321
9710
|
api: "openai-completions";
|
|
@@ -9332,13 +9721,13 @@ export declare const MODELS: {
|
|
|
9332
9721
|
contextWindow: number;
|
|
9333
9722
|
maxTokens: number;
|
|
9334
9723
|
};
|
|
9335
|
-
readonly "openai/gpt-5.
|
|
9724
|
+
readonly "openai/gpt-5.4": {
|
|
9336
9725
|
id: string;
|
|
9337
9726
|
name: string;
|
|
9338
9727
|
api: "openai-completions";
|
|
9339
9728
|
provider: string;
|
|
9340
9729
|
baseUrl: string;
|
|
9341
|
-
reasoning:
|
|
9730
|
+
reasoning: true;
|
|
9342
9731
|
input: ("image" | "text")[];
|
|
9343
9732
|
cost: {
|
|
9344
9733
|
input: number;
|
|
@@ -9349,7 +9738,7 @@ export declare const MODELS: {
|
|
|
9349
9738
|
contextWindow: number;
|
|
9350
9739
|
maxTokens: number;
|
|
9351
9740
|
};
|
|
9352
|
-
readonly "openai/gpt-5.
|
|
9741
|
+
readonly "openai/gpt-5.4-mini": {
|
|
9353
9742
|
id: string;
|
|
9354
9743
|
name: string;
|
|
9355
9744
|
api: "openai-completions";
|
|
@@ -9366,7 +9755,7 @@ export declare const MODELS: {
|
|
|
9366
9755
|
contextWindow: number;
|
|
9367
9756
|
maxTokens: number;
|
|
9368
9757
|
};
|
|
9369
|
-
readonly "openai/gpt-5.4": {
|
|
9758
|
+
readonly "openai/gpt-5.4-nano": {
|
|
9370
9759
|
id: string;
|
|
9371
9760
|
name: string;
|
|
9372
9761
|
api: "openai-completions";
|
|
@@ -9383,7 +9772,7 @@ export declare const MODELS: {
|
|
|
9383
9772
|
contextWindow: number;
|
|
9384
9773
|
maxTokens: number;
|
|
9385
9774
|
};
|
|
9386
|
-
readonly "openai/gpt-5.4-
|
|
9775
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9387
9776
|
id: string;
|
|
9388
9777
|
name: string;
|
|
9389
9778
|
api: "openai-completions";
|
|
@@ -9400,14 +9789,14 @@ export declare const MODELS: {
|
|
|
9400
9789
|
contextWindow: number;
|
|
9401
9790
|
maxTokens: number;
|
|
9402
9791
|
};
|
|
9403
|
-
readonly "openai/gpt-
|
|
9792
|
+
readonly "openai/gpt-audio": {
|
|
9404
9793
|
id: string;
|
|
9405
9794
|
name: string;
|
|
9406
9795
|
api: "openai-completions";
|
|
9407
9796
|
provider: string;
|
|
9408
9797
|
baseUrl: string;
|
|
9409
|
-
reasoning:
|
|
9410
|
-
input:
|
|
9798
|
+
reasoning: false;
|
|
9799
|
+
input: "text"[];
|
|
9411
9800
|
cost: {
|
|
9412
9801
|
input: number;
|
|
9413
9802
|
output: number;
|
|
@@ -9417,14 +9806,14 @@ export declare const MODELS: {
|
|
|
9417
9806
|
contextWindow: number;
|
|
9418
9807
|
maxTokens: number;
|
|
9419
9808
|
};
|
|
9420
|
-
readonly "openai/gpt-
|
|
9809
|
+
readonly "openai/gpt-audio-mini": {
|
|
9421
9810
|
id: string;
|
|
9422
9811
|
name: string;
|
|
9423
9812
|
api: "openai-completions";
|
|
9424
9813
|
provider: string;
|
|
9425
9814
|
baseUrl: string;
|
|
9426
|
-
reasoning:
|
|
9427
|
-
input:
|
|
9815
|
+
reasoning: false;
|
|
9816
|
+
input: "text"[];
|
|
9428
9817
|
cost: {
|
|
9429
9818
|
input: number;
|
|
9430
9819
|
output: number;
|
|
@@ -9899,7 +10288,7 @@ export declare const MODELS: {
|
|
|
9899
10288
|
api: "openai-completions";
|
|
9900
10289
|
provider: string;
|
|
9901
10290
|
baseUrl: string;
|
|
9902
|
-
reasoning:
|
|
10291
|
+
reasoning: false;
|
|
9903
10292
|
input: "text"[];
|
|
9904
10293
|
cost: {
|
|
9905
10294
|
input: number;
|
|
@@ -9995,23 +10384,6 @@ export declare const MODELS: {
|
|
|
9995
10384
|
contextWindow: number;
|
|
9996
10385
|
maxTokens: number;
|
|
9997
10386
|
};
|
|
9998
|
-
readonly "qwen/qwen3-4b:free": {
|
|
9999
|
-
id: string;
|
|
10000
|
-
name: string;
|
|
10001
|
-
api: "openai-completions";
|
|
10002
|
-
provider: string;
|
|
10003
|
-
baseUrl: string;
|
|
10004
|
-
reasoning: true;
|
|
10005
|
-
input: "text"[];
|
|
10006
|
-
cost: {
|
|
10007
|
-
input: number;
|
|
10008
|
-
output: number;
|
|
10009
|
-
cacheRead: number;
|
|
10010
|
-
cacheWrite: number;
|
|
10011
|
-
};
|
|
10012
|
-
contextWindow: number;
|
|
10013
|
-
maxTokens: number;
|
|
10014
|
-
};
|
|
10015
10387
|
readonly "qwen/qwen3-8b": {
|
|
10016
10388
|
id: string;
|
|
10017
10389
|
name: string;
|
|
@@ -10454,6 +10826,23 @@ export declare const MODELS: {
|
|
|
10454
10826
|
contextWindow: number;
|
|
10455
10827
|
maxTokens: number;
|
|
10456
10828
|
};
|
|
10829
|
+
readonly "qwen/qwen3.6-plus": {
|
|
10830
|
+
id: string;
|
|
10831
|
+
name: string;
|
|
10832
|
+
api: "openai-completions";
|
|
10833
|
+
provider: string;
|
|
10834
|
+
baseUrl: string;
|
|
10835
|
+
reasoning: true;
|
|
10836
|
+
input: ("image" | "text")[];
|
|
10837
|
+
cost: {
|
|
10838
|
+
input: number;
|
|
10839
|
+
output: number;
|
|
10840
|
+
cacheRead: number;
|
|
10841
|
+
cacheWrite: number;
|
|
10842
|
+
};
|
|
10843
|
+
contextWindow: number;
|
|
10844
|
+
maxTokens: number;
|
|
10845
|
+
};
|
|
10457
10846
|
readonly "qwen/qwq-32b": {
|
|
10458
10847
|
id: string;
|
|
10459
10848
|
name: string;
|
|
@@ -10471,6 +10860,23 @@ export declare const MODELS: {
|
|
|
10471
10860
|
contextWindow: number;
|
|
10472
10861
|
maxTokens: number;
|
|
10473
10862
|
};
|
|
10863
|
+
readonly "rekaai/reka-edge": {
|
|
10864
|
+
id: string;
|
|
10865
|
+
name: string;
|
|
10866
|
+
api: "openai-completions";
|
|
10867
|
+
provider: string;
|
|
10868
|
+
baseUrl: string;
|
|
10869
|
+
reasoning: false;
|
|
10870
|
+
input: ("image" | "text")[];
|
|
10871
|
+
cost: {
|
|
10872
|
+
input: number;
|
|
10873
|
+
output: number;
|
|
10874
|
+
cacheRead: number;
|
|
10875
|
+
cacheWrite: number;
|
|
10876
|
+
};
|
|
10877
|
+
contextWindow: number;
|
|
10878
|
+
maxTokens: number;
|
|
10879
|
+
};
|
|
10474
10880
|
readonly "relace/relace-search": {
|
|
10475
10881
|
id: string;
|
|
10476
10882
|
name: string;
|
|
@@ -10539,7 +10945,7 @@ export declare const MODELS: {
|
|
|
10539
10945
|
contextWindow: number;
|
|
10540
10946
|
maxTokens: number;
|
|
10541
10947
|
};
|
|
10542
|
-
readonly "
|
|
10948
|
+
readonly "tencent/hy3-preview:free": {
|
|
10543
10949
|
id: string;
|
|
10544
10950
|
name: string;
|
|
10545
10951
|
api: "openai-completions";
|
|
@@ -10743,7 +11149,7 @@ export declare const MODELS: {
|
|
|
10743
11149
|
contextWindow: number;
|
|
10744
11150
|
maxTokens: number;
|
|
10745
11151
|
};
|
|
10746
|
-
readonly "x-ai/grok-4.20
|
|
11152
|
+
readonly "x-ai/grok-4.20": {
|
|
10747
11153
|
id: string;
|
|
10748
11154
|
name: string;
|
|
10749
11155
|
api: "openai-completions";
|
|
@@ -10828,6 +11234,40 @@ export declare const MODELS: {
|
|
|
10828
11234
|
contextWindow: number;
|
|
10829
11235
|
maxTokens: number;
|
|
10830
11236
|
};
|
|
11237
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
11238
|
+
id: string;
|
|
11239
|
+
name: string;
|
|
11240
|
+
api: "openai-completions";
|
|
11241
|
+
provider: string;
|
|
11242
|
+
baseUrl: string;
|
|
11243
|
+
reasoning: true;
|
|
11244
|
+
input: ("image" | "text")[];
|
|
11245
|
+
cost: {
|
|
11246
|
+
input: number;
|
|
11247
|
+
output: number;
|
|
11248
|
+
cacheRead: number;
|
|
11249
|
+
cacheWrite: number;
|
|
11250
|
+
};
|
|
11251
|
+
contextWindow: number;
|
|
11252
|
+
maxTokens: number;
|
|
11253
|
+
};
|
|
11254
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
11255
|
+
id: string;
|
|
11256
|
+
name: string;
|
|
11257
|
+
api: "openai-completions";
|
|
11258
|
+
provider: string;
|
|
11259
|
+
baseUrl: string;
|
|
11260
|
+
reasoning: true;
|
|
11261
|
+
input: "text"[];
|
|
11262
|
+
cost: {
|
|
11263
|
+
input: number;
|
|
11264
|
+
output: number;
|
|
11265
|
+
cacheRead: number;
|
|
11266
|
+
cacheWrite: number;
|
|
11267
|
+
};
|
|
11268
|
+
contextWindow: number;
|
|
11269
|
+
maxTokens: number;
|
|
11270
|
+
};
|
|
10831
11271
|
readonly "z-ai/glm-4-32b": {
|
|
10832
11272
|
id: string;
|
|
10833
11273
|
name: string;
|
|
@@ -10845,14 +11285,65 @@ export declare const MODELS: {
|
|
|
10845
11285
|
contextWindow: number;
|
|
10846
11286
|
maxTokens: number;
|
|
10847
11287
|
};
|
|
10848
|
-
readonly "z-ai/glm-4.5": {
|
|
11288
|
+
readonly "z-ai/glm-4.5": {
|
|
11289
|
+
id: string;
|
|
11290
|
+
name: string;
|
|
11291
|
+
api: "openai-completions";
|
|
11292
|
+
provider: string;
|
|
11293
|
+
baseUrl: string;
|
|
11294
|
+
reasoning: true;
|
|
11295
|
+
input: "text"[];
|
|
11296
|
+
cost: {
|
|
11297
|
+
input: number;
|
|
11298
|
+
output: number;
|
|
11299
|
+
cacheRead: number;
|
|
11300
|
+
cacheWrite: number;
|
|
11301
|
+
};
|
|
11302
|
+
contextWindow: number;
|
|
11303
|
+
maxTokens: number;
|
|
11304
|
+
};
|
|
11305
|
+
readonly "z-ai/glm-4.5-air": {
|
|
11306
|
+
id: string;
|
|
11307
|
+
name: string;
|
|
11308
|
+
api: "openai-completions";
|
|
11309
|
+
provider: string;
|
|
11310
|
+
baseUrl: string;
|
|
11311
|
+
reasoning: true;
|
|
11312
|
+
input: "text"[];
|
|
11313
|
+
cost: {
|
|
11314
|
+
input: number;
|
|
11315
|
+
output: number;
|
|
11316
|
+
cacheRead: number;
|
|
11317
|
+
cacheWrite: number;
|
|
11318
|
+
};
|
|
11319
|
+
contextWindow: number;
|
|
11320
|
+
maxTokens: number;
|
|
11321
|
+
};
|
|
11322
|
+
readonly "z-ai/glm-4.5-air:free": {
|
|
11323
|
+
id: string;
|
|
11324
|
+
name: string;
|
|
11325
|
+
api: "openai-completions";
|
|
11326
|
+
provider: string;
|
|
11327
|
+
baseUrl: string;
|
|
11328
|
+
reasoning: true;
|
|
11329
|
+
input: "text"[];
|
|
11330
|
+
cost: {
|
|
11331
|
+
input: number;
|
|
11332
|
+
output: number;
|
|
11333
|
+
cacheRead: number;
|
|
11334
|
+
cacheWrite: number;
|
|
11335
|
+
};
|
|
11336
|
+
contextWindow: number;
|
|
11337
|
+
maxTokens: number;
|
|
11338
|
+
};
|
|
11339
|
+
readonly "z-ai/glm-4.5v": {
|
|
10849
11340
|
id: string;
|
|
10850
11341
|
name: string;
|
|
10851
11342
|
api: "openai-completions";
|
|
10852
11343
|
provider: string;
|
|
10853
11344
|
baseUrl: string;
|
|
10854
11345
|
reasoning: true;
|
|
10855
|
-
input: "text"[];
|
|
11346
|
+
input: ("image" | "text")[];
|
|
10856
11347
|
cost: {
|
|
10857
11348
|
input: number;
|
|
10858
11349
|
output: number;
|
|
@@ -10862,7 +11353,7 @@ export declare const MODELS: {
|
|
|
10862
11353
|
contextWindow: number;
|
|
10863
11354
|
maxTokens: number;
|
|
10864
11355
|
};
|
|
10865
|
-
readonly "z-ai/glm-4.
|
|
11356
|
+
readonly "z-ai/glm-4.6": {
|
|
10866
11357
|
id: string;
|
|
10867
11358
|
name: string;
|
|
10868
11359
|
api: "openai-completions";
|
|
@@ -10879,14 +11370,14 @@ export declare const MODELS: {
|
|
|
10879
11370
|
contextWindow: number;
|
|
10880
11371
|
maxTokens: number;
|
|
10881
11372
|
};
|
|
10882
|
-
readonly "z-ai/glm-4.
|
|
11373
|
+
readonly "z-ai/glm-4.6v": {
|
|
10883
11374
|
id: string;
|
|
10884
11375
|
name: string;
|
|
10885
11376
|
api: "openai-completions";
|
|
10886
11377
|
provider: string;
|
|
10887
11378
|
baseUrl: string;
|
|
10888
11379
|
reasoning: true;
|
|
10889
|
-
input: "text"[];
|
|
11380
|
+
input: ("image" | "text")[];
|
|
10890
11381
|
cost: {
|
|
10891
11382
|
input: number;
|
|
10892
11383
|
output: number;
|
|
@@ -10896,14 +11387,14 @@ export declare const MODELS: {
|
|
|
10896
11387
|
contextWindow: number;
|
|
10897
11388
|
maxTokens: number;
|
|
10898
11389
|
};
|
|
10899
|
-
readonly "z-ai/glm-4.
|
|
11390
|
+
readonly "z-ai/glm-4.7": {
|
|
10900
11391
|
id: string;
|
|
10901
11392
|
name: string;
|
|
10902
11393
|
api: "openai-completions";
|
|
10903
11394
|
provider: string;
|
|
10904
11395
|
baseUrl: string;
|
|
10905
11396
|
reasoning: true;
|
|
10906
|
-
input:
|
|
11397
|
+
input: "text"[];
|
|
10907
11398
|
cost: {
|
|
10908
11399
|
input: number;
|
|
10909
11400
|
output: number;
|
|
@@ -10913,7 +11404,7 @@ export declare const MODELS: {
|
|
|
10913
11404
|
contextWindow: number;
|
|
10914
11405
|
maxTokens: number;
|
|
10915
11406
|
};
|
|
10916
|
-
readonly "z-ai/glm-4.
|
|
11407
|
+
readonly "z-ai/glm-4.7-flash": {
|
|
10917
11408
|
id: string;
|
|
10918
11409
|
name: string;
|
|
10919
11410
|
api: "openai-completions";
|
|
@@ -10930,14 +11421,14 @@ export declare const MODELS: {
|
|
|
10930
11421
|
contextWindow: number;
|
|
10931
11422
|
maxTokens: number;
|
|
10932
11423
|
};
|
|
10933
|
-
readonly "z-ai/glm-
|
|
11424
|
+
readonly "z-ai/glm-5": {
|
|
10934
11425
|
id: string;
|
|
10935
11426
|
name: string;
|
|
10936
11427
|
api: "openai-completions";
|
|
10937
11428
|
provider: string;
|
|
10938
11429
|
baseUrl: string;
|
|
10939
11430
|
reasoning: true;
|
|
10940
|
-
input:
|
|
11431
|
+
input: "text"[];
|
|
10941
11432
|
cost: {
|
|
10942
11433
|
input: number;
|
|
10943
11434
|
output: number;
|
|
@@ -10947,7 +11438,7 @@ export declare const MODELS: {
|
|
|
10947
11438
|
contextWindow: number;
|
|
10948
11439
|
maxTokens: number;
|
|
10949
11440
|
};
|
|
10950
|
-
readonly "z-ai/glm-
|
|
11441
|
+
readonly "z-ai/glm-5-turbo": {
|
|
10951
11442
|
id: string;
|
|
10952
11443
|
name: string;
|
|
10953
11444
|
api: "openai-completions";
|
|
@@ -10964,7 +11455,7 @@ export declare const MODELS: {
|
|
|
10964
11455
|
contextWindow: number;
|
|
10965
11456
|
maxTokens: number;
|
|
10966
11457
|
};
|
|
10967
|
-
readonly "z-ai/glm-
|
|
11458
|
+
readonly "z-ai/glm-5.1": {
|
|
10968
11459
|
id: string;
|
|
10969
11460
|
name: string;
|
|
10970
11461
|
api: "openai-completions";
|
|
@@ -10981,14 +11472,14 @@ export declare const MODELS: {
|
|
|
10981
11472
|
contextWindow: number;
|
|
10982
11473
|
maxTokens: number;
|
|
10983
11474
|
};
|
|
10984
|
-
readonly "z-ai/glm-
|
|
11475
|
+
readonly "z-ai/glm-5v-turbo": {
|
|
10985
11476
|
id: string;
|
|
10986
11477
|
name: string;
|
|
10987
11478
|
api: "openai-completions";
|
|
10988
11479
|
provider: string;
|
|
10989
11480
|
baseUrl: string;
|
|
10990
11481
|
reasoning: true;
|
|
10991
|
-
input: "text"[];
|
|
11482
|
+
input: ("image" | "text")[];
|
|
10992
11483
|
cost: {
|
|
10993
11484
|
input: number;
|
|
10994
11485
|
output: number;
|
|
@@ -10998,14 +11489,14 @@ export declare const MODELS: {
|
|
|
10998
11489
|
contextWindow: number;
|
|
10999
11490
|
maxTokens: number;
|
|
11000
11491
|
};
|
|
11001
|
-
readonly "
|
|
11492
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
11002
11493
|
id: string;
|
|
11003
11494
|
name: string;
|
|
11004
11495
|
api: "openai-completions";
|
|
11005
11496
|
provider: string;
|
|
11006
11497
|
baseUrl: string;
|
|
11007
11498
|
reasoning: true;
|
|
11008
|
-
input: "text"[];
|
|
11499
|
+
input: ("image" | "text")[];
|
|
11009
11500
|
cost: {
|
|
11010
11501
|
input: number;
|
|
11011
11502
|
output: number;
|
|
@@ -11085,6 +11576,23 @@ export declare const MODELS: {
|
|
|
11085
11576
|
contextWindow: number;
|
|
11086
11577
|
maxTokens: number;
|
|
11087
11578
|
};
|
|
11579
|
+
readonly "alibaba/qwen-3.6-max-preview": {
|
|
11580
|
+
id: string;
|
|
11581
|
+
name: string;
|
|
11582
|
+
api: "anthropic-messages";
|
|
11583
|
+
provider: string;
|
|
11584
|
+
baseUrl: string;
|
|
11585
|
+
reasoning: true;
|
|
11586
|
+
input: ("image" | "text")[];
|
|
11587
|
+
cost: {
|
|
11588
|
+
input: number;
|
|
11589
|
+
output: number;
|
|
11590
|
+
cacheRead: number;
|
|
11591
|
+
cacheWrite: number;
|
|
11592
|
+
};
|
|
11593
|
+
contextWindow: number;
|
|
11594
|
+
maxTokens: number;
|
|
11595
|
+
};
|
|
11088
11596
|
readonly "alibaba/qwen3-235b-a22b-thinking": {
|
|
11089
11597
|
id: string;
|
|
11090
11598
|
name: string;
|
|
@@ -11272,13 +11780,13 @@ export declare const MODELS: {
|
|
|
11272
11780
|
contextWindow: number;
|
|
11273
11781
|
maxTokens: number;
|
|
11274
11782
|
};
|
|
11275
|
-
readonly "
|
|
11783
|
+
readonly "alibaba/qwen3.6-plus": {
|
|
11276
11784
|
id: string;
|
|
11277
11785
|
name: string;
|
|
11278
11786
|
api: "anthropic-messages";
|
|
11279
11787
|
provider: string;
|
|
11280
11788
|
baseUrl: string;
|
|
11281
|
-
reasoning:
|
|
11789
|
+
reasoning: true;
|
|
11282
11790
|
input: ("image" | "text")[];
|
|
11283
11791
|
cost: {
|
|
11284
11792
|
input: number;
|
|
@@ -11289,7 +11797,7 @@ export declare const MODELS: {
|
|
|
11289
11797
|
contextWindow: number;
|
|
11290
11798
|
maxTokens: number;
|
|
11291
11799
|
};
|
|
11292
|
-
readonly "anthropic/claude-3
|
|
11800
|
+
readonly "anthropic/claude-3-haiku": {
|
|
11293
11801
|
id: string;
|
|
11294
11802
|
name: string;
|
|
11295
11803
|
api: "anthropic-messages";
|
|
@@ -11306,7 +11814,7 @@ export declare const MODELS: {
|
|
|
11306
11814
|
contextWindow: number;
|
|
11307
11815
|
maxTokens: number;
|
|
11308
11816
|
};
|
|
11309
|
-
readonly "anthropic/claude-3.5-
|
|
11817
|
+
readonly "anthropic/claude-3.5-haiku": {
|
|
11310
11818
|
id: string;
|
|
11311
11819
|
name: string;
|
|
11312
11820
|
api: "anthropic-messages";
|
|
@@ -11323,13 +11831,13 @@ export declare const MODELS: {
|
|
|
11323
11831
|
contextWindow: number;
|
|
11324
11832
|
maxTokens: number;
|
|
11325
11833
|
};
|
|
11326
|
-
readonly "anthropic/claude-3.
|
|
11834
|
+
readonly "anthropic/claude-3.7-sonnet": {
|
|
11327
11835
|
id: string;
|
|
11328
11836
|
name: string;
|
|
11329
11837
|
api: "anthropic-messages";
|
|
11330
11838
|
provider: string;
|
|
11331
11839
|
baseUrl: string;
|
|
11332
|
-
reasoning:
|
|
11840
|
+
reasoning: true;
|
|
11333
11841
|
input: ("image" | "text")[];
|
|
11334
11842
|
cost: {
|
|
11335
11843
|
input: number;
|
|
@@ -11340,7 +11848,7 @@ export declare const MODELS: {
|
|
|
11340
11848
|
contextWindow: number;
|
|
11341
11849
|
maxTokens: number;
|
|
11342
11850
|
};
|
|
11343
|
-
readonly "anthropic/claude-
|
|
11851
|
+
readonly "anthropic/claude-haiku-4.5": {
|
|
11344
11852
|
id: string;
|
|
11345
11853
|
name: string;
|
|
11346
11854
|
api: "anthropic-messages";
|
|
@@ -11357,7 +11865,7 @@ export declare const MODELS: {
|
|
|
11357
11865
|
contextWindow: number;
|
|
11358
11866
|
maxTokens: number;
|
|
11359
11867
|
};
|
|
11360
|
-
readonly "anthropic/claude-
|
|
11868
|
+
readonly "anthropic/claude-opus-4": {
|
|
11361
11869
|
id: string;
|
|
11362
11870
|
name: string;
|
|
11363
11871
|
api: "anthropic-messages";
|
|
@@ -11374,7 +11882,7 @@ export declare const MODELS: {
|
|
|
11374
11882
|
contextWindow: number;
|
|
11375
11883
|
maxTokens: number;
|
|
11376
11884
|
};
|
|
11377
|
-
readonly "anthropic/claude-opus-4": {
|
|
11885
|
+
readonly "anthropic/claude-opus-4.1": {
|
|
11378
11886
|
id: string;
|
|
11379
11887
|
name: string;
|
|
11380
11888
|
api: "anthropic-messages";
|
|
@@ -11391,7 +11899,7 @@ export declare const MODELS: {
|
|
|
11391
11899
|
contextWindow: number;
|
|
11392
11900
|
maxTokens: number;
|
|
11393
11901
|
};
|
|
11394
|
-
readonly "anthropic/claude-opus-4.
|
|
11902
|
+
readonly "anthropic/claude-opus-4.5": {
|
|
11395
11903
|
id: string;
|
|
11396
11904
|
name: string;
|
|
11397
11905
|
api: "anthropic-messages";
|
|
@@ -11408,7 +11916,7 @@ export declare const MODELS: {
|
|
|
11408
11916
|
contextWindow: number;
|
|
11409
11917
|
maxTokens: number;
|
|
11410
11918
|
};
|
|
11411
|
-
readonly "anthropic/claude-opus-4.
|
|
11919
|
+
readonly "anthropic/claude-opus-4.6": {
|
|
11412
11920
|
id: string;
|
|
11413
11921
|
name: string;
|
|
11414
11922
|
api: "anthropic-messages";
|
|
@@ -11425,7 +11933,7 @@ export declare const MODELS: {
|
|
|
11425
11933
|
contextWindow: number;
|
|
11426
11934
|
maxTokens: number;
|
|
11427
11935
|
};
|
|
11428
|
-
readonly "anthropic/claude-opus-4.
|
|
11936
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
11429
11937
|
id: string;
|
|
11430
11938
|
name: string;
|
|
11431
11939
|
api: "anthropic-messages";
|
|
@@ -11510,6 +12018,23 @@ export declare const MODELS: {
|
|
|
11510
12018
|
contextWindow: number;
|
|
11511
12019
|
maxTokens: number;
|
|
11512
12020
|
};
|
|
12021
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
12022
|
+
id: string;
|
|
12023
|
+
name: string;
|
|
12024
|
+
api: "anthropic-messages";
|
|
12025
|
+
provider: string;
|
|
12026
|
+
baseUrl: string;
|
|
12027
|
+
reasoning: true;
|
|
12028
|
+
input: "text"[];
|
|
12029
|
+
cost: {
|
|
12030
|
+
input: number;
|
|
12031
|
+
output: number;
|
|
12032
|
+
cacheRead: number;
|
|
12033
|
+
cacheWrite: number;
|
|
12034
|
+
};
|
|
12035
|
+
contextWindow: number;
|
|
12036
|
+
maxTokens: number;
|
|
12037
|
+
};
|
|
11513
12038
|
readonly "bytedance/seed-1.6": {
|
|
11514
12039
|
id: string;
|
|
11515
12040
|
name: string;
|
|
@@ -11799,6 +12324,40 @@ export declare const MODELS: {
|
|
|
11799
12324
|
contextWindow: number;
|
|
11800
12325
|
maxTokens: number;
|
|
11801
12326
|
};
|
|
12327
|
+
readonly "google/gemma-4-26b-a4b-it": {
|
|
12328
|
+
id: string;
|
|
12329
|
+
name: string;
|
|
12330
|
+
api: "anthropic-messages";
|
|
12331
|
+
provider: string;
|
|
12332
|
+
baseUrl: string;
|
|
12333
|
+
reasoning: false;
|
|
12334
|
+
input: ("image" | "text")[];
|
|
12335
|
+
cost: {
|
|
12336
|
+
input: number;
|
|
12337
|
+
output: number;
|
|
12338
|
+
cacheRead: number;
|
|
12339
|
+
cacheWrite: number;
|
|
12340
|
+
};
|
|
12341
|
+
contextWindow: number;
|
|
12342
|
+
maxTokens: number;
|
|
12343
|
+
};
|
|
12344
|
+
readonly "google/gemma-4-31b-it": {
|
|
12345
|
+
id: string;
|
|
12346
|
+
name: string;
|
|
12347
|
+
api: "anthropic-messages";
|
|
12348
|
+
provider: string;
|
|
12349
|
+
baseUrl: string;
|
|
12350
|
+
reasoning: false;
|
|
12351
|
+
input: ("image" | "text")[];
|
|
12352
|
+
cost: {
|
|
12353
|
+
input: number;
|
|
12354
|
+
output: number;
|
|
12355
|
+
cacheRead: number;
|
|
12356
|
+
cacheWrite: number;
|
|
12357
|
+
};
|
|
12358
|
+
contextWindow: number;
|
|
12359
|
+
maxTokens: number;
|
|
12360
|
+
};
|
|
11802
12361
|
readonly "inception/mercury-2": {
|
|
11803
12362
|
id: string;
|
|
11804
12363
|
name: string;
|
|
@@ -11833,13 +12392,13 @@ export declare const MODELS: {
|
|
|
11833
12392
|
contextWindow: number;
|
|
11834
12393
|
maxTokens: number;
|
|
11835
12394
|
};
|
|
11836
|
-
readonly "
|
|
12395
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
11837
12396
|
id: string;
|
|
11838
12397
|
name: string;
|
|
11839
12398
|
api: "anthropic-messages";
|
|
11840
12399
|
provider: string;
|
|
11841
12400
|
baseUrl: string;
|
|
11842
|
-
reasoning:
|
|
12401
|
+
reasoning: true;
|
|
11843
12402
|
input: "text"[];
|
|
11844
12403
|
cost: {
|
|
11845
12404
|
input: number;
|
|
@@ -11850,13 +12409,13 @@ export declare const MODELS: {
|
|
|
11850
12409
|
contextWindow: number;
|
|
11851
12410
|
maxTokens: number;
|
|
11852
12411
|
};
|
|
11853
|
-
readonly "meituan/longcat-flash-
|
|
12412
|
+
readonly "meituan/longcat-flash-chat": {
|
|
11854
12413
|
id: string;
|
|
11855
12414
|
name: string;
|
|
11856
12415
|
api: "anthropic-messages";
|
|
11857
12416
|
provider: string;
|
|
11858
12417
|
baseUrl: string;
|
|
11859
|
-
reasoning:
|
|
12418
|
+
reasoning: false;
|
|
11860
12419
|
input: "text"[];
|
|
11861
12420
|
cost: {
|
|
11862
12421
|
input: number;
|
|
@@ -12377,6 +12936,23 @@ export declare const MODELS: {
|
|
|
12377
12936
|
contextWindow: number;
|
|
12378
12937
|
maxTokens: number;
|
|
12379
12938
|
};
|
|
12939
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
12940
|
+
id: string;
|
|
12941
|
+
name: string;
|
|
12942
|
+
api: "anthropic-messages";
|
|
12943
|
+
provider: string;
|
|
12944
|
+
baseUrl: string;
|
|
12945
|
+
reasoning: true;
|
|
12946
|
+
input: ("image" | "text")[];
|
|
12947
|
+
cost: {
|
|
12948
|
+
input: number;
|
|
12949
|
+
output: number;
|
|
12950
|
+
cacheRead: number;
|
|
12951
|
+
cacheWrite: number;
|
|
12952
|
+
};
|
|
12953
|
+
contextWindow: number;
|
|
12954
|
+
maxTokens: number;
|
|
12955
|
+
};
|
|
12380
12956
|
readonly "nvidia/nemotron-nano-12b-v2-vl": {
|
|
12381
12957
|
id: string;
|
|
12382
12958
|
name: string;
|
|
@@ -13057,23 +13633,6 @@ export declare const MODELS: {
|
|
|
13057
13633
|
contextWindow: number;
|
|
13058
13634
|
maxTokens: number;
|
|
13059
13635
|
};
|
|
13060
|
-
readonly "xai/grok-2-vision": {
|
|
13061
|
-
id: string;
|
|
13062
|
-
name: string;
|
|
13063
|
-
api: "anthropic-messages";
|
|
13064
|
-
provider: string;
|
|
13065
|
-
baseUrl: string;
|
|
13066
|
-
reasoning: false;
|
|
13067
|
-
input: ("image" | "text")[];
|
|
13068
|
-
cost: {
|
|
13069
|
-
input: number;
|
|
13070
|
-
output: number;
|
|
13071
|
-
cacheRead: number;
|
|
13072
|
-
cacheWrite: number;
|
|
13073
|
-
};
|
|
13074
|
-
contextWindow: number;
|
|
13075
|
-
maxTokens: number;
|
|
13076
|
-
};
|
|
13077
13636
|
readonly "xai/grok-3": {
|
|
13078
13637
|
id: string;
|
|
13079
13638
|
name: string;
|
|
@@ -13166,7 +13725,7 @@ export declare const MODELS: {
|
|
|
13166
13725
|
provider: string;
|
|
13167
13726
|
baseUrl: string;
|
|
13168
13727
|
reasoning: false;
|
|
13169
|
-
input: "text"[];
|
|
13728
|
+
input: ("image" | "text")[];
|
|
13170
13729
|
cost: {
|
|
13171
13730
|
input: number;
|
|
13172
13731
|
output: number;
|
|
@@ -13183,7 +13742,7 @@ export declare const MODELS: {
|
|
|
13183
13742
|
provider: string;
|
|
13184
13743
|
baseUrl: string;
|
|
13185
13744
|
reasoning: true;
|
|
13186
|
-
input: "text"[];
|
|
13745
|
+
input: ("image" | "text")[];
|
|
13187
13746
|
cost: {
|
|
13188
13747
|
input: number;
|
|
13189
13748
|
output: number;
|
|
@@ -13200,7 +13759,7 @@ export declare const MODELS: {
|
|
|
13200
13759
|
provider: string;
|
|
13201
13760
|
baseUrl: string;
|
|
13202
13761
|
reasoning: false;
|
|
13203
|
-
input: "text"[];
|
|
13762
|
+
input: ("image" | "text")[];
|
|
13204
13763
|
cost: {
|
|
13205
13764
|
input: number;
|
|
13206
13765
|
output: number;
|
|
@@ -13217,7 +13776,7 @@ export declare const MODELS: {
|
|
|
13217
13776
|
provider: string;
|
|
13218
13777
|
baseUrl: string;
|
|
13219
13778
|
reasoning: true;
|
|
13220
|
-
input: "text"[];
|
|
13779
|
+
input: ("image" | "text")[];
|
|
13221
13780
|
cost: {
|
|
13222
13781
|
input: number;
|
|
13223
13782
|
output: number;
|
|
@@ -13234,7 +13793,7 @@ export declare const MODELS: {
|
|
|
13234
13793
|
provider: string;
|
|
13235
13794
|
baseUrl: string;
|
|
13236
13795
|
reasoning: true;
|
|
13237
|
-
input: "text"[];
|
|
13796
|
+
input: ("image" | "text")[];
|
|
13238
13797
|
cost: {
|
|
13239
13798
|
input: number;
|
|
13240
13799
|
output: number;
|
|
@@ -13251,7 +13810,7 @@ export declare const MODELS: {
|
|
|
13251
13810
|
provider: string;
|
|
13252
13811
|
baseUrl: string;
|
|
13253
13812
|
reasoning: true;
|
|
13254
|
-
input: "text"[];
|
|
13813
|
+
input: ("image" | "text")[];
|
|
13255
13814
|
cost: {
|
|
13256
13815
|
input: number;
|
|
13257
13816
|
output: number;
|
|
@@ -13567,6 +14126,40 @@ export declare const MODELS: {
|
|
|
13567
14126
|
contextWindow: number;
|
|
13568
14127
|
maxTokens: number;
|
|
13569
14128
|
};
|
|
14129
|
+
readonly "zai/glm-5.1": {
|
|
14130
|
+
id: string;
|
|
14131
|
+
name: string;
|
|
14132
|
+
api: "anthropic-messages";
|
|
14133
|
+
provider: string;
|
|
14134
|
+
baseUrl: string;
|
|
14135
|
+
reasoning: true;
|
|
14136
|
+
input: "text"[];
|
|
14137
|
+
cost: {
|
|
14138
|
+
input: number;
|
|
14139
|
+
output: number;
|
|
14140
|
+
cacheRead: number;
|
|
14141
|
+
cacheWrite: number;
|
|
14142
|
+
};
|
|
14143
|
+
contextWindow: number;
|
|
14144
|
+
maxTokens: number;
|
|
14145
|
+
};
|
|
14146
|
+
readonly "zai/glm-5v-turbo": {
|
|
14147
|
+
id: string;
|
|
14148
|
+
name: string;
|
|
14149
|
+
api: "anthropic-messages";
|
|
14150
|
+
provider: string;
|
|
14151
|
+
baseUrl: string;
|
|
14152
|
+
reasoning: true;
|
|
14153
|
+
input: ("image" | "text")[];
|
|
14154
|
+
cost: {
|
|
14155
|
+
input: number;
|
|
14156
|
+
output: number;
|
|
14157
|
+
cacheRead: number;
|
|
14158
|
+
cacheWrite: number;
|
|
14159
|
+
};
|
|
14160
|
+
contextWindow: number;
|
|
14161
|
+
maxTokens: number;
|
|
14162
|
+
};
|
|
13570
14163
|
};
|
|
13571
14164
|
readonly xai: {
|
|
13572
14165
|
readonly "grok-2": {
|
|
@@ -14072,6 +14665,7 @@ export declare const MODELS: {
|
|
|
14072
14665
|
compat: {
|
|
14073
14666
|
supportsDeveloperRole: false;
|
|
14074
14667
|
thinkingFormat: "zai";
|
|
14668
|
+
zaiToolStream: true;
|
|
14075
14669
|
};
|
|
14076
14670
|
reasoning: true;
|
|
14077
14671
|
input: "text"[];
|
|
@@ -14093,6 +14687,7 @@ export declare const MODELS: {
|
|
|
14093
14687
|
compat: {
|
|
14094
14688
|
supportsDeveloperRole: false;
|
|
14095
14689
|
thinkingFormat: "zai";
|
|
14690
|
+
zaiToolStream: true;
|
|
14096
14691
|
};
|
|
14097
14692
|
reasoning: true;
|
|
14098
14693
|
input: ("image" | "text")[];
|
|
@@ -14114,6 +14709,7 @@ export declare const MODELS: {
|
|
|
14114
14709
|
compat: {
|
|
14115
14710
|
supportsDeveloperRole: false;
|
|
14116
14711
|
thinkingFormat: "zai";
|
|
14712
|
+
zaiToolStream: true;
|
|
14117
14713
|
};
|
|
14118
14714
|
reasoning: true;
|
|
14119
14715
|
input: "text"[];
|
|
@@ -14135,6 +14731,29 @@ export declare const MODELS: {
|
|
|
14135
14731
|
compat: {
|
|
14136
14732
|
supportsDeveloperRole: false;
|
|
14137
14733
|
thinkingFormat: "zai";
|
|
14734
|
+
zaiToolStream: true;
|
|
14735
|
+
};
|
|
14736
|
+
reasoning: true;
|
|
14737
|
+
input: "text"[];
|
|
14738
|
+
cost: {
|
|
14739
|
+
input: number;
|
|
14740
|
+
output: number;
|
|
14741
|
+
cacheRead: number;
|
|
14742
|
+
cacheWrite: number;
|
|
14743
|
+
};
|
|
14744
|
+
contextWindow: number;
|
|
14745
|
+
maxTokens: number;
|
|
14746
|
+
};
|
|
14747
|
+
readonly "glm-4.7-flashx": {
|
|
14748
|
+
id: string;
|
|
14749
|
+
name: string;
|
|
14750
|
+
api: "openai-completions";
|
|
14751
|
+
provider: string;
|
|
14752
|
+
baseUrl: string;
|
|
14753
|
+
compat: {
|
|
14754
|
+
supportsDeveloperRole: false;
|
|
14755
|
+
thinkingFormat: "zai";
|
|
14756
|
+
zaiToolStream: true;
|
|
14138
14757
|
};
|
|
14139
14758
|
reasoning: true;
|
|
14140
14759
|
input: "text"[];
|
|
@@ -14156,6 +14775,7 @@ export declare const MODELS: {
|
|
|
14156
14775
|
compat: {
|
|
14157
14776
|
supportsDeveloperRole: false;
|
|
14158
14777
|
thinkingFormat: "zai";
|
|
14778
|
+
zaiToolStream: true;
|
|
14159
14779
|
};
|
|
14160
14780
|
reasoning: true;
|
|
14161
14781
|
input: "text"[];
|
|
@@ -14177,6 +14797,29 @@ export declare const MODELS: {
|
|
|
14177
14797
|
compat: {
|
|
14178
14798
|
supportsDeveloperRole: false;
|
|
14179
14799
|
thinkingFormat: "zai";
|
|
14800
|
+
zaiToolStream: true;
|
|
14801
|
+
};
|
|
14802
|
+
reasoning: true;
|
|
14803
|
+
input: "text"[];
|
|
14804
|
+
cost: {
|
|
14805
|
+
input: number;
|
|
14806
|
+
output: number;
|
|
14807
|
+
cacheRead: number;
|
|
14808
|
+
cacheWrite: number;
|
|
14809
|
+
};
|
|
14810
|
+
contextWindow: number;
|
|
14811
|
+
maxTokens: number;
|
|
14812
|
+
};
|
|
14813
|
+
readonly "glm-5.1": {
|
|
14814
|
+
id: string;
|
|
14815
|
+
name: string;
|
|
14816
|
+
api: "openai-completions";
|
|
14817
|
+
provider: string;
|
|
14818
|
+
baseUrl: string;
|
|
14819
|
+
compat: {
|
|
14820
|
+
supportsDeveloperRole: false;
|
|
14821
|
+
thinkingFormat: "zai";
|
|
14822
|
+
zaiToolStream: true;
|
|
14180
14823
|
};
|
|
14181
14824
|
reasoning: true;
|
|
14182
14825
|
input: "text"[];
|
|
@@ -14189,6 +14832,28 @@ export declare const MODELS: {
|
|
|
14189
14832
|
contextWindow: number;
|
|
14190
14833
|
maxTokens: number;
|
|
14191
14834
|
};
|
|
14835
|
+
readonly "glm-5v-turbo": {
|
|
14836
|
+
id: string;
|
|
14837
|
+
name: string;
|
|
14838
|
+
api: "openai-completions";
|
|
14839
|
+
provider: string;
|
|
14840
|
+
baseUrl: string;
|
|
14841
|
+
compat: {
|
|
14842
|
+
supportsDeveloperRole: false;
|
|
14843
|
+
thinkingFormat: "zai";
|
|
14844
|
+
zaiToolStream: true;
|
|
14845
|
+
};
|
|
14846
|
+
reasoning: true;
|
|
14847
|
+
input: ("image" | "text")[];
|
|
14848
|
+
cost: {
|
|
14849
|
+
input: number;
|
|
14850
|
+
output: number;
|
|
14851
|
+
cacheRead: number;
|
|
14852
|
+
cacheWrite: number;
|
|
14853
|
+
};
|
|
14854
|
+
contextWindow: number;
|
|
14855
|
+
maxTokens: number;
|
|
14856
|
+
};
|
|
14192
14857
|
};
|
|
14193
14858
|
};
|
|
14194
14859
|
//# sourceMappingURL=models.generated.d.ts.map
|