@draht/ai 2026.4.5 → 2026.4.25
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +17 -8
- package/dist/cli.d.ts.map +1 -1
- package/dist/cli.js +1 -1
- package/dist/cli.js.map +1 -1
- package/dist/models.d.ts +7 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +996 -519
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +1255 -765
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +20 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +46 -14
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +44 -19
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +2 -2
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +2 -2
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +1 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js +29 -4
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +2 -2
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +29 -16
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +8 -1
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +4 -4
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts +6 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +10 -1
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +63 -6
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/index.d.ts +1 -0
- package/dist/utils/oauth/index.d.ts.map +1 -1
- package/dist/utils/oauth/index.js +4 -0
- package/dist/utils/oauth/index.js.map +1 -1
- package/dist/utils/oauth/opencode-go.d.ts +23 -0
- package/dist/utils/oauth/opencode-go.d.ts.map +1 -0
- package/dist/utils/oauth/opencode-go.js +46 -0
- package/dist/utils/oauth/opencode-go.js.map +1 -0
- package/dist/utils/overflow.d.ts +3 -2
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +24 -11
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +5 -8
|
@@ -255,6 +255,23 @@ export declare const MODELS: {
|
|
|
255
255
|
contextWindow: number;
|
|
256
256
|
maxTokens: number;
|
|
257
257
|
};
|
|
258
|
+
readonly "anthropic.claude-opus-4-7": {
|
|
259
|
+
id: string;
|
|
260
|
+
name: string;
|
|
261
|
+
api: "bedrock-converse-stream";
|
|
262
|
+
provider: string;
|
|
263
|
+
baseUrl: string;
|
|
264
|
+
reasoning: true;
|
|
265
|
+
input: ("image" | "text")[];
|
|
266
|
+
cost: {
|
|
267
|
+
input: number;
|
|
268
|
+
output: number;
|
|
269
|
+
cacheRead: number;
|
|
270
|
+
cacheWrite: number;
|
|
271
|
+
};
|
|
272
|
+
contextWindow: number;
|
|
273
|
+
maxTokens: number;
|
|
274
|
+
};
|
|
258
275
|
readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
259
276
|
id: string;
|
|
260
277
|
name: string;
|
|
@@ -306,6 +323,40 @@ export declare const MODELS: {
|
|
|
306
323
|
contextWindow: number;
|
|
307
324
|
maxTokens: number;
|
|
308
325
|
};
|
|
326
|
+
readonly "au.anthropic.claude-opus-4-6-v1": {
|
|
327
|
+
id: string;
|
|
328
|
+
name: string;
|
|
329
|
+
api: "bedrock-converse-stream";
|
|
330
|
+
provider: string;
|
|
331
|
+
baseUrl: string;
|
|
332
|
+
reasoning: true;
|
|
333
|
+
input: ("image" | "text")[];
|
|
334
|
+
cost: {
|
|
335
|
+
input: number;
|
|
336
|
+
output: number;
|
|
337
|
+
cacheRead: number;
|
|
338
|
+
cacheWrite: number;
|
|
339
|
+
};
|
|
340
|
+
contextWindow: number;
|
|
341
|
+
maxTokens: number;
|
|
342
|
+
};
|
|
343
|
+
readonly "au.anthropic.claude-sonnet-4-6": {
|
|
344
|
+
id: string;
|
|
345
|
+
name: string;
|
|
346
|
+
api: "bedrock-converse-stream";
|
|
347
|
+
provider: string;
|
|
348
|
+
baseUrl: string;
|
|
349
|
+
reasoning: true;
|
|
350
|
+
input: ("image" | "text")[];
|
|
351
|
+
cost: {
|
|
352
|
+
input: number;
|
|
353
|
+
output: number;
|
|
354
|
+
cacheRead: number;
|
|
355
|
+
cacheWrite: number;
|
|
356
|
+
};
|
|
357
|
+
contextWindow: number;
|
|
358
|
+
maxTokens: number;
|
|
359
|
+
};
|
|
309
360
|
readonly "deepseek.r1-v1:0": {
|
|
310
361
|
id: string;
|
|
311
362
|
name: string;
|
|
@@ -408,6 +459,23 @@ export declare const MODELS: {
|
|
|
408
459
|
contextWindow: number;
|
|
409
460
|
maxTokens: number;
|
|
410
461
|
};
|
|
462
|
+
readonly "eu.anthropic.claude-opus-4-7": {
|
|
463
|
+
id: string;
|
|
464
|
+
name: string;
|
|
465
|
+
api: "bedrock-converse-stream";
|
|
466
|
+
provider: string;
|
|
467
|
+
baseUrl: string;
|
|
468
|
+
reasoning: true;
|
|
469
|
+
input: ("image" | "text")[];
|
|
470
|
+
cost: {
|
|
471
|
+
input: number;
|
|
472
|
+
output: number;
|
|
473
|
+
cacheRead: number;
|
|
474
|
+
cacheWrite: number;
|
|
475
|
+
};
|
|
476
|
+
contextWindow: number;
|
|
477
|
+
maxTokens: number;
|
|
478
|
+
};
|
|
411
479
|
readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
412
480
|
id: string;
|
|
413
481
|
name: string;
|
|
@@ -510,6 +578,23 @@ export declare const MODELS: {
|
|
|
510
578
|
contextWindow: number;
|
|
511
579
|
maxTokens: number;
|
|
512
580
|
};
|
|
581
|
+
readonly "global.anthropic.claude-opus-4-7": {
|
|
582
|
+
id: string;
|
|
583
|
+
name: string;
|
|
584
|
+
api: "bedrock-converse-stream";
|
|
585
|
+
provider: string;
|
|
586
|
+
baseUrl: string;
|
|
587
|
+
reasoning: true;
|
|
588
|
+
input: ("image" | "text")[];
|
|
589
|
+
cost: {
|
|
590
|
+
input: number;
|
|
591
|
+
output: number;
|
|
592
|
+
cacheRead: number;
|
|
593
|
+
cacheWrite: number;
|
|
594
|
+
};
|
|
595
|
+
contextWindow: number;
|
|
596
|
+
maxTokens: number;
|
|
597
|
+
};
|
|
513
598
|
readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
514
599
|
id: string;
|
|
515
600
|
name: string;
|
|
@@ -1207,6 +1292,23 @@ export declare const MODELS: {
|
|
|
1207
1292
|
contextWindow: number;
|
|
1208
1293
|
maxTokens: number;
|
|
1209
1294
|
};
|
|
1295
|
+
readonly "qwen.qwen3-coder-next": {
|
|
1296
|
+
id: string;
|
|
1297
|
+
name: string;
|
|
1298
|
+
api: "bedrock-converse-stream";
|
|
1299
|
+
provider: string;
|
|
1300
|
+
baseUrl: string;
|
|
1301
|
+
reasoning: true;
|
|
1302
|
+
input: "text"[];
|
|
1303
|
+
cost: {
|
|
1304
|
+
input: number;
|
|
1305
|
+
output: number;
|
|
1306
|
+
cacheRead: number;
|
|
1307
|
+
cacheWrite: number;
|
|
1308
|
+
};
|
|
1309
|
+
contextWindow: number;
|
|
1310
|
+
maxTokens: number;
|
|
1311
|
+
};
|
|
1210
1312
|
readonly "qwen.qwen3-next-80b-a3b": {
|
|
1211
1313
|
id: string;
|
|
1212
1314
|
name: string;
|
|
@@ -1326,6 +1428,23 @@ export declare const MODELS: {
|
|
|
1326
1428
|
contextWindow: number;
|
|
1327
1429
|
maxTokens: number;
|
|
1328
1430
|
};
|
|
1431
|
+
readonly "us.anthropic.claude-opus-4-7": {
|
|
1432
|
+
id: string;
|
|
1433
|
+
name: string;
|
|
1434
|
+
api: "bedrock-converse-stream";
|
|
1435
|
+
provider: string;
|
|
1436
|
+
baseUrl: string;
|
|
1437
|
+
reasoning: true;
|
|
1438
|
+
input: ("image" | "text")[];
|
|
1439
|
+
cost: {
|
|
1440
|
+
input: number;
|
|
1441
|
+
output: number;
|
|
1442
|
+
cacheRead: number;
|
|
1443
|
+
cacheWrite: number;
|
|
1444
|
+
};
|
|
1445
|
+
contextWindow: number;
|
|
1446
|
+
maxTokens: number;
|
|
1447
|
+
};
|
|
1329
1448
|
readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
1330
1449
|
id: string;
|
|
1331
1450
|
name: string;
|
|
@@ -1753,7 +1872,7 @@ export declare const MODELS: {
|
|
|
1753
1872
|
contextWindow: number;
|
|
1754
1873
|
maxTokens: number;
|
|
1755
1874
|
};
|
|
1756
|
-
readonly "claude-
|
|
1875
|
+
readonly "claude-opus-4-7": {
|
|
1757
1876
|
id: string;
|
|
1758
1877
|
name: string;
|
|
1759
1878
|
api: "anthropic-messages";
|
|
@@ -1770,7 +1889,7 @@ export declare const MODELS: {
|
|
|
1770
1889
|
contextWindow: number;
|
|
1771
1890
|
maxTokens: number;
|
|
1772
1891
|
};
|
|
1773
|
-
readonly "claude-sonnet-4-
|
|
1892
|
+
readonly "claude-sonnet-4-0": {
|
|
1774
1893
|
id: string;
|
|
1775
1894
|
name: string;
|
|
1776
1895
|
api: "anthropic-messages";
|
|
@@ -1787,7 +1906,7 @@ export declare const MODELS: {
|
|
|
1787
1906
|
contextWindow: number;
|
|
1788
1907
|
maxTokens: number;
|
|
1789
1908
|
};
|
|
1790
|
-
readonly "claude-sonnet-4-
|
|
1909
|
+
readonly "claude-sonnet-4-20250514": {
|
|
1791
1910
|
id: string;
|
|
1792
1911
|
name: string;
|
|
1793
1912
|
api: "anthropic-messages";
|
|
@@ -1804,7 +1923,7 @@ export declare const MODELS: {
|
|
|
1804
1923
|
contextWindow: number;
|
|
1805
1924
|
maxTokens: number;
|
|
1806
1925
|
};
|
|
1807
|
-
readonly "claude-sonnet-4-5
|
|
1926
|
+
readonly "claude-sonnet-4-5": {
|
|
1808
1927
|
id: string;
|
|
1809
1928
|
name: string;
|
|
1810
1929
|
api: "anthropic-messages";
|
|
@@ -1821,7 +1940,7 @@ export declare const MODELS: {
|
|
|
1821
1940
|
contextWindow: number;
|
|
1822
1941
|
maxTokens: number;
|
|
1823
1942
|
};
|
|
1824
|
-
readonly "claude-sonnet-4-
|
|
1943
|
+
readonly "claude-sonnet-4-5-20250929": {
|
|
1825
1944
|
id: string;
|
|
1826
1945
|
name: string;
|
|
1827
1946
|
api: "anthropic-messages";
|
|
@@ -1838,16 +1957,14 @@ export declare const MODELS: {
|
|
|
1838
1957
|
contextWindow: number;
|
|
1839
1958
|
maxTokens: number;
|
|
1840
1959
|
};
|
|
1841
|
-
|
|
1842
|
-
readonly "azure-openai-responses": {
|
|
1843
|
-
readonly "codex-mini-latest": {
|
|
1960
|
+
readonly "claude-sonnet-4-6": {
|
|
1844
1961
|
id: string;
|
|
1845
1962
|
name: string;
|
|
1846
|
-
api: "
|
|
1963
|
+
api: "anthropic-messages";
|
|
1847
1964
|
provider: string;
|
|
1848
1965
|
baseUrl: string;
|
|
1849
1966
|
reasoning: true;
|
|
1850
|
-
input: "text"[];
|
|
1967
|
+
input: ("image" | "text")[];
|
|
1851
1968
|
cost: {
|
|
1852
1969
|
input: number;
|
|
1853
1970
|
output: number;
|
|
@@ -1857,6 +1974,8 @@ export declare const MODELS: {
|
|
|
1857
1974
|
contextWindow: number;
|
|
1858
1975
|
maxTokens: number;
|
|
1859
1976
|
};
|
|
1977
|
+
};
|
|
1978
|
+
readonly "azure-openai-responses": {
|
|
1860
1979
|
readonly "gpt-4": {
|
|
1861
1980
|
id: string;
|
|
1862
1981
|
name: string;
|
|
@@ -2401,6 +2520,23 @@ export declare const MODELS: {
|
|
|
2401
2520
|
contextWindow: number;
|
|
2402
2521
|
maxTokens: number;
|
|
2403
2522
|
};
|
|
2523
|
+
readonly "gpt-5.5": {
|
|
2524
|
+
id: string;
|
|
2525
|
+
name: string;
|
|
2526
|
+
api: "azure-openai-responses";
|
|
2527
|
+
provider: string;
|
|
2528
|
+
baseUrl: string;
|
|
2529
|
+
reasoning: true;
|
|
2530
|
+
input: ("image" | "text")[];
|
|
2531
|
+
cost: {
|
|
2532
|
+
input: number;
|
|
2533
|
+
output: number;
|
|
2534
|
+
cacheRead: number;
|
|
2535
|
+
cacheWrite: number;
|
|
2536
|
+
};
|
|
2537
|
+
contextWindow: number;
|
|
2538
|
+
maxTokens: number;
|
|
2539
|
+
};
|
|
2404
2540
|
readonly o1: {
|
|
2405
2541
|
id: string;
|
|
2406
2542
|
name: string;
|
|
@@ -2678,6 +2814,29 @@ export declare const MODELS: {
|
|
|
2678
2814
|
contextWindow: number;
|
|
2679
2815
|
maxTokens: number;
|
|
2680
2816
|
};
|
|
2817
|
+
readonly "claude-opus-4.7": {
|
|
2818
|
+
id: string;
|
|
2819
|
+
name: string;
|
|
2820
|
+
api: "anthropic-messages";
|
|
2821
|
+
provider: string;
|
|
2822
|
+
baseUrl: string;
|
|
2823
|
+
headers: {
|
|
2824
|
+
"User-Agent": string;
|
|
2825
|
+
"Editor-Version": string;
|
|
2826
|
+
"Editor-Plugin-Version": string;
|
|
2827
|
+
"Copilot-Integration-Id": string;
|
|
2828
|
+
};
|
|
2829
|
+
reasoning: true;
|
|
2830
|
+
input: ("image" | "text")[];
|
|
2831
|
+
cost: {
|
|
2832
|
+
input: number;
|
|
2833
|
+
output: number;
|
|
2834
|
+
cacheRead: number;
|
|
2835
|
+
cacheWrite: number;
|
|
2836
|
+
};
|
|
2837
|
+
contextWindow: number;
|
|
2838
|
+
maxTokens: number;
|
|
2839
|
+
};
|
|
2681
2840
|
readonly "claude-sonnet-4": {
|
|
2682
2841
|
id: string;
|
|
2683
2842
|
name: string;
|
|
@@ -3168,6 +3327,29 @@ export declare const MODELS: {
|
|
|
3168
3327
|
contextWindow: number;
|
|
3169
3328
|
maxTokens: number;
|
|
3170
3329
|
};
|
|
3330
|
+
readonly "gpt-5.5": {
|
|
3331
|
+
id: string;
|
|
3332
|
+
name: string;
|
|
3333
|
+
api: "openai-responses";
|
|
3334
|
+
provider: string;
|
|
3335
|
+
baseUrl: string;
|
|
3336
|
+
headers: {
|
|
3337
|
+
"User-Agent": string;
|
|
3338
|
+
"Editor-Version": string;
|
|
3339
|
+
"Editor-Plugin-Version": string;
|
|
3340
|
+
"Copilot-Integration-Id": string;
|
|
3341
|
+
};
|
|
3342
|
+
reasoning: true;
|
|
3343
|
+
input: ("image" | "text")[];
|
|
3344
|
+
cost: {
|
|
3345
|
+
input: number;
|
|
3346
|
+
output: number;
|
|
3347
|
+
cacheRead: number;
|
|
3348
|
+
cacheWrite: number;
|
|
3349
|
+
};
|
|
3350
|
+
contextWindow: number;
|
|
3351
|
+
maxTokens: number;
|
|
3352
|
+
};
|
|
3171
3353
|
readonly "grok-code-fast-1": {
|
|
3172
3354
|
id: string;
|
|
3173
3355
|
name: string;
|
|
@@ -3623,7 +3805,7 @@ export declare const MODELS: {
|
|
|
3623
3805
|
contextWindow: number;
|
|
3624
3806
|
maxTokens: number;
|
|
3625
3807
|
};
|
|
3626
|
-
readonly "gemma-4-26b": {
|
|
3808
|
+
readonly "gemma-4-26b-a4b-it": {
|
|
3627
3809
|
id: string;
|
|
3628
3810
|
name: string;
|
|
3629
3811
|
api: "google-generative-ai";
|
|
@@ -3640,7 +3822,7 @@ export declare const MODELS: {
|
|
|
3640
3822
|
contextWindow: number;
|
|
3641
3823
|
maxTokens: number;
|
|
3642
3824
|
};
|
|
3643
|
-
readonly "gemma-4-31b": {
|
|
3825
|
+
readonly "gemma-4-31b-it": {
|
|
3644
3826
|
id: string;
|
|
3645
3827
|
name: string;
|
|
3646
3828
|
api: "google-generative-ai";
|
|
@@ -4122,6 +4304,23 @@ export declare const MODELS: {
|
|
|
4122
4304
|
contextWindow: number;
|
|
4123
4305
|
maxTokens: number;
|
|
4124
4306
|
};
|
|
4307
|
+
readonly "gemini-3.1-pro-preview-customtools": {
|
|
4308
|
+
id: string;
|
|
4309
|
+
name: string;
|
|
4310
|
+
api: "google-vertex";
|
|
4311
|
+
provider: string;
|
|
4312
|
+
baseUrl: string;
|
|
4313
|
+
reasoning: true;
|
|
4314
|
+
input: ("image" | "text")[];
|
|
4315
|
+
cost: {
|
|
4316
|
+
input: number;
|
|
4317
|
+
output: number;
|
|
4318
|
+
cacheRead: number;
|
|
4319
|
+
cacheWrite: number;
|
|
4320
|
+
};
|
|
4321
|
+
contextWindow: number;
|
|
4322
|
+
maxTokens: number;
|
|
4323
|
+
};
|
|
4125
4324
|
};
|
|
4126
4325
|
readonly groq: {
|
|
4127
4326
|
readonly "deepseek-r1-distill-llama-70b": {
|
|
@@ -4472,6 +4671,26 @@ export declare const MODELS: {
|
|
|
4472
4671
|
contextWindow: number;
|
|
4473
4672
|
maxTokens: number;
|
|
4474
4673
|
};
|
|
4674
|
+
readonly "MiniMaxAI/MiniMax-M2.7": {
|
|
4675
|
+
id: string;
|
|
4676
|
+
name: string;
|
|
4677
|
+
api: "openai-completions";
|
|
4678
|
+
provider: string;
|
|
4679
|
+
baseUrl: string;
|
|
4680
|
+
compat: {
|
|
4681
|
+
supportsDeveloperRole: false;
|
|
4682
|
+
};
|
|
4683
|
+
reasoning: true;
|
|
4684
|
+
input: "text"[];
|
|
4685
|
+
cost: {
|
|
4686
|
+
input: number;
|
|
4687
|
+
output: number;
|
|
4688
|
+
cacheRead: number;
|
|
4689
|
+
cacheWrite: number;
|
|
4690
|
+
};
|
|
4691
|
+
contextWindow: number;
|
|
4692
|
+
maxTokens: number;
|
|
4693
|
+
};
|
|
4475
4694
|
readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4476
4695
|
id: string;
|
|
4477
4696
|
name: string;
|
|
@@ -4732,7 +4951,7 @@ export declare const MODELS: {
|
|
|
4732
4951
|
contextWindow: number;
|
|
4733
4952
|
maxTokens: number;
|
|
4734
4953
|
};
|
|
4735
|
-
readonly "
|
|
4954
|
+
readonly "moonshotai/Kimi-K2.6": {
|
|
4736
4955
|
id: string;
|
|
4737
4956
|
name: string;
|
|
4738
4957
|
api: "openai-completions";
|
|
@@ -4742,7 +4961,7 @@ export declare const MODELS: {
|
|
|
4742
4961
|
supportsDeveloperRole: false;
|
|
4743
4962
|
};
|
|
4744
4963
|
reasoning: true;
|
|
4745
|
-
input: "text"[];
|
|
4964
|
+
input: ("image" | "text")[];
|
|
4746
4965
|
cost: {
|
|
4747
4966
|
input: number;
|
|
4748
4967
|
output: number;
|
|
@@ -4752,7 +4971,7 @@ export declare const MODELS: {
|
|
|
4752
4971
|
contextWindow: number;
|
|
4753
4972
|
maxTokens: number;
|
|
4754
4973
|
};
|
|
4755
|
-
readonly "zai-org/GLM-4.7
|
|
4974
|
+
readonly "zai-org/GLM-4.7": {
|
|
4756
4975
|
id: string;
|
|
4757
4976
|
name: string;
|
|
4758
4977
|
api: "openai-completions";
|
|
@@ -4772,7 +4991,7 @@ export declare const MODELS: {
|
|
|
4772
4991
|
contextWindow: number;
|
|
4773
4992
|
maxTokens: number;
|
|
4774
4993
|
};
|
|
4775
|
-
readonly "zai-org/GLM-
|
|
4994
|
+
readonly "zai-org/GLM-4.7-Flash": {
|
|
4776
4995
|
id: string;
|
|
4777
4996
|
name: string;
|
|
4778
4997
|
api: "openai-completions";
|
|
@@ -4792,16 +5011,17 @@ export declare const MODELS: {
|
|
|
4792
5011
|
contextWindow: number;
|
|
4793
5012
|
maxTokens: number;
|
|
4794
5013
|
};
|
|
4795
|
-
|
|
4796
|
-
readonly "kimi-coding": {
|
|
4797
|
-
readonly k2p5: {
|
|
5014
|
+
readonly "zai-org/GLM-5": {
|
|
4798
5015
|
id: string;
|
|
4799
5016
|
name: string;
|
|
4800
|
-
api: "
|
|
5017
|
+
api: "openai-completions";
|
|
4801
5018
|
provider: string;
|
|
4802
5019
|
baseUrl: string;
|
|
5020
|
+
compat: {
|
|
5021
|
+
supportsDeveloperRole: false;
|
|
5022
|
+
};
|
|
4803
5023
|
reasoning: true;
|
|
4804
|
-
input:
|
|
5024
|
+
input: "text"[];
|
|
4805
5025
|
cost: {
|
|
4806
5026
|
input: number;
|
|
4807
5027
|
output: number;
|
|
@@ -4811,12 +5031,15 @@ export declare const MODELS: {
|
|
|
4811
5031
|
contextWindow: number;
|
|
4812
5032
|
maxTokens: number;
|
|
4813
5033
|
};
|
|
4814
|
-
readonly "
|
|
5034
|
+
readonly "zai-org/GLM-5.1": {
|
|
4815
5035
|
id: string;
|
|
4816
5036
|
name: string;
|
|
4817
|
-
api: "
|
|
5037
|
+
api: "openai-completions";
|
|
4818
5038
|
provider: string;
|
|
4819
5039
|
baseUrl: string;
|
|
5040
|
+
compat: {
|
|
5041
|
+
supportsDeveloperRole: false;
|
|
5042
|
+
};
|
|
4820
5043
|
reasoning: true;
|
|
4821
5044
|
input: "text"[];
|
|
4822
5045
|
cost: {
|
|
@@ -4829,15 +5052,15 @@ export declare const MODELS: {
|
|
|
4829
5052
|
maxTokens: number;
|
|
4830
5053
|
};
|
|
4831
5054
|
};
|
|
4832
|
-
readonly
|
|
4833
|
-
readonly
|
|
5055
|
+
readonly "kimi-coding": {
|
|
5056
|
+
readonly k2p5: {
|
|
4834
5057
|
id: string;
|
|
4835
5058
|
name: string;
|
|
4836
5059
|
api: "anthropic-messages";
|
|
4837
5060
|
provider: string;
|
|
4838
5061
|
baseUrl: string;
|
|
4839
5062
|
reasoning: true;
|
|
4840
|
-
input: "text"[];
|
|
5063
|
+
input: ("image" | "text")[];
|
|
4841
5064
|
cost: {
|
|
4842
5065
|
input: number;
|
|
4843
5066
|
output: number;
|
|
@@ -4847,48 +5070,14 @@ export declare const MODELS: {
|
|
|
4847
5070
|
contextWindow: number;
|
|
4848
5071
|
maxTokens: number;
|
|
4849
5072
|
};
|
|
4850
|
-
readonly
|
|
5073
|
+
readonly k2p6: {
|
|
4851
5074
|
id: string;
|
|
4852
5075
|
name: string;
|
|
4853
5076
|
api: "anthropic-messages";
|
|
4854
5077
|
provider: string;
|
|
4855
5078
|
baseUrl: string;
|
|
4856
5079
|
reasoning: true;
|
|
4857
|
-
input: "text"[];
|
|
4858
|
-
cost: {
|
|
4859
|
-
input: number;
|
|
4860
|
-
output: number;
|
|
4861
|
-
cacheRead: number;
|
|
4862
|
-
cacheWrite: number;
|
|
4863
|
-
};
|
|
4864
|
-
contextWindow: number;
|
|
4865
|
-
maxTokens: number;
|
|
4866
|
-
};
|
|
4867
|
-
readonly "MiniMax-M2.1-highspeed": {
|
|
4868
|
-
id: string;
|
|
4869
|
-
name: string;
|
|
4870
|
-
api: "anthropic-messages";
|
|
4871
|
-
provider: string;
|
|
4872
|
-
baseUrl: string;
|
|
4873
|
-
reasoning: true;
|
|
4874
|
-
input: "text"[];
|
|
4875
|
-
cost: {
|
|
4876
|
-
input: number;
|
|
4877
|
-
output: number;
|
|
4878
|
-
cacheRead: number;
|
|
4879
|
-
cacheWrite: number;
|
|
4880
|
-
};
|
|
4881
|
-
contextWindow: number;
|
|
4882
|
-
maxTokens: number;
|
|
4883
|
-
};
|
|
4884
|
-
readonly "MiniMax-M2.5": {
|
|
4885
|
-
id: string;
|
|
4886
|
-
name: string;
|
|
4887
|
-
api: "anthropic-messages";
|
|
4888
|
-
provider: string;
|
|
4889
|
-
baseUrl: string;
|
|
4890
|
-
reasoning: true;
|
|
4891
|
-
input: "text"[];
|
|
5080
|
+
input: ("image" | "text")[];
|
|
4892
5081
|
cost: {
|
|
4893
5082
|
input: number;
|
|
4894
5083
|
output: number;
|
|
@@ -4898,7 +5087,7 @@ export declare const MODELS: {
|
|
|
4898
5087
|
contextWindow: number;
|
|
4899
5088
|
maxTokens: number;
|
|
4900
5089
|
};
|
|
4901
|
-
readonly "
|
|
5090
|
+
readonly "kimi-k2-thinking": {
|
|
4902
5091
|
id: string;
|
|
4903
5092
|
name: string;
|
|
4904
5093
|
api: "anthropic-messages";
|
|
@@ -4915,6 +5104,8 @@ export declare const MODELS: {
|
|
|
4915
5104
|
contextWindow: number;
|
|
4916
5105
|
maxTokens: number;
|
|
4917
5106
|
};
|
|
5107
|
+
};
|
|
5108
|
+
readonly minimax: {
|
|
4918
5109
|
readonly "MiniMax-M2.7": {
|
|
4919
5110
|
id: string;
|
|
4920
5111
|
name: string;
|
|
@@ -4951,91 +5142,6 @@ export declare const MODELS: {
|
|
|
4951
5142
|
};
|
|
4952
5143
|
};
|
|
4953
5144
|
readonly "minimax-cn": {
|
|
4954
|
-
readonly "MiniMax-M2": {
|
|
4955
|
-
id: string;
|
|
4956
|
-
name: string;
|
|
4957
|
-
api: "anthropic-messages";
|
|
4958
|
-
provider: string;
|
|
4959
|
-
baseUrl: string;
|
|
4960
|
-
reasoning: true;
|
|
4961
|
-
input: "text"[];
|
|
4962
|
-
cost: {
|
|
4963
|
-
input: number;
|
|
4964
|
-
output: number;
|
|
4965
|
-
cacheRead: number;
|
|
4966
|
-
cacheWrite: number;
|
|
4967
|
-
};
|
|
4968
|
-
contextWindow: number;
|
|
4969
|
-
maxTokens: number;
|
|
4970
|
-
};
|
|
4971
|
-
readonly "MiniMax-M2.1": {
|
|
4972
|
-
id: string;
|
|
4973
|
-
name: string;
|
|
4974
|
-
api: "anthropic-messages";
|
|
4975
|
-
provider: string;
|
|
4976
|
-
baseUrl: string;
|
|
4977
|
-
reasoning: true;
|
|
4978
|
-
input: "text"[];
|
|
4979
|
-
cost: {
|
|
4980
|
-
input: number;
|
|
4981
|
-
output: number;
|
|
4982
|
-
cacheRead: number;
|
|
4983
|
-
cacheWrite: number;
|
|
4984
|
-
};
|
|
4985
|
-
contextWindow: number;
|
|
4986
|
-
maxTokens: number;
|
|
4987
|
-
};
|
|
4988
|
-
readonly "MiniMax-M2.1-highspeed": {
|
|
4989
|
-
id: string;
|
|
4990
|
-
name: string;
|
|
4991
|
-
api: "anthropic-messages";
|
|
4992
|
-
provider: string;
|
|
4993
|
-
baseUrl: string;
|
|
4994
|
-
reasoning: true;
|
|
4995
|
-
input: "text"[];
|
|
4996
|
-
cost: {
|
|
4997
|
-
input: number;
|
|
4998
|
-
output: number;
|
|
4999
|
-
cacheRead: number;
|
|
5000
|
-
cacheWrite: number;
|
|
5001
|
-
};
|
|
5002
|
-
contextWindow: number;
|
|
5003
|
-
maxTokens: number;
|
|
5004
|
-
};
|
|
5005
|
-
readonly "MiniMax-M2.5": {
|
|
5006
|
-
id: string;
|
|
5007
|
-
name: string;
|
|
5008
|
-
api: "anthropic-messages";
|
|
5009
|
-
provider: string;
|
|
5010
|
-
baseUrl: string;
|
|
5011
|
-
reasoning: true;
|
|
5012
|
-
input: "text"[];
|
|
5013
|
-
cost: {
|
|
5014
|
-
input: number;
|
|
5015
|
-
output: number;
|
|
5016
|
-
cacheRead: number;
|
|
5017
|
-
cacheWrite: number;
|
|
5018
|
-
};
|
|
5019
|
-
contextWindow: number;
|
|
5020
|
-
maxTokens: number;
|
|
5021
|
-
};
|
|
5022
|
-
readonly "MiniMax-M2.5-highspeed": {
|
|
5023
|
-
id: string;
|
|
5024
|
-
name: string;
|
|
5025
|
-
api: "anthropic-messages";
|
|
5026
|
-
provider: string;
|
|
5027
|
-
baseUrl: string;
|
|
5028
|
-
reasoning: true;
|
|
5029
|
-
input: "text"[];
|
|
5030
|
-
cost: {
|
|
5031
|
-
input: number;
|
|
5032
|
-
output: number;
|
|
5033
|
-
cacheRead: number;
|
|
5034
|
-
cacheWrite: number;
|
|
5035
|
-
};
|
|
5036
|
-
contextWindow: number;
|
|
5037
|
-
maxTokens: number;
|
|
5038
|
-
};
|
|
5039
5145
|
readonly "MiniMax-M2.7": {
|
|
5040
5146
|
id: string;
|
|
5041
5147
|
name: string;
|
|
@@ -5075,7 +5181,7 @@ export declare const MODELS: {
|
|
|
5075
5181
|
readonly "codestral-latest": {
|
|
5076
5182
|
id: string;
|
|
5077
5183
|
name: string;
|
|
5078
|
-
api: "
|
|
5184
|
+
api: "openai-completions";
|
|
5079
5185
|
provider: string;
|
|
5080
5186
|
baseUrl: string;
|
|
5081
5187
|
reasoning: false;
|
|
@@ -5092,7 +5198,7 @@ export declare const MODELS: {
|
|
|
5092
5198
|
readonly "devstral-2512": {
|
|
5093
5199
|
id: string;
|
|
5094
5200
|
name: string;
|
|
5095
|
-
api: "
|
|
5201
|
+
api: "openai-completions";
|
|
5096
5202
|
provider: string;
|
|
5097
5203
|
baseUrl: string;
|
|
5098
5204
|
reasoning: false;
|
|
@@ -5109,7 +5215,7 @@ export declare const MODELS: {
|
|
|
5109
5215
|
readonly "devstral-medium-2507": {
|
|
5110
5216
|
id: string;
|
|
5111
5217
|
name: string;
|
|
5112
|
-
api: "
|
|
5218
|
+
api: "openai-completions";
|
|
5113
5219
|
provider: string;
|
|
5114
5220
|
baseUrl: string;
|
|
5115
5221
|
reasoning: false;
|
|
@@ -5126,7 +5232,7 @@ export declare const MODELS: {
|
|
|
5126
5232
|
readonly "devstral-medium-latest": {
|
|
5127
5233
|
id: string;
|
|
5128
5234
|
name: string;
|
|
5129
|
-
api: "
|
|
5235
|
+
api: "openai-completions";
|
|
5130
5236
|
provider: string;
|
|
5131
5237
|
baseUrl: string;
|
|
5132
5238
|
reasoning: false;
|
|
@@ -5143,7 +5249,7 @@ export declare const MODELS: {
|
|
|
5143
5249
|
readonly "devstral-small-2505": {
|
|
5144
5250
|
id: string;
|
|
5145
5251
|
name: string;
|
|
5146
|
-
api: "
|
|
5252
|
+
api: "openai-completions";
|
|
5147
5253
|
provider: string;
|
|
5148
5254
|
baseUrl: string;
|
|
5149
5255
|
reasoning: false;
|
|
@@ -5160,7 +5266,7 @@ export declare const MODELS: {
|
|
|
5160
5266
|
readonly "devstral-small-2507": {
|
|
5161
5267
|
id: string;
|
|
5162
5268
|
name: string;
|
|
5163
|
-
api: "
|
|
5269
|
+
api: "openai-completions";
|
|
5164
5270
|
provider: string;
|
|
5165
5271
|
baseUrl: string;
|
|
5166
5272
|
reasoning: false;
|
|
@@ -5177,7 +5283,7 @@ export declare const MODELS: {
|
|
|
5177
5283
|
readonly "labs-devstral-small-2512": {
|
|
5178
5284
|
id: string;
|
|
5179
5285
|
name: string;
|
|
5180
|
-
api: "
|
|
5286
|
+
api: "openai-completions";
|
|
5181
5287
|
provider: string;
|
|
5182
5288
|
baseUrl: string;
|
|
5183
5289
|
reasoning: false;
|
|
@@ -5194,7 +5300,7 @@ export declare const MODELS: {
|
|
|
5194
5300
|
readonly "magistral-medium-latest": {
|
|
5195
5301
|
id: string;
|
|
5196
5302
|
name: string;
|
|
5197
|
-
api: "
|
|
5303
|
+
api: "openai-completions";
|
|
5198
5304
|
provider: string;
|
|
5199
5305
|
baseUrl: string;
|
|
5200
5306
|
reasoning: true;
|
|
@@ -5211,7 +5317,7 @@ export declare const MODELS: {
|
|
|
5211
5317
|
readonly "magistral-small": {
|
|
5212
5318
|
id: string;
|
|
5213
5319
|
name: string;
|
|
5214
|
-
api: "
|
|
5320
|
+
api: "openai-completions";
|
|
5215
5321
|
provider: string;
|
|
5216
5322
|
baseUrl: string;
|
|
5217
5323
|
reasoning: true;
|
|
@@ -5228,7 +5334,7 @@ export declare const MODELS: {
|
|
|
5228
5334
|
readonly "ministral-3b-latest": {
|
|
5229
5335
|
id: string;
|
|
5230
5336
|
name: string;
|
|
5231
|
-
api: "
|
|
5337
|
+
api: "openai-completions";
|
|
5232
5338
|
provider: string;
|
|
5233
5339
|
baseUrl: string;
|
|
5234
5340
|
reasoning: false;
|
|
@@ -5245,7 +5351,7 @@ export declare const MODELS: {
|
|
|
5245
5351
|
readonly "ministral-8b-latest": {
|
|
5246
5352
|
id: string;
|
|
5247
5353
|
name: string;
|
|
5248
|
-
api: "
|
|
5354
|
+
api: "openai-completions";
|
|
5249
5355
|
provider: string;
|
|
5250
5356
|
baseUrl: string;
|
|
5251
5357
|
reasoning: false;
|
|
@@ -5262,7 +5368,7 @@ export declare const MODELS: {
|
|
|
5262
5368
|
readonly "mistral-large-2411": {
|
|
5263
5369
|
id: string;
|
|
5264
5370
|
name: string;
|
|
5265
|
-
api: "
|
|
5371
|
+
api: "openai-completions";
|
|
5266
5372
|
provider: string;
|
|
5267
5373
|
baseUrl: string;
|
|
5268
5374
|
reasoning: false;
|
|
@@ -5279,7 +5385,7 @@ export declare const MODELS: {
|
|
|
5279
5385
|
readonly "mistral-large-2512": {
|
|
5280
5386
|
id: string;
|
|
5281
5387
|
name: string;
|
|
5282
|
-
api: "
|
|
5388
|
+
api: "openai-completions";
|
|
5283
5389
|
provider: string;
|
|
5284
5390
|
baseUrl: string;
|
|
5285
5391
|
reasoning: false;
|
|
@@ -5296,7 +5402,7 @@ export declare const MODELS: {
|
|
|
5296
5402
|
readonly "mistral-large-latest": {
|
|
5297
5403
|
id: string;
|
|
5298
5404
|
name: string;
|
|
5299
|
-
api: "
|
|
5405
|
+
api: "openai-completions";
|
|
5300
5406
|
provider: string;
|
|
5301
5407
|
baseUrl: string;
|
|
5302
5408
|
reasoning: false;
|
|
@@ -5313,7 +5419,7 @@ export declare const MODELS: {
|
|
|
5313
5419
|
readonly "mistral-medium-2505": {
|
|
5314
5420
|
id: string;
|
|
5315
5421
|
name: string;
|
|
5316
|
-
api: "
|
|
5422
|
+
api: "openai-completions";
|
|
5317
5423
|
provider: string;
|
|
5318
5424
|
baseUrl: string;
|
|
5319
5425
|
reasoning: false;
|
|
@@ -5330,7 +5436,7 @@ export declare const MODELS: {
|
|
|
5330
5436
|
readonly "mistral-medium-2508": {
|
|
5331
5437
|
id: string;
|
|
5332
5438
|
name: string;
|
|
5333
|
-
api: "
|
|
5439
|
+
api: "openai-completions";
|
|
5334
5440
|
provider: string;
|
|
5335
5441
|
baseUrl: string;
|
|
5336
5442
|
reasoning: false;
|
|
@@ -5347,7 +5453,7 @@ export declare const MODELS: {
|
|
|
5347
5453
|
readonly "mistral-medium-latest": {
|
|
5348
5454
|
id: string;
|
|
5349
5455
|
name: string;
|
|
5350
|
-
api: "
|
|
5456
|
+
api: "openai-completions";
|
|
5351
5457
|
provider: string;
|
|
5352
5458
|
baseUrl: string;
|
|
5353
5459
|
reasoning: false;
|
|
@@ -5364,7 +5470,7 @@ export declare const MODELS: {
|
|
|
5364
5470
|
readonly "mistral-nemo": {
|
|
5365
5471
|
id: string;
|
|
5366
5472
|
name: string;
|
|
5367
|
-
api: "
|
|
5473
|
+
api: "openai-completions";
|
|
5368
5474
|
provider: string;
|
|
5369
5475
|
baseUrl: string;
|
|
5370
5476
|
reasoning: false;
|
|
@@ -5381,7 +5487,7 @@ export declare const MODELS: {
|
|
|
5381
5487
|
readonly "mistral-small-2506": {
|
|
5382
5488
|
id: string;
|
|
5383
5489
|
name: string;
|
|
5384
|
-
api: "
|
|
5490
|
+
api: "openai-completions";
|
|
5385
5491
|
provider: string;
|
|
5386
5492
|
baseUrl: string;
|
|
5387
5493
|
reasoning: false;
|
|
@@ -5398,7 +5504,7 @@ export declare const MODELS: {
|
|
|
5398
5504
|
readonly "mistral-small-2603": {
|
|
5399
5505
|
id: string;
|
|
5400
5506
|
name: string;
|
|
5401
|
-
api: "
|
|
5507
|
+
api: "openai-completions";
|
|
5402
5508
|
provider: string;
|
|
5403
5509
|
baseUrl: string;
|
|
5404
5510
|
reasoning: true;
|
|
@@ -5415,7 +5521,7 @@ export declare const MODELS: {
|
|
|
5415
5521
|
readonly "mistral-small-latest": {
|
|
5416
5522
|
id: string;
|
|
5417
5523
|
name: string;
|
|
5418
|
-
api: "
|
|
5524
|
+
api: "openai-completions";
|
|
5419
5525
|
provider: string;
|
|
5420
5526
|
baseUrl: string;
|
|
5421
5527
|
reasoning: true;
|
|
@@ -5432,7 +5538,7 @@ export declare const MODELS: {
|
|
|
5432
5538
|
readonly "open-mistral-7b": {
|
|
5433
5539
|
id: string;
|
|
5434
5540
|
name: string;
|
|
5435
|
-
api: "
|
|
5541
|
+
api: "openai-completions";
|
|
5436
5542
|
provider: string;
|
|
5437
5543
|
baseUrl: string;
|
|
5438
5544
|
reasoning: false;
|
|
@@ -5449,7 +5555,7 @@ export declare const MODELS: {
|
|
|
5449
5555
|
readonly "open-mixtral-8x22b": {
|
|
5450
5556
|
id: string;
|
|
5451
5557
|
name: string;
|
|
5452
|
-
api: "
|
|
5558
|
+
api: "openai-completions";
|
|
5453
5559
|
provider: string;
|
|
5454
5560
|
baseUrl: string;
|
|
5455
5561
|
reasoning: false;
|
|
@@ -5466,7 +5572,7 @@ export declare const MODELS: {
|
|
|
5466
5572
|
readonly "open-mixtral-8x7b": {
|
|
5467
5573
|
id: string;
|
|
5468
5574
|
name: string;
|
|
5469
|
-
api: "
|
|
5575
|
+
api: "openai-completions";
|
|
5470
5576
|
provider: string;
|
|
5471
5577
|
baseUrl: string;
|
|
5472
5578
|
reasoning: false;
|
|
@@ -5483,7 +5589,7 @@ export declare const MODELS: {
|
|
|
5483
5589
|
readonly "pixtral-12b": {
|
|
5484
5590
|
id: string;
|
|
5485
5591
|
name: string;
|
|
5486
|
-
api: "
|
|
5592
|
+
api: "openai-completions";
|
|
5487
5593
|
provider: string;
|
|
5488
5594
|
baseUrl: string;
|
|
5489
5595
|
reasoning: false;
|
|
@@ -5500,7 +5606,7 @@ export declare const MODELS: {
|
|
|
5500
5606
|
readonly "pixtral-large-latest": {
|
|
5501
5607
|
id: string;
|
|
5502
5608
|
name: string;
|
|
5503
|
-
api: "
|
|
5609
|
+
api: "openai-completions";
|
|
5504
5610
|
provider: string;
|
|
5505
5611
|
baseUrl: string;
|
|
5506
5612
|
reasoning: false;
|
|
@@ -5516,23 +5622,6 @@ export declare const MODELS: {
|
|
|
5516
5622
|
};
|
|
5517
5623
|
};
|
|
5518
5624
|
readonly openai: {
|
|
5519
|
-
readonly "codex-mini-latest": {
|
|
5520
|
-
id: string;
|
|
5521
|
-
name: string;
|
|
5522
|
-
api: "openai-responses";
|
|
5523
|
-
provider: string;
|
|
5524
|
-
baseUrl: string;
|
|
5525
|
-
reasoning: true;
|
|
5526
|
-
input: "text"[];
|
|
5527
|
-
cost: {
|
|
5528
|
-
input: number;
|
|
5529
|
-
output: number;
|
|
5530
|
-
cacheRead: number;
|
|
5531
|
-
cacheWrite: number;
|
|
5532
|
-
};
|
|
5533
|
-
contextWindow: number;
|
|
5534
|
-
maxTokens: number;
|
|
5535
|
-
};
|
|
5536
5625
|
readonly "gpt-4": {
|
|
5537
5626
|
id: string;
|
|
5538
5627
|
name: string;
|
|
@@ -6077,6 +6166,23 @@ export declare const MODELS: {
|
|
|
6077
6166
|
contextWindow: number;
|
|
6078
6167
|
maxTokens: number;
|
|
6079
6168
|
};
|
|
6169
|
+
readonly "gpt-5.5": {
|
|
6170
|
+
id: string;
|
|
6171
|
+
name: string;
|
|
6172
|
+
api: "openai-responses";
|
|
6173
|
+
provider: string;
|
|
6174
|
+
baseUrl: string;
|
|
6175
|
+
reasoning: true;
|
|
6176
|
+
input: ("image" | "text")[];
|
|
6177
|
+
cost: {
|
|
6178
|
+
input: number;
|
|
6179
|
+
output: number;
|
|
6180
|
+
cacheRead: number;
|
|
6181
|
+
cacheWrite: number;
|
|
6182
|
+
};
|
|
6183
|
+
contextWindow: number;
|
|
6184
|
+
maxTokens: number;
|
|
6185
|
+
};
|
|
6080
6186
|
readonly o1: {
|
|
6081
6187
|
id: string;
|
|
6082
6188
|
name: string;
|
|
@@ -6472,6 +6578,23 @@ export declare const MODELS: {
|
|
|
6472
6578
|
contextWindow: number;
|
|
6473
6579
|
maxTokens: number;
|
|
6474
6580
|
};
|
|
6581
|
+
readonly "claude-opus-4-7": {
|
|
6582
|
+
id: string;
|
|
6583
|
+
name: string;
|
|
6584
|
+
api: "anthropic-messages";
|
|
6585
|
+
provider: string;
|
|
6586
|
+
baseUrl: string;
|
|
6587
|
+
reasoning: true;
|
|
6588
|
+
input: ("image" | "text")[];
|
|
6589
|
+
cost: {
|
|
6590
|
+
input: number;
|
|
6591
|
+
output: number;
|
|
6592
|
+
cacheRead: number;
|
|
6593
|
+
cacheWrite: number;
|
|
6594
|
+
};
|
|
6595
|
+
contextWindow: number;
|
|
6596
|
+
maxTokens: number;
|
|
6597
|
+
};
|
|
6475
6598
|
readonly "claude-sonnet-4": {
|
|
6476
6599
|
id: string;
|
|
6477
6600
|
name: string;
|
|
@@ -6574,6 +6697,23 @@ export declare const MODELS: {
|
|
|
6574
6697
|
contextWindow: number;
|
|
6575
6698
|
maxTokens: number;
|
|
6576
6699
|
};
|
|
6700
|
+
readonly "glm-5.1": {
|
|
6701
|
+
id: string;
|
|
6702
|
+
name: string;
|
|
6703
|
+
api: "openai-completions";
|
|
6704
|
+
provider: string;
|
|
6705
|
+
baseUrl: string;
|
|
6706
|
+
reasoning: true;
|
|
6707
|
+
input: "text"[];
|
|
6708
|
+
cost: {
|
|
6709
|
+
input: number;
|
|
6710
|
+
output: number;
|
|
6711
|
+
cacheRead: number;
|
|
6712
|
+
cacheWrite: number;
|
|
6713
|
+
};
|
|
6714
|
+
contextWindow: number;
|
|
6715
|
+
maxTokens: number;
|
|
6716
|
+
};
|
|
6577
6717
|
readonly "gpt-5": {
|
|
6578
6718
|
id: string;
|
|
6579
6719
|
name: string;
|
|
@@ -6812,10 +6952,10 @@ export declare const MODELS: {
|
|
|
6812
6952
|
contextWindow: number;
|
|
6813
6953
|
maxTokens: number;
|
|
6814
6954
|
};
|
|
6815
|
-
readonly "
|
|
6955
|
+
readonly "gpt-5.5": {
|
|
6816
6956
|
id: string;
|
|
6817
6957
|
name: string;
|
|
6818
|
-
api: "openai-
|
|
6958
|
+
api: "openai-responses";
|
|
6819
6959
|
provider: string;
|
|
6820
6960
|
baseUrl: string;
|
|
6821
6961
|
reasoning: true;
|
|
@@ -6829,14 +6969,14 @@ export declare const MODELS: {
|
|
|
6829
6969
|
contextWindow: number;
|
|
6830
6970
|
maxTokens: number;
|
|
6831
6971
|
};
|
|
6832
|
-
readonly "
|
|
6972
|
+
readonly "gpt-5.5-pro": {
|
|
6833
6973
|
id: string;
|
|
6834
6974
|
name: string;
|
|
6835
|
-
api: "openai-
|
|
6975
|
+
api: "openai-responses";
|
|
6836
6976
|
provider: string;
|
|
6837
6977
|
baseUrl: string;
|
|
6838
6978
|
reasoning: true;
|
|
6839
|
-
input: "text"[];
|
|
6979
|
+
input: ("image" | "text")[];
|
|
6840
6980
|
cost: {
|
|
6841
6981
|
input: number;
|
|
6842
6982
|
output: number;
|
|
@@ -6846,10 +6986,10 @@ export declare const MODELS: {
|
|
|
6846
6986
|
contextWindow: number;
|
|
6847
6987
|
maxTokens: number;
|
|
6848
6988
|
};
|
|
6849
|
-
readonly "
|
|
6989
|
+
readonly "hy3-preview-free": {
|
|
6850
6990
|
id: string;
|
|
6851
6991
|
name: string;
|
|
6852
|
-
api: "
|
|
6992
|
+
api: "openai-completions";
|
|
6853
6993
|
provider: string;
|
|
6854
6994
|
baseUrl: string;
|
|
6855
6995
|
reasoning: true;
|
|
@@ -6863,14 +7003,14 @@ export declare const MODELS: {
|
|
|
6863
7003
|
contextWindow: number;
|
|
6864
7004
|
maxTokens: number;
|
|
6865
7005
|
};
|
|
6866
|
-
readonly "
|
|
7006
|
+
readonly "kimi-k2.5": {
|
|
6867
7007
|
id: string;
|
|
6868
7008
|
name: string;
|
|
6869
7009
|
api: "openai-completions";
|
|
6870
7010
|
provider: string;
|
|
6871
7011
|
baseUrl: string;
|
|
6872
7012
|
reasoning: true;
|
|
6873
|
-
input: "text"[];
|
|
7013
|
+
input: ("image" | "text")[];
|
|
6874
7014
|
cost: {
|
|
6875
7015
|
input: number;
|
|
6876
7016
|
output: number;
|
|
@@ -6880,14 +7020,14 @@ export declare const MODELS: {
|
|
|
6880
7020
|
contextWindow: number;
|
|
6881
7021
|
maxTokens: number;
|
|
6882
7022
|
};
|
|
6883
|
-
readonly "
|
|
7023
|
+
readonly "kimi-k2.6": {
|
|
6884
7024
|
id: string;
|
|
6885
7025
|
name: string;
|
|
6886
7026
|
api: "openai-completions";
|
|
6887
7027
|
provider: string;
|
|
6888
7028
|
baseUrl: string;
|
|
6889
7029
|
reasoning: true;
|
|
6890
|
-
input: "text"[];
|
|
7030
|
+
input: ("image" | "text")[];
|
|
6891
7031
|
cost: {
|
|
6892
7032
|
input: number;
|
|
6893
7033
|
output: number;
|
|
@@ -6897,15 +7037,13 @@ export declare const MODELS: {
|
|
|
6897
7037
|
contextWindow: number;
|
|
6898
7038
|
maxTokens: number;
|
|
6899
7039
|
};
|
|
6900
|
-
|
|
6901
|
-
readonly "opencode-go": {
|
|
6902
|
-
readonly "glm-5": {
|
|
7040
|
+
readonly "ling-2.6-flash-free": {
|
|
6903
7041
|
id: string;
|
|
6904
7042
|
name: string;
|
|
6905
7043
|
api: "openai-completions";
|
|
6906
7044
|
provider: string;
|
|
6907
7045
|
baseUrl: string;
|
|
6908
|
-
reasoning:
|
|
7046
|
+
reasoning: false;
|
|
6909
7047
|
input: "text"[];
|
|
6910
7048
|
cost: {
|
|
6911
7049
|
input: number;
|
|
@@ -6916,14 +7054,14 @@ export declare const MODELS: {
|
|
|
6916
7054
|
contextWindow: number;
|
|
6917
7055
|
maxTokens: number;
|
|
6918
7056
|
};
|
|
6919
|
-
readonly "
|
|
7057
|
+
readonly "minimax-m2.5": {
|
|
6920
7058
|
id: string;
|
|
6921
7059
|
name: string;
|
|
6922
7060
|
api: "openai-completions";
|
|
6923
7061
|
provider: string;
|
|
6924
7062
|
baseUrl: string;
|
|
6925
7063
|
reasoning: true;
|
|
6926
|
-
input:
|
|
7064
|
+
input: "text"[];
|
|
6927
7065
|
cost: {
|
|
6928
7066
|
input: number;
|
|
6929
7067
|
output: number;
|
|
@@ -6933,14 +7071,14 @@ export declare const MODELS: {
|
|
|
6933
7071
|
contextWindow: number;
|
|
6934
7072
|
maxTokens: number;
|
|
6935
7073
|
};
|
|
6936
|
-
readonly "
|
|
7074
|
+
readonly "minimax-m2.5-free": {
|
|
6937
7075
|
id: string;
|
|
6938
7076
|
name: string;
|
|
6939
|
-
api: "
|
|
7077
|
+
api: "anthropic-messages";
|
|
6940
7078
|
provider: string;
|
|
6941
7079
|
baseUrl: string;
|
|
6942
7080
|
reasoning: true;
|
|
6943
|
-
input:
|
|
7081
|
+
input: "text"[];
|
|
6944
7082
|
cost: {
|
|
6945
7083
|
input: number;
|
|
6946
7084
|
output: number;
|
|
@@ -6950,7 +7088,7 @@ export declare const MODELS: {
|
|
|
6950
7088
|
contextWindow: number;
|
|
6951
7089
|
maxTokens: number;
|
|
6952
7090
|
};
|
|
6953
|
-
readonly "
|
|
7091
|
+
readonly "minimax-m2.7": {
|
|
6954
7092
|
id: string;
|
|
6955
7093
|
name: string;
|
|
6956
7094
|
api: "openai-completions";
|
|
@@ -6967,7 +7105,7 @@ export declare const MODELS: {
|
|
|
6967
7105
|
contextWindow: number;
|
|
6968
7106
|
maxTokens: number;
|
|
6969
7107
|
};
|
|
6970
|
-
readonly "
|
|
7108
|
+
readonly "nemotron-3-super-free": {
|
|
6971
7109
|
id: string;
|
|
6972
7110
|
name: string;
|
|
6973
7111
|
api: "openai-completions";
|
|
@@ -6984,14 +7122,14 @@ export declare const MODELS: {
|
|
|
6984
7122
|
contextWindow: number;
|
|
6985
7123
|
maxTokens: number;
|
|
6986
7124
|
};
|
|
6987
|
-
readonly "
|
|
7125
|
+
readonly "qwen3.5-plus": {
|
|
6988
7126
|
id: string;
|
|
6989
7127
|
name: string;
|
|
6990
7128
|
api: "anthropic-messages";
|
|
6991
7129
|
provider: string;
|
|
6992
7130
|
baseUrl: string;
|
|
6993
7131
|
reasoning: true;
|
|
6994
|
-
input: "text"[];
|
|
7132
|
+
input: ("image" | "text")[];
|
|
6995
7133
|
cost: {
|
|
6996
7134
|
input: number;
|
|
6997
7135
|
output: number;
|
|
@@ -7001,16 +7139,14 @@ export declare const MODELS: {
|
|
|
7001
7139
|
contextWindow: number;
|
|
7002
7140
|
maxTokens: number;
|
|
7003
7141
|
};
|
|
7004
|
-
|
|
7005
|
-
readonly openrouter: {
|
|
7006
|
-
readonly "ai21/jamba-large-1.7": {
|
|
7142
|
+
readonly "qwen3.6-plus": {
|
|
7007
7143
|
id: string;
|
|
7008
7144
|
name: string;
|
|
7009
|
-
api: "
|
|
7145
|
+
api: "anthropic-messages";
|
|
7010
7146
|
provider: string;
|
|
7011
7147
|
baseUrl: string;
|
|
7012
|
-
reasoning:
|
|
7013
|
-
input: "text"[];
|
|
7148
|
+
reasoning: true;
|
|
7149
|
+
input: ("image" | "text")[];
|
|
7014
7150
|
cost: {
|
|
7015
7151
|
input: number;
|
|
7016
7152
|
output: number;
|
|
@@ -7020,8 +7156,267 @@ export declare const MODELS: {
|
|
|
7020
7156
|
contextWindow: number;
|
|
7021
7157
|
maxTokens: number;
|
|
7022
7158
|
};
|
|
7023
|
-
|
|
7024
|
-
|
|
7159
|
+
};
|
|
7160
|
+
readonly "opencode-go": {
|
|
7161
|
+
readonly "deepseek-v4-flash": {
|
|
7162
|
+
id: string;
|
|
7163
|
+
name: string;
|
|
7164
|
+
api: "openai-completions";
|
|
7165
|
+
provider: string;
|
|
7166
|
+
baseUrl: string;
|
|
7167
|
+
reasoning: true;
|
|
7168
|
+
input: "text"[];
|
|
7169
|
+
cost: {
|
|
7170
|
+
input: number;
|
|
7171
|
+
output: number;
|
|
7172
|
+
cacheRead: number;
|
|
7173
|
+
cacheWrite: number;
|
|
7174
|
+
};
|
|
7175
|
+
contextWindow: number;
|
|
7176
|
+
maxTokens: number;
|
|
7177
|
+
};
|
|
7178
|
+
readonly "deepseek-v4-pro": {
|
|
7179
|
+
id: string;
|
|
7180
|
+
name: string;
|
|
7181
|
+
api: "openai-completions";
|
|
7182
|
+
provider: string;
|
|
7183
|
+
baseUrl: string;
|
|
7184
|
+
reasoning: true;
|
|
7185
|
+
input: "text"[];
|
|
7186
|
+
cost: {
|
|
7187
|
+
input: number;
|
|
7188
|
+
output: number;
|
|
7189
|
+
cacheRead: number;
|
|
7190
|
+
cacheWrite: number;
|
|
7191
|
+
};
|
|
7192
|
+
contextWindow: number;
|
|
7193
|
+
maxTokens: number;
|
|
7194
|
+
};
|
|
7195
|
+
readonly "glm-5": {
|
|
7196
|
+
id: string;
|
|
7197
|
+
name: string;
|
|
7198
|
+
api: "openai-completions";
|
|
7199
|
+
provider: string;
|
|
7200
|
+
baseUrl: string;
|
|
7201
|
+
reasoning: true;
|
|
7202
|
+
input: "text"[];
|
|
7203
|
+
cost: {
|
|
7204
|
+
input: number;
|
|
7205
|
+
output: number;
|
|
7206
|
+
cacheRead: number;
|
|
7207
|
+
cacheWrite: number;
|
|
7208
|
+
};
|
|
7209
|
+
contextWindow: number;
|
|
7210
|
+
maxTokens: number;
|
|
7211
|
+
};
|
|
7212
|
+
readonly "glm-5.1": {
|
|
7213
|
+
id: string;
|
|
7214
|
+
name: string;
|
|
7215
|
+
api: "openai-completions";
|
|
7216
|
+
provider: string;
|
|
7217
|
+
baseUrl: string;
|
|
7218
|
+
reasoning: true;
|
|
7219
|
+
input: "text"[];
|
|
7220
|
+
cost: {
|
|
7221
|
+
input: number;
|
|
7222
|
+
output: number;
|
|
7223
|
+
cacheRead: number;
|
|
7224
|
+
cacheWrite: number;
|
|
7225
|
+
};
|
|
7226
|
+
contextWindow: number;
|
|
7227
|
+
maxTokens: number;
|
|
7228
|
+
};
|
|
7229
|
+
readonly "kimi-k2.5": {
|
|
7230
|
+
id: string;
|
|
7231
|
+
name: string;
|
|
7232
|
+
api: "openai-completions";
|
|
7233
|
+
provider: string;
|
|
7234
|
+
baseUrl: string;
|
|
7235
|
+
reasoning: true;
|
|
7236
|
+
input: ("image" | "text")[];
|
|
7237
|
+
cost: {
|
|
7238
|
+
input: number;
|
|
7239
|
+
output: number;
|
|
7240
|
+
cacheRead: number;
|
|
7241
|
+
cacheWrite: number;
|
|
7242
|
+
};
|
|
7243
|
+
contextWindow: number;
|
|
7244
|
+
maxTokens: number;
|
|
7245
|
+
};
|
|
7246
|
+
readonly "kimi-k2.6": {
|
|
7247
|
+
id: string;
|
|
7248
|
+
name: string;
|
|
7249
|
+
api: "openai-completions";
|
|
7250
|
+
provider: string;
|
|
7251
|
+
baseUrl: string;
|
|
7252
|
+
reasoning: true;
|
|
7253
|
+
input: ("image" | "text")[];
|
|
7254
|
+
cost: {
|
|
7255
|
+
input: number;
|
|
7256
|
+
output: number;
|
|
7257
|
+
cacheRead: number;
|
|
7258
|
+
cacheWrite: number;
|
|
7259
|
+
};
|
|
7260
|
+
contextWindow: number;
|
|
7261
|
+
maxTokens: number;
|
|
7262
|
+
};
|
|
7263
|
+
readonly "mimo-v2-omni": {
|
|
7264
|
+
id: string;
|
|
7265
|
+
name: string;
|
|
7266
|
+
api: "openai-completions";
|
|
7267
|
+
provider: string;
|
|
7268
|
+
baseUrl: string;
|
|
7269
|
+
reasoning: true;
|
|
7270
|
+
input: ("image" | "text")[];
|
|
7271
|
+
cost: {
|
|
7272
|
+
input: number;
|
|
7273
|
+
output: number;
|
|
7274
|
+
cacheRead: number;
|
|
7275
|
+
cacheWrite: number;
|
|
7276
|
+
};
|
|
7277
|
+
contextWindow: number;
|
|
7278
|
+
maxTokens: number;
|
|
7279
|
+
};
|
|
7280
|
+
readonly "mimo-v2-pro": {
|
|
7281
|
+
id: string;
|
|
7282
|
+
name: string;
|
|
7283
|
+
api: "openai-completions";
|
|
7284
|
+
provider: string;
|
|
7285
|
+
baseUrl: string;
|
|
7286
|
+
reasoning: true;
|
|
7287
|
+
input: "text"[];
|
|
7288
|
+
cost: {
|
|
7289
|
+
input: number;
|
|
7290
|
+
output: number;
|
|
7291
|
+
cacheRead: number;
|
|
7292
|
+
cacheWrite: number;
|
|
7293
|
+
};
|
|
7294
|
+
contextWindow: number;
|
|
7295
|
+
maxTokens: number;
|
|
7296
|
+
};
|
|
7297
|
+
readonly "mimo-v2.5": {
|
|
7298
|
+
id: string;
|
|
7299
|
+
name: string;
|
|
7300
|
+
api: "openai-completions";
|
|
7301
|
+
provider: string;
|
|
7302
|
+
baseUrl: string;
|
|
7303
|
+
reasoning: true;
|
|
7304
|
+
input: ("image" | "text")[];
|
|
7305
|
+
cost: {
|
|
7306
|
+
input: number;
|
|
7307
|
+
output: number;
|
|
7308
|
+
cacheRead: number;
|
|
7309
|
+
cacheWrite: number;
|
|
7310
|
+
};
|
|
7311
|
+
contextWindow: number;
|
|
7312
|
+
maxTokens: number;
|
|
7313
|
+
};
|
|
7314
|
+
readonly "mimo-v2.5-pro": {
|
|
7315
|
+
id: string;
|
|
7316
|
+
name: string;
|
|
7317
|
+
api: "openai-completions";
|
|
7318
|
+
provider: string;
|
|
7319
|
+
baseUrl: string;
|
|
7320
|
+
reasoning: true;
|
|
7321
|
+
input: "text"[];
|
|
7322
|
+
cost: {
|
|
7323
|
+
input: number;
|
|
7324
|
+
output: number;
|
|
7325
|
+
cacheRead: number;
|
|
7326
|
+
cacheWrite: number;
|
|
7327
|
+
};
|
|
7328
|
+
contextWindow: number;
|
|
7329
|
+
maxTokens: number;
|
|
7330
|
+
};
|
|
7331
|
+
readonly "minimax-m2.5": {
|
|
7332
|
+
id: string;
|
|
7333
|
+
name: string;
|
|
7334
|
+
api: "openai-completions";
|
|
7335
|
+
provider: string;
|
|
7336
|
+
baseUrl: string;
|
|
7337
|
+
reasoning: true;
|
|
7338
|
+
input: "text"[];
|
|
7339
|
+
cost: {
|
|
7340
|
+
input: number;
|
|
7341
|
+
output: number;
|
|
7342
|
+
cacheRead: number;
|
|
7343
|
+
cacheWrite: number;
|
|
7344
|
+
};
|
|
7345
|
+
contextWindow: number;
|
|
7346
|
+
maxTokens: number;
|
|
7347
|
+
};
|
|
7348
|
+
readonly "minimax-m2.7": {
|
|
7349
|
+
id: string;
|
|
7350
|
+
name: string;
|
|
7351
|
+
api: "anthropic-messages";
|
|
7352
|
+
provider: string;
|
|
7353
|
+
baseUrl: string;
|
|
7354
|
+
reasoning: true;
|
|
7355
|
+
input: "text"[];
|
|
7356
|
+
cost: {
|
|
7357
|
+
input: number;
|
|
7358
|
+
output: number;
|
|
7359
|
+
cacheRead: number;
|
|
7360
|
+
cacheWrite: number;
|
|
7361
|
+
};
|
|
7362
|
+
contextWindow: number;
|
|
7363
|
+
maxTokens: number;
|
|
7364
|
+
};
|
|
7365
|
+
readonly "qwen3.5-plus": {
|
|
7366
|
+
id: string;
|
|
7367
|
+
name: string;
|
|
7368
|
+
api: "anthropic-messages";
|
|
7369
|
+
provider: string;
|
|
7370
|
+
baseUrl: string;
|
|
7371
|
+
reasoning: true;
|
|
7372
|
+
input: ("image" | "text")[];
|
|
7373
|
+
cost: {
|
|
7374
|
+
input: number;
|
|
7375
|
+
output: number;
|
|
7376
|
+
cacheRead: number;
|
|
7377
|
+
cacheWrite: number;
|
|
7378
|
+
};
|
|
7379
|
+
contextWindow: number;
|
|
7380
|
+
maxTokens: number;
|
|
7381
|
+
};
|
|
7382
|
+
readonly "qwen3.6-plus": {
|
|
7383
|
+
id: string;
|
|
7384
|
+
name: string;
|
|
7385
|
+
api: "anthropic-messages";
|
|
7386
|
+
provider: string;
|
|
7387
|
+
baseUrl: string;
|
|
7388
|
+
reasoning: true;
|
|
7389
|
+
input: ("image" | "text")[];
|
|
7390
|
+
cost: {
|
|
7391
|
+
input: number;
|
|
7392
|
+
output: number;
|
|
7393
|
+
cacheRead: number;
|
|
7394
|
+
cacheWrite: number;
|
|
7395
|
+
};
|
|
7396
|
+
contextWindow: number;
|
|
7397
|
+
maxTokens: number;
|
|
7398
|
+
};
|
|
7399
|
+
};
|
|
7400
|
+
readonly openrouter: {
|
|
7401
|
+
readonly "ai21/jamba-large-1.7": {
|
|
7402
|
+
id: string;
|
|
7403
|
+
name: string;
|
|
7404
|
+
api: "openai-completions";
|
|
7405
|
+
provider: string;
|
|
7406
|
+
baseUrl: string;
|
|
7407
|
+
reasoning: false;
|
|
7408
|
+
input: "text"[];
|
|
7409
|
+
cost: {
|
|
7410
|
+
input: number;
|
|
7411
|
+
output: number;
|
|
7412
|
+
cacheRead: number;
|
|
7413
|
+
cacheWrite: number;
|
|
7414
|
+
};
|
|
7415
|
+
contextWindow: number;
|
|
7416
|
+
maxTokens: number;
|
|
7417
|
+
};
|
|
7418
|
+
readonly "alibaba/tongyi-deepresearch-30b-a3b": {
|
|
7419
|
+
id: string;
|
|
7025
7420
|
name: string;
|
|
7026
7421
|
api: "openai-completions";
|
|
7027
7422
|
provider: string;
|
|
@@ -7292,7 +7687,7 @@ export declare const MODELS: {
|
|
|
7292
7687
|
contextWindow: number;
|
|
7293
7688
|
maxTokens: number;
|
|
7294
7689
|
};
|
|
7295
|
-
readonly "anthropic/claude-
|
|
7690
|
+
readonly "anthropic/claude-opus-4.6-fast": {
|
|
7296
7691
|
id: string;
|
|
7297
7692
|
name: string;
|
|
7298
7693
|
api: "openai-completions";
|
|
@@ -7309,7 +7704,7 @@ export declare const MODELS: {
|
|
|
7309
7704
|
contextWindow: number;
|
|
7310
7705
|
maxTokens: number;
|
|
7311
7706
|
};
|
|
7312
|
-
readonly "anthropic/claude-
|
|
7707
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
7313
7708
|
id: string;
|
|
7314
7709
|
name: string;
|
|
7315
7710
|
api: "openai-completions";
|
|
@@ -7326,7 +7721,7 @@ export declare const MODELS: {
|
|
|
7326
7721
|
contextWindow: number;
|
|
7327
7722
|
maxTokens: number;
|
|
7328
7723
|
};
|
|
7329
|
-
readonly "anthropic/claude-sonnet-4
|
|
7724
|
+
readonly "anthropic/claude-sonnet-4": {
|
|
7330
7725
|
id: string;
|
|
7331
7726
|
name: string;
|
|
7332
7727
|
api: "openai-completions";
|
|
@@ -7343,14 +7738,14 @@ export declare const MODELS: {
|
|
|
7343
7738
|
contextWindow: number;
|
|
7344
7739
|
maxTokens: number;
|
|
7345
7740
|
};
|
|
7346
|
-
readonly "
|
|
7741
|
+
readonly "anthropic/claude-sonnet-4.5": {
|
|
7347
7742
|
id: string;
|
|
7348
7743
|
name: string;
|
|
7349
7744
|
api: "openai-completions";
|
|
7350
7745
|
provider: string;
|
|
7351
7746
|
baseUrl: string;
|
|
7352
|
-
reasoning:
|
|
7353
|
-
input: "text"[];
|
|
7747
|
+
reasoning: true;
|
|
7748
|
+
input: ("image" | "text")[];
|
|
7354
7749
|
cost: {
|
|
7355
7750
|
input: number;
|
|
7356
7751
|
output: number;
|
|
@@ -7360,13 +7755,30 @@ export declare const MODELS: {
|
|
|
7360
7755
|
contextWindow: number;
|
|
7361
7756
|
maxTokens: number;
|
|
7362
7757
|
};
|
|
7363
|
-
readonly "
|
|
7758
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
7364
7759
|
id: string;
|
|
7365
7760
|
name: string;
|
|
7366
7761
|
api: "openai-completions";
|
|
7367
7762
|
provider: string;
|
|
7368
7763
|
baseUrl: string;
|
|
7369
7764
|
reasoning: true;
|
|
7765
|
+
input: ("image" | "text")[];
|
|
7766
|
+
cost: {
|
|
7767
|
+
input: number;
|
|
7768
|
+
output: number;
|
|
7769
|
+
cacheRead: number;
|
|
7770
|
+
cacheWrite: number;
|
|
7771
|
+
};
|
|
7772
|
+
contextWindow: number;
|
|
7773
|
+
maxTokens: number;
|
|
7774
|
+
};
|
|
7775
|
+
readonly "arcee-ai/trinity-large-preview": {
|
|
7776
|
+
id: string;
|
|
7777
|
+
name: string;
|
|
7778
|
+
api: "openai-completions";
|
|
7779
|
+
provider: string;
|
|
7780
|
+
baseUrl: string;
|
|
7781
|
+
reasoning: false;
|
|
7370
7782
|
input: "text"[];
|
|
7371
7783
|
cost: {
|
|
7372
7784
|
input: number;
|
|
@@ -7377,7 +7789,7 @@ export declare const MODELS: {
|
|
|
7377
7789
|
contextWindow: number;
|
|
7378
7790
|
maxTokens: number;
|
|
7379
7791
|
};
|
|
7380
|
-
readonly "arcee-ai/trinity-
|
|
7792
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
7381
7793
|
id: string;
|
|
7382
7794
|
name: string;
|
|
7383
7795
|
api: "openai-completions";
|
|
@@ -7394,7 +7806,7 @@ export declare const MODELS: {
|
|
|
7394
7806
|
contextWindow: number;
|
|
7395
7807
|
maxTokens: number;
|
|
7396
7808
|
};
|
|
7397
|
-
readonly "arcee-ai/trinity-mini
|
|
7809
|
+
readonly "arcee-ai/trinity-mini": {
|
|
7398
7810
|
id: string;
|
|
7399
7811
|
name: string;
|
|
7400
7812
|
api: "openai-completions";
|
|
@@ -7604,7 +8016,7 @@ export declare const MODELS: {
|
|
|
7604
8016
|
api: "openai-completions";
|
|
7605
8017
|
provider: string;
|
|
7606
8018
|
baseUrl: string;
|
|
7607
|
-
reasoning:
|
|
8019
|
+
reasoning: false;
|
|
7608
8020
|
input: "text"[];
|
|
7609
8021
|
cost: {
|
|
7610
8022
|
input: number;
|
|
@@ -7717,6 +8129,40 @@ export declare const MODELS: {
|
|
|
7717
8129
|
contextWindow: number;
|
|
7718
8130
|
maxTokens: number;
|
|
7719
8131
|
};
|
|
8132
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
8133
|
+
id: string;
|
|
8134
|
+
name: string;
|
|
8135
|
+
api: "openai-completions";
|
|
8136
|
+
provider: string;
|
|
8137
|
+
baseUrl: string;
|
|
8138
|
+
reasoning: true;
|
|
8139
|
+
input: "text"[];
|
|
8140
|
+
cost: {
|
|
8141
|
+
input: number;
|
|
8142
|
+
output: number;
|
|
8143
|
+
cacheRead: number;
|
|
8144
|
+
cacheWrite: number;
|
|
8145
|
+
};
|
|
8146
|
+
contextWindow: number;
|
|
8147
|
+
maxTokens: number;
|
|
8148
|
+
};
|
|
8149
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
8150
|
+
id: string;
|
|
8151
|
+
name: string;
|
|
8152
|
+
api: "openai-completions";
|
|
8153
|
+
provider: string;
|
|
8154
|
+
baseUrl: string;
|
|
8155
|
+
reasoning: true;
|
|
8156
|
+
input: "text"[];
|
|
8157
|
+
cost: {
|
|
8158
|
+
input: number;
|
|
8159
|
+
output: number;
|
|
8160
|
+
cacheRead: number;
|
|
8161
|
+
cacheWrite: number;
|
|
8162
|
+
};
|
|
8163
|
+
contextWindow: number;
|
|
8164
|
+
maxTokens: number;
|
|
8165
|
+
};
|
|
7720
8166
|
readonly "essentialai/rnj-1-instruct": {
|
|
7721
8167
|
id: string;
|
|
7722
8168
|
name: string;
|
|
@@ -7785,7 +8231,58 @@ export declare const MODELS: {
|
|
|
7785
8231
|
contextWindow: number;
|
|
7786
8232
|
maxTokens: number;
|
|
7787
8233
|
};
|
|
7788
|
-
readonly "google/gemini-2.5-flash-lite": {
|
|
8234
|
+
readonly "google/gemini-2.5-flash-lite": {
|
|
8235
|
+
id: string;
|
|
8236
|
+
name: string;
|
|
8237
|
+
api: "openai-completions";
|
|
8238
|
+
provider: string;
|
|
8239
|
+
baseUrl: string;
|
|
8240
|
+
reasoning: true;
|
|
8241
|
+
input: ("image" | "text")[];
|
|
8242
|
+
cost: {
|
|
8243
|
+
input: number;
|
|
8244
|
+
output: number;
|
|
8245
|
+
cacheRead: number;
|
|
8246
|
+
cacheWrite: number;
|
|
8247
|
+
};
|
|
8248
|
+
contextWindow: number;
|
|
8249
|
+
maxTokens: number;
|
|
8250
|
+
};
|
|
8251
|
+
readonly "google/gemini-2.5-flash-lite-preview-09-2025": {
|
|
8252
|
+
id: string;
|
|
8253
|
+
name: string;
|
|
8254
|
+
api: "openai-completions";
|
|
8255
|
+
provider: string;
|
|
8256
|
+
baseUrl: string;
|
|
8257
|
+
reasoning: true;
|
|
8258
|
+
input: ("image" | "text")[];
|
|
8259
|
+
cost: {
|
|
8260
|
+
input: number;
|
|
8261
|
+
output: number;
|
|
8262
|
+
cacheRead: number;
|
|
8263
|
+
cacheWrite: number;
|
|
8264
|
+
};
|
|
8265
|
+
contextWindow: number;
|
|
8266
|
+
maxTokens: number;
|
|
8267
|
+
};
|
|
8268
|
+
readonly "google/gemini-2.5-pro": {
|
|
8269
|
+
id: string;
|
|
8270
|
+
name: string;
|
|
8271
|
+
api: "openai-completions";
|
|
8272
|
+
provider: string;
|
|
8273
|
+
baseUrl: string;
|
|
8274
|
+
reasoning: true;
|
|
8275
|
+
input: ("image" | "text")[];
|
|
8276
|
+
cost: {
|
|
8277
|
+
input: number;
|
|
8278
|
+
output: number;
|
|
8279
|
+
cacheRead: number;
|
|
8280
|
+
cacheWrite: number;
|
|
8281
|
+
};
|
|
8282
|
+
contextWindow: number;
|
|
8283
|
+
maxTokens: number;
|
|
8284
|
+
};
|
|
8285
|
+
readonly "google/gemini-2.5-pro-preview": {
|
|
7789
8286
|
id: string;
|
|
7790
8287
|
name: string;
|
|
7791
8288
|
api: "openai-completions";
|
|
@@ -7802,7 +8299,7 @@ export declare const MODELS: {
|
|
|
7802
8299
|
contextWindow: number;
|
|
7803
8300
|
maxTokens: number;
|
|
7804
8301
|
};
|
|
7805
|
-
readonly "google/gemini-2.5-
|
|
8302
|
+
readonly "google/gemini-2.5-pro-preview-05-06": {
|
|
7806
8303
|
id: string;
|
|
7807
8304
|
name: string;
|
|
7808
8305
|
api: "openai-completions";
|
|
@@ -7819,7 +8316,7 @@ export declare const MODELS: {
|
|
|
7819
8316
|
contextWindow: number;
|
|
7820
8317
|
maxTokens: number;
|
|
7821
8318
|
};
|
|
7822
|
-
readonly "google/gemini-
|
|
8319
|
+
readonly "google/gemini-3-flash-preview": {
|
|
7823
8320
|
id: string;
|
|
7824
8321
|
name: string;
|
|
7825
8322
|
api: "openai-completions";
|
|
@@ -7836,7 +8333,7 @@ export declare const MODELS: {
|
|
|
7836
8333
|
contextWindow: number;
|
|
7837
8334
|
maxTokens: number;
|
|
7838
8335
|
};
|
|
7839
|
-
readonly "google/gemini-
|
|
8336
|
+
readonly "google/gemini-3.1-flash-lite-preview": {
|
|
7840
8337
|
id: string;
|
|
7841
8338
|
name: string;
|
|
7842
8339
|
api: "openai-completions";
|
|
@@ -7853,7 +8350,7 @@ export declare const MODELS: {
|
|
|
7853
8350
|
contextWindow: number;
|
|
7854
8351
|
maxTokens: number;
|
|
7855
8352
|
};
|
|
7856
|
-
readonly "google/gemini-
|
|
8353
|
+
readonly "google/gemini-3.1-pro-preview": {
|
|
7857
8354
|
id: string;
|
|
7858
8355
|
name: string;
|
|
7859
8356
|
api: "openai-completions";
|
|
@@ -7870,7 +8367,7 @@ export declare const MODELS: {
|
|
|
7870
8367
|
contextWindow: number;
|
|
7871
8368
|
maxTokens: number;
|
|
7872
8369
|
};
|
|
7873
|
-
readonly "google/gemini-3-
|
|
8370
|
+
readonly "google/gemini-3.1-pro-preview-customtools": {
|
|
7874
8371
|
id: string;
|
|
7875
8372
|
name: string;
|
|
7876
8373
|
api: "openai-completions";
|
|
@@ -7887,13 +8384,13 @@ export declare const MODELS: {
|
|
|
7887
8384
|
contextWindow: number;
|
|
7888
8385
|
maxTokens: number;
|
|
7889
8386
|
};
|
|
7890
|
-
readonly "google/
|
|
8387
|
+
readonly "google/gemma-3-12b-it": {
|
|
7891
8388
|
id: string;
|
|
7892
8389
|
name: string;
|
|
7893
8390
|
api: "openai-completions";
|
|
7894
8391
|
provider: string;
|
|
7895
8392
|
baseUrl: string;
|
|
7896
|
-
reasoning:
|
|
8393
|
+
reasoning: false;
|
|
7897
8394
|
input: ("image" | "text")[];
|
|
7898
8395
|
cost: {
|
|
7899
8396
|
input: number;
|
|
@@ -7904,13 +8401,13 @@ export declare const MODELS: {
|
|
|
7904
8401
|
contextWindow: number;
|
|
7905
8402
|
maxTokens: number;
|
|
7906
8403
|
};
|
|
7907
|
-
readonly "google/
|
|
8404
|
+
readonly "google/gemma-3-27b-it": {
|
|
7908
8405
|
id: string;
|
|
7909
8406
|
name: string;
|
|
7910
8407
|
api: "openai-completions";
|
|
7911
8408
|
provider: string;
|
|
7912
8409
|
baseUrl: string;
|
|
7913
|
-
reasoning:
|
|
8410
|
+
reasoning: false;
|
|
7914
8411
|
input: ("image" | "text")[];
|
|
7915
8412
|
cost: {
|
|
7916
8413
|
input: number;
|
|
@@ -7921,7 +8418,7 @@ export declare const MODELS: {
|
|
|
7921
8418
|
contextWindow: number;
|
|
7922
8419
|
maxTokens: number;
|
|
7923
8420
|
};
|
|
7924
|
-
readonly "google/
|
|
8421
|
+
readonly "google/gemma-4-26b-a4b-it": {
|
|
7925
8422
|
id: string;
|
|
7926
8423
|
name: string;
|
|
7927
8424
|
api: "openai-completions";
|
|
@@ -7938,7 +8435,7 @@ export declare const MODELS: {
|
|
|
7938
8435
|
contextWindow: number;
|
|
7939
8436
|
maxTokens: number;
|
|
7940
8437
|
};
|
|
7941
|
-
readonly "google/gemma-4-26b-a4b-it": {
|
|
8438
|
+
readonly "google/gemma-4-26b-a4b-it:free": {
|
|
7942
8439
|
id: string;
|
|
7943
8440
|
name: string;
|
|
7944
8441
|
api: "openai-completions";
|
|
@@ -7972,14 +8469,14 @@ export declare const MODELS: {
|
|
|
7972
8469
|
contextWindow: number;
|
|
7973
8470
|
maxTokens: number;
|
|
7974
8471
|
};
|
|
7975
|
-
readonly "
|
|
8472
|
+
readonly "google/gemma-4-31b-it:free": {
|
|
7976
8473
|
id: string;
|
|
7977
8474
|
name: string;
|
|
7978
8475
|
api: "openai-completions";
|
|
7979
8476
|
provider: string;
|
|
7980
8477
|
baseUrl: string;
|
|
7981
|
-
reasoning:
|
|
7982
|
-
input: "text"[];
|
|
8478
|
+
reasoning: true;
|
|
8479
|
+
input: ("image" | "text")[];
|
|
7983
8480
|
cost: {
|
|
7984
8481
|
input: number;
|
|
7985
8482
|
output: number;
|
|
@@ -8006,7 +8503,7 @@ export declare const MODELS: {
|
|
|
8006
8503
|
contextWindow: number;
|
|
8007
8504
|
maxTokens: number;
|
|
8008
8505
|
};
|
|
8009
|
-
readonly "
|
|
8506
|
+
readonly "inclusionai/ling-2.6-1t:free": {
|
|
8010
8507
|
id: string;
|
|
8011
8508
|
name: string;
|
|
8012
8509
|
api: "openai-completions";
|
|
@@ -8023,7 +8520,7 @@ export declare const MODELS: {
|
|
|
8023
8520
|
contextWindow: number;
|
|
8024
8521
|
maxTokens: number;
|
|
8025
8522
|
};
|
|
8026
|
-
readonly "
|
|
8523
|
+
readonly "inclusionai/ling-2.6-flash:free": {
|
|
8027
8524
|
id: string;
|
|
8028
8525
|
name: string;
|
|
8029
8526
|
api: "openai-completions";
|
|
@@ -8040,7 +8537,7 @@ export declare const MODELS: {
|
|
|
8040
8537
|
contextWindow: number;
|
|
8041
8538
|
maxTokens: number;
|
|
8042
8539
|
};
|
|
8043
|
-
readonly "
|
|
8540
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
8044
8541
|
id: string;
|
|
8045
8542
|
name: string;
|
|
8046
8543
|
api: "openai-completions";
|
|
@@ -8142,23 +8639,6 @@ export declare const MODELS: {
|
|
|
8142
8639
|
contextWindow: number;
|
|
8143
8640
|
maxTokens: number;
|
|
8144
8641
|
};
|
|
8145
|
-
readonly "meta-llama/llama-4-maverick": {
|
|
8146
|
-
id: string;
|
|
8147
|
-
name: string;
|
|
8148
|
-
api: "openai-completions";
|
|
8149
|
-
provider: string;
|
|
8150
|
-
baseUrl: string;
|
|
8151
|
-
reasoning: false;
|
|
8152
|
-
input: ("image" | "text")[];
|
|
8153
|
-
cost: {
|
|
8154
|
-
input: number;
|
|
8155
|
-
output: number;
|
|
8156
|
-
cacheRead: number;
|
|
8157
|
-
cacheWrite: number;
|
|
8158
|
-
};
|
|
8159
|
-
contextWindow: number;
|
|
8160
|
-
maxTokens: number;
|
|
8161
|
-
};
|
|
8162
8642
|
readonly "meta-llama/llama-4-scout": {
|
|
8163
8643
|
id: string;
|
|
8164
8644
|
name: string;
|
|
@@ -8720,6 +9200,23 @@ export declare const MODELS: {
|
|
|
8720
9200
|
contextWindow: number;
|
|
8721
9201
|
maxTokens: number;
|
|
8722
9202
|
};
|
|
9203
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
9204
|
+
id: string;
|
|
9205
|
+
name: string;
|
|
9206
|
+
api: "openai-completions";
|
|
9207
|
+
provider: string;
|
|
9208
|
+
baseUrl: string;
|
|
9209
|
+
reasoning: true;
|
|
9210
|
+
input: ("image" | "text")[];
|
|
9211
|
+
cost: {
|
|
9212
|
+
input: number;
|
|
9213
|
+
output: number;
|
|
9214
|
+
cacheRead: number;
|
|
9215
|
+
cacheWrite: number;
|
|
9216
|
+
};
|
|
9217
|
+
contextWindow: number;
|
|
9218
|
+
maxTokens: number;
|
|
9219
|
+
};
|
|
8723
9220
|
readonly "nex-agi/deepseek-v3.1-nex-n1": {
|
|
8724
9221
|
id: string;
|
|
8725
9222
|
name: string;
|
|
@@ -9196,23 +9693,6 @@ export declare const MODELS: {
|
|
|
9196
9693
|
contextWindow: number;
|
|
9197
9694
|
maxTokens: number;
|
|
9198
9695
|
};
|
|
9199
|
-
readonly "openai/gpt-4o:extended": {
|
|
9200
|
-
id: string;
|
|
9201
|
-
name: string;
|
|
9202
|
-
api: "openai-completions";
|
|
9203
|
-
provider: string;
|
|
9204
|
-
baseUrl: string;
|
|
9205
|
-
reasoning: false;
|
|
9206
|
-
input: ("image" | "text")[];
|
|
9207
|
-
cost: {
|
|
9208
|
-
input: number;
|
|
9209
|
-
output: number;
|
|
9210
|
-
cacheRead: number;
|
|
9211
|
-
cacheWrite: number;
|
|
9212
|
-
};
|
|
9213
|
-
contextWindow: number;
|
|
9214
|
-
maxTokens: number;
|
|
9215
|
-
};
|
|
9216
9696
|
readonly "openai/gpt-5": {
|
|
9217
9697
|
id: string;
|
|
9218
9698
|
name: string;
|
|
@@ -9247,7 +9727,7 @@ export declare const MODELS: {
|
|
|
9247
9727
|
contextWindow: number;
|
|
9248
9728
|
maxTokens: number;
|
|
9249
9729
|
};
|
|
9250
|
-
readonly "openai/gpt-5-
|
|
9730
|
+
readonly "openai/gpt-5-mini": {
|
|
9251
9731
|
id: string;
|
|
9252
9732
|
name: string;
|
|
9253
9733
|
api: "openai-completions";
|
|
@@ -9264,7 +9744,7 @@ export declare const MODELS: {
|
|
|
9264
9744
|
contextWindow: number;
|
|
9265
9745
|
maxTokens: number;
|
|
9266
9746
|
};
|
|
9267
|
-
readonly "openai/gpt-5-
|
|
9747
|
+
readonly "openai/gpt-5-nano": {
|
|
9268
9748
|
id: string;
|
|
9269
9749
|
name: string;
|
|
9270
9750
|
api: "openai-completions";
|
|
@@ -9281,7 +9761,7 @@ export declare const MODELS: {
|
|
|
9281
9761
|
contextWindow: number;
|
|
9282
9762
|
maxTokens: number;
|
|
9283
9763
|
};
|
|
9284
|
-
readonly "openai/gpt-5-
|
|
9764
|
+
readonly "openai/gpt-5-pro": {
|
|
9285
9765
|
id: string;
|
|
9286
9766
|
name: string;
|
|
9287
9767
|
api: "openai-completions";
|
|
@@ -9298,7 +9778,7 @@ export declare const MODELS: {
|
|
|
9298
9778
|
contextWindow: number;
|
|
9299
9779
|
maxTokens: number;
|
|
9300
9780
|
};
|
|
9301
|
-
readonly "openai/gpt-5
|
|
9781
|
+
readonly "openai/gpt-5.1": {
|
|
9302
9782
|
id: string;
|
|
9303
9783
|
name: string;
|
|
9304
9784
|
api: "openai-completions";
|
|
@@ -9315,13 +9795,13 @@ export declare const MODELS: {
|
|
|
9315
9795
|
contextWindow: number;
|
|
9316
9796
|
maxTokens: number;
|
|
9317
9797
|
};
|
|
9318
|
-
readonly "openai/gpt-5-
|
|
9798
|
+
readonly "openai/gpt-5.1-chat": {
|
|
9319
9799
|
id: string;
|
|
9320
9800
|
name: string;
|
|
9321
9801
|
api: "openai-completions";
|
|
9322
9802
|
provider: string;
|
|
9323
9803
|
baseUrl: string;
|
|
9324
|
-
reasoning:
|
|
9804
|
+
reasoning: false;
|
|
9325
9805
|
input: ("image" | "text")[];
|
|
9326
9806
|
cost: {
|
|
9327
9807
|
input: number;
|
|
@@ -9332,7 +9812,7 @@ export declare const MODELS: {
|
|
|
9332
9812
|
contextWindow: number;
|
|
9333
9813
|
maxTokens: number;
|
|
9334
9814
|
};
|
|
9335
|
-
readonly "openai/gpt-5.1": {
|
|
9815
|
+
readonly "openai/gpt-5.1-codex": {
|
|
9336
9816
|
id: string;
|
|
9337
9817
|
name: string;
|
|
9338
9818
|
api: "openai-completions";
|
|
@@ -9349,13 +9829,13 @@ export declare const MODELS: {
|
|
|
9349
9829
|
contextWindow: number;
|
|
9350
9830
|
maxTokens: number;
|
|
9351
9831
|
};
|
|
9352
|
-
readonly "openai/gpt-5.1-
|
|
9832
|
+
readonly "openai/gpt-5.1-codex-max": {
|
|
9353
9833
|
id: string;
|
|
9354
9834
|
name: string;
|
|
9355
9835
|
api: "openai-completions";
|
|
9356
9836
|
provider: string;
|
|
9357
9837
|
baseUrl: string;
|
|
9358
|
-
reasoning:
|
|
9838
|
+
reasoning: true;
|
|
9359
9839
|
input: ("image" | "text")[];
|
|
9360
9840
|
cost: {
|
|
9361
9841
|
input: number;
|
|
@@ -9366,7 +9846,7 @@ export declare const MODELS: {
|
|
|
9366
9846
|
contextWindow: number;
|
|
9367
9847
|
maxTokens: number;
|
|
9368
9848
|
};
|
|
9369
|
-
readonly "openai/gpt-5.1-codex": {
|
|
9849
|
+
readonly "openai/gpt-5.1-codex-mini": {
|
|
9370
9850
|
id: string;
|
|
9371
9851
|
name: string;
|
|
9372
9852
|
api: "openai-completions";
|
|
@@ -9383,7 +9863,7 @@ export declare const MODELS: {
|
|
|
9383
9863
|
contextWindow: number;
|
|
9384
9864
|
maxTokens: number;
|
|
9385
9865
|
};
|
|
9386
|
-
readonly "openai/gpt-5.
|
|
9866
|
+
readonly "openai/gpt-5.2": {
|
|
9387
9867
|
id: string;
|
|
9388
9868
|
name: string;
|
|
9389
9869
|
api: "openai-completions";
|
|
@@ -9400,13 +9880,13 @@ export declare const MODELS: {
|
|
|
9400
9880
|
contextWindow: number;
|
|
9401
9881
|
maxTokens: number;
|
|
9402
9882
|
};
|
|
9403
|
-
readonly "openai/gpt-5.
|
|
9883
|
+
readonly "openai/gpt-5.2-chat": {
|
|
9404
9884
|
id: string;
|
|
9405
9885
|
name: string;
|
|
9406
9886
|
api: "openai-completions";
|
|
9407
9887
|
provider: string;
|
|
9408
9888
|
baseUrl: string;
|
|
9409
|
-
reasoning:
|
|
9889
|
+
reasoning: false;
|
|
9410
9890
|
input: ("image" | "text")[];
|
|
9411
9891
|
cost: {
|
|
9412
9892
|
input: number;
|
|
@@ -9417,7 +9897,7 @@ export declare const MODELS: {
|
|
|
9417
9897
|
contextWindow: number;
|
|
9418
9898
|
maxTokens: number;
|
|
9419
9899
|
};
|
|
9420
|
-
readonly "openai/gpt-5.2": {
|
|
9900
|
+
readonly "openai/gpt-5.2-codex": {
|
|
9421
9901
|
id: string;
|
|
9422
9902
|
name: string;
|
|
9423
9903
|
api: "openai-completions";
|
|
@@ -9434,13 +9914,13 @@ export declare const MODELS: {
|
|
|
9434
9914
|
contextWindow: number;
|
|
9435
9915
|
maxTokens: number;
|
|
9436
9916
|
};
|
|
9437
|
-
readonly "openai/gpt-5.2-
|
|
9917
|
+
readonly "openai/gpt-5.2-pro": {
|
|
9438
9918
|
id: string;
|
|
9439
9919
|
name: string;
|
|
9440
9920
|
api: "openai-completions";
|
|
9441
9921
|
provider: string;
|
|
9442
9922
|
baseUrl: string;
|
|
9443
|
-
reasoning:
|
|
9923
|
+
reasoning: true;
|
|
9444
9924
|
input: ("image" | "text")[];
|
|
9445
9925
|
cost: {
|
|
9446
9926
|
input: number;
|
|
@@ -9451,13 +9931,13 @@ export declare const MODELS: {
|
|
|
9451
9931
|
contextWindow: number;
|
|
9452
9932
|
maxTokens: number;
|
|
9453
9933
|
};
|
|
9454
|
-
readonly "openai/gpt-5.
|
|
9934
|
+
readonly "openai/gpt-5.3-chat": {
|
|
9455
9935
|
id: string;
|
|
9456
9936
|
name: string;
|
|
9457
9937
|
api: "openai-completions";
|
|
9458
9938
|
provider: string;
|
|
9459
9939
|
baseUrl: string;
|
|
9460
|
-
reasoning:
|
|
9940
|
+
reasoning: false;
|
|
9461
9941
|
input: ("image" | "text")[];
|
|
9462
9942
|
cost: {
|
|
9463
9943
|
input: number;
|
|
@@ -9468,7 +9948,7 @@ export declare const MODELS: {
|
|
|
9468
9948
|
contextWindow: number;
|
|
9469
9949
|
maxTokens: number;
|
|
9470
9950
|
};
|
|
9471
|
-
readonly "openai/gpt-5.
|
|
9951
|
+
readonly "openai/gpt-5.3-codex": {
|
|
9472
9952
|
id: string;
|
|
9473
9953
|
name: string;
|
|
9474
9954
|
api: "openai-completions";
|
|
@@ -9485,13 +9965,13 @@ export declare const MODELS: {
|
|
|
9485
9965
|
contextWindow: number;
|
|
9486
9966
|
maxTokens: number;
|
|
9487
9967
|
};
|
|
9488
|
-
readonly "openai/gpt-5.
|
|
9968
|
+
readonly "openai/gpt-5.4": {
|
|
9489
9969
|
id: string;
|
|
9490
9970
|
name: string;
|
|
9491
9971
|
api: "openai-completions";
|
|
9492
9972
|
provider: string;
|
|
9493
9973
|
baseUrl: string;
|
|
9494
|
-
reasoning:
|
|
9974
|
+
reasoning: true;
|
|
9495
9975
|
input: ("image" | "text")[];
|
|
9496
9976
|
cost: {
|
|
9497
9977
|
input: number;
|
|
@@ -9502,7 +9982,7 @@ export declare const MODELS: {
|
|
|
9502
9982
|
contextWindow: number;
|
|
9503
9983
|
maxTokens: number;
|
|
9504
9984
|
};
|
|
9505
|
-
readonly "openai/gpt-5.
|
|
9985
|
+
readonly "openai/gpt-5.4-mini": {
|
|
9506
9986
|
id: string;
|
|
9507
9987
|
name: string;
|
|
9508
9988
|
api: "openai-completions";
|
|
@@ -9519,7 +9999,7 @@ export declare const MODELS: {
|
|
|
9519
9999
|
contextWindow: number;
|
|
9520
10000
|
maxTokens: number;
|
|
9521
10001
|
};
|
|
9522
|
-
readonly "openai/gpt-5.4": {
|
|
10002
|
+
readonly "openai/gpt-5.4-nano": {
|
|
9523
10003
|
id: string;
|
|
9524
10004
|
name: string;
|
|
9525
10005
|
api: "openai-completions";
|
|
@@ -9536,7 +10016,7 @@ export declare const MODELS: {
|
|
|
9536
10016
|
contextWindow: number;
|
|
9537
10017
|
maxTokens: number;
|
|
9538
10018
|
};
|
|
9539
|
-
readonly "openai/gpt-5.4-
|
|
10019
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9540
10020
|
id: string;
|
|
9541
10021
|
name: string;
|
|
9542
10022
|
api: "openai-completions";
|
|
@@ -9553,7 +10033,7 @@ export declare const MODELS: {
|
|
|
9553
10033
|
contextWindow: number;
|
|
9554
10034
|
maxTokens: number;
|
|
9555
10035
|
};
|
|
9556
|
-
readonly "openai/gpt-5.
|
|
10036
|
+
readonly "openai/gpt-5.5": {
|
|
9557
10037
|
id: string;
|
|
9558
10038
|
name: string;
|
|
9559
10039
|
api: "openai-completions";
|
|
@@ -9570,7 +10050,7 @@ export declare const MODELS: {
|
|
|
9570
10050
|
contextWindow: number;
|
|
9571
10051
|
maxTokens: number;
|
|
9572
10052
|
};
|
|
9573
|
-
readonly "openai/gpt-5.
|
|
10053
|
+
readonly "openai/gpt-5.5-pro": {
|
|
9574
10054
|
id: string;
|
|
9575
10055
|
name: string;
|
|
9576
10056
|
api: "openai-completions";
|
|
@@ -10086,7 +10566,7 @@ export declare const MODELS: {
|
|
|
10086
10566
|
api: "openai-completions";
|
|
10087
10567
|
provider: string;
|
|
10088
10568
|
baseUrl: string;
|
|
10089
|
-
reasoning:
|
|
10569
|
+
reasoning: false;
|
|
10090
10570
|
input: "text"[];
|
|
10091
10571
|
cost: {
|
|
10092
10572
|
input: number;
|
|
@@ -10624,7 +11104,7 @@ export declare const MODELS: {
|
|
|
10624
11104
|
contextWindow: number;
|
|
10625
11105
|
maxTokens: number;
|
|
10626
11106
|
};
|
|
10627
|
-
readonly "qwen/qwen3.6-plus
|
|
11107
|
+
readonly "qwen/qwen3.6-plus": {
|
|
10628
11108
|
id: string;
|
|
10629
11109
|
name: string;
|
|
10630
11110
|
api: "openai-completions";
|
|
@@ -10743,7 +11223,7 @@ export declare const MODELS: {
|
|
|
10743
11223
|
contextWindow: number;
|
|
10744
11224
|
maxTokens: number;
|
|
10745
11225
|
};
|
|
10746
|
-
readonly "
|
|
11226
|
+
readonly "tencent/hy3-preview:free": {
|
|
10747
11227
|
id: string;
|
|
10748
11228
|
name: string;
|
|
10749
11229
|
api: "openai-completions";
|
|
@@ -11032,6 +11512,40 @@ export declare const MODELS: {
|
|
|
11032
11512
|
contextWindow: number;
|
|
11033
11513
|
maxTokens: number;
|
|
11034
11514
|
};
|
|
11515
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
11516
|
+
id: string;
|
|
11517
|
+
name: string;
|
|
11518
|
+
api: "openai-completions";
|
|
11519
|
+
provider: string;
|
|
11520
|
+
baseUrl: string;
|
|
11521
|
+
reasoning: true;
|
|
11522
|
+
input: ("image" | "text")[];
|
|
11523
|
+
cost: {
|
|
11524
|
+
input: number;
|
|
11525
|
+
output: number;
|
|
11526
|
+
cacheRead: number;
|
|
11527
|
+
cacheWrite: number;
|
|
11528
|
+
};
|
|
11529
|
+
contextWindow: number;
|
|
11530
|
+
maxTokens: number;
|
|
11531
|
+
};
|
|
11532
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
11533
|
+
id: string;
|
|
11534
|
+
name: string;
|
|
11535
|
+
api: "openai-completions";
|
|
11536
|
+
provider: string;
|
|
11537
|
+
baseUrl: string;
|
|
11538
|
+
reasoning: true;
|
|
11539
|
+
input: "text"[];
|
|
11540
|
+
cost: {
|
|
11541
|
+
input: number;
|
|
11542
|
+
output: number;
|
|
11543
|
+
cacheRead: number;
|
|
11544
|
+
cacheWrite: number;
|
|
11545
|
+
};
|
|
11546
|
+
contextWindow: number;
|
|
11547
|
+
maxTokens: number;
|
|
11548
|
+
};
|
|
11035
11549
|
readonly "z-ai/glm-4-32b": {
|
|
11036
11550
|
id: string;
|
|
11037
11551
|
name: string;
|
|
@@ -11219,6 +11733,23 @@ export declare const MODELS: {
|
|
|
11219
11733
|
contextWindow: number;
|
|
11220
11734
|
maxTokens: number;
|
|
11221
11735
|
};
|
|
11736
|
+
readonly "z-ai/glm-5.1": {
|
|
11737
|
+
id: string;
|
|
11738
|
+
name: string;
|
|
11739
|
+
api: "openai-completions";
|
|
11740
|
+
provider: string;
|
|
11741
|
+
baseUrl: string;
|
|
11742
|
+
reasoning: true;
|
|
11743
|
+
input: "text"[];
|
|
11744
|
+
cost: {
|
|
11745
|
+
input: number;
|
|
11746
|
+
output: number;
|
|
11747
|
+
cacheRead: number;
|
|
11748
|
+
cacheWrite: number;
|
|
11749
|
+
};
|
|
11750
|
+
contextWindow: number;
|
|
11751
|
+
maxTokens: number;
|
|
11752
|
+
};
|
|
11222
11753
|
readonly "z-ai/glm-5v-turbo": {
|
|
11223
11754
|
id: string;
|
|
11224
11755
|
name: string;
|
|
@@ -11236,6 +11767,23 @@ export declare const MODELS: {
|
|
|
11236
11767
|
contextWindow: number;
|
|
11237
11768
|
maxTokens: number;
|
|
11238
11769
|
};
|
|
11770
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
11771
|
+
id: string;
|
|
11772
|
+
name: string;
|
|
11773
|
+
api: "openai-completions";
|
|
11774
|
+
provider: string;
|
|
11775
|
+
baseUrl: string;
|
|
11776
|
+
reasoning: true;
|
|
11777
|
+
input: ("image" | "text")[];
|
|
11778
|
+
cost: {
|
|
11779
|
+
input: number;
|
|
11780
|
+
output: number;
|
|
11781
|
+
cacheRead: number;
|
|
11782
|
+
cacheWrite: number;
|
|
11783
|
+
};
|
|
11784
|
+
contextWindow: number;
|
|
11785
|
+
maxTokens: number;
|
|
11786
|
+
};
|
|
11239
11787
|
};
|
|
11240
11788
|
readonly "vercel-ai-gateway": {
|
|
11241
11789
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -11306,6 +11854,23 @@ export declare const MODELS: {
|
|
|
11306
11854
|
contextWindow: number;
|
|
11307
11855
|
maxTokens: number;
|
|
11308
11856
|
};
|
|
11857
|
+
readonly "alibaba/qwen-3.6-max-preview": {
|
|
11858
|
+
id: string;
|
|
11859
|
+
name: string;
|
|
11860
|
+
api: "anthropic-messages";
|
|
11861
|
+
provider: string;
|
|
11862
|
+
baseUrl: string;
|
|
11863
|
+
reasoning: true;
|
|
11864
|
+
input: ("image" | "text")[];
|
|
11865
|
+
cost: {
|
|
11866
|
+
input: number;
|
|
11867
|
+
output: number;
|
|
11868
|
+
cacheRead: number;
|
|
11869
|
+
cacheWrite: number;
|
|
11870
|
+
};
|
|
11871
|
+
contextWindow: number;
|
|
11872
|
+
maxTokens: number;
|
|
11873
|
+
};
|
|
11309
11874
|
readonly "alibaba/qwen3-235b-a22b-thinking": {
|
|
11310
11875
|
id: string;
|
|
11311
11876
|
name: string;
|
|
@@ -11646,6 +12211,23 @@ export declare const MODELS: {
|
|
|
11646
12211
|
contextWindow: number;
|
|
11647
12212
|
maxTokens: number;
|
|
11648
12213
|
};
|
|
12214
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
12215
|
+
id: string;
|
|
12216
|
+
name: string;
|
|
12217
|
+
api: "anthropic-messages";
|
|
12218
|
+
provider: string;
|
|
12219
|
+
baseUrl: string;
|
|
12220
|
+
reasoning: true;
|
|
12221
|
+
input: ("image" | "text")[];
|
|
12222
|
+
cost: {
|
|
12223
|
+
input: number;
|
|
12224
|
+
output: number;
|
|
12225
|
+
cacheRead: number;
|
|
12226
|
+
cacheWrite: number;
|
|
12227
|
+
};
|
|
12228
|
+
contextWindow: number;
|
|
12229
|
+
maxTokens: number;
|
|
12230
|
+
};
|
|
11649
12231
|
readonly "anthropic/claude-sonnet-4": {
|
|
11650
12232
|
id: string;
|
|
11651
12233
|
name: string;
|
|
@@ -11805,7 +12387,41 @@ export declare const MODELS: {
|
|
|
11805
12387
|
api: "anthropic-messages";
|
|
11806
12388
|
provider: string;
|
|
11807
12389
|
baseUrl: string;
|
|
11808
|
-
reasoning: true;
|
|
12390
|
+
reasoning: true;
|
|
12391
|
+
input: "text"[];
|
|
12392
|
+
cost: {
|
|
12393
|
+
input: number;
|
|
12394
|
+
output: number;
|
|
12395
|
+
cacheRead: number;
|
|
12396
|
+
cacheWrite: number;
|
|
12397
|
+
};
|
|
12398
|
+
contextWindow: number;
|
|
12399
|
+
maxTokens: number;
|
|
12400
|
+
};
|
|
12401
|
+
readonly "deepseek/deepseek-v3.1-terminus": {
|
|
12402
|
+
id: string;
|
|
12403
|
+
name: string;
|
|
12404
|
+
api: "anthropic-messages";
|
|
12405
|
+
provider: string;
|
|
12406
|
+
baseUrl: string;
|
|
12407
|
+
reasoning: true;
|
|
12408
|
+
input: "text"[];
|
|
12409
|
+
cost: {
|
|
12410
|
+
input: number;
|
|
12411
|
+
output: number;
|
|
12412
|
+
cacheRead: number;
|
|
12413
|
+
cacheWrite: number;
|
|
12414
|
+
};
|
|
12415
|
+
contextWindow: number;
|
|
12416
|
+
maxTokens: number;
|
|
12417
|
+
};
|
|
12418
|
+
readonly "deepseek/deepseek-v3.2": {
|
|
12419
|
+
id: string;
|
|
12420
|
+
name: string;
|
|
12421
|
+
api: "anthropic-messages";
|
|
12422
|
+
provider: string;
|
|
12423
|
+
baseUrl: string;
|
|
12424
|
+
reasoning: false;
|
|
11809
12425
|
input: "text"[];
|
|
11810
12426
|
cost: {
|
|
11811
12427
|
input: number;
|
|
@@ -11816,7 +12432,7 @@ export declare const MODELS: {
|
|
|
11816
12432
|
contextWindow: number;
|
|
11817
12433
|
maxTokens: number;
|
|
11818
12434
|
};
|
|
11819
|
-
readonly "deepseek/deepseek-v3.
|
|
12435
|
+
readonly "deepseek/deepseek-v3.2-thinking": {
|
|
11820
12436
|
id: string;
|
|
11821
12437
|
name: string;
|
|
11822
12438
|
api: "anthropic-messages";
|
|
@@ -11833,13 +12449,13 @@ export declare const MODELS: {
|
|
|
11833
12449
|
contextWindow: number;
|
|
11834
12450
|
maxTokens: number;
|
|
11835
12451
|
};
|
|
11836
|
-
readonly "deepseek/deepseek-
|
|
12452
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
11837
12453
|
id: string;
|
|
11838
12454
|
name: string;
|
|
11839
12455
|
api: "anthropic-messages";
|
|
11840
12456
|
provider: string;
|
|
11841
12457
|
baseUrl: string;
|
|
11842
|
-
reasoning:
|
|
12458
|
+
reasoning: true;
|
|
11843
12459
|
input: "text"[];
|
|
11844
12460
|
cost: {
|
|
11845
12461
|
input: number;
|
|
@@ -11850,7 +12466,7 @@ export declare const MODELS: {
|
|
|
11850
12466
|
contextWindow: number;
|
|
11851
12467
|
maxTokens: number;
|
|
11852
12468
|
};
|
|
11853
|
-
readonly "deepseek/deepseek-
|
|
12469
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
11854
12470
|
id: string;
|
|
11855
12471
|
name: string;
|
|
11856
12472
|
api: "anthropic-messages";
|
|
@@ -12632,6 +13248,23 @@ export declare const MODELS: {
|
|
|
12632
13248
|
contextWindow: number;
|
|
12633
13249
|
maxTokens: number;
|
|
12634
13250
|
};
|
|
13251
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
13252
|
+
id: string;
|
|
13253
|
+
name: string;
|
|
13254
|
+
api: "anthropic-messages";
|
|
13255
|
+
provider: string;
|
|
13256
|
+
baseUrl: string;
|
|
13257
|
+
reasoning: true;
|
|
13258
|
+
input: ("image" | "text")[];
|
|
13259
|
+
cost: {
|
|
13260
|
+
input: number;
|
|
13261
|
+
output: number;
|
|
13262
|
+
cacheRead: number;
|
|
13263
|
+
cacheWrite: number;
|
|
13264
|
+
};
|
|
13265
|
+
contextWindow: number;
|
|
13266
|
+
maxTokens: number;
|
|
13267
|
+
};
|
|
12635
13268
|
readonly "nvidia/nemotron-nano-12b-v2-vl": {
|
|
12636
13269
|
id: string;
|
|
12637
13270
|
name: string;
|
|
@@ -13125,14 +13758,31 @@ export declare const MODELS: {
|
|
|
13125
13758
|
contextWindow: number;
|
|
13126
13759
|
maxTokens: number;
|
|
13127
13760
|
};
|
|
13128
|
-
readonly "openai/gpt-
|
|
13761
|
+
readonly "openai/gpt-5.5": {
|
|
13129
13762
|
id: string;
|
|
13130
13763
|
name: string;
|
|
13131
13764
|
api: "anthropic-messages";
|
|
13132
13765
|
provider: string;
|
|
13133
13766
|
baseUrl: string;
|
|
13134
13767
|
reasoning: true;
|
|
13135
|
-
input: "text"[];
|
|
13768
|
+
input: ("image" | "text")[];
|
|
13769
|
+
cost: {
|
|
13770
|
+
input: number;
|
|
13771
|
+
output: number;
|
|
13772
|
+
cacheRead: number;
|
|
13773
|
+
cacheWrite: number;
|
|
13774
|
+
};
|
|
13775
|
+
contextWindow: number;
|
|
13776
|
+
maxTokens: number;
|
|
13777
|
+
};
|
|
13778
|
+
readonly "openai/gpt-5.5-pro": {
|
|
13779
|
+
id: string;
|
|
13780
|
+
name: string;
|
|
13781
|
+
api: "anthropic-messages";
|
|
13782
|
+
provider: string;
|
|
13783
|
+
baseUrl: string;
|
|
13784
|
+
reasoning: true;
|
|
13785
|
+
input: ("image" | "text")[];
|
|
13136
13786
|
cost: {
|
|
13137
13787
|
input: number;
|
|
13138
13788
|
output: number;
|
|
@@ -13329,23 +13979,6 @@ export declare const MODELS: {
|
|
|
13329
13979
|
contextWindow: number;
|
|
13330
13980
|
maxTokens: number;
|
|
13331
13981
|
};
|
|
13332
|
-
readonly "xai/grok-2-vision": {
|
|
13333
|
-
id: string;
|
|
13334
|
-
name: string;
|
|
13335
|
-
api: "anthropic-messages";
|
|
13336
|
-
provider: string;
|
|
13337
|
-
baseUrl: string;
|
|
13338
|
-
reasoning: false;
|
|
13339
|
-
input: ("image" | "text")[];
|
|
13340
|
-
cost: {
|
|
13341
|
-
input: number;
|
|
13342
|
-
output: number;
|
|
13343
|
-
cacheRead: number;
|
|
13344
|
-
cacheWrite: number;
|
|
13345
|
-
};
|
|
13346
|
-
contextWindow: number;
|
|
13347
|
-
maxTokens: number;
|
|
13348
|
-
};
|
|
13349
13982
|
readonly "xai/grok-3": {
|
|
13350
13983
|
id: string;
|
|
13351
13984
|
name: string;
|
|
@@ -13438,7 +14071,7 @@ export declare const MODELS: {
|
|
|
13438
14071
|
provider: string;
|
|
13439
14072
|
baseUrl: string;
|
|
13440
14073
|
reasoning: false;
|
|
13441
|
-
input: "text"[];
|
|
14074
|
+
input: ("image" | "text")[];
|
|
13442
14075
|
cost: {
|
|
13443
14076
|
input: number;
|
|
13444
14077
|
output: number;
|
|
@@ -13455,7 +14088,7 @@ export declare const MODELS: {
|
|
|
13455
14088
|
provider: string;
|
|
13456
14089
|
baseUrl: string;
|
|
13457
14090
|
reasoning: true;
|
|
13458
|
-
input: "text"[];
|
|
14091
|
+
input: ("image" | "text")[];
|
|
13459
14092
|
cost: {
|
|
13460
14093
|
input: number;
|
|
13461
14094
|
output: number;
|
|
@@ -13472,7 +14105,7 @@ export declare const MODELS: {
|
|
|
13472
14105
|
provider: string;
|
|
13473
14106
|
baseUrl: string;
|
|
13474
14107
|
reasoning: false;
|
|
13475
|
-
input: "text"[];
|
|
14108
|
+
input: ("image" | "text")[];
|
|
13476
14109
|
cost: {
|
|
13477
14110
|
input: number;
|
|
13478
14111
|
output: number;
|
|
@@ -13489,7 +14122,7 @@ export declare const MODELS: {
|
|
|
13489
14122
|
provider: string;
|
|
13490
14123
|
baseUrl: string;
|
|
13491
14124
|
reasoning: true;
|
|
13492
|
-
input: "text"[];
|
|
14125
|
+
input: ("image" | "text")[];
|
|
13493
14126
|
cost: {
|
|
13494
14127
|
input: number;
|
|
13495
14128
|
output: number;
|
|
@@ -13506,7 +14139,7 @@ export declare const MODELS: {
|
|
|
13506
14139
|
provider: string;
|
|
13507
14140
|
baseUrl: string;
|
|
13508
14141
|
reasoning: true;
|
|
13509
|
-
input: "text"[];
|
|
14142
|
+
input: ("image" | "text")[];
|
|
13510
14143
|
cost: {
|
|
13511
14144
|
input: number;
|
|
13512
14145
|
output: number;
|
|
@@ -13523,7 +14156,7 @@ export declare const MODELS: {
|
|
|
13523
14156
|
provider: string;
|
|
13524
14157
|
baseUrl: string;
|
|
13525
14158
|
reasoning: true;
|
|
13526
|
-
input: "text"[];
|
|
14159
|
+
input: ("image" | "text")[];
|
|
13527
14160
|
cost: {
|
|
13528
14161
|
input: number;
|
|
13529
14162
|
output: number;
|
|
@@ -13839,6 +14472,23 @@ export declare const MODELS: {
|
|
|
13839
14472
|
contextWindow: number;
|
|
13840
14473
|
maxTokens: number;
|
|
13841
14474
|
};
|
|
14475
|
+
readonly "zai/glm-5.1": {
|
|
14476
|
+
id: string;
|
|
14477
|
+
name: string;
|
|
14478
|
+
api: "anthropic-messages";
|
|
14479
|
+
provider: string;
|
|
14480
|
+
baseUrl: string;
|
|
14481
|
+
reasoning: true;
|
|
14482
|
+
input: "text"[];
|
|
14483
|
+
cost: {
|
|
14484
|
+
input: number;
|
|
14485
|
+
output: number;
|
|
14486
|
+
cacheRead: number;
|
|
14487
|
+
cacheWrite: number;
|
|
14488
|
+
};
|
|
14489
|
+
contextWindow: number;
|
|
14490
|
+
maxTokens: number;
|
|
14491
|
+
};
|
|
13842
14492
|
readonly "zai/glm-5v-turbo": {
|
|
13843
14493
|
id: string;
|
|
13844
14494
|
name: string;
|
|
@@ -14268,27 +14918,6 @@ export declare const MODELS: {
|
|
|
14268
14918
|
};
|
|
14269
14919
|
};
|
|
14270
14920
|
readonly zai: {
|
|
14271
|
-
readonly "glm-4.5": {
|
|
14272
|
-
id: string;
|
|
14273
|
-
name: string;
|
|
14274
|
-
api: "openai-completions";
|
|
14275
|
-
provider: string;
|
|
14276
|
-
baseUrl: string;
|
|
14277
|
-
compat: {
|
|
14278
|
-
supportsDeveloperRole: false;
|
|
14279
|
-
thinkingFormat: "zai";
|
|
14280
|
-
};
|
|
14281
|
-
reasoning: true;
|
|
14282
|
-
input: "text"[];
|
|
14283
|
-
cost: {
|
|
14284
|
-
input: number;
|
|
14285
|
-
output: number;
|
|
14286
|
-
cacheRead: number;
|
|
14287
|
-
cacheWrite: number;
|
|
14288
|
-
};
|
|
14289
|
-
contextWindow: number;
|
|
14290
|
-
maxTokens: number;
|
|
14291
|
-
};
|
|
14292
14921
|
readonly "glm-4.5-air": {
|
|
14293
14922
|
id: string;
|
|
14294
14923
|
name: string;
|
|
@@ -14310,92 +14939,6 @@ export declare const MODELS: {
|
|
|
14310
14939
|
contextWindow: number;
|
|
14311
14940
|
maxTokens: number;
|
|
14312
14941
|
};
|
|
14313
|
-
readonly "glm-4.5-flash": {
|
|
14314
|
-
id: string;
|
|
14315
|
-
name: string;
|
|
14316
|
-
api: "openai-completions";
|
|
14317
|
-
provider: string;
|
|
14318
|
-
baseUrl: string;
|
|
14319
|
-
compat: {
|
|
14320
|
-
supportsDeveloperRole: false;
|
|
14321
|
-
thinkingFormat: "zai";
|
|
14322
|
-
};
|
|
14323
|
-
reasoning: true;
|
|
14324
|
-
input: "text"[];
|
|
14325
|
-
cost: {
|
|
14326
|
-
input: number;
|
|
14327
|
-
output: number;
|
|
14328
|
-
cacheRead: number;
|
|
14329
|
-
cacheWrite: number;
|
|
14330
|
-
};
|
|
14331
|
-
contextWindow: number;
|
|
14332
|
-
maxTokens: number;
|
|
14333
|
-
};
|
|
14334
|
-
readonly "glm-4.5v": {
|
|
14335
|
-
id: string;
|
|
14336
|
-
name: string;
|
|
14337
|
-
api: "openai-completions";
|
|
14338
|
-
provider: string;
|
|
14339
|
-
baseUrl: string;
|
|
14340
|
-
compat: {
|
|
14341
|
-
supportsDeveloperRole: false;
|
|
14342
|
-
thinkingFormat: "zai";
|
|
14343
|
-
};
|
|
14344
|
-
reasoning: true;
|
|
14345
|
-
input: ("image" | "text")[];
|
|
14346
|
-
cost: {
|
|
14347
|
-
input: number;
|
|
14348
|
-
output: number;
|
|
14349
|
-
cacheRead: number;
|
|
14350
|
-
cacheWrite: number;
|
|
14351
|
-
};
|
|
14352
|
-
contextWindow: number;
|
|
14353
|
-
maxTokens: number;
|
|
14354
|
-
};
|
|
14355
|
-
readonly "glm-4.6": {
|
|
14356
|
-
id: string;
|
|
14357
|
-
name: string;
|
|
14358
|
-
api: "openai-completions";
|
|
14359
|
-
provider: string;
|
|
14360
|
-
baseUrl: string;
|
|
14361
|
-
compat: {
|
|
14362
|
-
supportsDeveloperRole: false;
|
|
14363
|
-
thinkingFormat: "zai";
|
|
14364
|
-
zaiToolStream: true;
|
|
14365
|
-
};
|
|
14366
|
-
reasoning: true;
|
|
14367
|
-
input: "text"[];
|
|
14368
|
-
cost: {
|
|
14369
|
-
input: number;
|
|
14370
|
-
output: number;
|
|
14371
|
-
cacheRead: number;
|
|
14372
|
-
cacheWrite: number;
|
|
14373
|
-
};
|
|
14374
|
-
contextWindow: number;
|
|
14375
|
-
maxTokens: number;
|
|
14376
|
-
};
|
|
14377
|
-
readonly "glm-4.6v": {
|
|
14378
|
-
id: string;
|
|
14379
|
-
name: string;
|
|
14380
|
-
api: "openai-completions";
|
|
14381
|
-
provider: string;
|
|
14382
|
-
baseUrl: string;
|
|
14383
|
-
compat: {
|
|
14384
|
-
supportsDeveloperRole: false;
|
|
14385
|
-
thinkingFormat: "zai";
|
|
14386
|
-
zaiToolStream: true;
|
|
14387
|
-
};
|
|
14388
|
-
reasoning: true;
|
|
14389
|
-
input: ("image" | "text")[];
|
|
14390
|
-
cost: {
|
|
14391
|
-
input: number;
|
|
14392
|
-
output: number;
|
|
14393
|
-
cacheRead: number;
|
|
14394
|
-
cacheWrite: number;
|
|
14395
|
-
};
|
|
14396
|
-
contextWindow: number;
|
|
14397
|
-
maxTokens: number;
|
|
14398
|
-
};
|
|
14399
14942
|
readonly "glm-4.7": {
|
|
14400
14943
|
id: string;
|
|
14401
14944
|
name: string;
|
|
@@ -14418,72 +14961,6 @@ export declare const MODELS: {
|
|
|
14418
14961
|
contextWindow: number;
|
|
14419
14962
|
maxTokens: number;
|
|
14420
14963
|
};
|
|
14421
|
-
readonly "glm-4.7-flash": {
|
|
14422
|
-
id: string;
|
|
14423
|
-
name: string;
|
|
14424
|
-
api: "openai-completions";
|
|
14425
|
-
provider: string;
|
|
14426
|
-
baseUrl: string;
|
|
14427
|
-
compat: {
|
|
14428
|
-
supportsDeveloperRole: false;
|
|
14429
|
-
thinkingFormat: "zai";
|
|
14430
|
-
zaiToolStream: true;
|
|
14431
|
-
};
|
|
14432
|
-
reasoning: true;
|
|
14433
|
-
input: "text"[];
|
|
14434
|
-
cost: {
|
|
14435
|
-
input: number;
|
|
14436
|
-
output: number;
|
|
14437
|
-
cacheRead: number;
|
|
14438
|
-
cacheWrite: number;
|
|
14439
|
-
};
|
|
14440
|
-
contextWindow: number;
|
|
14441
|
-
maxTokens: number;
|
|
14442
|
-
};
|
|
14443
|
-
readonly "glm-4.7-flashx": {
|
|
14444
|
-
id: string;
|
|
14445
|
-
name: string;
|
|
14446
|
-
api: "openai-completions";
|
|
14447
|
-
provider: string;
|
|
14448
|
-
baseUrl: string;
|
|
14449
|
-
compat: {
|
|
14450
|
-
supportsDeveloperRole: false;
|
|
14451
|
-
thinkingFormat: "zai";
|
|
14452
|
-
zaiToolStream: true;
|
|
14453
|
-
};
|
|
14454
|
-
reasoning: true;
|
|
14455
|
-
input: "text"[];
|
|
14456
|
-
cost: {
|
|
14457
|
-
input: number;
|
|
14458
|
-
output: number;
|
|
14459
|
-
cacheRead: number;
|
|
14460
|
-
cacheWrite: number;
|
|
14461
|
-
};
|
|
14462
|
-
contextWindow: number;
|
|
14463
|
-
maxTokens: number;
|
|
14464
|
-
};
|
|
14465
|
-
readonly "glm-5": {
|
|
14466
|
-
id: string;
|
|
14467
|
-
name: string;
|
|
14468
|
-
api: "openai-completions";
|
|
14469
|
-
provider: string;
|
|
14470
|
-
baseUrl: string;
|
|
14471
|
-
compat: {
|
|
14472
|
-
supportsDeveloperRole: false;
|
|
14473
|
-
thinkingFormat: "zai";
|
|
14474
|
-
zaiToolStream: true;
|
|
14475
|
-
};
|
|
14476
|
-
reasoning: true;
|
|
14477
|
-
input: "text"[];
|
|
14478
|
-
cost: {
|
|
14479
|
-
input: number;
|
|
14480
|
-
output: number;
|
|
14481
|
-
cacheRead: number;
|
|
14482
|
-
cacheWrite: number;
|
|
14483
|
-
};
|
|
14484
|
-
contextWindow: number;
|
|
14485
|
-
maxTokens: number;
|
|
14486
|
-
};
|
|
14487
14964
|
readonly "glm-5-turbo": {
|
|
14488
14965
|
id: string;
|
|
14489
14966
|
name: string;
|
|
@@ -14506,7 +14983,7 @@ export declare const MODELS: {
|
|
|
14506
14983
|
contextWindow: number;
|
|
14507
14984
|
maxTokens: number;
|
|
14508
14985
|
};
|
|
14509
|
-
readonly "glm-
|
|
14986
|
+
readonly "glm-5.1": {
|
|
14510
14987
|
id: string;
|
|
14511
14988
|
name: string;
|
|
14512
14989
|
api: "openai-completions";
|
|
@@ -14518,7 +14995,7 @@ export declare const MODELS: {
|
|
|
14518
14995
|
zaiToolStream: true;
|
|
14519
14996
|
};
|
|
14520
14997
|
reasoning: true;
|
|
14521
|
-
input:
|
|
14998
|
+
input: "text"[];
|
|
14522
14999
|
cost: {
|
|
14523
15000
|
input: number;
|
|
14524
15001
|
output: number;
|