@draht/ai 2026.3.11-1 → 2026.3.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cli.js +0 -0
- package/dist/models.generated.d.ts +316 -146
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +422 -252
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
package/dist/cli.js
CHANGED
|
File without changes
|
|
@@ -85,40 +85,6 @@ export declare const MODELS: {
|
|
|
85
85
|
contextWindow: number;
|
|
86
86
|
maxTokens: number;
|
|
87
87
|
};
|
|
88
|
-
readonly "amazon.titan-text-express-v1": {
|
|
89
|
-
id: string;
|
|
90
|
-
name: string;
|
|
91
|
-
api: "bedrock-converse-stream";
|
|
92
|
-
provider: string;
|
|
93
|
-
baseUrl: string;
|
|
94
|
-
reasoning: false;
|
|
95
|
-
input: "text"[];
|
|
96
|
-
cost: {
|
|
97
|
-
input: number;
|
|
98
|
-
output: number;
|
|
99
|
-
cacheRead: number;
|
|
100
|
-
cacheWrite: number;
|
|
101
|
-
};
|
|
102
|
-
contextWindow: number;
|
|
103
|
-
maxTokens: number;
|
|
104
|
-
};
|
|
105
|
-
readonly "amazon.titan-text-express-v1:0:8k": {
|
|
106
|
-
id: string;
|
|
107
|
-
name: string;
|
|
108
|
-
api: "bedrock-converse-stream";
|
|
109
|
-
provider: string;
|
|
110
|
-
baseUrl: string;
|
|
111
|
-
reasoning: false;
|
|
112
|
-
input: "text"[];
|
|
113
|
-
cost: {
|
|
114
|
-
input: number;
|
|
115
|
-
output: number;
|
|
116
|
-
cacheRead: number;
|
|
117
|
-
cacheWrite: number;
|
|
118
|
-
};
|
|
119
|
-
contextWindow: number;
|
|
120
|
-
maxTokens: number;
|
|
121
|
-
};
|
|
122
88
|
readonly "anthropic.claude-3-5-haiku-20241022-v1:0": {
|
|
123
89
|
id: string;
|
|
124
90
|
name: string;
|
|
@@ -204,40 +170,6 @@ export declare const MODELS: {
|
|
|
204
170
|
contextWindow: number;
|
|
205
171
|
maxTokens: number;
|
|
206
172
|
};
|
|
207
|
-
readonly "anthropic.claude-3-opus-20240229-v1:0": {
|
|
208
|
-
id: string;
|
|
209
|
-
name: string;
|
|
210
|
-
api: "bedrock-converse-stream";
|
|
211
|
-
provider: string;
|
|
212
|
-
baseUrl: string;
|
|
213
|
-
reasoning: false;
|
|
214
|
-
input: ("image" | "text")[];
|
|
215
|
-
cost: {
|
|
216
|
-
input: number;
|
|
217
|
-
output: number;
|
|
218
|
-
cacheRead: number;
|
|
219
|
-
cacheWrite: number;
|
|
220
|
-
};
|
|
221
|
-
contextWindow: number;
|
|
222
|
-
maxTokens: number;
|
|
223
|
-
};
|
|
224
|
-
readonly "anthropic.claude-3-sonnet-20240229-v1:0": {
|
|
225
|
-
id: string;
|
|
226
|
-
name: string;
|
|
227
|
-
api: "bedrock-converse-stream";
|
|
228
|
-
provider: string;
|
|
229
|
-
baseUrl: string;
|
|
230
|
-
reasoning: false;
|
|
231
|
-
input: ("image" | "text")[];
|
|
232
|
-
cost: {
|
|
233
|
-
input: number;
|
|
234
|
-
output: number;
|
|
235
|
-
cacheRead: number;
|
|
236
|
-
cacheWrite: number;
|
|
237
|
-
};
|
|
238
|
-
contextWindow: number;
|
|
239
|
-
maxTokens: number;
|
|
240
|
-
};
|
|
241
173
|
readonly "anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
242
174
|
id: string;
|
|
243
175
|
name: string;
|
|
@@ -374,40 +306,6 @@ export declare const MODELS: {
|
|
|
374
306
|
contextWindow: number;
|
|
375
307
|
maxTokens: number;
|
|
376
308
|
};
|
|
377
|
-
readonly "cohere.command-r-plus-v1:0": {
|
|
378
|
-
id: string;
|
|
379
|
-
name: string;
|
|
380
|
-
api: "bedrock-converse-stream";
|
|
381
|
-
provider: string;
|
|
382
|
-
baseUrl: string;
|
|
383
|
-
reasoning: false;
|
|
384
|
-
input: "text"[];
|
|
385
|
-
cost: {
|
|
386
|
-
input: number;
|
|
387
|
-
output: number;
|
|
388
|
-
cacheRead: number;
|
|
389
|
-
cacheWrite: number;
|
|
390
|
-
};
|
|
391
|
-
contextWindow: number;
|
|
392
|
-
maxTokens: number;
|
|
393
|
-
};
|
|
394
|
-
readonly "cohere.command-r-v1:0": {
|
|
395
|
-
id: string;
|
|
396
|
-
name: string;
|
|
397
|
-
api: "bedrock-converse-stream";
|
|
398
|
-
provider: string;
|
|
399
|
-
baseUrl: string;
|
|
400
|
-
reasoning: false;
|
|
401
|
-
input: "text"[];
|
|
402
|
-
cost: {
|
|
403
|
-
input: number;
|
|
404
|
-
output: number;
|
|
405
|
-
cacheRead: number;
|
|
406
|
-
cacheWrite: number;
|
|
407
|
-
};
|
|
408
|
-
contextWindow: number;
|
|
409
|
-
maxTokens: number;
|
|
410
|
-
};
|
|
411
309
|
readonly "deepseek.r1-v1:0": {
|
|
412
310
|
id: string;
|
|
413
311
|
name: string;
|
|
@@ -442,7 +340,7 @@ export declare const MODELS: {
|
|
|
442
340
|
contextWindow: number;
|
|
443
341
|
maxTokens: number;
|
|
444
342
|
};
|
|
445
|
-
readonly "deepseek.v3.2
|
|
343
|
+
readonly "deepseek.v3.2": {
|
|
446
344
|
id: string;
|
|
447
345
|
name: string;
|
|
448
346
|
api: "bedrock-converse-stream";
|
|
@@ -697,6 +595,23 @@ export declare const MODELS: {
|
|
|
697
595
|
contextWindow: number;
|
|
698
596
|
maxTokens: number;
|
|
699
597
|
};
|
|
598
|
+
readonly "meta.llama3-1-405b-instruct-v1:0": {
|
|
599
|
+
id: string;
|
|
600
|
+
name: string;
|
|
601
|
+
api: "bedrock-converse-stream";
|
|
602
|
+
provider: string;
|
|
603
|
+
baseUrl: string;
|
|
604
|
+
reasoning: false;
|
|
605
|
+
input: "text"[];
|
|
606
|
+
cost: {
|
|
607
|
+
input: number;
|
|
608
|
+
output: number;
|
|
609
|
+
cacheRead: number;
|
|
610
|
+
cacheWrite: number;
|
|
611
|
+
};
|
|
612
|
+
contextWindow: number;
|
|
613
|
+
maxTokens: number;
|
|
614
|
+
};
|
|
700
615
|
readonly "meta.llama3-1-70b-instruct-v1:0": {
|
|
701
616
|
id: string;
|
|
702
617
|
name: string;
|
|
@@ -901,6 +816,23 @@ export declare const MODELS: {
|
|
|
901
816
|
contextWindow: number;
|
|
902
817
|
maxTokens: number;
|
|
903
818
|
};
|
|
819
|
+
readonly "mistral.magistral-small-2509": {
|
|
820
|
+
id: string;
|
|
821
|
+
name: string;
|
|
822
|
+
api: "bedrock-converse-stream";
|
|
823
|
+
provider: string;
|
|
824
|
+
baseUrl: string;
|
|
825
|
+
reasoning: true;
|
|
826
|
+
input: ("image" | "text")[];
|
|
827
|
+
cost: {
|
|
828
|
+
input: number;
|
|
829
|
+
output: number;
|
|
830
|
+
cacheRead: number;
|
|
831
|
+
cacheWrite: number;
|
|
832
|
+
};
|
|
833
|
+
contextWindow: number;
|
|
834
|
+
maxTokens: number;
|
|
835
|
+
};
|
|
904
836
|
readonly "mistral.ministral-3-14b-instruct": {
|
|
905
837
|
id: string;
|
|
906
838
|
name: string;
|
|
@@ -918,6 +850,23 @@ export declare const MODELS: {
|
|
|
918
850
|
contextWindow: number;
|
|
919
851
|
maxTokens: number;
|
|
920
852
|
};
|
|
853
|
+
readonly "mistral.ministral-3-3b-instruct": {
|
|
854
|
+
id: string;
|
|
855
|
+
name: string;
|
|
856
|
+
api: "bedrock-converse-stream";
|
|
857
|
+
provider: string;
|
|
858
|
+
baseUrl: string;
|
|
859
|
+
reasoning: false;
|
|
860
|
+
input: ("image" | "text")[];
|
|
861
|
+
cost: {
|
|
862
|
+
input: number;
|
|
863
|
+
output: number;
|
|
864
|
+
cacheRead: number;
|
|
865
|
+
cacheWrite: number;
|
|
866
|
+
};
|
|
867
|
+
contextWindow: number;
|
|
868
|
+
maxTokens: number;
|
|
869
|
+
};
|
|
921
870
|
readonly "mistral.ministral-3-8b-instruct": {
|
|
922
871
|
id: string;
|
|
923
872
|
name: string;
|
|
@@ -935,14 +884,31 @@ export declare const MODELS: {
|
|
|
935
884
|
contextWindow: number;
|
|
936
885
|
maxTokens: number;
|
|
937
886
|
};
|
|
938
|
-
readonly "mistral.mistral-large-
|
|
887
|
+
readonly "mistral.mistral-large-3-675b-instruct": {
|
|
939
888
|
id: string;
|
|
940
889
|
name: string;
|
|
941
890
|
api: "bedrock-converse-stream";
|
|
942
891
|
provider: string;
|
|
943
892
|
baseUrl: string;
|
|
944
893
|
reasoning: false;
|
|
945
|
-
input: "text"[];
|
|
894
|
+
input: ("image" | "text")[];
|
|
895
|
+
cost: {
|
|
896
|
+
input: number;
|
|
897
|
+
output: number;
|
|
898
|
+
cacheRead: number;
|
|
899
|
+
cacheWrite: number;
|
|
900
|
+
};
|
|
901
|
+
contextWindow: number;
|
|
902
|
+
maxTokens: number;
|
|
903
|
+
};
|
|
904
|
+
readonly "mistral.pixtral-large-2502-v1:0": {
|
|
905
|
+
id: string;
|
|
906
|
+
name: string;
|
|
907
|
+
api: "bedrock-converse-stream";
|
|
908
|
+
provider: string;
|
|
909
|
+
baseUrl: string;
|
|
910
|
+
reasoning: false;
|
|
911
|
+
input: ("image" | "text")[];
|
|
946
912
|
cost: {
|
|
947
913
|
input: number;
|
|
948
914
|
output: number;
|
|
@@ -1037,6 +1003,23 @@ export declare const MODELS: {
|
|
|
1037
1003
|
contextWindow: number;
|
|
1038
1004
|
maxTokens: number;
|
|
1039
1005
|
};
|
|
1006
|
+
readonly "nvidia.nemotron-nano-3-30b": {
|
|
1007
|
+
id: string;
|
|
1008
|
+
name: string;
|
|
1009
|
+
api: "bedrock-converse-stream";
|
|
1010
|
+
provider: string;
|
|
1011
|
+
baseUrl: string;
|
|
1012
|
+
reasoning: true;
|
|
1013
|
+
input: "text"[];
|
|
1014
|
+
cost: {
|
|
1015
|
+
input: number;
|
|
1016
|
+
output: number;
|
|
1017
|
+
cacheRead: number;
|
|
1018
|
+
cacheWrite: number;
|
|
1019
|
+
};
|
|
1020
|
+
contextWindow: number;
|
|
1021
|
+
maxTokens: number;
|
|
1022
|
+
};
|
|
1040
1023
|
readonly "nvidia.nemotron-nano-9b-v2": {
|
|
1041
1024
|
id: string;
|
|
1042
1025
|
name: string;
|
|
@@ -6534,6 +6517,23 @@ export declare const MODELS: {
|
|
|
6534
6517
|
contextWindow: number;
|
|
6535
6518
|
maxTokens: number;
|
|
6536
6519
|
};
|
|
6520
|
+
readonly "nemotron-3-super-free": {
|
|
6521
|
+
id: string;
|
|
6522
|
+
name: string;
|
|
6523
|
+
api: "openai-completions";
|
|
6524
|
+
provider: string;
|
|
6525
|
+
baseUrl: string;
|
|
6526
|
+
reasoning: true;
|
|
6527
|
+
input: "text"[];
|
|
6528
|
+
cost: {
|
|
6529
|
+
input: number;
|
|
6530
|
+
output: number;
|
|
6531
|
+
cacheRead: number;
|
|
6532
|
+
cacheWrite: number;
|
|
6533
|
+
};
|
|
6534
|
+
contextWindow: number;
|
|
6535
|
+
maxTokens: number;
|
|
6536
|
+
};
|
|
6537
6537
|
};
|
|
6538
6538
|
readonly "opencode-go": {
|
|
6539
6539
|
readonly "glm-5": {
|
|
@@ -7677,23 +7677,6 @@ export declare const MODELS: {
|
|
|
7677
7677
|
contextWindow: number;
|
|
7678
7678
|
maxTokens: number;
|
|
7679
7679
|
};
|
|
7680
|
-
readonly "meta-llama/llama-3.1-405b-instruct": {
|
|
7681
|
-
id: string;
|
|
7682
|
-
name: string;
|
|
7683
|
-
api: "openai-completions";
|
|
7684
|
-
provider: string;
|
|
7685
|
-
baseUrl: string;
|
|
7686
|
-
reasoning: false;
|
|
7687
|
-
input: "text"[];
|
|
7688
|
-
cost: {
|
|
7689
|
-
input: number;
|
|
7690
|
-
output: number;
|
|
7691
|
-
cacheRead: number;
|
|
7692
|
-
cacheWrite: number;
|
|
7693
|
-
};
|
|
7694
|
-
contextWindow: number;
|
|
7695
|
-
maxTokens: number;
|
|
7696
|
-
};
|
|
7697
7680
|
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
7698
7681
|
id: string;
|
|
7699
7682
|
name: string;
|
|
@@ -8408,6 +8391,23 @@ export declare const MODELS: {
|
|
|
8408
8391
|
contextWindow: number;
|
|
8409
8392
|
maxTokens: number;
|
|
8410
8393
|
};
|
|
8394
|
+
readonly "nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8395
|
+
id: string;
|
|
8396
|
+
name: string;
|
|
8397
|
+
api: "openai-completions";
|
|
8398
|
+
provider: string;
|
|
8399
|
+
baseUrl: string;
|
|
8400
|
+
reasoning: true;
|
|
8401
|
+
input: "text"[];
|
|
8402
|
+
cost: {
|
|
8403
|
+
input: number;
|
|
8404
|
+
output: number;
|
|
8405
|
+
cacheRead: number;
|
|
8406
|
+
cacheWrite: number;
|
|
8407
|
+
};
|
|
8408
|
+
contextWindow: number;
|
|
8409
|
+
maxTokens: number;
|
|
8410
|
+
};
|
|
8411
8411
|
readonly "nvidia/nemotron-nano-12b-v2-vl:free": {
|
|
8412
8412
|
id: string;
|
|
8413
8413
|
name: string;
|
|
@@ -9394,6 +9394,40 @@ export declare const MODELS: {
|
|
|
9394
9394
|
contextWindow: number;
|
|
9395
9395
|
maxTokens: number;
|
|
9396
9396
|
};
|
|
9397
|
+
readonly "openrouter/healer-alpha": {
|
|
9398
|
+
id: string;
|
|
9399
|
+
name: string;
|
|
9400
|
+
api: "openai-completions";
|
|
9401
|
+
provider: string;
|
|
9402
|
+
baseUrl: string;
|
|
9403
|
+
reasoning: true;
|
|
9404
|
+
input: ("image" | "text")[];
|
|
9405
|
+
cost: {
|
|
9406
|
+
input: number;
|
|
9407
|
+
output: number;
|
|
9408
|
+
cacheRead: number;
|
|
9409
|
+
cacheWrite: number;
|
|
9410
|
+
};
|
|
9411
|
+
contextWindow: number;
|
|
9412
|
+
maxTokens: number;
|
|
9413
|
+
};
|
|
9414
|
+
readonly "openrouter/hunter-alpha": {
|
|
9415
|
+
id: string;
|
|
9416
|
+
name: string;
|
|
9417
|
+
api: "openai-completions";
|
|
9418
|
+
provider: string;
|
|
9419
|
+
baseUrl: string;
|
|
9420
|
+
reasoning: true;
|
|
9421
|
+
input: "text"[];
|
|
9422
|
+
cost: {
|
|
9423
|
+
input: number;
|
|
9424
|
+
output: number;
|
|
9425
|
+
cacheRead: number;
|
|
9426
|
+
cacheWrite: number;
|
|
9427
|
+
};
|
|
9428
|
+
contextWindow: number;
|
|
9429
|
+
maxTokens: number;
|
|
9430
|
+
};
|
|
9397
9431
|
readonly "prime-intellect/intellect-3": {
|
|
9398
9432
|
id: string;
|
|
9399
9433
|
name: string;
|
|
@@ -10431,6 +10465,23 @@ export declare const MODELS: {
|
|
|
10431
10465
|
contextWindow: number;
|
|
10432
10466
|
maxTokens: number;
|
|
10433
10467
|
};
|
|
10468
|
+
readonly "x-ai/grok-4.20-beta": {
|
|
10469
|
+
id: string;
|
|
10470
|
+
name: string;
|
|
10471
|
+
api: "openai-completions";
|
|
10472
|
+
provider: string;
|
|
10473
|
+
baseUrl: string;
|
|
10474
|
+
reasoning: true;
|
|
10475
|
+
input: ("image" | "text")[];
|
|
10476
|
+
cost: {
|
|
10477
|
+
input: number;
|
|
10478
|
+
output: number;
|
|
10479
|
+
cacheRead: number;
|
|
10480
|
+
cacheWrite: number;
|
|
10481
|
+
};
|
|
10482
|
+
contextWindow: number;
|
|
10483
|
+
maxTokens: number;
|
|
10484
|
+
};
|
|
10434
10485
|
readonly "x-ai/grok-code-fast-1": {
|
|
10435
10486
|
id: string;
|
|
10436
10487
|
name: string;
|
|
@@ -10790,6 +10841,23 @@ export declare const MODELS: {
|
|
|
10790
10841
|
contextWindow: number;
|
|
10791
10842
|
maxTokens: number;
|
|
10792
10843
|
};
|
|
10844
|
+
readonly "alibaba/qwen3-max": {
|
|
10845
|
+
id: string;
|
|
10846
|
+
name: string;
|
|
10847
|
+
api: "anthropic-messages";
|
|
10848
|
+
provider: string;
|
|
10849
|
+
baseUrl: string;
|
|
10850
|
+
reasoning: false;
|
|
10851
|
+
input: "text"[];
|
|
10852
|
+
cost: {
|
|
10853
|
+
input: number;
|
|
10854
|
+
output: number;
|
|
10855
|
+
cacheRead: number;
|
|
10856
|
+
cacheWrite: number;
|
|
10857
|
+
};
|
|
10858
|
+
contextWindow: number;
|
|
10859
|
+
maxTokens: number;
|
|
10860
|
+
};
|
|
10793
10861
|
readonly "alibaba/qwen3-max-preview": {
|
|
10794
10862
|
id: string;
|
|
10795
10863
|
name: string;
|
|
@@ -11147,6 +11215,23 @@ export declare const MODELS: {
|
|
|
11147
11215
|
contextWindow: number;
|
|
11148
11216
|
maxTokens: number;
|
|
11149
11217
|
};
|
|
11218
|
+
readonly "deepseek/deepseek-r1": {
|
|
11219
|
+
id: string;
|
|
11220
|
+
name: string;
|
|
11221
|
+
api: "anthropic-messages";
|
|
11222
|
+
provider: string;
|
|
11223
|
+
baseUrl: string;
|
|
11224
|
+
reasoning: true;
|
|
11225
|
+
input: "text"[];
|
|
11226
|
+
cost: {
|
|
11227
|
+
input: number;
|
|
11228
|
+
output: number;
|
|
11229
|
+
cacheRead: number;
|
|
11230
|
+
cacheWrite: number;
|
|
11231
|
+
};
|
|
11232
|
+
contextWindow: number;
|
|
11233
|
+
maxTokens: number;
|
|
11234
|
+
};
|
|
11150
11235
|
readonly "deepseek/deepseek-v3": {
|
|
11151
11236
|
id: string;
|
|
11152
11237
|
name: string;
|
|
@@ -11232,6 +11317,40 @@ export declare const MODELS: {
|
|
|
11232
11317
|
contextWindow: number;
|
|
11233
11318
|
maxTokens: number;
|
|
11234
11319
|
};
|
|
11320
|
+
readonly "google/gemini-2.0-flash": {
|
|
11321
|
+
id: string;
|
|
11322
|
+
name: string;
|
|
11323
|
+
api: "anthropic-messages";
|
|
11324
|
+
provider: string;
|
|
11325
|
+
baseUrl: string;
|
|
11326
|
+
reasoning: false;
|
|
11327
|
+
input: ("image" | "text")[];
|
|
11328
|
+
cost: {
|
|
11329
|
+
input: number;
|
|
11330
|
+
output: number;
|
|
11331
|
+
cacheRead: number;
|
|
11332
|
+
cacheWrite: number;
|
|
11333
|
+
};
|
|
11334
|
+
contextWindow: number;
|
|
11335
|
+
maxTokens: number;
|
|
11336
|
+
};
|
|
11337
|
+
readonly "google/gemini-2.0-flash-lite": {
|
|
11338
|
+
id: string;
|
|
11339
|
+
name: string;
|
|
11340
|
+
api: "anthropic-messages";
|
|
11341
|
+
provider: string;
|
|
11342
|
+
baseUrl: string;
|
|
11343
|
+
reasoning: false;
|
|
11344
|
+
input: ("image" | "text")[];
|
|
11345
|
+
cost: {
|
|
11346
|
+
input: number;
|
|
11347
|
+
output: number;
|
|
11348
|
+
cacheRead: number;
|
|
11349
|
+
cacheWrite: number;
|
|
11350
|
+
};
|
|
11351
|
+
contextWindow: number;
|
|
11352
|
+
maxTokens: number;
|
|
11353
|
+
};
|
|
11235
11354
|
readonly "google/gemini-2.5-flash": {
|
|
11236
11355
|
id: string;
|
|
11237
11356
|
name: string;
|
|
@@ -11239,7 +11358,7 @@ export declare const MODELS: {
|
|
|
11239
11358
|
provider: string;
|
|
11240
11359
|
baseUrl: string;
|
|
11241
11360
|
reasoning: true;
|
|
11242
|
-
input: "text"[];
|
|
11361
|
+
input: ("image" | "text")[];
|
|
11243
11362
|
cost: {
|
|
11244
11363
|
input: number;
|
|
11245
11364
|
output: number;
|
|
@@ -11307,7 +11426,7 @@ export declare const MODELS: {
|
|
|
11307
11426
|
provider: string;
|
|
11308
11427
|
baseUrl: string;
|
|
11309
11428
|
reasoning: true;
|
|
11310
|
-
input: "text"[];
|
|
11429
|
+
input: ("image" | "text")[];
|
|
11311
11430
|
cost: {
|
|
11312
11431
|
input: number;
|
|
11313
11432
|
output: number;
|
|
@@ -11844,6 +11963,23 @@ export declare const MODELS: {
|
|
|
11844
11963
|
contextWindow: number;
|
|
11845
11964
|
maxTokens: number;
|
|
11846
11965
|
};
|
|
11966
|
+
readonly "moonshotai/kimi-k2-0905": {
|
|
11967
|
+
id: string;
|
|
11968
|
+
name: string;
|
|
11969
|
+
api: "anthropic-messages";
|
|
11970
|
+
provider: string;
|
|
11971
|
+
baseUrl: string;
|
|
11972
|
+
reasoning: false;
|
|
11973
|
+
input: "text"[];
|
|
11974
|
+
cost: {
|
|
11975
|
+
input: number;
|
|
11976
|
+
output: number;
|
|
11977
|
+
cacheRead: number;
|
|
11978
|
+
cacheWrite: number;
|
|
11979
|
+
};
|
|
11980
|
+
contextWindow: number;
|
|
11981
|
+
maxTokens: number;
|
|
11982
|
+
};
|
|
11847
11983
|
readonly "moonshotai/kimi-k2-thinking": {
|
|
11848
11984
|
id: string;
|
|
11849
11985
|
name: string;
|
|
@@ -12089,7 +12225,7 @@ export declare const MODELS: {
|
|
|
12089
12225
|
provider: string;
|
|
12090
12226
|
baseUrl: string;
|
|
12091
12227
|
reasoning: true;
|
|
12092
|
-
input:
|
|
12228
|
+
input: "text"[];
|
|
12093
12229
|
cost: {
|
|
12094
12230
|
input: number;
|
|
12095
12231
|
output: number;
|
|
@@ -12371,23 +12507,6 @@ export declare const MODELS: {
|
|
|
12371
12507
|
contextWindow: number;
|
|
12372
12508
|
maxTokens: number;
|
|
12373
12509
|
};
|
|
12374
|
-
readonly "openai/gpt-oss-120b": {
|
|
12375
|
-
id: string;
|
|
12376
|
-
name: string;
|
|
12377
|
-
api: "anthropic-messages";
|
|
12378
|
-
provider: string;
|
|
12379
|
-
baseUrl: string;
|
|
12380
|
-
reasoning: true;
|
|
12381
|
-
input: "text"[];
|
|
12382
|
-
cost: {
|
|
12383
|
-
input: number;
|
|
12384
|
-
output: number;
|
|
12385
|
-
cacheRead: number;
|
|
12386
|
-
cacheWrite: number;
|
|
12387
|
-
};
|
|
12388
|
-
contextWindow: number;
|
|
12389
|
-
maxTokens: number;
|
|
12390
|
-
};
|
|
12391
12510
|
readonly "openai/gpt-oss-20b": {
|
|
12392
12511
|
id: string;
|
|
12393
12512
|
name: string;
|
|
@@ -12779,6 +12898,57 @@ export declare const MODELS: {
|
|
|
12779
12898
|
contextWindow: number;
|
|
12780
12899
|
maxTokens: number;
|
|
12781
12900
|
};
|
|
12901
|
+
readonly "xai/grok-4.20-multi-agent-beta": {
|
|
12902
|
+
id: string;
|
|
12903
|
+
name: string;
|
|
12904
|
+
api: "anthropic-messages";
|
|
12905
|
+
provider: string;
|
|
12906
|
+
baseUrl: string;
|
|
12907
|
+
reasoning: true;
|
|
12908
|
+
input: "text"[];
|
|
12909
|
+
cost: {
|
|
12910
|
+
input: number;
|
|
12911
|
+
output: number;
|
|
12912
|
+
cacheRead: number;
|
|
12913
|
+
cacheWrite: number;
|
|
12914
|
+
};
|
|
12915
|
+
contextWindow: number;
|
|
12916
|
+
maxTokens: number;
|
|
12917
|
+
};
|
|
12918
|
+
readonly "xai/grok-4.20-non-reasoning-beta": {
|
|
12919
|
+
id: string;
|
|
12920
|
+
name: string;
|
|
12921
|
+
api: "anthropic-messages";
|
|
12922
|
+
provider: string;
|
|
12923
|
+
baseUrl: string;
|
|
12924
|
+
reasoning: false;
|
|
12925
|
+
input: ("image" | "text")[];
|
|
12926
|
+
cost: {
|
|
12927
|
+
input: number;
|
|
12928
|
+
output: number;
|
|
12929
|
+
cacheRead: number;
|
|
12930
|
+
cacheWrite: number;
|
|
12931
|
+
};
|
|
12932
|
+
contextWindow: number;
|
|
12933
|
+
maxTokens: number;
|
|
12934
|
+
};
|
|
12935
|
+
readonly "xai/grok-4.20-reasoning-beta": {
|
|
12936
|
+
id: string;
|
|
12937
|
+
name: string;
|
|
12938
|
+
api: "anthropic-messages";
|
|
12939
|
+
provider: string;
|
|
12940
|
+
baseUrl: string;
|
|
12941
|
+
reasoning: true;
|
|
12942
|
+
input: ("image" | "text")[];
|
|
12943
|
+
cost: {
|
|
12944
|
+
input: number;
|
|
12945
|
+
output: number;
|
|
12946
|
+
cacheRead: number;
|
|
12947
|
+
cacheWrite: number;
|
|
12948
|
+
};
|
|
12949
|
+
contextWindow: number;
|
|
12950
|
+
maxTokens: number;
|
|
12951
|
+
};
|
|
12782
12952
|
readonly "xai/grok-code-fast-1": {
|
|
12783
12953
|
id: string;
|
|
12784
12954
|
name: string;
|
|
@@ -12853,7 +13023,7 @@ export declare const MODELS: {
|
|
|
12853
13023
|
api: "anthropic-messages";
|
|
12854
13024
|
provider: string;
|
|
12855
13025
|
baseUrl: string;
|
|
12856
|
-
reasoning:
|
|
13026
|
+
reasoning: false;
|
|
12857
13027
|
input: ("image" | "text")[];
|
|
12858
13028
|
cost: {
|
|
12859
13029
|
input: number;
|
|
@@ -13308,7 +13478,7 @@ export declare const MODELS: {
|
|
|
13308
13478
|
contextWindow: number;
|
|
13309
13479
|
maxTokens: number;
|
|
13310
13480
|
};
|
|
13311
|
-
readonly "grok-4.20-
|
|
13481
|
+
readonly "grok-4.20-beta-latest-non-reasoning": {
|
|
13312
13482
|
id: string;
|
|
13313
13483
|
name: string;
|
|
13314
13484
|
api: "openai-completions";
|
|
@@ -13325,7 +13495,7 @@ export declare const MODELS: {
|
|
|
13325
13495
|
contextWindow: number;
|
|
13326
13496
|
maxTokens: number;
|
|
13327
13497
|
};
|
|
13328
|
-
readonly "grok-4.20-
|
|
13498
|
+
readonly "grok-4.20-beta-latest-reasoning": {
|
|
13329
13499
|
id: string;
|
|
13330
13500
|
name: string;
|
|
13331
13501
|
api: "openai-completions";
|
|
@@ -13342,7 +13512,7 @@ export declare const MODELS: {
|
|
|
13342
13512
|
contextWindow: number;
|
|
13343
13513
|
maxTokens: number;
|
|
13344
13514
|
};
|
|
13345
|
-
readonly "grok-4.20-multi-agent-
|
|
13515
|
+
readonly "grok-4.20-multi-agent-beta-latest": {
|
|
13346
13516
|
id: string;
|
|
13347
13517
|
name: string;
|
|
13348
13518
|
api: "openai-completions";
|