@draht/ai 2026.3.11-1 → 2026.3.14

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cli.js CHANGED
File without changes
@@ -85,40 +85,6 @@ export declare const MODELS: {
85
85
  contextWindow: number;
86
86
  maxTokens: number;
87
87
  };
88
- readonly "amazon.titan-text-express-v1": {
89
- id: string;
90
- name: string;
91
- api: "bedrock-converse-stream";
92
- provider: string;
93
- baseUrl: string;
94
- reasoning: false;
95
- input: "text"[];
96
- cost: {
97
- input: number;
98
- output: number;
99
- cacheRead: number;
100
- cacheWrite: number;
101
- };
102
- contextWindow: number;
103
- maxTokens: number;
104
- };
105
- readonly "amazon.titan-text-express-v1:0:8k": {
106
- id: string;
107
- name: string;
108
- api: "bedrock-converse-stream";
109
- provider: string;
110
- baseUrl: string;
111
- reasoning: false;
112
- input: "text"[];
113
- cost: {
114
- input: number;
115
- output: number;
116
- cacheRead: number;
117
- cacheWrite: number;
118
- };
119
- contextWindow: number;
120
- maxTokens: number;
121
- };
122
88
  readonly "anthropic.claude-3-5-haiku-20241022-v1:0": {
123
89
  id: string;
124
90
  name: string;
@@ -204,40 +170,6 @@ export declare const MODELS: {
204
170
  contextWindow: number;
205
171
  maxTokens: number;
206
172
  };
207
- readonly "anthropic.claude-3-opus-20240229-v1:0": {
208
- id: string;
209
- name: string;
210
- api: "bedrock-converse-stream";
211
- provider: string;
212
- baseUrl: string;
213
- reasoning: false;
214
- input: ("image" | "text")[];
215
- cost: {
216
- input: number;
217
- output: number;
218
- cacheRead: number;
219
- cacheWrite: number;
220
- };
221
- contextWindow: number;
222
- maxTokens: number;
223
- };
224
- readonly "anthropic.claude-3-sonnet-20240229-v1:0": {
225
- id: string;
226
- name: string;
227
- api: "bedrock-converse-stream";
228
- provider: string;
229
- baseUrl: string;
230
- reasoning: false;
231
- input: ("image" | "text")[];
232
- cost: {
233
- input: number;
234
- output: number;
235
- cacheRead: number;
236
- cacheWrite: number;
237
- };
238
- contextWindow: number;
239
- maxTokens: number;
240
- };
241
173
  readonly "anthropic.claude-haiku-4-5-20251001-v1:0": {
242
174
  id: string;
243
175
  name: string;
@@ -374,40 +306,6 @@ export declare const MODELS: {
374
306
  contextWindow: number;
375
307
  maxTokens: number;
376
308
  };
377
- readonly "cohere.command-r-plus-v1:0": {
378
- id: string;
379
- name: string;
380
- api: "bedrock-converse-stream";
381
- provider: string;
382
- baseUrl: string;
383
- reasoning: false;
384
- input: "text"[];
385
- cost: {
386
- input: number;
387
- output: number;
388
- cacheRead: number;
389
- cacheWrite: number;
390
- };
391
- contextWindow: number;
392
- maxTokens: number;
393
- };
394
- readonly "cohere.command-r-v1:0": {
395
- id: string;
396
- name: string;
397
- api: "bedrock-converse-stream";
398
- provider: string;
399
- baseUrl: string;
400
- reasoning: false;
401
- input: "text"[];
402
- cost: {
403
- input: number;
404
- output: number;
405
- cacheRead: number;
406
- cacheWrite: number;
407
- };
408
- contextWindow: number;
409
- maxTokens: number;
410
- };
411
309
  readonly "deepseek.r1-v1:0": {
412
310
  id: string;
413
311
  name: string;
@@ -442,7 +340,7 @@ export declare const MODELS: {
442
340
  contextWindow: number;
443
341
  maxTokens: number;
444
342
  };
445
- readonly "deepseek.v3.2-v1:0": {
343
+ readonly "deepseek.v3.2": {
446
344
  id: string;
447
345
  name: string;
448
346
  api: "bedrock-converse-stream";
@@ -697,6 +595,23 @@ export declare const MODELS: {
697
595
  contextWindow: number;
698
596
  maxTokens: number;
699
597
  };
598
+ readonly "meta.llama3-1-405b-instruct-v1:0": {
599
+ id: string;
600
+ name: string;
601
+ api: "bedrock-converse-stream";
602
+ provider: string;
603
+ baseUrl: string;
604
+ reasoning: false;
605
+ input: "text"[];
606
+ cost: {
607
+ input: number;
608
+ output: number;
609
+ cacheRead: number;
610
+ cacheWrite: number;
611
+ };
612
+ contextWindow: number;
613
+ maxTokens: number;
614
+ };
700
615
  readonly "meta.llama3-1-70b-instruct-v1:0": {
701
616
  id: string;
702
617
  name: string;
@@ -901,6 +816,23 @@ export declare const MODELS: {
901
816
  contextWindow: number;
902
817
  maxTokens: number;
903
818
  };
819
+ readonly "mistral.magistral-small-2509": {
820
+ id: string;
821
+ name: string;
822
+ api: "bedrock-converse-stream";
823
+ provider: string;
824
+ baseUrl: string;
825
+ reasoning: true;
826
+ input: ("image" | "text")[];
827
+ cost: {
828
+ input: number;
829
+ output: number;
830
+ cacheRead: number;
831
+ cacheWrite: number;
832
+ };
833
+ contextWindow: number;
834
+ maxTokens: number;
835
+ };
904
836
  readonly "mistral.ministral-3-14b-instruct": {
905
837
  id: string;
906
838
  name: string;
@@ -918,6 +850,23 @@ export declare const MODELS: {
918
850
  contextWindow: number;
919
851
  maxTokens: number;
920
852
  };
853
+ readonly "mistral.ministral-3-3b-instruct": {
854
+ id: string;
855
+ name: string;
856
+ api: "bedrock-converse-stream";
857
+ provider: string;
858
+ baseUrl: string;
859
+ reasoning: false;
860
+ input: ("image" | "text")[];
861
+ cost: {
862
+ input: number;
863
+ output: number;
864
+ cacheRead: number;
865
+ cacheWrite: number;
866
+ };
867
+ contextWindow: number;
868
+ maxTokens: number;
869
+ };
921
870
  readonly "mistral.ministral-3-8b-instruct": {
922
871
  id: string;
923
872
  name: string;
@@ -935,14 +884,31 @@ export declare const MODELS: {
935
884
  contextWindow: number;
936
885
  maxTokens: number;
937
886
  };
938
- readonly "mistral.mistral-large-2402-v1:0": {
887
+ readonly "mistral.mistral-large-3-675b-instruct": {
939
888
  id: string;
940
889
  name: string;
941
890
  api: "bedrock-converse-stream";
942
891
  provider: string;
943
892
  baseUrl: string;
944
893
  reasoning: false;
945
- input: "text"[];
894
+ input: ("image" | "text")[];
895
+ cost: {
896
+ input: number;
897
+ output: number;
898
+ cacheRead: number;
899
+ cacheWrite: number;
900
+ };
901
+ contextWindow: number;
902
+ maxTokens: number;
903
+ };
904
+ readonly "mistral.pixtral-large-2502-v1:0": {
905
+ id: string;
906
+ name: string;
907
+ api: "bedrock-converse-stream";
908
+ provider: string;
909
+ baseUrl: string;
910
+ reasoning: false;
911
+ input: ("image" | "text")[];
946
912
  cost: {
947
913
  input: number;
948
914
  output: number;
@@ -1037,6 +1003,23 @@ export declare const MODELS: {
1037
1003
  contextWindow: number;
1038
1004
  maxTokens: number;
1039
1005
  };
1006
+ readonly "nvidia.nemotron-nano-3-30b": {
1007
+ id: string;
1008
+ name: string;
1009
+ api: "bedrock-converse-stream";
1010
+ provider: string;
1011
+ baseUrl: string;
1012
+ reasoning: true;
1013
+ input: "text"[];
1014
+ cost: {
1015
+ input: number;
1016
+ output: number;
1017
+ cacheRead: number;
1018
+ cacheWrite: number;
1019
+ };
1020
+ contextWindow: number;
1021
+ maxTokens: number;
1022
+ };
1040
1023
  readonly "nvidia.nemotron-nano-9b-v2": {
1041
1024
  id: string;
1042
1025
  name: string;
@@ -6534,6 +6517,23 @@ export declare const MODELS: {
6534
6517
  contextWindow: number;
6535
6518
  maxTokens: number;
6536
6519
  };
6520
+ readonly "nemotron-3-super-free": {
6521
+ id: string;
6522
+ name: string;
6523
+ api: "openai-completions";
6524
+ provider: string;
6525
+ baseUrl: string;
6526
+ reasoning: true;
6527
+ input: "text"[];
6528
+ cost: {
6529
+ input: number;
6530
+ output: number;
6531
+ cacheRead: number;
6532
+ cacheWrite: number;
6533
+ };
6534
+ contextWindow: number;
6535
+ maxTokens: number;
6536
+ };
6537
6537
  };
6538
6538
  readonly "opencode-go": {
6539
6539
  readonly "glm-5": {
@@ -7677,23 +7677,6 @@ export declare const MODELS: {
7677
7677
  contextWindow: number;
7678
7678
  maxTokens: number;
7679
7679
  };
7680
- readonly "meta-llama/llama-3.1-405b-instruct": {
7681
- id: string;
7682
- name: string;
7683
- api: "openai-completions";
7684
- provider: string;
7685
- baseUrl: string;
7686
- reasoning: false;
7687
- input: "text"[];
7688
- cost: {
7689
- input: number;
7690
- output: number;
7691
- cacheRead: number;
7692
- cacheWrite: number;
7693
- };
7694
- contextWindow: number;
7695
- maxTokens: number;
7696
- };
7697
7680
  readonly "meta-llama/llama-3.1-70b-instruct": {
7698
7681
  id: string;
7699
7682
  name: string;
@@ -8408,6 +8391,23 @@ export declare const MODELS: {
8408
8391
  contextWindow: number;
8409
8392
  maxTokens: number;
8410
8393
  };
8394
+ readonly "nvidia/nemotron-3-super-120b-a12b:free": {
8395
+ id: string;
8396
+ name: string;
8397
+ api: "openai-completions";
8398
+ provider: string;
8399
+ baseUrl: string;
8400
+ reasoning: true;
8401
+ input: "text"[];
8402
+ cost: {
8403
+ input: number;
8404
+ output: number;
8405
+ cacheRead: number;
8406
+ cacheWrite: number;
8407
+ };
8408
+ contextWindow: number;
8409
+ maxTokens: number;
8410
+ };
8411
8411
  readonly "nvidia/nemotron-nano-12b-v2-vl:free": {
8412
8412
  id: string;
8413
8413
  name: string;
@@ -9394,6 +9394,40 @@ export declare const MODELS: {
9394
9394
  contextWindow: number;
9395
9395
  maxTokens: number;
9396
9396
  };
9397
+ readonly "openrouter/healer-alpha": {
9398
+ id: string;
9399
+ name: string;
9400
+ api: "openai-completions";
9401
+ provider: string;
9402
+ baseUrl: string;
9403
+ reasoning: true;
9404
+ input: ("image" | "text")[];
9405
+ cost: {
9406
+ input: number;
9407
+ output: number;
9408
+ cacheRead: number;
9409
+ cacheWrite: number;
9410
+ };
9411
+ contextWindow: number;
9412
+ maxTokens: number;
9413
+ };
9414
+ readonly "openrouter/hunter-alpha": {
9415
+ id: string;
9416
+ name: string;
9417
+ api: "openai-completions";
9418
+ provider: string;
9419
+ baseUrl: string;
9420
+ reasoning: true;
9421
+ input: "text"[];
9422
+ cost: {
9423
+ input: number;
9424
+ output: number;
9425
+ cacheRead: number;
9426
+ cacheWrite: number;
9427
+ };
9428
+ contextWindow: number;
9429
+ maxTokens: number;
9430
+ };
9397
9431
  readonly "prime-intellect/intellect-3": {
9398
9432
  id: string;
9399
9433
  name: string;
@@ -10431,6 +10465,23 @@ export declare const MODELS: {
10431
10465
  contextWindow: number;
10432
10466
  maxTokens: number;
10433
10467
  };
10468
+ readonly "x-ai/grok-4.20-beta": {
10469
+ id: string;
10470
+ name: string;
10471
+ api: "openai-completions";
10472
+ provider: string;
10473
+ baseUrl: string;
10474
+ reasoning: true;
10475
+ input: ("image" | "text")[];
10476
+ cost: {
10477
+ input: number;
10478
+ output: number;
10479
+ cacheRead: number;
10480
+ cacheWrite: number;
10481
+ };
10482
+ contextWindow: number;
10483
+ maxTokens: number;
10484
+ };
10434
10485
  readonly "x-ai/grok-code-fast-1": {
10435
10486
  id: string;
10436
10487
  name: string;
@@ -10790,6 +10841,23 @@ export declare const MODELS: {
10790
10841
  contextWindow: number;
10791
10842
  maxTokens: number;
10792
10843
  };
10844
+ readonly "alibaba/qwen3-max": {
10845
+ id: string;
10846
+ name: string;
10847
+ api: "anthropic-messages";
10848
+ provider: string;
10849
+ baseUrl: string;
10850
+ reasoning: false;
10851
+ input: "text"[];
10852
+ cost: {
10853
+ input: number;
10854
+ output: number;
10855
+ cacheRead: number;
10856
+ cacheWrite: number;
10857
+ };
10858
+ contextWindow: number;
10859
+ maxTokens: number;
10860
+ };
10793
10861
  readonly "alibaba/qwen3-max-preview": {
10794
10862
  id: string;
10795
10863
  name: string;
@@ -11147,6 +11215,23 @@ export declare const MODELS: {
11147
11215
  contextWindow: number;
11148
11216
  maxTokens: number;
11149
11217
  };
11218
+ readonly "deepseek/deepseek-r1": {
11219
+ id: string;
11220
+ name: string;
11221
+ api: "anthropic-messages";
11222
+ provider: string;
11223
+ baseUrl: string;
11224
+ reasoning: true;
11225
+ input: "text"[];
11226
+ cost: {
11227
+ input: number;
11228
+ output: number;
11229
+ cacheRead: number;
11230
+ cacheWrite: number;
11231
+ };
11232
+ contextWindow: number;
11233
+ maxTokens: number;
11234
+ };
11150
11235
  readonly "deepseek/deepseek-v3": {
11151
11236
  id: string;
11152
11237
  name: string;
@@ -11232,6 +11317,40 @@ export declare const MODELS: {
11232
11317
  contextWindow: number;
11233
11318
  maxTokens: number;
11234
11319
  };
11320
+ readonly "google/gemini-2.0-flash": {
11321
+ id: string;
11322
+ name: string;
11323
+ api: "anthropic-messages";
11324
+ provider: string;
11325
+ baseUrl: string;
11326
+ reasoning: false;
11327
+ input: ("image" | "text")[];
11328
+ cost: {
11329
+ input: number;
11330
+ output: number;
11331
+ cacheRead: number;
11332
+ cacheWrite: number;
11333
+ };
11334
+ contextWindow: number;
11335
+ maxTokens: number;
11336
+ };
11337
+ readonly "google/gemini-2.0-flash-lite": {
11338
+ id: string;
11339
+ name: string;
11340
+ api: "anthropic-messages";
11341
+ provider: string;
11342
+ baseUrl: string;
11343
+ reasoning: false;
11344
+ input: ("image" | "text")[];
11345
+ cost: {
11346
+ input: number;
11347
+ output: number;
11348
+ cacheRead: number;
11349
+ cacheWrite: number;
11350
+ };
11351
+ contextWindow: number;
11352
+ maxTokens: number;
11353
+ };
11235
11354
  readonly "google/gemini-2.5-flash": {
11236
11355
  id: string;
11237
11356
  name: string;
@@ -11239,7 +11358,7 @@ export declare const MODELS: {
11239
11358
  provider: string;
11240
11359
  baseUrl: string;
11241
11360
  reasoning: true;
11242
- input: "text"[];
11361
+ input: ("image" | "text")[];
11243
11362
  cost: {
11244
11363
  input: number;
11245
11364
  output: number;
@@ -11307,7 +11426,7 @@ export declare const MODELS: {
11307
11426
  provider: string;
11308
11427
  baseUrl: string;
11309
11428
  reasoning: true;
11310
- input: "text"[];
11429
+ input: ("image" | "text")[];
11311
11430
  cost: {
11312
11431
  input: number;
11313
11432
  output: number;
@@ -11844,6 +11963,23 @@ export declare const MODELS: {
11844
11963
  contextWindow: number;
11845
11964
  maxTokens: number;
11846
11965
  };
11966
+ readonly "moonshotai/kimi-k2-0905": {
11967
+ id: string;
11968
+ name: string;
11969
+ api: "anthropic-messages";
11970
+ provider: string;
11971
+ baseUrl: string;
11972
+ reasoning: false;
11973
+ input: "text"[];
11974
+ cost: {
11975
+ input: number;
11976
+ output: number;
11977
+ cacheRead: number;
11978
+ cacheWrite: number;
11979
+ };
11980
+ contextWindow: number;
11981
+ maxTokens: number;
11982
+ };
11847
11983
  readonly "moonshotai/kimi-k2-thinking": {
11848
11984
  id: string;
11849
11985
  name: string;
@@ -12089,7 +12225,7 @@ export declare const MODELS: {
12089
12225
  provider: string;
12090
12226
  baseUrl: string;
12091
12227
  reasoning: true;
12092
- input: ("image" | "text")[];
12228
+ input: "text"[];
12093
12229
  cost: {
12094
12230
  input: number;
12095
12231
  output: number;
@@ -12371,23 +12507,6 @@ export declare const MODELS: {
12371
12507
  contextWindow: number;
12372
12508
  maxTokens: number;
12373
12509
  };
12374
- readonly "openai/gpt-oss-120b": {
12375
- id: string;
12376
- name: string;
12377
- api: "anthropic-messages";
12378
- provider: string;
12379
- baseUrl: string;
12380
- reasoning: true;
12381
- input: "text"[];
12382
- cost: {
12383
- input: number;
12384
- output: number;
12385
- cacheRead: number;
12386
- cacheWrite: number;
12387
- };
12388
- contextWindow: number;
12389
- maxTokens: number;
12390
- };
12391
12510
  readonly "openai/gpt-oss-20b": {
12392
12511
  id: string;
12393
12512
  name: string;
@@ -12779,6 +12898,57 @@ export declare const MODELS: {
12779
12898
  contextWindow: number;
12780
12899
  maxTokens: number;
12781
12900
  };
12901
+ readonly "xai/grok-4.20-multi-agent-beta": {
12902
+ id: string;
12903
+ name: string;
12904
+ api: "anthropic-messages";
12905
+ provider: string;
12906
+ baseUrl: string;
12907
+ reasoning: true;
12908
+ input: "text"[];
12909
+ cost: {
12910
+ input: number;
12911
+ output: number;
12912
+ cacheRead: number;
12913
+ cacheWrite: number;
12914
+ };
12915
+ contextWindow: number;
12916
+ maxTokens: number;
12917
+ };
12918
+ readonly "xai/grok-4.20-non-reasoning-beta": {
12919
+ id: string;
12920
+ name: string;
12921
+ api: "anthropic-messages";
12922
+ provider: string;
12923
+ baseUrl: string;
12924
+ reasoning: false;
12925
+ input: ("image" | "text")[];
12926
+ cost: {
12927
+ input: number;
12928
+ output: number;
12929
+ cacheRead: number;
12930
+ cacheWrite: number;
12931
+ };
12932
+ contextWindow: number;
12933
+ maxTokens: number;
12934
+ };
12935
+ readonly "xai/grok-4.20-reasoning-beta": {
12936
+ id: string;
12937
+ name: string;
12938
+ api: "anthropic-messages";
12939
+ provider: string;
12940
+ baseUrl: string;
12941
+ reasoning: true;
12942
+ input: ("image" | "text")[];
12943
+ cost: {
12944
+ input: number;
12945
+ output: number;
12946
+ cacheRead: number;
12947
+ cacheWrite: number;
12948
+ };
12949
+ contextWindow: number;
12950
+ maxTokens: number;
12951
+ };
12782
12952
  readonly "xai/grok-code-fast-1": {
12783
12953
  id: string;
12784
12954
  name: string;
@@ -12853,7 +13023,7 @@ export declare const MODELS: {
12853
13023
  api: "anthropic-messages";
12854
13024
  provider: string;
12855
13025
  baseUrl: string;
12856
- reasoning: true;
13026
+ reasoning: false;
12857
13027
  input: ("image" | "text")[];
12858
13028
  cost: {
12859
13029
  input: number;
@@ -13308,7 +13478,7 @@ export declare const MODELS: {
13308
13478
  contextWindow: number;
13309
13479
  maxTokens: number;
13310
13480
  };
13311
- readonly "grok-4.20-experimental-beta-0304-non-reasoning": {
13481
+ readonly "grok-4.20-beta-latest-non-reasoning": {
13312
13482
  id: string;
13313
13483
  name: string;
13314
13484
  api: "openai-completions";
@@ -13325,7 +13495,7 @@ export declare const MODELS: {
13325
13495
  contextWindow: number;
13326
13496
  maxTokens: number;
13327
13497
  };
13328
- readonly "grok-4.20-experimental-beta-0304-reasoning": {
13498
+ readonly "grok-4.20-beta-latest-reasoning": {
13329
13499
  id: string;
13330
13500
  name: string;
13331
13501
  api: "openai-completions";
@@ -13342,7 +13512,7 @@ export declare const MODELS: {
13342
13512
  contextWindow: number;
13343
13513
  maxTokens: number;
13344
13514
  };
13345
- readonly "grok-4.20-multi-agent-experimental-beta-0304": {
13515
+ readonly "grok-4.20-multi-agent-beta-latest": {
13346
13516
  id: string;
13347
13517
  name: string;
13348
13518
  api: "openai-completions";