@draht/ai 2026.3.4 → 2026.3.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2282,6 +2282,40 @@ export declare const MODELS: {
2282
2282
  contextWindow: number;
2283
2283
  maxTokens: number;
2284
2284
  };
2285
+ readonly "gpt-5.4": {
2286
+ id: string;
2287
+ name: string;
2288
+ api: "azure-openai-responses";
2289
+ provider: string;
2290
+ baseUrl: string;
2291
+ reasoning: true;
2292
+ input: ("image" | "text")[];
2293
+ cost: {
2294
+ input: number;
2295
+ output: number;
2296
+ cacheRead: number;
2297
+ cacheWrite: number;
2298
+ };
2299
+ contextWindow: number;
2300
+ maxTokens: number;
2301
+ };
2302
+ readonly "gpt-5.4-pro": {
2303
+ id: string;
2304
+ name: string;
2305
+ api: "azure-openai-responses";
2306
+ provider: string;
2307
+ baseUrl: string;
2308
+ reasoning: true;
2309
+ input: ("image" | "text")[];
2310
+ cost: {
2311
+ input: number;
2312
+ output: number;
2313
+ cacheRead: number;
2314
+ cacheWrite: number;
2315
+ };
2316
+ contextWindow: number;
2317
+ maxTokens: number;
2318
+ };
2285
2319
  readonly o1: {
2286
2320
  id: string;
2287
2321
  name: string;
@@ -2980,6 +3014,52 @@ export declare const MODELS: {
2980
3014
  contextWindow: number;
2981
3015
  maxTokens: number;
2982
3016
  };
3017
+ readonly "gpt-5.3-codex": {
3018
+ id: string;
3019
+ name: string;
3020
+ api: "openai-responses";
3021
+ provider: string;
3022
+ baseUrl: string;
3023
+ headers: {
3024
+ "User-Agent": string;
3025
+ "Editor-Version": string;
3026
+ "Editor-Plugin-Version": string;
3027
+ "Copilot-Integration-Id": string;
3028
+ };
3029
+ reasoning: true;
3030
+ input: ("image" | "text")[];
3031
+ cost: {
3032
+ input: number;
3033
+ output: number;
3034
+ cacheRead: number;
3035
+ cacheWrite: number;
3036
+ };
3037
+ contextWindow: number;
3038
+ maxTokens: number;
3039
+ };
3040
+ readonly "gpt-5.4": {
3041
+ id: string;
3042
+ name: string;
3043
+ api: "openai-responses";
3044
+ provider: string;
3045
+ baseUrl: string;
3046
+ headers: {
3047
+ "User-Agent": string;
3048
+ "Editor-Version": string;
3049
+ "Editor-Plugin-Version": string;
3050
+ "Copilot-Integration-Id": string;
3051
+ };
3052
+ reasoning: true;
3053
+ input: ("image" | "text")[];
3054
+ cost: {
3055
+ input: number;
3056
+ output: number;
3057
+ cacheRead: number;
3058
+ cacheWrite: number;
3059
+ };
3060
+ contextWindow: number;
3061
+ maxTokens: number;
3062
+ };
2983
3063
  readonly "grok-code-fast-1": {
2984
3064
  id: string;
2985
3065
  name: string;
@@ -5566,6 +5646,40 @@ export declare const MODELS: {
5566
5646
  contextWindow: number;
5567
5647
  maxTokens: number;
5568
5648
  };
5649
+ readonly "gpt-5.4": {
5650
+ id: string;
5651
+ name: string;
5652
+ api: "openai-responses";
5653
+ provider: string;
5654
+ baseUrl: string;
5655
+ reasoning: true;
5656
+ input: ("image" | "text")[];
5657
+ cost: {
5658
+ input: number;
5659
+ output: number;
5660
+ cacheRead: number;
5661
+ cacheWrite: number;
5662
+ };
5663
+ contextWindow: number;
5664
+ maxTokens: number;
5665
+ };
5666
+ readonly "gpt-5.4-pro": {
5667
+ id: string;
5668
+ name: string;
5669
+ api: "openai-responses";
5670
+ provider: string;
5671
+ baseUrl: string;
5672
+ reasoning: true;
5673
+ input: ("image" | "text")[];
5674
+ cost: {
5675
+ input: number;
5676
+ output: number;
5677
+ cacheRead: number;
5678
+ cacheWrite: number;
5679
+ };
5680
+ contextWindow: number;
5681
+ maxTokens: number;
5682
+ };
5569
5683
  readonly o1: {
5570
5684
  id: string;
5571
5685
  name: string;
@@ -6250,10 +6364,27 @@ export declare const MODELS: {
6250
6364
  contextWindow: number;
6251
6365
  maxTokens: number;
6252
6366
  };
6253
- readonly "kimi-k2.5": {
6367
+ readonly "gpt-5.3-codex-spark": {
6254
6368
  id: string;
6255
6369
  name: string;
6256
- api: "openai-completions";
6370
+ api: "openai-responses";
6371
+ provider: string;
6372
+ baseUrl: string;
6373
+ reasoning: true;
6374
+ input: "text"[];
6375
+ cost: {
6376
+ input: number;
6377
+ output: number;
6378
+ cacheRead: number;
6379
+ cacheWrite: number;
6380
+ };
6381
+ contextWindow: number;
6382
+ maxTokens: number;
6383
+ };
6384
+ readonly "gpt-5.4": {
6385
+ id: string;
6386
+ name: string;
6387
+ api: "openai-responses";
6257
6388
  provider: string;
6258
6389
  baseUrl: string;
6259
6390
  reasoning: true;
@@ -6267,14 +6398,31 @@ export declare const MODELS: {
6267
6398
  contextWindow: number;
6268
6399
  maxTokens: number;
6269
6400
  };
6270
- readonly "minimax-m2.1": {
6401
+ readonly "gpt-5.4-pro": {
6402
+ id: string;
6403
+ name: string;
6404
+ api: "openai-responses";
6405
+ provider: string;
6406
+ baseUrl: string;
6407
+ reasoning: true;
6408
+ input: ("image" | "text")[];
6409
+ cost: {
6410
+ input: number;
6411
+ output: number;
6412
+ cacheRead: number;
6413
+ cacheWrite: number;
6414
+ };
6415
+ contextWindow: number;
6416
+ maxTokens: number;
6417
+ };
6418
+ readonly "kimi-k2.5": {
6271
6419
  id: string;
6272
6420
  name: string;
6273
6421
  api: "openai-completions";
6274
6422
  provider: string;
6275
6423
  baseUrl: string;
6276
6424
  reasoning: true;
6277
- input: "text"[];
6425
+ input: ("image" | "text")[];
6278
6426
  cost: {
6279
6427
  input: number;
6280
6428
  output: number;
@@ -6284,7 +6432,7 @@ export declare const MODELS: {
6284
6432
  contextWindow: number;
6285
6433
  maxTokens: number;
6286
6434
  };
6287
- readonly "minimax-m2.5": {
6435
+ readonly "minimax-m2.1": {
6288
6436
  id: string;
6289
6437
  name: string;
6290
6438
  api: "openai-completions";
@@ -6301,10 +6449,10 @@ export declare const MODELS: {
6301
6449
  contextWindow: number;
6302
6450
  maxTokens: number;
6303
6451
  };
6304
- readonly "minimax-m2.5-free": {
6452
+ readonly "minimax-m2.5": {
6305
6453
  id: string;
6306
6454
  name: string;
6307
- api: "anthropic-messages";
6455
+ api: "openai-completions";
6308
6456
  provider: string;
6309
6457
  baseUrl: string;
6310
6458
  reasoning: true;
@@ -6318,13 +6466,13 @@ export declare const MODELS: {
6318
6466
  contextWindow: number;
6319
6467
  maxTokens: number;
6320
6468
  };
6321
- readonly "trinity-large-preview-free": {
6469
+ readonly "minimax-m2.5-free": {
6322
6470
  id: string;
6323
6471
  name: string;
6324
- api: "openai-completions";
6472
+ api: "anthropic-messages";
6325
6473
  provider: string;
6326
6474
  baseUrl: string;
6327
- reasoning: false;
6475
+ reasoning: true;
6328
6476
  input: "text"[];
6329
6477
  cost: {
6330
6478
  input: number;
@@ -7393,6 +7541,23 @@ export declare const MODELS: {
7393
7541
  contextWindow: number;
7394
7542
  maxTokens: number;
7395
7543
  };
7544
+ readonly "inception/mercury-2": {
7545
+ id: string;
7546
+ name: string;
7547
+ api: "openai-completions";
7548
+ provider: string;
7549
+ baseUrl: string;
7550
+ reasoning: true;
7551
+ input: "text"[];
7552
+ cost: {
7553
+ input: number;
7554
+ output: number;
7555
+ cacheRead: number;
7556
+ cacheWrite: number;
7557
+ };
7558
+ contextWindow: number;
7559
+ maxTokens: number;
7560
+ };
7396
7561
  readonly "inception/mercury-coder": {
7397
7562
  id: string;
7398
7563
  name: string;
@@ -8889,6 +9054,40 @@ export declare const MODELS: {
8889
9054
  contextWindow: number;
8890
9055
  maxTokens: number;
8891
9056
  };
9057
+ readonly "openai/gpt-5.4": {
9058
+ id: string;
9059
+ name: string;
9060
+ api: "openai-completions";
9061
+ provider: string;
9062
+ baseUrl: string;
9063
+ reasoning: true;
9064
+ input: ("image" | "text")[];
9065
+ cost: {
9066
+ input: number;
9067
+ output: number;
9068
+ cacheRead: number;
9069
+ cacheWrite: number;
9070
+ };
9071
+ contextWindow: number;
9072
+ maxTokens: number;
9073
+ };
9074
+ readonly "openai/gpt-5.4-pro": {
9075
+ id: string;
9076
+ name: string;
9077
+ api: "openai-completions";
9078
+ provider: string;
9079
+ baseUrl: string;
9080
+ reasoning: true;
9081
+ input: ("image" | "text")[];
9082
+ cost: {
9083
+ input: number;
9084
+ output: number;
9085
+ cacheRead: number;
9086
+ cacheWrite: number;
9087
+ };
9088
+ contextWindow: number;
9089
+ maxTokens: number;
9090
+ };
8892
9091
  readonly "openai/gpt-oss-120b": {
8893
9092
  id: string;
8894
9093
  name: string;
@@ -12138,6 +12337,40 @@ export declare const MODELS: {
12138
12337
  contextWindow: number;
12139
12338
  maxTokens: number;
12140
12339
  };
12340
+ readonly "openai/gpt-5.4": {
12341
+ id: string;
12342
+ name: string;
12343
+ api: "anthropic-messages";
12344
+ provider: string;
12345
+ baseUrl: string;
12346
+ reasoning: true;
12347
+ input: ("image" | "text")[];
12348
+ cost: {
12349
+ input: number;
12350
+ output: number;
12351
+ cacheRead: number;
12352
+ cacheWrite: number;
12353
+ };
12354
+ contextWindow: number;
12355
+ maxTokens: number;
12356
+ };
12357
+ readonly "openai/gpt-5.4-pro": {
12358
+ id: string;
12359
+ name: string;
12360
+ api: "anthropic-messages";
12361
+ provider: string;
12362
+ baseUrl: string;
12363
+ reasoning: true;
12364
+ input: ("image" | "text")[];
12365
+ cost: {
12366
+ input: number;
12367
+ output: number;
12368
+ cacheRead: number;
12369
+ cacheWrite: number;
12370
+ };
12371
+ contextWindow: number;
12372
+ maxTokens: number;
12373
+ };
12141
12374
  readonly "openai/gpt-oss-120b": {
12142
12375
  id: string;
12143
12376
  name: string;