@draht/ai 2026.3.4 → 2026.3.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +243 -10
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +259 -36
- package/dist/models.generated.js.map +1 -1
- package/package.json +3 -3
|
@@ -2282,6 +2282,40 @@ export declare const MODELS: {
|
|
|
2282
2282
|
contextWindow: number;
|
|
2283
2283
|
maxTokens: number;
|
|
2284
2284
|
};
|
|
2285
|
+
readonly "gpt-5.4": {
|
|
2286
|
+
id: string;
|
|
2287
|
+
name: string;
|
|
2288
|
+
api: "azure-openai-responses";
|
|
2289
|
+
provider: string;
|
|
2290
|
+
baseUrl: string;
|
|
2291
|
+
reasoning: true;
|
|
2292
|
+
input: ("image" | "text")[];
|
|
2293
|
+
cost: {
|
|
2294
|
+
input: number;
|
|
2295
|
+
output: number;
|
|
2296
|
+
cacheRead: number;
|
|
2297
|
+
cacheWrite: number;
|
|
2298
|
+
};
|
|
2299
|
+
contextWindow: number;
|
|
2300
|
+
maxTokens: number;
|
|
2301
|
+
};
|
|
2302
|
+
readonly "gpt-5.4-pro": {
|
|
2303
|
+
id: string;
|
|
2304
|
+
name: string;
|
|
2305
|
+
api: "azure-openai-responses";
|
|
2306
|
+
provider: string;
|
|
2307
|
+
baseUrl: string;
|
|
2308
|
+
reasoning: true;
|
|
2309
|
+
input: ("image" | "text")[];
|
|
2310
|
+
cost: {
|
|
2311
|
+
input: number;
|
|
2312
|
+
output: number;
|
|
2313
|
+
cacheRead: number;
|
|
2314
|
+
cacheWrite: number;
|
|
2315
|
+
};
|
|
2316
|
+
contextWindow: number;
|
|
2317
|
+
maxTokens: number;
|
|
2318
|
+
};
|
|
2285
2319
|
readonly o1: {
|
|
2286
2320
|
id: string;
|
|
2287
2321
|
name: string;
|
|
@@ -2980,6 +3014,52 @@ export declare const MODELS: {
|
|
|
2980
3014
|
contextWindow: number;
|
|
2981
3015
|
maxTokens: number;
|
|
2982
3016
|
};
|
|
3017
|
+
readonly "gpt-5.3-codex": {
|
|
3018
|
+
id: string;
|
|
3019
|
+
name: string;
|
|
3020
|
+
api: "openai-responses";
|
|
3021
|
+
provider: string;
|
|
3022
|
+
baseUrl: string;
|
|
3023
|
+
headers: {
|
|
3024
|
+
"User-Agent": string;
|
|
3025
|
+
"Editor-Version": string;
|
|
3026
|
+
"Editor-Plugin-Version": string;
|
|
3027
|
+
"Copilot-Integration-Id": string;
|
|
3028
|
+
};
|
|
3029
|
+
reasoning: true;
|
|
3030
|
+
input: ("image" | "text")[];
|
|
3031
|
+
cost: {
|
|
3032
|
+
input: number;
|
|
3033
|
+
output: number;
|
|
3034
|
+
cacheRead: number;
|
|
3035
|
+
cacheWrite: number;
|
|
3036
|
+
};
|
|
3037
|
+
contextWindow: number;
|
|
3038
|
+
maxTokens: number;
|
|
3039
|
+
};
|
|
3040
|
+
readonly "gpt-5.4": {
|
|
3041
|
+
id: string;
|
|
3042
|
+
name: string;
|
|
3043
|
+
api: "openai-responses";
|
|
3044
|
+
provider: string;
|
|
3045
|
+
baseUrl: string;
|
|
3046
|
+
headers: {
|
|
3047
|
+
"User-Agent": string;
|
|
3048
|
+
"Editor-Version": string;
|
|
3049
|
+
"Editor-Plugin-Version": string;
|
|
3050
|
+
"Copilot-Integration-Id": string;
|
|
3051
|
+
};
|
|
3052
|
+
reasoning: true;
|
|
3053
|
+
input: ("image" | "text")[];
|
|
3054
|
+
cost: {
|
|
3055
|
+
input: number;
|
|
3056
|
+
output: number;
|
|
3057
|
+
cacheRead: number;
|
|
3058
|
+
cacheWrite: number;
|
|
3059
|
+
};
|
|
3060
|
+
contextWindow: number;
|
|
3061
|
+
maxTokens: number;
|
|
3062
|
+
};
|
|
2983
3063
|
readonly "grok-code-fast-1": {
|
|
2984
3064
|
id: string;
|
|
2985
3065
|
name: string;
|
|
@@ -5566,6 +5646,40 @@ export declare const MODELS: {
|
|
|
5566
5646
|
contextWindow: number;
|
|
5567
5647
|
maxTokens: number;
|
|
5568
5648
|
};
|
|
5649
|
+
readonly "gpt-5.4": {
|
|
5650
|
+
id: string;
|
|
5651
|
+
name: string;
|
|
5652
|
+
api: "openai-responses";
|
|
5653
|
+
provider: string;
|
|
5654
|
+
baseUrl: string;
|
|
5655
|
+
reasoning: true;
|
|
5656
|
+
input: ("image" | "text")[];
|
|
5657
|
+
cost: {
|
|
5658
|
+
input: number;
|
|
5659
|
+
output: number;
|
|
5660
|
+
cacheRead: number;
|
|
5661
|
+
cacheWrite: number;
|
|
5662
|
+
};
|
|
5663
|
+
contextWindow: number;
|
|
5664
|
+
maxTokens: number;
|
|
5665
|
+
};
|
|
5666
|
+
readonly "gpt-5.4-pro": {
|
|
5667
|
+
id: string;
|
|
5668
|
+
name: string;
|
|
5669
|
+
api: "openai-responses";
|
|
5670
|
+
provider: string;
|
|
5671
|
+
baseUrl: string;
|
|
5672
|
+
reasoning: true;
|
|
5673
|
+
input: ("image" | "text")[];
|
|
5674
|
+
cost: {
|
|
5675
|
+
input: number;
|
|
5676
|
+
output: number;
|
|
5677
|
+
cacheRead: number;
|
|
5678
|
+
cacheWrite: number;
|
|
5679
|
+
};
|
|
5680
|
+
contextWindow: number;
|
|
5681
|
+
maxTokens: number;
|
|
5682
|
+
};
|
|
5569
5683
|
readonly o1: {
|
|
5570
5684
|
id: string;
|
|
5571
5685
|
name: string;
|
|
@@ -6250,10 +6364,27 @@ export declare const MODELS: {
|
|
|
6250
6364
|
contextWindow: number;
|
|
6251
6365
|
maxTokens: number;
|
|
6252
6366
|
};
|
|
6253
|
-
readonly "
|
|
6367
|
+
readonly "gpt-5.3-codex-spark": {
|
|
6254
6368
|
id: string;
|
|
6255
6369
|
name: string;
|
|
6256
|
-
api: "openai-
|
|
6370
|
+
api: "openai-responses";
|
|
6371
|
+
provider: string;
|
|
6372
|
+
baseUrl: string;
|
|
6373
|
+
reasoning: true;
|
|
6374
|
+
input: "text"[];
|
|
6375
|
+
cost: {
|
|
6376
|
+
input: number;
|
|
6377
|
+
output: number;
|
|
6378
|
+
cacheRead: number;
|
|
6379
|
+
cacheWrite: number;
|
|
6380
|
+
};
|
|
6381
|
+
contextWindow: number;
|
|
6382
|
+
maxTokens: number;
|
|
6383
|
+
};
|
|
6384
|
+
readonly "gpt-5.4": {
|
|
6385
|
+
id: string;
|
|
6386
|
+
name: string;
|
|
6387
|
+
api: "openai-responses";
|
|
6257
6388
|
provider: string;
|
|
6258
6389
|
baseUrl: string;
|
|
6259
6390
|
reasoning: true;
|
|
@@ -6267,14 +6398,31 @@ export declare const MODELS: {
|
|
|
6267
6398
|
contextWindow: number;
|
|
6268
6399
|
maxTokens: number;
|
|
6269
6400
|
};
|
|
6270
|
-
readonly "
|
|
6401
|
+
readonly "gpt-5.4-pro": {
|
|
6402
|
+
id: string;
|
|
6403
|
+
name: string;
|
|
6404
|
+
api: "openai-responses";
|
|
6405
|
+
provider: string;
|
|
6406
|
+
baseUrl: string;
|
|
6407
|
+
reasoning: true;
|
|
6408
|
+
input: ("image" | "text")[];
|
|
6409
|
+
cost: {
|
|
6410
|
+
input: number;
|
|
6411
|
+
output: number;
|
|
6412
|
+
cacheRead: number;
|
|
6413
|
+
cacheWrite: number;
|
|
6414
|
+
};
|
|
6415
|
+
contextWindow: number;
|
|
6416
|
+
maxTokens: number;
|
|
6417
|
+
};
|
|
6418
|
+
readonly "kimi-k2.5": {
|
|
6271
6419
|
id: string;
|
|
6272
6420
|
name: string;
|
|
6273
6421
|
api: "openai-completions";
|
|
6274
6422
|
provider: string;
|
|
6275
6423
|
baseUrl: string;
|
|
6276
6424
|
reasoning: true;
|
|
6277
|
-
input: "text"[];
|
|
6425
|
+
input: ("image" | "text")[];
|
|
6278
6426
|
cost: {
|
|
6279
6427
|
input: number;
|
|
6280
6428
|
output: number;
|
|
@@ -6284,7 +6432,7 @@ export declare const MODELS: {
|
|
|
6284
6432
|
contextWindow: number;
|
|
6285
6433
|
maxTokens: number;
|
|
6286
6434
|
};
|
|
6287
|
-
readonly "minimax-m2.
|
|
6435
|
+
readonly "minimax-m2.1": {
|
|
6288
6436
|
id: string;
|
|
6289
6437
|
name: string;
|
|
6290
6438
|
api: "openai-completions";
|
|
@@ -6301,10 +6449,10 @@ export declare const MODELS: {
|
|
|
6301
6449
|
contextWindow: number;
|
|
6302
6450
|
maxTokens: number;
|
|
6303
6451
|
};
|
|
6304
|
-
readonly "minimax-m2.5
|
|
6452
|
+
readonly "minimax-m2.5": {
|
|
6305
6453
|
id: string;
|
|
6306
6454
|
name: string;
|
|
6307
|
-
api: "
|
|
6455
|
+
api: "openai-completions";
|
|
6308
6456
|
provider: string;
|
|
6309
6457
|
baseUrl: string;
|
|
6310
6458
|
reasoning: true;
|
|
@@ -6318,13 +6466,13 @@ export declare const MODELS: {
|
|
|
6318
6466
|
contextWindow: number;
|
|
6319
6467
|
maxTokens: number;
|
|
6320
6468
|
};
|
|
6321
|
-
readonly "
|
|
6469
|
+
readonly "minimax-m2.5-free": {
|
|
6322
6470
|
id: string;
|
|
6323
6471
|
name: string;
|
|
6324
|
-
api: "
|
|
6472
|
+
api: "anthropic-messages";
|
|
6325
6473
|
provider: string;
|
|
6326
6474
|
baseUrl: string;
|
|
6327
|
-
reasoning:
|
|
6475
|
+
reasoning: true;
|
|
6328
6476
|
input: "text"[];
|
|
6329
6477
|
cost: {
|
|
6330
6478
|
input: number;
|
|
@@ -7393,6 +7541,23 @@ export declare const MODELS: {
|
|
|
7393
7541
|
contextWindow: number;
|
|
7394
7542
|
maxTokens: number;
|
|
7395
7543
|
};
|
|
7544
|
+
readonly "inception/mercury-2": {
|
|
7545
|
+
id: string;
|
|
7546
|
+
name: string;
|
|
7547
|
+
api: "openai-completions";
|
|
7548
|
+
provider: string;
|
|
7549
|
+
baseUrl: string;
|
|
7550
|
+
reasoning: true;
|
|
7551
|
+
input: "text"[];
|
|
7552
|
+
cost: {
|
|
7553
|
+
input: number;
|
|
7554
|
+
output: number;
|
|
7555
|
+
cacheRead: number;
|
|
7556
|
+
cacheWrite: number;
|
|
7557
|
+
};
|
|
7558
|
+
contextWindow: number;
|
|
7559
|
+
maxTokens: number;
|
|
7560
|
+
};
|
|
7396
7561
|
readonly "inception/mercury-coder": {
|
|
7397
7562
|
id: string;
|
|
7398
7563
|
name: string;
|
|
@@ -8889,6 +9054,40 @@ export declare const MODELS: {
|
|
|
8889
9054
|
contextWindow: number;
|
|
8890
9055
|
maxTokens: number;
|
|
8891
9056
|
};
|
|
9057
|
+
readonly "openai/gpt-5.4": {
|
|
9058
|
+
id: string;
|
|
9059
|
+
name: string;
|
|
9060
|
+
api: "openai-completions";
|
|
9061
|
+
provider: string;
|
|
9062
|
+
baseUrl: string;
|
|
9063
|
+
reasoning: true;
|
|
9064
|
+
input: ("image" | "text")[];
|
|
9065
|
+
cost: {
|
|
9066
|
+
input: number;
|
|
9067
|
+
output: number;
|
|
9068
|
+
cacheRead: number;
|
|
9069
|
+
cacheWrite: number;
|
|
9070
|
+
};
|
|
9071
|
+
contextWindow: number;
|
|
9072
|
+
maxTokens: number;
|
|
9073
|
+
};
|
|
9074
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9075
|
+
id: string;
|
|
9076
|
+
name: string;
|
|
9077
|
+
api: "openai-completions";
|
|
9078
|
+
provider: string;
|
|
9079
|
+
baseUrl: string;
|
|
9080
|
+
reasoning: true;
|
|
9081
|
+
input: ("image" | "text")[];
|
|
9082
|
+
cost: {
|
|
9083
|
+
input: number;
|
|
9084
|
+
output: number;
|
|
9085
|
+
cacheRead: number;
|
|
9086
|
+
cacheWrite: number;
|
|
9087
|
+
};
|
|
9088
|
+
contextWindow: number;
|
|
9089
|
+
maxTokens: number;
|
|
9090
|
+
};
|
|
8892
9091
|
readonly "openai/gpt-oss-120b": {
|
|
8893
9092
|
id: string;
|
|
8894
9093
|
name: string;
|
|
@@ -12138,6 +12337,40 @@ export declare const MODELS: {
|
|
|
12138
12337
|
contextWindow: number;
|
|
12139
12338
|
maxTokens: number;
|
|
12140
12339
|
};
|
|
12340
|
+
readonly "openai/gpt-5.4": {
|
|
12341
|
+
id: string;
|
|
12342
|
+
name: string;
|
|
12343
|
+
api: "anthropic-messages";
|
|
12344
|
+
provider: string;
|
|
12345
|
+
baseUrl: string;
|
|
12346
|
+
reasoning: true;
|
|
12347
|
+
input: ("image" | "text")[];
|
|
12348
|
+
cost: {
|
|
12349
|
+
input: number;
|
|
12350
|
+
output: number;
|
|
12351
|
+
cacheRead: number;
|
|
12352
|
+
cacheWrite: number;
|
|
12353
|
+
};
|
|
12354
|
+
contextWindow: number;
|
|
12355
|
+
maxTokens: number;
|
|
12356
|
+
};
|
|
12357
|
+
readonly "openai/gpt-5.4-pro": {
|
|
12358
|
+
id: string;
|
|
12359
|
+
name: string;
|
|
12360
|
+
api: "anthropic-messages";
|
|
12361
|
+
provider: string;
|
|
12362
|
+
baseUrl: string;
|
|
12363
|
+
reasoning: true;
|
|
12364
|
+
input: ("image" | "text")[];
|
|
12365
|
+
cost: {
|
|
12366
|
+
input: number;
|
|
12367
|
+
output: number;
|
|
12368
|
+
cacheRead: number;
|
|
12369
|
+
cacheWrite: number;
|
|
12370
|
+
};
|
|
12371
|
+
contextWindow: number;
|
|
12372
|
+
maxTokens: number;
|
|
12373
|
+
};
|
|
12141
12374
|
readonly "openai/gpt-oss-120b": {
|
|
12142
12375
|
id: string;
|
|
12143
12376
|
name: string;
|