@draht/ai 2026.3.4 → 2026.3.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -6250,10 +6250,10 @@ export declare const MODELS: {
6250
6250
  contextWindow: number;
6251
6251
  maxTokens: number;
6252
6252
  };
6253
- readonly "kimi-k2.5": {
6253
+ readonly "gpt-5.4": {
6254
6254
  id: string;
6255
6255
  name: string;
6256
- api: "openai-completions";
6256
+ api: "openai-responses";
6257
6257
  provider: string;
6258
6258
  baseUrl: string;
6259
6259
  reasoning: true;
@@ -6267,14 +6267,14 @@ export declare const MODELS: {
6267
6267
  contextWindow: number;
6268
6268
  maxTokens: number;
6269
6269
  };
6270
- readonly "minimax-m2.1": {
6270
+ readonly "kimi-k2.5": {
6271
6271
  id: string;
6272
6272
  name: string;
6273
6273
  api: "openai-completions";
6274
6274
  provider: string;
6275
6275
  baseUrl: string;
6276
6276
  reasoning: true;
6277
- input: "text"[];
6277
+ input: ("image" | "text")[];
6278
6278
  cost: {
6279
6279
  input: number;
6280
6280
  output: number;
@@ -6284,7 +6284,7 @@ export declare const MODELS: {
6284
6284
  contextWindow: number;
6285
6285
  maxTokens: number;
6286
6286
  };
6287
- readonly "minimax-m2.5": {
6287
+ readonly "minimax-m2.1": {
6288
6288
  id: string;
6289
6289
  name: string;
6290
6290
  api: "openai-completions";
@@ -6301,10 +6301,10 @@ export declare const MODELS: {
6301
6301
  contextWindow: number;
6302
6302
  maxTokens: number;
6303
6303
  };
6304
- readonly "minimax-m2.5-free": {
6304
+ readonly "minimax-m2.5": {
6305
6305
  id: string;
6306
6306
  name: string;
6307
- api: "anthropic-messages";
6307
+ api: "openai-completions";
6308
6308
  provider: string;
6309
6309
  baseUrl: string;
6310
6310
  reasoning: true;
@@ -6318,13 +6318,13 @@ export declare const MODELS: {
6318
6318
  contextWindow: number;
6319
6319
  maxTokens: number;
6320
6320
  };
6321
- readonly "trinity-large-preview-free": {
6321
+ readonly "minimax-m2.5-free": {
6322
6322
  id: string;
6323
6323
  name: string;
6324
- api: "openai-completions";
6324
+ api: "anthropic-messages";
6325
6325
  provider: string;
6326
6326
  baseUrl: string;
6327
- reasoning: false;
6327
+ reasoning: true;
6328
6328
  input: "text"[];
6329
6329
  cost: {
6330
6330
  input: number;
@@ -7393,6 +7393,23 @@ export declare const MODELS: {
7393
7393
  contextWindow: number;
7394
7394
  maxTokens: number;
7395
7395
  };
7396
+ readonly "inception/mercury-2": {
7397
+ id: string;
7398
+ name: string;
7399
+ api: "openai-completions";
7400
+ provider: string;
7401
+ baseUrl: string;
7402
+ reasoning: true;
7403
+ input: "text"[];
7404
+ cost: {
7405
+ input: number;
7406
+ output: number;
7407
+ cacheRead: number;
7408
+ cacheWrite: number;
7409
+ };
7410
+ contextWindow: number;
7411
+ maxTokens: number;
7412
+ };
7396
7413
  readonly "inception/mercury-coder": {
7397
7414
  id: string;
7398
7415
  name: string;
@@ -8889,6 +8906,40 @@ export declare const MODELS: {
8889
8906
  contextWindow: number;
8890
8907
  maxTokens: number;
8891
8908
  };
8909
+ readonly "openai/gpt-5.4": {
8910
+ id: string;
8911
+ name: string;
8912
+ api: "openai-completions";
8913
+ provider: string;
8914
+ baseUrl: string;
8915
+ reasoning: true;
8916
+ input: ("image" | "text")[];
8917
+ cost: {
8918
+ input: number;
8919
+ output: number;
8920
+ cacheRead: number;
8921
+ cacheWrite: number;
8922
+ };
8923
+ contextWindow: number;
8924
+ maxTokens: number;
8925
+ };
8926
+ readonly "openai/gpt-5.4-pro": {
8927
+ id: string;
8928
+ name: string;
8929
+ api: "openai-completions";
8930
+ provider: string;
8931
+ baseUrl: string;
8932
+ reasoning: true;
8933
+ input: ("image" | "text")[];
8934
+ cost: {
8935
+ input: number;
8936
+ output: number;
8937
+ cacheRead: number;
8938
+ cacheWrite: number;
8939
+ };
8940
+ contextWindow: number;
8941
+ maxTokens: number;
8942
+ };
8892
8943
  readonly "openai/gpt-oss-120b": {
8893
8944
  id: string;
8894
8945
  name: string;
@@ -12138,6 +12189,23 @@ export declare const MODELS: {
12138
12189
  contextWindow: number;
12139
12190
  maxTokens: number;
12140
12191
  };
12192
+ readonly "openai/gpt-5.4": {
12193
+ id: string;
12194
+ name: string;
12195
+ api: "anthropic-messages";
12196
+ provider: string;
12197
+ baseUrl: string;
12198
+ reasoning: true;
12199
+ input: ("image" | "text")[];
12200
+ cost: {
12201
+ input: number;
12202
+ output: number;
12203
+ cacheRead: number;
12204
+ cacheWrite: number;
12205
+ };
12206
+ contextWindow: number;
12207
+ maxTokens: number;
12208
+ };
12141
12209
  readonly "openai/gpt-oss-120b": {
12142
12210
  id: string;
12143
12211
  name: string;