@avadisabelle/ava-pi-ai 0.65.2 → 0.65.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -5470,13 +5487,30 @@ export declare const MODELS: {
5470
5487
  contextWindow: number;
5471
5488
  maxTokens: number;
5472
5489
  };
5490
+ readonly "mistral-medium-2604": {
5491
+ id: string;
5492
+ name: string;
5493
+ api: "mistral-conversations";
5494
+ provider: string;
5495
+ baseUrl: string;
5496
+ reasoning: true;
5497
+ input: ("image" | "text")[];
5498
+ cost: {
5499
+ input: number;
5500
+ output: number;
5501
+ cacheRead: number;
5502
+ cacheWrite: number;
5503
+ };
5504
+ contextWindow: number;
5505
+ maxTokens: number;
5506
+ };
5473
5507
  readonly "mistral-medium-latest": {
5474
5508
  id: string;
5475
5509
  name: string;
5476
5510
  api: "mistral-conversations";
5477
5511
  provider: string;
5478
5512
  baseUrl: string;
5479
- reasoning: false;
5513
+ reasoning: true;
5480
5514
  input: ("image" | "text")[];
5481
5515
  cost: {
5482
5516
  input: number;
@@ -6203,6 +6237,23 @@ export declare const MODELS: {
6203
6237
  contextWindow: number;
6204
6238
  maxTokens: number;
6205
6239
  };
6240
+ readonly "gpt-5.5-pro": {
6241
+ id: string;
6242
+ name: string;
6243
+ api: "openai-responses";
6244
+ provider: string;
6245
+ baseUrl: string;
6246
+ reasoning: true;
6247
+ input: ("image" | "text")[];
6248
+ cost: {
6249
+ input: number;
6250
+ output: number;
6251
+ cacheRead: number;
6252
+ cacheWrite: number;
6253
+ };
6254
+ contextWindow: number;
6255
+ maxTokens: number;
6256
+ };
6206
6257
  readonly o1: {
6207
6258
  id: string;
6208
6259
  name: string;
@@ -7040,23 +7091,6 @@ export declare const MODELS: {
7040
7091
  contextWindow: number;
7041
7092
  maxTokens: number;
7042
7093
  };
7043
- readonly "ling-2.6-flash-free": {
7044
- id: string;
7045
- name: string;
7046
- api: "openai-completions";
7047
- provider: string;
7048
- baseUrl: string;
7049
- reasoning: false;
7050
- input: "text"[];
7051
- cost: {
7052
- input: number;
7053
- output: number;
7054
- cacheRead: number;
7055
- cacheWrite: number;
7056
- };
7057
- contextWindow: number;
7058
- maxTokens: number;
7059
- };
7060
7094
  readonly "minimax-m2.5": {
7061
7095
  id: string;
7062
7096
  name: string;
@@ -8489,6 +8523,23 @@ export declare const MODELS: {
8489
8523
  contextWindow: number;
8490
8524
  maxTokens: number;
8491
8525
  };
8526
+ readonly "ibm-granite/granite-4.1-8b": {
8527
+ id: string;
8528
+ name: string;
8529
+ api: "openai-completions";
8530
+ provider: string;
8531
+ baseUrl: string;
8532
+ reasoning: false;
8533
+ input: "text"[];
8534
+ cost: {
8535
+ input: number;
8536
+ output: number;
8537
+ cacheRead: number;
8538
+ cacheWrite: number;
8539
+ };
8540
+ contextWindow: number;
8541
+ maxTokens: number;
8542
+ };
8492
8543
  readonly "inception/mercury-2": {
8493
8544
  id: string;
8494
8545
  name: string;
@@ -8523,7 +8574,7 @@ export declare const MODELS: {
8523
8574
  contextWindow: number;
8524
8575
  maxTokens: number;
8525
8576
  };
8526
- readonly "inclusionai/ling-2.6-flash:free": {
8577
+ readonly "inclusionai/ling-2.6-flash": {
8527
8578
  id: string;
8528
8579
  name: string;
8529
8580
  api: "openai-completions";
@@ -9050,23 +9101,6 @@ export declare const MODELS: {
9050
9101
  contextWindow: number;
9051
9102
  maxTokens: number;
9052
9103
  };
9053
- readonly "mistralai/mistral-small-creative": {
9054
- id: string;
9055
- name: string;
9056
- api: "openai-completions";
9057
- provider: string;
9058
- baseUrl: string;
9059
- reasoning: false;
9060
- input: "text"[];
9061
- cost: {
9062
- input: number;
9063
- output: number;
9064
- cacheRead: number;
9065
- cacheWrite: number;
9066
- };
9067
- contextWindow: number;
9068
- maxTokens: number;
9069
- };
9070
9104
  readonly "mistralai/mixtral-8x22b-instruct": {
9071
9105
  id: string;
9072
9106
  name: string;
@@ -9305,6 +9339,23 @@ export declare const MODELS: {
9305
9339
  contextWindow: number;
9306
9340
  maxTokens: number;
9307
9341
  };
9342
+ readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
9343
+ id: string;
9344
+ name: string;
9345
+ api: "openai-completions";
9346
+ provider: string;
9347
+ baseUrl: string;
9348
+ reasoning: true;
9349
+ input: ("image" | "text")[];
9350
+ cost: {
9351
+ input: number;
9352
+ output: number;
9353
+ cacheRead: number;
9354
+ cacheWrite: number;
9355
+ };
9356
+ contextWindow: number;
9357
+ maxTokens: number;
9358
+ };
9308
9359
  readonly "nvidia/nemotron-3-super-120b-a12b": {
9309
9360
  id: string;
9310
9361
  name: string;
@@ -10376,6 +10427,57 @@ export declare const MODELS: {
10376
10427
  contextWindow: number;
10377
10428
  maxTokens: number;
10378
10429
  };
10430
+ readonly "openrouter/owl-alpha": {
10431
+ id: string;
10432
+ name: string;
10433
+ api: "openai-completions";
10434
+ provider: string;
10435
+ baseUrl: string;
10436
+ reasoning: false;
10437
+ input: "text"[];
10438
+ cost: {
10439
+ input: number;
10440
+ output: number;
10441
+ cacheRead: number;
10442
+ cacheWrite: number;
10443
+ };
10444
+ contextWindow: number;
10445
+ maxTokens: number;
10446
+ };
10447
+ readonly "poolside/laguna-m.1:free": {
10448
+ id: string;
10449
+ name: string;
10450
+ api: "openai-completions";
10451
+ provider: string;
10452
+ baseUrl: string;
10453
+ reasoning: true;
10454
+ input: "text"[];
10455
+ cost: {
10456
+ input: number;
10457
+ output: number;
10458
+ cacheRead: number;
10459
+ cacheWrite: number;
10460
+ };
10461
+ contextWindow: number;
10462
+ maxTokens: number;
10463
+ };
10464
+ readonly "poolside/laguna-xs.2:free": {
10465
+ id: string;
10466
+ name: string;
10467
+ api: "openai-completions";
10468
+ provider: string;
10469
+ baseUrl: string;
10470
+ reasoning: true;
10471
+ input: "text"[];
10472
+ cost: {
10473
+ input: number;
10474
+ output: number;
10475
+ cacheRead: number;
10476
+ cacheWrite: number;
10477
+ };
10478
+ contextWindow: number;
10479
+ maxTokens: number;
10480
+ };
10379
10481
  readonly "prime-intellect/intellect-3": {
10380
10482
  id: string;
10381
10483
  name: string;
@@ -11192,23 +11294,6 @@ export declare const MODELS: {
11192
11294
  contextWindow: number;
11193
11295
  maxTokens: number;
11194
11296
  };
11195
- readonly "qwen/qwq-32b": {
11196
- id: string;
11197
- name: string;
11198
- api: "openai-completions";
11199
- provider: string;
11200
- baseUrl: string;
11201
- reasoning: true;
11202
- input: "text"[];
11203
- cost: {
11204
- input: number;
11205
- output: number;
11206
- cacheRead: number;
11207
- cacheWrite: number;
11208
- };
11209
- contextWindow: number;
11210
- maxTokens: number;
11211
- };
11212
11297
  readonly "rekaai/reka-edge": {
11213
11298
  id: string;
11214
11299
  name: string;
@@ -11515,6 +11600,23 @@ export declare const MODELS: {
11515
11600
  contextWindow: number;
11516
11601
  maxTokens: number;
11517
11602
  };
11603
+ readonly "x-ai/grok-4.3": {
11604
+ id: string;
11605
+ name: string;
11606
+ api: "openai-completions";
11607
+ provider: string;
11608
+ baseUrl: string;
11609
+ reasoning: true;
11610
+ input: ("image" | "text")[];
11611
+ cost: {
11612
+ input: number;
11613
+ output: number;
11614
+ cacheRead: number;
11615
+ cacheWrite: number;
11616
+ };
11617
+ contextWindow: number;
11618
+ maxTokens: number;
11619
+ };
11518
11620
  readonly "x-ai/grok-code-fast-1": {
11519
11621
  id: string;
11520
11622
  name: string;
@@ -12248,6 +12350,23 @@ export declare const MODELS: {
12248
12350
  contextWindow: number;
12249
12351
  maxTokens: number;
12250
12352
  };
12353
+ readonly "alibaba/qwen3.6-27b": {
12354
+ id: string;
12355
+ name: string;
12356
+ api: "anthropic-messages";
12357
+ provider: string;
12358
+ baseUrl: string;
12359
+ reasoning: true;
12360
+ input: ("image" | "text")[];
12361
+ cost: {
12362
+ input: number;
12363
+ output: number;
12364
+ cacheRead: number;
12365
+ cacheWrite: number;
12366
+ };
12367
+ contextWindow: number;
12368
+ maxTokens: number;
12369
+ };
12251
12370
  readonly "alibaba/qwen3.6-plus": {
12252
12371
  id: string;
12253
12372
  name: string;
@@ -12628,7 +12747,7 @@ export declare const MODELS: {
12628
12747
  api: "anthropic-messages";
12629
12748
  provider: string;
12630
12749
  baseUrl: string;
12631
- reasoning: true;
12750
+ reasoning: false;
12632
12751
  input: "text"[];
12633
12752
  cost: {
12634
12753
  input: number;
@@ -14152,23 +14271,6 @@ export declare const MODELS: {
14152
14271
  contextWindow: number;
14153
14272
  maxTokens: number;
14154
14273
  };
14155
- readonly "prime-intellect/intellect-3": {
14156
- id: string;
14157
- name: string;
14158
- api: "anthropic-messages";
14159
- provider: string;
14160
- baseUrl: string;
14161
- reasoning: true;
14162
- input: "text"[];
14163
- cost: {
14164
- input: number;
14165
- output: number;
14166
- cacheRead: number;
14167
- cacheWrite: number;
14168
- };
14169
- contextWindow: number;
14170
- maxTokens: number;
14171
- };
14172
14274
  readonly "xai/grok-3": {
14173
14275
  id: string;
14174
14276
  name: string;
@@ -14424,6 +14526,23 @@ export declare const MODELS: {
14424
14526
  contextWindow: number;
14425
14527
  maxTokens: number;
14426
14528
  };
14529
+ readonly "xai/grok-4.3": {
14530
+ id: string;
14531
+ name: string;
14532
+ api: "anthropic-messages";
14533
+ provider: string;
14534
+ baseUrl: string;
14535
+ reasoning: true;
14536
+ input: ("image" | "text")[];
14537
+ cost: {
14538
+ input: number;
14539
+ output: number;
14540
+ cacheRead: number;
14541
+ cacheWrite: number;
14542
+ };
14543
+ contextWindow: number;
14544
+ maxTokens: number;
14545
+ };
14427
14546
  readonly "xai/grok-code-fast-1": {
14428
14547
  id: string;
14429
14548
  name: string;
@@ -14475,6 +14594,40 @@ export declare const MODELS: {
14475
14594
  contextWindow: number;
14476
14595
  maxTokens: number;
14477
14596
  };
14597
+ readonly "xiaomi/mimo-v2.5": {
14598
+ id: string;
14599
+ name: string;
14600
+ api: "anthropic-messages";
14601
+ provider: string;
14602
+ baseUrl: string;
14603
+ reasoning: true;
14604
+ input: ("image" | "text")[];
14605
+ cost: {
14606
+ input: number;
14607
+ output: number;
14608
+ cacheRead: number;
14609
+ cacheWrite: number;
14610
+ };
14611
+ contextWindow: number;
14612
+ maxTokens: number;
14613
+ };
14614
+ readonly "xiaomi/mimo-v2.5-pro": {
14615
+ id: string;
14616
+ name: string;
14617
+ api: "anthropic-messages";
14618
+ provider: string;
14619
+ baseUrl: string;
14620
+ reasoning: true;
14621
+ input: ("image" | "text")[];
14622
+ cost: {
14623
+ input: number;
14624
+ output: number;
14625
+ cacheRead: number;
14626
+ cacheWrite: number;
14627
+ };
14628
+ contextWindow: number;
14629
+ maxTokens: number;
14630
+ };
14478
14631
  readonly "zai/glm-4.5": {
14479
14632
  id: string;
14480
14633
  name: string;
@@ -15055,6 +15208,23 @@ export declare const MODELS: {
15055
15208
  contextWindow: number;
15056
15209
  maxTokens: number;
15057
15210
  };
15211
+ readonly "grok-4.3": {
15212
+ id: string;
15213
+ name: string;
15214
+ api: "openai-completions";
15215
+ provider: string;
15216
+ baseUrl: string;
15217
+ reasoning: true;
15218
+ input: ("image" | "text")[];
15219
+ cost: {
15220
+ input: number;
15221
+ output: number;
15222
+ cacheRead: number;
15223
+ cacheWrite: number;
15224
+ };
15225
+ contextWindow: number;
15226
+ maxTokens: number;
15227
+ };
15058
15228
  readonly "grok-beta": {
15059
15229
  id: string;
15060
15230
  name: string;
@@ -15195,6 +15365,28 @@ export declare const MODELS: {
15195
15365
  contextWindow: number;
15196
15366
  maxTokens: number;
15197
15367
  };
15368
+ readonly "glm-5v-turbo": {
15369
+ id: string;
15370
+ name: string;
15371
+ api: "openai-completions";
15372
+ provider: string;
15373
+ baseUrl: string;
15374
+ compat: {
15375
+ supportsDeveloperRole: false;
15376
+ thinkingFormat: "zai";
15377
+ zaiToolStream: true;
15378
+ };
15379
+ reasoning: true;
15380
+ input: ("image" | "text")[];
15381
+ cost: {
15382
+ input: number;
15383
+ output: number;
15384
+ cacheRead: number;
15385
+ cacheWrite: number;
15386
+ };
15387
+ contextWindow: number;
15388
+ maxTokens: number;
15389
+ };
15198
15390
  };
15199
15391
  };
15200
15392
  //# sourceMappingURL=models.generated.d.ts.map