@avadisabelle/ava-pi-ai 0.65.2 → 0.65.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +263 -71
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +314 -126
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
|
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
|
|
|
2537
2537
|
contextWindow: number;
|
|
2538
2538
|
maxTokens: number;
|
|
2539
2539
|
};
|
|
2540
|
+
readonly "gpt-5.5-pro": {
|
|
2541
|
+
id: string;
|
|
2542
|
+
name: string;
|
|
2543
|
+
api: "azure-openai-responses";
|
|
2544
|
+
provider: string;
|
|
2545
|
+
baseUrl: string;
|
|
2546
|
+
reasoning: true;
|
|
2547
|
+
input: ("image" | "text")[];
|
|
2548
|
+
cost: {
|
|
2549
|
+
input: number;
|
|
2550
|
+
output: number;
|
|
2551
|
+
cacheRead: number;
|
|
2552
|
+
cacheWrite: number;
|
|
2553
|
+
};
|
|
2554
|
+
contextWindow: number;
|
|
2555
|
+
maxTokens: number;
|
|
2556
|
+
};
|
|
2540
2557
|
readonly o1: {
|
|
2541
2558
|
id: string;
|
|
2542
2559
|
name: string;
|
|
@@ -5470,13 +5487,30 @@ export declare const MODELS: {
|
|
|
5470
5487
|
contextWindow: number;
|
|
5471
5488
|
maxTokens: number;
|
|
5472
5489
|
};
|
|
5490
|
+
readonly "mistral-medium-2604": {
|
|
5491
|
+
id: string;
|
|
5492
|
+
name: string;
|
|
5493
|
+
api: "mistral-conversations";
|
|
5494
|
+
provider: string;
|
|
5495
|
+
baseUrl: string;
|
|
5496
|
+
reasoning: true;
|
|
5497
|
+
input: ("image" | "text")[];
|
|
5498
|
+
cost: {
|
|
5499
|
+
input: number;
|
|
5500
|
+
output: number;
|
|
5501
|
+
cacheRead: number;
|
|
5502
|
+
cacheWrite: number;
|
|
5503
|
+
};
|
|
5504
|
+
contextWindow: number;
|
|
5505
|
+
maxTokens: number;
|
|
5506
|
+
};
|
|
5473
5507
|
readonly "mistral-medium-latest": {
|
|
5474
5508
|
id: string;
|
|
5475
5509
|
name: string;
|
|
5476
5510
|
api: "mistral-conversations";
|
|
5477
5511
|
provider: string;
|
|
5478
5512
|
baseUrl: string;
|
|
5479
|
-
reasoning:
|
|
5513
|
+
reasoning: true;
|
|
5480
5514
|
input: ("image" | "text")[];
|
|
5481
5515
|
cost: {
|
|
5482
5516
|
input: number;
|
|
@@ -6203,6 +6237,23 @@ export declare const MODELS: {
|
|
|
6203
6237
|
contextWindow: number;
|
|
6204
6238
|
maxTokens: number;
|
|
6205
6239
|
};
|
|
6240
|
+
readonly "gpt-5.5-pro": {
|
|
6241
|
+
id: string;
|
|
6242
|
+
name: string;
|
|
6243
|
+
api: "openai-responses";
|
|
6244
|
+
provider: string;
|
|
6245
|
+
baseUrl: string;
|
|
6246
|
+
reasoning: true;
|
|
6247
|
+
input: ("image" | "text")[];
|
|
6248
|
+
cost: {
|
|
6249
|
+
input: number;
|
|
6250
|
+
output: number;
|
|
6251
|
+
cacheRead: number;
|
|
6252
|
+
cacheWrite: number;
|
|
6253
|
+
};
|
|
6254
|
+
contextWindow: number;
|
|
6255
|
+
maxTokens: number;
|
|
6256
|
+
};
|
|
6206
6257
|
readonly o1: {
|
|
6207
6258
|
id: string;
|
|
6208
6259
|
name: string;
|
|
@@ -7040,23 +7091,6 @@ export declare const MODELS: {
|
|
|
7040
7091
|
contextWindow: number;
|
|
7041
7092
|
maxTokens: number;
|
|
7042
7093
|
};
|
|
7043
|
-
readonly "ling-2.6-flash-free": {
|
|
7044
|
-
id: string;
|
|
7045
|
-
name: string;
|
|
7046
|
-
api: "openai-completions";
|
|
7047
|
-
provider: string;
|
|
7048
|
-
baseUrl: string;
|
|
7049
|
-
reasoning: false;
|
|
7050
|
-
input: "text"[];
|
|
7051
|
-
cost: {
|
|
7052
|
-
input: number;
|
|
7053
|
-
output: number;
|
|
7054
|
-
cacheRead: number;
|
|
7055
|
-
cacheWrite: number;
|
|
7056
|
-
};
|
|
7057
|
-
contextWindow: number;
|
|
7058
|
-
maxTokens: number;
|
|
7059
|
-
};
|
|
7060
7094
|
readonly "minimax-m2.5": {
|
|
7061
7095
|
id: string;
|
|
7062
7096
|
name: string;
|
|
@@ -8489,6 +8523,23 @@ export declare const MODELS: {
|
|
|
8489
8523
|
contextWindow: number;
|
|
8490
8524
|
maxTokens: number;
|
|
8491
8525
|
};
|
|
8526
|
+
readonly "ibm-granite/granite-4.1-8b": {
|
|
8527
|
+
id: string;
|
|
8528
|
+
name: string;
|
|
8529
|
+
api: "openai-completions";
|
|
8530
|
+
provider: string;
|
|
8531
|
+
baseUrl: string;
|
|
8532
|
+
reasoning: false;
|
|
8533
|
+
input: "text"[];
|
|
8534
|
+
cost: {
|
|
8535
|
+
input: number;
|
|
8536
|
+
output: number;
|
|
8537
|
+
cacheRead: number;
|
|
8538
|
+
cacheWrite: number;
|
|
8539
|
+
};
|
|
8540
|
+
contextWindow: number;
|
|
8541
|
+
maxTokens: number;
|
|
8542
|
+
};
|
|
8492
8543
|
readonly "inception/mercury-2": {
|
|
8493
8544
|
id: string;
|
|
8494
8545
|
name: string;
|
|
@@ -8523,7 +8574,7 @@ export declare const MODELS: {
|
|
|
8523
8574
|
contextWindow: number;
|
|
8524
8575
|
maxTokens: number;
|
|
8525
8576
|
};
|
|
8526
|
-
readonly "inclusionai/ling-2.6-flash
|
|
8577
|
+
readonly "inclusionai/ling-2.6-flash": {
|
|
8527
8578
|
id: string;
|
|
8528
8579
|
name: string;
|
|
8529
8580
|
api: "openai-completions";
|
|
@@ -9050,23 +9101,6 @@ export declare const MODELS: {
|
|
|
9050
9101
|
contextWindow: number;
|
|
9051
9102
|
maxTokens: number;
|
|
9052
9103
|
};
|
|
9053
|
-
readonly "mistralai/mistral-small-creative": {
|
|
9054
|
-
id: string;
|
|
9055
|
-
name: string;
|
|
9056
|
-
api: "openai-completions";
|
|
9057
|
-
provider: string;
|
|
9058
|
-
baseUrl: string;
|
|
9059
|
-
reasoning: false;
|
|
9060
|
-
input: "text"[];
|
|
9061
|
-
cost: {
|
|
9062
|
-
input: number;
|
|
9063
|
-
output: number;
|
|
9064
|
-
cacheRead: number;
|
|
9065
|
-
cacheWrite: number;
|
|
9066
|
-
};
|
|
9067
|
-
contextWindow: number;
|
|
9068
|
-
maxTokens: number;
|
|
9069
|
-
};
|
|
9070
9104
|
readonly "mistralai/mixtral-8x22b-instruct": {
|
|
9071
9105
|
id: string;
|
|
9072
9106
|
name: string;
|
|
@@ -9305,6 +9339,23 @@ export declare const MODELS: {
|
|
|
9305
9339
|
contextWindow: number;
|
|
9306
9340
|
maxTokens: number;
|
|
9307
9341
|
};
|
|
9342
|
+
readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
|
|
9343
|
+
id: string;
|
|
9344
|
+
name: string;
|
|
9345
|
+
api: "openai-completions";
|
|
9346
|
+
provider: string;
|
|
9347
|
+
baseUrl: string;
|
|
9348
|
+
reasoning: true;
|
|
9349
|
+
input: ("image" | "text")[];
|
|
9350
|
+
cost: {
|
|
9351
|
+
input: number;
|
|
9352
|
+
output: number;
|
|
9353
|
+
cacheRead: number;
|
|
9354
|
+
cacheWrite: number;
|
|
9355
|
+
};
|
|
9356
|
+
contextWindow: number;
|
|
9357
|
+
maxTokens: number;
|
|
9358
|
+
};
|
|
9308
9359
|
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
9309
9360
|
id: string;
|
|
9310
9361
|
name: string;
|
|
@@ -10376,6 +10427,57 @@ export declare const MODELS: {
|
|
|
10376
10427
|
contextWindow: number;
|
|
10377
10428
|
maxTokens: number;
|
|
10378
10429
|
};
|
|
10430
|
+
readonly "openrouter/owl-alpha": {
|
|
10431
|
+
id: string;
|
|
10432
|
+
name: string;
|
|
10433
|
+
api: "openai-completions";
|
|
10434
|
+
provider: string;
|
|
10435
|
+
baseUrl: string;
|
|
10436
|
+
reasoning: false;
|
|
10437
|
+
input: "text"[];
|
|
10438
|
+
cost: {
|
|
10439
|
+
input: number;
|
|
10440
|
+
output: number;
|
|
10441
|
+
cacheRead: number;
|
|
10442
|
+
cacheWrite: number;
|
|
10443
|
+
};
|
|
10444
|
+
contextWindow: number;
|
|
10445
|
+
maxTokens: number;
|
|
10446
|
+
};
|
|
10447
|
+
readonly "poolside/laguna-m.1:free": {
|
|
10448
|
+
id: string;
|
|
10449
|
+
name: string;
|
|
10450
|
+
api: "openai-completions";
|
|
10451
|
+
provider: string;
|
|
10452
|
+
baseUrl: string;
|
|
10453
|
+
reasoning: true;
|
|
10454
|
+
input: "text"[];
|
|
10455
|
+
cost: {
|
|
10456
|
+
input: number;
|
|
10457
|
+
output: number;
|
|
10458
|
+
cacheRead: number;
|
|
10459
|
+
cacheWrite: number;
|
|
10460
|
+
};
|
|
10461
|
+
contextWindow: number;
|
|
10462
|
+
maxTokens: number;
|
|
10463
|
+
};
|
|
10464
|
+
readonly "poolside/laguna-xs.2:free": {
|
|
10465
|
+
id: string;
|
|
10466
|
+
name: string;
|
|
10467
|
+
api: "openai-completions";
|
|
10468
|
+
provider: string;
|
|
10469
|
+
baseUrl: string;
|
|
10470
|
+
reasoning: true;
|
|
10471
|
+
input: "text"[];
|
|
10472
|
+
cost: {
|
|
10473
|
+
input: number;
|
|
10474
|
+
output: number;
|
|
10475
|
+
cacheRead: number;
|
|
10476
|
+
cacheWrite: number;
|
|
10477
|
+
};
|
|
10478
|
+
contextWindow: number;
|
|
10479
|
+
maxTokens: number;
|
|
10480
|
+
};
|
|
10379
10481
|
readonly "prime-intellect/intellect-3": {
|
|
10380
10482
|
id: string;
|
|
10381
10483
|
name: string;
|
|
@@ -11192,23 +11294,6 @@ export declare const MODELS: {
|
|
|
11192
11294
|
contextWindow: number;
|
|
11193
11295
|
maxTokens: number;
|
|
11194
11296
|
};
|
|
11195
|
-
readonly "qwen/qwq-32b": {
|
|
11196
|
-
id: string;
|
|
11197
|
-
name: string;
|
|
11198
|
-
api: "openai-completions";
|
|
11199
|
-
provider: string;
|
|
11200
|
-
baseUrl: string;
|
|
11201
|
-
reasoning: true;
|
|
11202
|
-
input: "text"[];
|
|
11203
|
-
cost: {
|
|
11204
|
-
input: number;
|
|
11205
|
-
output: number;
|
|
11206
|
-
cacheRead: number;
|
|
11207
|
-
cacheWrite: number;
|
|
11208
|
-
};
|
|
11209
|
-
contextWindow: number;
|
|
11210
|
-
maxTokens: number;
|
|
11211
|
-
};
|
|
11212
11297
|
readonly "rekaai/reka-edge": {
|
|
11213
11298
|
id: string;
|
|
11214
11299
|
name: string;
|
|
@@ -11515,6 +11600,23 @@ export declare const MODELS: {
|
|
|
11515
11600
|
contextWindow: number;
|
|
11516
11601
|
maxTokens: number;
|
|
11517
11602
|
};
|
|
11603
|
+
readonly "x-ai/grok-4.3": {
|
|
11604
|
+
id: string;
|
|
11605
|
+
name: string;
|
|
11606
|
+
api: "openai-completions";
|
|
11607
|
+
provider: string;
|
|
11608
|
+
baseUrl: string;
|
|
11609
|
+
reasoning: true;
|
|
11610
|
+
input: ("image" | "text")[];
|
|
11611
|
+
cost: {
|
|
11612
|
+
input: number;
|
|
11613
|
+
output: number;
|
|
11614
|
+
cacheRead: number;
|
|
11615
|
+
cacheWrite: number;
|
|
11616
|
+
};
|
|
11617
|
+
contextWindow: number;
|
|
11618
|
+
maxTokens: number;
|
|
11619
|
+
};
|
|
11518
11620
|
readonly "x-ai/grok-code-fast-1": {
|
|
11519
11621
|
id: string;
|
|
11520
11622
|
name: string;
|
|
@@ -12248,6 +12350,23 @@ export declare const MODELS: {
|
|
|
12248
12350
|
contextWindow: number;
|
|
12249
12351
|
maxTokens: number;
|
|
12250
12352
|
};
|
|
12353
|
+
readonly "alibaba/qwen3.6-27b": {
|
|
12354
|
+
id: string;
|
|
12355
|
+
name: string;
|
|
12356
|
+
api: "anthropic-messages";
|
|
12357
|
+
provider: string;
|
|
12358
|
+
baseUrl: string;
|
|
12359
|
+
reasoning: true;
|
|
12360
|
+
input: ("image" | "text")[];
|
|
12361
|
+
cost: {
|
|
12362
|
+
input: number;
|
|
12363
|
+
output: number;
|
|
12364
|
+
cacheRead: number;
|
|
12365
|
+
cacheWrite: number;
|
|
12366
|
+
};
|
|
12367
|
+
contextWindow: number;
|
|
12368
|
+
maxTokens: number;
|
|
12369
|
+
};
|
|
12251
12370
|
readonly "alibaba/qwen3.6-plus": {
|
|
12252
12371
|
id: string;
|
|
12253
12372
|
name: string;
|
|
@@ -12628,7 +12747,7 @@ export declare const MODELS: {
|
|
|
12628
12747
|
api: "anthropic-messages";
|
|
12629
12748
|
provider: string;
|
|
12630
12749
|
baseUrl: string;
|
|
12631
|
-
reasoning:
|
|
12750
|
+
reasoning: false;
|
|
12632
12751
|
input: "text"[];
|
|
12633
12752
|
cost: {
|
|
12634
12753
|
input: number;
|
|
@@ -14152,23 +14271,6 @@ export declare const MODELS: {
|
|
|
14152
14271
|
contextWindow: number;
|
|
14153
14272
|
maxTokens: number;
|
|
14154
14273
|
};
|
|
14155
|
-
readonly "prime-intellect/intellect-3": {
|
|
14156
|
-
id: string;
|
|
14157
|
-
name: string;
|
|
14158
|
-
api: "anthropic-messages";
|
|
14159
|
-
provider: string;
|
|
14160
|
-
baseUrl: string;
|
|
14161
|
-
reasoning: true;
|
|
14162
|
-
input: "text"[];
|
|
14163
|
-
cost: {
|
|
14164
|
-
input: number;
|
|
14165
|
-
output: number;
|
|
14166
|
-
cacheRead: number;
|
|
14167
|
-
cacheWrite: number;
|
|
14168
|
-
};
|
|
14169
|
-
contextWindow: number;
|
|
14170
|
-
maxTokens: number;
|
|
14171
|
-
};
|
|
14172
14274
|
readonly "xai/grok-3": {
|
|
14173
14275
|
id: string;
|
|
14174
14276
|
name: string;
|
|
@@ -14424,6 +14526,23 @@ export declare const MODELS: {
|
|
|
14424
14526
|
contextWindow: number;
|
|
14425
14527
|
maxTokens: number;
|
|
14426
14528
|
};
|
|
14529
|
+
readonly "xai/grok-4.3": {
|
|
14530
|
+
id: string;
|
|
14531
|
+
name: string;
|
|
14532
|
+
api: "anthropic-messages";
|
|
14533
|
+
provider: string;
|
|
14534
|
+
baseUrl: string;
|
|
14535
|
+
reasoning: true;
|
|
14536
|
+
input: ("image" | "text")[];
|
|
14537
|
+
cost: {
|
|
14538
|
+
input: number;
|
|
14539
|
+
output: number;
|
|
14540
|
+
cacheRead: number;
|
|
14541
|
+
cacheWrite: number;
|
|
14542
|
+
};
|
|
14543
|
+
contextWindow: number;
|
|
14544
|
+
maxTokens: number;
|
|
14545
|
+
};
|
|
14427
14546
|
readonly "xai/grok-code-fast-1": {
|
|
14428
14547
|
id: string;
|
|
14429
14548
|
name: string;
|
|
@@ -14475,6 +14594,40 @@ export declare const MODELS: {
|
|
|
14475
14594
|
contextWindow: number;
|
|
14476
14595
|
maxTokens: number;
|
|
14477
14596
|
};
|
|
14597
|
+
readonly "xiaomi/mimo-v2.5": {
|
|
14598
|
+
id: string;
|
|
14599
|
+
name: string;
|
|
14600
|
+
api: "anthropic-messages";
|
|
14601
|
+
provider: string;
|
|
14602
|
+
baseUrl: string;
|
|
14603
|
+
reasoning: true;
|
|
14604
|
+
input: ("image" | "text")[];
|
|
14605
|
+
cost: {
|
|
14606
|
+
input: number;
|
|
14607
|
+
output: number;
|
|
14608
|
+
cacheRead: number;
|
|
14609
|
+
cacheWrite: number;
|
|
14610
|
+
};
|
|
14611
|
+
contextWindow: number;
|
|
14612
|
+
maxTokens: number;
|
|
14613
|
+
};
|
|
14614
|
+
readonly "xiaomi/mimo-v2.5-pro": {
|
|
14615
|
+
id: string;
|
|
14616
|
+
name: string;
|
|
14617
|
+
api: "anthropic-messages";
|
|
14618
|
+
provider: string;
|
|
14619
|
+
baseUrl: string;
|
|
14620
|
+
reasoning: true;
|
|
14621
|
+
input: ("image" | "text")[];
|
|
14622
|
+
cost: {
|
|
14623
|
+
input: number;
|
|
14624
|
+
output: number;
|
|
14625
|
+
cacheRead: number;
|
|
14626
|
+
cacheWrite: number;
|
|
14627
|
+
};
|
|
14628
|
+
contextWindow: number;
|
|
14629
|
+
maxTokens: number;
|
|
14630
|
+
};
|
|
14478
14631
|
readonly "zai/glm-4.5": {
|
|
14479
14632
|
id: string;
|
|
14480
14633
|
name: string;
|
|
@@ -15055,6 +15208,23 @@ export declare const MODELS: {
|
|
|
15055
15208
|
contextWindow: number;
|
|
15056
15209
|
maxTokens: number;
|
|
15057
15210
|
};
|
|
15211
|
+
readonly "grok-4.3": {
|
|
15212
|
+
id: string;
|
|
15213
|
+
name: string;
|
|
15214
|
+
api: "openai-completions";
|
|
15215
|
+
provider: string;
|
|
15216
|
+
baseUrl: string;
|
|
15217
|
+
reasoning: true;
|
|
15218
|
+
input: ("image" | "text")[];
|
|
15219
|
+
cost: {
|
|
15220
|
+
input: number;
|
|
15221
|
+
output: number;
|
|
15222
|
+
cacheRead: number;
|
|
15223
|
+
cacheWrite: number;
|
|
15224
|
+
};
|
|
15225
|
+
contextWindow: number;
|
|
15226
|
+
maxTokens: number;
|
|
15227
|
+
};
|
|
15058
15228
|
readonly "grok-beta": {
|
|
15059
15229
|
id: string;
|
|
15060
15230
|
name: string;
|
|
@@ -15195,6 +15365,28 @@ export declare const MODELS: {
|
|
|
15195
15365
|
contextWindow: number;
|
|
15196
15366
|
maxTokens: number;
|
|
15197
15367
|
};
|
|
15368
|
+
readonly "glm-5v-turbo": {
|
|
15369
|
+
id: string;
|
|
15370
|
+
name: string;
|
|
15371
|
+
api: "openai-completions";
|
|
15372
|
+
provider: string;
|
|
15373
|
+
baseUrl: string;
|
|
15374
|
+
compat: {
|
|
15375
|
+
supportsDeveloperRole: false;
|
|
15376
|
+
thinkingFormat: "zai";
|
|
15377
|
+
zaiToolStream: true;
|
|
15378
|
+
};
|
|
15379
|
+
reasoning: true;
|
|
15380
|
+
input: ("image" | "text")[];
|
|
15381
|
+
cost: {
|
|
15382
|
+
input: number;
|
|
15383
|
+
output: number;
|
|
15384
|
+
cacheRead: number;
|
|
15385
|
+
cacheWrite: number;
|
|
15386
|
+
};
|
|
15387
|
+
contextWindow: number;
|
|
15388
|
+
maxTokens: number;
|
|
15389
|
+
};
|
|
15198
15390
|
};
|
|
15199
15391
|
};
|
|
15200
15392
|
//# sourceMappingURL=models.generated.d.ts.map
|