@mariozechner/pi-ai 0.70.2 → 0.70.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +2 -2
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +49 -7
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.d.ts +1 -0
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +442 -242
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +460 -242
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +3 -1
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +49 -25
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +12 -4
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +18 -2
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +2 -2
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +1 -1
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +1 -1
- package/dist/providers/openai-responses.js.map +1 -1
- package/package.json +1 -1
|
@@ -3702,6 +3702,29 @@ export declare const MODELS: {
|
|
|
3702
3702
|
contextWindow: number;
|
|
3703
3703
|
maxTokens: number;
|
|
3704
3704
|
};
|
|
3705
|
+
readonly "gpt-5.5": {
|
|
3706
|
+
id: string;
|
|
3707
|
+
name: string;
|
|
3708
|
+
api: "openai-responses";
|
|
3709
|
+
provider: string;
|
|
3710
|
+
baseUrl: string;
|
|
3711
|
+
headers: {
|
|
3712
|
+
"User-Agent": string;
|
|
3713
|
+
"Editor-Version": string;
|
|
3714
|
+
"Editor-Plugin-Version": string;
|
|
3715
|
+
"Copilot-Integration-Id": string;
|
|
3716
|
+
};
|
|
3717
|
+
reasoning: true;
|
|
3718
|
+
input: ("image" | "text")[];
|
|
3719
|
+
cost: {
|
|
3720
|
+
input: number;
|
|
3721
|
+
output: number;
|
|
3722
|
+
cacheRead: number;
|
|
3723
|
+
cacheWrite: number;
|
|
3724
|
+
};
|
|
3725
|
+
contextWindow: number;
|
|
3726
|
+
maxTokens: number;
|
|
3727
|
+
};
|
|
3705
3728
|
readonly "grok-code-fast-1": {
|
|
3706
3729
|
id: string;
|
|
3707
3730
|
name: string;
|
|
@@ -5240,6 +5263,26 @@ export declare const MODELS: {
|
|
|
5240
5263
|
contextWindow: number;
|
|
5241
5264
|
maxTokens: number;
|
|
5242
5265
|
};
|
|
5266
|
+
readonly "deepseek-ai/DeepSeek-V4-Pro": {
|
|
5267
|
+
id: string;
|
|
5268
|
+
name: string;
|
|
5269
|
+
api: "openai-completions";
|
|
5270
|
+
provider: string;
|
|
5271
|
+
baseUrl: string;
|
|
5272
|
+
compat: {
|
|
5273
|
+
supportsDeveloperRole: false;
|
|
5274
|
+
};
|
|
5275
|
+
reasoning: true;
|
|
5276
|
+
input: "text"[];
|
|
5277
|
+
cost: {
|
|
5278
|
+
input: number;
|
|
5279
|
+
output: number;
|
|
5280
|
+
cacheRead: number;
|
|
5281
|
+
cacheWrite: number;
|
|
5282
|
+
};
|
|
5283
|
+
contextWindow: number;
|
|
5284
|
+
maxTokens: number;
|
|
5285
|
+
};
|
|
5243
5286
|
readonly "moonshotai/Kimi-K2-Instruct": {
|
|
5244
5287
|
id: string;
|
|
5245
5288
|
name: string;
|
|
@@ -6888,23 +6931,6 @@ export declare const MODELS: {
|
|
|
6888
6931
|
contextWindow: number;
|
|
6889
6932
|
maxTokens: number;
|
|
6890
6933
|
};
|
|
6891
|
-
readonly "claude-3-5-haiku": {
|
|
6892
|
-
id: string;
|
|
6893
|
-
name: string;
|
|
6894
|
-
api: "anthropic-messages";
|
|
6895
|
-
provider: string;
|
|
6896
|
-
baseUrl: string;
|
|
6897
|
-
reasoning: false;
|
|
6898
|
-
input: ("image" | "text")[];
|
|
6899
|
-
cost: {
|
|
6900
|
-
input: number;
|
|
6901
|
-
output: number;
|
|
6902
|
-
cacheRead: number;
|
|
6903
|
-
cacheWrite: number;
|
|
6904
|
-
};
|
|
6905
|
-
contextWindow: number;
|
|
6906
|
-
maxTokens: number;
|
|
6907
|
-
};
|
|
6908
6934
|
readonly "claude-haiku-4-5": {
|
|
6909
6935
|
id: string;
|
|
6910
6936
|
name: string;
|
|
@@ -7347,6 +7373,40 @@ export declare const MODELS: {
|
|
|
7347
7373
|
contextWindow: number;
|
|
7348
7374
|
maxTokens: number;
|
|
7349
7375
|
};
|
|
7376
|
+
readonly "gpt-5.5": {
|
|
7377
|
+
id: string;
|
|
7378
|
+
name: string;
|
|
7379
|
+
api: "openai-responses";
|
|
7380
|
+
provider: string;
|
|
7381
|
+
baseUrl: string;
|
|
7382
|
+
reasoning: true;
|
|
7383
|
+
input: ("image" | "text")[];
|
|
7384
|
+
cost: {
|
|
7385
|
+
input: number;
|
|
7386
|
+
output: number;
|
|
7387
|
+
cacheRead: number;
|
|
7388
|
+
cacheWrite: number;
|
|
7389
|
+
};
|
|
7390
|
+
contextWindow: number;
|
|
7391
|
+
maxTokens: number;
|
|
7392
|
+
};
|
|
7393
|
+
readonly "gpt-5.5-pro": {
|
|
7394
|
+
id: string;
|
|
7395
|
+
name: string;
|
|
7396
|
+
api: "openai-responses";
|
|
7397
|
+
provider: string;
|
|
7398
|
+
baseUrl: string;
|
|
7399
|
+
reasoning: true;
|
|
7400
|
+
input: ("image" | "text")[];
|
|
7401
|
+
cost: {
|
|
7402
|
+
input: number;
|
|
7403
|
+
output: number;
|
|
7404
|
+
cacheRead: number;
|
|
7405
|
+
cacheWrite: number;
|
|
7406
|
+
};
|
|
7407
|
+
contextWindow: number;
|
|
7408
|
+
maxTokens: number;
|
|
7409
|
+
};
|
|
7350
7410
|
readonly "hy3-preview-free": {
|
|
7351
7411
|
id: string;
|
|
7352
7412
|
name: string;
|
|
@@ -7519,6 +7579,46 @@ export declare const MODELS: {
|
|
|
7519
7579
|
};
|
|
7520
7580
|
};
|
|
7521
7581
|
readonly "opencode-go": {
|
|
7582
|
+
readonly "deepseek-v4-flash": {
|
|
7583
|
+
id: string;
|
|
7584
|
+
name: string;
|
|
7585
|
+
api: "openai-completions";
|
|
7586
|
+
provider: string;
|
|
7587
|
+
baseUrl: string;
|
|
7588
|
+
compat: {
|
|
7589
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
7590
|
+
};
|
|
7591
|
+
reasoning: true;
|
|
7592
|
+
input: "text"[];
|
|
7593
|
+
cost: {
|
|
7594
|
+
input: number;
|
|
7595
|
+
output: number;
|
|
7596
|
+
cacheRead: number;
|
|
7597
|
+
cacheWrite: number;
|
|
7598
|
+
};
|
|
7599
|
+
contextWindow: number;
|
|
7600
|
+
maxTokens: number;
|
|
7601
|
+
};
|
|
7602
|
+
readonly "deepseek-v4-pro": {
|
|
7603
|
+
id: string;
|
|
7604
|
+
name: string;
|
|
7605
|
+
api: "openai-completions";
|
|
7606
|
+
provider: string;
|
|
7607
|
+
baseUrl: string;
|
|
7608
|
+
compat: {
|
|
7609
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
7610
|
+
};
|
|
7611
|
+
reasoning: true;
|
|
7612
|
+
input: "text"[];
|
|
7613
|
+
cost: {
|
|
7614
|
+
input: number;
|
|
7615
|
+
output: number;
|
|
7616
|
+
cacheRead: number;
|
|
7617
|
+
cacheWrite: number;
|
|
7618
|
+
};
|
|
7619
|
+
contextWindow: number;
|
|
7620
|
+
maxTokens: number;
|
|
7621
|
+
};
|
|
7522
7622
|
readonly "glm-5": {
|
|
7523
7623
|
id: string;
|
|
7524
7624
|
name: string;
|
|
@@ -8462,6 +8562,9 @@ export declare const MODELS: {
|
|
|
8462
8562
|
api: "openai-completions";
|
|
8463
8563
|
provider: string;
|
|
8464
8564
|
baseUrl: string;
|
|
8565
|
+
compat: {
|
|
8566
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
8567
|
+
};
|
|
8465
8568
|
reasoning: true;
|
|
8466
8569
|
input: "text"[];
|
|
8467
8570
|
cost: {
|
|
@@ -8479,6 +8582,9 @@ export declare const MODELS: {
|
|
|
8479
8582
|
api: "openai-completions";
|
|
8480
8583
|
provider: string;
|
|
8481
8584
|
baseUrl: string;
|
|
8585
|
+
compat: {
|
|
8586
|
+
requiresReasoningContentOnAssistantMessages: true;
|
|
8587
|
+
};
|
|
8482
8588
|
reasoning: true;
|
|
8483
8589
|
input: "text"[];
|
|
8484
8590
|
cost: {
|
|
@@ -8711,6 +8817,40 @@ export declare const MODELS: {
|
|
|
8711
8817
|
contextWindow: number;
|
|
8712
8818
|
maxTokens: number;
|
|
8713
8819
|
};
|
|
8820
|
+
readonly "google/gemma-3-12b-it": {
|
|
8821
|
+
id: string;
|
|
8822
|
+
name: string;
|
|
8823
|
+
api: "openai-completions";
|
|
8824
|
+
provider: string;
|
|
8825
|
+
baseUrl: string;
|
|
8826
|
+
reasoning: false;
|
|
8827
|
+
input: ("image" | "text")[];
|
|
8828
|
+
cost: {
|
|
8829
|
+
input: number;
|
|
8830
|
+
output: number;
|
|
8831
|
+
cacheRead: number;
|
|
8832
|
+
cacheWrite: number;
|
|
8833
|
+
};
|
|
8834
|
+
contextWindow: number;
|
|
8835
|
+
maxTokens: number;
|
|
8836
|
+
};
|
|
8837
|
+
readonly "google/gemma-3-27b-it": {
|
|
8838
|
+
id: string;
|
|
8839
|
+
name: string;
|
|
8840
|
+
api: "openai-completions";
|
|
8841
|
+
provider: string;
|
|
8842
|
+
baseUrl: string;
|
|
8843
|
+
reasoning: false;
|
|
8844
|
+
input: ("image" | "text")[];
|
|
8845
|
+
cost: {
|
|
8846
|
+
input: number;
|
|
8847
|
+
output: number;
|
|
8848
|
+
cacheRead: number;
|
|
8849
|
+
cacheWrite: number;
|
|
8850
|
+
};
|
|
8851
|
+
contextWindow: number;
|
|
8852
|
+
maxTokens: number;
|
|
8853
|
+
};
|
|
8714
8854
|
readonly "google/gemma-4-26b-a4b-it": {
|
|
8715
8855
|
id: string;
|
|
8716
8856
|
name: string;
|
|
@@ -10326,6 +10466,40 @@ export declare const MODELS: {
|
|
|
10326
10466
|
contextWindow: number;
|
|
10327
10467
|
maxTokens: number;
|
|
10328
10468
|
};
|
|
10469
|
+
readonly "openai/gpt-5.5": {
|
|
10470
|
+
id: string;
|
|
10471
|
+
name: string;
|
|
10472
|
+
api: "openai-completions";
|
|
10473
|
+
provider: string;
|
|
10474
|
+
baseUrl: string;
|
|
10475
|
+
reasoning: true;
|
|
10476
|
+
input: ("image" | "text")[];
|
|
10477
|
+
cost: {
|
|
10478
|
+
input: number;
|
|
10479
|
+
output: number;
|
|
10480
|
+
cacheRead: number;
|
|
10481
|
+
cacheWrite: number;
|
|
10482
|
+
};
|
|
10483
|
+
contextWindow: number;
|
|
10484
|
+
maxTokens: number;
|
|
10485
|
+
};
|
|
10486
|
+
readonly "openai/gpt-5.5-pro": {
|
|
10487
|
+
id: string;
|
|
10488
|
+
name: string;
|
|
10489
|
+
api: "openai-completions";
|
|
10490
|
+
provider: string;
|
|
10491
|
+
baseUrl: string;
|
|
10492
|
+
reasoning: true;
|
|
10493
|
+
input: ("image" | "text")[];
|
|
10494
|
+
cost: {
|
|
10495
|
+
input: number;
|
|
10496
|
+
output: number;
|
|
10497
|
+
cacheRead: number;
|
|
10498
|
+
cacheWrite: number;
|
|
10499
|
+
};
|
|
10500
|
+
contextWindow: number;
|
|
10501
|
+
maxTokens: number;
|
|
10502
|
+
};
|
|
10329
10503
|
readonly "openai/gpt-audio": {
|
|
10330
10504
|
id: string;
|
|
10331
10505
|
name: string;
|
|
@@ -11363,7 +11537,7 @@ export declare const MODELS: {
|
|
|
11363
11537
|
contextWindow: number;
|
|
11364
11538
|
maxTokens: number;
|
|
11365
11539
|
};
|
|
11366
|
-
readonly "qwen/qwen3.
|
|
11540
|
+
readonly "qwen/qwen3.5-plus-20260420": {
|
|
11367
11541
|
id: string;
|
|
11368
11542
|
name: string;
|
|
11369
11543
|
api: "openai-completions";
|
|
@@ -11380,14 +11554,14 @@ export declare const MODELS: {
|
|
|
11380
11554
|
contextWindow: number;
|
|
11381
11555
|
maxTokens: number;
|
|
11382
11556
|
};
|
|
11383
|
-
readonly "qwen/
|
|
11557
|
+
readonly "qwen/qwen3.6-27b": {
|
|
11384
11558
|
id: string;
|
|
11385
11559
|
name: string;
|
|
11386
11560
|
api: "openai-completions";
|
|
11387
11561
|
provider: string;
|
|
11388
11562
|
baseUrl: string;
|
|
11389
11563
|
reasoning: true;
|
|
11390
|
-
input: "text"[];
|
|
11564
|
+
input: ("image" | "text")[];
|
|
11391
11565
|
cost: {
|
|
11392
11566
|
input: number;
|
|
11393
11567
|
output: number;
|
|
@@ -11397,13 +11571,13 @@ export declare const MODELS: {
|
|
|
11397
11571
|
contextWindow: number;
|
|
11398
11572
|
maxTokens: number;
|
|
11399
11573
|
};
|
|
11400
|
-
readonly "
|
|
11574
|
+
readonly "qwen/qwen3.6-flash": {
|
|
11401
11575
|
id: string;
|
|
11402
11576
|
name: string;
|
|
11403
11577
|
api: "openai-completions";
|
|
11404
11578
|
provider: string;
|
|
11405
11579
|
baseUrl: string;
|
|
11406
|
-
reasoning:
|
|
11580
|
+
reasoning: true;
|
|
11407
11581
|
input: ("image" | "text")[];
|
|
11408
11582
|
cost: {
|
|
11409
11583
|
input: number;
|
|
@@ -11414,13 +11588,13 @@ export declare const MODELS: {
|
|
|
11414
11588
|
contextWindow: number;
|
|
11415
11589
|
maxTokens: number;
|
|
11416
11590
|
};
|
|
11417
|
-
readonly "
|
|
11591
|
+
readonly "qwen/qwen3.6-max-preview": {
|
|
11418
11592
|
id: string;
|
|
11419
11593
|
name: string;
|
|
11420
11594
|
api: "openai-completions";
|
|
11421
11595
|
provider: string;
|
|
11422
11596
|
baseUrl: string;
|
|
11423
|
-
reasoning:
|
|
11597
|
+
reasoning: true;
|
|
11424
11598
|
input: "text"[];
|
|
11425
11599
|
cost: {
|
|
11426
11600
|
input: number;
|
|
@@ -11431,14 +11605,14 @@ export declare const MODELS: {
|
|
|
11431
11605
|
contextWindow: number;
|
|
11432
11606
|
maxTokens: number;
|
|
11433
11607
|
};
|
|
11434
|
-
readonly "
|
|
11608
|
+
readonly "qwen/qwen3.6-plus": {
|
|
11435
11609
|
id: string;
|
|
11436
11610
|
name: string;
|
|
11437
11611
|
api: "openai-completions";
|
|
11438
11612
|
provider: string;
|
|
11439
11613
|
baseUrl: string;
|
|
11440
|
-
reasoning:
|
|
11441
|
-
input: "text"[];
|
|
11614
|
+
reasoning: true;
|
|
11615
|
+
input: ("image" | "text")[];
|
|
11442
11616
|
cost: {
|
|
11443
11617
|
input: number;
|
|
11444
11618
|
output: number;
|
|
@@ -11448,13 +11622,13 @@ export declare const MODELS: {
|
|
|
11448
11622
|
contextWindow: number;
|
|
11449
11623
|
maxTokens: number;
|
|
11450
11624
|
};
|
|
11451
|
-
readonly "
|
|
11625
|
+
readonly "qwen/qwq-32b": {
|
|
11452
11626
|
id: string;
|
|
11453
11627
|
name: string;
|
|
11454
11628
|
api: "openai-completions";
|
|
11455
11629
|
provider: string;
|
|
11456
11630
|
baseUrl: string;
|
|
11457
|
-
reasoning:
|
|
11631
|
+
reasoning: true;
|
|
11458
11632
|
input: "text"[];
|
|
11459
11633
|
cost: {
|
|
11460
11634
|
input: number;
|
|
@@ -11465,14 +11639,14 @@ export declare const MODELS: {
|
|
|
11465
11639
|
contextWindow: number;
|
|
11466
11640
|
maxTokens: number;
|
|
11467
11641
|
};
|
|
11468
|
-
readonly "
|
|
11642
|
+
readonly "rekaai/reka-edge": {
|
|
11469
11643
|
id: string;
|
|
11470
11644
|
name: string;
|
|
11471
11645
|
api: "openai-completions";
|
|
11472
11646
|
provider: string;
|
|
11473
11647
|
baseUrl: string;
|
|
11474
|
-
reasoning:
|
|
11475
|
-
input: "text"[];
|
|
11648
|
+
reasoning: false;
|
|
11649
|
+
input: ("image" | "text")[];
|
|
11476
11650
|
cost: {
|
|
11477
11651
|
input: number;
|
|
11478
11652
|
output: number;
|
|
@@ -11482,13 +11656,13 @@ export declare const MODELS: {
|
|
|
11482
11656
|
contextWindow: number;
|
|
11483
11657
|
maxTokens: number;
|
|
11484
11658
|
};
|
|
11485
|
-
readonly "
|
|
11659
|
+
readonly "relace/relace-search": {
|
|
11486
11660
|
id: string;
|
|
11487
11661
|
name: string;
|
|
11488
11662
|
api: "openai-completions";
|
|
11489
11663
|
provider: string;
|
|
11490
11664
|
baseUrl: string;
|
|
11491
|
-
reasoning:
|
|
11665
|
+
reasoning: false;
|
|
11492
11666
|
input: "text"[];
|
|
11493
11667
|
cost: {
|
|
11494
11668
|
input: number;
|
|
@@ -11499,7 +11673,7 @@ export declare const MODELS: {
|
|
|
11499
11673
|
contextWindow: number;
|
|
11500
11674
|
maxTokens: number;
|
|
11501
11675
|
};
|
|
11502
|
-
readonly "
|
|
11676
|
+
readonly "sao10k/l3-euryale-70b": {
|
|
11503
11677
|
id: string;
|
|
11504
11678
|
name: string;
|
|
11505
11679
|
api: "openai-completions";
|
|
@@ -11516,7 +11690,7 @@ export declare const MODELS: {
|
|
|
11516
11690
|
contextWindow: number;
|
|
11517
11691
|
maxTokens: number;
|
|
11518
11692
|
};
|
|
11519
|
-
readonly "
|
|
11693
|
+
readonly "sao10k/l3.1-euryale-70b": {
|
|
11520
11694
|
id: string;
|
|
11521
11695
|
name: string;
|
|
11522
11696
|
api: "openai-completions";
|
|
@@ -11533,7 +11707,7 @@ export declare const MODELS: {
|
|
|
11533
11707
|
contextWindow: number;
|
|
11534
11708
|
maxTokens: number;
|
|
11535
11709
|
};
|
|
11536
|
-
readonly "
|
|
11710
|
+
readonly "stepfun/step-3.5-flash": {
|
|
11537
11711
|
id: string;
|
|
11538
11712
|
name: string;
|
|
11539
11713
|
api: "openai-completions";
|
|
@@ -11550,7 +11724,7 @@ export declare const MODELS: {
|
|
|
11550
11724
|
contextWindow: number;
|
|
11551
11725
|
maxTokens: number;
|
|
11552
11726
|
};
|
|
11553
|
-
readonly "
|
|
11727
|
+
readonly "tencent/hy3-preview:free": {
|
|
11554
11728
|
id: string;
|
|
11555
11729
|
name: string;
|
|
11556
11730
|
api: "openai-completions";
|
|
@@ -11567,7 +11741,7 @@ export declare const MODELS: {
|
|
|
11567
11741
|
contextWindow: number;
|
|
11568
11742
|
maxTokens: number;
|
|
11569
11743
|
};
|
|
11570
|
-
readonly "
|
|
11744
|
+
readonly "thedrummer/rocinante-12b": {
|
|
11571
11745
|
id: string;
|
|
11572
11746
|
name: string;
|
|
11573
11747
|
api: "openai-completions";
|
|
@@ -11584,7 +11758,7 @@ export declare const MODELS: {
|
|
|
11584
11758
|
contextWindow: number;
|
|
11585
11759
|
maxTokens: number;
|
|
11586
11760
|
};
|
|
11587
|
-
readonly "
|
|
11761
|
+
readonly "thedrummer/unslopnemo-12b": {
|
|
11588
11762
|
id: string;
|
|
11589
11763
|
name: string;
|
|
11590
11764
|
api: "openai-completions";
|
|
@@ -11601,7 +11775,7 @@ export declare const MODELS: {
|
|
|
11601
11775
|
contextWindow: number;
|
|
11602
11776
|
maxTokens: number;
|
|
11603
11777
|
};
|
|
11604
|
-
readonly "
|
|
11778
|
+
readonly "tngtech/deepseek-r1t2-chimera": {
|
|
11605
11779
|
id: string;
|
|
11606
11780
|
name: string;
|
|
11607
11781
|
api: "openai-completions";
|
|
@@ -11618,7 +11792,7 @@ export declare const MODELS: {
|
|
|
11618
11792
|
contextWindow: number;
|
|
11619
11793
|
maxTokens: number;
|
|
11620
11794
|
};
|
|
11621
|
-
readonly "
|
|
11795
|
+
readonly "upstage/solar-pro-3": {
|
|
11622
11796
|
id: string;
|
|
11623
11797
|
name: string;
|
|
11624
11798
|
api: "openai-completions";
|
|
@@ -11635,14 +11809,14 @@ export declare const MODELS: {
|
|
|
11635
11809
|
contextWindow: number;
|
|
11636
11810
|
maxTokens: number;
|
|
11637
11811
|
};
|
|
11638
|
-
readonly "x-ai/grok-
|
|
11812
|
+
readonly "x-ai/grok-3": {
|
|
11639
11813
|
id: string;
|
|
11640
11814
|
name: string;
|
|
11641
11815
|
api: "openai-completions";
|
|
11642
11816
|
provider: string;
|
|
11643
11817
|
baseUrl: string;
|
|
11644
|
-
reasoning:
|
|
11645
|
-
input:
|
|
11818
|
+
reasoning: false;
|
|
11819
|
+
input: "text"[];
|
|
11646
11820
|
cost: {
|
|
11647
11821
|
input: number;
|
|
11648
11822
|
output: number;
|
|
@@ -11652,13 +11826,81 @@ export declare const MODELS: {
|
|
|
11652
11826
|
contextWindow: number;
|
|
11653
11827
|
maxTokens: number;
|
|
11654
11828
|
};
|
|
11655
|
-
readonly "x-ai/grok-
|
|
11829
|
+
readonly "x-ai/grok-3-beta": {
|
|
11656
11830
|
id: string;
|
|
11657
11831
|
name: string;
|
|
11658
11832
|
api: "openai-completions";
|
|
11659
11833
|
provider: string;
|
|
11660
11834
|
baseUrl: string;
|
|
11661
|
-
reasoning:
|
|
11835
|
+
reasoning: false;
|
|
11836
|
+
input: "text"[];
|
|
11837
|
+
cost: {
|
|
11838
|
+
input: number;
|
|
11839
|
+
output: number;
|
|
11840
|
+
cacheRead: number;
|
|
11841
|
+
cacheWrite: number;
|
|
11842
|
+
};
|
|
11843
|
+
contextWindow: number;
|
|
11844
|
+
maxTokens: number;
|
|
11845
|
+
};
|
|
11846
|
+
readonly "x-ai/grok-3-mini": {
|
|
11847
|
+
id: string;
|
|
11848
|
+
name: string;
|
|
11849
|
+
api: "openai-completions";
|
|
11850
|
+
provider: string;
|
|
11851
|
+
baseUrl: string;
|
|
11852
|
+
reasoning: true;
|
|
11853
|
+
input: "text"[];
|
|
11854
|
+
cost: {
|
|
11855
|
+
input: number;
|
|
11856
|
+
output: number;
|
|
11857
|
+
cacheRead: number;
|
|
11858
|
+
cacheWrite: number;
|
|
11859
|
+
};
|
|
11860
|
+
contextWindow: number;
|
|
11861
|
+
maxTokens: number;
|
|
11862
|
+
};
|
|
11863
|
+
readonly "x-ai/grok-3-mini-beta": {
|
|
11864
|
+
id: string;
|
|
11865
|
+
name: string;
|
|
11866
|
+
api: "openai-completions";
|
|
11867
|
+
provider: string;
|
|
11868
|
+
baseUrl: string;
|
|
11869
|
+
reasoning: true;
|
|
11870
|
+
input: "text"[];
|
|
11871
|
+
cost: {
|
|
11872
|
+
input: number;
|
|
11873
|
+
output: number;
|
|
11874
|
+
cacheRead: number;
|
|
11875
|
+
cacheWrite: number;
|
|
11876
|
+
};
|
|
11877
|
+
contextWindow: number;
|
|
11878
|
+
maxTokens: number;
|
|
11879
|
+
};
|
|
11880
|
+
readonly "x-ai/grok-4": {
|
|
11881
|
+
id: string;
|
|
11882
|
+
name: string;
|
|
11883
|
+
api: "openai-completions";
|
|
11884
|
+
provider: string;
|
|
11885
|
+
baseUrl: string;
|
|
11886
|
+
reasoning: true;
|
|
11887
|
+
input: ("image" | "text")[];
|
|
11888
|
+
cost: {
|
|
11889
|
+
input: number;
|
|
11890
|
+
output: number;
|
|
11891
|
+
cacheRead: number;
|
|
11892
|
+
cacheWrite: number;
|
|
11893
|
+
};
|
|
11894
|
+
contextWindow: number;
|
|
11895
|
+
maxTokens: number;
|
|
11896
|
+
};
|
|
11897
|
+
readonly "x-ai/grok-4-fast": {
|
|
11898
|
+
id: string;
|
|
11899
|
+
name: string;
|
|
11900
|
+
api: "openai-completions";
|
|
11901
|
+
provider: string;
|
|
11902
|
+
baseUrl: string;
|
|
11903
|
+
reasoning: true;
|
|
11662
11904
|
input: ("image" | "text")[];
|
|
11663
11905
|
cost: {
|
|
11664
11906
|
input: number;
|
|
@@ -12026,6 +12268,23 @@ export declare const MODELS: {
|
|
|
12026
12268
|
contextWindow: number;
|
|
12027
12269
|
maxTokens: number;
|
|
12028
12270
|
};
|
|
12271
|
+
readonly "~anthropic/claude-haiku-latest": {
|
|
12272
|
+
id: string;
|
|
12273
|
+
name: string;
|
|
12274
|
+
api: "openai-completions";
|
|
12275
|
+
provider: string;
|
|
12276
|
+
baseUrl: string;
|
|
12277
|
+
reasoning: true;
|
|
12278
|
+
input: ("image" | "text")[];
|
|
12279
|
+
cost: {
|
|
12280
|
+
input: number;
|
|
12281
|
+
output: number;
|
|
12282
|
+
cacheRead: number;
|
|
12283
|
+
cacheWrite: number;
|
|
12284
|
+
};
|
|
12285
|
+
contextWindow: number;
|
|
12286
|
+
maxTokens: number;
|
|
12287
|
+
};
|
|
12029
12288
|
readonly "~anthropic/claude-opus-latest": {
|
|
12030
12289
|
id: string;
|
|
12031
12290
|
name: string;
|
|
@@ -12043,6 +12302,108 @@ export declare const MODELS: {
|
|
|
12043
12302
|
contextWindow: number;
|
|
12044
12303
|
maxTokens: number;
|
|
12045
12304
|
};
|
|
12305
|
+
readonly "~anthropic/claude-sonnet-latest": {
|
|
12306
|
+
id: string;
|
|
12307
|
+
name: string;
|
|
12308
|
+
api: "openai-completions";
|
|
12309
|
+
provider: string;
|
|
12310
|
+
baseUrl: string;
|
|
12311
|
+
reasoning: true;
|
|
12312
|
+
input: ("image" | "text")[];
|
|
12313
|
+
cost: {
|
|
12314
|
+
input: number;
|
|
12315
|
+
output: number;
|
|
12316
|
+
cacheRead: number;
|
|
12317
|
+
cacheWrite: number;
|
|
12318
|
+
};
|
|
12319
|
+
contextWindow: number;
|
|
12320
|
+
maxTokens: number;
|
|
12321
|
+
};
|
|
12322
|
+
readonly "~google/gemini-flash-latest": {
|
|
12323
|
+
id: string;
|
|
12324
|
+
name: string;
|
|
12325
|
+
api: "openai-completions";
|
|
12326
|
+
provider: string;
|
|
12327
|
+
baseUrl: string;
|
|
12328
|
+
reasoning: true;
|
|
12329
|
+
input: ("image" | "text")[];
|
|
12330
|
+
cost: {
|
|
12331
|
+
input: number;
|
|
12332
|
+
output: number;
|
|
12333
|
+
cacheRead: number;
|
|
12334
|
+
cacheWrite: number;
|
|
12335
|
+
};
|
|
12336
|
+
contextWindow: number;
|
|
12337
|
+
maxTokens: number;
|
|
12338
|
+
};
|
|
12339
|
+
readonly "~google/gemini-pro-latest": {
|
|
12340
|
+
id: string;
|
|
12341
|
+
name: string;
|
|
12342
|
+
api: "openai-completions";
|
|
12343
|
+
provider: string;
|
|
12344
|
+
baseUrl: string;
|
|
12345
|
+
reasoning: true;
|
|
12346
|
+
input: ("image" | "text")[];
|
|
12347
|
+
cost: {
|
|
12348
|
+
input: number;
|
|
12349
|
+
output: number;
|
|
12350
|
+
cacheRead: number;
|
|
12351
|
+
cacheWrite: number;
|
|
12352
|
+
};
|
|
12353
|
+
contextWindow: number;
|
|
12354
|
+
maxTokens: number;
|
|
12355
|
+
};
|
|
12356
|
+
readonly "~moonshotai/kimi-latest": {
|
|
12357
|
+
id: string;
|
|
12358
|
+
name: string;
|
|
12359
|
+
api: "openai-completions";
|
|
12360
|
+
provider: string;
|
|
12361
|
+
baseUrl: string;
|
|
12362
|
+
reasoning: true;
|
|
12363
|
+
input: ("image" | "text")[];
|
|
12364
|
+
cost: {
|
|
12365
|
+
input: number;
|
|
12366
|
+
output: number;
|
|
12367
|
+
cacheRead: number;
|
|
12368
|
+
cacheWrite: number;
|
|
12369
|
+
};
|
|
12370
|
+
contextWindow: number;
|
|
12371
|
+
maxTokens: number;
|
|
12372
|
+
};
|
|
12373
|
+
readonly "~openai/gpt-latest": {
|
|
12374
|
+
id: string;
|
|
12375
|
+
name: string;
|
|
12376
|
+
api: "openai-completions";
|
|
12377
|
+
provider: string;
|
|
12378
|
+
baseUrl: string;
|
|
12379
|
+
reasoning: true;
|
|
12380
|
+
input: ("image" | "text")[];
|
|
12381
|
+
cost: {
|
|
12382
|
+
input: number;
|
|
12383
|
+
output: number;
|
|
12384
|
+
cacheRead: number;
|
|
12385
|
+
cacheWrite: number;
|
|
12386
|
+
};
|
|
12387
|
+
contextWindow: number;
|
|
12388
|
+
maxTokens: number;
|
|
12389
|
+
};
|
|
12390
|
+
readonly "~openai/gpt-mini-latest": {
|
|
12391
|
+
id: string;
|
|
12392
|
+
name: string;
|
|
12393
|
+
api: "openai-completions";
|
|
12394
|
+
provider: string;
|
|
12395
|
+
baseUrl: string;
|
|
12396
|
+
reasoning: true;
|
|
12397
|
+
input: ("image" | "text")[];
|
|
12398
|
+
cost: {
|
|
12399
|
+
input: number;
|
|
12400
|
+
output: number;
|
|
12401
|
+
cacheRead: number;
|
|
12402
|
+
cacheWrite: number;
|
|
12403
|
+
};
|
|
12404
|
+
contextWindow: number;
|
|
12405
|
+
maxTokens: number;
|
|
12406
|
+
};
|
|
12046
12407
|
};
|
|
12047
12408
|
readonly "vercel-ai-gateway": {
|
|
12048
12409
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -14017,6 +14378,40 @@ export declare const MODELS: {
|
|
|
14017
14378
|
contextWindow: number;
|
|
14018
14379
|
maxTokens: number;
|
|
14019
14380
|
};
|
|
14381
|
+
readonly "openai/gpt-5.5": {
|
|
14382
|
+
id: string;
|
|
14383
|
+
name: string;
|
|
14384
|
+
api: "anthropic-messages";
|
|
14385
|
+
provider: string;
|
|
14386
|
+
baseUrl: string;
|
|
14387
|
+
reasoning: true;
|
|
14388
|
+
input: ("image" | "text")[];
|
|
14389
|
+
cost: {
|
|
14390
|
+
input: number;
|
|
14391
|
+
output: number;
|
|
14392
|
+
cacheRead: number;
|
|
14393
|
+
cacheWrite: number;
|
|
14394
|
+
};
|
|
14395
|
+
contextWindow: number;
|
|
14396
|
+
maxTokens: number;
|
|
14397
|
+
};
|
|
14398
|
+
readonly "openai/gpt-5.5-pro": {
|
|
14399
|
+
id: string;
|
|
14400
|
+
name: string;
|
|
14401
|
+
api: "anthropic-messages";
|
|
14402
|
+
provider: string;
|
|
14403
|
+
baseUrl: string;
|
|
14404
|
+
reasoning: true;
|
|
14405
|
+
input: ("image" | "text")[];
|
|
14406
|
+
cost: {
|
|
14407
|
+
input: number;
|
|
14408
|
+
output: number;
|
|
14409
|
+
cacheRead: number;
|
|
14410
|
+
cacheWrite: number;
|
|
14411
|
+
};
|
|
14412
|
+
contextWindow: number;
|
|
14413
|
+
maxTokens: number;
|
|
14414
|
+
};
|
|
14020
14415
|
readonly "openai/gpt-oss-20b": {
|
|
14021
14416
|
id: string;
|
|
14022
14417
|
name: string;
|
|
@@ -15143,27 +15538,6 @@ export declare const MODELS: {
|
|
|
15143
15538
|
};
|
|
15144
15539
|
};
|
|
15145
15540
|
readonly zai: {
|
|
15146
|
-
readonly "glm-4.5": {
|
|
15147
|
-
id: string;
|
|
15148
|
-
name: string;
|
|
15149
|
-
api: "openai-completions";
|
|
15150
|
-
provider: string;
|
|
15151
|
-
baseUrl: string;
|
|
15152
|
-
compat: {
|
|
15153
|
-
supportsDeveloperRole: false;
|
|
15154
|
-
thinkingFormat: "zai";
|
|
15155
|
-
};
|
|
15156
|
-
reasoning: true;
|
|
15157
|
-
input: "text"[];
|
|
15158
|
-
cost: {
|
|
15159
|
-
input: number;
|
|
15160
|
-
output: number;
|
|
15161
|
-
cacheRead: number;
|
|
15162
|
-
cacheWrite: number;
|
|
15163
|
-
};
|
|
15164
|
-
contextWindow: number;
|
|
15165
|
-
maxTokens: number;
|
|
15166
|
-
};
|
|
15167
15541
|
readonly "glm-4.5-air": {
|
|
15168
15542
|
id: string;
|
|
15169
15543
|
name: string;
|
|
@@ -15185,92 +15559,6 @@ export declare const MODELS: {
|
|
|
15185
15559
|
contextWindow: number;
|
|
15186
15560
|
maxTokens: number;
|
|
15187
15561
|
};
|
|
15188
|
-
readonly "glm-4.5-flash": {
|
|
15189
|
-
id: string;
|
|
15190
|
-
name: string;
|
|
15191
|
-
api: "openai-completions";
|
|
15192
|
-
provider: string;
|
|
15193
|
-
baseUrl: string;
|
|
15194
|
-
compat: {
|
|
15195
|
-
supportsDeveloperRole: false;
|
|
15196
|
-
thinkingFormat: "zai";
|
|
15197
|
-
};
|
|
15198
|
-
reasoning: true;
|
|
15199
|
-
input: "text"[];
|
|
15200
|
-
cost: {
|
|
15201
|
-
input: number;
|
|
15202
|
-
output: number;
|
|
15203
|
-
cacheRead: number;
|
|
15204
|
-
cacheWrite: number;
|
|
15205
|
-
};
|
|
15206
|
-
contextWindow: number;
|
|
15207
|
-
maxTokens: number;
|
|
15208
|
-
};
|
|
15209
|
-
readonly "glm-4.5v": {
|
|
15210
|
-
id: string;
|
|
15211
|
-
name: string;
|
|
15212
|
-
api: "openai-completions";
|
|
15213
|
-
provider: string;
|
|
15214
|
-
baseUrl: string;
|
|
15215
|
-
compat: {
|
|
15216
|
-
supportsDeveloperRole: false;
|
|
15217
|
-
thinkingFormat: "zai";
|
|
15218
|
-
};
|
|
15219
|
-
reasoning: true;
|
|
15220
|
-
input: ("image" | "text")[];
|
|
15221
|
-
cost: {
|
|
15222
|
-
input: number;
|
|
15223
|
-
output: number;
|
|
15224
|
-
cacheRead: number;
|
|
15225
|
-
cacheWrite: number;
|
|
15226
|
-
};
|
|
15227
|
-
contextWindow: number;
|
|
15228
|
-
maxTokens: number;
|
|
15229
|
-
};
|
|
15230
|
-
readonly "glm-4.6": {
|
|
15231
|
-
id: string;
|
|
15232
|
-
name: string;
|
|
15233
|
-
api: "openai-completions";
|
|
15234
|
-
provider: string;
|
|
15235
|
-
baseUrl: string;
|
|
15236
|
-
compat: {
|
|
15237
|
-
supportsDeveloperRole: false;
|
|
15238
|
-
thinkingFormat: "zai";
|
|
15239
|
-
zaiToolStream: true;
|
|
15240
|
-
};
|
|
15241
|
-
reasoning: true;
|
|
15242
|
-
input: "text"[];
|
|
15243
|
-
cost: {
|
|
15244
|
-
input: number;
|
|
15245
|
-
output: number;
|
|
15246
|
-
cacheRead: number;
|
|
15247
|
-
cacheWrite: number;
|
|
15248
|
-
};
|
|
15249
|
-
contextWindow: number;
|
|
15250
|
-
maxTokens: number;
|
|
15251
|
-
};
|
|
15252
|
-
readonly "glm-4.6v": {
|
|
15253
|
-
id: string;
|
|
15254
|
-
name: string;
|
|
15255
|
-
api: "openai-completions";
|
|
15256
|
-
provider: string;
|
|
15257
|
-
baseUrl: string;
|
|
15258
|
-
compat: {
|
|
15259
|
-
supportsDeveloperRole: false;
|
|
15260
|
-
thinkingFormat: "zai";
|
|
15261
|
-
zaiToolStream: true;
|
|
15262
|
-
};
|
|
15263
|
-
reasoning: true;
|
|
15264
|
-
input: ("image" | "text")[];
|
|
15265
|
-
cost: {
|
|
15266
|
-
input: number;
|
|
15267
|
-
output: number;
|
|
15268
|
-
cacheRead: number;
|
|
15269
|
-
cacheWrite: number;
|
|
15270
|
-
};
|
|
15271
|
-
contextWindow: number;
|
|
15272
|
-
maxTokens: number;
|
|
15273
|
-
};
|
|
15274
15562
|
readonly "glm-4.7": {
|
|
15275
15563
|
id: string;
|
|
15276
15564
|
name: string;
|
|
@@ -15293,72 +15581,6 @@ export declare const MODELS: {
|
|
|
15293
15581
|
contextWindow: number;
|
|
15294
15582
|
maxTokens: number;
|
|
15295
15583
|
};
|
|
15296
|
-
readonly "glm-4.7-flash": {
|
|
15297
|
-
id: string;
|
|
15298
|
-
name: string;
|
|
15299
|
-
api: "openai-completions";
|
|
15300
|
-
provider: string;
|
|
15301
|
-
baseUrl: string;
|
|
15302
|
-
compat: {
|
|
15303
|
-
supportsDeveloperRole: false;
|
|
15304
|
-
thinkingFormat: "zai";
|
|
15305
|
-
zaiToolStream: true;
|
|
15306
|
-
};
|
|
15307
|
-
reasoning: true;
|
|
15308
|
-
input: "text"[];
|
|
15309
|
-
cost: {
|
|
15310
|
-
input: number;
|
|
15311
|
-
output: number;
|
|
15312
|
-
cacheRead: number;
|
|
15313
|
-
cacheWrite: number;
|
|
15314
|
-
};
|
|
15315
|
-
contextWindow: number;
|
|
15316
|
-
maxTokens: number;
|
|
15317
|
-
};
|
|
15318
|
-
readonly "glm-4.7-flashx": {
|
|
15319
|
-
id: string;
|
|
15320
|
-
name: string;
|
|
15321
|
-
api: "openai-completions";
|
|
15322
|
-
provider: string;
|
|
15323
|
-
baseUrl: string;
|
|
15324
|
-
compat: {
|
|
15325
|
-
supportsDeveloperRole: false;
|
|
15326
|
-
thinkingFormat: "zai";
|
|
15327
|
-
zaiToolStream: true;
|
|
15328
|
-
};
|
|
15329
|
-
reasoning: true;
|
|
15330
|
-
input: "text"[];
|
|
15331
|
-
cost: {
|
|
15332
|
-
input: number;
|
|
15333
|
-
output: number;
|
|
15334
|
-
cacheRead: number;
|
|
15335
|
-
cacheWrite: number;
|
|
15336
|
-
};
|
|
15337
|
-
contextWindow: number;
|
|
15338
|
-
maxTokens: number;
|
|
15339
|
-
};
|
|
15340
|
-
readonly "glm-5": {
|
|
15341
|
-
id: string;
|
|
15342
|
-
name: string;
|
|
15343
|
-
api: "openai-completions";
|
|
15344
|
-
provider: string;
|
|
15345
|
-
baseUrl: string;
|
|
15346
|
-
compat: {
|
|
15347
|
-
supportsDeveloperRole: false;
|
|
15348
|
-
thinkingFormat: "zai";
|
|
15349
|
-
zaiToolStream: true;
|
|
15350
|
-
};
|
|
15351
|
-
reasoning: true;
|
|
15352
|
-
input: "text"[];
|
|
15353
|
-
cost: {
|
|
15354
|
-
input: number;
|
|
15355
|
-
output: number;
|
|
15356
|
-
cacheRead: number;
|
|
15357
|
-
cacheWrite: number;
|
|
15358
|
-
};
|
|
15359
|
-
contextWindow: number;
|
|
15360
|
-
maxTokens: number;
|
|
15361
|
-
};
|
|
15362
15584
|
readonly "glm-5-turbo": {
|
|
15363
15585
|
id: string;
|
|
15364
15586
|
name: string;
|
|
@@ -15403,28 +15625,6 @@ export declare const MODELS: {
|
|
|
15403
15625
|
contextWindow: number;
|
|
15404
15626
|
maxTokens: number;
|
|
15405
15627
|
};
|
|
15406
|
-
readonly "glm-5v-turbo": {
|
|
15407
|
-
id: string;
|
|
15408
|
-
name: string;
|
|
15409
|
-
api: "openai-completions";
|
|
15410
|
-
provider: string;
|
|
15411
|
-
baseUrl: string;
|
|
15412
|
-
compat: {
|
|
15413
|
-
supportsDeveloperRole: false;
|
|
15414
|
-
thinkingFormat: "zai";
|
|
15415
|
-
zaiToolStream: true;
|
|
15416
|
-
};
|
|
15417
|
-
reasoning: true;
|
|
15418
|
-
input: ("image" | "text")[];
|
|
15419
|
-
cost: {
|
|
15420
|
-
input: number;
|
|
15421
|
-
output: number;
|
|
15422
|
-
cacheRead: number;
|
|
15423
|
-
cacheWrite: number;
|
|
15424
|
-
};
|
|
15425
|
-
contextWindow: number;
|
|
15426
|
-
maxTokens: number;
|
|
15427
|
-
};
|
|
15428
15628
|
};
|
|
15429
15629
|
};
|
|
15430
15630
|
//# sourceMappingURL=models.generated.d.ts.map
|