@mariozechner/pi-ai 0.59.0 → 0.61.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +398 -86
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +445 -138
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +13 -1
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +20 -8
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +8 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +34 -13
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +11 -11
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/register-builtins.d.ts.map +1 -1
- package/dist/providers/register-builtins.js +11 -21
- package/dist/providers/register-builtins.js.map +1 -1
- package/dist/types.d.ts +2 -2
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +20 -28
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
- package/dist/utils/oauth/google-antigravity.js +22 -19
- package/dist/utils/oauth/google-antigravity.js.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.js +22 -19
- package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
- package/dist/utils/oauth/oauth-page.d.ts +3 -0
- package/dist/utils/oauth/oauth-page.d.ts.map +1 -0
- package/dist/utils/oauth/oauth-page.js +105 -0
- package/dist/utils/oauth/oauth-page.js.map +1 -0
- package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
- package/dist/utils/oauth/openai-codex.js +24 -31
- package/dist/utils/oauth/openai-codex.js.map +1 -1
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +17 -9
- package/dist/utils/validation.js.map +1 -1
- package/package.json +1 -1
|
@@ -2299,6 +2299,40 @@ export declare const MODELS: {
|
|
|
2299
2299
|
contextWindow: number;
|
|
2300
2300
|
maxTokens: number;
|
|
2301
2301
|
};
|
|
2302
|
+
readonly "gpt-5.4-mini": {
|
|
2303
|
+
id: string;
|
|
2304
|
+
name: string;
|
|
2305
|
+
api: "azure-openai-responses";
|
|
2306
|
+
provider: string;
|
|
2307
|
+
baseUrl: string;
|
|
2308
|
+
reasoning: true;
|
|
2309
|
+
input: ("image" | "text")[];
|
|
2310
|
+
cost: {
|
|
2311
|
+
input: number;
|
|
2312
|
+
output: number;
|
|
2313
|
+
cacheRead: number;
|
|
2314
|
+
cacheWrite: number;
|
|
2315
|
+
};
|
|
2316
|
+
contextWindow: number;
|
|
2317
|
+
maxTokens: number;
|
|
2318
|
+
};
|
|
2319
|
+
readonly "gpt-5.4-nano": {
|
|
2320
|
+
id: string;
|
|
2321
|
+
name: string;
|
|
2322
|
+
api: "azure-openai-responses";
|
|
2323
|
+
provider: string;
|
|
2324
|
+
baseUrl: string;
|
|
2325
|
+
reasoning: true;
|
|
2326
|
+
input: ("image" | "text")[];
|
|
2327
|
+
cost: {
|
|
2328
|
+
input: number;
|
|
2329
|
+
output: number;
|
|
2330
|
+
cacheRead: number;
|
|
2331
|
+
cacheWrite: number;
|
|
2332
|
+
};
|
|
2333
|
+
contextWindow: number;
|
|
2334
|
+
maxTokens: number;
|
|
2335
|
+
};
|
|
2302
2336
|
readonly "gpt-5.4-pro": {
|
|
2303
2337
|
id: string;
|
|
2304
2338
|
name: string;
|
|
@@ -3060,6 +3094,29 @@ export declare const MODELS: {
|
|
|
3060
3094
|
contextWindow: number;
|
|
3061
3095
|
maxTokens: number;
|
|
3062
3096
|
};
|
|
3097
|
+
readonly "gpt-5.4-mini": {
|
|
3098
|
+
id: string;
|
|
3099
|
+
name: string;
|
|
3100
|
+
api: "openai-responses";
|
|
3101
|
+
provider: string;
|
|
3102
|
+
baseUrl: string;
|
|
3103
|
+
headers: {
|
|
3104
|
+
"User-Agent": string;
|
|
3105
|
+
"Editor-Version": string;
|
|
3106
|
+
"Editor-Plugin-Version": string;
|
|
3107
|
+
"Copilot-Integration-Id": string;
|
|
3108
|
+
};
|
|
3109
|
+
reasoning: true;
|
|
3110
|
+
input: ("image" | "text")[];
|
|
3111
|
+
cost: {
|
|
3112
|
+
input: number;
|
|
3113
|
+
output: number;
|
|
3114
|
+
cacheRead: number;
|
|
3115
|
+
cacheWrite: number;
|
|
3116
|
+
};
|
|
3117
|
+
contextWindow: number;
|
|
3118
|
+
maxTokens: number;
|
|
3119
|
+
};
|
|
3063
3120
|
readonly "grok-code-fast-1": {
|
|
3064
3121
|
id: string;
|
|
3065
3122
|
name: string;
|
|
@@ -4688,6 +4745,40 @@ export declare const MODELS: {
|
|
|
4688
4745
|
contextWindow: number;
|
|
4689
4746
|
maxTokens: number;
|
|
4690
4747
|
};
|
|
4748
|
+
readonly "MiniMax-M2.7": {
|
|
4749
|
+
id: string;
|
|
4750
|
+
name: string;
|
|
4751
|
+
api: "anthropic-messages";
|
|
4752
|
+
provider: string;
|
|
4753
|
+
baseUrl: string;
|
|
4754
|
+
reasoning: true;
|
|
4755
|
+
input: "text"[];
|
|
4756
|
+
cost: {
|
|
4757
|
+
input: number;
|
|
4758
|
+
output: number;
|
|
4759
|
+
cacheRead: number;
|
|
4760
|
+
cacheWrite: number;
|
|
4761
|
+
};
|
|
4762
|
+
contextWindow: number;
|
|
4763
|
+
maxTokens: number;
|
|
4764
|
+
};
|
|
4765
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4766
|
+
id: string;
|
|
4767
|
+
name: string;
|
|
4768
|
+
api: "anthropic-messages";
|
|
4769
|
+
provider: string;
|
|
4770
|
+
baseUrl: string;
|
|
4771
|
+
reasoning: true;
|
|
4772
|
+
input: "text"[];
|
|
4773
|
+
cost: {
|
|
4774
|
+
input: number;
|
|
4775
|
+
output: number;
|
|
4776
|
+
cacheRead: number;
|
|
4777
|
+
cacheWrite: number;
|
|
4778
|
+
};
|
|
4779
|
+
contextWindow: number;
|
|
4780
|
+
maxTokens: number;
|
|
4781
|
+
};
|
|
4691
4782
|
};
|
|
4692
4783
|
readonly "minimax-cn": {
|
|
4693
4784
|
readonly "MiniMax-M2": {
|
|
@@ -4758,6 +4849,40 @@ export declare const MODELS: {
|
|
|
4758
4849
|
contextWindow: number;
|
|
4759
4850
|
maxTokens: number;
|
|
4760
4851
|
};
|
|
4852
|
+
readonly "MiniMax-M2.7": {
|
|
4853
|
+
id: string;
|
|
4854
|
+
name: string;
|
|
4855
|
+
api: "anthropic-messages";
|
|
4856
|
+
provider: string;
|
|
4857
|
+
baseUrl: string;
|
|
4858
|
+
reasoning: true;
|
|
4859
|
+
input: "text"[];
|
|
4860
|
+
cost: {
|
|
4861
|
+
input: number;
|
|
4862
|
+
output: number;
|
|
4863
|
+
cacheRead: number;
|
|
4864
|
+
cacheWrite: number;
|
|
4865
|
+
};
|
|
4866
|
+
contextWindow: number;
|
|
4867
|
+
maxTokens: number;
|
|
4868
|
+
};
|
|
4869
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4870
|
+
id: string;
|
|
4871
|
+
name: string;
|
|
4872
|
+
api: "anthropic-messages";
|
|
4873
|
+
provider: string;
|
|
4874
|
+
baseUrl: string;
|
|
4875
|
+
reasoning: true;
|
|
4876
|
+
input: "text"[];
|
|
4877
|
+
cost: {
|
|
4878
|
+
input: number;
|
|
4879
|
+
output: number;
|
|
4880
|
+
cacheRead: number;
|
|
4881
|
+
cacheWrite: number;
|
|
4882
|
+
};
|
|
4883
|
+
contextWindow: number;
|
|
4884
|
+
maxTokens: number;
|
|
4885
|
+
};
|
|
4761
4886
|
};
|
|
4762
4887
|
readonly mistral: {
|
|
4763
4888
|
readonly "codestral-latest": {
|
|
@@ -5680,6 +5805,40 @@ export declare const MODELS: {
|
|
|
5680
5805
|
contextWindow: number;
|
|
5681
5806
|
maxTokens: number;
|
|
5682
5807
|
};
|
|
5808
|
+
readonly "gpt-5.4-mini": {
|
|
5809
|
+
id: string;
|
|
5810
|
+
name: string;
|
|
5811
|
+
api: "openai-responses";
|
|
5812
|
+
provider: string;
|
|
5813
|
+
baseUrl: string;
|
|
5814
|
+
reasoning: true;
|
|
5815
|
+
input: ("image" | "text")[];
|
|
5816
|
+
cost: {
|
|
5817
|
+
input: number;
|
|
5818
|
+
output: number;
|
|
5819
|
+
cacheRead: number;
|
|
5820
|
+
cacheWrite: number;
|
|
5821
|
+
};
|
|
5822
|
+
contextWindow: number;
|
|
5823
|
+
maxTokens: number;
|
|
5824
|
+
};
|
|
5825
|
+
readonly "gpt-5.4-nano": {
|
|
5826
|
+
id: string;
|
|
5827
|
+
name: string;
|
|
5828
|
+
api: "openai-responses";
|
|
5829
|
+
provider: string;
|
|
5830
|
+
baseUrl: string;
|
|
5831
|
+
reasoning: true;
|
|
5832
|
+
input: ("image" | "text")[];
|
|
5833
|
+
cost: {
|
|
5834
|
+
input: number;
|
|
5835
|
+
output: number;
|
|
5836
|
+
cacheRead: number;
|
|
5837
|
+
cacheWrite: number;
|
|
5838
|
+
};
|
|
5839
|
+
contextWindow: number;
|
|
5840
|
+
maxTokens: number;
|
|
5841
|
+
};
|
|
5683
5842
|
readonly "gpt-5.4-pro": {
|
|
5684
5843
|
id: string;
|
|
5685
5844
|
name: string;
|
|
@@ -5971,6 +6130,23 @@ export declare const MODELS: {
|
|
|
5971
6130
|
contextWindow: number;
|
|
5972
6131
|
maxTokens: number;
|
|
5973
6132
|
};
|
|
6133
|
+
readonly "gpt-5.4-mini": {
|
|
6134
|
+
id: string;
|
|
6135
|
+
name: string;
|
|
6136
|
+
api: "openai-codex-responses";
|
|
6137
|
+
provider: string;
|
|
6138
|
+
baseUrl: string;
|
|
6139
|
+
reasoning: true;
|
|
6140
|
+
input: ("image" | "text")[];
|
|
6141
|
+
cost: {
|
|
6142
|
+
input: number;
|
|
6143
|
+
output: number;
|
|
6144
|
+
cacheRead: number;
|
|
6145
|
+
cacheWrite: number;
|
|
6146
|
+
};
|
|
6147
|
+
contextWindow: number;
|
|
6148
|
+
maxTokens: number;
|
|
6149
|
+
};
|
|
5974
6150
|
};
|
|
5975
6151
|
readonly opencode: {
|
|
5976
6152
|
readonly "big-pickle": {
|
|
@@ -6364,6 +6540,40 @@ export declare const MODELS: {
|
|
|
6364
6540
|
contextWindow: number;
|
|
6365
6541
|
maxTokens: number;
|
|
6366
6542
|
};
|
|
6543
|
+
readonly "gpt-5.4-mini": {
|
|
6544
|
+
id: string;
|
|
6545
|
+
name: string;
|
|
6546
|
+
api: "openai-responses";
|
|
6547
|
+
provider: string;
|
|
6548
|
+
baseUrl: string;
|
|
6549
|
+
reasoning: true;
|
|
6550
|
+
input: ("image" | "text")[];
|
|
6551
|
+
cost: {
|
|
6552
|
+
input: number;
|
|
6553
|
+
output: number;
|
|
6554
|
+
cacheRead: number;
|
|
6555
|
+
cacheWrite: number;
|
|
6556
|
+
};
|
|
6557
|
+
contextWindow: number;
|
|
6558
|
+
maxTokens: number;
|
|
6559
|
+
};
|
|
6560
|
+
readonly "gpt-5.4-nano": {
|
|
6561
|
+
id: string;
|
|
6562
|
+
name: string;
|
|
6563
|
+
api: "openai-responses";
|
|
6564
|
+
provider: string;
|
|
6565
|
+
baseUrl: string;
|
|
6566
|
+
reasoning: true;
|
|
6567
|
+
input: ("image" | "text")[];
|
|
6568
|
+
cost: {
|
|
6569
|
+
input: number;
|
|
6570
|
+
output: number;
|
|
6571
|
+
cacheRead: number;
|
|
6572
|
+
cacheWrite: number;
|
|
6573
|
+
};
|
|
6574
|
+
contextWindow: number;
|
|
6575
|
+
maxTokens: number;
|
|
6576
|
+
};
|
|
6367
6577
|
readonly "gpt-5.4-pro": {
|
|
6368
6578
|
id: string;
|
|
6369
6579
|
name: string;
|
|
@@ -6398,7 +6608,24 @@ export declare const MODELS: {
|
|
|
6398
6608
|
contextWindow: number;
|
|
6399
6609
|
maxTokens: number;
|
|
6400
6610
|
};
|
|
6401
|
-
readonly "mimo-v2-
|
|
6611
|
+
readonly "mimo-v2-omni-free": {
|
|
6612
|
+
id: string;
|
|
6613
|
+
name: string;
|
|
6614
|
+
api: "openai-completions";
|
|
6615
|
+
provider: string;
|
|
6616
|
+
baseUrl: string;
|
|
6617
|
+
reasoning: true;
|
|
6618
|
+
input: ("image" | "text")[];
|
|
6619
|
+
cost: {
|
|
6620
|
+
input: number;
|
|
6621
|
+
output: number;
|
|
6622
|
+
cacheRead: number;
|
|
6623
|
+
cacheWrite: number;
|
|
6624
|
+
};
|
|
6625
|
+
contextWindow: number;
|
|
6626
|
+
maxTokens: number;
|
|
6627
|
+
};
|
|
6628
|
+
readonly "mimo-v2-pro-free": {
|
|
6402
6629
|
id: string;
|
|
6403
6630
|
name: string;
|
|
6404
6631
|
api: "openai-completions";
|
|
@@ -6519,6 +6746,23 @@ export declare const MODELS: {
|
|
|
6519
6746
|
contextWindow: number;
|
|
6520
6747
|
maxTokens: number;
|
|
6521
6748
|
};
|
|
6749
|
+
readonly "minimax-m2.7": {
|
|
6750
|
+
id: string;
|
|
6751
|
+
name: string;
|
|
6752
|
+
api: "anthropic-messages";
|
|
6753
|
+
provider: string;
|
|
6754
|
+
baseUrl: string;
|
|
6755
|
+
reasoning: true;
|
|
6756
|
+
input: "text"[];
|
|
6757
|
+
cost: {
|
|
6758
|
+
input: number;
|
|
6759
|
+
output: number;
|
|
6760
|
+
cacheRead: number;
|
|
6761
|
+
cacheWrite: number;
|
|
6762
|
+
};
|
|
6763
|
+
contextWindow: number;
|
|
6764
|
+
maxTokens: number;
|
|
6765
|
+
};
|
|
6522
6766
|
};
|
|
6523
6767
|
readonly openrouter: {
|
|
6524
6768
|
readonly "ai21/jamba-large-1.7": {
|
|
@@ -7473,40 +7717,6 @@ export declare const MODELS: {
|
|
|
7473
7717
|
contextWindow: number;
|
|
7474
7718
|
maxTokens: number;
|
|
7475
7719
|
};
|
|
7476
|
-
readonly "google/gemma-3-27b-it": {
|
|
7477
|
-
id: string;
|
|
7478
|
-
name: string;
|
|
7479
|
-
api: "openai-completions";
|
|
7480
|
-
provider: string;
|
|
7481
|
-
baseUrl: string;
|
|
7482
|
-
reasoning: false;
|
|
7483
|
-
input: ("image" | "text")[];
|
|
7484
|
-
cost: {
|
|
7485
|
-
input: number;
|
|
7486
|
-
output: number;
|
|
7487
|
-
cacheRead: number;
|
|
7488
|
-
cacheWrite: number;
|
|
7489
|
-
};
|
|
7490
|
-
contextWindow: number;
|
|
7491
|
-
maxTokens: number;
|
|
7492
|
-
};
|
|
7493
|
-
readonly "google/gemma-3-27b-it:free": {
|
|
7494
|
-
id: string;
|
|
7495
|
-
name: string;
|
|
7496
|
-
api: "openai-completions";
|
|
7497
|
-
provider: string;
|
|
7498
|
-
baseUrl: string;
|
|
7499
|
-
reasoning: false;
|
|
7500
|
-
input: ("image" | "text")[];
|
|
7501
|
-
cost: {
|
|
7502
|
-
input: number;
|
|
7503
|
-
output: number;
|
|
7504
|
-
cacheRead: number;
|
|
7505
|
-
cacheWrite: number;
|
|
7506
|
-
};
|
|
7507
|
-
contextWindow: number;
|
|
7508
|
-
maxTokens: number;
|
|
7509
|
-
};
|
|
7510
7720
|
readonly "inception/mercury": {
|
|
7511
7721
|
id: string;
|
|
7512
7722
|
name: string;
|
|
@@ -7796,6 +8006,23 @@ export declare const MODELS: {
|
|
|
7796
8006
|
contextWindow: number;
|
|
7797
8007
|
maxTokens: number;
|
|
7798
8008
|
};
|
|
8009
|
+
readonly "minimax/minimax-m2.7": {
|
|
8010
|
+
id: string;
|
|
8011
|
+
name: string;
|
|
8012
|
+
api: "openai-completions";
|
|
8013
|
+
provider: string;
|
|
8014
|
+
baseUrl: string;
|
|
8015
|
+
reasoning: true;
|
|
8016
|
+
input: "text"[];
|
|
8017
|
+
cost: {
|
|
8018
|
+
input: number;
|
|
8019
|
+
output: number;
|
|
8020
|
+
cacheRead: number;
|
|
8021
|
+
cacheWrite: number;
|
|
8022
|
+
};
|
|
8023
|
+
contextWindow: number;
|
|
8024
|
+
maxTokens: number;
|
|
8025
|
+
};
|
|
7799
8026
|
readonly "mistralai/codestral-2508": {
|
|
7800
8027
|
id: string;
|
|
7801
8028
|
name: string;
|
|
@@ -8357,6 +8584,23 @@ export declare const MODELS: {
|
|
|
8357
8584
|
contextWindow: number;
|
|
8358
8585
|
maxTokens: number;
|
|
8359
8586
|
};
|
|
8587
|
+
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
8588
|
+
id: string;
|
|
8589
|
+
name: string;
|
|
8590
|
+
api: "openai-completions";
|
|
8591
|
+
provider: string;
|
|
8592
|
+
baseUrl: string;
|
|
8593
|
+
reasoning: true;
|
|
8594
|
+
input: "text"[];
|
|
8595
|
+
cost: {
|
|
8596
|
+
input: number;
|
|
8597
|
+
output: number;
|
|
8598
|
+
cacheRead: number;
|
|
8599
|
+
cacheWrite: number;
|
|
8600
|
+
};
|
|
8601
|
+
contextWindow: number;
|
|
8602
|
+
maxTokens: number;
|
|
8603
|
+
};
|
|
8360
8604
|
readonly "nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8361
8605
|
id: string;
|
|
8362
8606
|
name: string;
|
|
@@ -9394,40 +9638,6 @@ export declare const MODELS: {
|
|
|
9394
9638
|
contextWindow: number;
|
|
9395
9639
|
maxTokens: number;
|
|
9396
9640
|
};
|
|
9397
|
-
readonly "openrouter/healer-alpha": {
|
|
9398
|
-
id: string;
|
|
9399
|
-
name: string;
|
|
9400
|
-
api: "openai-completions";
|
|
9401
|
-
provider: string;
|
|
9402
|
-
baseUrl: string;
|
|
9403
|
-
reasoning: true;
|
|
9404
|
-
input: ("image" | "text")[];
|
|
9405
|
-
cost: {
|
|
9406
|
-
input: number;
|
|
9407
|
-
output: number;
|
|
9408
|
-
cacheRead: number;
|
|
9409
|
-
cacheWrite: number;
|
|
9410
|
-
};
|
|
9411
|
-
contextWindow: number;
|
|
9412
|
-
maxTokens: number;
|
|
9413
|
-
};
|
|
9414
|
-
readonly "openrouter/hunter-alpha": {
|
|
9415
|
-
id: string;
|
|
9416
|
-
name: string;
|
|
9417
|
-
api: "openai-completions";
|
|
9418
|
-
provider: string;
|
|
9419
|
-
baseUrl: string;
|
|
9420
|
-
reasoning: true;
|
|
9421
|
-
input: ("image" | "text")[];
|
|
9422
|
-
cost: {
|
|
9423
|
-
input: number;
|
|
9424
|
-
output: number;
|
|
9425
|
-
cacheRead: number;
|
|
9426
|
-
cacheWrite: number;
|
|
9427
|
-
};
|
|
9428
|
-
contextWindow: number;
|
|
9429
|
-
maxTokens: number;
|
|
9430
|
-
};
|
|
9431
9641
|
readonly "prime-intellect/intellect-3": {
|
|
9432
9642
|
id: string;
|
|
9433
9643
|
name: string;
|
|
@@ -10516,6 +10726,40 @@ export declare const MODELS: {
|
|
|
10516
10726
|
contextWindow: number;
|
|
10517
10727
|
maxTokens: number;
|
|
10518
10728
|
};
|
|
10729
|
+
readonly "xiaomi/mimo-v2-omni": {
|
|
10730
|
+
id: string;
|
|
10731
|
+
name: string;
|
|
10732
|
+
api: "openai-completions";
|
|
10733
|
+
provider: string;
|
|
10734
|
+
baseUrl: string;
|
|
10735
|
+
reasoning: true;
|
|
10736
|
+
input: ("image" | "text")[];
|
|
10737
|
+
cost: {
|
|
10738
|
+
input: number;
|
|
10739
|
+
output: number;
|
|
10740
|
+
cacheRead: number;
|
|
10741
|
+
cacheWrite: number;
|
|
10742
|
+
};
|
|
10743
|
+
contextWindow: number;
|
|
10744
|
+
maxTokens: number;
|
|
10745
|
+
};
|
|
10746
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
10747
|
+
id: string;
|
|
10748
|
+
name: string;
|
|
10749
|
+
api: "openai-completions";
|
|
10750
|
+
provider: string;
|
|
10751
|
+
baseUrl: string;
|
|
10752
|
+
reasoning: true;
|
|
10753
|
+
input: "text"[];
|
|
10754
|
+
cost: {
|
|
10755
|
+
input: number;
|
|
10756
|
+
output: number;
|
|
10757
|
+
cacheRead: number;
|
|
10758
|
+
cacheWrite: number;
|
|
10759
|
+
};
|
|
10760
|
+
contextWindow: number;
|
|
10761
|
+
maxTokens: number;
|
|
10762
|
+
};
|
|
10519
10763
|
readonly "z-ai/glm-4-32b": {
|
|
10520
10764
|
id: string;
|
|
10521
10765
|
name: string;
|
|
@@ -11759,6 +12003,40 @@ export declare const MODELS: {
|
|
|
11759
12003
|
contextWindow: number;
|
|
11760
12004
|
maxTokens: number;
|
|
11761
12005
|
};
|
|
12006
|
+
readonly "minimax/minimax-m2.7": {
|
|
12007
|
+
id: string;
|
|
12008
|
+
name: string;
|
|
12009
|
+
api: "anthropic-messages";
|
|
12010
|
+
provider: string;
|
|
12011
|
+
baseUrl: string;
|
|
12012
|
+
reasoning: true;
|
|
12013
|
+
input: ("image" | "text")[];
|
|
12014
|
+
cost: {
|
|
12015
|
+
input: number;
|
|
12016
|
+
output: number;
|
|
12017
|
+
cacheRead: number;
|
|
12018
|
+
cacheWrite: number;
|
|
12019
|
+
};
|
|
12020
|
+
contextWindow: number;
|
|
12021
|
+
maxTokens: number;
|
|
12022
|
+
};
|
|
12023
|
+
readonly "minimax/minimax-m2.7-highspeed": {
|
|
12024
|
+
id: string;
|
|
12025
|
+
name: string;
|
|
12026
|
+
api: "anthropic-messages";
|
|
12027
|
+
provider: string;
|
|
12028
|
+
baseUrl: string;
|
|
12029
|
+
reasoning: true;
|
|
12030
|
+
input: ("image" | "text")[];
|
|
12031
|
+
cost: {
|
|
12032
|
+
input: number;
|
|
12033
|
+
output: number;
|
|
12034
|
+
cacheRead: number;
|
|
12035
|
+
cacheWrite: number;
|
|
12036
|
+
};
|
|
12037
|
+
contextWindow: number;
|
|
12038
|
+
maxTokens: number;
|
|
12039
|
+
};
|
|
11762
12040
|
readonly "mistral/codestral": {
|
|
11763
12041
|
id: string;
|
|
11764
12042
|
name: string;
|
|
@@ -12473,6 +12751,40 @@ export declare const MODELS: {
|
|
|
12473
12751
|
contextWindow: number;
|
|
12474
12752
|
maxTokens: number;
|
|
12475
12753
|
};
|
|
12754
|
+
readonly "openai/gpt-5.4-mini": {
|
|
12755
|
+
id: string;
|
|
12756
|
+
name: string;
|
|
12757
|
+
api: "anthropic-messages";
|
|
12758
|
+
provider: string;
|
|
12759
|
+
baseUrl: string;
|
|
12760
|
+
reasoning: true;
|
|
12761
|
+
input: ("image" | "text")[];
|
|
12762
|
+
cost: {
|
|
12763
|
+
input: number;
|
|
12764
|
+
output: number;
|
|
12765
|
+
cacheRead: number;
|
|
12766
|
+
cacheWrite: number;
|
|
12767
|
+
};
|
|
12768
|
+
contextWindow: number;
|
|
12769
|
+
maxTokens: number;
|
|
12770
|
+
};
|
|
12771
|
+
readonly "openai/gpt-5.4-nano": {
|
|
12772
|
+
id: string;
|
|
12773
|
+
name: string;
|
|
12774
|
+
api: "anthropic-messages";
|
|
12775
|
+
provider: string;
|
|
12776
|
+
baseUrl: string;
|
|
12777
|
+
reasoning: true;
|
|
12778
|
+
input: ("image" | "text")[];
|
|
12779
|
+
cost: {
|
|
12780
|
+
input: number;
|
|
12781
|
+
output: number;
|
|
12782
|
+
cacheRead: number;
|
|
12783
|
+
cacheWrite: number;
|
|
12784
|
+
};
|
|
12785
|
+
contextWindow: number;
|
|
12786
|
+
maxTokens: number;
|
|
12787
|
+
};
|
|
12476
12788
|
readonly "openai/gpt-5.4-pro": {
|
|
12477
12789
|
id: string;
|
|
12478
12790
|
name: string;
|
|
@@ -12932,6 +13244,23 @@ export declare const MODELS: {
|
|
|
12932
13244
|
contextWindow: number;
|
|
12933
13245
|
maxTokens: number;
|
|
12934
13246
|
};
|
|
13247
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
13248
|
+
id: string;
|
|
13249
|
+
name: string;
|
|
13250
|
+
api: "anthropic-messages";
|
|
13251
|
+
provider: string;
|
|
13252
|
+
baseUrl: string;
|
|
13253
|
+
reasoning: true;
|
|
13254
|
+
input: "text"[];
|
|
13255
|
+
cost: {
|
|
13256
|
+
input: number;
|
|
13257
|
+
output: number;
|
|
13258
|
+
cacheRead: number;
|
|
13259
|
+
cacheWrite: number;
|
|
13260
|
+
};
|
|
13261
|
+
contextWindow: number;
|
|
13262
|
+
maxTokens: number;
|
|
13263
|
+
};
|
|
12935
13264
|
readonly "zai/glm-4.5": {
|
|
12936
13265
|
id: string;
|
|
12937
13266
|
name: string;
|
|
@@ -13478,23 +13807,6 @@ export declare const MODELS: {
|
|
|
13478
13807
|
contextWindow: number;
|
|
13479
13808
|
maxTokens: number;
|
|
13480
13809
|
};
|
|
13481
|
-
readonly "grok-4.20-multi-agent-beta-latest": {
|
|
13482
|
-
id: string;
|
|
13483
|
-
name: string;
|
|
13484
|
-
api: "openai-completions";
|
|
13485
|
-
provider: string;
|
|
13486
|
-
baseUrl: string;
|
|
13487
|
-
reasoning: true;
|
|
13488
|
-
input: ("image" | "text")[];
|
|
13489
|
-
cost: {
|
|
13490
|
-
input: number;
|
|
13491
|
-
output: number;
|
|
13492
|
-
cacheRead: number;
|
|
13493
|
-
cacheWrite: number;
|
|
13494
|
-
};
|
|
13495
|
-
contextWindow: number;
|
|
13496
|
-
maxTokens: number;
|
|
13497
|
-
};
|
|
13498
13810
|
readonly "grok-beta": {
|
|
13499
13811
|
id: string;
|
|
13500
13812
|
name: string;
|