@mariozechner/pi-ai 0.60.0 → 0.61.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +307 -109
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +358 -155
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +13 -1
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +8 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +12 -1
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/types.d.ts +2 -2
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/validation.d.ts.map +1 -1
- package/dist/utils/validation.js +17 -9
- package/dist/utils/validation.js.map +1 -1
- package/package.json +1 -1
|
@@ -3117,29 +3117,6 @@ export declare const MODELS: {
|
|
|
3117
3117
|
contextWindow: number;
|
|
3118
3118
|
maxTokens: number;
|
|
3119
3119
|
};
|
|
3120
|
-
readonly "gpt-5.4-nano": {
|
|
3121
|
-
id: string;
|
|
3122
|
-
name: string;
|
|
3123
|
-
api: "openai-responses";
|
|
3124
|
-
provider: string;
|
|
3125
|
-
baseUrl: string;
|
|
3126
|
-
headers: {
|
|
3127
|
-
"User-Agent": string;
|
|
3128
|
-
"Editor-Version": string;
|
|
3129
|
-
"Editor-Plugin-Version": string;
|
|
3130
|
-
"Copilot-Integration-Id": string;
|
|
3131
|
-
};
|
|
3132
|
-
reasoning: true;
|
|
3133
|
-
input: ("image" | "text")[];
|
|
3134
|
-
cost: {
|
|
3135
|
-
input: number;
|
|
3136
|
-
output: number;
|
|
3137
|
-
cacheRead: number;
|
|
3138
|
-
cacheWrite: number;
|
|
3139
|
-
};
|
|
3140
|
-
contextWindow: number;
|
|
3141
|
-
maxTokens: number;
|
|
3142
|
-
};
|
|
3143
3120
|
readonly "grok-code-fast-1": {
|
|
3144
3121
|
id: string;
|
|
3145
3122
|
name: string;
|
|
@@ -4734,6 +4711,23 @@ export declare const MODELS: {
|
|
|
4734
4711
|
contextWindow: number;
|
|
4735
4712
|
maxTokens: number;
|
|
4736
4713
|
};
|
|
4714
|
+
readonly "MiniMax-M2.1-highspeed": {
|
|
4715
|
+
id: string;
|
|
4716
|
+
name: string;
|
|
4717
|
+
api: "anthropic-messages";
|
|
4718
|
+
provider: string;
|
|
4719
|
+
baseUrl: string;
|
|
4720
|
+
reasoning: true;
|
|
4721
|
+
input: "text"[];
|
|
4722
|
+
cost: {
|
|
4723
|
+
input: number;
|
|
4724
|
+
output: number;
|
|
4725
|
+
cacheRead: number;
|
|
4726
|
+
cacheWrite: number;
|
|
4727
|
+
};
|
|
4728
|
+
contextWindow: number;
|
|
4729
|
+
maxTokens: number;
|
|
4730
|
+
};
|
|
4737
4731
|
readonly "MiniMax-M2.5": {
|
|
4738
4732
|
id: string;
|
|
4739
4733
|
name: string;
|
|
@@ -4768,6 +4762,40 @@ export declare const MODELS: {
|
|
|
4768
4762
|
contextWindow: number;
|
|
4769
4763
|
maxTokens: number;
|
|
4770
4764
|
};
|
|
4765
|
+
readonly "MiniMax-M2.7": {
|
|
4766
|
+
id: string;
|
|
4767
|
+
name: string;
|
|
4768
|
+
api: "anthropic-messages";
|
|
4769
|
+
provider: string;
|
|
4770
|
+
baseUrl: string;
|
|
4771
|
+
reasoning: true;
|
|
4772
|
+
input: "text"[];
|
|
4773
|
+
cost: {
|
|
4774
|
+
input: number;
|
|
4775
|
+
output: number;
|
|
4776
|
+
cacheRead: number;
|
|
4777
|
+
cacheWrite: number;
|
|
4778
|
+
};
|
|
4779
|
+
contextWindow: number;
|
|
4780
|
+
maxTokens: number;
|
|
4781
|
+
};
|
|
4782
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4783
|
+
id: string;
|
|
4784
|
+
name: string;
|
|
4785
|
+
api: "anthropic-messages";
|
|
4786
|
+
provider: string;
|
|
4787
|
+
baseUrl: string;
|
|
4788
|
+
reasoning: true;
|
|
4789
|
+
input: "text"[];
|
|
4790
|
+
cost: {
|
|
4791
|
+
input: number;
|
|
4792
|
+
output: number;
|
|
4793
|
+
cacheRead: number;
|
|
4794
|
+
cacheWrite: number;
|
|
4795
|
+
};
|
|
4796
|
+
contextWindow: number;
|
|
4797
|
+
maxTokens: number;
|
|
4798
|
+
};
|
|
4771
4799
|
};
|
|
4772
4800
|
readonly "minimax-cn": {
|
|
4773
4801
|
readonly "MiniMax-M2": {
|
|
@@ -4804,6 +4832,23 @@ export declare const MODELS: {
|
|
|
4804
4832
|
contextWindow: number;
|
|
4805
4833
|
maxTokens: number;
|
|
4806
4834
|
};
|
|
4835
|
+
readonly "MiniMax-M2.1-highspeed": {
|
|
4836
|
+
id: string;
|
|
4837
|
+
name: string;
|
|
4838
|
+
api: "anthropic-messages";
|
|
4839
|
+
provider: string;
|
|
4840
|
+
baseUrl: string;
|
|
4841
|
+
reasoning: true;
|
|
4842
|
+
input: "text"[];
|
|
4843
|
+
cost: {
|
|
4844
|
+
input: number;
|
|
4845
|
+
output: number;
|
|
4846
|
+
cacheRead: number;
|
|
4847
|
+
cacheWrite: number;
|
|
4848
|
+
};
|
|
4849
|
+
contextWindow: number;
|
|
4850
|
+
maxTokens: number;
|
|
4851
|
+
};
|
|
4807
4852
|
readonly "MiniMax-M2.5": {
|
|
4808
4853
|
id: string;
|
|
4809
4854
|
name: string;
|
|
@@ -4838,6 +4883,40 @@ export declare const MODELS: {
|
|
|
4838
4883
|
contextWindow: number;
|
|
4839
4884
|
maxTokens: number;
|
|
4840
4885
|
};
|
|
4886
|
+
readonly "MiniMax-M2.7": {
|
|
4887
|
+
id: string;
|
|
4888
|
+
name: string;
|
|
4889
|
+
api: "anthropic-messages";
|
|
4890
|
+
provider: string;
|
|
4891
|
+
baseUrl: string;
|
|
4892
|
+
reasoning: true;
|
|
4893
|
+
input: "text"[];
|
|
4894
|
+
cost: {
|
|
4895
|
+
input: number;
|
|
4896
|
+
output: number;
|
|
4897
|
+
cacheRead: number;
|
|
4898
|
+
cacheWrite: number;
|
|
4899
|
+
};
|
|
4900
|
+
contextWindow: number;
|
|
4901
|
+
maxTokens: number;
|
|
4902
|
+
};
|
|
4903
|
+
readonly "MiniMax-M2.7-highspeed": {
|
|
4904
|
+
id: string;
|
|
4905
|
+
name: string;
|
|
4906
|
+
api: "anthropic-messages";
|
|
4907
|
+
provider: string;
|
|
4908
|
+
baseUrl: string;
|
|
4909
|
+
reasoning: true;
|
|
4910
|
+
input: "text"[];
|
|
4911
|
+
cost: {
|
|
4912
|
+
input: number;
|
|
4913
|
+
output: number;
|
|
4914
|
+
cacheRead: number;
|
|
4915
|
+
cacheWrite: number;
|
|
4916
|
+
};
|
|
4917
|
+
contextWindow: number;
|
|
4918
|
+
maxTokens: number;
|
|
4919
|
+
};
|
|
4841
4920
|
};
|
|
4842
4921
|
readonly mistral: {
|
|
4843
4922
|
readonly "codestral-latest": {
|
|
@@ -6085,6 +6164,23 @@ export declare const MODELS: {
|
|
|
6085
6164
|
contextWindow: number;
|
|
6086
6165
|
maxTokens: number;
|
|
6087
6166
|
};
|
|
6167
|
+
readonly "gpt-5.4-mini": {
|
|
6168
|
+
id: string;
|
|
6169
|
+
name: string;
|
|
6170
|
+
api: "openai-codex-responses";
|
|
6171
|
+
provider: string;
|
|
6172
|
+
baseUrl: string;
|
|
6173
|
+
reasoning: true;
|
|
6174
|
+
input: ("image" | "text")[];
|
|
6175
|
+
cost: {
|
|
6176
|
+
input: number;
|
|
6177
|
+
output: number;
|
|
6178
|
+
cacheRead: number;
|
|
6179
|
+
cacheWrite: number;
|
|
6180
|
+
};
|
|
6181
|
+
contextWindow: number;
|
|
6182
|
+
maxTokens: number;
|
|
6183
|
+
};
|
|
6088
6184
|
};
|
|
6089
6185
|
readonly opencode: {
|
|
6090
6186
|
readonly "big-pickle": {
|
|
@@ -6478,6 +6574,40 @@ export declare const MODELS: {
|
|
|
6478
6574
|
contextWindow: number;
|
|
6479
6575
|
maxTokens: number;
|
|
6480
6576
|
};
|
|
6577
|
+
readonly "gpt-5.4-mini": {
|
|
6578
|
+
id: string;
|
|
6579
|
+
name: string;
|
|
6580
|
+
api: "openai-responses";
|
|
6581
|
+
provider: string;
|
|
6582
|
+
baseUrl: string;
|
|
6583
|
+
reasoning: true;
|
|
6584
|
+
input: ("image" | "text")[];
|
|
6585
|
+
cost: {
|
|
6586
|
+
input: number;
|
|
6587
|
+
output: number;
|
|
6588
|
+
cacheRead: number;
|
|
6589
|
+
cacheWrite: number;
|
|
6590
|
+
};
|
|
6591
|
+
contextWindow: number;
|
|
6592
|
+
maxTokens: number;
|
|
6593
|
+
};
|
|
6594
|
+
readonly "gpt-5.4-nano": {
|
|
6595
|
+
id: string;
|
|
6596
|
+
name: string;
|
|
6597
|
+
api: "openai-responses";
|
|
6598
|
+
provider: string;
|
|
6599
|
+
baseUrl: string;
|
|
6600
|
+
reasoning: true;
|
|
6601
|
+
input: ("image" | "text")[];
|
|
6602
|
+
cost: {
|
|
6603
|
+
input: number;
|
|
6604
|
+
output: number;
|
|
6605
|
+
cacheRead: number;
|
|
6606
|
+
cacheWrite: number;
|
|
6607
|
+
};
|
|
6608
|
+
contextWindow: number;
|
|
6609
|
+
maxTokens: number;
|
|
6610
|
+
};
|
|
6481
6611
|
readonly "gpt-5.4-pro": {
|
|
6482
6612
|
id: string;
|
|
6483
6613
|
name: string;
|
|
@@ -6512,7 +6642,24 @@ export declare const MODELS: {
|
|
|
6512
6642
|
contextWindow: number;
|
|
6513
6643
|
maxTokens: number;
|
|
6514
6644
|
};
|
|
6515
|
-
readonly "mimo-v2-
|
|
6645
|
+
readonly "mimo-v2-omni-free": {
|
|
6646
|
+
id: string;
|
|
6647
|
+
name: string;
|
|
6648
|
+
api: "openai-completions";
|
|
6649
|
+
provider: string;
|
|
6650
|
+
baseUrl: string;
|
|
6651
|
+
reasoning: true;
|
|
6652
|
+
input: ("image" | "text")[];
|
|
6653
|
+
cost: {
|
|
6654
|
+
input: number;
|
|
6655
|
+
output: number;
|
|
6656
|
+
cacheRead: number;
|
|
6657
|
+
cacheWrite: number;
|
|
6658
|
+
};
|
|
6659
|
+
contextWindow: number;
|
|
6660
|
+
maxTokens: number;
|
|
6661
|
+
};
|
|
6662
|
+
readonly "mimo-v2-pro-free": {
|
|
6516
6663
|
id: string;
|
|
6517
6664
|
name: string;
|
|
6518
6665
|
api: "openai-completions";
|
|
@@ -6633,6 +6780,23 @@ export declare const MODELS: {
|
|
|
6633
6780
|
contextWindow: number;
|
|
6634
6781
|
maxTokens: number;
|
|
6635
6782
|
};
|
|
6783
|
+
readonly "minimax-m2.7": {
|
|
6784
|
+
id: string;
|
|
6785
|
+
name: string;
|
|
6786
|
+
api: "anthropic-messages";
|
|
6787
|
+
provider: string;
|
|
6788
|
+
baseUrl: string;
|
|
6789
|
+
reasoning: true;
|
|
6790
|
+
input: "text"[];
|
|
6791
|
+
cost: {
|
|
6792
|
+
input: number;
|
|
6793
|
+
output: number;
|
|
6794
|
+
cacheRead: number;
|
|
6795
|
+
cacheWrite: number;
|
|
6796
|
+
};
|
|
6797
|
+
contextWindow: number;
|
|
6798
|
+
maxTokens: number;
|
|
6799
|
+
};
|
|
6636
6800
|
};
|
|
6637
6801
|
readonly openrouter: {
|
|
6638
6802
|
readonly "ai21/jamba-large-1.7": {
|
|
@@ -7587,40 +7751,6 @@ export declare const MODELS: {
|
|
|
7587
7751
|
contextWindow: number;
|
|
7588
7752
|
maxTokens: number;
|
|
7589
7753
|
};
|
|
7590
|
-
readonly "google/gemma-3-27b-it": {
|
|
7591
|
-
id: string;
|
|
7592
|
-
name: string;
|
|
7593
|
-
api: "openai-completions";
|
|
7594
|
-
provider: string;
|
|
7595
|
-
baseUrl: string;
|
|
7596
|
-
reasoning: false;
|
|
7597
|
-
input: ("image" | "text")[];
|
|
7598
|
-
cost: {
|
|
7599
|
-
input: number;
|
|
7600
|
-
output: number;
|
|
7601
|
-
cacheRead: number;
|
|
7602
|
-
cacheWrite: number;
|
|
7603
|
-
};
|
|
7604
|
-
contextWindow: number;
|
|
7605
|
-
maxTokens: number;
|
|
7606
|
-
};
|
|
7607
|
-
readonly "google/gemma-3-27b-it:free": {
|
|
7608
|
-
id: string;
|
|
7609
|
-
name: string;
|
|
7610
|
-
api: "openai-completions";
|
|
7611
|
-
provider: string;
|
|
7612
|
-
baseUrl: string;
|
|
7613
|
-
reasoning: false;
|
|
7614
|
-
input: ("image" | "text")[];
|
|
7615
|
-
cost: {
|
|
7616
|
-
input: number;
|
|
7617
|
-
output: number;
|
|
7618
|
-
cacheRead: number;
|
|
7619
|
-
cacheWrite: number;
|
|
7620
|
-
};
|
|
7621
|
-
contextWindow: number;
|
|
7622
|
-
maxTokens: number;
|
|
7623
|
-
};
|
|
7624
7754
|
readonly "inception/mercury": {
|
|
7625
7755
|
id: string;
|
|
7626
7756
|
name: string;
|
|
@@ -7910,6 +8040,23 @@ export declare const MODELS: {
|
|
|
7910
8040
|
contextWindow: number;
|
|
7911
8041
|
maxTokens: number;
|
|
7912
8042
|
};
|
|
8043
|
+
readonly "minimax/minimax-m2.7": {
|
|
8044
|
+
id: string;
|
|
8045
|
+
name: string;
|
|
8046
|
+
api: "openai-completions";
|
|
8047
|
+
provider: string;
|
|
8048
|
+
baseUrl: string;
|
|
8049
|
+
reasoning: true;
|
|
8050
|
+
input: "text"[];
|
|
8051
|
+
cost: {
|
|
8052
|
+
input: number;
|
|
8053
|
+
output: number;
|
|
8054
|
+
cacheRead: number;
|
|
8055
|
+
cacheWrite: number;
|
|
8056
|
+
};
|
|
8057
|
+
contextWindow: number;
|
|
8058
|
+
maxTokens: number;
|
|
8059
|
+
};
|
|
7913
8060
|
readonly "mistralai/codestral-2508": {
|
|
7914
8061
|
id: string;
|
|
7915
8062
|
name: string;
|
|
@@ -8471,6 +8618,23 @@ export declare const MODELS: {
|
|
|
8471
8618
|
contextWindow: number;
|
|
8472
8619
|
maxTokens: number;
|
|
8473
8620
|
};
|
|
8621
|
+
readonly "nvidia/nemotron-3-super-120b-a12b": {
|
|
8622
|
+
id: string;
|
|
8623
|
+
name: string;
|
|
8624
|
+
api: "openai-completions";
|
|
8625
|
+
provider: string;
|
|
8626
|
+
baseUrl: string;
|
|
8627
|
+
reasoning: true;
|
|
8628
|
+
input: "text"[];
|
|
8629
|
+
cost: {
|
|
8630
|
+
input: number;
|
|
8631
|
+
output: number;
|
|
8632
|
+
cacheRead: number;
|
|
8633
|
+
cacheWrite: number;
|
|
8634
|
+
};
|
|
8635
|
+
contextWindow: number;
|
|
8636
|
+
maxTokens: number;
|
|
8637
|
+
};
|
|
8474
8638
|
readonly "nvidia/nemotron-3-super-120b-a12b:free": {
|
|
8475
8639
|
id: string;
|
|
8476
8640
|
name: string;
|
|
@@ -9508,40 +9672,6 @@ export declare const MODELS: {
|
|
|
9508
9672
|
contextWindow: number;
|
|
9509
9673
|
maxTokens: number;
|
|
9510
9674
|
};
|
|
9511
|
-
readonly "openrouter/healer-alpha": {
|
|
9512
|
-
id: string;
|
|
9513
|
-
name: string;
|
|
9514
|
-
api: "openai-completions";
|
|
9515
|
-
provider: string;
|
|
9516
|
-
baseUrl: string;
|
|
9517
|
-
reasoning: true;
|
|
9518
|
-
input: ("image" | "text")[];
|
|
9519
|
-
cost: {
|
|
9520
|
-
input: number;
|
|
9521
|
-
output: number;
|
|
9522
|
-
cacheRead: number;
|
|
9523
|
-
cacheWrite: number;
|
|
9524
|
-
};
|
|
9525
|
-
contextWindow: number;
|
|
9526
|
-
maxTokens: number;
|
|
9527
|
-
};
|
|
9528
|
-
readonly "openrouter/hunter-alpha": {
|
|
9529
|
-
id: string;
|
|
9530
|
-
name: string;
|
|
9531
|
-
api: "openai-completions";
|
|
9532
|
-
provider: string;
|
|
9533
|
-
baseUrl: string;
|
|
9534
|
-
reasoning: true;
|
|
9535
|
-
input: ("image" | "text")[];
|
|
9536
|
-
cost: {
|
|
9537
|
-
input: number;
|
|
9538
|
-
output: number;
|
|
9539
|
-
cacheRead: number;
|
|
9540
|
-
cacheWrite: number;
|
|
9541
|
-
};
|
|
9542
|
-
contextWindow: number;
|
|
9543
|
-
maxTokens: number;
|
|
9544
|
-
};
|
|
9545
9675
|
readonly "prime-intellect/intellect-3": {
|
|
9546
9676
|
id: string;
|
|
9547
9677
|
name: string;
|
|
@@ -10630,6 +10760,40 @@ export declare const MODELS: {
|
|
|
10630
10760
|
contextWindow: number;
|
|
10631
10761
|
maxTokens: number;
|
|
10632
10762
|
};
|
|
10763
|
+
readonly "xiaomi/mimo-v2-omni": {
|
|
10764
|
+
id: string;
|
|
10765
|
+
name: string;
|
|
10766
|
+
api: "openai-completions";
|
|
10767
|
+
provider: string;
|
|
10768
|
+
baseUrl: string;
|
|
10769
|
+
reasoning: true;
|
|
10770
|
+
input: ("image" | "text")[];
|
|
10771
|
+
cost: {
|
|
10772
|
+
input: number;
|
|
10773
|
+
output: number;
|
|
10774
|
+
cacheRead: number;
|
|
10775
|
+
cacheWrite: number;
|
|
10776
|
+
};
|
|
10777
|
+
contextWindow: number;
|
|
10778
|
+
maxTokens: number;
|
|
10779
|
+
};
|
|
10780
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
10781
|
+
id: string;
|
|
10782
|
+
name: string;
|
|
10783
|
+
api: "openai-completions";
|
|
10784
|
+
provider: string;
|
|
10785
|
+
baseUrl: string;
|
|
10786
|
+
reasoning: true;
|
|
10787
|
+
input: "text"[];
|
|
10788
|
+
cost: {
|
|
10789
|
+
input: number;
|
|
10790
|
+
output: number;
|
|
10791
|
+
cacheRead: number;
|
|
10792
|
+
cacheWrite: number;
|
|
10793
|
+
};
|
|
10794
|
+
contextWindow: number;
|
|
10795
|
+
maxTokens: number;
|
|
10796
|
+
};
|
|
10633
10797
|
readonly "z-ai/glm-4-32b": {
|
|
10634
10798
|
id: string;
|
|
10635
10799
|
name: string;
|
|
@@ -11873,6 +12037,40 @@ export declare const MODELS: {
|
|
|
11873
12037
|
contextWindow: number;
|
|
11874
12038
|
maxTokens: number;
|
|
11875
12039
|
};
|
|
12040
|
+
readonly "minimax/minimax-m2.7": {
|
|
12041
|
+
id: string;
|
|
12042
|
+
name: string;
|
|
12043
|
+
api: "anthropic-messages";
|
|
12044
|
+
provider: string;
|
|
12045
|
+
baseUrl: string;
|
|
12046
|
+
reasoning: true;
|
|
12047
|
+
input: ("image" | "text")[];
|
|
12048
|
+
cost: {
|
|
12049
|
+
input: number;
|
|
12050
|
+
output: number;
|
|
12051
|
+
cacheRead: number;
|
|
12052
|
+
cacheWrite: number;
|
|
12053
|
+
};
|
|
12054
|
+
contextWindow: number;
|
|
12055
|
+
maxTokens: number;
|
|
12056
|
+
};
|
|
12057
|
+
readonly "minimax/minimax-m2.7-highspeed": {
|
|
12058
|
+
id: string;
|
|
12059
|
+
name: string;
|
|
12060
|
+
api: "anthropic-messages";
|
|
12061
|
+
provider: string;
|
|
12062
|
+
baseUrl: string;
|
|
12063
|
+
reasoning: true;
|
|
12064
|
+
input: ("image" | "text")[];
|
|
12065
|
+
cost: {
|
|
12066
|
+
input: number;
|
|
12067
|
+
output: number;
|
|
12068
|
+
cacheRead: number;
|
|
12069
|
+
cacheWrite: number;
|
|
12070
|
+
};
|
|
12071
|
+
contextWindow: number;
|
|
12072
|
+
maxTokens: number;
|
|
12073
|
+
};
|
|
11876
12074
|
readonly "mistral/codestral": {
|
|
11877
12075
|
id: string;
|
|
11878
12076
|
name: string;
|
|
@@ -13080,6 +13278,23 @@ export declare const MODELS: {
|
|
|
13080
13278
|
contextWindow: number;
|
|
13081
13279
|
maxTokens: number;
|
|
13082
13280
|
};
|
|
13281
|
+
readonly "xiaomi/mimo-v2-pro": {
|
|
13282
|
+
id: string;
|
|
13283
|
+
name: string;
|
|
13284
|
+
api: "anthropic-messages";
|
|
13285
|
+
provider: string;
|
|
13286
|
+
baseUrl: string;
|
|
13287
|
+
reasoning: true;
|
|
13288
|
+
input: "text"[];
|
|
13289
|
+
cost: {
|
|
13290
|
+
input: number;
|
|
13291
|
+
output: number;
|
|
13292
|
+
cacheRead: number;
|
|
13293
|
+
cacheWrite: number;
|
|
13294
|
+
};
|
|
13295
|
+
contextWindow: number;
|
|
13296
|
+
maxTokens: number;
|
|
13297
|
+
};
|
|
13083
13298
|
readonly "zai/glm-4.5": {
|
|
13084
13299
|
id: string;
|
|
13085
13300
|
name: string;
|
|
@@ -13626,23 +13841,6 @@ export declare const MODELS: {
|
|
|
13626
13841
|
contextWindow: number;
|
|
13627
13842
|
maxTokens: number;
|
|
13628
13843
|
};
|
|
13629
|
-
readonly "grok-4.20-multi-agent-beta-latest": {
|
|
13630
|
-
id: string;
|
|
13631
|
-
name: string;
|
|
13632
|
-
api: "openai-completions";
|
|
13633
|
-
provider: string;
|
|
13634
|
-
baseUrl: string;
|
|
13635
|
-
reasoning: true;
|
|
13636
|
-
input: ("image" | "text")[];
|
|
13637
|
-
cost: {
|
|
13638
|
-
input: number;
|
|
13639
|
-
output: number;
|
|
13640
|
-
cacheRead: number;
|
|
13641
|
-
cacheWrite: number;
|
|
13642
|
-
};
|
|
13643
|
-
contextWindow: number;
|
|
13644
|
-
maxTokens: number;
|
|
13645
|
-
};
|
|
13646
13844
|
readonly "grok-beta": {
|
|
13647
13845
|
id: string;
|
|
13648
13846
|
name: string;
|