@mariozechner/pi-ai 0.56.1 → 0.56.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +4 -3
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -0
- package/dist/index.js.map +1 -1
- package/dist/models.d.ts +1 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +285 -35
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +327 -87
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +2 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js +25 -15
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +16 -22
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/mistral.d.ts +22 -0
- package/dist/providers/mistral.d.ts.map +1 -0
- package/dist/providers/mistral.js +495 -0
- package/dist/providers/mistral.js.map +1 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +1 -1
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +8 -33
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +28 -13
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/register-builtins.d.ts.map +1 -1
- package/dist/providers/register-builtins.js +6 -0
- package/dist/providers/register-builtins.js.map +1 -1
- package/dist/types.d.ts +6 -3
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/hash.d.ts +3 -0
- package/dist/utils/hash.d.ts.map +1 -0
- package/dist/utils/hash.js +14 -0
- package/dist/utils/hash.js.map +1 -0
- package/dist/utils/overflow.d.ts +1 -1
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +4 -3
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +5 -7
- package/oauth.d.ts +0 -1
- package/oauth.js +0 -1
|
@@ -2282,6 +2282,40 @@ export declare const MODELS: {
|
|
|
2282
2282
|
contextWindow: number;
|
|
2283
2283
|
maxTokens: number;
|
|
2284
2284
|
};
|
|
2285
|
+
readonly "gpt-5.4": {
|
|
2286
|
+
id: string;
|
|
2287
|
+
name: string;
|
|
2288
|
+
api: "azure-openai-responses";
|
|
2289
|
+
provider: string;
|
|
2290
|
+
baseUrl: string;
|
|
2291
|
+
reasoning: true;
|
|
2292
|
+
input: ("image" | "text")[];
|
|
2293
|
+
cost: {
|
|
2294
|
+
input: number;
|
|
2295
|
+
output: number;
|
|
2296
|
+
cacheRead: number;
|
|
2297
|
+
cacheWrite: number;
|
|
2298
|
+
};
|
|
2299
|
+
contextWindow: number;
|
|
2300
|
+
maxTokens: number;
|
|
2301
|
+
};
|
|
2302
|
+
readonly "gpt-5.4-pro": {
|
|
2303
|
+
id: string;
|
|
2304
|
+
name: string;
|
|
2305
|
+
api: "azure-openai-responses";
|
|
2306
|
+
provider: string;
|
|
2307
|
+
baseUrl: string;
|
|
2308
|
+
reasoning: true;
|
|
2309
|
+
input: ("image" | "text")[];
|
|
2310
|
+
cost: {
|
|
2311
|
+
input: number;
|
|
2312
|
+
output: number;
|
|
2313
|
+
cacheRead: number;
|
|
2314
|
+
cacheWrite: number;
|
|
2315
|
+
};
|
|
2316
|
+
contextWindow: number;
|
|
2317
|
+
maxTokens: number;
|
|
2318
|
+
};
|
|
2285
2319
|
readonly o1: {
|
|
2286
2320
|
id: string;
|
|
2287
2321
|
name: string;
|
|
@@ -2980,6 +3014,52 @@ export declare const MODELS: {
|
|
|
2980
3014
|
contextWindow: number;
|
|
2981
3015
|
maxTokens: number;
|
|
2982
3016
|
};
|
|
3017
|
+
readonly "gpt-5.3-codex": {
|
|
3018
|
+
id: string;
|
|
3019
|
+
name: string;
|
|
3020
|
+
api: "openai-responses";
|
|
3021
|
+
provider: string;
|
|
3022
|
+
baseUrl: string;
|
|
3023
|
+
headers: {
|
|
3024
|
+
"User-Agent": string;
|
|
3025
|
+
"Editor-Version": string;
|
|
3026
|
+
"Editor-Plugin-Version": string;
|
|
3027
|
+
"Copilot-Integration-Id": string;
|
|
3028
|
+
};
|
|
3029
|
+
reasoning: true;
|
|
3030
|
+
input: ("image" | "text")[];
|
|
3031
|
+
cost: {
|
|
3032
|
+
input: number;
|
|
3033
|
+
output: number;
|
|
3034
|
+
cacheRead: number;
|
|
3035
|
+
cacheWrite: number;
|
|
3036
|
+
};
|
|
3037
|
+
contextWindow: number;
|
|
3038
|
+
maxTokens: number;
|
|
3039
|
+
};
|
|
3040
|
+
readonly "gpt-5.4": {
|
|
3041
|
+
id: string;
|
|
3042
|
+
name: string;
|
|
3043
|
+
api: "openai-responses";
|
|
3044
|
+
provider: string;
|
|
3045
|
+
baseUrl: string;
|
|
3046
|
+
headers: {
|
|
3047
|
+
"User-Agent": string;
|
|
3048
|
+
"Editor-Version": string;
|
|
3049
|
+
"Editor-Plugin-Version": string;
|
|
3050
|
+
"Copilot-Integration-Id": string;
|
|
3051
|
+
};
|
|
3052
|
+
reasoning: true;
|
|
3053
|
+
input: ("image" | "text")[];
|
|
3054
|
+
cost: {
|
|
3055
|
+
input: number;
|
|
3056
|
+
output: number;
|
|
3057
|
+
cacheRead: number;
|
|
3058
|
+
cacheWrite: number;
|
|
3059
|
+
};
|
|
3060
|
+
contextWindow: number;
|
|
3061
|
+
maxTokens: number;
|
|
3062
|
+
};
|
|
2983
3063
|
readonly "grok-code-fast-1": {
|
|
2984
3064
|
id: string;
|
|
2985
3065
|
name: string;
|
|
@@ -3488,6 +3568,23 @@ export declare const MODELS: {
|
|
|
3488
3568
|
contextWindow: number;
|
|
3489
3569
|
maxTokens: number;
|
|
3490
3570
|
};
|
|
3571
|
+
readonly "claude-sonnet-4-6": {
|
|
3572
|
+
id: string;
|
|
3573
|
+
name: string;
|
|
3574
|
+
api: "google-gemini-cli";
|
|
3575
|
+
provider: string;
|
|
3576
|
+
baseUrl: string;
|
|
3577
|
+
reasoning: true;
|
|
3578
|
+
input: ("image" | "text")[];
|
|
3579
|
+
cost: {
|
|
3580
|
+
input: number;
|
|
3581
|
+
output: number;
|
|
3582
|
+
cacheRead: number;
|
|
3583
|
+
cacheWrite: number;
|
|
3584
|
+
};
|
|
3585
|
+
contextWindow: number;
|
|
3586
|
+
maxTokens: number;
|
|
3587
|
+
};
|
|
3491
3588
|
readonly "gemini-3-flash": {
|
|
3492
3589
|
id: string;
|
|
3493
3590
|
name: string;
|
|
@@ -4666,7 +4763,7 @@ export declare const MODELS: {
|
|
|
4666
4763
|
readonly "codestral-latest": {
|
|
4667
4764
|
id: string;
|
|
4668
4765
|
name: string;
|
|
4669
|
-
api: "
|
|
4766
|
+
api: "mistral-conversations";
|
|
4670
4767
|
provider: string;
|
|
4671
4768
|
baseUrl: string;
|
|
4672
4769
|
reasoning: false;
|
|
@@ -4683,7 +4780,7 @@ export declare const MODELS: {
|
|
|
4683
4780
|
readonly "devstral-2512": {
|
|
4684
4781
|
id: string;
|
|
4685
4782
|
name: string;
|
|
4686
|
-
api: "
|
|
4783
|
+
api: "mistral-conversations";
|
|
4687
4784
|
provider: string;
|
|
4688
4785
|
baseUrl: string;
|
|
4689
4786
|
reasoning: false;
|
|
@@ -4700,7 +4797,7 @@ export declare const MODELS: {
|
|
|
4700
4797
|
readonly "devstral-medium-2507": {
|
|
4701
4798
|
id: string;
|
|
4702
4799
|
name: string;
|
|
4703
|
-
api: "
|
|
4800
|
+
api: "mistral-conversations";
|
|
4704
4801
|
provider: string;
|
|
4705
4802
|
baseUrl: string;
|
|
4706
4803
|
reasoning: false;
|
|
@@ -4717,7 +4814,7 @@ export declare const MODELS: {
|
|
|
4717
4814
|
readonly "devstral-medium-latest": {
|
|
4718
4815
|
id: string;
|
|
4719
4816
|
name: string;
|
|
4720
|
-
api: "
|
|
4817
|
+
api: "mistral-conversations";
|
|
4721
4818
|
provider: string;
|
|
4722
4819
|
baseUrl: string;
|
|
4723
4820
|
reasoning: false;
|
|
@@ -4734,7 +4831,7 @@ export declare const MODELS: {
|
|
|
4734
4831
|
readonly "devstral-small-2505": {
|
|
4735
4832
|
id: string;
|
|
4736
4833
|
name: string;
|
|
4737
|
-
api: "
|
|
4834
|
+
api: "mistral-conversations";
|
|
4738
4835
|
provider: string;
|
|
4739
4836
|
baseUrl: string;
|
|
4740
4837
|
reasoning: false;
|
|
@@ -4751,7 +4848,7 @@ export declare const MODELS: {
|
|
|
4751
4848
|
readonly "devstral-small-2507": {
|
|
4752
4849
|
id: string;
|
|
4753
4850
|
name: string;
|
|
4754
|
-
api: "
|
|
4851
|
+
api: "mistral-conversations";
|
|
4755
4852
|
provider: string;
|
|
4756
4853
|
baseUrl: string;
|
|
4757
4854
|
reasoning: false;
|
|
@@ -4768,7 +4865,7 @@ export declare const MODELS: {
|
|
|
4768
4865
|
readonly "labs-devstral-small-2512": {
|
|
4769
4866
|
id: string;
|
|
4770
4867
|
name: string;
|
|
4771
|
-
api: "
|
|
4868
|
+
api: "mistral-conversations";
|
|
4772
4869
|
provider: string;
|
|
4773
4870
|
baseUrl: string;
|
|
4774
4871
|
reasoning: false;
|
|
@@ -4785,7 +4882,7 @@ export declare const MODELS: {
|
|
|
4785
4882
|
readonly "magistral-medium-latest": {
|
|
4786
4883
|
id: string;
|
|
4787
4884
|
name: string;
|
|
4788
|
-
api: "
|
|
4885
|
+
api: "mistral-conversations";
|
|
4789
4886
|
provider: string;
|
|
4790
4887
|
baseUrl: string;
|
|
4791
4888
|
reasoning: true;
|
|
@@ -4802,7 +4899,7 @@ export declare const MODELS: {
|
|
|
4802
4899
|
readonly "magistral-small": {
|
|
4803
4900
|
id: string;
|
|
4804
4901
|
name: string;
|
|
4805
|
-
api: "
|
|
4902
|
+
api: "mistral-conversations";
|
|
4806
4903
|
provider: string;
|
|
4807
4904
|
baseUrl: string;
|
|
4808
4905
|
reasoning: true;
|
|
@@ -4819,7 +4916,7 @@ export declare const MODELS: {
|
|
|
4819
4916
|
readonly "ministral-3b-latest": {
|
|
4820
4917
|
id: string;
|
|
4821
4918
|
name: string;
|
|
4822
|
-
api: "
|
|
4919
|
+
api: "mistral-conversations";
|
|
4823
4920
|
provider: string;
|
|
4824
4921
|
baseUrl: string;
|
|
4825
4922
|
reasoning: false;
|
|
@@ -4836,7 +4933,7 @@ export declare const MODELS: {
|
|
|
4836
4933
|
readonly "ministral-8b-latest": {
|
|
4837
4934
|
id: string;
|
|
4838
4935
|
name: string;
|
|
4839
|
-
api: "
|
|
4936
|
+
api: "mistral-conversations";
|
|
4840
4937
|
provider: string;
|
|
4841
4938
|
baseUrl: string;
|
|
4842
4939
|
reasoning: false;
|
|
@@ -4853,7 +4950,7 @@ export declare const MODELS: {
|
|
|
4853
4950
|
readonly "mistral-large-2411": {
|
|
4854
4951
|
id: string;
|
|
4855
4952
|
name: string;
|
|
4856
|
-
api: "
|
|
4953
|
+
api: "mistral-conversations";
|
|
4857
4954
|
provider: string;
|
|
4858
4955
|
baseUrl: string;
|
|
4859
4956
|
reasoning: false;
|
|
@@ -4870,7 +4967,7 @@ export declare const MODELS: {
|
|
|
4870
4967
|
readonly "mistral-large-2512": {
|
|
4871
4968
|
id: string;
|
|
4872
4969
|
name: string;
|
|
4873
|
-
api: "
|
|
4970
|
+
api: "mistral-conversations";
|
|
4874
4971
|
provider: string;
|
|
4875
4972
|
baseUrl: string;
|
|
4876
4973
|
reasoning: false;
|
|
@@ -4887,7 +4984,7 @@ export declare const MODELS: {
|
|
|
4887
4984
|
readonly "mistral-large-latest": {
|
|
4888
4985
|
id: string;
|
|
4889
4986
|
name: string;
|
|
4890
|
-
api: "
|
|
4987
|
+
api: "mistral-conversations";
|
|
4891
4988
|
provider: string;
|
|
4892
4989
|
baseUrl: string;
|
|
4893
4990
|
reasoning: false;
|
|
@@ -4904,7 +5001,7 @@ export declare const MODELS: {
|
|
|
4904
5001
|
readonly "mistral-medium-2505": {
|
|
4905
5002
|
id: string;
|
|
4906
5003
|
name: string;
|
|
4907
|
-
api: "
|
|
5004
|
+
api: "mistral-conversations";
|
|
4908
5005
|
provider: string;
|
|
4909
5006
|
baseUrl: string;
|
|
4910
5007
|
reasoning: false;
|
|
@@ -4921,7 +5018,7 @@ export declare const MODELS: {
|
|
|
4921
5018
|
readonly "mistral-medium-2508": {
|
|
4922
5019
|
id: string;
|
|
4923
5020
|
name: string;
|
|
4924
|
-
api: "
|
|
5021
|
+
api: "mistral-conversations";
|
|
4925
5022
|
provider: string;
|
|
4926
5023
|
baseUrl: string;
|
|
4927
5024
|
reasoning: false;
|
|
@@ -4938,7 +5035,7 @@ export declare const MODELS: {
|
|
|
4938
5035
|
readonly "mistral-medium-latest": {
|
|
4939
5036
|
id: string;
|
|
4940
5037
|
name: string;
|
|
4941
|
-
api: "
|
|
5038
|
+
api: "mistral-conversations";
|
|
4942
5039
|
provider: string;
|
|
4943
5040
|
baseUrl: string;
|
|
4944
5041
|
reasoning: false;
|
|
@@ -4955,7 +5052,7 @@ export declare const MODELS: {
|
|
|
4955
5052
|
readonly "mistral-nemo": {
|
|
4956
5053
|
id: string;
|
|
4957
5054
|
name: string;
|
|
4958
|
-
api: "
|
|
5055
|
+
api: "mistral-conversations";
|
|
4959
5056
|
provider: string;
|
|
4960
5057
|
baseUrl: string;
|
|
4961
5058
|
reasoning: false;
|
|
@@ -4972,7 +5069,7 @@ export declare const MODELS: {
|
|
|
4972
5069
|
readonly "mistral-small-2506": {
|
|
4973
5070
|
id: string;
|
|
4974
5071
|
name: string;
|
|
4975
|
-
api: "
|
|
5072
|
+
api: "mistral-conversations";
|
|
4976
5073
|
provider: string;
|
|
4977
5074
|
baseUrl: string;
|
|
4978
5075
|
reasoning: false;
|
|
@@ -4989,7 +5086,7 @@ export declare const MODELS: {
|
|
|
4989
5086
|
readonly "mistral-small-latest": {
|
|
4990
5087
|
id: string;
|
|
4991
5088
|
name: string;
|
|
4992
|
-
api: "
|
|
5089
|
+
api: "mistral-conversations";
|
|
4993
5090
|
provider: string;
|
|
4994
5091
|
baseUrl: string;
|
|
4995
5092
|
reasoning: false;
|
|
@@ -5006,7 +5103,7 @@ export declare const MODELS: {
|
|
|
5006
5103
|
readonly "open-mistral-7b": {
|
|
5007
5104
|
id: string;
|
|
5008
5105
|
name: string;
|
|
5009
|
-
api: "
|
|
5106
|
+
api: "mistral-conversations";
|
|
5010
5107
|
provider: string;
|
|
5011
5108
|
baseUrl: string;
|
|
5012
5109
|
reasoning: false;
|
|
@@ -5023,7 +5120,7 @@ export declare const MODELS: {
|
|
|
5023
5120
|
readonly "open-mixtral-8x22b": {
|
|
5024
5121
|
id: string;
|
|
5025
5122
|
name: string;
|
|
5026
|
-
api: "
|
|
5123
|
+
api: "mistral-conversations";
|
|
5027
5124
|
provider: string;
|
|
5028
5125
|
baseUrl: string;
|
|
5029
5126
|
reasoning: false;
|
|
@@ -5040,7 +5137,7 @@ export declare const MODELS: {
|
|
|
5040
5137
|
readonly "open-mixtral-8x7b": {
|
|
5041
5138
|
id: string;
|
|
5042
5139
|
name: string;
|
|
5043
|
-
api: "
|
|
5140
|
+
api: "mistral-conversations";
|
|
5044
5141
|
provider: string;
|
|
5045
5142
|
baseUrl: string;
|
|
5046
5143
|
reasoning: false;
|
|
@@ -5057,7 +5154,7 @@ export declare const MODELS: {
|
|
|
5057
5154
|
readonly "pixtral-12b": {
|
|
5058
5155
|
id: string;
|
|
5059
5156
|
name: string;
|
|
5060
|
-
api: "
|
|
5157
|
+
api: "mistral-conversations";
|
|
5061
5158
|
provider: string;
|
|
5062
5159
|
baseUrl: string;
|
|
5063
5160
|
reasoning: false;
|
|
@@ -5074,7 +5171,7 @@ export declare const MODELS: {
|
|
|
5074
5171
|
readonly "pixtral-large-latest": {
|
|
5075
5172
|
id: string;
|
|
5076
5173
|
name: string;
|
|
5077
|
-
api: "
|
|
5174
|
+
api: "mistral-conversations";
|
|
5078
5175
|
provider: string;
|
|
5079
5176
|
baseUrl: string;
|
|
5080
5177
|
reasoning: false;
|
|
@@ -5566,6 +5663,40 @@ export declare const MODELS: {
|
|
|
5566
5663
|
contextWindow: number;
|
|
5567
5664
|
maxTokens: number;
|
|
5568
5665
|
};
|
|
5666
|
+
readonly "gpt-5.4": {
|
|
5667
|
+
id: string;
|
|
5668
|
+
name: string;
|
|
5669
|
+
api: "openai-responses";
|
|
5670
|
+
provider: string;
|
|
5671
|
+
baseUrl: string;
|
|
5672
|
+
reasoning: true;
|
|
5673
|
+
input: ("image" | "text")[];
|
|
5674
|
+
cost: {
|
|
5675
|
+
input: number;
|
|
5676
|
+
output: number;
|
|
5677
|
+
cacheRead: number;
|
|
5678
|
+
cacheWrite: number;
|
|
5679
|
+
};
|
|
5680
|
+
contextWindow: number;
|
|
5681
|
+
maxTokens: number;
|
|
5682
|
+
};
|
|
5683
|
+
readonly "gpt-5.4-pro": {
|
|
5684
|
+
id: string;
|
|
5685
|
+
name: string;
|
|
5686
|
+
api: "openai-responses";
|
|
5687
|
+
provider: string;
|
|
5688
|
+
baseUrl: string;
|
|
5689
|
+
reasoning: true;
|
|
5690
|
+
input: ("image" | "text")[];
|
|
5691
|
+
cost: {
|
|
5692
|
+
input: number;
|
|
5693
|
+
output: number;
|
|
5694
|
+
cacheRead: number;
|
|
5695
|
+
cacheWrite: number;
|
|
5696
|
+
};
|
|
5697
|
+
contextWindow: number;
|
|
5698
|
+
maxTokens: number;
|
|
5699
|
+
};
|
|
5569
5700
|
readonly o1: {
|
|
5570
5701
|
id: string;
|
|
5571
5702
|
name: string;
|
|
@@ -5823,6 +5954,23 @@ export declare const MODELS: {
|
|
|
5823
5954
|
contextWindow: number;
|
|
5824
5955
|
maxTokens: number;
|
|
5825
5956
|
};
|
|
5957
|
+
readonly "gpt-5.4": {
|
|
5958
|
+
id: string;
|
|
5959
|
+
name: string;
|
|
5960
|
+
api: "openai-codex-responses";
|
|
5961
|
+
provider: string;
|
|
5962
|
+
baseUrl: string;
|
|
5963
|
+
reasoning: true;
|
|
5964
|
+
input: ("image" | "text")[];
|
|
5965
|
+
cost: {
|
|
5966
|
+
input: number;
|
|
5967
|
+
output: number;
|
|
5968
|
+
cacheRead: number;
|
|
5969
|
+
cacheWrite: number;
|
|
5970
|
+
};
|
|
5971
|
+
contextWindow: number;
|
|
5972
|
+
maxTokens: number;
|
|
5973
|
+
};
|
|
5826
5974
|
};
|
|
5827
5975
|
readonly opencode: {
|
|
5828
5976
|
readonly "big-pickle": {
|
|
@@ -6250,10 +6398,10 @@ export declare const MODELS: {
|
|
|
6250
6398
|
contextWindow: number;
|
|
6251
6399
|
maxTokens: number;
|
|
6252
6400
|
};
|
|
6253
|
-
readonly "
|
|
6401
|
+
readonly "gpt-5.4": {
|
|
6254
6402
|
id: string;
|
|
6255
6403
|
name: string;
|
|
6256
|
-
api: "openai-
|
|
6404
|
+
api: "openai-responses";
|
|
6257
6405
|
provider: string;
|
|
6258
6406
|
baseUrl: string;
|
|
6259
6407
|
reasoning: true;
|
|
@@ -6267,14 +6415,31 @@ export declare const MODELS: {
|
|
|
6267
6415
|
contextWindow: number;
|
|
6268
6416
|
maxTokens: number;
|
|
6269
6417
|
};
|
|
6270
|
-
readonly "
|
|
6418
|
+
readonly "gpt-5.4-pro": {
|
|
6419
|
+
id: string;
|
|
6420
|
+
name: string;
|
|
6421
|
+
api: "openai-responses";
|
|
6422
|
+
provider: string;
|
|
6423
|
+
baseUrl: string;
|
|
6424
|
+
reasoning: true;
|
|
6425
|
+
input: ("image" | "text")[];
|
|
6426
|
+
cost: {
|
|
6427
|
+
input: number;
|
|
6428
|
+
output: number;
|
|
6429
|
+
cacheRead: number;
|
|
6430
|
+
cacheWrite: number;
|
|
6431
|
+
};
|
|
6432
|
+
contextWindow: number;
|
|
6433
|
+
maxTokens: number;
|
|
6434
|
+
};
|
|
6435
|
+
readonly "kimi-k2.5": {
|
|
6271
6436
|
id: string;
|
|
6272
6437
|
name: string;
|
|
6273
6438
|
api: "openai-completions";
|
|
6274
6439
|
provider: string;
|
|
6275
6440
|
baseUrl: string;
|
|
6276
6441
|
reasoning: true;
|
|
6277
|
-
input: "text"[];
|
|
6442
|
+
input: ("image" | "text")[];
|
|
6278
6443
|
cost: {
|
|
6279
6444
|
input: number;
|
|
6280
6445
|
output: number;
|
|
@@ -6284,7 +6449,7 @@ export declare const MODELS: {
|
|
|
6284
6449
|
contextWindow: number;
|
|
6285
6450
|
maxTokens: number;
|
|
6286
6451
|
};
|
|
6287
|
-
readonly "minimax-m2.
|
|
6452
|
+
readonly "minimax-m2.1": {
|
|
6288
6453
|
id: string;
|
|
6289
6454
|
name: string;
|
|
6290
6455
|
api: "openai-completions";
|
|
@@ -6301,10 +6466,10 @@ export declare const MODELS: {
|
|
|
6301
6466
|
contextWindow: number;
|
|
6302
6467
|
maxTokens: number;
|
|
6303
6468
|
};
|
|
6304
|
-
readonly "minimax-m2.5
|
|
6469
|
+
readonly "minimax-m2.5": {
|
|
6305
6470
|
id: string;
|
|
6306
6471
|
name: string;
|
|
6307
|
-
api: "
|
|
6472
|
+
api: "openai-completions";
|
|
6308
6473
|
provider: string;
|
|
6309
6474
|
baseUrl: string;
|
|
6310
6475
|
reasoning: true;
|
|
@@ -6318,13 +6483,13 @@ export declare const MODELS: {
|
|
|
6318
6483
|
contextWindow: number;
|
|
6319
6484
|
maxTokens: number;
|
|
6320
6485
|
};
|
|
6321
|
-
readonly "
|
|
6486
|
+
readonly "minimax-m2.5-free": {
|
|
6322
6487
|
id: string;
|
|
6323
6488
|
name: string;
|
|
6324
|
-
api: "
|
|
6489
|
+
api: "anthropic-messages";
|
|
6325
6490
|
provider: string;
|
|
6326
6491
|
baseUrl: string;
|
|
6327
|
-
reasoning:
|
|
6492
|
+
reasoning: true;
|
|
6328
6493
|
input: "text"[];
|
|
6329
6494
|
cost: {
|
|
6330
6495
|
input: number;
|
|
@@ -7393,6 +7558,23 @@ export declare const MODELS: {
|
|
|
7393
7558
|
contextWindow: number;
|
|
7394
7559
|
maxTokens: number;
|
|
7395
7560
|
};
|
|
7561
|
+
readonly "inception/mercury-2": {
|
|
7562
|
+
id: string;
|
|
7563
|
+
name: string;
|
|
7564
|
+
api: "openai-completions";
|
|
7565
|
+
provider: string;
|
|
7566
|
+
baseUrl: string;
|
|
7567
|
+
reasoning: true;
|
|
7568
|
+
input: "text"[];
|
|
7569
|
+
cost: {
|
|
7570
|
+
input: number;
|
|
7571
|
+
output: number;
|
|
7572
|
+
cacheRead: number;
|
|
7573
|
+
cacheWrite: number;
|
|
7574
|
+
};
|
|
7575
|
+
contextWindow: number;
|
|
7576
|
+
maxTokens: number;
|
|
7577
|
+
};
|
|
7396
7578
|
readonly "inception/mercury-coder": {
|
|
7397
7579
|
id: string;
|
|
7398
7580
|
name: string;
|
|
@@ -8889,6 +9071,40 @@ export declare const MODELS: {
|
|
|
8889
9071
|
contextWindow: number;
|
|
8890
9072
|
maxTokens: number;
|
|
8891
9073
|
};
|
|
9074
|
+
readonly "openai/gpt-5.4": {
|
|
9075
|
+
id: string;
|
|
9076
|
+
name: string;
|
|
9077
|
+
api: "openai-completions";
|
|
9078
|
+
provider: string;
|
|
9079
|
+
baseUrl: string;
|
|
9080
|
+
reasoning: true;
|
|
9081
|
+
input: ("image" | "text")[];
|
|
9082
|
+
cost: {
|
|
9083
|
+
input: number;
|
|
9084
|
+
output: number;
|
|
9085
|
+
cacheRead: number;
|
|
9086
|
+
cacheWrite: number;
|
|
9087
|
+
};
|
|
9088
|
+
contextWindow: number;
|
|
9089
|
+
maxTokens: number;
|
|
9090
|
+
};
|
|
9091
|
+
readonly "openai/gpt-5.4-pro": {
|
|
9092
|
+
id: string;
|
|
9093
|
+
name: string;
|
|
9094
|
+
api: "openai-completions";
|
|
9095
|
+
provider: string;
|
|
9096
|
+
baseUrl: string;
|
|
9097
|
+
reasoning: true;
|
|
9098
|
+
input: ("image" | "text")[];
|
|
9099
|
+
cost: {
|
|
9100
|
+
input: number;
|
|
9101
|
+
output: number;
|
|
9102
|
+
cacheRead: number;
|
|
9103
|
+
cacheWrite: number;
|
|
9104
|
+
};
|
|
9105
|
+
contextWindow: number;
|
|
9106
|
+
maxTokens: number;
|
|
9107
|
+
};
|
|
8892
9108
|
readonly "openai/gpt-oss-120b": {
|
|
8893
9109
|
id: string;
|
|
8894
9110
|
name: string;
|
|
@@ -12138,6 +12354,40 @@ export declare const MODELS: {
|
|
|
12138
12354
|
contextWindow: number;
|
|
12139
12355
|
maxTokens: number;
|
|
12140
12356
|
};
|
|
12357
|
+
readonly "openai/gpt-5.4": {
|
|
12358
|
+
id: string;
|
|
12359
|
+
name: string;
|
|
12360
|
+
api: "anthropic-messages";
|
|
12361
|
+
provider: string;
|
|
12362
|
+
baseUrl: string;
|
|
12363
|
+
reasoning: true;
|
|
12364
|
+
input: ("image" | "text")[];
|
|
12365
|
+
cost: {
|
|
12366
|
+
input: number;
|
|
12367
|
+
output: number;
|
|
12368
|
+
cacheRead: number;
|
|
12369
|
+
cacheWrite: number;
|
|
12370
|
+
};
|
|
12371
|
+
contextWindow: number;
|
|
12372
|
+
maxTokens: number;
|
|
12373
|
+
};
|
|
12374
|
+
readonly "openai/gpt-5.4-pro": {
|
|
12375
|
+
id: string;
|
|
12376
|
+
name: string;
|
|
12377
|
+
api: "anthropic-messages";
|
|
12378
|
+
provider: string;
|
|
12379
|
+
baseUrl: string;
|
|
12380
|
+
reasoning: true;
|
|
12381
|
+
input: ("image" | "text")[];
|
|
12382
|
+
cost: {
|
|
12383
|
+
input: number;
|
|
12384
|
+
output: number;
|
|
12385
|
+
cacheRead: number;
|
|
12386
|
+
cacheWrite: number;
|
|
12387
|
+
};
|
|
12388
|
+
contextWindow: number;
|
|
12389
|
+
maxTokens: number;
|
|
12390
|
+
};
|
|
12141
12391
|
readonly "openai/gpt-oss-120b": {
|
|
12142
12392
|
id: string;
|
|
12143
12393
|
name: string;
|