@mariozechner/pi-ai 0.45.6 → 0.46.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +144 -68
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +143 -72
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +33 -10
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +22 -10
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +1 -1
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/stream.d.ts.map +1 -1
- package/dist/stream.js +1 -0
- package/dist/stream.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
|
@@ -1802,6 +1802,29 @@ export declare const MODELS: {
|
|
|
1802
1802
|
contextWindow: number;
|
|
1803
1803
|
maxTokens: number;
|
|
1804
1804
|
};
|
|
1805
|
+
readonly "gpt-5.2-codex": {
|
|
1806
|
+
id: string;
|
|
1807
|
+
name: string;
|
|
1808
|
+
api: "openai-responses";
|
|
1809
|
+
provider: string;
|
|
1810
|
+
baseUrl: string;
|
|
1811
|
+
headers: {
|
|
1812
|
+
"User-Agent": string;
|
|
1813
|
+
"Editor-Version": string;
|
|
1814
|
+
"Editor-Plugin-Version": string;
|
|
1815
|
+
"Copilot-Integration-Id": string;
|
|
1816
|
+
};
|
|
1817
|
+
reasoning: true;
|
|
1818
|
+
input: ("image" | "text")[];
|
|
1819
|
+
cost: {
|
|
1820
|
+
input: number;
|
|
1821
|
+
output: number;
|
|
1822
|
+
cacheRead: number;
|
|
1823
|
+
cacheWrite: number;
|
|
1824
|
+
};
|
|
1825
|
+
contextWindow: number;
|
|
1826
|
+
maxTokens: number;
|
|
1827
|
+
};
|
|
1805
1828
|
readonly "grok-code-fast-1": {
|
|
1806
1829
|
id: string;
|
|
1807
1830
|
name: string;
|
|
@@ -2903,6 +2926,42 @@ export declare const MODELS: {
|
|
|
2903
2926
|
maxTokens: number;
|
|
2904
2927
|
};
|
|
2905
2928
|
};
|
|
2929
|
+
readonly "minimax-cn": {
|
|
2930
|
+
readonly "MiniMax-M2": {
|
|
2931
|
+
id: string;
|
|
2932
|
+
name: string;
|
|
2933
|
+
api: "anthropic-messages";
|
|
2934
|
+
provider: string;
|
|
2935
|
+
baseUrl: string;
|
|
2936
|
+
reasoning: true;
|
|
2937
|
+
input: "text"[];
|
|
2938
|
+
cost: {
|
|
2939
|
+
input: number;
|
|
2940
|
+
output: number;
|
|
2941
|
+
cacheRead: number;
|
|
2942
|
+
cacheWrite: number;
|
|
2943
|
+
};
|
|
2944
|
+
contextWindow: number;
|
|
2945
|
+
maxTokens: number;
|
|
2946
|
+
};
|
|
2947
|
+
readonly "MiniMax-M2.1": {
|
|
2948
|
+
id: string;
|
|
2949
|
+
name: string;
|
|
2950
|
+
api: "anthropic-messages";
|
|
2951
|
+
provider: string;
|
|
2952
|
+
baseUrl: string;
|
|
2953
|
+
reasoning: true;
|
|
2954
|
+
input: "text"[];
|
|
2955
|
+
cost: {
|
|
2956
|
+
input: number;
|
|
2957
|
+
output: number;
|
|
2958
|
+
cacheRead: number;
|
|
2959
|
+
cacheWrite: number;
|
|
2960
|
+
};
|
|
2961
|
+
contextWindow: number;
|
|
2962
|
+
maxTokens: number;
|
|
2963
|
+
};
|
|
2964
|
+
};
|
|
2906
2965
|
readonly mistral: {
|
|
2907
2966
|
readonly "codestral-latest": {
|
|
2908
2967
|
id: string;
|
|
@@ -3739,6 +3798,23 @@ export declare const MODELS: {
|
|
|
3739
3798
|
contextWindow: number;
|
|
3740
3799
|
maxTokens: number;
|
|
3741
3800
|
};
|
|
3801
|
+
readonly "gpt-5.2-codex": {
|
|
3802
|
+
id: string;
|
|
3803
|
+
name: string;
|
|
3804
|
+
api: "openai-responses";
|
|
3805
|
+
provider: string;
|
|
3806
|
+
baseUrl: string;
|
|
3807
|
+
reasoning: true;
|
|
3808
|
+
input: ("image" | "text")[];
|
|
3809
|
+
cost: {
|
|
3810
|
+
input: number;
|
|
3811
|
+
output: number;
|
|
3812
|
+
cacheRead: number;
|
|
3813
|
+
cacheWrite: number;
|
|
3814
|
+
};
|
|
3815
|
+
contextWindow: number;
|
|
3816
|
+
maxTokens: number;
|
|
3817
|
+
};
|
|
3742
3818
|
readonly "gpt-5.2-pro": {
|
|
3743
3819
|
id: string;
|
|
3744
3820
|
name: string;
|
|
@@ -4338,6 +4414,23 @@ export declare const MODELS: {
|
|
|
4338
4414
|
contextWindow: number;
|
|
4339
4415
|
maxTokens: number;
|
|
4340
4416
|
};
|
|
4417
|
+
readonly "gpt-5.2-codex": {
|
|
4418
|
+
id: string;
|
|
4419
|
+
name: string;
|
|
4420
|
+
api: "openai-responses";
|
|
4421
|
+
provider: string;
|
|
4422
|
+
baseUrl: string;
|
|
4423
|
+
reasoning: true;
|
|
4424
|
+
input: ("image" | "text")[];
|
|
4425
|
+
cost: {
|
|
4426
|
+
input: number;
|
|
4427
|
+
output: number;
|
|
4428
|
+
cacheRead: number;
|
|
4429
|
+
cacheWrite: number;
|
|
4430
|
+
};
|
|
4431
|
+
contextWindow: number;
|
|
4432
|
+
maxTokens: number;
|
|
4433
|
+
};
|
|
4341
4434
|
readonly "grok-code": {
|
|
4342
4435
|
id: string;
|
|
4343
4436
|
name: string;
|
|
@@ -4629,23 +4722,6 @@ export declare const MODELS: {
|
|
|
4629
4722
|
contextWindow: number;
|
|
4630
4723
|
maxTokens: number;
|
|
4631
4724
|
};
|
|
4632
|
-
readonly "anthropic/claude-3.5-haiku-20241022": {
|
|
4633
|
-
id: string;
|
|
4634
|
-
name: string;
|
|
4635
|
-
api: "openai-completions";
|
|
4636
|
-
provider: string;
|
|
4637
|
-
baseUrl: string;
|
|
4638
|
-
reasoning: false;
|
|
4639
|
-
input: ("image" | "text")[];
|
|
4640
|
-
cost: {
|
|
4641
|
-
input: number;
|
|
4642
|
-
output: number;
|
|
4643
|
-
cacheRead: number;
|
|
4644
|
-
cacheWrite: number;
|
|
4645
|
-
};
|
|
4646
|
-
contextWindow: number;
|
|
4647
|
-
maxTokens: number;
|
|
4648
|
-
};
|
|
4649
4725
|
readonly "anthropic/claude-3.5-sonnet": {
|
|
4650
4726
|
id: string;
|
|
4651
4727
|
name: string;
|
|
@@ -5275,23 +5351,6 @@ export declare const MODELS: {
|
|
|
5275
5351
|
contextWindow: number;
|
|
5276
5352
|
maxTokens: number;
|
|
5277
5353
|
};
|
|
5278
|
-
readonly "google/gemini-2.5-flash-preview-09-2025": {
|
|
5279
|
-
id: string;
|
|
5280
|
-
name: string;
|
|
5281
|
-
api: "openai-completions";
|
|
5282
|
-
provider: string;
|
|
5283
|
-
baseUrl: string;
|
|
5284
|
-
reasoning: true;
|
|
5285
|
-
input: ("image" | "text")[];
|
|
5286
|
-
cost: {
|
|
5287
|
-
input: number;
|
|
5288
|
-
output: number;
|
|
5289
|
-
cacheRead: number;
|
|
5290
|
-
cacheWrite: number;
|
|
5291
|
-
};
|
|
5292
|
-
contextWindow: number;
|
|
5293
|
-
maxTokens: number;
|
|
5294
|
-
};
|
|
5295
5354
|
readonly "google/gemini-2.5-pro": {
|
|
5296
5355
|
id: string;
|
|
5297
5356
|
name: string;
|
|
@@ -5547,23 +5606,6 @@ export declare const MODELS: {
|
|
|
5547
5606
|
contextWindow: number;
|
|
5548
5607
|
maxTokens: number;
|
|
5549
5608
|
};
|
|
5550
|
-
readonly "meta-llama/llama-3.2-3b-instruct": {
|
|
5551
|
-
id: string;
|
|
5552
|
-
name: string;
|
|
5553
|
-
api: "openai-completions";
|
|
5554
|
-
provider: string;
|
|
5555
|
-
baseUrl: string;
|
|
5556
|
-
reasoning: false;
|
|
5557
|
-
input: "text"[];
|
|
5558
|
-
cost: {
|
|
5559
|
-
input: number;
|
|
5560
|
-
output: number;
|
|
5561
|
-
cacheRead: number;
|
|
5562
|
-
cacheWrite: number;
|
|
5563
|
-
};
|
|
5564
|
-
contextWindow: number;
|
|
5565
|
-
maxTokens: number;
|
|
5566
|
-
};
|
|
5567
5609
|
readonly "meta-llama/llama-3.3-70b-instruct": {
|
|
5568
5610
|
id: string;
|
|
5569
5611
|
name: string;
|
|
@@ -7026,6 +7068,23 @@ export declare const MODELS: {
|
|
|
7026
7068
|
contextWindow: number;
|
|
7027
7069
|
maxTokens: number;
|
|
7028
7070
|
};
|
|
7071
|
+
readonly "openai/gpt-5.2-codex": {
|
|
7072
|
+
id: string;
|
|
7073
|
+
name: string;
|
|
7074
|
+
api: "openai-completions";
|
|
7075
|
+
provider: string;
|
|
7076
|
+
baseUrl: string;
|
|
7077
|
+
reasoning: true;
|
|
7078
|
+
input: ("image" | "text")[];
|
|
7079
|
+
cost: {
|
|
7080
|
+
input: number;
|
|
7081
|
+
output: number;
|
|
7082
|
+
cacheRead: number;
|
|
7083
|
+
cacheWrite: number;
|
|
7084
|
+
};
|
|
7085
|
+
contextWindow: number;
|
|
7086
|
+
maxTokens: number;
|
|
7087
|
+
};
|
|
7029
7088
|
readonly "openai/gpt-5.2-pro": {
|
|
7030
7089
|
id: string;
|
|
7031
7090
|
name: string;
|
|
@@ -8199,6 +8258,23 @@ export declare const MODELS: {
|
|
|
8199
8258
|
contextWindow: number;
|
|
8200
8259
|
maxTokens: number;
|
|
8201
8260
|
};
|
|
8261
|
+
readonly "xiaomi/mimo-v2-flash": {
|
|
8262
|
+
id: string;
|
|
8263
|
+
name: string;
|
|
8264
|
+
api: "openai-completions";
|
|
8265
|
+
provider: string;
|
|
8266
|
+
baseUrl: string;
|
|
8267
|
+
reasoning: true;
|
|
8268
|
+
input: "text"[];
|
|
8269
|
+
cost: {
|
|
8270
|
+
input: number;
|
|
8271
|
+
output: number;
|
|
8272
|
+
cacheRead: number;
|
|
8273
|
+
cacheWrite: number;
|
|
8274
|
+
};
|
|
8275
|
+
contextWindow: number;
|
|
8276
|
+
maxTokens: number;
|
|
8277
|
+
};
|
|
8202
8278
|
readonly "xiaomi/mimo-v2-flash:free": {
|
|
8203
8279
|
id: string;
|
|
8204
8280
|
name: string;
|
|
@@ -9357,23 +9433,6 @@ export declare const MODELS: {
|
|
|
9357
9433
|
contextWindow: number;
|
|
9358
9434
|
maxTokens: number;
|
|
9359
9435
|
};
|
|
9360
|
-
readonly "mistral/mistral-nemo": {
|
|
9361
|
-
id: string;
|
|
9362
|
-
name: string;
|
|
9363
|
-
api: "anthropic-messages";
|
|
9364
|
-
provider: string;
|
|
9365
|
-
baseUrl: string;
|
|
9366
|
-
reasoning: false;
|
|
9367
|
-
input: "text"[];
|
|
9368
|
-
cost: {
|
|
9369
|
-
input: number;
|
|
9370
|
-
output: number;
|
|
9371
|
-
cacheRead: number;
|
|
9372
|
-
cacheWrite: number;
|
|
9373
|
-
};
|
|
9374
|
-
contextWindow: number;
|
|
9375
|
-
maxTokens: number;
|
|
9376
|
-
};
|
|
9377
9436
|
readonly "mistral/mistral-small": {
|
|
9378
9437
|
id: string;
|
|
9379
9438
|
name: string;
|
|
@@ -9867,6 +9926,23 @@ export declare const MODELS: {
|
|
|
9867
9926
|
contextWindow: number;
|
|
9868
9927
|
maxTokens: number;
|
|
9869
9928
|
};
|
|
9929
|
+
readonly "openai/gpt-5.2-codex": {
|
|
9930
|
+
id: string;
|
|
9931
|
+
name: string;
|
|
9932
|
+
api: "anthropic-messages";
|
|
9933
|
+
provider: string;
|
|
9934
|
+
baseUrl: string;
|
|
9935
|
+
reasoning: true;
|
|
9936
|
+
input: ("image" | "text")[];
|
|
9937
|
+
cost: {
|
|
9938
|
+
input: number;
|
|
9939
|
+
output: number;
|
|
9940
|
+
cacheRead: number;
|
|
9941
|
+
cacheWrite: number;
|
|
9942
|
+
};
|
|
9943
|
+
contextWindow: number;
|
|
9944
|
+
maxTokens: number;
|
|
9945
|
+
};
|
|
9870
9946
|
readonly "openai/gpt-5.2-pro": {
|
|
9871
9947
|
id: string;
|
|
9872
9948
|
name: string;
|