@mariozechner/pi-ai 0.45.6 → 0.46.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +144 -68
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +143 -72
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +33 -10
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +22 -10
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +1 -1
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/stream.d.ts.map +1 -1
- package/dist/stream.js +1 -0
- package/dist/stream.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -1683,6 +1683,24 @@ export const MODELS = {
|
|
|
1683
1683
|
contextWindow: 128000,
|
|
1684
1684
|
maxTokens: 64000,
|
|
1685
1685
|
},
|
|
1686
|
+
"gpt-5.2-codex": {
|
|
1687
|
+
id: "gpt-5.2-codex",
|
|
1688
|
+
name: "GPT-5.2-Codex",
|
|
1689
|
+
api: "openai-responses",
|
|
1690
|
+
provider: "github-copilot",
|
|
1691
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
1692
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
1693
|
+
reasoning: true,
|
|
1694
|
+
input: ["text", "image"],
|
|
1695
|
+
cost: {
|
|
1696
|
+
input: 0,
|
|
1697
|
+
output: 0,
|
|
1698
|
+
cacheRead: 0,
|
|
1699
|
+
cacheWrite: 0,
|
|
1700
|
+
},
|
|
1701
|
+
contextWindow: 272000,
|
|
1702
|
+
maxTokens: 128000,
|
|
1703
|
+
},
|
|
1686
1704
|
"grok-code-fast-1": {
|
|
1687
1705
|
id: "grok-code-fast-1",
|
|
1688
1706
|
name: "Grok Code Fast 1",
|
|
@@ -2770,6 +2788,42 @@ export const MODELS = {
|
|
|
2770
2788
|
maxTokens: 131072,
|
|
2771
2789
|
},
|
|
2772
2790
|
},
|
|
2791
|
+
"minimax-cn": {
|
|
2792
|
+
"MiniMax-M2": {
|
|
2793
|
+
id: "MiniMax-M2",
|
|
2794
|
+
name: "MiniMax-M2",
|
|
2795
|
+
api: "anthropic-messages",
|
|
2796
|
+
provider: "minimax-cn",
|
|
2797
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
2798
|
+
reasoning: true,
|
|
2799
|
+
input: ["text"],
|
|
2800
|
+
cost: {
|
|
2801
|
+
input: 0.3,
|
|
2802
|
+
output: 1.2,
|
|
2803
|
+
cacheRead: 0,
|
|
2804
|
+
cacheWrite: 0,
|
|
2805
|
+
},
|
|
2806
|
+
contextWindow: 196608,
|
|
2807
|
+
maxTokens: 128000,
|
|
2808
|
+
},
|
|
2809
|
+
"MiniMax-M2.1": {
|
|
2810
|
+
id: "MiniMax-M2.1",
|
|
2811
|
+
name: "MiniMax-M2.1",
|
|
2812
|
+
api: "anthropic-messages",
|
|
2813
|
+
provider: "minimax-cn",
|
|
2814
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
2815
|
+
reasoning: true,
|
|
2816
|
+
input: ["text"],
|
|
2817
|
+
cost: {
|
|
2818
|
+
input: 0.3,
|
|
2819
|
+
output: 1.2,
|
|
2820
|
+
cacheRead: 0,
|
|
2821
|
+
cacheWrite: 0,
|
|
2822
|
+
},
|
|
2823
|
+
contextWindow: 204800,
|
|
2824
|
+
maxTokens: 131072,
|
|
2825
|
+
},
|
|
2826
|
+
},
|
|
2773
2827
|
"mistral": {
|
|
2774
2828
|
"codestral-latest": {
|
|
2775
2829
|
id: "codestral-latest",
|
|
@@ -3606,6 +3660,23 @@ export const MODELS = {
|
|
|
3606
3660
|
contextWindow: 128000,
|
|
3607
3661
|
maxTokens: 16384,
|
|
3608
3662
|
},
|
|
3663
|
+
"gpt-5.2-codex": {
|
|
3664
|
+
id: "gpt-5.2-codex",
|
|
3665
|
+
name: "GPT-5.2 Codex",
|
|
3666
|
+
api: "openai-responses",
|
|
3667
|
+
provider: "openai",
|
|
3668
|
+
baseUrl: "https://api.openai.com/v1",
|
|
3669
|
+
reasoning: true,
|
|
3670
|
+
input: ["text", "image"],
|
|
3671
|
+
cost: {
|
|
3672
|
+
input: 1.75,
|
|
3673
|
+
output: 14,
|
|
3674
|
+
cacheRead: 0.175,
|
|
3675
|
+
cacheWrite: 0,
|
|
3676
|
+
},
|
|
3677
|
+
contextWindow: 400000,
|
|
3678
|
+
maxTokens: 128000,
|
|
3679
|
+
},
|
|
3609
3680
|
"gpt-5.2-pro": {
|
|
3610
3681
|
id: "gpt-5.2-pro",
|
|
3611
3682
|
name: "GPT-5.2 Pro",
|
|
@@ -4205,6 +4276,23 @@ export const MODELS = {
|
|
|
4205
4276
|
contextWindow: 400000,
|
|
4206
4277
|
maxTokens: 128000,
|
|
4207
4278
|
},
|
|
4279
|
+
"gpt-5.2-codex": {
|
|
4280
|
+
id: "gpt-5.2-codex",
|
|
4281
|
+
name: "GPT-5.2 Codex",
|
|
4282
|
+
api: "openai-responses",
|
|
4283
|
+
provider: "opencode",
|
|
4284
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
4285
|
+
reasoning: true,
|
|
4286
|
+
input: ["text", "image"],
|
|
4287
|
+
cost: {
|
|
4288
|
+
input: 1.75,
|
|
4289
|
+
output: 14,
|
|
4290
|
+
cacheRead: 0.175,
|
|
4291
|
+
cacheWrite: 0,
|
|
4292
|
+
},
|
|
4293
|
+
contextWindow: 400000,
|
|
4294
|
+
maxTokens: 128000,
|
|
4295
|
+
},
|
|
4208
4296
|
"grok-code": {
|
|
4209
4297
|
id: "grok-code",
|
|
4210
4298
|
name: "Grok Code Fast 1",
|
|
@@ -4496,23 +4584,6 @@ export const MODELS = {
|
|
|
4496
4584
|
contextWindow: 200000,
|
|
4497
4585
|
maxTokens: 8192,
|
|
4498
4586
|
},
|
|
4499
|
-
"anthropic/claude-3.5-haiku-20241022": {
|
|
4500
|
-
id: "anthropic/claude-3.5-haiku-20241022",
|
|
4501
|
-
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
4502
|
-
api: "openai-completions",
|
|
4503
|
-
provider: "openrouter",
|
|
4504
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4505
|
-
reasoning: false,
|
|
4506
|
-
input: ["text", "image"],
|
|
4507
|
-
cost: {
|
|
4508
|
-
input: 0.7999999999999999,
|
|
4509
|
-
output: 4,
|
|
4510
|
-
cacheRead: 0.08,
|
|
4511
|
-
cacheWrite: 1,
|
|
4512
|
-
},
|
|
4513
|
-
contextWindow: 200000,
|
|
4514
|
-
maxTokens: 8192,
|
|
4515
|
-
},
|
|
4516
4587
|
"anthropic/claude-3.5-sonnet": {
|
|
4517
4588
|
id: "anthropic/claude-3.5-sonnet",
|
|
4518
4589
|
name: "Anthropic: Claude 3.5 Sonnet",
|
|
@@ -5142,23 +5213,6 @@ export const MODELS = {
|
|
|
5142
5213
|
contextWindow: 1048576,
|
|
5143
5214
|
maxTokens: 65536,
|
|
5144
5215
|
},
|
|
5145
|
-
"google/gemini-2.5-flash-preview-09-2025": {
|
|
5146
|
-
id: "google/gemini-2.5-flash-preview-09-2025",
|
|
5147
|
-
name: "Google: Gemini 2.5 Flash Preview 09-2025",
|
|
5148
|
-
api: "openai-completions",
|
|
5149
|
-
provider: "openrouter",
|
|
5150
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5151
|
-
reasoning: true,
|
|
5152
|
-
input: ["text", "image"],
|
|
5153
|
-
cost: {
|
|
5154
|
-
input: 0.3,
|
|
5155
|
-
output: 2.5,
|
|
5156
|
-
cacheRead: 0.075,
|
|
5157
|
-
cacheWrite: 0.3833,
|
|
5158
|
-
},
|
|
5159
|
-
contextWindow: 1048576,
|
|
5160
|
-
maxTokens: 65536,
|
|
5161
|
-
},
|
|
5162
5216
|
"google/gemini-2.5-pro": {
|
|
5163
5217
|
id: "google/gemini-2.5-pro",
|
|
5164
5218
|
name: "Google: Gemini 2.5 Pro",
|
|
@@ -5414,23 +5468,6 @@ export const MODELS = {
|
|
|
5414
5468
|
contextWindow: 16384,
|
|
5415
5469
|
maxTokens: 16384,
|
|
5416
5470
|
},
|
|
5417
|
-
"meta-llama/llama-3.2-3b-instruct": {
|
|
5418
|
-
id: "meta-llama/llama-3.2-3b-instruct",
|
|
5419
|
-
name: "Meta: Llama 3.2 3B Instruct",
|
|
5420
|
-
api: "openai-completions",
|
|
5421
|
-
provider: "openrouter",
|
|
5422
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5423
|
-
reasoning: false,
|
|
5424
|
-
input: ["text"],
|
|
5425
|
-
cost: {
|
|
5426
|
-
input: 0.02,
|
|
5427
|
-
output: 0.02,
|
|
5428
|
-
cacheRead: 0,
|
|
5429
|
-
cacheWrite: 0,
|
|
5430
|
-
},
|
|
5431
|
-
contextWindow: 131072,
|
|
5432
|
-
maxTokens: 16384,
|
|
5433
|
-
},
|
|
5434
5471
|
"meta-llama/llama-3.3-70b-instruct": {
|
|
5435
5472
|
id: "meta-llama/llama-3.3-70b-instruct",
|
|
5436
5473
|
name: "Meta: Llama 3.3 70B Instruct",
|
|
@@ -6893,6 +6930,23 @@ export const MODELS = {
|
|
|
6893
6930
|
contextWindow: 128000,
|
|
6894
6931
|
maxTokens: 16384,
|
|
6895
6932
|
},
|
|
6933
|
+
"openai/gpt-5.2-codex": {
|
|
6934
|
+
id: "openai/gpt-5.2-codex",
|
|
6935
|
+
name: "OpenAI: GPT-5.2-Codex",
|
|
6936
|
+
api: "openai-completions",
|
|
6937
|
+
provider: "openrouter",
|
|
6938
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6939
|
+
reasoning: true,
|
|
6940
|
+
input: ["text", "image"],
|
|
6941
|
+
cost: {
|
|
6942
|
+
input: 1.75,
|
|
6943
|
+
output: 14,
|
|
6944
|
+
cacheRead: 0.175,
|
|
6945
|
+
cacheWrite: 0,
|
|
6946
|
+
},
|
|
6947
|
+
contextWindow: 400000,
|
|
6948
|
+
maxTokens: 128000,
|
|
6949
|
+
},
|
|
6896
6950
|
"openai/gpt-5.2-pro": {
|
|
6897
6951
|
id: "openai/gpt-5.2-pro",
|
|
6898
6952
|
name: "OpenAI: GPT-5.2 Pro",
|
|
@@ -7497,13 +7551,13 @@ export const MODELS = {
|
|
|
7497
7551
|
reasoning: true,
|
|
7498
7552
|
input: ["text"],
|
|
7499
7553
|
cost: {
|
|
7500
|
-
input: 0.
|
|
7501
|
-
output: 0.
|
|
7554
|
+
input: 0.049999999999999996,
|
|
7555
|
+
output: 0.25,
|
|
7502
7556
|
cacheRead: 0,
|
|
7503
7557
|
cacheWrite: 0,
|
|
7504
7558
|
},
|
|
7505
|
-
contextWindow:
|
|
7506
|
-
maxTokens:
|
|
7559
|
+
contextWindow: 32000,
|
|
7560
|
+
maxTokens: 8192,
|
|
7507
7561
|
},
|
|
7508
7562
|
"qwen/qwen3-coder": {
|
|
7509
7563
|
id: "qwen/qwen3-coder",
|
|
@@ -8066,6 +8120,23 @@ export const MODELS = {
|
|
|
8066
8120
|
contextWindow: 256000,
|
|
8067
8121
|
maxTokens: 10000,
|
|
8068
8122
|
},
|
|
8123
|
+
"xiaomi/mimo-v2-flash": {
|
|
8124
|
+
id: "xiaomi/mimo-v2-flash",
|
|
8125
|
+
name: "Xiaomi: MiMo-V2-Flash",
|
|
8126
|
+
api: "openai-completions",
|
|
8127
|
+
provider: "openrouter",
|
|
8128
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8129
|
+
reasoning: true,
|
|
8130
|
+
input: ["text"],
|
|
8131
|
+
cost: {
|
|
8132
|
+
input: 0.09999999999999999,
|
|
8133
|
+
output: 0.3,
|
|
8134
|
+
cacheRead: 0.02,
|
|
8135
|
+
cacheWrite: 0,
|
|
8136
|
+
},
|
|
8137
|
+
contextWindow: 262144,
|
|
8138
|
+
maxTokens: 32000,
|
|
8139
|
+
},
|
|
8069
8140
|
"xiaomi/mimo-v2-flash:free": {
|
|
8070
8141
|
id: "xiaomi/mimo-v2-flash:free",
|
|
8071
8142
|
name: "Xiaomi: MiMo-V2-Flash (free)",
|
|
@@ -9224,23 +9295,6 @@ export const MODELS = {
|
|
|
9224
9295
|
contextWindow: 128000,
|
|
9225
9296
|
maxTokens: 64000,
|
|
9226
9297
|
},
|
|
9227
|
-
"mistral/mistral-nemo": {
|
|
9228
|
-
id: "mistral/mistral-nemo",
|
|
9229
|
-
name: "Mistral Nemo",
|
|
9230
|
-
api: "anthropic-messages",
|
|
9231
|
-
provider: "vercel-ai-gateway",
|
|
9232
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
9233
|
-
reasoning: false,
|
|
9234
|
-
input: ["text"],
|
|
9235
|
-
cost: {
|
|
9236
|
-
input: 0.04,
|
|
9237
|
-
output: 0.16999999999999998,
|
|
9238
|
-
cacheRead: 0,
|
|
9239
|
-
cacheWrite: 0,
|
|
9240
|
-
},
|
|
9241
|
-
contextWindow: 60288,
|
|
9242
|
-
maxTokens: 16000,
|
|
9243
|
-
},
|
|
9244
9298
|
"mistral/mistral-small": {
|
|
9245
9299
|
id: "mistral/mistral-small",
|
|
9246
9300
|
name: "Mistral Small",
|
|
@@ -9734,6 +9788,23 @@ export const MODELS = {
|
|
|
9734
9788
|
contextWindow: 128000,
|
|
9735
9789
|
maxTokens: 16384,
|
|
9736
9790
|
},
|
|
9791
|
+
"openai/gpt-5.2-codex": {
|
|
9792
|
+
id: "openai/gpt-5.2-codex",
|
|
9793
|
+
name: "GPT-5.2-Codex",
|
|
9794
|
+
api: "anthropic-messages",
|
|
9795
|
+
provider: "vercel-ai-gateway",
|
|
9796
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
9797
|
+
reasoning: true,
|
|
9798
|
+
input: ["text", "image"],
|
|
9799
|
+
cost: {
|
|
9800
|
+
input: 1.75,
|
|
9801
|
+
output: 14,
|
|
9802
|
+
cacheRead: 0.175,
|
|
9803
|
+
cacheWrite: 0,
|
|
9804
|
+
},
|
|
9805
|
+
contextWindow: 400000,
|
|
9806
|
+
maxTokens: 128000,
|
|
9807
|
+
},
|
|
9737
9808
|
"openai/gpt-5.2-pro": {
|
|
9738
9809
|
id: "openai/gpt-5.2-pro",
|
|
9739
9810
|
name: "GPT 5.2 ",
|