@mariozechner/pi-ai 0.45.7 → 0.46.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1683,6 +1683,24 @@ export const MODELS = {
1683
1683
  contextWindow: 128000,
1684
1684
  maxTokens: 64000,
1685
1685
  },
1686
+ "gpt-5.2-codex": {
1687
+ id: "gpt-5.2-codex",
1688
+ name: "GPT-5.2-Codex",
1689
+ api: "openai-responses",
1690
+ provider: "github-copilot",
1691
+ baseUrl: "https://api.individual.githubcopilot.com",
1692
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
1693
+ reasoning: true,
1694
+ input: ["text", "image"],
1695
+ cost: {
1696
+ input: 0,
1697
+ output: 0,
1698
+ cacheRead: 0,
1699
+ cacheWrite: 0,
1700
+ },
1701
+ contextWindow: 272000,
1702
+ maxTokens: 128000,
1703
+ },
1686
1704
  "grok-code-fast-1": {
1687
1705
  id: "grok-code-fast-1",
1688
1706
  name: "Grok Code Fast 1",
@@ -2770,6 +2788,42 @@ export const MODELS = {
2770
2788
  maxTokens: 131072,
2771
2789
  },
2772
2790
  },
2791
+ "minimax-cn": {
2792
+ "MiniMax-M2": {
2793
+ id: "MiniMax-M2",
2794
+ name: "MiniMax-M2",
2795
+ api: "anthropic-messages",
2796
+ provider: "minimax-cn",
2797
+ baseUrl: "https://api.minimaxi.com/anthropic",
2798
+ reasoning: true,
2799
+ input: ["text"],
2800
+ cost: {
2801
+ input: 0.3,
2802
+ output: 1.2,
2803
+ cacheRead: 0,
2804
+ cacheWrite: 0,
2805
+ },
2806
+ contextWindow: 196608,
2807
+ maxTokens: 128000,
2808
+ },
2809
+ "MiniMax-M2.1": {
2810
+ id: "MiniMax-M2.1",
2811
+ name: "MiniMax-M2.1",
2812
+ api: "anthropic-messages",
2813
+ provider: "minimax-cn",
2814
+ baseUrl: "https://api.minimaxi.com/anthropic",
2815
+ reasoning: true,
2816
+ input: ["text"],
2817
+ cost: {
2818
+ input: 0.3,
2819
+ output: 1.2,
2820
+ cacheRead: 0,
2821
+ cacheWrite: 0,
2822
+ },
2823
+ contextWindow: 204800,
2824
+ maxTokens: 131072,
2825
+ },
2826
+ },
2773
2827
  "mistral": {
2774
2828
  "codestral-latest": {
2775
2829
  id: "codestral-latest",
@@ -3606,6 +3660,23 @@ export const MODELS = {
3606
3660
  contextWindow: 128000,
3607
3661
  maxTokens: 16384,
3608
3662
  },
3663
+ "gpt-5.2-codex": {
3664
+ id: "gpt-5.2-codex",
3665
+ name: "GPT-5.2 Codex",
3666
+ api: "openai-responses",
3667
+ provider: "openai",
3668
+ baseUrl: "https://api.openai.com/v1",
3669
+ reasoning: true,
3670
+ input: ["text", "image"],
3671
+ cost: {
3672
+ input: 1.75,
3673
+ output: 14,
3674
+ cacheRead: 0.175,
3675
+ cacheWrite: 0,
3676
+ },
3677
+ contextWindow: 400000,
3678
+ maxTokens: 128000,
3679
+ },
3609
3680
  "gpt-5.2-pro": {
3610
3681
  id: "gpt-5.2-pro",
3611
3682
  name: "GPT-5.2 Pro",
@@ -4205,6 +4276,23 @@ export const MODELS = {
4205
4276
  contextWindow: 400000,
4206
4277
  maxTokens: 128000,
4207
4278
  },
4279
+ "gpt-5.2-codex": {
4280
+ id: "gpt-5.2-codex",
4281
+ name: "GPT-5.2 Codex",
4282
+ api: "openai-responses",
4283
+ provider: "opencode",
4284
+ baseUrl: "https://opencode.ai/zen/v1",
4285
+ reasoning: true,
4286
+ input: ["text", "image"],
4287
+ cost: {
4288
+ input: 1.75,
4289
+ output: 14,
4290
+ cacheRead: 0.175,
4291
+ cacheWrite: 0,
4292
+ },
4293
+ contextWindow: 400000,
4294
+ maxTokens: 128000,
4295
+ },
4208
4296
  "grok-code": {
4209
4297
  id: "grok-code",
4210
4298
  name: "Grok Code Fast 1",
@@ -4496,23 +4584,6 @@ export const MODELS = {
4496
4584
  contextWindow: 200000,
4497
4585
  maxTokens: 8192,
4498
4586
  },
4499
- "anthropic/claude-3.5-haiku-20241022": {
4500
- id: "anthropic/claude-3.5-haiku-20241022",
4501
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4502
- api: "openai-completions",
4503
- provider: "openrouter",
4504
- baseUrl: "https://openrouter.ai/api/v1",
4505
- reasoning: false,
4506
- input: ["text", "image"],
4507
- cost: {
4508
- input: 0.7999999999999999,
4509
- output: 4,
4510
- cacheRead: 0.08,
4511
- cacheWrite: 1,
4512
- },
4513
- contextWindow: 200000,
4514
- maxTokens: 8192,
4515
- },
4516
4587
  "anthropic/claude-3.5-sonnet": {
4517
4588
  id: "anthropic/claude-3.5-sonnet",
4518
4589
  name: "Anthropic: Claude 3.5 Sonnet",
@@ -5142,23 +5213,6 @@ export const MODELS = {
5142
5213
  contextWindow: 1048576,
5143
5214
  maxTokens: 65536,
5144
5215
  },
5145
- "google/gemini-2.5-flash-preview-09-2025": {
5146
- id: "google/gemini-2.5-flash-preview-09-2025",
5147
- name: "Google: Gemini 2.5 Flash Preview 09-2025",
5148
- api: "openai-completions",
5149
- provider: "openrouter",
5150
- baseUrl: "https://openrouter.ai/api/v1",
5151
- reasoning: true,
5152
- input: ["text", "image"],
5153
- cost: {
5154
- input: 0.3,
5155
- output: 2.5,
5156
- cacheRead: 0.075,
5157
- cacheWrite: 0.3833,
5158
- },
5159
- contextWindow: 1048576,
5160
- maxTokens: 65536,
5161
- },
5162
5216
  "google/gemini-2.5-pro": {
5163
5217
  id: "google/gemini-2.5-pro",
5164
5218
  name: "Google: Gemini 2.5 Pro",
@@ -5414,23 +5468,6 @@ export const MODELS = {
5414
5468
  contextWindow: 16384,
5415
5469
  maxTokens: 16384,
5416
5470
  },
5417
- "meta-llama/llama-3.2-3b-instruct": {
5418
- id: "meta-llama/llama-3.2-3b-instruct",
5419
- name: "Meta: Llama 3.2 3B Instruct",
5420
- api: "openai-completions",
5421
- provider: "openrouter",
5422
- baseUrl: "https://openrouter.ai/api/v1",
5423
- reasoning: false,
5424
- input: ["text"],
5425
- cost: {
5426
- input: 0.02,
5427
- output: 0.02,
5428
- cacheRead: 0,
5429
- cacheWrite: 0,
5430
- },
5431
- contextWindow: 131072,
5432
- maxTokens: 16384,
5433
- },
5434
5471
  "meta-llama/llama-3.3-70b-instruct": {
5435
5472
  id: "meta-llama/llama-3.3-70b-instruct",
5436
5473
  name: "Meta: Llama 3.3 70B Instruct",
@@ -6893,6 +6930,23 @@ export const MODELS = {
6893
6930
  contextWindow: 128000,
6894
6931
  maxTokens: 16384,
6895
6932
  },
6933
+ "openai/gpt-5.2-codex": {
6934
+ id: "openai/gpt-5.2-codex",
6935
+ name: "OpenAI: GPT-5.2-Codex",
6936
+ api: "openai-completions",
6937
+ provider: "openrouter",
6938
+ baseUrl: "https://openrouter.ai/api/v1",
6939
+ reasoning: true,
6940
+ input: ["text", "image"],
6941
+ cost: {
6942
+ input: 1.75,
6943
+ output: 14,
6944
+ cacheRead: 0.175,
6945
+ cacheWrite: 0,
6946
+ },
6947
+ contextWindow: 400000,
6948
+ maxTokens: 128000,
6949
+ },
6896
6950
  "openai/gpt-5.2-pro": {
6897
6951
  id: "openai/gpt-5.2-pro",
6898
6952
  name: "OpenAI: GPT-5.2 Pro",
@@ -7497,13 +7551,13 @@ export const MODELS = {
7497
7551
  reasoning: true,
7498
7552
  input: ["text"],
7499
7553
  cost: {
7500
- input: 0.035,
7501
- output: 0.13799999999999998,
7554
+ input: 0.049999999999999996,
7555
+ output: 0.25,
7502
7556
  cacheRead: 0,
7503
7557
  cacheWrite: 0,
7504
7558
  },
7505
- contextWindow: 128000,
7506
- maxTokens: 20000,
7559
+ contextWindow: 32000,
7560
+ maxTokens: 8192,
7507
7561
  },
7508
7562
  "qwen/qwen3-coder": {
7509
7563
  id: "qwen/qwen3-coder",
@@ -8066,6 +8120,23 @@ export const MODELS = {
8066
8120
  contextWindow: 256000,
8067
8121
  maxTokens: 10000,
8068
8122
  },
8123
+ "xiaomi/mimo-v2-flash": {
8124
+ id: "xiaomi/mimo-v2-flash",
8125
+ name: "Xiaomi: MiMo-V2-Flash",
8126
+ api: "openai-completions",
8127
+ provider: "openrouter",
8128
+ baseUrl: "https://openrouter.ai/api/v1",
8129
+ reasoning: true,
8130
+ input: ["text"],
8131
+ cost: {
8132
+ input: 0.09999999999999999,
8133
+ output: 0.3,
8134
+ cacheRead: 0.02,
8135
+ cacheWrite: 0,
8136
+ },
8137
+ contextWindow: 262144,
8138
+ maxTokens: 32000,
8139
+ },
8069
8140
  "xiaomi/mimo-v2-flash:free": {
8070
8141
  id: "xiaomi/mimo-v2-flash:free",
8071
8142
  name: "Xiaomi: MiMo-V2-Flash (free)",
@@ -9224,23 +9295,6 @@ export const MODELS = {
9224
9295
  contextWindow: 128000,
9225
9296
  maxTokens: 64000,
9226
9297
  },
9227
- "mistral/mistral-nemo": {
9228
- id: "mistral/mistral-nemo",
9229
- name: "Mistral Nemo",
9230
- api: "anthropic-messages",
9231
- provider: "vercel-ai-gateway",
9232
- baseUrl: "https://ai-gateway.vercel.sh",
9233
- reasoning: false,
9234
- input: ["text"],
9235
- cost: {
9236
- input: 0.04,
9237
- output: 0.16999999999999998,
9238
- cacheRead: 0,
9239
- cacheWrite: 0,
9240
- },
9241
- contextWindow: 60288,
9242
- maxTokens: 16000,
9243
- },
9244
9298
  "mistral/mistral-small": {
9245
9299
  id: "mistral/mistral-small",
9246
9300
  name: "Mistral Small",
@@ -9734,6 +9788,23 @@ export const MODELS = {
9734
9788
  contextWindow: 128000,
9735
9789
  maxTokens: 16384,
9736
9790
  },
9791
+ "openai/gpt-5.2-codex": {
9792
+ id: "openai/gpt-5.2-codex",
9793
+ name: "GPT-5.2-Codex",
9794
+ api: "anthropic-messages",
9795
+ provider: "vercel-ai-gateway",
9796
+ baseUrl: "https://ai-gateway.vercel.sh",
9797
+ reasoning: true,
9798
+ input: ["text", "image"],
9799
+ cost: {
9800
+ input: 1.75,
9801
+ output: 14,
9802
+ cacheRead: 0.175,
9803
+ cacheWrite: 0,
9804
+ },
9805
+ contextWindow: 400000,
9806
+ maxTokens: 128000,
9807
+ },
9737
9808
  "openai/gpt-5.2-pro": {
9738
9809
  id: "openai/gpt-5.2-pro",
9739
9810
  name: "GPT 5.2 ",