@mariozechner/pi-ai 0.45.7 → 0.47.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/constants.d.ts +6 -0
- package/dist/constants.d.ts.map +1 -0
- package/dist/constants.js +14 -0
- package/dist/constants.js.map +1 -0
- package/dist/index.d.ts +1 -1
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -1
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +134 -302
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +165 -333
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +27 -5
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +22 -10
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts +0 -2
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +476 -489
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +2 -1
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/stream.d.ts.map +1 -1
- package/dist/stream.js +1 -0
- package/dist/stream.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
- package/dist/providers/openai-codex/constants.d.ts +0 -21
- package/dist/providers/openai-codex/constants.d.ts.map +0 -1
- package/dist/providers/openai-codex/constants.js +0 -21
- package/dist/providers/openai-codex/constants.js.map +0 -1
- package/dist/providers/openai-codex/index.d.ts +0 -7
- package/dist/providers/openai-codex/index.d.ts.map +0 -1
- package/dist/providers/openai-codex/index.js +0 -7
- package/dist/providers/openai-codex/index.js.map +0 -1
- package/dist/providers/openai-codex/prompts/codex.d.ts +0 -3
- package/dist/providers/openai-codex/prompts/codex.d.ts.map +0 -1
- package/dist/providers/openai-codex/prompts/codex.js +0 -323
- package/dist/providers/openai-codex/prompts/codex.js.map +0 -1
- package/dist/providers/openai-codex/prompts/pi-codex-bridge.d.ts +0 -7
- package/dist/providers/openai-codex/prompts/pi-codex-bridge.d.ts.map +0 -1
- package/dist/providers/openai-codex/prompts/pi-codex-bridge.js +0 -50
- package/dist/providers/openai-codex/prompts/pi-codex-bridge.js.map +0 -1
- package/dist/providers/openai-codex/prompts/system-prompt.d.ts +0 -10
- package/dist/providers/openai-codex/prompts/system-prompt.d.ts.map +0 -1
- package/dist/providers/openai-codex/prompts/system-prompt.js +0 -15
- package/dist/providers/openai-codex/prompts/system-prompt.js.map +0 -1
- package/dist/providers/openai-codex/request-transformer.d.ts +0 -44
- package/dist/providers/openai-codex/request-transformer.d.ts.map +0 -1
- package/dist/providers/openai-codex/request-transformer.js +0 -99
- package/dist/providers/openai-codex/request-transformer.js.map +0 -1
- package/dist/providers/openai-codex/response-handler.d.ts +0 -19
- package/dist/providers/openai-codex/response-handler.d.ts.map +0 -1
- package/dist/providers/openai-codex/response-handler.js +0 -107
- package/dist/providers/openai-codex/response-handler.js.map +0 -1
package/dist/models.generated.js
CHANGED
|
@@ -1683,41 +1683,41 @@ export const MODELS = {
|
|
|
1683
1683
|
contextWindow: 128000,
|
|
1684
1684
|
maxTokens: 64000,
|
|
1685
1685
|
},
|
|
1686
|
-
"
|
|
1687
|
-
id: "
|
|
1688
|
-
name: "
|
|
1689
|
-
api: "openai-
|
|
1686
|
+
"gpt-5.2-codex": {
|
|
1687
|
+
id: "gpt-5.2-codex",
|
|
1688
|
+
name: "GPT-5.2-Codex",
|
|
1689
|
+
api: "openai-responses",
|
|
1690
1690
|
provider: "github-copilot",
|
|
1691
1691
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
1692
1692
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
1693
|
-
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
|
|
1694
1693
|
reasoning: true,
|
|
1695
|
-
input: ["text"],
|
|
1694
|
+
input: ["text", "image"],
|
|
1696
1695
|
cost: {
|
|
1697
1696
|
input: 0,
|
|
1698
1697
|
output: 0,
|
|
1699
1698
|
cacheRead: 0,
|
|
1700
1699
|
cacheWrite: 0,
|
|
1701
1700
|
},
|
|
1702
|
-
contextWindow:
|
|
1703
|
-
maxTokens:
|
|
1701
|
+
contextWindow: 272000,
|
|
1702
|
+
maxTokens: 128000,
|
|
1704
1703
|
},
|
|
1705
|
-
"
|
|
1706
|
-
id: "
|
|
1707
|
-
name: "
|
|
1708
|
-
api: "openai-
|
|
1704
|
+
"grok-code-fast-1": {
|
|
1705
|
+
id: "grok-code-fast-1",
|
|
1706
|
+
name: "Grok Code Fast 1",
|
|
1707
|
+
api: "openai-completions",
|
|
1709
1708
|
provider: "github-copilot",
|
|
1710
1709
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
1711
1710
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
1711
|
+
compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
|
|
1712
1712
|
reasoning: true,
|
|
1713
|
-
input: ["text"
|
|
1713
|
+
input: ["text"],
|
|
1714
1714
|
cost: {
|
|
1715
1715
|
input: 0,
|
|
1716
1716
|
output: 0,
|
|
1717
1717
|
cacheRead: 0,
|
|
1718
1718
|
cacheWrite: 0,
|
|
1719
1719
|
},
|
|
1720
|
-
contextWindow:
|
|
1720
|
+
contextWindow: 128000,
|
|
1721
1721
|
maxTokens: 64000,
|
|
1722
1722
|
},
|
|
1723
1723
|
},
|
|
@@ -2770,6 +2770,42 @@ export const MODELS = {
|
|
|
2770
2770
|
maxTokens: 131072,
|
|
2771
2771
|
},
|
|
2772
2772
|
},
|
|
2773
|
+
"minimax-cn": {
|
|
2774
|
+
"MiniMax-M2": {
|
|
2775
|
+
id: "MiniMax-M2",
|
|
2776
|
+
name: "MiniMax-M2",
|
|
2777
|
+
api: "anthropic-messages",
|
|
2778
|
+
provider: "minimax-cn",
|
|
2779
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
2780
|
+
reasoning: true,
|
|
2781
|
+
input: ["text"],
|
|
2782
|
+
cost: {
|
|
2783
|
+
input: 0.3,
|
|
2784
|
+
output: 1.2,
|
|
2785
|
+
cacheRead: 0,
|
|
2786
|
+
cacheWrite: 0,
|
|
2787
|
+
},
|
|
2788
|
+
contextWindow: 196608,
|
|
2789
|
+
maxTokens: 128000,
|
|
2790
|
+
},
|
|
2791
|
+
"MiniMax-M2.1": {
|
|
2792
|
+
id: "MiniMax-M2.1",
|
|
2793
|
+
name: "MiniMax-M2.1",
|
|
2794
|
+
api: "anthropic-messages",
|
|
2795
|
+
provider: "minimax-cn",
|
|
2796
|
+
baseUrl: "https://api.minimaxi.com/anthropic",
|
|
2797
|
+
reasoning: true,
|
|
2798
|
+
input: ["text"],
|
|
2799
|
+
cost: {
|
|
2800
|
+
input: 0.3,
|
|
2801
|
+
output: 1.2,
|
|
2802
|
+
cacheRead: 0,
|
|
2803
|
+
cacheWrite: 0,
|
|
2804
|
+
},
|
|
2805
|
+
contextWindow: 204800,
|
|
2806
|
+
maxTokens: 131072,
|
|
2807
|
+
},
|
|
2808
|
+
},
|
|
2773
2809
|
"mistral": {
|
|
2774
2810
|
"codestral-latest": {
|
|
2775
2811
|
id: "codestral-latest",
|
|
@@ -3606,6 +3642,23 @@ export const MODELS = {
|
|
|
3606
3642
|
contextWindow: 128000,
|
|
3607
3643
|
maxTokens: 16384,
|
|
3608
3644
|
},
|
|
3645
|
+
"gpt-5.2-codex": {
|
|
3646
|
+
id: "gpt-5.2-codex",
|
|
3647
|
+
name: "GPT-5.2 Codex",
|
|
3648
|
+
api: "openai-responses",
|
|
3649
|
+
provider: "openai",
|
|
3650
|
+
baseUrl: "https://api.openai.com/v1",
|
|
3651
|
+
reasoning: true,
|
|
3652
|
+
input: ["text", "image"],
|
|
3653
|
+
cost: {
|
|
3654
|
+
input: 1.75,
|
|
3655
|
+
output: 14,
|
|
3656
|
+
cacheRead: 0.175,
|
|
3657
|
+
cacheWrite: 0,
|
|
3658
|
+
},
|
|
3659
|
+
contextWindow: 400000,
|
|
3660
|
+
maxTokens: 128000,
|
|
3661
|
+
},
|
|
3609
3662
|
"gpt-5.2-pro": {
|
|
3610
3663
|
id: "gpt-5.2-pro",
|
|
3611
3664
|
name: "GPT-5.2 Pro",
|
|
@@ -4205,6 +4258,23 @@ export const MODELS = {
|
|
|
4205
4258
|
contextWindow: 400000,
|
|
4206
4259
|
maxTokens: 128000,
|
|
4207
4260
|
},
|
|
4261
|
+
"gpt-5.2-codex": {
|
|
4262
|
+
id: "gpt-5.2-codex",
|
|
4263
|
+
name: "GPT-5.2 Codex",
|
|
4264
|
+
api: "openai-responses",
|
|
4265
|
+
provider: "opencode",
|
|
4266
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
4267
|
+
reasoning: true,
|
|
4268
|
+
input: ["text", "image"],
|
|
4269
|
+
cost: {
|
|
4270
|
+
input: 1.75,
|
|
4271
|
+
output: 14,
|
|
4272
|
+
cacheRead: 0.175,
|
|
4273
|
+
cacheWrite: 0,
|
|
4274
|
+
},
|
|
4275
|
+
contextWindow: 400000,
|
|
4276
|
+
maxTokens: 128000,
|
|
4277
|
+
},
|
|
4208
4278
|
"grok-code": {
|
|
4209
4279
|
id: "grok-code",
|
|
4210
4280
|
name: "Grok Code Fast 1",
|
|
@@ -4343,23 +4413,6 @@ export const MODELS = {
|
|
|
4343
4413
|
contextWindow: 131072,
|
|
4344
4414
|
maxTokens: 131072,
|
|
4345
4415
|
},
|
|
4346
|
-
"allenai/olmo-3-7b-instruct": {
|
|
4347
|
-
id: "allenai/olmo-3-7b-instruct",
|
|
4348
|
-
name: "AllenAI: Olmo 3 7B Instruct",
|
|
4349
|
-
api: "openai-completions",
|
|
4350
|
-
provider: "openrouter",
|
|
4351
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4352
|
-
reasoning: false,
|
|
4353
|
-
input: ["text"],
|
|
4354
|
-
cost: {
|
|
4355
|
-
input: 0.09999999999999999,
|
|
4356
|
-
output: 0.19999999999999998,
|
|
4357
|
-
cacheRead: 0,
|
|
4358
|
-
cacheWrite: 0,
|
|
4359
|
-
},
|
|
4360
|
-
contextWindow: 65536,
|
|
4361
|
-
maxTokens: 65536,
|
|
4362
|
-
},
|
|
4363
4416
|
"allenai/olmo-3.1-32b-instruct": {
|
|
4364
4417
|
id: "allenai/olmo-3.1-32b-instruct",
|
|
4365
4418
|
name: "AllenAI: Olmo 3.1 32B Instruct",
|
|
@@ -4496,23 +4549,6 @@ export const MODELS = {
|
|
|
4496
4549
|
contextWindow: 200000,
|
|
4497
4550
|
maxTokens: 8192,
|
|
4498
4551
|
},
|
|
4499
|
-
"anthropic/claude-3.5-haiku-20241022": {
|
|
4500
|
-
id: "anthropic/claude-3.5-haiku-20241022",
|
|
4501
|
-
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
4502
|
-
api: "openai-completions",
|
|
4503
|
-
provider: "openrouter",
|
|
4504
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4505
|
-
reasoning: false,
|
|
4506
|
-
input: ["text", "image"],
|
|
4507
|
-
cost: {
|
|
4508
|
-
input: 0.7999999999999999,
|
|
4509
|
-
output: 4,
|
|
4510
|
-
cacheRead: 0.08,
|
|
4511
|
-
cacheWrite: 1,
|
|
4512
|
-
},
|
|
4513
|
-
contextWindow: 200000,
|
|
4514
|
-
maxTokens: 8192,
|
|
4515
|
-
},
|
|
4516
4552
|
"anthropic/claude-3.5-sonnet": {
|
|
4517
4553
|
id: "anthropic/claude-3.5-sonnet",
|
|
4518
4554
|
name: "Anthropic: Claude 3.5 Sonnet",
|
|
@@ -4931,12 +4967,12 @@ export const MODELS = {
|
|
|
4931
4967
|
input: ["text"],
|
|
4932
4968
|
cost: {
|
|
4933
4969
|
input: 0.7,
|
|
4934
|
-
output: 2.
|
|
4970
|
+
output: 2.5,
|
|
4935
4971
|
cacheRead: 0,
|
|
4936
4972
|
cacheWrite: 0,
|
|
4937
4973
|
},
|
|
4938
|
-
contextWindow:
|
|
4939
|
-
maxTokens:
|
|
4974
|
+
contextWindow: 64000,
|
|
4975
|
+
maxTokens: 16000,
|
|
4940
4976
|
},
|
|
4941
4977
|
"deepseek/deepseek-r1-0528": {
|
|
4942
4978
|
id: "deepseek/deepseek-r1-0528",
|
|
@@ -5136,8 +5172,8 @@ export const MODELS = {
|
|
|
5136
5172
|
cost: {
|
|
5137
5173
|
input: 0.09999999999999999,
|
|
5138
5174
|
output: 0.39999999999999997,
|
|
5139
|
-
cacheRead: 0,
|
|
5140
|
-
cacheWrite:
|
|
5175
|
+
cacheRead: 0.01,
|
|
5176
|
+
cacheWrite: 1,
|
|
5141
5177
|
},
|
|
5142
5178
|
contextWindow: 1048576,
|
|
5143
5179
|
maxTokens: 65536,
|
|
@@ -5157,7 +5193,7 @@ export const MODELS = {
|
|
|
5157
5193
|
cacheWrite: 0.3833,
|
|
5158
5194
|
},
|
|
5159
5195
|
contextWindow: 1048576,
|
|
5160
|
-
maxTokens:
|
|
5196
|
+
maxTokens: 65535,
|
|
5161
5197
|
},
|
|
5162
5198
|
"google/gemini-2.5-pro": {
|
|
5163
5199
|
id: "google/gemini-2.5-pro",
|
|
@@ -5329,23 +5365,6 @@ export const MODELS = {
|
|
|
5329
5365
|
contextWindow: 256000,
|
|
5330
5366
|
maxTokens: 128000,
|
|
5331
5367
|
},
|
|
5332
|
-
"meta-llama/llama-3-70b-instruct": {
|
|
5333
|
-
id: "meta-llama/llama-3-70b-instruct",
|
|
5334
|
-
name: "Meta: Llama 3 70B Instruct",
|
|
5335
|
-
api: "openai-completions",
|
|
5336
|
-
provider: "openrouter",
|
|
5337
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5338
|
-
reasoning: false,
|
|
5339
|
-
input: ["text"],
|
|
5340
|
-
cost: {
|
|
5341
|
-
input: 0.3,
|
|
5342
|
-
output: 0.39999999999999997,
|
|
5343
|
-
cacheRead: 0,
|
|
5344
|
-
cacheWrite: 0,
|
|
5345
|
-
},
|
|
5346
|
-
contextWindow: 8192,
|
|
5347
|
-
maxTokens: 16384,
|
|
5348
|
-
},
|
|
5349
5368
|
"meta-llama/llama-3-8b-instruct": {
|
|
5350
5369
|
id: "meta-llama/llama-3-8b-instruct",
|
|
5351
5370
|
name: "Meta: Llama 3 8B Instruct",
|
|
@@ -5414,23 +5433,6 @@ export const MODELS = {
|
|
|
5414
5433
|
contextWindow: 16384,
|
|
5415
5434
|
maxTokens: 16384,
|
|
5416
5435
|
},
|
|
5417
|
-
"meta-llama/llama-3.2-3b-instruct": {
|
|
5418
|
-
id: "meta-llama/llama-3.2-3b-instruct",
|
|
5419
|
-
name: "Meta: Llama 3.2 3B Instruct",
|
|
5420
|
-
api: "openai-completions",
|
|
5421
|
-
provider: "openrouter",
|
|
5422
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5423
|
-
reasoning: false,
|
|
5424
|
-
input: ["text"],
|
|
5425
|
-
cost: {
|
|
5426
|
-
input: 0.02,
|
|
5427
|
-
output: 0.02,
|
|
5428
|
-
cacheRead: 0,
|
|
5429
|
-
cacheWrite: 0,
|
|
5430
|
-
},
|
|
5431
|
-
contextWindow: 131072,
|
|
5432
|
-
maxTokens: 16384,
|
|
5433
|
-
},
|
|
5434
5436
|
"meta-llama/llama-3.3-70b-instruct": {
|
|
5435
5437
|
id: "meta-llama/llama-3.3-70b-instruct",
|
|
5436
5438
|
name: "Meta: Llama 3.3 70B Instruct",
|
|
@@ -5627,12 +5629,12 @@ export const MODELS = {
|
|
|
5627
5629
|
reasoning: false,
|
|
5628
5630
|
input: ["text"],
|
|
5629
5631
|
cost: {
|
|
5630
|
-
input: 0.
|
|
5631
|
-
output: 0.
|
|
5632
|
+
input: 0.09999999999999999,
|
|
5633
|
+
output: 0.3,
|
|
5632
5634
|
cacheRead: 0,
|
|
5633
5635
|
cacheWrite: 0,
|
|
5634
5636
|
},
|
|
5635
|
-
contextWindow:
|
|
5637
|
+
contextWindow: 131072,
|
|
5636
5638
|
maxTokens: 4096,
|
|
5637
5639
|
},
|
|
5638
5640
|
"mistralai/ministral-14b-2512": {
|
|
@@ -5720,40 +5722,6 @@ export const MODELS = {
|
|
|
5720
5722
|
contextWindow: 262144,
|
|
5721
5723
|
maxTokens: 4096,
|
|
5722
5724
|
},
|
|
5723
|
-
"mistralai/mistral-7b-instruct": {
|
|
5724
|
-
id: "mistralai/mistral-7b-instruct",
|
|
5725
|
-
name: "Mistral: Mistral 7B Instruct",
|
|
5726
|
-
api: "openai-completions",
|
|
5727
|
-
provider: "openrouter",
|
|
5728
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5729
|
-
reasoning: false,
|
|
5730
|
-
input: ["text"],
|
|
5731
|
-
cost: {
|
|
5732
|
-
input: 0.028,
|
|
5733
|
-
output: 0.054,
|
|
5734
|
-
cacheRead: 0,
|
|
5735
|
-
cacheWrite: 0,
|
|
5736
|
-
},
|
|
5737
|
-
contextWindow: 32768,
|
|
5738
|
-
maxTokens: 16384,
|
|
5739
|
-
},
|
|
5740
|
-
"mistralai/mistral-7b-instruct:free": {
|
|
5741
|
-
id: "mistralai/mistral-7b-instruct:free",
|
|
5742
|
-
name: "Mistral: Mistral 7B Instruct (free)",
|
|
5743
|
-
api: "openai-completions",
|
|
5744
|
-
provider: "openrouter",
|
|
5745
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5746
|
-
reasoning: false,
|
|
5747
|
-
input: ["text"],
|
|
5748
|
-
cost: {
|
|
5749
|
-
input: 0,
|
|
5750
|
-
output: 0,
|
|
5751
|
-
cacheRead: 0,
|
|
5752
|
-
cacheWrite: 0,
|
|
5753
|
-
},
|
|
5754
|
-
contextWindow: 32768,
|
|
5755
|
-
maxTokens: 16384,
|
|
5756
|
-
},
|
|
5757
5725
|
"mistralai/mistral-large": {
|
|
5758
5726
|
id: "mistralai/mistral-large",
|
|
5759
5727
|
name: "Mistral Large",
|
|
@@ -5871,7 +5839,7 @@ export const MODELS = {
|
|
|
5871
5839
|
cacheWrite: 0,
|
|
5872
5840
|
},
|
|
5873
5841
|
contextWindow: 131072,
|
|
5874
|
-
maxTokens:
|
|
5842
|
+
maxTokens: 131072,
|
|
5875
5843
|
},
|
|
5876
5844
|
"mistralai/mistral-saba": {
|
|
5877
5845
|
id: "mistralai/mistral-saba",
|
|
@@ -6315,23 +6283,6 @@ export const MODELS = {
|
|
|
6315
6283
|
contextWindow: 128000,
|
|
6316
6284
|
maxTokens: 4096,
|
|
6317
6285
|
},
|
|
6318
|
-
"openai/codex-mini": {
|
|
6319
|
-
id: "openai/codex-mini",
|
|
6320
|
-
name: "OpenAI: Codex Mini",
|
|
6321
|
-
api: "openai-completions",
|
|
6322
|
-
provider: "openrouter",
|
|
6323
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
6324
|
-
reasoning: true,
|
|
6325
|
-
input: ["text", "image"],
|
|
6326
|
-
cost: {
|
|
6327
|
-
input: 1.5,
|
|
6328
|
-
output: 6,
|
|
6329
|
-
cacheRead: 0.375,
|
|
6330
|
-
cacheWrite: 0,
|
|
6331
|
-
},
|
|
6332
|
-
contextWindow: 200000,
|
|
6333
|
-
maxTokens: 100000,
|
|
6334
|
-
},
|
|
6335
6286
|
"openai/gpt-3.5-turbo": {
|
|
6336
6287
|
id: "openai/gpt-3.5-turbo",
|
|
6337
6288
|
name: "OpenAI: GPT-3.5 Turbo",
|
|
@@ -6893,6 +6844,23 @@ export const MODELS = {
|
|
|
6893
6844
|
contextWindow: 128000,
|
|
6894
6845
|
maxTokens: 16384,
|
|
6895
6846
|
},
|
|
6847
|
+
"openai/gpt-5.2-codex": {
|
|
6848
|
+
id: "openai/gpt-5.2-codex",
|
|
6849
|
+
name: "OpenAI: GPT-5.2-Codex",
|
|
6850
|
+
api: "openai-completions",
|
|
6851
|
+
provider: "openrouter",
|
|
6852
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
6853
|
+
reasoning: true,
|
|
6854
|
+
input: ["text", "image"],
|
|
6855
|
+
cost: {
|
|
6856
|
+
input: 1.75,
|
|
6857
|
+
output: 14,
|
|
6858
|
+
cacheRead: 0.175,
|
|
6859
|
+
cacheWrite: 0,
|
|
6860
|
+
},
|
|
6861
|
+
contextWindow: 400000,
|
|
6862
|
+
maxTokens: 128000,
|
|
6863
|
+
},
|
|
6896
6864
|
"openai/gpt-5.2-pro": {
|
|
6897
6865
|
id: "openai/gpt-5.2-pro",
|
|
6898
6866
|
name: "OpenAI: GPT-5.2 Pro",
|
|
@@ -7233,108 +7201,6 @@ export const MODELS = {
|
|
|
7233
7201
|
contextWindow: 32768,
|
|
7234
7202
|
maxTokens: 4096,
|
|
7235
7203
|
},
|
|
7236
|
-
"qwen/qwen-max": {
|
|
7237
|
-
id: "qwen/qwen-max",
|
|
7238
|
-
name: "Qwen: Qwen-Max ",
|
|
7239
|
-
api: "openai-completions",
|
|
7240
|
-
provider: "openrouter",
|
|
7241
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7242
|
-
reasoning: false,
|
|
7243
|
-
input: ["text"],
|
|
7244
|
-
cost: {
|
|
7245
|
-
input: 1.5999999999999999,
|
|
7246
|
-
output: 6.3999999999999995,
|
|
7247
|
-
cacheRead: 0.64,
|
|
7248
|
-
cacheWrite: 0,
|
|
7249
|
-
},
|
|
7250
|
-
contextWindow: 32768,
|
|
7251
|
-
maxTokens: 8192,
|
|
7252
|
-
},
|
|
7253
|
-
"qwen/qwen-plus": {
|
|
7254
|
-
id: "qwen/qwen-plus",
|
|
7255
|
-
name: "Qwen: Qwen-Plus",
|
|
7256
|
-
api: "openai-completions",
|
|
7257
|
-
provider: "openrouter",
|
|
7258
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7259
|
-
reasoning: false,
|
|
7260
|
-
input: ["text"],
|
|
7261
|
-
cost: {
|
|
7262
|
-
input: 0.39999999999999997,
|
|
7263
|
-
output: 1.2,
|
|
7264
|
-
cacheRead: 0.16,
|
|
7265
|
-
cacheWrite: 0,
|
|
7266
|
-
},
|
|
7267
|
-
contextWindow: 131072,
|
|
7268
|
-
maxTokens: 8192,
|
|
7269
|
-
},
|
|
7270
|
-
"qwen/qwen-plus-2025-07-28": {
|
|
7271
|
-
id: "qwen/qwen-plus-2025-07-28",
|
|
7272
|
-
name: "Qwen: Qwen Plus 0728",
|
|
7273
|
-
api: "openai-completions",
|
|
7274
|
-
provider: "openrouter",
|
|
7275
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7276
|
-
reasoning: false,
|
|
7277
|
-
input: ["text"],
|
|
7278
|
-
cost: {
|
|
7279
|
-
input: 0.39999999999999997,
|
|
7280
|
-
output: 1.2,
|
|
7281
|
-
cacheRead: 0,
|
|
7282
|
-
cacheWrite: 0,
|
|
7283
|
-
},
|
|
7284
|
-
contextWindow: 1000000,
|
|
7285
|
-
maxTokens: 32768,
|
|
7286
|
-
},
|
|
7287
|
-
"qwen/qwen-plus-2025-07-28:thinking": {
|
|
7288
|
-
id: "qwen/qwen-plus-2025-07-28:thinking",
|
|
7289
|
-
name: "Qwen: Qwen Plus 0728 (thinking)",
|
|
7290
|
-
api: "openai-completions",
|
|
7291
|
-
provider: "openrouter",
|
|
7292
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7293
|
-
reasoning: true,
|
|
7294
|
-
input: ["text"],
|
|
7295
|
-
cost: {
|
|
7296
|
-
input: 0.39999999999999997,
|
|
7297
|
-
output: 4,
|
|
7298
|
-
cacheRead: 0,
|
|
7299
|
-
cacheWrite: 0,
|
|
7300
|
-
},
|
|
7301
|
-
contextWindow: 1000000,
|
|
7302
|
-
maxTokens: 32768,
|
|
7303
|
-
},
|
|
7304
|
-
"qwen/qwen-turbo": {
|
|
7305
|
-
id: "qwen/qwen-turbo",
|
|
7306
|
-
name: "Qwen: Qwen-Turbo",
|
|
7307
|
-
api: "openai-completions",
|
|
7308
|
-
provider: "openrouter",
|
|
7309
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7310
|
-
reasoning: false,
|
|
7311
|
-
input: ["text"],
|
|
7312
|
-
cost: {
|
|
7313
|
-
input: 0.049999999999999996,
|
|
7314
|
-
output: 0.19999999999999998,
|
|
7315
|
-
cacheRead: 0.02,
|
|
7316
|
-
cacheWrite: 0,
|
|
7317
|
-
},
|
|
7318
|
-
contextWindow: 1000000,
|
|
7319
|
-
maxTokens: 8192,
|
|
7320
|
-
},
|
|
7321
|
-
"qwen/qwen-vl-max": {
|
|
7322
|
-
id: "qwen/qwen-vl-max",
|
|
7323
|
-
name: "Qwen: Qwen VL Max",
|
|
7324
|
-
api: "openai-completions",
|
|
7325
|
-
provider: "openrouter",
|
|
7326
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7327
|
-
reasoning: false,
|
|
7328
|
-
input: ["text", "image"],
|
|
7329
|
-
cost: {
|
|
7330
|
-
input: 0.7999999999999999,
|
|
7331
|
-
output: 3.1999999999999997,
|
|
7332
|
-
cacheRead: 0,
|
|
7333
|
-
cacheWrite: 0,
|
|
7334
|
-
},
|
|
7335
|
-
contextWindow: 131072,
|
|
7336
|
-
maxTokens: 8192,
|
|
7337
|
-
},
|
|
7338
7204
|
"qwen/qwen3-14b": {
|
|
7339
7205
|
id: "qwen/qwen3-14b",
|
|
7340
7206
|
name: "Qwen: Qwen3 14B",
|
|
@@ -7361,13 +7227,13 @@ export const MODELS = {
|
|
|
7361
7227
|
reasoning: true,
|
|
7362
7228
|
input: ["text"],
|
|
7363
7229
|
cost: {
|
|
7364
|
-
input: 0.
|
|
7365
|
-
output: 0.
|
|
7230
|
+
input: 0.19999999999999998,
|
|
7231
|
+
output: 0.6,
|
|
7366
7232
|
cacheRead: 0,
|
|
7367
7233
|
cacheWrite: 0,
|
|
7368
7234
|
},
|
|
7369
7235
|
contextWindow: 40960,
|
|
7370
|
-
maxTokens:
|
|
7236
|
+
maxTokens: 4096,
|
|
7371
7237
|
},
|
|
7372
7238
|
"qwen/qwen3-235b-a22b-2507": {
|
|
7373
7239
|
id: "qwen/qwen3-235b-a22b-2507",
|
|
@@ -7497,13 +7363,13 @@ export const MODELS = {
|
|
|
7497
7363
|
reasoning: true,
|
|
7498
7364
|
input: ["text"],
|
|
7499
7365
|
cost: {
|
|
7500
|
-
input: 0.
|
|
7501
|
-
output: 0.
|
|
7366
|
+
input: 0.049999999999999996,
|
|
7367
|
+
output: 0.25,
|
|
7502
7368
|
cacheRead: 0,
|
|
7503
7369
|
cacheWrite: 0,
|
|
7504
7370
|
},
|
|
7505
|
-
contextWindow:
|
|
7506
|
-
maxTokens:
|
|
7371
|
+
contextWindow: 32000,
|
|
7372
|
+
maxTokens: 8192,
|
|
7507
7373
|
},
|
|
7508
7374
|
"qwen/qwen3-coder": {
|
|
7509
7375
|
id: "qwen/qwen3-coder",
|
|
@@ -7539,40 +7405,6 @@ export const MODELS = {
|
|
|
7539
7405
|
contextWindow: 160000,
|
|
7540
7406
|
maxTokens: 32768,
|
|
7541
7407
|
},
|
|
7542
|
-
"qwen/qwen3-coder-flash": {
|
|
7543
|
-
id: "qwen/qwen3-coder-flash",
|
|
7544
|
-
name: "Qwen: Qwen3 Coder Flash",
|
|
7545
|
-
api: "openai-completions",
|
|
7546
|
-
provider: "openrouter",
|
|
7547
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7548
|
-
reasoning: false,
|
|
7549
|
-
input: ["text"],
|
|
7550
|
-
cost: {
|
|
7551
|
-
input: 0.3,
|
|
7552
|
-
output: 1.5,
|
|
7553
|
-
cacheRead: 0.08,
|
|
7554
|
-
cacheWrite: 0,
|
|
7555
|
-
},
|
|
7556
|
-
contextWindow: 128000,
|
|
7557
|
-
maxTokens: 65536,
|
|
7558
|
-
},
|
|
7559
|
-
"qwen/qwen3-coder-plus": {
|
|
7560
|
-
id: "qwen/qwen3-coder-plus",
|
|
7561
|
-
name: "Qwen: Qwen3 Coder Plus",
|
|
7562
|
-
api: "openai-completions",
|
|
7563
|
-
provider: "openrouter",
|
|
7564
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7565
|
-
reasoning: false,
|
|
7566
|
-
input: ["text"],
|
|
7567
|
-
cost: {
|
|
7568
|
-
input: 1,
|
|
7569
|
-
output: 5,
|
|
7570
|
-
cacheRead: 0.09999999999999999,
|
|
7571
|
-
cacheWrite: 0,
|
|
7572
|
-
},
|
|
7573
|
-
contextWindow: 128000,
|
|
7574
|
-
maxTokens: 65536,
|
|
7575
|
-
},
|
|
7576
7408
|
"qwen/qwen3-coder:exacto": {
|
|
7577
7409
|
id: "qwen/qwen3-coder:exacto",
|
|
7578
7410
|
name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
|
|
@@ -7607,34 +7439,34 @@ export const MODELS = {
|
|
|
7607
7439
|
contextWindow: 262000,
|
|
7608
7440
|
maxTokens: 262000,
|
|
7609
7441
|
},
|
|
7610
|
-
"qwen/qwen3-
|
|
7611
|
-
id: "qwen/qwen3-
|
|
7612
|
-
name: "Qwen: Qwen3
|
|
7442
|
+
"qwen/qwen3-next-80b-a3b-instruct": {
|
|
7443
|
+
id: "qwen/qwen3-next-80b-a3b-instruct",
|
|
7444
|
+
name: "Qwen: Qwen3 Next 80B A3B Instruct",
|
|
7613
7445
|
api: "openai-completions",
|
|
7614
7446
|
provider: "openrouter",
|
|
7615
7447
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7616
7448
|
reasoning: false,
|
|
7617
7449
|
input: ["text"],
|
|
7618
7450
|
cost: {
|
|
7619
|
-
input:
|
|
7620
|
-
output:
|
|
7621
|
-
cacheRead: 0
|
|
7451
|
+
input: 0.09,
|
|
7452
|
+
output: 1.1,
|
|
7453
|
+
cacheRead: 0,
|
|
7622
7454
|
cacheWrite: 0,
|
|
7623
7455
|
},
|
|
7624
|
-
contextWindow:
|
|
7625
|
-
maxTokens:
|
|
7456
|
+
contextWindow: 262144,
|
|
7457
|
+
maxTokens: 4096,
|
|
7626
7458
|
},
|
|
7627
|
-
"qwen/qwen3-next-80b-a3b-instruct": {
|
|
7628
|
-
id: "qwen/qwen3-next-80b-a3b-instruct",
|
|
7629
|
-
name: "Qwen: Qwen3 Next 80B A3B Instruct",
|
|
7459
|
+
"qwen/qwen3-next-80b-a3b-instruct:free": {
|
|
7460
|
+
id: "qwen/qwen3-next-80b-a3b-instruct:free",
|
|
7461
|
+
name: "Qwen: Qwen3 Next 80B A3B Instruct (free)",
|
|
7630
7462
|
api: "openai-completions",
|
|
7631
7463
|
provider: "openrouter",
|
|
7632
7464
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
7633
7465
|
reasoning: false,
|
|
7634
7466
|
input: ["text"],
|
|
7635
7467
|
cost: {
|
|
7636
|
-
input: 0
|
|
7637
|
-
output:
|
|
7468
|
+
input: 0,
|
|
7469
|
+
output: 0,
|
|
7638
7470
|
cacheRead: 0,
|
|
7639
7471
|
cacheWrite: 0,
|
|
7640
7472
|
},
|
|
@@ -7743,23 +7575,6 @@ export const MODELS = {
|
|
|
7743
7575
|
contextWindow: 131072,
|
|
7744
7576
|
maxTokens: 32768,
|
|
7745
7577
|
},
|
|
7746
|
-
"qwen/qwen3-vl-8b-thinking": {
|
|
7747
|
-
id: "qwen/qwen3-vl-8b-thinking",
|
|
7748
|
-
name: "Qwen: Qwen3 VL 8B Thinking",
|
|
7749
|
-
api: "openai-completions",
|
|
7750
|
-
provider: "openrouter",
|
|
7751
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7752
|
-
reasoning: true,
|
|
7753
|
-
input: ["text", "image"],
|
|
7754
|
-
cost: {
|
|
7755
|
-
input: 0.18,
|
|
7756
|
-
output: 2.0999999999999996,
|
|
7757
|
-
cacheRead: 0,
|
|
7758
|
-
cacheWrite: 0,
|
|
7759
|
-
},
|
|
7760
|
-
contextWindow: 256000,
|
|
7761
|
-
maxTokens: 32768,
|
|
7762
|
-
},
|
|
7763
7578
|
"qwen/qwq-32b": {
|
|
7764
7579
|
id: "qwen/qwq-32b",
|
|
7765
7580
|
name: "Qwen: QwQ 32B",
|
|
@@ -8066,6 +7881,23 @@ export const MODELS = {
|
|
|
8066
7881
|
contextWindow: 256000,
|
|
8067
7882
|
maxTokens: 10000,
|
|
8068
7883
|
},
|
|
7884
|
+
"xiaomi/mimo-v2-flash": {
|
|
7885
|
+
id: "xiaomi/mimo-v2-flash",
|
|
7886
|
+
name: "Xiaomi: MiMo-V2-Flash",
|
|
7887
|
+
api: "openai-completions",
|
|
7888
|
+
provider: "openrouter",
|
|
7889
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7890
|
+
reasoning: true,
|
|
7891
|
+
input: ["text"],
|
|
7892
|
+
cost: {
|
|
7893
|
+
input: 0.09999999999999999,
|
|
7894
|
+
output: 0.3,
|
|
7895
|
+
cacheRead: 0.02,
|
|
7896
|
+
cacheWrite: 0,
|
|
7897
|
+
},
|
|
7898
|
+
contextWindow: 262144,
|
|
7899
|
+
maxTokens: 32000,
|
|
7900
|
+
},
|
|
8069
7901
|
"xiaomi/mimo-v2-flash:free": {
|
|
8070
7902
|
id: "xiaomi/mimo-v2-flash:free",
|
|
8071
7903
|
name: "Xiaomi: MiMo-V2-Flash (free)",
|
|
@@ -9224,23 +9056,6 @@ export const MODELS = {
|
|
|
9224
9056
|
contextWindow: 128000,
|
|
9225
9057
|
maxTokens: 64000,
|
|
9226
9058
|
},
|
|
9227
|
-
"mistral/mistral-nemo": {
|
|
9228
|
-
id: "mistral/mistral-nemo",
|
|
9229
|
-
name: "Mistral Nemo",
|
|
9230
|
-
api: "anthropic-messages",
|
|
9231
|
-
provider: "vercel-ai-gateway",
|
|
9232
|
-
baseUrl: "https://ai-gateway.vercel.sh",
|
|
9233
|
-
reasoning: false,
|
|
9234
|
-
input: ["text"],
|
|
9235
|
-
cost: {
|
|
9236
|
-
input: 0.04,
|
|
9237
|
-
output: 0.16999999999999998,
|
|
9238
|
-
cacheRead: 0,
|
|
9239
|
-
cacheWrite: 0,
|
|
9240
|
-
},
|
|
9241
|
-
contextWindow: 60288,
|
|
9242
|
-
maxTokens: 16000,
|
|
9243
|
-
},
|
|
9244
9059
|
"mistral/mistral-small": {
|
|
9245
9060
|
id: "mistral/mistral-small",
|
|
9246
9061
|
name: "Mistral Small",
|
|
@@ -9734,6 +9549,23 @@ export const MODELS = {
|
|
|
9734
9549
|
contextWindow: 128000,
|
|
9735
9550
|
maxTokens: 16384,
|
|
9736
9551
|
},
|
|
9552
|
+
"openai/gpt-5.2-codex": {
|
|
9553
|
+
id: "openai/gpt-5.2-codex",
|
|
9554
|
+
name: "GPT-5.2-Codex",
|
|
9555
|
+
api: "anthropic-messages",
|
|
9556
|
+
provider: "vercel-ai-gateway",
|
|
9557
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
9558
|
+
reasoning: true,
|
|
9559
|
+
input: ["text", "image"],
|
|
9560
|
+
cost: {
|
|
9561
|
+
input: 1.75,
|
|
9562
|
+
output: 14,
|
|
9563
|
+
cacheRead: 0.175,
|
|
9564
|
+
cacheWrite: 0,
|
|
9565
|
+
},
|
|
9566
|
+
contextWindow: 400000,
|
|
9567
|
+
maxTokens: 128000,
|
|
9568
|
+
},
|
|
9737
9569
|
"openai/gpt-5.2-pro": {
|
|
9738
9570
|
id: "openai/gpt-5.2-pro",
|
|
9739
9571
|
name: "GPT 5.2 ",
|