@mariozechner/pi-ai 0.45.7 → 0.47.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (60) hide show
  1. package/dist/constants.d.ts +6 -0
  2. package/dist/constants.d.ts.map +1 -0
  3. package/dist/constants.js +14 -0
  4. package/dist/constants.js.map +1 -0
  5. package/dist/index.d.ts +1 -1
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -1
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.generated.d.ts +134 -302
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +165 -333
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +27 -5
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/google-shared.d.ts.map +1 -1
  17. package/dist/providers/google-shared.js +22 -10
  18. package/dist/providers/google-shared.js.map +1 -1
  19. package/dist/providers/openai-codex-responses.d.ts +0 -2
  20. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  21. package/dist/providers/openai-codex-responses.js +476 -489
  22. package/dist/providers/openai-codex-responses.js.map +1 -1
  23. package/dist/providers/openai-completions.d.ts.map +1 -1
  24. package/dist/providers/openai-completions.js +2 -1
  25. package/dist/providers/openai-completions.js.map +1 -1
  26. package/dist/stream.d.ts.map +1 -1
  27. package/dist/stream.js +1 -0
  28. package/dist/stream.js.map +1 -1
  29. package/dist/types.d.ts +1 -1
  30. package/dist/types.d.ts.map +1 -1
  31. package/dist/types.js.map +1 -1
  32. package/package.json +1 -1
  33. package/dist/providers/openai-codex/constants.d.ts +0 -21
  34. package/dist/providers/openai-codex/constants.d.ts.map +0 -1
  35. package/dist/providers/openai-codex/constants.js +0 -21
  36. package/dist/providers/openai-codex/constants.js.map +0 -1
  37. package/dist/providers/openai-codex/index.d.ts +0 -7
  38. package/dist/providers/openai-codex/index.d.ts.map +0 -1
  39. package/dist/providers/openai-codex/index.js +0 -7
  40. package/dist/providers/openai-codex/index.js.map +0 -1
  41. package/dist/providers/openai-codex/prompts/codex.d.ts +0 -3
  42. package/dist/providers/openai-codex/prompts/codex.d.ts.map +0 -1
  43. package/dist/providers/openai-codex/prompts/codex.js +0 -323
  44. package/dist/providers/openai-codex/prompts/codex.js.map +0 -1
  45. package/dist/providers/openai-codex/prompts/pi-codex-bridge.d.ts +0 -7
  46. package/dist/providers/openai-codex/prompts/pi-codex-bridge.d.ts.map +0 -1
  47. package/dist/providers/openai-codex/prompts/pi-codex-bridge.js +0 -50
  48. package/dist/providers/openai-codex/prompts/pi-codex-bridge.js.map +0 -1
  49. package/dist/providers/openai-codex/prompts/system-prompt.d.ts +0 -10
  50. package/dist/providers/openai-codex/prompts/system-prompt.d.ts.map +0 -1
  51. package/dist/providers/openai-codex/prompts/system-prompt.js +0 -15
  52. package/dist/providers/openai-codex/prompts/system-prompt.js.map +0 -1
  53. package/dist/providers/openai-codex/request-transformer.d.ts +0 -44
  54. package/dist/providers/openai-codex/request-transformer.d.ts.map +0 -1
  55. package/dist/providers/openai-codex/request-transformer.js +0 -99
  56. package/dist/providers/openai-codex/request-transformer.js.map +0 -1
  57. package/dist/providers/openai-codex/response-handler.d.ts +0 -19
  58. package/dist/providers/openai-codex/response-handler.d.ts.map +0 -1
  59. package/dist/providers/openai-codex/response-handler.js +0 -107
  60. package/dist/providers/openai-codex/response-handler.js.map +0 -1
@@ -1683,41 +1683,41 @@ export const MODELS = {
1683
1683
  contextWindow: 128000,
1684
1684
  maxTokens: 64000,
1685
1685
  },
1686
- "grok-code-fast-1": {
1687
- id: "grok-code-fast-1",
1688
- name: "Grok Code Fast 1",
1689
- api: "openai-completions",
1686
+ "gpt-5.2-codex": {
1687
+ id: "gpt-5.2-codex",
1688
+ name: "GPT-5.2-Codex",
1689
+ api: "openai-responses",
1690
1690
  provider: "github-copilot",
1691
1691
  baseUrl: "https://api.individual.githubcopilot.com",
1692
1692
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
1693
- compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
1694
1693
  reasoning: true,
1695
- input: ["text"],
1694
+ input: ["text", "image"],
1696
1695
  cost: {
1697
1696
  input: 0,
1698
1697
  output: 0,
1699
1698
  cacheRead: 0,
1700
1699
  cacheWrite: 0,
1701
1700
  },
1702
- contextWindow: 128000,
1703
- maxTokens: 64000,
1701
+ contextWindow: 272000,
1702
+ maxTokens: 128000,
1704
1703
  },
1705
- "oswe-vscode-prime": {
1706
- id: "oswe-vscode-prime",
1707
- name: "Raptor Mini (Preview)",
1708
- api: "openai-responses",
1704
+ "grok-code-fast-1": {
1705
+ id: "grok-code-fast-1",
1706
+ name: "Grok Code Fast 1",
1707
+ api: "openai-completions",
1709
1708
  provider: "github-copilot",
1710
1709
  baseUrl: "https://api.individual.githubcopilot.com",
1711
1710
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
1711
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false },
1712
1712
  reasoning: true,
1713
- input: ["text", "image"],
1713
+ input: ["text"],
1714
1714
  cost: {
1715
1715
  input: 0,
1716
1716
  output: 0,
1717
1717
  cacheRead: 0,
1718
1718
  cacheWrite: 0,
1719
1719
  },
1720
- contextWindow: 200000,
1720
+ contextWindow: 128000,
1721
1721
  maxTokens: 64000,
1722
1722
  },
1723
1723
  },
@@ -2770,6 +2770,42 @@ export const MODELS = {
2770
2770
  maxTokens: 131072,
2771
2771
  },
2772
2772
  },
2773
+ "minimax-cn": {
2774
+ "MiniMax-M2": {
2775
+ id: "MiniMax-M2",
2776
+ name: "MiniMax-M2",
2777
+ api: "anthropic-messages",
2778
+ provider: "minimax-cn",
2779
+ baseUrl: "https://api.minimaxi.com/anthropic",
2780
+ reasoning: true,
2781
+ input: ["text"],
2782
+ cost: {
2783
+ input: 0.3,
2784
+ output: 1.2,
2785
+ cacheRead: 0,
2786
+ cacheWrite: 0,
2787
+ },
2788
+ contextWindow: 196608,
2789
+ maxTokens: 128000,
2790
+ },
2791
+ "MiniMax-M2.1": {
2792
+ id: "MiniMax-M2.1",
2793
+ name: "MiniMax-M2.1",
2794
+ api: "anthropic-messages",
2795
+ provider: "minimax-cn",
2796
+ baseUrl: "https://api.minimaxi.com/anthropic",
2797
+ reasoning: true,
2798
+ input: ["text"],
2799
+ cost: {
2800
+ input: 0.3,
2801
+ output: 1.2,
2802
+ cacheRead: 0,
2803
+ cacheWrite: 0,
2804
+ },
2805
+ contextWindow: 204800,
2806
+ maxTokens: 131072,
2807
+ },
2808
+ },
2773
2809
  "mistral": {
2774
2810
  "codestral-latest": {
2775
2811
  id: "codestral-latest",
@@ -3606,6 +3642,23 @@ export const MODELS = {
3606
3642
  contextWindow: 128000,
3607
3643
  maxTokens: 16384,
3608
3644
  },
3645
+ "gpt-5.2-codex": {
3646
+ id: "gpt-5.2-codex",
3647
+ name: "GPT-5.2 Codex",
3648
+ api: "openai-responses",
3649
+ provider: "openai",
3650
+ baseUrl: "https://api.openai.com/v1",
3651
+ reasoning: true,
3652
+ input: ["text", "image"],
3653
+ cost: {
3654
+ input: 1.75,
3655
+ output: 14,
3656
+ cacheRead: 0.175,
3657
+ cacheWrite: 0,
3658
+ },
3659
+ contextWindow: 400000,
3660
+ maxTokens: 128000,
3661
+ },
3609
3662
  "gpt-5.2-pro": {
3610
3663
  id: "gpt-5.2-pro",
3611
3664
  name: "GPT-5.2 Pro",
@@ -4205,6 +4258,23 @@ export const MODELS = {
4205
4258
  contextWindow: 400000,
4206
4259
  maxTokens: 128000,
4207
4260
  },
4261
+ "gpt-5.2-codex": {
4262
+ id: "gpt-5.2-codex",
4263
+ name: "GPT-5.2 Codex",
4264
+ api: "openai-responses",
4265
+ provider: "opencode",
4266
+ baseUrl: "https://opencode.ai/zen/v1",
4267
+ reasoning: true,
4268
+ input: ["text", "image"],
4269
+ cost: {
4270
+ input: 1.75,
4271
+ output: 14,
4272
+ cacheRead: 0.175,
4273
+ cacheWrite: 0,
4274
+ },
4275
+ contextWindow: 400000,
4276
+ maxTokens: 128000,
4277
+ },
4208
4278
  "grok-code": {
4209
4279
  id: "grok-code",
4210
4280
  name: "Grok Code Fast 1",
@@ -4343,23 +4413,6 @@ export const MODELS = {
4343
4413
  contextWindow: 131072,
4344
4414
  maxTokens: 131072,
4345
4415
  },
4346
- "allenai/olmo-3-7b-instruct": {
4347
- id: "allenai/olmo-3-7b-instruct",
4348
- name: "AllenAI: Olmo 3 7B Instruct",
4349
- api: "openai-completions",
4350
- provider: "openrouter",
4351
- baseUrl: "https://openrouter.ai/api/v1",
4352
- reasoning: false,
4353
- input: ["text"],
4354
- cost: {
4355
- input: 0.09999999999999999,
4356
- output: 0.19999999999999998,
4357
- cacheRead: 0,
4358
- cacheWrite: 0,
4359
- },
4360
- contextWindow: 65536,
4361
- maxTokens: 65536,
4362
- },
4363
4416
  "allenai/olmo-3.1-32b-instruct": {
4364
4417
  id: "allenai/olmo-3.1-32b-instruct",
4365
4418
  name: "AllenAI: Olmo 3.1 32B Instruct",
@@ -4496,23 +4549,6 @@ export const MODELS = {
4496
4549
  contextWindow: 200000,
4497
4550
  maxTokens: 8192,
4498
4551
  },
4499
- "anthropic/claude-3.5-haiku-20241022": {
4500
- id: "anthropic/claude-3.5-haiku-20241022",
4501
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4502
- api: "openai-completions",
4503
- provider: "openrouter",
4504
- baseUrl: "https://openrouter.ai/api/v1",
4505
- reasoning: false,
4506
- input: ["text", "image"],
4507
- cost: {
4508
- input: 0.7999999999999999,
4509
- output: 4,
4510
- cacheRead: 0.08,
4511
- cacheWrite: 1,
4512
- },
4513
- contextWindow: 200000,
4514
- maxTokens: 8192,
4515
- },
4516
4552
  "anthropic/claude-3.5-sonnet": {
4517
4553
  id: "anthropic/claude-3.5-sonnet",
4518
4554
  name: "Anthropic: Claude 3.5 Sonnet",
@@ -4931,12 +4967,12 @@ export const MODELS = {
4931
4967
  input: ["text"],
4932
4968
  cost: {
4933
4969
  input: 0.7,
4934
- output: 2.4,
4970
+ output: 2.5,
4935
4971
  cacheRead: 0,
4936
4972
  cacheWrite: 0,
4937
4973
  },
4938
- contextWindow: 163840,
4939
- maxTokens: 163840,
4974
+ contextWindow: 64000,
4975
+ maxTokens: 16000,
4940
4976
  },
4941
4977
  "deepseek/deepseek-r1-0528": {
4942
4978
  id: "deepseek/deepseek-r1-0528",
@@ -5136,8 +5172,8 @@ export const MODELS = {
5136
5172
  cost: {
5137
5173
  input: 0.09999999999999999,
5138
5174
  output: 0.39999999999999997,
5139
- cacheRead: 0,
5140
- cacheWrite: 0,
5175
+ cacheRead: 0.01,
5176
+ cacheWrite: 1,
5141
5177
  },
5142
5178
  contextWindow: 1048576,
5143
5179
  maxTokens: 65536,
@@ -5157,7 +5193,7 @@ export const MODELS = {
5157
5193
  cacheWrite: 0.3833,
5158
5194
  },
5159
5195
  contextWindow: 1048576,
5160
- maxTokens: 65536,
5196
+ maxTokens: 65535,
5161
5197
  },
5162
5198
  "google/gemini-2.5-pro": {
5163
5199
  id: "google/gemini-2.5-pro",
@@ -5329,23 +5365,6 @@ export const MODELS = {
5329
5365
  contextWindow: 256000,
5330
5366
  maxTokens: 128000,
5331
5367
  },
5332
- "meta-llama/llama-3-70b-instruct": {
5333
- id: "meta-llama/llama-3-70b-instruct",
5334
- name: "Meta: Llama 3 70B Instruct",
5335
- api: "openai-completions",
5336
- provider: "openrouter",
5337
- baseUrl: "https://openrouter.ai/api/v1",
5338
- reasoning: false,
5339
- input: ["text"],
5340
- cost: {
5341
- input: 0.3,
5342
- output: 0.39999999999999997,
5343
- cacheRead: 0,
5344
- cacheWrite: 0,
5345
- },
5346
- contextWindow: 8192,
5347
- maxTokens: 16384,
5348
- },
5349
5368
  "meta-llama/llama-3-8b-instruct": {
5350
5369
  id: "meta-llama/llama-3-8b-instruct",
5351
5370
  name: "Meta: Llama 3 8B Instruct",
@@ -5414,23 +5433,6 @@ export const MODELS = {
5414
5433
  contextWindow: 16384,
5415
5434
  maxTokens: 16384,
5416
5435
  },
5417
- "meta-llama/llama-3.2-3b-instruct": {
5418
- id: "meta-llama/llama-3.2-3b-instruct",
5419
- name: "Meta: Llama 3.2 3B Instruct",
5420
- api: "openai-completions",
5421
- provider: "openrouter",
5422
- baseUrl: "https://openrouter.ai/api/v1",
5423
- reasoning: false,
5424
- input: ["text"],
5425
- cost: {
5426
- input: 0.02,
5427
- output: 0.02,
5428
- cacheRead: 0,
5429
- cacheWrite: 0,
5430
- },
5431
- contextWindow: 131072,
5432
- maxTokens: 16384,
5433
- },
5434
5436
  "meta-llama/llama-3.3-70b-instruct": {
5435
5437
  id: "meta-llama/llama-3.3-70b-instruct",
5436
5438
  name: "Meta: Llama 3.3 70B Instruct",
@@ -5627,12 +5629,12 @@ export const MODELS = {
5627
5629
  reasoning: false,
5628
5630
  input: ["text"],
5629
5631
  cost: {
5630
- input: 0.07,
5631
- output: 0.28,
5632
+ input: 0.09999999999999999,
5633
+ output: 0.3,
5632
5634
  cacheRead: 0,
5633
5635
  cacheWrite: 0,
5634
5636
  },
5635
- contextWindow: 128000,
5637
+ contextWindow: 131072,
5636
5638
  maxTokens: 4096,
5637
5639
  },
5638
5640
  "mistralai/ministral-14b-2512": {
@@ -5720,40 +5722,6 @@ export const MODELS = {
5720
5722
  contextWindow: 262144,
5721
5723
  maxTokens: 4096,
5722
5724
  },
5723
- "mistralai/mistral-7b-instruct": {
5724
- id: "mistralai/mistral-7b-instruct",
5725
- name: "Mistral: Mistral 7B Instruct",
5726
- api: "openai-completions",
5727
- provider: "openrouter",
5728
- baseUrl: "https://openrouter.ai/api/v1",
5729
- reasoning: false,
5730
- input: ["text"],
5731
- cost: {
5732
- input: 0.028,
5733
- output: 0.054,
5734
- cacheRead: 0,
5735
- cacheWrite: 0,
5736
- },
5737
- contextWindow: 32768,
5738
- maxTokens: 16384,
5739
- },
5740
- "mistralai/mistral-7b-instruct:free": {
5741
- id: "mistralai/mistral-7b-instruct:free",
5742
- name: "Mistral: Mistral 7B Instruct (free)",
5743
- api: "openai-completions",
5744
- provider: "openrouter",
5745
- baseUrl: "https://openrouter.ai/api/v1",
5746
- reasoning: false,
5747
- input: ["text"],
5748
- cost: {
5749
- input: 0,
5750
- output: 0,
5751
- cacheRead: 0,
5752
- cacheWrite: 0,
5753
- },
5754
- contextWindow: 32768,
5755
- maxTokens: 16384,
5756
- },
5757
5725
  "mistralai/mistral-large": {
5758
5726
  id: "mistralai/mistral-large",
5759
5727
  name: "Mistral Large",
@@ -5871,7 +5839,7 @@ export const MODELS = {
5871
5839
  cacheWrite: 0,
5872
5840
  },
5873
5841
  contextWindow: 131072,
5874
- maxTokens: 16384,
5842
+ maxTokens: 131072,
5875
5843
  },
5876
5844
  "mistralai/mistral-saba": {
5877
5845
  id: "mistralai/mistral-saba",
@@ -6315,23 +6283,6 @@ export const MODELS = {
6315
6283
  contextWindow: 128000,
6316
6284
  maxTokens: 4096,
6317
6285
  },
6318
- "openai/codex-mini": {
6319
- id: "openai/codex-mini",
6320
- name: "OpenAI: Codex Mini",
6321
- api: "openai-completions",
6322
- provider: "openrouter",
6323
- baseUrl: "https://openrouter.ai/api/v1",
6324
- reasoning: true,
6325
- input: ["text", "image"],
6326
- cost: {
6327
- input: 1.5,
6328
- output: 6,
6329
- cacheRead: 0.375,
6330
- cacheWrite: 0,
6331
- },
6332
- contextWindow: 200000,
6333
- maxTokens: 100000,
6334
- },
6335
6286
  "openai/gpt-3.5-turbo": {
6336
6287
  id: "openai/gpt-3.5-turbo",
6337
6288
  name: "OpenAI: GPT-3.5 Turbo",
@@ -6893,6 +6844,23 @@ export const MODELS = {
6893
6844
  contextWindow: 128000,
6894
6845
  maxTokens: 16384,
6895
6846
  },
6847
+ "openai/gpt-5.2-codex": {
6848
+ id: "openai/gpt-5.2-codex",
6849
+ name: "OpenAI: GPT-5.2-Codex",
6850
+ api: "openai-completions",
6851
+ provider: "openrouter",
6852
+ baseUrl: "https://openrouter.ai/api/v1",
6853
+ reasoning: true,
6854
+ input: ["text", "image"],
6855
+ cost: {
6856
+ input: 1.75,
6857
+ output: 14,
6858
+ cacheRead: 0.175,
6859
+ cacheWrite: 0,
6860
+ },
6861
+ contextWindow: 400000,
6862
+ maxTokens: 128000,
6863
+ },
6896
6864
  "openai/gpt-5.2-pro": {
6897
6865
  id: "openai/gpt-5.2-pro",
6898
6866
  name: "OpenAI: GPT-5.2 Pro",
@@ -7233,108 +7201,6 @@ export const MODELS = {
7233
7201
  contextWindow: 32768,
7234
7202
  maxTokens: 4096,
7235
7203
  },
7236
- "qwen/qwen-max": {
7237
- id: "qwen/qwen-max",
7238
- name: "Qwen: Qwen-Max ",
7239
- api: "openai-completions",
7240
- provider: "openrouter",
7241
- baseUrl: "https://openrouter.ai/api/v1",
7242
- reasoning: false,
7243
- input: ["text"],
7244
- cost: {
7245
- input: 1.5999999999999999,
7246
- output: 6.3999999999999995,
7247
- cacheRead: 0.64,
7248
- cacheWrite: 0,
7249
- },
7250
- contextWindow: 32768,
7251
- maxTokens: 8192,
7252
- },
7253
- "qwen/qwen-plus": {
7254
- id: "qwen/qwen-plus",
7255
- name: "Qwen: Qwen-Plus",
7256
- api: "openai-completions",
7257
- provider: "openrouter",
7258
- baseUrl: "https://openrouter.ai/api/v1",
7259
- reasoning: false,
7260
- input: ["text"],
7261
- cost: {
7262
- input: 0.39999999999999997,
7263
- output: 1.2,
7264
- cacheRead: 0.16,
7265
- cacheWrite: 0,
7266
- },
7267
- contextWindow: 131072,
7268
- maxTokens: 8192,
7269
- },
7270
- "qwen/qwen-plus-2025-07-28": {
7271
- id: "qwen/qwen-plus-2025-07-28",
7272
- name: "Qwen: Qwen Plus 0728",
7273
- api: "openai-completions",
7274
- provider: "openrouter",
7275
- baseUrl: "https://openrouter.ai/api/v1",
7276
- reasoning: false,
7277
- input: ["text"],
7278
- cost: {
7279
- input: 0.39999999999999997,
7280
- output: 1.2,
7281
- cacheRead: 0,
7282
- cacheWrite: 0,
7283
- },
7284
- contextWindow: 1000000,
7285
- maxTokens: 32768,
7286
- },
7287
- "qwen/qwen-plus-2025-07-28:thinking": {
7288
- id: "qwen/qwen-plus-2025-07-28:thinking",
7289
- name: "Qwen: Qwen Plus 0728 (thinking)",
7290
- api: "openai-completions",
7291
- provider: "openrouter",
7292
- baseUrl: "https://openrouter.ai/api/v1",
7293
- reasoning: true,
7294
- input: ["text"],
7295
- cost: {
7296
- input: 0.39999999999999997,
7297
- output: 4,
7298
- cacheRead: 0,
7299
- cacheWrite: 0,
7300
- },
7301
- contextWindow: 1000000,
7302
- maxTokens: 32768,
7303
- },
7304
- "qwen/qwen-turbo": {
7305
- id: "qwen/qwen-turbo",
7306
- name: "Qwen: Qwen-Turbo",
7307
- api: "openai-completions",
7308
- provider: "openrouter",
7309
- baseUrl: "https://openrouter.ai/api/v1",
7310
- reasoning: false,
7311
- input: ["text"],
7312
- cost: {
7313
- input: 0.049999999999999996,
7314
- output: 0.19999999999999998,
7315
- cacheRead: 0.02,
7316
- cacheWrite: 0,
7317
- },
7318
- contextWindow: 1000000,
7319
- maxTokens: 8192,
7320
- },
7321
- "qwen/qwen-vl-max": {
7322
- id: "qwen/qwen-vl-max",
7323
- name: "Qwen: Qwen VL Max",
7324
- api: "openai-completions",
7325
- provider: "openrouter",
7326
- baseUrl: "https://openrouter.ai/api/v1",
7327
- reasoning: false,
7328
- input: ["text", "image"],
7329
- cost: {
7330
- input: 0.7999999999999999,
7331
- output: 3.1999999999999997,
7332
- cacheRead: 0,
7333
- cacheWrite: 0,
7334
- },
7335
- contextWindow: 131072,
7336
- maxTokens: 8192,
7337
- },
7338
7204
  "qwen/qwen3-14b": {
7339
7205
  id: "qwen/qwen3-14b",
7340
7206
  name: "Qwen: Qwen3 14B",
@@ -7361,13 +7227,13 @@ export const MODELS = {
7361
7227
  reasoning: true,
7362
7228
  input: ["text"],
7363
7229
  cost: {
7364
- input: 0.18,
7365
- output: 0.54,
7230
+ input: 0.19999999999999998,
7231
+ output: 0.6,
7366
7232
  cacheRead: 0,
7367
7233
  cacheWrite: 0,
7368
7234
  },
7369
7235
  contextWindow: 40960,
7370
- maxTokens: 40960,
7236
+ maxTokens: 4096,
7371
7237
  },
7372
7238
  "qwen/qwen3-235b-a22b-2507": {
7373
7239
  id: "qwen/qwen3-235b-a22b-2507",
@@ -7497,13 +7363,13 @@ export const MODELS = {
7497
7363
  reasoning: true,
7498
7364
  input: ["text"],
7499
7365
  cost: {
7500
- input: 0.035,
7501
- output: 0.13799999999999998,
7366
+ input: 0.049999999999999996,
7367
+ output: 0.25,
7502
7368
  cacheRead: 0,
7503
7369
  cacheWrite: 0,
7504
7370
  },
7505
- contextWindow: 128000,
7506
- maxTokens: 20000,
7371
+ contextWindow: 32000,
7372
+ maxTokens: 8192,
7507
7373
  },
7508
7374
  "qwen/qwen3-coder": {
7509
7375
  id: "qwen/qwen3-coder",
@@ -7539,40 +7405,6 @@ export const MODELS = {
7539
7405
  contextWindow: 160000,
7540
7406
  maxTokens: 32768,
7541
7407
  },
7542
- "qwen/qwen3-coder-flash": {
7543
- id: "qwen/qwen3-coder-flash",
7544
- name: "Qwen: Qwen3 Coder Flash",
7545
- api: "openai-completions",
7546
- provider: "openrouter",
7547
- baseUrl: "https://openrouter.ai/api/v1",
7548
- reasoning: false,
7549
- input: ["text"],
7550
- cost: {
7551
- input: 0.3,
7552
- output: 1.5,
7553
- cacheRead: 0.08,
7554
- cacheWrite: 0,
7555
- },
7556
- contextWindow: 128000,
7557
- maxTokens: 65536,
7558
- },
7559
- "qwen/qwen3-coder-plus": {
7560
- id: "qwen/qwen3-coder-plus",
7561
- name: "Qwen: Qwen3 Coder Plus",
7562
- api: "openai-completions",
7563
- provider: "openrouter",
7564
- baseUrl: "https://openrouter.ai/api/v1",
7565
- reasoning: false,
7566
- input: ["text"],
7567
- cost: {
7568
- input: 1,
7569
- output: 5,
7570
- cacheRead: 0.09999999999999999,
7571
- cacheWrite: 0,
7572
- },
7573
- contextWindow: 128000,
7574
- maxTokens: 65536,
7575
- },
7576
7408
  "qwen/qwen3-coder:exacto": {
7577
7409
  id: "qwen/qwen3-coder:exacto",
7578
7410
  name: "Qwen: Qwen3 Coder 480B A35B (exacto)",
@@ -7607,34 +7439,34 @@ export const MODELS = {
7607
7439
  contextWindow: 262000,
7608
7440
  maxTokens: 262000,
7609
7441
  },
7610
- "qwen/qwen3-max": {
7611
- id: "qwen/qwen3-max",
7612
- name: "Qwen: Qwen3 Max",
7442
+ "qwen/qwen3-next-80b-a3b-instruct": {
7443
+ id: "qwen/qwen3-next-80b-a3b-instruct",
7444
+ name: "Qwen: Qwen3 Next 80B A3B Instruct",
7613
7445
  api: "openai-completions",
7614
7446
  provider: "openrouter",
7615
7447
  baseUrl: "https://openrouter.ai/api/v1",
7616
7448
  reasoning: false,
7617
7449
  input: ["text"],
7618
7450
  cost: {
7619
- input: 1.2,
7620
- output: 6,
7621
- cacheRead: 0.24,
7451
+ input: 0.09,
7452
+ output: 1.1,
7453
+ cacheRead: 0,
7622
7454
  cacheWrite: 0,
7623
7455
  },
7624
- contextWindow: 256000,
7625
- maxTokens: 32768,
7456
+ contextWindow: 262144,
7457
+ maxTokens: 4096,
7626
7458
  },
7627
- "qwen/qwen3-next-80b-a3b-instruct": {
7628
- id: "qwen/qwen3-next-80b-a3b-instruct",
7629
- name: "Qwen: Qwen3 Next 80B A3B Instruct",
7459
+ "qwen/qwen3-next-80b-a3b-instruct:free": {
7460
+ id: "qwen/qwen3-next-80b-a3b-instruct:free",
7461
+ name: "Qwen: Qwen3 Next 80B A3B Instruct (free)",
7630
7462
  api: "openai-completions",
7631
7463
  provider: "openrouter",
7632
7464
  baseUrl: "https://openrouter.ai/api/v1",
7633
7465
  reasoning: false,
7634
7466
  input: ["text"],
7635
7467
  cost: {
7636
- input: 0.09,
7637
- output: 1.1,
7468
+ input: 0,
7469
+ output: 0,
7638
7470
  cacheRead: 0,
7639
7471
  cacheWrite: 0,
7640
7472
  },
@@ -7743,23 +7575,6 @@ export const MODELS = {
7743
7575
  contextWindow: 131072,
7744
7576
  maxTokens: 32768,
7745
7577
  },
7746
- "qwen/qwen3-vl-8b-thinking": {
7747
- id: "qwen/qwen3-vl-8b-thinking",
7748
- name: "Qwen: Qwen3 VL 8B Thinking",
7749
- api: "openai-completions",
7750
- provider: "openrouter",
7751
- baseUrl: "https://openrouter.ai/api/v1",
7752
- reasoning: true,
7753
- input: ["text", "image"],
7754
- cost: {
7755
- input: 0.18,
7756
- output: 2.0999999999999996,
7757
- cacheRead: 0,
7758
- cacheWrite: 0,
7759
- },
7760
- contextWindow: 256000,
7761
- maxTokens: 32768,
7762
- },
7763
7578
  "qwen/qwq-32b": {
7764
7579
  id: "qwen/qwq-32b",
7765
7580
  name: "Qwen: QwQ 32B",
@@ -8066,6 +7881,23 @@ export const MODELS = {
8066
7881
  contextWindow: 256000,
8067
7882
  maxTokens: 10000,
8068
7883
  },
7884
+ "xiaomi/mimo-v2-flash": {
7885
+ id: "xiaomi/mimo-v2-flash",
7886
+ name: "Xiaomi: MiMo-V2-Flash",
7887
+ api: "openai-completions",
7888
+ provider: "openrouter",
7889
+ baseUrl: "https://openrouter.ai/api/v1",
7890
+ reasoning: true,
7891
+ input: ["text"],
7892
+ cost: {
7893
+ input: 0.09999999999999999,
7894
+ output: 0.3,
7895
+ cacheRead: 0.02,
7896
+ cacheWrite: 0,
7897
+ },
7898
+ contextWindow: 262144,
7899
+ maxTokens: 32000,
7900
+ },
8069
7901
  "xiaomi/mimo-v2-flash:free": {
8070
7902
  id: "xiaomi/mimo-v2-flash:free",
8071
7903
  name: "Xiaomi: MiMo-V2-Flash (free)",
@@ -9224,23 +9056,6 @@ export const MODELS = {
9224
9056
  contextWindow: 128000,
9225
9057
  maxTokens: 64000,
9226
9058
  },
9227
- "mistral/mistral-nemo": {
9228
- id: "mistral/mistral-nemo",
9229
- name: "Mistral Nemo",
9230
- api: "anthropic-messages",
9231
- provider: "vercel-ai-gateway",
9232
- baseUrl: "https://ai-gateway.vercel.sh",
9233
- reasoning: false,
9234
- input: ["text"],
9235
- cost: {
9236
- input: 0.04,
9237
- output: 0.16999999999999998,
9238
- cacheRead: 0,
9239
- cacheWrite: 0,
9240
- },
9241
- contextWindow: 60288,
9242
- maxTokens: 16000,
9243
- },
9244
9059
  "mistral/mistral-small": {
9245
9060
  id: "mistral/mistral-small",
9246
9061
  name: "Mistral Small",
@@ -9734,6 +9549,23 @@ export const MODELS = {
9734
9549
  contextWindow: 128000,
9735
9550
  maxTokens: 16384,
9736
9551
  },
9552
+ "openai/gpt-5.2-codex": {
9553
+ id: "openai/gpt-5.2-codex",
9554
+ name: "GPT-5.2-Codex",
9555
+ api: "anthropic-messages",
9556
+ provider: "vercel-ai-gateway",
9557
+ baseUrl: "https://ai-gateway.vercel.sh",
9558
+ reasoning: true,
9559
+ input: ["text", "image"],
9560
+ cost: {
9561
+ input: 1.75,
9562
+ output: 14,
9563
+ cacheRead: 0.175,
9564
+ cacheWrite: 0,
9565
+ },
9566
+ contextWindow: 400000,
9567
+ maxTokens: 128000,
9568
+ },
9737
9569
  "openai/gpt-5.2-pro": {
9738
9570
  id: "openai/gpt-5.2-pro",
9739
9571
  name: "GPT 5.2 ",