@mariozechner/pi-ai 0.58.4 → 0.60.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (60) hide show
  1. package/README.md +5 -0
  2. package/dist/index.d.ts +9 -8
  3. package/dist/index.d.ts.map +1 -1
  4. package/dist/index.js +0 -8
  5. package/dist/index.js.map +1 -1
  6. package/dist/models.generated.d.ts +194 -131
  7. package/dist/models.generated.d.ts.map +1 -1
  8. package/dist/models.generated.js +259 -206
  9. package/dist/models.generated.js.map +1 -1
  10. package/dist/providers/anthropic.d.ts +7 -0
  11. package/dist/providers/anthropic.d.ts.map +1 -1
  12. package/dist/providers/anthropic.js +22 -11
  13. package/dist/providers/anthropic.js.map +1 -1
  14. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  15. package/dist/providers/google-gemini-cli.js +3 -0
  16. package/dist/providers/google-gemini-cli.js.map +1 -1
  17. package/dist/providers/google-shared.d.ts.map +1 -1
  18. package/dist/providers/google-shared.js +20 -8
  19. package/dist/providers/google-shared.js.map +1 -1
  20. package/dist/providers/google-vertex.d.ts.map +1 -1
  21. package/dist/providers/google-vertex.js +3 -0
  22. package/dist/providers/google-vertex.js.map +1 -1
  23. package/dist/providers/google.d.ts.map +1 -1
  24. package/dist/providers/google.js +3 -0
  25. package/dist/providers/google.js.map +1 -1
  26. package/dist/providers/mistral.d.ts.map +1 -1
  27. package/dist/providers/mistral.js +3 -0
  28. package/dist/providers/mistral.js.map +1 -1
  29. package/dist/providers/openai-completions.d.ts.map +1 -1
  30. package/dist/providers/openai-completions.js +25 -12
  31. package/dist/providers/openai-completions.js.map +1 -1
  32. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  33. package/dist/providers/openai-responses-shared.js +18 -12
  34. package/dist/providers/openai-responses-shared.js.map +1 -1
  35. package/dist/providers/register-builtins.d.ts +28 -1
  36. package/dist/providers/register-builtins.d.ts.map +1 -1
  37. package/dist/providers/register-builtins.js +170 -47
  38. package/dist/providers/register-builtins.js.map +1 -1
  39. package/dist/types.d.ts +1 -0
  40. package/dist/types.d.ts.map +1 -1
  41. package/dist/types.js.map +1 -1
  42. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  43. package/dist/utils/oauth/anthropic.js +20 -28
  44. package/dist/utils/oauth/anthropic.js.map +1 -1
  45. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  46. package/dist/utils/oauth/google-antigravity.js +22 -19
  47. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  48. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  49. package/dist/utils/oauth/google-gemini-cli.js +22 -19
  50. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  51. package/dist/utils/oauth/oauth-page.d.ts +3 -0
  52. package/dist/utils/oauth/oauth-page.d.ts.map +1 -0
  53. package/dist/utils/oauth/oauth-page.js +105 -0
  54. package/dist/utils/oauth/oauth-page.js.map +1 -0
  55. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  56. package/dist/utils/oauth/openai-codex.js +24 -31
  57. package/dist/utils/oauth/openai-codex.js.map +1 -1
  58. package/package.json +39 -5
  59. package/bedrock-provider.d.ts +0 -1
  60. package/bedrock-provider.js +0 -1
@@ -254,7 +254,7 @@ export const MODELS = {
254
254
  cacheRead: 0.5,
255
255
  cacheWrite: 6.25,
256
256
  },
257
- contextWindow: 1000000,
257
+ contextWindow: 200000,
258
258
  maxTokens: 128000,
259
259
  },
260
260
  "anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -305,7 +305,7 @@ export const MODELS = {
305
305
  cacheRead: 0.3,
306
306
  cacheWrite: 3.75,
307
307
  },
308
- contextWindow: 1000000,
308
+ contextWindow: 200000,
309
309
  maxTokens: 64000,
310
310
  },
311
311
  "deepseek.r1-v1:0": {
@@ -407,7 +407,7 @@ export const MODELS = {
407
407
  cacheRead: 0.5,
408
408
  cacheWrite: 6.25,
409
409
  },
410
- contextWindow: 1000000,
410
+ contextWindow: 200000,
411
411
  maxTokens: 128000,
412
412
  },
413
413
  "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -458,7 +458,7 @@ export const MODELS = {
458
458
  cacheRead: 0.3,
459
459
  cacheWrite: 3.75,
460
460
  },
461
- contextWindow: 1000000,
461
+ contextWindow: 200000,
462
462
  maxTokens: 64000,
463
463
  },
464
464
  "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
@@ -509,7 +509,7 @@ export const MODELS = {
509
509
  cacheRead: 0.5,
510
510
  cacheWrite: 6.25,
511
511
  },
512
- contextWindow: 1000000,
512
+ contextWindow: 200000,
513
513
  maxTokens: 128000,
514
514
  },
515
515
  "global.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -560,7 +560,7 @@ export const MODELS = {
560
560
  cacheRead: 0.3,
561
561
  cacheWrite: 3.75,
562
562
  },
563
- contextWindow: 1000000,
563
+ contextWindow: 200000,
564
564
  maxTokens: 64000,
565
565
  },
566
566
  "google.gemma-3-27b-it": {
@@ -1291,7 +1291,7 @@ export const MODELS = {
1291
1291
  cacheRead: 0.5,
1292
1292
  cacheWrite: 6.25,
1293
1293
  },
1294
- contextWindow: 1000000,
1294
+ contextWindow: 200000,
1295
1295
  maxTokens: 128000,
1296
1296
  },
1297
1297
  "us.anthropic.claude-sonnet-4-20250514-v1:0": {
@@ -1342,7 +1342,7 @@ export const MODELS = {
1342
1342
  cacheRead: 0.3,
1343
1343
  cacheWrite: 3.75,
1344
1344
  },
1345
- contextWindow: 1000000,
1345
+ contextWindow: 200000,
1346
1346
  maxTokens: 64000,
1347
1347
  },
1348
1348
  "writer.palmyra-x4-v1:0": {
@@ -2301,6 +2301,40 @@ export const MODELS = {
2301
2301
  contextWindow: 272000,
2302
2302
  maxTokens: 128000,
2303
2303
  },
2304
+ "gpt-5.4-mini": {
2305
+ id: "gpt-5.4-mini",
2306
+ name: "GPT-5.4 mini",
2307
+ api: "azure-openai-responses",
2308
+ provider: "azure-openai-responses",
2309
+ baseUrl: "",
2310
+ reasoning: true,
2311
+ input: ["text", "image"],
2312
+ cost: {
2313
+ input: 0.75,
2314
+ output: 4.5,
2315
+ cacheRead: 0.075,
2316
+ cacheWrite: 0,
2317
+ },
2318
+ contextWindow: 400000,
2319
+ maxTokens: 128000,
2320
+ },
2321
+ "gpt-5.4-nano": {
2322
+ id: "gpt-5.4-nano",
2323
+ name: "GPT-5.4 nano",
2324
+ api: "azure-openai-responses",
2325
+ provider: "azure-openai-responses",
2326
+ baseUrl: "",
2327
+ reasoning: true,
2328
+ input: ["text", "image"],
2329
+ cost: {
2330
+ input: 0.2,
2331
+ output: 1.25,
2332
+ cacheRead: 0.02,
2333
+ cacheWrite: 0,
2334
+ },
2335
+ contextWindow: 400000,
2336
+ maxTokens: 128000,
2337
+ },
2304
2338
  "gpt-5.4-pro": {
2305
2339
  id: "gpt-5.4-pro",
2306
2340
  name: "GPT-5.4 Pro",
@@ -2577,7 +2611,7 @@ export const MODELS = {
2577
2611
  cacheRead: 0,
2578
2612
  cacheWrite: 0,
2579
2613
  },
2580
- contextWindow: 128000,
2614
+ contextWindow: 1000000,
2581
2615
  maxTokens: 64000,
2582
2616
  },
2583
2617
  "claude-sonnet-4": {
@@ -2631,7 +2665,7 @@ export const MODELS = {
2631
2665
  cacheRead: 0,
2632
2666
  cacheWrite: 0,
2633
2667
  },
2634
- contextWindow: 128000,
2668
+ contextWindow: 1000000,
2635
2669
  maxTokens: 32000,
2636
2670
  },
2637
2671
  "gemini-2.5-pro": {
@@ -2928,6 +2962,42 @@ export const MODELS = {
2928
2962
  contextWindow: 400000,
2929
2963
  maxTokens: 128000,
2930
2964
  },
2965
+ "gpt-5.4-mini": {
2966
+ id: "gpt-5.4-mini",
2967
+ name: "GPT-5.4 mini",
2968
+ api: "openai-responses",
2969
+ provider: "github-copilot",
2970
+ baseUrl: "https://api.individual.githubcopilot.com",
2971
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
2972
+ reasoning: true,
2973
+ input: ["text", "image"],
2974
+ cost: {
2975
+ input: 0,
2976
+ output: 0,
2977
+ cacheRead: 0,
2978
+ cacheWrite: 0,
2979
+ },
2980
+ contextWindow: 400000,
2981
+ maxTokens: 128000,
2982
+ },
2983
+ "gpt-5.4-nano": {
2984
+ id: "gpt-5.4-nano",
2985
+ name: "GPT-5.4 nano",
2986
+ api: "openai-responses",
2987
+ provider: "github-copilot",
2988
+ baseUrl: "https://api.individual.githubcopilot.com",
2989
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
2990
+ reasoning: true,
2991
+ input: ["text", "image"],
2992
+ cost: {
2993
+ input: 0,
2994
+ output: 0,
2995
+ cacheRead: 0,
2996
+ cacheWrite: 0,
2997
+ },
2998
+ contextWindow: 400000,
2999
+ maxTokens: 128000,
3000
+ },
2931
3001
  "grok-code-fast-1": {
2932
3002
  id: "grok-code-fast-1",
2933
3003
  name: "Grok Code Fast 1",
@@ -5503,6 +5573,40 @@ export const MODELS = {
5503
5573
  contextWindow: 272000,
5504
5574
  maxTokens: 128000,
5505
5575
  },
5576
+ "gpt-5.4-mini": {
5577
+ id: "gpt-5.4-mini",
5578
+ name: "GPT-5.4 mini",
5579
+ api: "openai-responses",
5580
+ provider: "openai",
5581
+ baseUrl: "https://api.openai.com/v1",
5582
+ reasoning: true,
5583
+ input: ["text", "image"],
5584
+ cost: {
5585
+ input: 0.75,
5586
+ output: 4.5,
5587
+ cacheRead: 0.075,
5588
+ cacheWrite: 0,
5589
+ },
5590
+ contextWindow: 400000,
5591
+ maxTokens: 128000,
5592
+ },
5593
+ "gpt-5.4-nano": {
5594
+ id: "gpt-5.4-nano",
5595
+ name: "GPT-5.4 nano",
5596
+ api: "openai-responses",
5597
+ provider: "openai",
5598
+ baseUrl: "https://api.openai.com/v1",
5599
+ reasoning: true,
5600
+ input: ["text", "image"],
5601
+ cost: {
5602
+ input: 0.2,
5603
+ output: 1.25,
5604
+ cacheRead: 0.02,
5605
+ cacheWrite: 0,
5606
+ },
5607
+ contextWindow: 400000,
5608
+ maxTokens: 128000,
5609
+ },
5506
5610
  "gpt-5.4-pro": {
5507
5611
  id: "gpt-5.4-pro",
5508
5612
  name: "GPT-5.4 Pro",
@@ -5966,23 +6070,6 @@ export const MODELS = {
5966
6070
  contextWindow: 1048576,
5967
6071
  maxTokens: 65536,
5968
6072
  },
5969
- "gemini-3-pro": {
5970
- id: "gemini-3-pro",
5971
- name: "Gemini 3 Pro",
5972
- api: "google-generative-ai",
5973
- provider: "opencode",
5974
- baseUrl: "https://opencode.ai/zen/v1",
5975
- reasoning: true,
5976
- input: ["text", "image"],
5977
- cost: {
5978
- input: 2,
5979
- output: 12,
5980
- cacheRead: 0.2,
5981
- cacheWrite: 0,
5982
- },
5983
- contextWindow: 1048576,
5984
- maxTokens: 65536,
5985
- },
5986
6073
  "gemini-3.1-pro": {
5987
6074
  id: "gemini-3.1-pro",
5988
6075
  name: "Gemini 3.1 Pro Preview",
@@ -6000,40 +6087,6 @@ export const MODELS = {
6000
6087
  contextWindow: 1048576,
6001
6088
  maxTokens: 65536,
6002
6089
  },
6003
- "glm-4.6": {
6004
- id: "glm-4.6",
6005
- name: "GLM-4.6",
6006
- api: "openai-completions",
6007
- provider: "opencode",
6008
- baseUrl: "https://opencode.ai/zen/v1",
6009
- reasoning: true,
6010
- input: ["text"],
6011
- cost: {
6012
- input: 0.6,
6013
- output: 2.2,
6014
- cacheRead: 0.1,
6015
- cacheWrite: 0,
6016
- },
6017
- contextWindow: 204800,
6018
- maxTokens: 131072,
6019
- },
6020
- "glm-4.7": {
6021
- id: "glm-4.7",
6022
- name: "GLM-4.7",
6023
- api: "openai-completions",
6024
- provider: "opencode",
6025
- baseUrl: "https://opencode.ai/zen/v1",
6026
- reasoning: true,
6027
- input: ["text"],
6028
- cost: {
6029
- input: 0.6,
6030
- output: 2.2,
6031
- cacheRead: 0.1,
6032
- cacheWrite: 0,
6033
- },
6034
- contextWindow: 204800,
6035
- maxTokens: 131072,
6036
- },
6037
6090
  "glm-5": {
6038
6091
  id: "glm-5",
6039
6092
  name: "GLM-5",
@@ -6289,23 +6342,6 @@ export const MODELS = {
6289
6342
  contextWindow: 262144,
6290
6343
  maxTokens: 65536,
6291
6344
  },
6292
- "minimax-m2.1": {
6293
- id: "minimax-m2.1",
6294
- name: "MiniMax M2.1",
6295
- api: "openai-completions",
6296
- provider: "opencode",
6297
- baseUrl: "https://opencode.ai/zen/v1",
6298
- reasoning: true,
6299
- input: ["text"],
6300
- cost: {
6301
- input: 0.3,
6302
- output: 1.2,
6303
- cacheRead: 0.1,
6304
- cacheWrite: 0,
6305
- },
6306
- contextWindow: 204800,
6307
- maxTokens: 131072,
6308
- },
6309
6345
  "minimax-m2.5": {
6310
6346
  id: "minimax-m2.5",
6311
6347
  name: "MiniMax M2.5",
@@ -7101,13 +7137,13 @@ export const MODELS = {
7101
7137
  reasoning: true,
7102
7138
  input: ["text"],
7103
7139
  cost: {
7104
- input: 0.26,
7105
- output: 0.38,
7106
- cacheRead: 0.13,
7140
+ input: 0.255,
7141
+ output: 0.39999999999999997,
7142
+ cacheRead: 0,
7107
7143
  cacheWrite: 0,
7108
7144
  },
7109
7145
  contextWindow: 163840,
7110
- maxTokens: 4096,
7146
+ maxTokens: 163840,
7111
7147
  },
7112
7148
  "deepseek/deepseek-v3.2-exp": {
7113
7149
  id: "deepseek/deepseek-v3.2-exp",
@@ -7662,7 +7698,7 @@ export const MODELS = {
7662
7698
  reasoning: true,
7663
7699
  input: ["text"],
7664
7700
  cost: {
7665
- input: 0.25,
7701
+ input: 0.19999999999999998,
7666
7702
  output: 1.2,
7667
7703
  cacheRead: 0,
7668
7704
  cacheWrite: 0,
@@ -7698,7 +7734,7 @@ export const MODELS = {
7698
7734
  cost: {
7699
7735
  input: 0.3,
7700
7736
  output: 0.8999999999999999,
7701
- cacheRead: 0,
7737
+ cacheRead: 0.03,
7702
7738
  cacheWrite: 0,
7703
7739
  },
7704
7740
  contextWindow: 256000,
@@ -7715,7 +7751,7 @@ export const MODELS = {
7715
7751
  cost: {
7716
7752
  input: 0.39999999999999997,
7717
7753
  output: 2,
7718
- cacheRead: 0,
7754
+ cacheRead: 0.04,
7719
7755
  cacheWrite: 0,
7720
7756
  },
7721
7757
  contextWindow: 262144,
@@ -7732,7 +7768,7 @@ export const MODELS = {
7732
7768
  cost: {
7733
7769
  input: 0.39999999999999997,
7734
7770
  output: 2,
7735
- cacheRead: 0,
7771
+ cacheRead: 0.04,
7736
7772
  cacheWrite: 0,
7737
7773
  },
7738
7774
  contextWindow: 131072,
@@ -7749,7 +7785,7 @@ export const MODELS = {
7749
7785
  cost: {
7750
7786
  input: 0.09999999999999999,
7751
7787
  output: 0.3,
7752
- cacheRead: 0,
7788
+ cacheRead: 0.01,
7753
7789
  cacheWrite: 0,
7754
7790
  },
7755
7791
  contextWindow: 131072,
@@ -7766,7 +7802,7 @@ export const MODELS = {
7766
7802
  cost: {
7767
7803
  input: 0.19999999999999998,
7768
7804
  output: 0.19999999999999998,
7769
- cacheRead: 0,
7805
+ cacheRead: 0.02,
7770
7806
  cacheWrite: 0,
7771
7807
  },
7772
7808
  contextWindow: 262144,
@@ -7783,7 +7819,7 @@ export const MODELS = {
7783
7819
  cost: {
7784
7820
  input: 0.09999999999999999,
7785
7821
  output: 0.09999999999999999,
7786
- cacheRead: 0,
7822
+ cacheRead: 0.01,
7787
7823
  cacheWrite: 0,
7788
7824
  },
7789
7825
  contextWindow: 131072,
@@ -7800,7 +7836,7 @@ export const MODELS = {
7800
7836
  cost: {
7801
7837
  input: 0.15,
7802
7838
  output: 0.15,
7803
- cacheRead: 0,
7839
+ cacheRead: 0.015,
7804
7840
  cacheWrite: 0,
7805
7841
  },
7806
7842
  contextWindow: 262144,
@@ -7817,7 +7853,7 @@ export const MODELS = {
7817
7853
  cost: {
7818
7854
  input: 2,
7819
7855
  output: 6,
7820
- cacheRead: 0,
7856
+ cacheRead: 0.19999999999999998,
7821
7857
  cacheWrite: 0,
7822
7858
  },
7823
7859
  contextWindow: 128000,
@@ -7834,7 +7870,7 @@ export const MODELS = {
7834
7870
  cost: {
7835
7871
  input: 2,
7836
7872
  output: 6,
7837
- cacheRead: 0,
7873
+ cacheRead: 0.19999999999999998,
7838
7874
  cacheWrite: 0,
7839
7875
  },
7840
7876
  contextWindow: 131072,
@@ -7851,7 +7887,7 @@ export const MODELS = {
7851
7887
  cost: {
7852
7888
  input: 2,
7853
7889
  output: 6,
7854
- cacheRead: 0,
7890
+ cacheRead: 0.19999999999999998,
7855
7891
  cacheWrite: 0,
7856
7892
  },
7857
7893
  contextWindow: 131072,
@@ -7868,7 +7904,7 @@ export const MODELS = {
7868
7904
  cost: {
7869
7905
  input: 0.5,
7870
7906
  output: 1.5,
7871
- cacheRead: 0,
7907
+ cacheRead: 0.049999999999999996,
7872
7908
  cacheWrite: 0,
7873
7909
  },
7874
7910
  contextWindow: 262144,
@@ -7885,7 +7921,7 @@ export const MODELS = {
7885
7921
  cost: {
7886
7922
  input: 0.39999999999999997,
7887
7923
  output: 2,
7888
- cacheRead: 0,
7924
+ cacheRead: 0.04,
7889
7925
  cacheWrite: 0,
7890
7926
  },
7891
7927
  contextWindow: 131072,
@@ -7902,7 +7938,7 @@ export const MODELS = {
7902
7938
  cost: {
7903
7939
  input: 0.39999999999999997,
7904
7940
  output: 2,
7905
- cacheRead: 0,
7941
+ cacheRead: 0.04,
7906
7942
  cacheWrite: 0,
7907
7943
  },
7908
7944
  contextWindow: 131072,
@@ -7936,7 +7972,7 @@ export const MODELS = {
7936
7972
  cost: {
7937
7973
  input: 0.19999999999999998,
7938
7974
  output: 0.6,
7939
- cacheRead: 0,
7975
+ cacheRead: 0.02,
7940
7976
  cacheWrite: 0,
7941
7977
  },
7942
7978
  contextWindow: 32768,
@@ -7959,6 +7995,23 @@ export const MODELS = {
7959
7995
  contextWindow: 32768,
7960
7996
  maxTokens: 16384,
7961
7997
  },
7998
+ "mistralai/mistral-small-2603": {
7999
+ id: "mistralai/mistral-small-2603",
8000
+ name: "Mistral: Mistral Small 4",
8001
+ api: "openai-completions",
8002
+ provider: "openrouter",
8003
+ baseUrl: "https://openrouter.ai/api/v1",
8004
+ reasoning: true,
8005
+ input: ["text", "image"],
8006
+ cost: {
8007
+ input: 0.15,
8008
+ output: 0.6,
8009
+ cacheRead: 0.015,
8010
+ cacheWrite: 0,
8011
+ },
8012
+ contextWindow: 262144,
8013
+ maxTokens: 4096,
8014
+ },
7962
8015
  "mistralai/mistral-small-3.1-24b-instruct:free": {
7963
8016
  id: "mistralai/mistral-small-3.1-24b-instruct:free",
7964
8017
  name: "Mistral: Mistral Small 3.1 24B (free)",
@@ -8004,7 +8057,7 @@ export const MODELS = {
8004
8057
  cost: {
8005
8058
  input: 0.09999999999999999,
8006
8059
  output: 0.3,
8007
- cacheRead: 0,
8060
+ cacheRead: 0.01,
8008
8061
  cacheWrite: 0,
8009
8062
  },
8010
8063
  contextWindow: 32768,
@@ -8021,7 +8074,7 @@ export const MODELS = {
8021
8074
  cost: {
8022
8075
  input: 2,
8023
8076
  output: 6,
8024
- cacheRead: 0,
8077
+ cacheRead: 0.19999999999999998,
8025
8078
  cacheWrite: 0,
8026
8079
  },
8027
8080
  contextWindow: 65536,
@@ -8055,7 +8108,7 @@ export const MODELS = {
8055
8108
  cost: {
8056
8109
  input: 2,
8057
8110
  output: 6,
8058
- cacheRead: 0,
8111
+ cacheRead: 0.19999999999999998,
8059
8112
  cacheWrite: 0,
8060
8113
  },
8061
8114
  contextWindow: 131072,
@@ -8072,7 +8125,7 @@ export const MODELS = {
8072
8125
  cost: {
8073
8126
  input: 0.09999999999999999,
8074
8127
  output: 0.3,
8075
- cacheRead: 0,
8128
+ cacheRead: 0.01,
8076
8129
  cacheWrite: 0,
8077
8130
  },
8078
8131
  contextWindow: 32000,
@@ -8945,6 +8998,40 @@ export const MODELS = {
8945
8998
  contextWindow: 1050000,
8946
8999
  maxTokens: 128000,
8947
9000
  },
9001
+ "openai/gpt-5.4-mini": {
9002
+ id: "openai/gpt-5.4-mini",
9003
+ name: "OpenAI: GPT-5.4 Mini",
9004
+ api: "openai-completions",
9005
+ provider: "openrouter",
9006
+ baseUrl: "https://openrouter.ai/api/v1",
9007
+ reasoning: true,
9008
+ input: ["text", "image"],
9009
+ cost: {
9010
+ input: 0.75,
9011
+ output: 4.5,
9012
+ cacheRead: 0.075,
9013
+ cacheWrite: 0,
9014
+ },
9015
+ contextWindow: 400000,
9016
+ maxTokens: 128000,
9017
+ },
9018
+ "openai/gpt-5.4-nano": {
9019
+ id: "openai/gpt-5.4-nano",
9020
+ name: "OpenAI: GPT-5.4 Nano",
9021
+ api: "openai-completions",
9022
+ provider: "openrouter",
9023
+ baseUrl: "https://openrouter.ai/api/v1",
9024
+ reasoning: true,
9025
+ input: ["text", "image"],
9026
+ cost: {
9027
+ input: 0.19999999999999998,
9028
+ output: 1.25,
9029
+ cacheRead: 0.02,
9030
+ cacheWrite: 0,
9031
+ },
9032
+ contextWindow: 400000,
9033
+ maxTokens: 128000,
9034
+ },
8948
9035
  "openai/gpt-5.4-pro": {
8949
9036
  id: "openai/gpt-5.4-pro",
8950
9037
  name: "OpenAI: GPT-5.4 Pro",
@@ -9053,7 +9140,7 @@ export const MODELS = {
9053
9140
  api: "openai-completions",
9054
9141
  provider: "openrouter",
9055
9142
  baseUrl: "https://openrouter.ai/api/v1",
9056
- reasoning: false,
9143
+ reasoning: true,
9057
9144
  input: ["text", "image"],
9058
9145
  cost: {
9059
9146
  input: 15,
@@ -9104,7 +9191,7 @@ export const MODELS = {
9104
9191
  api: "openai-completions",
9105
9192
  provider: "openrouter",
9106
9193
  baseUrl: "https://openrouter.ai/api/v1",
9107
- reasoning: false,
9194
+ reasoning: true,
9108
9195
  input: ["text"],
9109
9196
  cost: {
9110
9197
  input: 1.1,
@@ -9121,7 +9208,7 @@ export const MODELS = {
9121
9208
  api: "openai-completions",
9122
9209
  provider: "openrouter",
9123
9210
  baseUrl: "https://openrouter.ai/api/v1",
9124
- reasoning: false,
9211
+ reasoning: true,
9125
9212
  input: ["text"],
9126
9213
  cost: {
9127
9214
  input: 1.1,
@@ -10523,7 +10610,7 @@ export const MODELS = {
10523
10610
  cacheRead: 0.119,
10524
10611
  cacheWrite: 0,
10525
10612
  },
10526
- contextWindow: 202752,
10613
+ contextWindow: 80000,
10527
10614
  maxTokens: 131072,
10528
10615
  },
10529
10616
  "z-ai/glm-5-turbo": {
@@ -10554,7 +10641,7 @@ export const MODELS = {
10554
10641
  reasoning: true,
10555
10642
  input: ["text"],
10556
10643
  cost: {
10557
- input: 0.06,
10644
+ input: 0.12,
10558
10645
  output: 0.24,
10559
10646
  cacheRead: 0,
10560
10647
  cacheWrite: 0,
@@ -10607,7 +10694,7 @@ export const MODELS = {
10607
10694
  cost: {
10608
10695
  input: 0.29,
10609
10696
  output: 0.59,
10610
- cacheRead: 0,
10697
+ cacheRead: 0.145,
10611
10698
  cacheWrite: 0,
10612
10699
  },
10613
10700
  contextWindow: 131072,
@@ -10622,9 +10709,9 @@ export const MODELS = {
10622
10709
  reasoning: true,
10623
10710
  input: ["text", "image"],
10624
10711
  cost: {
10625
- input: 0.3,
10626
- output: 2.9000000000000004,
10627
- cacheRead: 0,
10712
+ input: 0.22999999999999998,
10713
+ output: 2.3,
10714
+ cacheRead: 0.19999999999999998,
10628
10715
  cacheWrite: 0,
10629
10716
  },
10630
10717
  contextWindow: 262114,
@@ -10641,7 +10728,7 @@ export const MODELS = {
10641
10728
  cost: {
10642
10729
  input: 0.39999999999999997,
10643
10730
  output: 1.5999999999999999,
10644
- cacheRead: 0,
10731
+ cacheRead: 0.022,
10645
10732
  cacheWrite: 0,
10646
10733
  },
10647
10734
  contextWindow: 262144,
@@ -11134,7 +11221,7 @@ export const MODELS = {
11134
11221
  cost: {
11135
11222
  input: 0.27,
11136
11223
  output: 1,
11137
- cacheRead: 0,
11224
+ cacheRead: 0.135,
11138
11225
  cacheWrite: 0,
11139
11226
  },
11140
11227
  contextWindow: 131072,
@@ -11149,9 +11236,9 @@ export const MODELS = {
11149
11236
  reasoning: false,
11150
11237
  input: ["text"],
11151
11238
  cost: {
11152
- input: 0.26,
11153
- output: 0.38,
11154
- cacheRead: 0.13,
11239
+ input: 0.28,
11240
+ output: 0.42,
11241
+ cacheRead: 0.028,
11155
11242
  cacheWrite: 0,
11156
11243
  },
11157
11244
  contextWindow: 128000,
@@ -11185,7 +11272,7 @@ export const MODELS = {
11185
11272
  cost: {
11186
11273
  input: 0.15,
11187
11274
  output: 0.6,
11188
- cacheRead: 0,
11275
+ cacheRead: 0.024999999999999998,
11189
11276
  cacheWrite: 0,
11190
11277
  },
11191
11278
  contextWindow: 1048576,
@@ -11202,7 +11289,7 @@ export const MODELS = {
11202
11289
  cost: {
11203
11290
  input: 0.075,
11204
11291
  output: 0.3,
11205
- cacheRead: 0,
11292
+ cacheRead: 0.02,
11206
11293
  cacheWrite: 0,
11207
11294
  },
11208
11295
  contextWindow: 1048576,
@@ -11242,40 +11329,6 @@ export const MODELS = {
11242
11329
  contextWindow: 1048576,
11243
11330
  maxTokens: 65536,
11244
11331
  },
11245
- "google/gemini-2.5-flash-lite-preview-09-2025": {
11246
- id: "google/gemini-2.5-flash-lite-preview-09-2025",
11247
- name: "Gemini 2.5 Flash Lite Preview 09-2025",
11248
- api: "anthropic-messages",
11249
- provider: "vercel-ai-gateway",
11250
- baseUrl: "https://ai-gateway.vercel.sh",
11251
- reasoning: true,
11252
- input: ["text", "image"],
11253
- cost: {
11254
- input: 0.09999999999999999,
11255
- output: 0.39999999999999997,
11256
- cacheRead: 0.01,
11257
- cacheWrite: 0,
11258
- },
11259
- contextWindow: 1048576,
11260
- maxTokens: 65536,
11261
- },
11262
- "google/gemini-2.5-flash-preview-09-2025": {
11263
- id: "google/gemini-2.5-flash-preview-09-2025",
11264
- name: "Gemini 2.5 Flash Preview 09-2025",
11265
- api: "anthropic-messages",
11266
- provider: "vercel-ai-gateway",
11267
- baseUrl: "https://ai-gateway.vercel.sh",
11268
- reasoning: true,
11269
- input: ["text", "image"],
11270
- cost: {
11271
- input: 0.3,
11272
- output: 2.5,
11273
- cacheRead: 0.03,
11274
- cacheWrite: 0,
11275
- },
11276
- contextWindow: 1000000,
11277
- maxTokens: 65536,
11278
- },
11279
11332
  "google/gemini-2.5-pro": {
11280
11333
  id: "google/gemini-2.5-pro",
11281
11334
  name: "Gemini 2.5 Pro",
@@ -11457,7 +11510,7 @@ export const MODELS = {
11457
11510
  cost: {
11458
11511
  input: 0.09999999999999999,
11459
11512
  output: 0.09999999999999999,
11460
- cacheRead: 0,
11513
+ cacheRead: 0.09999999999999999,
11461
11514
  cacheWrite: 0,
11462
11515
  },
11463
11516
  contextWindow: 128000,
@@ -11630,8 +11683,8 @@ export const MODELS = {
11630
11683
  cacheRead: 0.03,
11631
11684
  cacheWrite: 0.375,
11632
11685
  },
11633
- contextWindow: 4096,
11634
- maxTokens: 4096,
11686
+ contextWindow: 204800,
11687
+ maxTokens: 131000,
11635
11688
  },
11636
11689
  "mistral/codestral": {
11637
11690
  id: "mistral/codestral",
@@ -11659,8 +11712,8 @@ export const MODELS = {
11659
11712
  reasoning: false,
11660
11713
  input: ["text"],
11661
11714
  cost: {
11662
- input: 0,
11663
- output: 0,
11715
+ input: 0.39999999999999997,
11716
+ output: 2,
11664
11717
  cacheRead: 0,
11665
11718
  cacheWrite: 0,
11666
11719
  },
@@ -11693,8 +11746,8 @@ export const MODELS = {
11693
11746
  reasoning: false,
11694
11747
  input: ["text"],
11695
11748
  cost: {
11696
- input: 0,
11697
- output: 0,
11749
+ input: 0.09999999999999999,
11750
+ output: 0.3,
11698
11751
  cacheRead: 0,
11699
11752
  cacheWrite: 0,
11700
11753
  },
@@ -11710,8 +11763,8 @@ export const MODELS = {
11710
11763
  reasoning: false,
11711
11764
  input: ["text"],
11712
11765
  cost: {
11713
- input: 0.04,
11714
- output: 0.04,
11766
+ input: 0.09999999999999999,
11767
+ output: 0.09999999999999999,
11715
11768
  cacheRead: 0,
11716
11769
  cacheWrite: 0,
11717
11770
  },
@@ -11727,8 +11780,8 @@ export const MODELS = {
11727
11780
  reasoning: false,
11728
11781
  input: ["text"],
11729
11782
  cost: {
11730
- input: 0.09999999999999999,
11731
- output: 0.09999999999999999,
11783
+ input: 0.15,
11784
+ output: 0.15,
11732
11785
  cacheRead: 0,
11733
11786
  cacheWrite: 0,
11734
11787
  },
@@ -11814,7 +11867,7 @@ export const MODELS = {
11814
11867
  cost: {
11815
11868
  input: 0.6,
11816
11869
  output: 2.5,
11817
- cacheRead: 0,
11870
+ cacheRead: 0.15,
11818
11871
  cacheWrite: 0,
11819
11872
  },
11820
11873
  contextWindow: 131072,
@@ -11831,7 +11884,7 @@ export const MODELS = {
11831
11884
  cost: {
11832
11885
  input: 0.6,
11833
11886
  output: 2.5,
11834
- cacheRead: 0,
11887
+ cacheRead: 0.15,
11835
11888
  cacheWrite: 0,
11836
11889
  },
11837
11890
  contextWindow: 256000,
@@ -11880,9 +11933,9 @@ export const MODELS = {
11880
11933
  reasoning: false,
11881
11934
  input: ["text"],
11882
11935
  cost: {
11883
- input: 2.4,
11884
- output: 10,
11885
- cacheRead: 0,
11936
+ input: 1.15,
11937
+ output: 8,
11938
+ cacheRead: 0.15,
11886
11939
  cacheWrite: 0,
11887
11940
  },
11888
11941
  contextWindow: 256000,
@@ -12347,6 +12400,40 @@ export const MODELS = {
12347
12400
  contextWindow: 1050000,
12348
12401
  maxTokens: 128000,
12349
12402
  },
12403
+ "openai/gpt-5.4-mini": {
12404
+ id: "openai/gpt-5.4-mini",
12405
+ name: "GPT 5.4 Mini",
12406
+ api: "anthropic-messages",
12407
+ provider: "vercel-ai-gateway",
12408
+ baseUrl: "https://ai-gateway.vercel.sh",
12409
+ reasoning: true,
12410
+ input: ["text", "image"],
12411
+ cost: {
12412
+ input: 0.75,
12413
+ output: 4.5,
12414
+ cacheRead: 0.075,
12415
+ cacheWrite: 0,
12416
+ },
12417
+ contextWindow: 400000,
12418
+ maxTokens: 128000,
12419
+ },
12420
+ "openai/gpt-5.4-nano": {
12421
+ id: "openai/gpt-5.4-nano",
12422
+ name: "GPT 5.4 Nano",
12423
+ api: "anthropic-messages",
12424
+ provider: "vercel-ai-gateway",
12425
+ baseUrl: "https://ai-gateway.vercel.sh",
12426
+ reasoning: true,
12427
+ input: ["text", "image"],
12428
+ cost: {
12429
+ input: 0.19999999999999998,
12430
+ output: 1.25,
12431
+ cacheRead: 0.02,
12432
+ cacheWrite: 0,
12433
+ },
12434
+ contextWindow: 400000,
12435
+ maxTokens: 128000,
12436
+ },
12350
12437
  "openai/gpt-5.4-pro": {
12351
12438
  id: "openai/gpt-5.4-pro",
12352
12439
  name: "GPT 5.4 Pro",
@@ -12551,40 +12638,6 @@ export const MODELS = {
12551
12638
  contextWindow: 131072,
12552
12639
  maxTokens: 131072,
12553
12640
  },
12554
- "vercel/v0-1.0-md": {
12555
- id: "vercel/v0-1.0-md",
12556
- name: "v0-1.0-md",
12557
- api: "anthropic-messages",
12558
- provider: "vercel-ai-gateway",
12559
- baseUrl: "https://ai-gateway.vercel.sh",
12560
- reasoning: false,
12561
- input: ["text", "image"],
12562
- cost: {
12563
- input: 3,
12564
- output: 15,
12565
- cacheRead: 0,
12566
- cacheWrite: 0,
12567
- },
12568
- contextWindow: 128000,
12569
- maxTokens: 32000,
12570
- },
12571
- "vercel/v0-1.5-md": {
12572
- id: "vercel/v0-1.5-md",
12573
- name: "v0-1.5-md",
12574
- api: "anthropic-messages",
12575
- provider: "vercel-ai-gateway",
12576
- baseUrl: "https://ai-gateway.vercel.sh",
12577
- reasoning: false,
12578
- input: ["text", "image"],
12579
- cost: {
12580
- input: 3,
12581
- output: 15,
12582
- cacheRead: 0,
12583
- cacheWrite: 0,
12584
- },
12585
- contextWindow: 128000,
12586
- maxTokens: 32768,
12587
- },
12588
12641
  "xai/grok-2-vision": {
12589
12642
  id: "xai/grok-2-vision",
12590
12643
  name: "Grok 2 Vision",
@@ -12613,7 +12666,7 @@ export const MODELS = {
12613
12666
  cost: {
12614
12667
  input: 3,
12615
12668
  output: 15,
12616
- cacheRead: 0,
12669
+ cacheRead: 0.75,
12617
12670
  cacheWrite: 0,
12618
12671
  },
12619
12672
  contextWindow: 131072,
@@ -12630,7 +12683,7 @@ export const MODELS = {
12630
12683
  cost: {
12631
12684
  input: 5,
12632
12685
  output: 25,
12633
- cacheRead: 0,
12686
+ cacheRead: 1.25,
12634
12687
  cacheWrite: 0,
12635
12688
  },
12636
12689
  contextWindow: 131072,
@@ -12647,7 +12700,7 @@ export const MODELS = {
12647
12700
  cost: {
12648
12701
  input: 0.3,
12649
12702
  output: 0.5,
12650
- cacheRead: 0,
12703
+ cacheRead: 0.075,
12651
12704
  cacheWrite: 0,
12652
12705
  },
12653
12706
  contextWindow: 131072,
@@ -12900,8 +12953,8 @@ export const MODELS = {
12900
12953
  reasoning: true,
12901
12954
  input: ["text"],
12902
12955
  cost: {
12903
- input: 0.44999999999999996,
12904
- output: 1.7999999999999998,
12956
+ input: 0.6,
12957
+ output: 2.2,
12905
12958
  cacheRead: 0.11,
12906
12959
  cacheWrite: 0,
12907
12960
  },