@mariozechner/pi-ai 0.68.0 → 0.69.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/README.md +3 -1
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +1 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +2 -2
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -1
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.generated.d.ts +561 -75
  10. package/dist/models.generated.d.ts.map +1 -1
  11. package/dist/models.generated.js +584 -92
  12. package/dist/models.generated.js.map +1 -1
  13. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  14. package/dist/providers/amazon-bedrock.js +49 -9
  15. package/dist/providers/amazon-bedrock.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +140 -8
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/transform-messages.d.ts.map +1 -1
  20. package/dist/providers/transform-messages.js +2 -0
  21. package/dist/providers/transform-messages.js.map +1 -1
  22. package/dist/types.d.ts +2 -2
  23. package/dist/types.d.ts.map +1 -1
  24. package/dist/types.js.map +1 -1
  25. package/dist/utils/json-parse.d.ts +8 -1
  26. package/dist/utils/json-parse.d.ts.map +1 -1
  27. package/dist/utils/json-parse.js +89 -5
  28. package/dist/utils/json-parse.js.map +1 -1
  29. package/dist/utils/typebox-helpers.d.ts +1 -1
  30. package/dist/utils/typebox-helpers.d.ts.map +1 -1
  31. package/dist/utils/typebox-helpers.js +1 -1
  32. package/dist/utils/typebox-helpers.js.map +1 -1
  33. package/dist/utils/validation.d.ts.map +1 -1
  34. package/dist/utils/validation.js +242 -41
  35. package/dist/utils/validation.js.map +1 -1
  36. package/package.json +2 -4
@@ -381,7 +381,7 @@ export const MODELS = {
381
381
  name: "Claude Haiku 4.5 (EU)",
382
382
  api: "bedrock-converse-stream",
383
383
  provider: "amazon-bedrock",
384
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
384
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
385
385
  reasoning: true,
386
386
  input: ["text", "image"],
387
387
  cost: {
@@ -398,7 +398,7 @@ export const MODELS = {
398
398
  name: "Claude Opus 4.5 (EU)",
399
399
  api: "bedrock-converse-stream",
400
400
  provider: "amazon-bedrock",
401
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
401
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
402
402
  reasoning: true,
403
403
  input: ["text", "image"],
404
404
  cost: {
@@ -415,7 +415,7 @@ export const MODELS = {
415
415
  name: "Claude Opus 4.6 (EU)",
416
416
  api: "bedrock-converse-stream",
417
417
  provider: "amazon-bedrock",
418
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
418
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
419
419
  reasoning: true,
420
420
  input: ["text", "image"],
421
421
  cost: {
@@ -432,7 +432,7 @@ export const MODELS = {
432
432
  name: "Claude Opus 4.7 (EU)",
433
433
  api: "bedrock-converse-stream",
434
434
  provider: "amazon-bedrock",
435
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
435
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
436
436
  reasoning: true,
437
437
  input: ["text", "image"],
438
438
  cost: {
@@ -449,7 +449,7 @@ export const MODELS = {
449
449
  name: "Claude Sonnet 4 (EU)",
450
450
  api: "bedrock-converse-stream",
451
451
  provider: "amazon-bedrock",
452
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
452
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
453
453
  reasoning: true,
454
454
  input: ["text", "image"],
455
455
  cost: {
@@ -466,7 +466,7 @@ export const MODELS = {
466
466
  name: "Claude Sonnet 4.5 (EU)",
467
467
  api: "bedrock-converse-stream",
468
468
  provider: "amazon-bedrock",
469
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
469
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
470
470
  reasoning: true,
471
471
  input: ["text", "image"],
472
472
  cost: {
@@ -483,7 +483,7 @@ export const MODELS = {
483
483
  name: "Claude Sonnet 4.6 (EU)",
484
484
  api: "bedrock-converse-stream",
485
485
  provider: "amazon-bedrock",
486
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
486
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
487
487
  reasoning: true,
488
488
  input: ["text", "image"],
489
489
  cost: {
@@ -2695,6 +2695,314 @@ export const MODELS = {
2695
2695
  maxTokens: 40000,
2696
2696
  },
2697
2697
  },
2698
+ "fireworks": {
2699
+ "accounts/fireworks/models/deepseek-v3p1": {
2700
+ id: "accounts/fireworks/models/deepseek-v3p1",
2701
+ name: "DeepSeek V3.1",
2702
+ api: "anthropic-messages",
2703
+ provider: "fireworks",
2704
+ baseUrl: "https://api.fireworks.ai/inference",
2705
+ reasoning: true,
2706
+ input: ["text"],
2707
+ cost: {
2708
+ input: 0.56,
2709
+ output: 1.68,
2710
+ cacheRead: 0,
2711
+ cacheWrite: 0,
2712
+ },
2713
+ contextWindow: 163840,
2714
+ maxTokens: 163840,
2715
+ },
2716
+ "accounts/fireworks/models/deepseek-v3p2": {
2717
+ id: "accounts/fireworks/models/deepseek-v3p2",
2718
+ name: "DeepSeek V3.2",
2719
+ api: "anthropic-messages",
2720
+ provider: "fireworks",
2721
+ baseUrl: "https://api.fireworks.ai/inference",
2722
+ reasoning: true,
2723
+ input: ["text"],
2724
+ cost: {
2725
+ input: 0.56,
2726
+ output: 1.68,
2727
+ cacheRead: 0.28,
2728
+ cacheWrite: 0,
2729
+ },
2730
+ contextWindow: 160000,
2731
+ maxTokens: 160000,
2732
+ },
2733
+ "accounts/fireworks/models/glm-4p5": {
2734
+ id: "accounts/fireworks/models/glm-4p5",
2735
+ name: "GLM 4.5",
2736
+ api: "anthropic-messages",
2737
+ provider: "fireworks",
2738
+ baseUrl: "https://api.fireworks.ai/inference",
2739
+ reasoning: true,
2740
+ input: ["text"],
2741
+ cost: {
2742
+ input: 0.55,
2743
+ output: 2.19,
2744
+ cacheRead: 0,
2745
+ cacheWrite: 0,
2746
+ },
2747
+ contextWindow: 131072,
2748
+ maxTokens: 131072,
2749
+ },
2750
+ "accounts/fireworks/models/glm-4p5-air": {
2751
+ id: "accounts/fireworks/models/glm-4p5-air",
2752
+ name: "GLM 4.5 Air",
2753
+ api: "anthropic-messages",
2754
+ provider: "fireworks",
2755
+ baseUrl: "https://api.fireworks.ai/inference",
2756
+ reasoning: true,
2757
+ input: ["text"],
2758
+ cost: {
2759
+ input: 0.22,
2760
+ output: 0.88,
2761
+ cacheRead: 0,
2762
+ cacheWrite: 0,
2763
+ },
2764
+ contextWindow: 131072,
2765
+ maxTokens: 131072,
2766
+ },
2767
+ "accounts/fireworks/models/glm-4p7": {
2768
+ id: "accounts/fireworks/models/glm-4p7",
2769
+ name: "GLM 4.7",
2770
+ api: "anthropic-messages",
2771
+ provider: "fireworks",
2772
+ baseUrl: "https://api.fireworks.ai/inference",
2773
+ reasoning: true,
2774
+ input: ["text"],
2775
+ cost: {
2776
+ input: 0.6,
2777
+ output: 2.2,
2778
+ cacheRead: 0.3,
2779
+ cacheWrite: 0,
2780
+ },
2781
+ contextWindow: 198000,
2782
+ maxTokens: 198000,
2783
+ },
2784
+ "accounts/fireworks/models/glm-5": {
2785
+ id: "accounts/fireworks/models/glm-5",
2786
+ name: "GLM 5",
2787
+ api: "anthropic-messages",
2788
+ provider: "fireworks",
2789
+ baseUrl: "https://api.fireworks.ai/inference",
2790
+ reasoning: true,
2791
+ input: ["text"],
2792
+ cost: {
2793
+ input: 1,
2794
+ output: 3.2,
2795
+ cacheRead: 0.5,
2796
+ cacheWrite: 0,
2797
+ },
2798
+ contextWindow: 202752,
2799
+ maxTokens: 131072,
2800
+ },
2801
+ "accounts/fireworks/models/glm-5p1": {
2802
+ id: "accounts/fireworks/models/glm-5p1",
2803
+ name: "GLM 5.1",
2804
+ api: "anthropic-messages",
2805
+ provider: "fireworks",
2806
+ baseUrl: "https://api.fireworks.ai/inference",
2807
+ reasoning: true,
2808
+ input: ["text"],
2809
+ cost: {
2810
+ input: 1.4,
2811
+ output: 4.4,
2812
+ cacheRead: 0.26,
2813
+ cacheWrite: 0,
2814
+ },
2815
+ contextWindow: 202800,
2816
+ maxTokens: 131072,
2817
+ },
2818
+ "accounts/fireworks/models/gpt-oss-120b": {
2819
+ id: "accounts/fireworks/models/gpt-oss-120b",
2820
+ name: "GPT OSS 120B",
2821
+ api: "anthropic-messages",
2822
+ provider: "fireworks",
2823
+ baseUrl: "https://api.fireworks.ai/inference",
2824
+ reasoning: true,
2825
+ input: ["text"],
2826
+ cost: {
2827
+ input: 0.15,
2828
+ output: 0.6,
2829
+ cacheRead: 0,
2830
+ cacheWrite: 0,
2831
+ },
2832
+ contextWindow: 131072,
2833
+ maxTokens: 32768,
2834
+ },
2835
+ "accounts/fireworks/models/gpt-oss-20b": {
2836
+ id: "accounts/fireworks/models/gpt-oss-20b",
2837
+ name: "GPT OSS 20B",
2838
+ api: "anthropic-messages",
2839
+ provider: "fireworks",
2840
+ baseUrl: "https://api.fireworks.ai/inference",
2841
+ reasoning: true,
2842
+ input: ["text"],
2843
+ cost: {
2844
+ input: 0.05,
2845
+ output: 0.2,
2846
+ cacheRead: 0,
2847
+ cacheWrite: 0,
2848
+ },
2849
+ contextWindow: 131072,
2850
+ maxTokens: 32768,
2851
+ },
2852
+ "accounts/fireworks/models/kimi-k2-instruct": {
2853
+ id: "accounts/fireworks/models/kimi-k2-instruct",
2854
+ name: "Kimi K2 Instruct",
2855
+ api: "anthropic-messages",
2856
+ provider: "fireworks",
2857
+ baseUrl: "https://api.fireworks.ai/inference",
2858
+ reasoning: false,
2859
+ input: ["text"],
2860
+ cost: {
2861
+ input: 1,
2862
+ output: 3,
2863
+ cacheRead: 0,
2864
+ cacheWrite: 0,
2865
+ },
2866
+ contextWindow: 128000,
2867
+ maxTokens: 16384,
2868
+ },
2869
+ "accounts/fireworks/models/kimi-k2-thinking": {
2870
+ id: "accounts/fireworks/models/kimi-k2-thinking",
2871
+ name: "Kimi K2 Thinking",
2872
+ api: "anthropic-messages",
2873
+ provider: "fireworks",
2874
+ baseUrl: "https://api.fireworks.ai/inference",
2875
+ reasoning: true,
2876
+ input: ["text"],
2877
+ cost: {
2878
+ input: 0.6,
2879
+ output: 2.5,
2880
+ cacheRead: 0.3,
2881
+ cacheWrite: 0,
2882
+ },
2883
+ contextWindow: 256000,
2884
+ maxTokens: 256000,
2885
+ },
2886
+ "accounts/fireworks/models/kimi-k2p5": {
2887
+ id: "accounts/fireworks/models/kimi-k2p5",
2888
+ name: "Kimi K2.5",
2889
+ api: "anthropic-messages",
2890
+ provider: "fireworks",
2891
+ baseUrl: "https://api.fireworks.ai/inference",
2892
+ reasoning: true,
2893
+ input: ["text", "image"],
2894
+ cost: {
2895
+ input: 0.6,
2896
+ output: 3,
2897
+ cacheRead: 0.1,
2898
+ cacheWrite: 0,
2899
+ },
2900
+ contextWindow: 256000,
2901
+ maxTokens: 256000,
2902
+ },
2903
+ "accounts/fireworks/models/kimi-k2p6": {
2904
+ id: "accounts/fireworks/models/kimi-k2p6",
2905
+ name: "Kimi K2.6",
2906
+ api: "anthropic-messages",
2907
+ provider: "fireworks",
2908
+ baseUrl: "https://api.fireworks.ai/inference",
2909
+ reasoning: true,
2910
+ input: ["text", "image"],
2911
+ cost: {
2912
+ input: 0.95,
2913
+ output: 4,
2914
+ cacheRead: 0.16,
2915
+ cacheWrite: 0,
2916
+ },
2917
+ contextWindow: 262000,
2918
+ maxTokens: 262000,
2919
+ },
2920
+ "accounts/fireworks/models/minimax-m2p1": {
2921
+ id: "accounts/fireworks/models/minimax-m2p1",
2922
+ name: "MiniMax-M2.1",
2923
+ api: "anthropic-messages",
2924
+ provider: "fireworks",
2925
+ baseUrl: "https://api.fireworks.ai/inference",
2926
+ reasoning: true,
2927
+ input: ["text"],
2928
+ cost: {
2929
+ input: 0.3,
2930
+ output: 1.2,
2931
+ cacheRead: 0.03,
2932
+ cacheWrite: 0,
2933
+ },
2934
+ contextWindow: 200000,
2935
+ maxTokens: 200000,
2936
+ },
2937
+ "accounts/fireworks/models/minimax-m2p5": {
2938
+ id: "accounts/fireworks/models/minimax-m2p5",
2939
+ name: "MiniMax-M2.5",
2940
+ api: "anthropic-messages",
2941
+ provider: "fireworks",
2942
+ baseUrl: "https://api.fireworks.ai/inference",
2943
+ reasoning: true,
2944
+ input: ["text"],
2945
+ cost: {
2946
+ input: 0.3,
2947
+ output: 1.2,
2948
+ cacheRead: 0.03,
2949
+ cacheWrite: 0,
2950
+ },
2951
+ contextWindow: 196608,
2952
+ maxTokens: 196608,
2953
+ },
2954
+ "accounts/fireworks/models/minimax-m2p7": {
2955
+ id: "accounts/fireworks/models/minimax-m2p7",
2956
+ name: "MiniMax-M2.7",
2957
+ api: "anthropic-messages",
2958
+ provider: "fireworks",
2959
+ baseUrl: "https://api.fireworks.ai/inference",
2960
+ reasoning: true,
2961
+ input: ["text"],
2962
+ cost: {
2963
+ input: 0.3,
2964
+ output: 1.2,
2965
+ cacheRead: 0.03,
2966
+ cacheWrite: 0,
2967
+ },
2968
+ contextWindow: 196608,
2969
+ maxTokens: 196608,
2970
+ },
2971
+ "accounts/fireworks/models/qwen3p6-plus": {
2972
+ id: "accounts/fireworks/models/qwen3p6-plus",
2973
+ name: "Qwen 3.6 Plus",
2974
+ api: "anthropic-messages",
2975
+ provider: "fireworks",
2976
+ baseUrl: "https://api.fireworks.ai/inference",
2977
+ reasoning: true,
2978
+ input: ["text", "image"],
2979
+ cost: {
2980
+ input: 0.5,
2981
+ output: 3,
2982
+ cacheRead: 0.1,
2983
+ cacheWrite: 0,
2984
+ },
2985
+ contextWindow: 128000,
2986
+ maxTokens: 8192,
2987
+ },
2988
+ "accounts/fireworks/routers/kimi-k2p5-turbo": {
2989
+ id: "accounts/fireworks/routers/kimi-k2p5-turbo",
2990
+ name: "Kimi K2.5 Turbo (firepass)",
2991
+ api: "anthropic-messages",
2992
+ provider: "fireworks",
2993
+ baseUrl: "https://api.fireworks.ai/inference",
2994
+ reasoning: true,
2995
+ input: ["text", "image"],
2996
+ cost: {
2997
+ input: 0,
2998
+ output: 0,
2999
+ cacheRead: 0,
3000
+ cacheWrite: 0,
3001
+ },
3002
+ contextWindow: 256000,
3003
+ maxTokens: 256000,
3004
+ },
3005
+ },
2698
3006
  "github-copilot": {
2699
3007
  "claude-haiku-4.5": {
2700
3008
  id: "claude-haiku-4.5",
@@ -3856,6 +4164,23 @@ export const MODELS = {
3856
4164
  contextWindow: 1048576,
3857
4165
  maxTokens: 65535,
3858
4166
  },
4167
+ "gemini-3.1-flash-lite-preview": {
4168
+ id: "gemini-3.1-flash-lite-preview",
4169
+ name: "Gemini 3.1 Flash Lite Preview (Cloud Code Assist)",
4170
+ api: "google-gemini-cli",
4171
+ provider: "google-gemini-cli",
4172
+ baseUrl: "https://cloudcode-pa.googleapis.com",
4173
+ reasoning: true,
4174
+ input: ["text", "image"],
4175
+ cost: {
4176
+ input: 0,
4177
+ output: 0,
4178
+ cacheRead: 0,
4179
+ cacheWrite: 0,
4180
+ },
4181
+ contextWindow: 1048576,
4182
+ maxTokens: 65535,
4183
+ },
3859
4184
  "gemini-3.1-pro-preview": {
3860
4185
  id: "gemini-3.1-pro-preview",
3861
4186
  name: "Gemini 3.1 Pro Preview (Cloud Code Assist)",
@@ -4694,6 +5019,24 @@ export const MODELS = {
4694
5019
  contextWindow: 262144,
4695
5020
  maxTokens: 262144,
4696
5021
  },
5022
+ "moonshotai/Kimi-K2.6": {
5023
+ id: "moonshotai/Kimi-K2.6",
5024
+ name: "Kimi-K2.6",
5025
+ api: "openai-completions",
5026
+ provider: "huggingface",
5027
+ baseUrl: "https://router.huggingface.co/v1",
5028
+ compat: { "supportsDeveloperRole": false },
5029
+ reasoning: true,
5030
+ input: ["text", "image"],
5031
+ cost: {
5032
+ input: 0.95,
5033
+ output: 4,
5034
+ cacheRead: 0.16,
5035
+ cacheWrite: 0,
5036
+ },
5037
+ contextWindow: 262144,
5038
+ maxTokens: 262144,
5039
+ },
4697
5040
  "zai-org/GLM-4.7": {
4698
5041
  id: "zai-org/GLM-4.7",
4699
5042
  name: "GLM-4.7",
@@ -4763,11 +5106,28 @@ export const MODELS = {
4763
5106
  cacheRead: 0.2,
4764
5107
  cacheWrite: 0,
4765
5108
  },
4766
- contextWindow: 202752,
4767
- maxTokens: 131072,
5109
+ contextWindow: 202752,
5110
+ maxTokens: 131072,
5111
+ },
5112
+ },
5113
+ "kimi-coding": {
5114
+ "k2p6": {
5115
+ id: "k2p6",
5116
+ name: "Kimi K2.6",
5117
+ api: "anthropic-messages",
5118
+ provider: "kimi-coding",
5119
+ baseUrl: "https://api.kimi.com/coding",
5120
+ reasoning: true,
5121
+ input: ["text", "image"],
5122
+ cost: {
5123
+ input: 0,
5124
+ output: 0,
5125
+ cacheRead: 0,
5126
+ cacheWrite: 0,
5127
+ },
5128
+ contextWindow: 262144,
5129
+ maxTokens: 32768,
4768
5130
  },
4769
- },
4770
- "kimi-coding": {
4771
5131
  "kimi-for-coding": {
4772
5132
  id: "kimi-for-coding",
4773
5133
  name: "Kimi For Coding",
@@ -6650,6 +7010,40 @@ export const MODELS = {
6650
7010
  contextWindow: 262144,
6651
7011
  maxTokens: 65536,
6652
7012
  },
7013
+ "kimi-k2.6": {
7014
+ id: "kimi-k2.6",
7015
+ name: "Kimi K2.6",
7016
+ api: "openai-completions",
7017
+ provider: "opencode",
7018
+ baseUrl: "https://opencode.ai/zen/v1",
7019
+ reasoning: true,
7020
+ input: ["text", "image"],
7021
+ cost: {
7022
+ input: 0.95,
7023
+ output: 4,
7024
+ cacheRead: 0.16,
7025
+ cacheWrite: 0,
7026
+ },
7027
+ contextWindow: 262144,
7028
+ maxTokens: 65536,
7029
+ },
7030
+ "ling-2.6-flash-free": {
7031
+ id: "ling-2.6-flash-free",
7032
+ name: "Ling 2.6 Flash Free",
7033
+ api: "openai-completions",
7034
+ provider: "opencode",
7035
+ baseUrl: "https://opencode.ai/zen/v1",
7036
+ reasoning: false,
7037
+ input: ["text"],
7038
+ cost: {
7039
+ input: 0,
7040
+ output: 0,
7041
+ cacheRead: 0,
7042
+ cacheWrite: 0,
7043
+ },
7044
+ contextWindow: 262100,
7045
+ maxTokens: 32800,
7046
+ },
6653
7047
  "minimax-m2.5": {
6654
7048
  id: "minimax-m2.5",
6655
7049
  name: "MiniMax M2.5",
@@ -6684,6 +7078,23 @@ export const MODELS = {
6684
7078
  contextWindow: 204800,
6685
7079
  maxTokens: 131072,
6686
7080
  },
7081
+ "minimax-m2.7": {
7082
+ id: "minimax-m2.7",
7083
+ name: "MiniMax M2.7",
7084
+ api: "openai-completions",
7085
+ provider: "opencode",
7086
+ baseUrl: "https://opencode.ai/zen/v1",
7087
+ reasoning: true,
7088
+ input: ["text"],
7089
+ cost: {
7090
+ input: 0.3,
7091
+ output: 1.2,
7092
+ cacheRead: 0.06,
7093
+ cacheWrite: 0,
7094
+ },
7095
+ contextWindow: 204800,
7096
+ maxTokens: 131072,
7097
+ },
6687
7098
  "nemotron-3-super-free": {
6688
7099
  id: "nemotron-3-super-free",
6689
7100
  name: "Nemotron 3 Super Free",
@@ -6704,10 +7115,9 @@ export const MODELS = {
6704
7115
  "qwen3.5-plus": {
6705
7116
  id: "qwen3.5-plus",
6706
7117
  name: "Qwen3.5 Plus",
6707
- api: "openai-completions",
7118
+ api: "anthropic-messages",
6708
7119
  provider: "opencode",
6709
- baseUrl: "https://opencode.ai/zen/v1",
6710
- compat: { "cacheControlFormat": "anthropic" },
7120
+ baseUrl: "https://opencode.ai/zen",
6711
7121
  reasoning: true,
6712
7122
  input: ["text", "image"],
6713
7123
  cost: {
@@ -6722,10 +7132,9 @@ export const MODELS = {
6722
7132
  "qwen3.6-plus": {
6723
7133
  id: "qwen3.6-plus",
6724
7134
  name: "Qwen3.6 Plus",
6725
- api: "openai-completions",
7135
+ api: "anthropic-messages",
6726
7136
  provider: "opencode",
6727
- baseUrl: "https://opencode.ai/zen/v1",
6728
- compat: { "cacheControlFormat": "anthropic" },
7137
+ baseUrl: "https://opencode.ai/zen",
6729
7138
  reasoning: true,
6730
7139
  input: ["text", "image"],
6731
7140
  cost: {
@@ -6792,16 +7201,16 @@ export const MODELS = {
6792
7201
  },
6793
7202
  "kimi-k2.6": {
6794
7203
  id: "kimi-k2.6",
6795
- name: "Kimi K2.6",
7204
+ name: "Kimi K2.6 (3x limits)",
6796
7205
  api: "openai-completions",
6797
7206
  provider: "opencode-go",
6798
7207
  baseUrl: "https://opencode.ai/zen/go/v1",
6799
7208
  reasoning: true,
6800
7209
  input: ["text", "image"],
6801
7210
  cost: {
6802
- input: 0.95,
6803
- output: 4,
6804
- cacheRead: 0.16,
7211
+ input: 0.32,
7212
+ output: 1.34,
7213
+ cacheRead: 0.054,
6805
7214
  cacheWrite: 0,
6806
7215
  },
6807
7216
  contextWindow: 262144,
@@ -6822,7 +7231,7 @@ export const MODELS = {
6822
7231
  cacheWrite: 0,
6823
7232
  },
6824
7233
  contextWindow: 262144,
6825
- maxTokens: 64000,
7234
+ maxTokens: 128000,
6826
7235
  },
6827
7236
  "mimo-v2-pro": {
6828
7237
  id: "mimo-v2-pro",
@@ -6839,7 +7248,41 @@ export const MODELS = {
6839
7248
  cacheWrite: 0,
6840
7249
  },
6841
7250
  contextWindow: 1048576,
6842
- maxTokens: 64000,
7251
+ maxTokens: 128000,
7252
+ },
7253
+ "mimo-v2.5": {
7254
+ id: "mimo-v2.5",
7255
+ name: "MiMo V2.5",
7256
+ api: "openai-completions",
7257
+ provider: "opencode-go",
7258
+ baseUrl: "https://opencode.ai/zen/go/v1",
7259
+ reasoning: true,
7260
+ input: ["text", "image"],
7261
+ cost: {
7262
+ input: 0.4,
7263
+ output: 2,
7264
+ cacheRead: 0.08,
7265
+ cacheWrite: 0,
7266
+ },
7267
+ contextWindow: 262144,
7268
+ maxTokens: 128000,
7269
+ },
7270
+ "mimo-v2.5-pro": {
7271
+ id: "mimo-v2.5-pro",
7272
+ name: "MiMo V2.5 Pro",
7273
+ api: "openai-completions",
7274
+ provider: "opencode-go",
7275
+ baseUrl: "https://opencode.ai/zen/go/v1",
7276
+ reasoning: true,
7277
+ input: ["text"],
7278
+ cost: {
7279
+ input: 1,
7280
+ output: 3,
7281
+ cacheRead: 0.2,
7282
+ cacheWrite: 0,
7283
+ },
7284
+ contextWindow: 1048576,
7285
+ maxTokens: 128000,
6843
7286
  },
6844
7287
  "minimax-m2.5": {
6845
7288
  id: "minimax-m2.5",
@@ -6878,10 +7321,9 @@ export const MODELS = {
6878
7321
  "qwen3.5-plus": {
6879
7322
  id: "qwen3.5-plus",
6880
7323
  name: "Qwen3.5 Plus",
6881
- api: "openai-completions",
7324
+ api: "anthropic-messages",
6882
7325
  provider: "opencode-go",
6883
- baseUrl: "https://opencode.ai/zen/go/v1",
6884
- compat: { "cacheControlFormat": "anthropic" },
7326
+ baseUrl: "https://opencode.ai/zen/go",
6885
7327
  reasoning: true,
6886
7328
  input: ["text", "image"],
6887
7329
  cost: {
@@ -6896,10 +7338,9 @@ export const MODELS = {
6896
7338
  "qwen3.6-plus": {
6897
7339
  id: "qwen3.6-plus",
6898
7340
  name: "Qwen3.6 Plus",
6899
- api: "openai-completions",
7341
+ api: "anthropic-messages",
6900
7342
  provider: "opencode-go",
6901
- baseUrl: "https://opencode.ai/zen/go/v1",
6902
- compat: { "cacheControlFormat": "anthropic" },
7343
+ baseUrl: "https://opencode.ai/zen/go",
6903
7344
  reasoning: true,
6904
7345
  input: ["text", "image"],
6905
7346
  cost: {
@@ -7287,6 +7728,23 @@ export const MODELS = {
7287
7728
  contextWindow: 1000000,
7288
7729
  maxTokens: 128000,
7289
7730
  },
7731
+ "arcee-ai/trinity-large-preview": {
7732
+ id: "arcee-ai/trinity-large-preview",
7733
+ name: "Arcee AI: Trinity Large Preview",
7734
+ api: "openai-completions",
7735
+ provider: "openrouter",
7736
+ baseUrl: "https://openrouter.ai/api/v1",
7737
+ reasoning: false,
7738
+ input: ["text"],
7739
+ cost: {
7740
+ input: 0.15,
7741
+ output: 0.44999999999999996,
7742
+ cacheRead: 0,
7743
+ cacheWrite: 0,
7744
+ },
7745
+ contextWindow: 131000,
7746
+ maxTokens: 4096,
7747
+ },
7290
7748
  "arcee-ai/trinity-large-preview:free": {
7291
7749
  id: "arcee-ai/trinity-large-preview:free",
7292
7750
  name: "Arcee AI: Trinity Large Preview (free)",
@@ -7874,13 +8332,13 @@ export const MODELS = {
7874
8332
  reasoning: true,
7875
8333
  input: ["text", "image"],
7876
8334
  cost: {
7877
- input: 0.08,
8335
+ input: 0.07,
7878
8336
  output: 0.35,
7879
- cacheRead: 0.01,
8337
+ cacheRead: 0.04,
7880
8338
  cacheWrite: 0,
7881
8339
  },
7882
8340
  contextWindow: 262144,
7883
- maxTokens: 4096,
8341
+ maxTokens: 262144,
7884
8342
  },
7885
8343
  "google/gemma-4-26b-a4b-it:free": {
7886
8344
  id: "google/gemma-4-26b-a4b-it:free",
@@ -7950,6 +8408,23 @@ export const MODELS = {
7950
8408
  contextWindow: 128000,
7951
8409
  maxTokens: 50000,
7952
8410
  },
8411
+ "inclusionai/ling-2.6-flash:free": {
8412
+ id: "inclusionai/ling-2.6-flash:free",
8413
+ name: "inclusionAI: Ling-2.6-flash (free)",
8414
+ api: "openai-completions",
8415
+ provider: "openrouter",
8416
+ baseUrl: "https://openrouter.ai/api/v1",
8417
+ reasoning: false,
8418
+ input: ["text"],
8419
+ cost: {
8420
+ input: 0,
8421
+ output: 0,
8422
+ cacheRead: 0,
8423
+ cacheWrite: 0,
8424
+ },
8425
+ contextWindow: 262144,
8426
+ maxTokens: 32768,
8427
+ },
7953
8428
  "kwaipilot/kat-coder-pro-v2": {
7954
8429
  id: "kwaipilot/kat-coder-pro-v2",
7955
8430
  name: "Kwaipilot: KAT-Coder-Pro V2",
@@ -8027,13 +8502,13 @@ export const MODELS = {
8027
8502
  reasoning: false,
8028
8503
  input: ["text"],
8029
8504
  cost: {
8030
- input: 0.12,
8031
- output: 0.38,
8505
+ input: 0.09999999999999999,
8506
+ output: 0.32,
8032
8507
  cacheRead: 0,
8033
8508
  cacheWrite: 0,
8034
8509
  },
8035
8510
  contextWindow: 131072,
8036
- maxTokens: 131072,
8511
+ maxTokens: 16384,
8037
8512
  },
8038
8513
  "meta-llama/llama-3.3-70b-instruct:free": {
8039
8514
  id: "meta-llama/llama-3.3-70b-instruct:free",
@@ -8622,9 +9097,9 @@ export const MODELS = {
8622
9097
  reasoning: true,
8623
9098
  input: ["text", "image"],
8624
9099
  cost: {
8625
- input: 0,
8626
- output: 0,
8627
- cacheRead: 0,
9100
+ input: 0.7999999999999999,
9101
+ output: 3.5,
9102
+ cacheRead: 0.19999999999999998,
8628
9103
  cacheWrite: 0,
8629
9104
  },
8630
9105
  contextWindow: 262144,
@@ -9140,40 +9615,6 @@ export const MODELS = {
9140
9615
  contextWindow: 400000,
9141
9616
  maxTokens: 128000,
9142
9617
  },
9143
- "openai/gpt-5-image": {
9144
- id: "openai/gpt-5-image",
9145
- name: "OpenAI: GPT-5 Image",
9146
- api: "openai-completions",
9147
- provider: "openrouter",
9148
- baseUrl: "https://openrouter.ai/api/v1",
9149
- reasoning: true,
9150
- input: ["text", "image"],
9151
- cost: {
9152
- input: 10,
9153
- output: 10,
9154
- cacheRead: 1.25,
9155
- cacheWrite: 0,
9156
- },
9157
- contextWindow: 400000,
9158
- maxTokens: 128000,
9159
- },
9160
- "openai/gpt-5-image-mini": {
9161
- id: "openai/gpt-5-image-mini",
9162
- name: "OpenAI: GPT-5 Image Mini",
9163
- api: "openai-completions",
9164
- provider: "openrouter",
9165
- baseUrl: "https://openrouter.ai/api/v1",
9166
- reasoning: true,
9167
- input: ["text", "image"],
9168
- cost: {
9169
- input: 2.5,
9170
- output: 2,
9171
- cacheRead: 0.25,
9172
- cacheWrite: 0,
9173
- },
9174
- contextWindow: 400000,
9175
- maxTokens: 128000,
9176
- },
9177
9618
  "openai/gpt-5-mini": {
9178
9619
  id: "openai/gpt-5-mini",
9179
9620
  name: "OpenAI: GPT-5 Mini",
@@ -9769,23 +10210,6 @@ export const MODELS = {
9769
10210
  contextWindow: 2000000,
9770
10211
  maxTokens: 4096,
9771
10212
  },
9772
- "openrouter/elephant-alpha": {
9773
- id: "openrouter/elephant-alpha",
9774
- name: "Elephant",
9775
- api: "openai-completions",
9776
- provider: "openrouter",
9777
- baseUrl: "https://openrouter.ai/api/v1",
9778
- reasoning: false,
9779
- input: ["text"],
9780
- cost: {
9781
- input: 0,
9782
- output: 0,
9783
- cacheRead: 0,
9784
- cacheWrite: 0,
9785
- },
9786
- contextWindow: 262144,
9787
- maxTokens: 32768,
9788
- },
9789
10213
  "openrouter/free": {
9790
10214
  id: "openrouter/free",
9791
10215
  name: "Free Models Router",
@@ -10171,7 +10595,7 @@ export const MODELS = {
10171
10595
  cost: {
10172
10596
  input: 0.15,
10173
10597
  output: 0.7999999999999999,
10174
- cacheRead: 0.12,
10598
+ cacheRead: 0.11,
10175
10599
  cacheWrite: 0,
10176
10600
  },
10177
10601
  contextWindow: 262144,
@@ -10925,6 +11349,40 @@ export const MODELS = {
10925
11349
  contextWindow: 1048576,
10926
11350
  maxTokens: 131072,
10927
11351
  },
11352
+ "xiaomi/mimo-v2.5": {
11353
+ id: "xiaomi/mimo-v2.5",
11354
+ name: "Xiaomi: MiMo-V2.5",
11355
+ api: "openai-completions",
11356
+ provider: "openrouter",
11357
+ baseUrl: "https://openrouter.ai/api/v1",
11358
+ reasoning: true,
11359
+ input: ["text", "image"],
11360
+ cost: {
11361
+ input: 0.39999999999999997,
11362
+ output: 2,
11363
+ cacheRead: 0.08,
11364
+ cacheWrite: 0,
11365
+ },
11366
+ contextWindow: 1048576,
11367
+ maxTokens: 131072,
11368
+ },
11369
+ "xiaomi/mimo-v2.5-pro": {
11370
+ id: "xiaomi/mimo-v2.5-pro",
11371
+ name: "Xiaomi: MiMo-V2.5-Pro",
11372
+ api: "openai-completions",
11373
+ provider: "openrouter",
11374
+ baseUrl: "https://openrouter.ai/api/v1",
11375
+ reasoning: true,
11376
+ input: ["text"],
11377
+ cost: {
11378
+ input: 1,
11379
+ output: 3,
11380
+ cacheRead: 0.19999999999999998,
11381
+ cacheWrite: 0,
11382
+ },
11383
+ contextWindow: 1048576,
11384
+ maxTokens: 131072,
11385
+ },
10928
11386
  "z-ai/glm-4-32b": {
10929
11387
  id: "z-ai/glm-4-32b",
10930
11388
  name: "Z.ai: GLM 4 32B ",
@@ -11092,8 +11550,8 @@ export const MODELS = {
11092
11550
  cacheRead: 0.119,
11093
11551
  cacheWrite: 0,
11094
11552
  },
11095
- contextWindow: 80000,
11096
- maxTokens: 131072,
11553
+ contextWindow: 202752,
11554
+ maxTokens: 4096,
11097
11555
  },
11098
11556
  "z-ai/glm-5-turbo": {
11099
11557
  id: "z-ai/glm-5-turbo",
@@ -11146,6 +11604,23 @@ export const MODELS = {
11146
11604
  contextWindow: 202752,
11147
11605
  maxTokens: 131072,
11148
11606
  },
11607
+ "~anthropic/claude-opus-latest": {
11608
+ id: "~anthropic/claude-opus-latest",
11609
+ name: "Anthropic: Claude Opus Latest",
11610
+ api: "openai-completions",
11611
+ provider: "openrouter",
11612
+ baseUrl: "https://openrouter.ai/api/v1",
11613
+ reasoning: true,
11614
+ input: ["text", "image"],
11615
+ cost: {
11616
+ input: 5,
11617
+ output: 25,
11618
+ cacheRead: 0.5,
11619
+ cacheWrite: 6.25,
11620
+ },
11621
+ contextWindow: 1000000,
11622
+ maxTokens: 128000,
11623
+ },
11149
11624
  },
11150
11625
  "vercel-ai-gateway": {
11151
11626
  "alibaba/qwen-3-14b": {
@@ -11216,6 +11691,23 @@ export const MODELS = {
11216
11691
  contextWindow: 128000,
11217
11692
  maxTokens: 8192,
11218
11693
  },
11694
+ "alibaba/qwen-3.6-max-preview": {
11695
+ id: "alibaba/qwen-3.6-max-preview",
11696
+ name: "Qwen 3.6 Max Preview",
11697
+ api: "anthropic-messages",
11698
+ provider: "vercel-ai-gateway",
11699
+ baseUrl: "https://ai-gateway.vercel.sh",
11700
+ reasoning: true,
11701
+ input: ["text", "image"],
11702
+ cost: {
11703
+ input: 1.3,
11704
+ output: 7.8,
11705
+ cacheRead: 0.26,
11706
+ cacheWrite: 1.625,
11707
+ },
11708
+ contextWindow: 240000,
11709
+ maxTokens: 64000,
11710
+ },
11219
11711
  "alibaba/qwen3-235b-a22b-thinking": {
11220
11712
  id: "alibaba/qwen3-235b-a22b-thinking",
11221
11713
  name: "Qwen3 235B A22B Thinking 2507",