@mariozechner/pi-ai 0.67.68 → 0.68.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (50) hide show
  1. package/README.md +5 -1
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +1 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/models.generated.d.ts +502 -77
  6. package/dist/models.generated.d.ts.map +1 -1
  7. package/dist/models.generated.js +532 -117
  8. package/dist/models.generated.js.map +1 -1
  9. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  10. package/dist/providers/amazon-bedrock.js +55 -7
  11. package/dist/providers/amazon-bedrock.js.map +1 -1
  12. package/dist/providers/anthropic.d.ts.map +1 -1
  13. package/dist/providers/anthropic.js +141 -10
  14. package/dist/providers/anthropic.js.map +1 -1
  15. package/dist/providers/google-shared.d.ts.map +1 -1
  16. package/dist/providers/google-shared.js +30 -4
  17. package/dist/providers/google-shared.js.map +1 -1
  18. package/dist/providers/openai-completions.d.ts +5 -1
  19. package/dist/providers/openai-completions.d.ts.map +1 -1
  20. package/dist/providers/openai-completions.js +149 -60
  21. package/dist/providers/openai-completions.js.map +1 -1
  22. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  23. package/dist/providers/openai-responses-shared.js +2 -5
  24. package/dist/providers/openai-responses-shared.js.map +1 -1
  25. package/dist/providers/simple-options.d.ts.map +1 -1
  26. package/dist/providers/simple-options.js +1 -1
  27. package/dist/providers/simple-options.js.map +1 -1
  28. package/dist/providers/transform-messages.d.ts.map +1 -1
  29. package/dist/providers/transform-messages.js +41 -2
  30. package/dist/providers/transform-messages.js.map +1 -1
  31. package/dist/types.d.ts +5 -1
  32. package/dist/types.d.ts.map +1 -1
  33. package/dist/types.js.map +1 -1
  34. package/dist/utils/json-parse.d.ts +8 -1
  35. package/dist/utils/json-parse.d.ts.map +1 -1
  36. package/dist/utils/json-parse.js +89 -5
  37. package/dist/utils/json-parse.js.map +1 -1
  38. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  39. package/dist/utils/oauth/anthropic.js +1 -1
  40. package/dist/utils/oauth/anthropic.js.map +1 -1
  41. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  42. package/dist/utils/oauth/google-antigravity.js +2 -1
  43. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  44. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  45. package/dist/utils/oauth/google-gemini-cli.js +2 -1
  46. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  47. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  48. package/dist/utils/oauth/openai-codex.js +3 -2
  49. package/dist/utils/oauth/openai-codex.js.map +1 -1
  50. package/package.json +1 -1
@@ -381,7 +381,7 @@ export const MODELS = {
381
381
  name: "Claude Haiku 4.5 (EU)",
382
382
  api: "bedrock-converse-stream",
383
383
  provider: "amazon-bedrock",
384
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
384
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
385
385
  reasoning: true,
386
386
  input: ["text", "image"],
387
387
  cost: {
@@ -398,7 +398,7 @@ export const MODELS = {
398
398
  name: "Claude Opus 4.5 (EU)",
399
399
  api: "bedrock-converse-stream",
400
400
  provider: "amazon-bedrock",
401
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
401
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
402
402
  reasoning: true,
403
403
  input: ["text", "image"],
404
404
  cost: {
@@ -415,7 +415,7 @@ export const MODELS = {
415
415
  name: "Claude Opus 4.6 (EU)",
416
416
  api: "bedrock-converse-stream",
417
417
  provider: "amazon-bedrock",
418
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
418
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
419
419
  reasoning: true,
420
420
  input: ["text", "image"],
421
421
  cost: {
@@ -432,7 +432,7 @@ export const MODELS = {
432
432
  name: "Claude Opus 4.7 (EU)",
433
433
  api: "bedrock-converse-stream",
434
434
  provider: "amazon-bedrock",
435
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
435
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
436
436
  reasoning: true,
437
437
  input: ["text", "image"],
438
438
  cost: {
@@ -449,7 +449,7 @@ export const MODELS = {
449
449
  name: "Claude Sonnet 4 (EU)",
450
450
  api: "bedrock-converse-stream",
451
451
  provider: "amazon-bedrock",
452
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
452
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
453
453
  reasoning: true,
454
454
  input: ["text", "image"],
455
455
  cost: {
@@ -466,7 +466,7 @@ export const MODELS = {
466
466
  name: "Claude Sonnet 4.5 (EU)",
467
467
  api: "bedrock-converse-stream",
468
468
  provider: "amazon-bedrock",
469
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
469
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
470
470
  reasoning: true,
471
471
  input: ["text", "image"],
472
472
  cost: {
@@ -483,7 +483,7 @@ export const MODELS = {
483
483
  name: "Claude Sonnet 4.6 (EU)",
484
484
  api: "bedrock-converse-stream",
485
485
  provider: "amazon-bedrock",
486
- baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
486
+ baseUrl: "https://bedrock-runtime.eu-central-1.amazonaws.com",
487
487
  reasoning: true,
488
488
  input: ["text", "image"],
489
489
  cost: {
@@ -2695,6 +2695,314 @@ export const MODELS = {
2695
2695
  maxTokens: 40000,
2696
2696
  },
2697
2697
  },
2698
+ "fireworks": {
2699
+ "accounts/fireworks/models/deepseek-v3p1": {
2700
+ id: "accounts/fireworks/models/deepseek-v3p1",
2701
+ name: "DeepSeek V3.1",
2702
+ api: "anthropic-messages",
2703
+ provider: "fireworks",
2704
+ baseUrl: "https://api.fireworks.ai/inference",
2705
+ reasoning: true,
2706
+ input: ["text"],
2707
+ cost: {
2708
+ input: 0.56,
2709
+ output: 1.68,
2710
+ cacheRead: 0,
2711
+ cacheWrite: 0,
2712
+ },
2713
+ contextWindow: 163840,
2714
+ maxTokens: 163840,
2715
+ },
2716
+ "accounts/fireworks/models/deepseek-v3p2": {
2717
+ id: "accounts/fireworks/models/deepseek-v3p2",
2718
+ name: "DeepSeek V3.2",
2719
+ api: "anthropic-messages",
2720
+ provider: "fireworks",
2721
+ baseUrl: "https://api.fireworks.ai/inference",
2722
+ reasoning: true,
2723
+ input: ["text"],
2724
+ cost: {
2725
+ input: 0.56,
2726
+ output: 1.68,
2727
+ cacheRead: 0.28,
2728
+ cacheWrite: 0,
2729
+ },
2730
+ contextWindow: 160000,
2731
+ maxTokens: 160000,
2732
+ },
2733
+ "accounts/fireworks/models/glm-4p5": {
2734
+ id: "accounts/fireworks/models/glm-4p5",
2735
+ name: "GLM 4.5",
2736
+ api: "anthropic-messages",
2737
+ provider: "fireworks",
2738
+ baseUrl: "https://api.fireworks.ai/inference",
2739
+ reasoning: true,
2740
+ input: ["text"],
2741
+ cost: {
2742
+ input: 0.55,
2743
+ output: 2.19,
2744
+ cacheRead: 0,
2745
+ cacheWrite: 0,
2746
+ },
2747
+ contextWindow: 131072,
2748
+ maxTokens: 131072,
2749
+ },
2750
+ "accounts/fireworks/models/glm-4p5-air": {
2751
+ id: "accounts/fireworks/models/glm-4p5-air",
2752
+ name: "GLM 4.5 Air",
2753
+ api: "anthropic-messages",
2754
+ provider: "fireworks",
2755
+ baseUrl: "https://api.fireworks.ai/inference",
2756
+ reasoning: true,
2757
+ input: ["text"],
2758
+ cost: {
2759
+ input: 0.22,
2760
+ output: 0.88,
2761
+ cacheRead: 0,
2762
+ cacheWrite: 0,
2763
+ },
2764
+ contextWindow: 131072,
2765
+ maxTokens: 131072,
2766
+ },
2767
+ "accounts/fireworks/models/glm-4p7": {
2768
+ id: "accounts/fireworks/models/glm-4p7",
2769
+ name: "GLM 4.7",
2770
+ api: "anthropic-messages",
2771
+ provider: "fireworks",
2772
+ baseUrl: "https://api.fireworks.ai/inference",
2773
+ reasoning: true,
2774
+ input: ["text"],
2775
+ cost: {
2776
+ input: 0.6,
2777
+ output: 2.2,
2778
+ cacheRead: 0.3,
2779
+ cacheWrite: 0,
2780
+ },
2781
+ contextWindow: 198000,
2782
+ maxTokens: 198000,
2783
+ },
2784
+ "accounts/fireworks/models/glm-5": {
2785
+ id: "accounts/fireworks/models/glm-5",
2786
+ name: "GLM 5",
2787
+ api: "anthropic-messages",
2788
+ provider: "fireworks",
2789
+ baseUrl: "https://api.fireworks.ai/inference",
2790
+ reasoning: true,
2791
+ input: ["text"],
2792
+ cost: {
2793
+ input: 1,
2794
+ output: 3.2,
2795
+ cacheRead: 0.5,
2796
+ cacheWrite: 0,
2797
+ },
2798
+ contextWindow: 202752,
2799
+ maxTokens: 131072,
2800
+ },
2801
+ "accounts/fireworks/models/glm-5p1": {
2802
+ id: "accounts/fireworks/models/glm-5p1",
2803
+ name: "GLM 5.1",
2804
+ api: "anthropic-messages",
2805
+ provider: "fireworks",
2806
+ baseUrl: "https://api.fireworks.ai/inference",
2807
+ reasoning: true,
2808
+ input: ["text"],
2809
+ cost: {
2810
+ input: 1.4,
2811
+ output: 4.4,
2812
+ cacheRead: 0.26,
2813
+ cacheWrite: 0,
2814
+ },
2815
+ contextWindow: 202800,
2816
+ maxTokens: 131072,
2817
+ },
2818
+ "accounts/fireworks/models/gpt-oss-120b": {
2819
+ id: "accounts/fireworks/models/gpt-oss-120b",
2820
+ name: "GPT OSS 120B",
2821
+ api: "anthropic-messages",
2822
+ provider: "fireworks",
2823
+ baseUrl: "https://api.fireworks.ai/inference",
2824
+ reasoning: true,
2825
+ input: ["text"],
2826
+ cost: {
2827
+ input: 0.15,
2828
+ output: 0.6,
2829
+ cacheRead: 0,
2830
+ cacheWrite: 0,
2831
+ },
2832
+ contextWindow: 131072,
2833
+ maxTokens: 32768,
2834
+ },
2835
+ "accounts/fireworks/models/gpt-oss-20b": {
2836
+ id: "accounts/fireworks/models/gpt-oss-20b",
2837
+ name: "GPT OSS 20B",
2838
+ api: "anthropic-messages",
2839
+ provider: "fireworks",
2840
+ baseUrl: "https://api.fireworks.ai/inference",
2841
+ reasoning: true,
2842
+ input: ["text"],
2843
+ cost: {
2844
+ input: 0.05,
2845
+ output: 0.2,
2846
+ cacheRead: 0,
2847
+ cacheWrite: 0,
2848
+ },
2849
+ contextWindow: 131072,
2850
+ maxTokens: 32768,
2851
+ },
2852
+ "accounts/fireworks/models/kimi-k2-instruct": {
2853
+ id: "accounts/fireworks/models/kimi-k2-instruct",
2854
+ name: "Kimi K2 Instruct",
2855
+ api: "anthropic-messages",
2856
+ provider: "fireworks",
2857
+ baseUrl: "https://api.fireworks.ai/inference",
2858
+ reasoning: false,
2859
+ input: ["text"],
2860
+ cost: {
2861
+ input: 1,
2862
+ output: 3,
2863
+ cacheRead: 0,
2864
+ cacheWrite: 0,
2865
+ },
2866
+ contextWindow: 128000,
2867
+ maxTokens: 16384,
2868
+ },
2869
+ "accounts/fireworks/models/kimi-k2-thinking": {
2870
+ id: "accounts/fireworks/models/kimi-k2-thinking",
2871
+ name: "Kimi K2 Thinking",
2872
+ api: "anthropic-messages",
2873
+ provider: "fireworks",
2874
+ baseUrl: "https://api.fireworks.ai/inference",
2875
+ reasoning: true,
2876
+ input: ["text"],
2877
+ cost: {
2878
+ input: 0.6,
2879
+ output: 2.5,
2880
+ cacheRead: 0.3,
2881
+ cacheWrite: 0,
2882
+ },
2883
+ contextWindow: 256000,
2884
+ maxTokens: 256000,
2885
+ },
2886
+ "accounts/fireworks/models/kimi-k2p5": {
2887
+ id: "accounts/fireworks/models/kimi-k2p5",
2888
+ name: "Kimi K2.5",
2889
+ api: "anthropic-messages",
2890
+ provider: "fireworks",
2891
+ baseUrl: "https://api.fireworks.ai/inference",
2892
+ reasoning: true,
2893
+ input: ["text", "image"],
2894
+ cost: {
2895
+ input: 0.6,
2896
+ output: 3,
2897
+ cacheRead: 0.1,
2898
+ cacheWrite: 0,
2899
+ },
2900
+ contextWindow: 256000,
2901
+ maxTokens: 256000,
2902
+ },
2903
+ "accounts/fireworks/models/kimi-k2p6": {
2904
+ id: "accounts/fireworks/models/kimi-k2p6",
2905
+ name: "Kimi K2.6",
2906
+ api: "anthropic-messages",
2907
+ provider: "fireworks",
2908
+ baseUrl: "https://api.fireworks.ai/inference",
2909
+ reasoning: true,
2910
+ input: ["text", "image"],
2911
+ cost: {
2912
+ input: 0.95,
2913
+ output: 4,
2914
+ cacheRead: 0.16,
2915
+ cacheWrite: 0,
2916
+ },
2917
+ contextWindow: 262000,
2918
+ maxTokens: 262000,
2919
+ },
2920
+ "accounts/fireworks/models/minimax-m2p1": {
2921
+ id: "accounts/fireworks/models/minimax-m2p1",
2922
+ name: "MiniMax-M2.1",
2923
+ api: "anthropic-messages",
2924
+ provider: "fireworks",
2925
+ baseUrl: "https://api.fireworks.ai/inference",
2926
+ reasoning: true,
2927
+ input: ["text"],
2928
+ cost: {
2929
+ input: 0.3,
2930
+ output: 1.2,
2931
+ cacheRead: 0.03,
2932
+ cacheWrite: 0,
2933
+ },
2934
+ contextWindow: 200000,
2935
+ maxTokens: 200000,
2936
+ },
2937
+ "accounts/fireworks/models/minimax-m2p5": {
2938
+ id: "accounts/fireworks/models/minimax-m2p5",
2939
+ name: "MiniMax-M2.5",
2940
+ api: "anthropic-messages",
2941
+ provider: "fireworks",
2942
+ baseUrl: "https://api.fireworks.ai/inference",
2943
+ reasoning: true,
2944
+ input: ["text"],
2945
+ cost: {
2946
+ input: 0.3,
2947
+ output: 1.2,
2948
+ cacheRead: 0.03,
2949
+ cacheWrite: 0,
2950
+ },
2951
+ contextWindow: 196608,
2952
+ maxTokens: 196608,
2953
+ },
2954
+ "accounts/fireworks/models/minimax-m2p7": {
2955
+ id: "accounts/fireworks/models/minimax-m2p7",
2956
+ name: "MiniMax-M2.7",
2957
+ api: "anthropic-messages",
2958
+ provider: "fireworks",
2959
+ baseUrl: "https://api.fireworks.ai/inference",
2960
+ reasoning: true,
2961
+ input: ["text"],
2962
+ cost: {
2963
+ input: 0.3,
2964
+ output: 1.2,
2965
+ cacheRead: 0.03,
2966
+ cacheWrite: 0,
2967
+ },
2968
+ contextWindow: 196608,
2969
+ maxTokens: 196608,
2970
+ },
2971
+ "accounts/fireworks/models/qwen3p6-plus": {
2972
+ id: "accounts/fireworks/models/qwen3p6-plus",
2973
+ name: "Qwen 3.6 Plus",
2974
+ api: "anthropic-messages",
2975
+ provider: "fireworks",
2976
+ baseUrl: "https://api.fireworks.ai/inference",
2977
+ reasoning: true,
2978
+ input: ["text", "image"],
2979
+ cost: {
2980
+ input: 0.5,
2981
+ output: 3,
2982
+ cacheRead: 0.1,
2983
+ cacheWrite: 0,
2984
+ },
2985
+ contextWindow: 128000,
2986
+ maxTokens: 8192,
2987
+ },
2988
+ "accounts/fireworks/routers/kimi-k2p5-turbo": {
2989
+ id: "accounts/fireworks/routers/kimi-k2p5-turbo",
2990
+ name: "Kimi K2.5 Turbo (firepass)",
2991
+ api: "anthropic-messages",
2992
+ provider: "fireworks",
2993
+ baseUrl: "https://api.fireworks.ai/inference",
2994
+ reasoning: true,
2995
+ input: ["text", "image"],
2996
+ cost: {
2997
+ input: 0,
2998
+ output: 0,
2999
+ cacheRead: 0,
3000
+ cacheWrite: 0,
3001
+ },
3002
+ contextWindow: 256000,
3003
+ maxTokens: 256000,
3004
+ },
3005
+ },
2698
3006
  "github-copilot": {
2699
3007
  "claude-haiku-4.5": {
2700
3008
  id: "claude-haiku-4.5",
@@ -3251,7 +3559,7 @@ export const MODELS = {
3251
3559
  cost: {
3252
3560
  input: 0.3,
3253
3561
  output: 2.5,
3254
- cacheRead: 0.075,
3562
+ cacheRead: 0.03,
3255
3563
  cacheWrite: 0,
3256
3564
  },
3257
3565
  contextWindow: 1048576,
@@ -3370,7 +3678,7 @@ export const MODELS = {
3370
3678
  cost: {
3371
3679
  input: 1.25,
3372
3680
  output: 10,
3373
- cacheRead: 0.31,
3681
+ cacheRead: 0.125,
3374
3682
  cacheWrite: 0,
3375
3683
  },
3376
3684
  contextWindow: 1048576,
@@ -4694,6 +5002,24 @@ export const MODELS = {
4694
5002
  contextWindow: 262144,
4695
5003
  maxTokens: 262144,
4696
5004
  },
5005
+ "moonshotai/Kimi-K2.6": {
5006
+ id: "moonshotai/Kimi-K2.6",
5007
+ name: "Kimi-K2.6",
5008
+ api: "openai-completions",
5009
+ provider: "huggingface",
5010
+ baseUrl: "https://router.huggingface.co/v1",
5011
+ compat: { "supportsDeveloperRole": false },
5012
+ reasoning: true,
5013
+ input: ["text", "image"],
5014
+ cost: {
5015
+ input: 0.95,
5016
+ output: 4,
5017
+ cacheRead: 0.16,
5018
+ cacheWrite: 0,
5019
+ },
5020
+ contextWindow: 262144,
5021
+ maxTokens: 262144,
5022
+ },
4697
5023
  "zai-org/GLM-4.7": {
4698
5024
  id: "zai-org/GLM-4.7",
4699
5025
  name: "GLM-4.7",
@@ -4768,6 +5094,23 @@ export const MODELS = {
4768
5094
  },
4769
5095
  },
4770
5096
  "kimi-coding": {
5097
+ "k2p6": {
5098
+ id: "k2p6",
5099
+ name: "Kimi K2.6",
5100
+ api: "anthropic-messages",
5101
+ provider: "kimi-coding",
5102
+ baseUrl: "https://api.kimi.com/coding",
5103
+ reasoning: true,
5104
+ input: ["text", "image"],
5105
+ cost: {
5106
+ input: 0,
5107
+ output: 0,
5108
+ cacheRead: 0,
5109
+ cacheWrite: 0,
5110
+ },
5111
+ contextWindow: 262144,
5112
+ maxTokens: 32768,
5113
+ },
4771
5114
  "kimi-for-coding": {
4772
5115
  id: "kimi-for-coding",
4773
5116
  name: "Kimi For Coding",
@@ -6650,6 +6993,40 @@ export const MODELS = {
6650
6993
  contextWindow: 262144,
6651
6994
  maxTokens: 65536,
6652
6995
  },
6996
+ "kimi-k2.6": {
6997
+ id: "kimi-k2.6",
6998
+ name: "Kimi K2.6",
6999
+ api: "openai-completions",
7000
+ provider: "opencode",
7001
+ baseUrl: "https://opencode.ai/zen/v1",
7002
+ reasoning: true,
7003
+ input: ["text", "image"],
7004
+ cost: {
7005
+ input: 0.95,
7006
+ output: 4,
7007
+ cacheRead: 0.16,
7008
+ cacheWrite: 0,
7009
+ },
7010
+ contextWindow: 262144,
7011
+ maxTokens: 65536,
7012
+ },
7013
+ "ling-2.6-flash-free": {
7014
+ id: "ling-2.6-flash-free",
7015
+ name: "Ling 2.6 Flash Free",
7016
+ api: "openai-completions",
7017
+ provider: "opencode",
7018
+ baseUrl: "https://opencode.ai/zen/v1",
7019
+ reasoning: false,
7020
+ input: ["text"],
7021
+ cost: {
7022
+ input: 0,
7023
+ output: 0,
7024
+ cacheRead: 0,
7025
+ cacheWrite: 0,
7026
+ },
7027
+ contextWindow: 262100,
7028
+ maxTokens: 32800,
7029
+ },
6653
7030
  "minimax-m2.5": {
6654
7031
  id: "minimax-m2.5",
6655
7032
  name: "MiniMax M2.5",
@@ -6684,6 +7061,23 @@ export const MODELS = {
6684
7061
  contextWindow: 204800,
6685
7062
  maxTokens: 131072,
6686
7063
  },
7064
+ "minimax-m2.7": {
7065
+ id: "minimax-m2.7",
7066
+ name: "MiniMax M2.7",
7067
+ api: "openai-completions",
7068
+ provider: "opencode",
7069
+ baseUrl: "https://opencode.ai/zen/v1",
7070
+ reasoning: true,
7071
+ input: ["text"],
7072
+ cost: {
7073
+ input: 0.3,
7074
+ output: 1.2,
7075
+ cacheRead: 0.06,
7076
+ cacheWrite: 0,
7077
+ },
7078
+ contextWindow: 204800,
7079
+ maxTokens: 131072,
7080
+ },
6687
7081
  "nemotron-3-super-free": {
6688
7082
  id: "nemotron-3-super-free",
6689
7083
  name: "Nemotron 3 Super Free",
@@ -6707,6 +7101,7 @@ export const MODELS = {
6707
7101
  api: "openai-completions",
6708
7102
  provider: "opencode",
6709
7103
  baseUrl: "https://opencode.ai/zen/v1",
7104
+ compat: { "cacheControlFormat": "anthropic" },
6710
7105
  reasoning: true,
6711
7106
  input: ["text", "image"],
6712
7107
  cost: {
@@ -6724,6 +7119,7 @@ export const MODELS = {
6724
7119
  api: "openai-completions",
6725
7120
  provider: "opencode",
6726
7121
  baseUrl: "https://opencode.ai/zen/v1",
7122
+ compat: { "cacheControlFormat": "anthropic" },
6727
7123
  reasoning: true,
6728
7124
  input: ["text", "image"],
6729
7125
  cost: {
@@ -6788,6 +7184,23 @@ export const MODELS = {
6788
7184
  contextWindow: 262144,
6789
7185
  maxTokens: 65536,
6790
7186
  },
7187
+ "kimi-k2.6": {
7188
+ id: "kimi-k2.6",
7189
+ name: "Kimi K2.6 (3x limits)",
7190
+ api: "openai-completions",
7191
+ provider: "opencode-go",
7192
+ baseUrl: "https://opencode.ai/zen/go/v1",
7193
+ reasoning: true,
7194
+ input: ["text", "image"],
7195
+ cost: {
7196
+ input: 0.32,
7197
+ output: 1.34,
7198
+ cacheRead: 0.054,
7199
+ cacheWrite: 0,
7200
+ },
7201
+ contextWindow: 262144,
7202
+ maxTokens: 65536,
7203
+ },
6791
7204
  "mimo-v2-omni": {
6792
7205
  id: "mimo-v2-omni",
6793
7206
  name: "MiMo V2 Omni",
@@ -6862,6 +7275,7 @@ export const MODELS = {
6862
7275
  api: "openai-completions",
6863
7276
  provider: "opencode-go",
6864
7277
  baseUrl: "https://opencode.ai/zen/go/v1",
7278
+ compat: { "cacheControlFormat": "anthropic" },
6865
7279
  reasoning: true,
6866
7280
  input: ["text", "image"],
6867
7281
  cost: {
@@ -6879,6 +7293,7 @@ export const MODELS = {
6879
7293
  api: "openai-completions",
6880
7294
  provider: "opencode-go",
6881
7295
  baseUrl: "https://opencode.ai/zen/go/v1",
7296
+ compat: { "cacheControlFormat": "anthropic" },
6882
7297
  reasoning: true,
6883
7298
  input: ["text", "image"],
6884
7299
  cost: {
@@ -7598,13 +8013,13 @@ export const MODELS = {
7598
8013
  reasoning: true,
7599
8014
  input: ["text"],
7600
8015
  cost: {
7601
- input: 0.25899999999999995,
7602
- output: 0.42,
7603
- cacheRead: 0.135,
8016
+ input: 0.252,
8017
+ output: 0.378,
8018
+ cacheRead: 0.0252,
7604
8019
  cacheWrite: 0,
7605
8020
  },
7606
- contextWindow: 163840,
7607
- maxTokens: 163840,
8021
+ contextWindow: 131072,
8022
+ maxTokens: 32768,
7608
8023
  },
7609
8024
  "deepseek/deepseek-v3.2-exp": {
7610
8025
  id: "deepseek/deepseek-v3.2-exp",
@@ -7853,13 +8268,13 @@ export const MODELS = {
7853
8268
  reasoning: true,
7854
8269
  input: ["text", "image"],
7855
8270
  cost: {
7856
- input: 0.08,
8271
+ input: 0.07,
7857
8272
  output: 0.35,
7858
- cacheRead: 0.01,
8273
+ cacheRead: 0.04,
7859
8274
  cacheWrite: 0,
7860
8275
  },
7861
8276
  contextWindow: 262144,
7862
- maxTokens: 4096,
8277
+ maxTokens: 262144,
7863
8278
  },
7864
8279
  "google/gemma-4-26b-a4b-it:free": {
7865
8280
  id: "google/gemma-4-26b-a4b-it:free",
@@ -7929,6 +8344,23 @@ export const MODELS = {
7929
8344
  contextWindow: 128000,
7930
8345
  maxTokens: 50000,
7931
8346
  },
8347
+ "inclusionai/ling-2.6-flash:free": {
8348
+ id: "inclusionai/ling-2.6-flash:free",
8349
+ name: "inclusionAI: Ling-2.6-flash (free)",
8350
+ api: "openai-completions",
8351
+ provider: "openrouter",
8352
+ baseUrl: "https://openrouter.ai/api/v1",
8353
+ reasoning: false,
8354
+ input: ["text"],
8355
+ cost: {
8356
+ input: 0,
8357
+ output: 0,
8358
+ cacheRead: 0,
8359
+ cacheWrite: 0,
8360
+ },
8361
+ contextWindow: 262144,
8362
+ maxTokens: 32768,
8363
+ },
7932
8364
  "kwaipilot/kat-coder-pro-v2": {
7933
8365
  id: "kwaipilot/kat-coder-pro-v2",
7934
8366
  name: "Kwaipilot: KAT-Coder-Pro V2",
@@ -8031,23 +8463,6 @@ export const MODELS = {
8031
8463
  contextWindow: 65536,
8032
8464
  maxTokens: 4096,
8033
8465
  },
8034
- "meta-llama/llama-4-maverick": {
8035
- id: "meta-llama/llama-4-maverick",
8036
- name: "Meta: Llama 4 Maverick",
8037
- api: "openai-completions",
8038
- provider: "openrouter",
8039
- baseUrl: "https://openrouter.ai/api/v1",
8040
- reasoning: false,
8041
- input: ["text", "image"],
8042
- cost: {
8043
- input: 0.15,
8044
- output: 0.6,
8045
- cacheRead: 0,
8046
- cacheWrite: 0,
8047
- },
8048
- contextWindow: 1048576,
8049
- maxTokens: 16384,
8050
- },
8051
8466
  "meta-llama/llama-4-scout": {
8052
8467
  id: "meta-llama/llama-4-scout",
8053
8468
  name: "Meta: Llama 4 Scout",
@@ -8125,9 +8540,9 @@ export const MODELS = {
8125
8540
  reasoning: true,
8126
8541
  input: ["text"],
8127
8542
  cost: {
8128
- input: 0.118,
8129
- output: 0.9900000000000001,
8130
- cacheRead: 0.059,
8543
+ input: 0.15,
8544
+ output: 1.2,
8545
+ cacheRead: 0.075,
8131
8546
  cacheWrite: 0,
8132
8547
  },
8133
8548
  contextWindow: 196608,
@@ -8606,9 +9021,26 @@ export const MODELS = {
8606
9021
  cacheRead: 0.07,
8607
9022
  cacheWrite: 0,
8608
9023
  },
8609
- contextWindow: 256000,
9024
+ contextWindow: 262144,
8610
9025
  maxTokens: 4096,
8611
9026
  },
9027
+ "moonshotai/kimi-k2.6": {
9028
+ id: "moonshotai/kimi-k2.6",
9029
+ name: "MoonshotAI: Kimi K2.6",
9030
+ api: "openai-completions",
9031
+ provider: "openrouter",
9032
+ baseUrl: "https://openrouter.ai/api/v1",
9033
+ reasoning: true,
9034
+ input: ["text", "image"],
9035
+ cost: {
9036
+ input: 0.7999999999999999,
9037
+ output: 3.5,
9038
+ cacheRead: 0.19999999999999998,
9039
+ cacheWrite: 0,
9040
+ },
9041
+ contextWindow: 262144,
9042
+ maxTokens: 262144,
9043
+ },
8612
9044
  "nex-agi/deepseek-v3.1-nex-n1": {
8613
9045
  id: "nex-agi/deepseek-v3.1-nex-n1",
8614
9046
  name: "Nex AGI: DeepSeek V3.1 Nex N1",
@@ -9085,23 +9517,6 @@ export const MODELS = {
9085
9517
  contextWindow: 128000,
9086
9518
  maxTokens: 16384,
9087
9519
  },
9088
- "openai/gpt-4o:extended": {
9089
- id: "openai/gpt-4o:extended",
9090
- name: "OpenAI: GPT-4o (extended)",
9091
- api: "openai-completions",
9092
- provider: "openrouter",
9093
- baseUrl: "https://openrouter.ai/api/v1",
9094
- reasoning: false,
9095
- input: ["text", "image"],
9096
- cost: {
9097
- input: 6,
9098
- output: 18,
9099
- cacheRead: 0,
9100
- cacheWrite: 0,
9101
- },
9102
- contextWindow: 128000,
9103
- maxTokens: 64000,
9104
- },
9105
9520
  "openai/gpt-5": {
9106
9521
  id: "openai/gpt-5",
9107
9522
  name: "OpenAI: GPT-5",
@@ -9136,40 +9551,6 @@ export const MODELS = {
9136
9551
  contextWindow: 400000,
9137
9552
  maxTokens: 128000,
9138
9553
  },
9139
- "openai/gpt-5-image": {
9140
- id: "openai/gpt-5-image",
9141
- name: "OpenAI: GPT-5 Image",
9142
- api: "openai-completions",
9143
- provider: "openrouter",
9144
- baseUrl: "https://openrouter.ai/api/v1",
9145
- reasoning: true,
9146
- input: ["text", "image"],
9147
- cost: {
9148
- input: 10,
9149
- output: 10,
9150
- cacheRead: 1.25,
9151
- cacheWrite: 0,
9152
- },
9153
- contextWindow: 400000,
9154
- maxTokens: 128000,
9155
- },
9156
- "openai/gpt-5-image-mini": {
9157
- id: "openai/gpt-5-image-mini",
9158
- name: "OpenAI: GPT-5 Image Mini",
9159
- api: "openai-completions",
9160
- provider: "openrouter",
9161
- baseUrl: "https://openrouter.ai/api/v1",
9162
- reasoning: true,
9163
- input: ["text", "image"],
9164
- cost: {
9165
- input: 2.5,
9166
- output: 2,
9167
- cacheRead: 0.25,
9168
- cacheWrite: 0,
9169
- },
9170
- contextWindow: 400000,
9171
- maxTokens: 128000,
9172
- },
9173
9554
  "openai/gpt-5-mini": {
9174
9555
  id: "openai/gpt-5-mini",
9175
9556
  name: "OpenAI: GPT-5 Mini",
@@ -9765,23 +10146,6 @@ export const MODELS = {
9765
10146
  contextWindow: 2000000,
9766
10147
  maxTokens: 4096,
9767
10148
  },
9768
- "openrouter/elephant-alpha": {
9769
- id: "openrouter/elephant-alpha",
9770
- name: "Elephant",
9771
- api: "openai-completions",
9772
- provider: "openrouter",
9773
- baseUrl: "https://openrouter.ai/api/v1",
9774
- reasoning: false,
9775
- input: ["text"],
9776
- cost: {
9777
- input: 0,
9778
- output: 0,
9779
- cacheRead: 0,
9780
- cacheWrite: 0,
9781
- },
9782
- contextWindow: 262144,
9783
- maxTokens: 32768,
9784
- },
9785
10149
  "openrouter/free": {
9786
10150
  id: "openrouter/free",
9787
10151
  name: "Free Models Router",
@@ -9992,7 +10356,7 @@ export const MODELS = {
9992
10356
  api: "openai-completions",
9993
10357
  provider: "openrouter",
9994
10358
  baseUrl: "https://openrouter.ai/api/v1",
9995
- reasoning: true,
10359
+ reasoning: false,
9996
10360
  input: ["text"],
9997
10361
  cost: {
9998
10362
  input: 0.071,
@@ -10167,7 +10531,7 @@ export const MODELS = {
10167
10531
  cost: {
10168
10532
  input: 0.15,
10169
10533
  output: 0.7999999999999999,
10170
- cacheRead: 0.12,
10534
+ cacheRead: 0.11,
10171
10535
  cacheWrite: 0,
10172
10536
  },
10173
10537
  contextWindow: 262144,
@@ -11049,13 +11413,13 @@ export const MODELS = {
11049
11413
  reasoning: true,
11050
11414
  input: ["text"],
11051
11415
  cost: {
11052
- input: 0.39,
11053
- output: 1.75,
11054
- cacheRead: 0.195,
11416
+ input: 0.38,
11417
+ output: 1.74,
11418
+ cacheRead: 0,
11055
11419
  cacheWrite: 0,
11056
11420
  },
11057
11421
  contextWindow: 202752,
11058
- maxTokens: 65535,
11422
+ maxTokens: 4096,
11059
11423
  },
11060
11424
  "z-ai/glm-4.7-flash": {
11061
11425
  id: "z-ai/glm-4.7-flash",
@@ -11088,8 +11452,8 @@ export const MODELS = {
11088
11452
  cacheRead: 0.119,
11089
11453
  cacheWrite: 0,
11090
11454
  },
11091
- contextWindow: 80000,
11092
- maxTokens: 131072,
11455
+ contextWindow: 202752,
11456
+ maxTokens: 4096,
11093
11457
  },
11094
11458
  "z-ai/glm-5-turbo": {
11095
11459
  id: "z-ai/glm-5-turbo",
@@ -11117,9 +11481,9 @@ export const MODELS = {
11117
11481
  reasoning: true,
11118
11482
  input: ["text"],
11119
11483
  cost: {
11120
- input: 0.95,
11121
- output: 3.15,
11122
- cacheRead: 0.475,
11484
+ input: 1.0499999999999998,
11485
+ output: 3.5,
11486
+ cacheRead: 0.5249999999999999,
11123
11487
  cacheWrite: 0,
11124
11488
  },
11125
11489
  contextWindow: 202752,
@@ -11142,6 +11506,23 @@ export const MODELS = {
11142
11506
  contextWindow: 202752,
11143
11507
  maxTokens: 131072,
11144
11508
  },
11509
+ "~anthropic/claude-opus-latest": {
11510
+ id: "~anthropic/claude-opus-latest",
11511
+ name: "Anthropic: Claude Opus Latest",
11512
+ api: "openai-completions",
11513
+ provider: "openrouter",
11514
+ baseUrl: "https://openrouter.ai/api/v1",
11515
+ reasoning: true,
11516
+ input: ["text", "image"],
11517
+ cost: {
11518
+ input: 5,
11519
+ output: 25,
11520
+ cacheRead: 0.5,
11521
+ cacheWrite: 6.25,
11522
+ },
11523
+ contextWindow: 1000000,
11524
+ maxTokens: 128000,
11525
+ },
11145
11526
  },
11146
11527
  "vercel-ai-gateway": {
11147
11528
  "alibaba/qwen-3-14b": {
@@ -11212,6 +11593,23 @@ export const MODELS = {
11212
11593
  contextWindow: 128000,
11213
11594
  maxTokens: 8192,
11214
11595
  },
11596
+ "alibaba/qwen-3.6-max-preview": {
11597
+ id: "alibaba/qwen-3.6-max-preview",
11598
+ name: "Qwen 3.6 Max Preview",
11599
+ api: "anthropic-messages",
11600
+ provider: "vercel-ai-gateway",
11601
+ baseUrl: "https://ai-gateway.vercel.sh",
11602
+ reasoning: true,
11603
+ input: ["text", "image"],
11604
+ cost: {
11605
+ input: 1.3,
11606
+ output: 7.8,
11607
+ cacheRead: 0.26,
11608
+ cacheWrite: 1.625,
11609
+ },
11610
+ contextWindow: 240000,
11611
+ maxTokens: 64000,
11612
+ },
11215
11613
  "alibaba/qwen3-235b-a22b-thinking": {
11216
11614
  id: "alibaba/qwen3-235b-a22b-thinking",
11217
11615
  name: "Qwen3 235B A22B Thinking 2507",
@@ -12555,6 +12953,23 @@ export const MODELS = {
12555
12953
  contextWindow: 262114,
12556
12954
  maxTokens: 262114,
12557
12955
  },
12956
+ "moonshotai/kimi-k2.6": {
12957
+ id: "moonshotai/kimi-k2.6",
12958
+ name: "Kimi K2.6",
12959
+ api: "anthropic-messages",
12960
+ provider: "vercel-ai-gateway",
12961
+ baseUrl: "https://ai-gateway.vercel.sh",
12962
+ reasoning: true,
12963
+ input: ["text", "image"],
12964
+ cost: {
12965
+ input: 0.95,
12966
+ output: 4,
12967
+ cacheRead: 0.16,
12968
+ cacheWrite: 0,
12969
+ },
12970
+ contextWindow: 262000,
12971
+ maxTokens: 262000,
12972
+ },
12558
12973
  "nvidia/nemotron-nano-12b-v2-vl": {
12559
12974
  id: "nvidia/nemotron-nano-12b-v2-vl",
12560
12975
  name: "Nvidia Nemotron Nano 12B V2 VL",