@mariozechner/pi-ai 0.70.5 → 0.71.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (68) hide show
  1. package/README.md +6 -31
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +4 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +1 -1
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js.map +1 -1
  8. package/dist/models.d.ts +1 -1
  9. package/dist/models.d.ts.map +1 -1
  10. package/dist/models.generated.d.ts +1653 -596
  11. package/dist/models.generated.d.ts.map +1 -1
  12. package/dist/models.generated.js +1337 -422
  13. package/dist/models.generated.js.map +1 -1
  14. package/dist/models.js +3 -2
  15. package/dist/models.js.map +1 -1
  16. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  17. package/dist/providers/amazon-bedrock.js +14 -15
  18. package/dist/providers/amazon-bedrock.js.map +1 -1
  19. package/dist/providers/anthropic.d.ts.map +1 -1
  20. package/dist/providers/anthropic.js +38 -15
  21. package/dist/providers/anthropic.js.map +1 -1
  22. package/dist/providers/cloudflare.d.ts +13 -0
  23. package/dist/providers/cloudflare.d.ts.map +1 -0
  24. package/dist/providers/cloudflare.js +26 -0
  25. package/dist/providers/cloudflare.js.map +1 -0
  26. package/dist/providers/google-shared.d.ts +7 -2
  27. package/dist/providers/google-shared.d.ts.map +1 -1
  28. package/dist/providers/google-shared.js +4 -13
  29. package/dist/providers/google-shared.js.map +1 -1
  30. package/dist/providers/google-vertex.d.ts +1 -1
  31. package/dist/providers/google-vertex.d.ts.map +1 -1
  32. package/dist/providers/google-vertex.js.map +1 -1
  33. package/dist/providers/google.d.ts +1 -1
  34. package/dist/providers/google.d.ts.map +1 -1
  35. package/dist/providers/google.js.map +1 -1
  36. package/dist/providers/mistral.d.ts.map +1 -1
  37. package/dist/providers/mistral.js +1 -1
  38. package/dist/providers/mistral.js.map +1 -1
  39. package/dist/providers/openai-completions.d.ts.map +1 -1
  40. package/dist/providers/openai-completions.js +25 -8
  41. package/dist/providers/openai-completions.js.map +1 -1
  42. package/dist/providers/openai-responses.d.ts.map +1 -1
  43. package/dist/providers/openai-responses.js +10 -2
  44. package/dist/providers/openai-responses.js.map +1 -1
  45. package/dist/providers/register-builtins.d.ts +0 -3
  46. package/dist/providers/register-builtins.d.ts.map +1 -1
  47. package/dist/providers/register-builtins.js +0 -18
  48. package/dist/providers/register-builtins.js.map +1 -1
  49. package/dist/types.d.ts +3 -2
  50. package/dist/types.d.ts.map +1 -1
  51. package/dist/types.js.map +1 -1
  52. package/dist/utils/oauth/index.d.ts +0 -4
  53. package/dist/utils/oauth/index.d.ts.map +1 -1
  54. package/dist/utils/oauth/index.js +0 -10
  55. package/dist/utils/oauth/index.js.map +1 -1
  56. package/package.json +2 -6
  57. package/dist/providers/google-gemini-cli.d.ts +0 -74
  58. package/dist/providers/google-gemini-cli.d.ts.map +0 -1
  59. package/dist/providers/google-gemini-cli.js +0 -779
  60. package/dist/providers/google-gemini-cli.js.map +0 -1
  61. package/dist/utils/oauth/google-antigravity.d.ts +0 -26
  62. package/dist/utils/oauth/google-antigravity.d.ts.map +0 -1
  63. package/dist/utils/oauth/google-antigravity.js +0 -377
  64. package/dist/utils/oauth/google-antigravity.js.map +0 -1
  65. package/dist/utils/oauth/google-gemini-cli.d.ts +0 -26
  66. package/dist/utils/oauth/google-gemini-cli.d.ts.map +0 -1
  67. package/dist/utils/oauth/google-gemini-cli.js +0 -482
  68. package/dist/utils/oauth/google-gemini-cli.js.map +0 -1
@@ -2539,6 +2539,23 @@ export const MODELS = {
2539
2539
  contextWindow: 272000,
2540
2540
  maxTokens: 128000,
2541
2541
  },
2542
+ "gpt-5.5-pro": {
2543
+ id: "gpt-5.5-pro",
2544
+ name: "GPT-5.5 Pro",
2545
+ api: "azure-openai-responses",
2546
+ provider: "azure-openai-responses",
2547
+ baseUrl: "",
2548
+ reasoning: true,
2549
+ input: ["text", "image"],
2550
+ cost: {
2551
+ input: 30,
2552
+ output: 180,
2553
+ cacheRead: 0,
2554
+ cacheWrite: 0,
2555
+ },
2556
+ contextWindow: 1050000,
2557
+ maxTokens: 128000,
2558
+ },
2542
2559
  "o1": {
2543
2560
  id: "o1",
2544
2561
  name: "o1",
@@ -2715,35 +2732,782 @@ export const MODELS = {
2715
2732
  id: "qwen-3-235b-a22b-instruct-2507",
2716
2733
  name: "Qwen 3 235B Instruct",
2717
2734
  api: "openai-completions",
2718
- provider: "cerebras",
2719
- baseUrl: "https://api.cerebras.ai/v1",
2720
- reasoning: false,
2735
+ provider: "cerebras",
2736
+ baseUrl: "https://api.cerebras.ai/v1",
2737
+ reasoning: false,
2738
+ input: ["text"],
2739
+ cost: {
2740
+ input: 0.6,
2741
+ output: 1.2,
2742
+ cacheRead: 0,
2743
+ cacheWrite: 0,
2744
+ },
2745
+ contextWindow: 131000,
2746
+ maxTokens: 32000,
2747
+ },
2748
+ "zai-glm-4.7": {
2749
+ id: "zai-glm-4.7",
2750
+ name: "Z.AI GLM-4.7",
2751
+ api: "openai-completions",
2752
+ provider: "cerebras",
2753
+ baseUrl: "https://api.cerebras.ai/v1",
2754
+ reasoning: false,
2755
+ input: ["text"],
2756
+ cost: {
2757
+ input: 2.25,
2758
+ output: 2.75,
2759
+ cacheRead: 0,
2760
+ cacheWrite: 0,
2761
+ },
2762
+ contextWindow: 131072,
2763
+ maxTokens: 40000,
2764
+ },
2765
+ },
2766
+ "cloudflare-ai-gateway": {
2767
+ "claude-3-5-haiku": {
2768
+ id: "claude-3-5-haiku",
2769
+ name: "Claude Haiku 3.5 (latest)",
2770
+ api: "anthropic-messages",
2771
+ provider: "cloudflare-ai-gateway",
2772
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2773
+ reasoning: false,
2774
+ input: ["text", "image"],
2775
+ cost: {
2776
+ input: 0.8,
2777
+ output: 4,
2778
+ cacheRead: 0.08,
2779
+ cacheWrite: 1,
2780
+ },
2781
+ contextWindow: 200000,
2782
+ maxTokens: 8192,
2783
+ },
2784
+ "claude-3-haiku": {
2785
+ id: "claude-3-haiku",
2786
+ name: "Claude Haiku 3",
2787
+ api: "anthropic-messages",
2788
+ provider: "cloudflare-ai-gateway",
2789
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2790
+ reasoning: false,
2791
+ input: ["text", "image"],
2792
+ cost: {
2793
+ input: 0.25,
2794
+ output: 1.25,
2795
+ cacheRead: 0.03,
2796
+ cacheWrite: 0.3,
2797
+ },
2798
+ contextWindow: 200000,
2799
+ maxTokens: 4096,
2800
+ },
2801
+ "claude-3-opus": {
2802
+ id: "claude-3-opus",
2803
+ name: "Claude Opus 3",
2804
+ api: "anthropic-messages",
2805
+ provider: "cloudflare-ai-gateway",
2806
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2807
+ reasoning: false,
2808
+ input: ["text", "image"],
2809
+ cost: {
2810
+ input: 15,
2811
+ output: 75,
2812
+ cacheRead: 1.5,
2813
+ cacheWrite: 18.75,
2814
+ },
2815
+ contextWindow: 200000,
2816
+ maxTokens: 4096,
2817
+ },
2818
+ "claude-3-sonnet": {
2819
+ id: "claude-3-sonnet",
2820
+ name: "Claude Sonnet 3",
2821
+ api: "anthropic-messages",
2822
+ provider: "cloudflare-ai-gateway",
2823
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2824
+ reasoning: false,
2825
+ input: ["text", "image"],
2826
+ cost: {
2827
+ input: 3,
2828
+ output: 15,
2829
+ cacheRead: 0.3,
2830
+ cacheWrite: 0.3,
2831
+ },
2832
+ contextWindow: 200000,
2833
+ maxTokens: 4096,
2834
+ },
2835
+ "claude-3.5-haiku": {
2836
+ id: "claude-3.5-haiku",
2837
+ name: "Claude Haiku 3.5 (latest)",
2838
+ api: "anthropic-messages",
2839
+ provider: "cloudflare-ai-gateway",
2840
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2841
+ reasoning: false,
2842
+ input: ["text", "image"],
2843
+ cost: {
2844
+ input: 0.8,
2845
+ output: 4,
2846
+ cacheRead: 0.08,
2847
+ cacheWrite: 1,
2848
+ },
2849
+ contextWindow: 200000,
2850
+ maxTokens: 8192,
2851
+ },
2852
+ "claude-3.5-sonnet": {
2853
+ id: "claude-3.5-sonnet",
2854
+ name: "Claude Sonnet 3.5 v2",
2855
+ api: "anthropic-messages",
2856
+ provider: "cloudflare-ai-gateway",
2857
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2858
+ reasoning: false,
2859
+ input: ["text", "image"],
2860
+ cost: {
2861
+ input: 3,
2862
+ output: 15,
2863
+ cacheRead: 0.3,
2864
+ cacheWrite: 3.75,
2865
+ },
2866
+ contextWindow: 200000,
2867
+ maxTokens: 8192,
2868
+ },
2869
+ "claude-haiku-4-5": {
2870
+ id: "claude-haiku-4-5",
2871
+ name: "Claude Haiku 4.5 (latest)",
2872
+ api: "anthropic-messages",
2873
+ provider: "cloudflare-ai-gateway",
2874
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2875
+ reasoning: true,
2876
+ input: ["text", "image"],
2877
+ cost: {
2878
+ input: 1,
2879
+ output: 5,
2880
+ cacheRead: 0.1,
2881
+ cacheWrite: 1.25,
2882
+ },
2883
+ contextWindow: 200000,
2884
+ maxTokens: 64000,
2885
+ },
2886
+ "claude-opus-4": {
2887
+ id: "claude-opus-4",
2888
+ name: "Claude Opus 4 (latest)",
2889
+ api: "anthropic-messages",
2890
+ provider: "cloudflare-ai-gateway",
2891
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2892
+ reasoning: true,
2893
+ input: ["text", "image"],
2894
+ cost: {
2895
+ input: 15,
2896
+ output: 75,
2897
+ cacheRead: 1.5,
2898
+ cacheWrite: 18.75,
2899
+ },
2900
+ contextWindow: 200000,
2901
+ maxTokens: 32000,
2902
+ },
2903
+ "claude-opus-4-1": {
2904
+ id: "claude-opus-4-1",
2905
+ name: "Claude Opus 4.1 (latest)",
2906
+ api: "anthropic-messages",
2907
+ provider: "cloudflare-ai-gateway",
2908
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2909
+ reasoning: true,
2910
+ input: ["text", "image"],
2911
+ cost: {
2912
+ input: 15,
2913
+ output: 75,
2914
+ cacheRead: 1.5,
2915
+ cacheWrite: 18.75,
2916
+ },
2917
+ contextWindow: 200000,
2918
+ maxTokens: 32000,
2919
+ },
2920
+ "claude-opus-4-5": {
2921
+ id: "claude-opus-4-5",
2922
+ name: "Claude Opus 4.5 (latest)",
2923
+ api: "anthropic-messages",
2924
+ provider: "cloudflare-ai-gateway",
2925
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2926
+ reasoning: true,
2927
+ input: ["text", "image"],
2928
+ cost: {
2929
+ input: 5,
2930
+ output: 25,
2931
+ cacheRead: 0.5,
2932
+ cacheWrite: 6.25,
2933
+ },
2934
+ contextWindow: 200000,
2935
+ maxTokens: 64000,
2936
+ },
2937
+ "claude-opus-4-6": {
2938
+ id: "claude-opus-4-6",
2939
+ name: "Claude Opus 4.6 (latest)",
2940
+ api: "anthropic-messages",
2941
+ provider: "cloudflare-ai-gateway",
2942
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2943
+ reasoning: true,
2944
+ input: ["text", "image"],
2945
+ cost: {
2946
+ input: 5,
2947
+ output: 25,
2948
+ cacheRead: 0.5,
2949
+ cacheWrite: 6.25,
2950
+ },
2951
+ contextWindow: 1000000,
2952
+ maxTokens: 128000,
2953
+ },
2954
+ "claude-opus-4-7": {
2955
+ id: "claude-opus-4-7",
2956
+ name: "Claude Opus 4.7",
2957
+ api: "anthropic-messages",
2958
+ provider: "cloudflare-ai-gateway",
2959
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2960
+ reasoning: true,
2961
+ input: ["text", "image"],
2962
+ cost: {
2963
+ input: 5,
2964
+ output: 25,
2965
+ cacheRead: 0.5,
2966
+ cacheWrite: 6.25,
2967
+ },
2968
+ contextWindow: 1000000,
2969
+ maxTokens: 128000,
2970
+ },
2971
+ "claude-sonnet-4": {
2972
+ id: "claude-sonnet-4",
2973
+ name: "Claude Sonnet 4 (latest)",
2974
+ api: "anthropic-messages",
2975
+ provider: "cloudflare-ai-gateway",
2976
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2977
+ reasoning: true,
2978
+ input: ["text", "image"],
2979
+ cost: {
2980
+ input: 3,
2981
+ output: 15,
2982
+ cacheRead: 0.3,
2983
+ cacheWrite: 3.75,
2984
+ },
2985
+ contextWindow: 200000,
2986
+ maxTokens: 64000,
2987
+ },
2988
+ "claude-sonnet-4-5": {
2989
+ id: "claude-sonnet-4-5",
2990
+ name: "Claude Sonnet 4.5 (latest)",
2991
+ api: "anthropic-messages",
2992
+ provider: "cloudflare-ai-gateway",
2993
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
2994
+ reasoning: true,
2995
+ input: ["text", "image"],
2996
+ cost: {
2997
+ input: 3,
2998
+ output: 15,
2999
+ cacheRead: 0.3,
3000
+ cacheWrite: 3.75,
3001
+ },
3002
+ contextWindow: 200000,
3003
+ maxTokens: 64000,
3004
+ },
3005
+ "claude-sonnet-4-6": {
3006
+ id: "claude-sonnet-4-6",
3007
+ name: "Claude Sonnet 4.6",
3008
+ api: "anthropic-messages",
3009
+ provider: "cloudflare-ai-gateway",
3010
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/anthropic",
3011
+ reasoning: true,
3012
+ input: ["text", "image"],
3013
+ cost: {
3014
+ input: 3,
3015
+ output: 15,
3016
+ cacheRead: 0.3,
3017
+ cacheWrite: 3.75,
3018
+ },
3019
+ contextWindow: 1000000,
3020
+ maxTokens: 64000,
3021
+ },
3022
+ "gpt-4": {
3023
+ id: "gpt-4",
3024
+ name: "GPT-4",
3025
+ api: "openai-responses",
3026
+ provider: "cloudflare-ai-gateway",
3027
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3028
+ reasoning: false,
3029
+ input: ["text"],
3030
+ cost: {
3031
+ input: 30,
3032
+ output: 60,
3033
+ cacheRead: 0,
3034
+ cacheWrite: 0,
3035
+ },
3036
+ contextWindow: 8192,
3037
+ maxTokens: 8192,
3038
+ },
3039
+ "gpt-4-turbo": {
3040
+ id: "gpt-4-turbo",
3041
+ name: "GPT-4 Turbo",
3042
+ api: "openai-responses",
3043
+ provider: "cloudflare-ai-gateway",
3044
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3045
+ reasoning: false,
3046
+ input: ["text", "image"],
3047
+ cost: {
3048
+ input: 10,
3049
+ output: 30,
3050
+ cacheRead: 0,
3051
+ cacheWrite: 0,
3052
+ },
3053
+ contextWindow: 128000,
3054
+ maxTokens: 4096,
3055
+ },
3056
+ "gpt-4o": {
3057
+ id: "gpt-4o",
3058
+ name: "GPT-4o",
3059
+ api: "openai-responses",
3060
+ provider: "cloudflare-ai-gateway",
3061
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3062
+ reasoning: false,
3063
+ input: ["text", "image"],
3064
+ cost: {
3065
+ input: 2.5,
3066
+ output: 10,
3067
+ cacheRead: 1.25,
3068
+ cacheWrite: 0,
3069
+ },
3070
+ contextWindow: 128000,
3071
+ maxTokens: 16384,
3072
+ },
3073
+ "gpt-4o-mini": {
3074
+ id: "gpt-4o-mini",
3075
+ name: "GPT-4o mini",
3076
+ api: "openai-responses",
3077
+ provider: "cloudflare-ai-gateway",
3078
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3079
+ reasoning: false,
3080
+ input: ["text", "image"],
3081
+ cost: {
3082
+ input: 0.15,
3083
+ output: 0.6,
3084
+ cacheRead: 0.08,
3085
+ cacheWrite: 0,
3086
+ },
3087
+ contextWindow: 128000,
3088
+ maxTokens: 16384,
3089
+ },
3090
+ "gpt-5.1": {
3091
+ id: "gpt-5.1",
3092
+ name: "GPT-5.1",
3093
+ api: "openai-responses",
3094
+ provider: "cloudflare-ai-gateway",
3095
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3096
+ reasoning: true,
3097
+ input: ["text", "image"],
3098
+ cost: {
3099
+ input: 1.25,
3100
+ output: 10,
3101
+ cacheRead: 0.13,
3102
+ cacheWrite: 0,
3103
+ },
3104
+ contextWindow: 400000,
3105
+ maxTokens: 128000,
3106
+ },
3107
+ "gpt-5.1-codex": {
3108
+ id: "gpt-5.1-codex",
3109
+ name: "GPT-5.1 Codex",
3110
+ api: "openai-responses",
3111
+ provider: "cloudflare-ai-gateway",
3112
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3113
+ reasoning: true,
3114
+ input: ["text", "image"],
3115
+ cost: {
3116
+ input: 1.25,
3117
+ output: 10,
3118
+ cacheRead: 0.125,
3119
+ cacheWrite: 0,
3120
+ },
3121
+ contextWindow: 400000,
3122
+ maxTokens: 128000,
3123
+ },
3124
+ "gpt-5.2": {
3125
+ id: "gpt-5.2",
3126
+ name: "GPT-5.2",
3127
+ api: "openai-responses",
3128
+ provider: "cloudflare-ai-gateway",
3129
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3130
+ reasoning: true,
3131
+ input: ["text", "image"],
3132
+ cost: {
3133
+ input: 1.75,
3134
+ output: 14,
3135
+ cacheRead: 0.175,
3136
+ cacheWrite: 0,
3137
+ },
3138
+ contextWindow: 400000,
3139
+ maxTokens: 128000,
3140
+ },
3141
+ "gpt-5.2-codex": {
3142
+ id: "gpt-5.2-codex",
3143
+ name: "GPT-5.2 Codex",
3144
+ api: "openai-responses",
3145
+ provider: "cloudflare-ai-gateway",
3146
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3147
+ reasoning: true,
3148
+ input: ["text", "image"],
3149
+ cost: {
3150
+ input: 1.75,
3151
+ output: 14,
3152
+ cacheRead: 0.175,
3153
+ cacheWrite: 0,
3154
+ },
3155
+ contextWindow: 400000,
3156
+ maxTokens: 128000,
3157
+ },
3158
+ "gpt-5.3-codex": {
3159
+ id: "gpt-5.3-codex",
3160
+ name: "GPT-5.3 Codex",
3161
+ api: "openai-responses",
3162
+ provider: "cloudflare-ai-gateway",
3163
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3164
+ reasoning: true,
3165
+ input: ["text", "image"],
3166
+ cost: {
3167
+ input: 1.75,
3168
+ output: 14,
3169
+ cacheRead: 0.175,
3170
+ cacheWrite: 0,
3171
+ },
3172
+ contextWindow: 400000,
3173
+ maxTokens: 128000,
3174
+ },
3175
+ "gpt-5.4": {
3176
+ id: "gpt-5.4",
3177
+ name: "GPT-5.4",
3178
+ api: "openai-responses",
3179
+ provider: "cloudflare-ai-gateway",
3180
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3181
+ reasoning: true,
3182
+ input: ["text", "image"],
3183
+ cost: {
3184
+ input: 2.5,
3185
+ output: 15,
3186
+ cacheRead: 0.25,
3187
+ cacheWrite: 0,
3188
+ },
3189
+ contextWindow: 1050000,
3190
+ maxTokens: 128000,
3191
+ },
3192
+ "gpt-5.5": {
3193
+ id: "gpt-5.5",
3194
+ name: "GPT-5.5",
3195
+ api: "openai-responses",
3196
+ provider: "cloudflare-ai-gateway",
3197
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3198
+ reasoning: true,
3199
+ input: ["text", "image"],
3200
+ cost: {
3201
+ input: 5,
3202
+ output: 30,
3203
+ cacheRead: 0.5,
3204
+ cacheWrite: 0,
3205
+ },
3206
+ contextWindow: 1050000,
3207
+ maxTokens: 128000,
3208
+ },
3209
+ "o1": {
3210
+ id: "o1",
3211
+ name: "o1",
3212
+ api: "openai-responses",
3213
+ provider: "cloudflare-ai-gateway",
3214
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3215
+ reasoning: true,
3216
+ input: ["text", "image"],
3217
+ cost: {
3218
+ input: 15,
3219
+ output: 60,
3220
+ cacheRead: 7.5,
3221
+ cacheWrite: 0,
3222
+ },
3223
+ contextWindow: 200000,
3224
+ maxTokens: 100000,
3225
+ },
3226
+ "o3": {
3227
+ id: "o3",
3228
+ name: "o3",
3229
+ api: "openai-responses",
3230
+ provider: "cloudflare-ai-gateway",
3231
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3232
+ reasoning: true,
3233
+ input: ["text", "image"],
3234
+ cost: {
3235
+ input: 2,
3236
+ output: 8,
3237
+ cacheRead: 0.5,
3238
+ cacheWrite: 0,
3239
+ },
3240
+ contextWindow: 200000,
3241
+ maxTokens: 100000,
3242
+ },
3243
+ "o3-mini": {
3244
+ id: "o3-mini",
3245
+ name: "o3-mini",
3246
+ api: "openai-responses",
3247
+ provider: "cloudflare-ai-gateway",
3248
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3249
+ reasoning: true,
3250
+ input: ["text"],
3251
+ cost: {
3252
+ input: 1.1,
3253
+ output: 4.4,
3254
+ cacheRead: 0.55,
3255
+ cacheWrite: 0,
3256
+ },
3257
+ contextWindow: 200000,
3258
+ maxTokens: 100000,
3259
+ },
3260
+ "o3-pro": {
3261
+ id: "o3-pro",
3262
+ name: "o3-pro",
3263
+ api: "openai-responses",
3264
+ provider: "cloudflare-ai-gateway",
3265
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3266
+ reasoning: true,
3267
+ input: ["text", "image"],
3268
+ cost: {
3269
+ input: 20,
3270
+ output: 80,
3271
+ cacheRead: 0,
3272
+ cacheWrite: 0,
3273
+ },
3274
+ contextWindow: 200000,
3275
+ maxTokens: 100000,
3276
+ },
3277
+ "o4-mini": {
3278
+ id: "o4-mini",
3279
+ name: "o4-mini",
3280
+ api: "openai-responses",
3281
+ provider: "cloudflare-ai-gateway",
3282
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/openai",
3283
+ reasoning: true,
3284
+ input: ["text", "image"],
3285
+ cost: {
3286
+ input: 1.1,
3287
+ output: 4.4,
3288
+ cacheRead: 0.28,
3289
+ cacheWrite: 0,
3290
+ },
3291
+ contextWindow: 200000,
3292
+ maxTokens: 100000,
3293
+ },
3294
+ "workers-ai/@cf/moonshotai/kimi-k2.5": {
3295
+ id: "workers-ai/@cf/moonshotai/kimi-k2.5",
3296
+ name: "Kimi K2.5",
3297
+ api: "openai-completions",
3298
+ provider: "cloudflare-ai-gateway",
3299
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/compat",
3300
+ compat: { "sendSessionAffinityHeaders": true },
3301
+ reasoning: true,
3302
+ input: ["text", "image"],
3303
+ cost: {
3304
+ input: 0.6,
3305
+ output: 3,
3306
+ cacheRead: 0.1,
3307
+ cacheWrite: 0,
3308
+ },
3309
+ contextWindow: 256000,
3310
+ maxTokens: 256000,
3311
+ },
3312
+ "workers-ai/@cf/moonshotai/kimi-k2.6": {
3313
+ id: "workers-ai/@cf/moonshotai/kimi-k2.6",
3314
+ name: "Kimi K2.6",
3315
+ api: "openai-completions",
3316
+ provider: "cloudflare-ai-gateway",
3317
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/compat",
3318
+ compat: { "sendSessionAffinityHeaders": true },
3319
+ reasoning: true,
3320
+ input: ["text", "image"],
3321
+ cost: {
3322
+ input: 0.95,
3323
+ output: 4,
3324
+ cacheRead: 0.16,
3325
+ cacheWrite: 0,
3326
+ },
3327
+ contextWindow: 256000,
3328
+ maxTokens: 256000,
3329
+ },
3330
+ "workers-ai/@cf/nvidia/nemotron-3-120b-a12b": {
3331
+ id: "workers-ai/@cf/nvidia/nemotron-3-120b-a12b",
3332
+ name: "Nemotron 3 Super 120B",
3333
+ api: "openai-completions",
3334
+ provider: "cloudflare-ai-gateway",
3335
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/compat",
3336
+ compat: { "sendSessionAffinityHeaders": true },
3337
+ reasoning: true,
3338
+ input: ["text"],
3339
+ cost: {
3340
+ input: 0.5,
3341
+ output: 1.5,
3342
+ cacheRead: 0,
3343
+ cacheWrite: 0,
3344
+ },
3345
+ contextWindow: 256000,
3346
+ maxTokens: 256000,
3347
+ },
3348
+ "workers-ai/@cf/zai-org/glm-4.7-flash": {
3349
+ id: "workers-ai/@cf/zai-org/glm-4.7-flash",
3350
+ name: "GLM-4.7-Flash",
3351
+ api: "openai-completions",
3352
+ provider: "cloudflare-ai-gateway",
3353
+ baseUrl: "https://gateway.ai.cloudflare.com/v1/{CLOUDFLARE_ACCOUNT_ID}/{CLOUDFLARE_GATEWAY_ID}/compat",
3354
+ compat: { "sendSessionAffinityHeaders": true },
3355
+ reasoning: true,
3356
+ input: ["text"],
3357
+ cost: {
3358
+ input: 0.06,
3359
+ output: 0.4,
3360
+ cacheRead: 0,
3361
+ cacheWrite: 0,
3362
+ },
3363
+ contextWindow: 131072,
3364
+ maxTokens: 131072,
3365
+ },
3366
+ },
3367
+ "cloudflare-workers-ai": {
3368
+ "@cf/google/gemma-4-26b-a4b-it": {
3369
+ id: "@cf/google/gemma-4-26b-a4b-it",
3370
+ name: "Gemma 4 26B A4B IT",
3371
+ api: "openai-completions",
3372
+ provider: "cloudflare-workers-ai",
3373
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3374
+ compat: { "sendSessionAffinityHeaders": true },
3375
+ reasoning: true,
3376
+ input: ["text", "image"],
3377
+ cost: {
3378
+ input: 0.1,
3379
+ output: 0.3,
3380
+ cacheRead: 0,
3381
+ cacheWrite: 0,
3382
+ },
3383
+ contextWindow: 256000,
3384
+ maxTokens: 16384,
3385
+ },
3386
+ "@cf/meta/llama-4-scout-17b-16e-instruct": {
3387
+ id: "@cf/meta/llama-4-scout-17b-16e-instruct",
3388
+ name: "Llama 4 Scout 17B 16E Instruct",
3389
+ api: "openai-completions",
3390
+ provider: "cloudflare-workers-ai",
3391
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3392
+ compat: { "sendSessionAffinityHeaders": true },
3393
+ reasoning: false,
3394
+ input: ["text", "image"],
3395
+ cost: {
3396
+ input: 0.27,
3397
+ output: 0.85,
3398
+ cacheRead: 0,
3399
+ cacheWrite: 0,
3400
+ },
3401
+ contextWindow: 128000,
3402
+ maxTokens: 16384,
3403
+ },
3404
+ "@cf/moonshotai/kimi-k2.5": {
3405
+ id: "@cf/moonshotai/kimi-k2.5",
3406
+ name: "Kimi K2.5",
3407
+ api: "openai-completions",
3408
+ provider: "cloudflare-workers-ai",
3409
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3410
+ compat: { "sendSessionAffinityHeaders": true },
3411
+ reasoning: true,
3412
+ input: ["text", "image"],
3413
+ cost: {
3414
+ input: 0.6,
3415
+ output: 3,
3416
+ cacheRead: 0.1,
3417
+ cacheWrite: 0,
3418
+ },
3419
+ contextWindow: 256000,
3420
+ maxTokens: 256000,
3421
+ },
3422
+ "@cf/moonshotai/kimi-k2.6": {
3423
+ id: "@cf/moonshotai/kimi-k2.6",
3424
+ name: "Kimi K2.6",
3425
+ api: "openai-completions",
3426
+ provider: "cloudflare-workers-ai",
3427
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3428
+ compat: { "sendSessionAffinityHeaders": true },
3429
+ reasoning: true,
3430
+ input: ["text", "image"],
3431
+ cost: {
3432
+ input: 0.95,
3433
+ output: 4,
3434
+ cacheRead: 0.16,
3435
+ cacheWrite: 0,
3436
+ },
3437
+ contextWindow: 256000,
3438
+ maxTokens: 256000,
3439
+ },
3440
+ "@cf/nvidia/nemotron-3-120b-a12b": {
3441
+ id: "@cf/nvidia/nemotron-3-120b-a12b",
3442
+ name: "Nemotron 3 Super 120B",
3443
+ api: "openai-completions",
3444
+ provider: "cloudflare-workers-ai",
3445
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3446
+ compat: { "sendSessionAffinityHeaders": true },
3447
+ reasoning: true,
3448
+ input: ["text"],
3449
+ cost: {
3450
+ input: 0.5,
3451
+ output: 1.5,
3452
+ cacheRead: 0,
3453
+ cacheWrite: 0,
3454
+ },
3455
+ contextWindow: 256000,
3456
+ maxTokens: 256000,
3457
+ },
3458
+ "@cf/openai/gpt-oss-120b": {
3459
+ id: "@cf/openai/gpt-oss-120b",
3460
+ name: "GPT OSS 120B",
3461
+ api: "openai-completions",
3462
+ provider: "cloudflare-workers-ai",
3463
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3464
+ compat: { "sendSessionAffinityHeaders": true },
3465
+ reasoning: true,
3466
+ input: ["text"],
3467
+ cost: {
3468
+ input: 0.35,
3469
+ output: 0.75,
3470
+ cacheRead: 0,
3471
+ cacheWrite: 0,
3472
+ },
3473
+ contextWindow: 128000,
3474
+ maxTokens: 16384,
3475
+ },
3476
+ "@cf/openai/gpt-oss-20b": {
3477
+ id: "@cf/openai/gpt-oss-20b",
3478
+ name: "GPT OSS 20B",
3479
+ api: "openai-completions",
3480
+ provider: "cloudflare-workers-ai",
3481
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3482
+ compat: { "sendSessionAffinityHeaders": true },
3483
+ reasoning: true,
2721
3484
  input: ["text"],
2722
3485
  cost: {
2723
- input: 0.6,
2724
- output: 1.2,
3486
+ input: 0.2,
3487
+ output: 0.3,
2725
3488
  cacheRead: 0,
2726
3489
  cacheWrite: 0,
2727
3490
  },
2728
- contextWindow: 131000,
2729
- maxTokens: 32000,
3491
+ contextWindow: 128000,
3492
+ maxTokens: 16384,
2730
3493
  },
2731
- "zai-glm-4.7": {
2732
- id: "zai-glm-4.7",
2733
- name: "Z.AI GLM-4.7",
3494
+ "@cf/zai-org/glm-4.7-flash": {
3495
+ id: "@cf/zai-org/glm-4.7-flash",
3496
+ name: "GLM-4.7-Flash",
2734
3497
  api: "openai-completions",
2735
- provider: "cerebras",
2736
- baseUrl: "https://api.cerebras.ai/v1",
2737
- reasoning: false,
3498
+ provider: "cloudflare-workers-ai",
3499
+ baseUrl: "https://api.cloudflare.com/client/v4/accounts/{CLOUDFLARE_ACCOUNT_ID}/ai/v1",
3500
+ compat: { "sendSessionAffinityHeaders": true },
3501
+ reasoning: true,
2738
3502
  input: ["text"],
2739
3503
  cost: {
2740
- input: 2.25,
2741
- output: 2.75,
3504
+ input: 0.06,
3505
+ output: 0.4,
2742
3506
  cacheRead: 0,
2743
3507
  cacheWrite: 0,
2744
3508
  },
2745
3509
  contextWindow: 131072,
2746
- maxTokens: 40000,
3510
+ maxTokens: 131072,
2747
3511
  },
2748
3512
  },
2749
3513
  "deepseek": {
@@ -2759,7 +3523,7 @@ export const MODELS = {
2759
3523
  cost: {
2760
3524
  input: 0.14,
2761
3525
  output: 0.28,
2762
- cacheRead: 0.028,
3526
+ cacheRead: 0.0028,
2763
3527
  cacheWrite: 0,
2764
3528
  },
2765
3529
  contextWindow: 1000000,
@@ -2775,9 +3539,9 @@ export const MODELS = {
2775
3539
  reasoning: true,
2776
3540
  input: ["text"],
2777
3541
  cost: {
2778
- input: 1.74,
2779
- output: 3.48,
2780
- cacheRead: 0.145,
3542
+ input: 0.435,
3543
+ output: 0.87,
3544
+ cacheRead: 0.003625,
2781
3545
  cacheWrite: 0,
2782
3546
  },
2783
3547
  contextWindow: 1000000,
@@ -2819,6 +3583,23 @@ export const MODELS = {
2819
3583
  contextWindow: 160000,
2820
3584
  maxTokens: 160000,
2821
3585
  },
3586
+ "accounts/fireworks/models/deepseek-v4-pro": {
3587
+ id: "accounts/fireworks/models/deepseek-v4-pro",
3588
+ name: "DeepSeek V4 Pro",
3589
+ api: "anthropic-messages",
3590
+ provider: "fireworks",
3591
+ baseUrl: "https://api.fireworks.ai/inference",
3592
+ reasoning: true,
3593
+ input: ["text"],
3594
+ cost: {
3595
+ input: 1.74,
3596
+ output: 3.48,
3597
+ cacheRead: 0.15,
3598
+ cacheWrite: 0,
3599
+ },
3600
+ contextWindow: 1000000,
3601
+ maxTokens: 384000,
3602
+ },
2822
3603
  "accounts/fireworks/models/glm-4p5": {
2823
3604
  id: "accounts/fireworks/models/glm-4p5",
2824
3605
  name: "GLM 4.5",
@@ -3076,7 +3857,7 @@ export const MODELS = {
3076
3857
  },
3077
3858
  "accounts/fireworks/routers/kimi-k2p5-turbo": {
3078
3859
  id: "accounts/fireworks/routers/kimi-k2p5-turbo",
3079
- name: "Kimi K2.5 Turbo (firepass)",
3860
+ name: "Kimi K2.5 Turbo",
3080
3861
  api: "anthropic-messages",
3081
3862
  provider: "fireworks",
3082
3863
  baseUrl: "https://api.fireworks.ai/inference",
@@ -3996,290 +4777,14 @@ export const MODELS = {
3996
4777
  cacheWrite: 0,
3997
4778
  },
3998
4779
  contextWindow: 131072,
3999
- maxTokens: 8192,
4000
- },
4001
- "gemma-4-26b-a4b-it": {
4002
- id: "gemma-4-26b-a4b-it",
4003
- name: "Gemma 4 26B",
4004
- api: "google-generative-ai",
4005
- provider: "google",
4006
- baseUrl: "https://generativelanguage.googleapis.com/v1beta",
4007
- reasoning: true,
4008
- input: ["text", "image"],
4009
- cost: {
4010
- input: 0,
4011
- output: 0,
4012
- cacheRead: 0,
4013
- cacheWrite: 0,
4014
- },
4015
- contextWindow: 256000,
4016
- maxTokens: 8192,
4017
- },
4018
- "gemma-4-31b-it": {
4019
- id: "gemma-4-31b-it",
4020
- name: "Gemma 4 31B",
4021
- api: "google-generative-ai",
4022
- provider: "google",
4023
- baseUrl: "https://generativelanguage.googleapis.com/v1beta",
4024
- reasoning: true,
4025
- input: ["text", "image"],
4026
- cost: {
4027
- input: 0,
4028
- output: 0,
4029
- cacheRead: 0,
4030
- cacheWrite: 0,
4031
- },
4032
- contextWindow: 256000,
4033
- maxTokens: 8192,
4034
- },
4035
- },
4036
- "google-antigravity": {
4037
- "claude-opus-4-5-thinking": {
4038
- id: "claude-opus-4-5-thinking",
4039
- name: "Claude Opus 4.5 Thinking (Antigravity)",
4040
- api: "google-gemini-cli",
4041
- provider: "google-antigravity",
4042
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4043
- reasoning: true,
4044
- input: ["text", "image"],
4045
- cost: {
4046
- input: 5,
4047
- output: 25,
4048
- cacheRead: 0.5,
4049
- cacheWrite: 6.25,
4050
- },
4051
- contextWindow: 200000,
4052
- maxTokens: 64000,
4053
- },
4054
- "claude-opus-4-6-thinking": {
4055
- id: "claude-opus-4-6-thinking",
4056
- name: "Claude Opus 4.6 Thinking (Antigravity)",
4057
- api: "google-gemini-cli",
4058
- provider: "google-antigravity",
4059
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4060
- reasoning: true,
4061
- input: ["text", "image"],
4062
- cost: {
4063
- input: 5,
4064
- output: 25,
4065
- cacheRead: 0.5,
4066
- cacheWrite: 6.25,
4067
- },
4068
- contextWindow: 200000,
4069
- maxTokens: 128000,
4070
- },
4071
- "claude-sonnet-4-5": {
4072
- id: "claude-sonnet-4-5",
4073
- name: "Claude Sonnet 4.5 (Antigravity)",
4074
- api: "google-gemini-cli",
4075
- provider: "google-antigravity",
4076
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4077
- reasoning: false,
4078
- input: ["text", "image"],
4079
- cost: {
4080
- input: 3,
4081
- output: 15,
4082
- cacheRead: 0.3,
4083
- cacheWrite: 3.75,
4084
- },
4085
- contextWindow: 200000,
4086
- maxTokens: 64000,
4087
- },
4088
- "claude-sonnet-4-5-thinking": {
4089
- id: "claude-sonnet-4-5-thinking",
4090
- name: "Claude Sonnet 4.5 Thinking (Antigravity)",
4091
- api: "google-gemini-cli",
4092
- provider: "google-antigravity",
4093
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4094
- reasoning: true,
4095
- input: ["text", "image"],
4096
- cost: {
4097
- input: 3,
4098
- output: 15,
4099
- cacheRead: 0.3,
4100
- cacheWrite: 3.75,
4101
- },
4102
- contextWindow: 200000,
4103
- maxTokens: 64000,
4104
- },
4105
- "claude-sonnet-4-6": {
4106
- id: "claude-sonnet-4-6",
4107
- name: "Claude Sonnet 4.6 (Antigravity)",
4108
- api: "google-gemini-cli",
4109
- provider: "google-antigravity",
4110
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4111
- reasoning: true,
4112
- input: ["text", "image"],
4113
- cost: {
4114
- input: 3,
4115
- output: 15,
4116
- cacheRead: 0.3,
4117
- cacheWrite: 3.75,
4118
- },
4119
- contextWindow: 200000,
4120
- maxTokens: 64000,
4121
- },
4122
- "gemini-3-flash": {
4123
- id: "gemini-3-flash",
4124
- name: "Gemini 3 Flash (Antigravity)",
4125
- api: "google-gemini-cli",
4126
- provider: "google-antigravity",
4127
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4128
- reasoning: true,
4129
- input: ["text", "image"],
4130
- cost: {
4131
- input: 0.5,
4132
- output: 3,
4133
- cacheRead: 0.5,
4134
- cacheWrite: 0,
4135
- },
4136
- contextWindow: 1048576,
4137
- maxTokens: 65535,
4138
- },
4139
- "gemini-3.1-pro-high": {
4140
- id: "gemini-3.1-pro-high",
4141
- name: "Gemini 3.1 Pro High (Antigravity)",
4142
- api: "google-gemini-cli",
4143
- provider: "google-antigravity",
4144
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4145
- reasoning: true,
4146
- input: ["text", "image"],
4147
- cost: {
4148
- input: 2,
4149
- output: 12,
4150
- cacheRead: 0.2,
4151
- cacheWrite: 2.375,
4152
- },
4153
- contextWindow: 1048576,
4154
- maxTokens: 65535,
4155
- },
4156
- "gemini-3.1-pro-low": {
4157
- id: "gemini-3.1-pro-low",
4158
- name: "Gemini 3.1 Pro Low (Antigravity)",
4159
- api: "google-gemini-cli",
4160
- provider: "google-antigravity",
4161
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4162
- reasoning: true,
4163
- input: ["text", "image"],
4164
- cost: {
4165
- input: 2,
4166
- output: 12,
4167
- cacheRead: 0.2,
4168
- cacheWrite: 2.375,
4169
- },
4170
- contextWindow: 1048576,
4171
- maxTokens: 65535,
4172
- },
4173
- "gpt-oss-120b-medium": {
4174
- id: "gpt-oss-120b-medium",
4175
- name: "GPT-OSS 120B Medium (Antigravity)",
4176
- api: "google-gemini-cli",
4177
- provider: "google-antigravity",
4178
- baseUrl: "https://daily-cloudcode-pa.sandbox.googleapis.com",
4179
- reasoning: false,
4180
- input: ["text"],
4181
- cost: {
4182
- input: 0.09,
4183
- output: 0.36,
4184
- cacheRead: 0,
4185
- cacheWrite: 0,
4186
- },
4187
- contextWindow: 131072,
4188
- maxTokens: 32768,
4189
- },
4190
- },
4191
- "google-gemini-cli": {
4192
- "gemini-2.0-flash": {
4193
- id: "gemini-2.0-flash",
4194
- name: "Gemini 2.0 Flash (Cloud Code Assist)",
4195
- api: "google-gemini-cli",
4196
- provider: "google-gemini-cli",
4197
- baseUrl: "https://cloudcode-pa.googleapis.com",
4198
- reasoning: false,
4199
- input: ["text", "image"],
4200
- cost: {
4201
- input: 0,
4202
- output: 0,
4203
- cacheRead: 0,
4204
- cacheWrite: 0,
4205
- },
4206
- contextWindow: 1048576,
4207
- maxTokens: 8192,
4208
- },
4209
- "gemini-2.5-flash": {
4210
- id: "gemini-2.5-flash",
4211
- name: "Gemini 2.5 Flash (Cloud Code Assist)",
4212
- api: "google-gemini-cli",
4213
- provider: "google-gemini-cli",
4214
- baseUrl: "https://cloudcode-pa.googleapis.com",
4215
- reasoning: true,
4216
- input: ["text", "image"],
4217
- cost: {
4218
- input: 0,
4219
- output: 0,
4220
- cacheRead: 0,
4221
- cacheWrite: 0,
4222
- },
4223
- contextWindow: 1048576,
4224
- maxTokens: 65535,
4225
- },
4226
- "gemini-2.5-pro": {
4227
- id: "gemini-2.5-pro",
4228
- name: "Gemini 2.5 Pro (Cloud Code Assist)",
4229
- api: "google-gemini-cli",
4230
- provider: "google-gemini-cli",
4231
- baseUrl: "https://cloudcode-pa.googleapis.com",
4232
- reasoning: true,
4233
- input: ["text", "image"],
4234
- cost: {
4235
- input: 0,
4236
- output: 0,
4237
- cacheRead: 0,
4238
- cacheWrite: 0,
4239
- },
4240
- contextWindow: 1048576,
4241
- maxTokens: 65535,
4242
- },
4243
- "gemini-3-flash-preview": {
4244
- id: "gemini-3-flash-preview",
4245
- name: "Gemini 3 Flash Preview (Cloud Code Assist)",
4246
- api: "google-gemini-cli",
4247
- provider: "google-gemini-cli",
4248
- baseUrl: "https://cloudcode-pa.googleapis.com",
4249
- reasoning: true,
4250
- input: ["text", "image"],
4251
- cost: {
4252
- input: 0,
4253
- output: 0,
4254
- cacheRead: 0,
4255
- cacheWrite: 0,
4256
- },
4257
- contextWindow: 1048576,
4258
- maxTokens: 65535,
4259
- },
4260
- "gemini-3-pro-preview": {
4261
- id: "gemini-3-pro-preview",
4262
- name: "Gemini 3 Pro Preview (Cloud Code Assist)",
4263
- api: "google-gemini-cli",
4264
- provider: "google-gemini-cli",
4265
- baseUrl: "https://cloudcode-pa.googleapis.com",
4266
- reasoning: true,
4267
- input: ["text", "image"],
4268
- cost: {
4269
- input: 0,
4270
- output: 0,
4271
- cacheRead: 0,
4272
- cacheWrite: 0,
4273
- },
4274
- contextWindow: 1048576,
4275
- maxTokens: 65535,
4780
+ maxTokens: 8192,
4276
4781
  },
4277
- "gemini-3.1-flash-lite-preview": {
4278
- id: "gemini-3.1-flash-lite-preview",
4279
- name: "Gemini 3.1 Flash Lite Preview (Cloud Code Assist)",
4280
- api: "google-gemini-cli",
4281
- provider: "google-gemini-cli",
4282
- baseUrl: "https://cloudcode-pa.googleapis.com",
4782
+ "gemma-4-26b-a4b-it": {
4783
+ id: "gemma-4-26b-a4b-it",
4784
+ name: "Gemma 4 26B",
4785
+ api: "google-generative-ai",
4786
+ provider: "google",
4787
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
4283
4788
  reasoning: true,
4284
4789
  input: ["text", "image"],
4285
4790
  cost: {
@@ -4288,15 +4793,15 @@ export const MODELS = {
4288
4793
  cacheRead: 0,
4289
4794
  cacheWrite: 0,
4290
4795
  },
4291
- contextWindow: 1048576,
4292
- maxTokens: 65535,
4796
+ contextWindow: 256000,
4797
+ maxTokens: 8192,
4293
4798
  },
4294
- "gemini-3.1-pro-preview": {
4295
- id: "gemini-3.1-pro-preview",
4296
- name: "Gemini 3.1 Pro Preview (Cloud Code Assist)",
4297
- api: "google-gemini-cli",
4298
- provider: "google-gemini-cli",
4299
- baseUrl: "https://cloudcode-pa.googleapis.com",
4799
+ "gemma-4-31b-it": {
4800
+ id: "gemma-4-31b-it",
4801
+ name: "Gemma 4 31B",
4802
+ api: "google-generative-ai",
4803
+ provider: "google",
4804
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta",
4300
4805
  reasoning: true,
4301
4806
  input: ["text", "image"],
4302
4807
  cost: {
@@ -4305,8 +4810,8 @@ export const MODELS = {
4305
4810
  cacheRead: 0,
4306
4811
  cacheWrite: 0,
4307
4812
  },
4308
- contextWindow: 1048576,
4309
- maxTokens: 65535,
4813
+ contextWindow: 256000,
4814
+ maxTokens: 8192,
4310
4815
  },
4311
4816
  },
4312
4817
  "google-vertex": {
@@ -5639,22 +6144,56 @@ export const MODELS = {
5639
6144
  contextWindow: 262144,
5640
6145
  maxTokens: 262144,
5641
6146
  },
6147
+ "mistral-medium-2604": {
6148
+ id: "mistral-medium-2604",
6149
+ name: "Mistral Medium 3.5",
6150
+ api: "mistral-conversations",
6151
+ provider: "mistral",
6152
+ baseUrl: "https://api.mistral.ai",
6153
+ reasoning: true,
6154
+ input: ["text", "image"],
6155
+ cost: {
6156
+ input: 1.5,
6157
+ output: 7.5,
6158
+ cacheRead: 0,
6159
+ cacheWrite: 0,
6160
+ },
6161
+ contextWindow: 262144,
6162
+ maxTokens: 262144,
6163
+ },
6164
+ "mistral-medium-3.5": {
6165
+ id: "mistral-medium-3.5",
6166
+ name: "Mistral Medium 3.5",
6167
+ api: "mistral-conversations",
6168
+ provider: "mistral",
6169
+ baseUrl: "https://api.mistral.ai",
6170
+ reasoning: true,
6171
+ input: ["text", "image"],
6172
+ cost: {
6173
+ input: 1.5,
6174
+ output: 7.5,
6175
+ cacheRead: 0,
6176
+ cacheWrite: 0,
6177
+ },
6178
+ contextWindow: 262144,
6179
+ maxTokens: 262144,
6180
+ },
5642
6181
  "mistral-medium-latest": {
5643
6182
  id: "mistral-medium-latest",
5644
6183
  name: "Mistral Medium (latest)",
5645
6184
  api: "mistral-conversations",
5646
6185
  provider: "mistral",
5647
6186
  baseUrl: "https://api.mistral.ai",
5648
- reasoning: false,
6187
+ reasoning: true,
5649
6188
  input: ["text", "image"],
5650
6189
  cost: {
5651
- input: 0.4,
5652
- output: 2,
6190
+ input: 1.5,
6191
+ output: 7.5,
5653
6192
  cacheRead: 0,
5654
6193
  cacheWrite: 0,
5655
6194
  },
5656
- contextWindow: 128000,
5657
- maxTokens: 16384,
6195
+ contextWindow: 262144,
6196
+ maxTokens: 262144,
5658
6197
  },
5659
6198
  "mistral-nemo": {
5660
6199
  id: "mistral-nemo",
@@ -5801,13 +6340,269 @@ export const MODELS = {
5801
6340
  reasoning: false,
5802
6341
  input: ["text", "image"],
5803
6342
  cost: {
5804
- input: 2,
5805
- output: 6,
5806
- cacheRead: 0,
6343
+ input: 2,
6344
+ output: 6,
6345
+ cacheRead: 0,
6346
+ cacheWrite: 0,
6347
+ },
6348
+ contextWindow: 128000,
6349
+ maxTokens: 128000,
6350
+ },
6351
+ },
6352
+ "moonshotai": {
6353
+ "kimi-k2-0711-preview": {
6354
+ id: "kimi-k2-0711-preview",
6355
+ name: "Kimi K2 0711",
6356
+ api: "openai-completions",
6357
+ provider: "moonshotai",
6358
+ baseUrl: "https://api.moonshot.ai/v1",
6359
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6360
+ reasoning: false,
6361
+ input: ["text"],
6362
+ cost: {
6363
+ input: 0.6,
6364
+ output: 2.5,
6365
+ cacheRead: 0.15,
6366
+ cacheWrite: 0,
6367
+ },
6368
+ contextWindow: 131072,
6369
+ maxTokens: 16384,
6370
+ },
6371
+ "kimi-k2-0905-preview": {
6372
+ id: "kimi-k2-0905-preview",
6373
+ name: "Kimi K2 0905",
6374
+ api: "openai-completions",
6375
+ provider: "moonshotai",
6376
+ baseUrl: "https://api.moonshot.ai/v1",
6377
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6378
+ reasoning: false,
6379
+ input: ["text"],
6380
+ cost: {
6381
+ input: 0.6,
6382
+ output: 2.5,
6383
+ cacheRead: 0.15,
6384
+ cacheWrite: 0,
6385
+ },
6386
+ contextWindow: 262144,
6387
+ maxTokens: 262144,
6388
+ },
6389
+ "kimi-k2-thinking": {
6390
+ id: "kimi-k2-thinking",
6391
+ name: "Kimi K2 Thinking",
6392
+ api: "openai-completions",
6393
+ provider: "moonshotai",
6394
+ baseUrl: "https://api.moonshot.ai/v1",
6395
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6396
+ reasoning: true,
6397
+ input: ["text"],
6398
+ cost: {
6399
+ input: 0.6,
6400
+ output: 2.5,
6401
+ cacheRead: 0.15,
6402
+ cacheWrite: 0,
6403
+ },
6404
+ contextWindow: 262144,
6405
+ maxTokens: 262144,
6406
+ },
6407
+ "kimi-k2-thinking-turbo": {
6408
+ id: "kimi-k2-thinking-turbo",
6409
+ name: "Kimi K2 Thinking Turbo",
6410
+ api: "openai-completions",
6411
+ provider: "moonshotai",
6412
+ baseUrl: "https://api.moonshot.ai/v1",
6413
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6414
+ reasoning: true,
6415
+ input: ["text"],
6416
+ cost: {
6417
+ input: 1.15,
6418
+ output: 8,
6419
+ cacheRead: 0.15,
6420
+ cacheWrite: 0,
6421
+ },
6422
+ contextWindow: 262144,
6423
+ maxTokens: 262144,
6424
+ },
6425
+ "kimi-k2-turbo-preview": {
6426
+ id: "kimi-k2-turbo-preview",
6427
+ name: "Kimi K2 Turbo",
6428
+ api: "openai-completions",
6429
+ provider: "moonshotai",
6430
+ baseUrl: "https://api.moonshot.ai/v1",
6431
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6432
+ reasoning: false,
6433
+ input: ["text"],
6434
+ cost: {
6435
+ input: 2.4,
6436
+ output: 10,
6437
+ cacheRead: 0.6,
6438
+ cacheWrite: 0,
6439
+ },
6440
+ contextWindow: 262144,
6441
+ maxTokens: 262144,
6442
+ },
6443
+ "kimi-k2.5": {
6444
+ id: "kimi-k2.5",
6445
+ name: "Kimi K2.5",
6446
+ api: "openai-completions",
6447
+ provider: "moonshotai",
6448
+ baseUrl: "https://api.moonshot.ai/v1",
6449
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6450
+ reasoning: true,
6451
+ input: ["text", "image"],
6452
+ cost: {
6453
+ input: 0.6,
6454
+ output: 3,
6455
+ cacheRead: 0.1,
6456
+ cacheWrite: 0,
6457
+ },
6458
+ contextWindow: 262144,
6459
+ maxTokens: 262144,
6460
+ },
6461
+ "kimi-k2.6": {
6462
+ id: "kimi-k2.6",
6463
+ name: "Kimi K2.6",
6464
+ api: "openai-completions",
6465
+ provider: "moonshotai",
6466
+ baseUrl: "https://api.moonshot.ai/v1",
6467
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6468
+ reasoning: true,
6469
+ input: ["text", "image"],
6470
+ cost: {
6471
+ input: 0.95,
6472
+ output: 4,
6473
+ cacheRead: 0.16,
6474
+ cacheWrite: 0,
6475
+ },
6476
+ contextWindow: 262144,
6477
+ maxTokens: 262144,
6478
+ },
6479
+ },
6480
+ "moonshotai-cn": {
6481
+ "kimi-k2-0711-preview": {
6482
+ id: "kimi-k2-0711-preview",
6483
+ name: "Kimi K2 0711",
6484
+ api: "openai-completions",
6485
+ provider: "moonshotai-cn",
6486
+ baseUrl: "https://api.moonshot.cn/v1",
6487
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6488
+ reasoning: false,
6489
+ input: ["text"],
6490
+ cost: {
6491
+ input: 0.6,
6492
+ output: 2.5,
6493
+ cacheRead: 0.15,
6494
+ cacheWrite: 0,
6495
+ },
6496
+ contextWindow: 131072,
6497
+ maxTokens: 16384,
6498
+ },
6499
+ "kimi-k2-0905-preview": {
6500
+ id: "kimi-k2-0905-preview",
6501
+ name: "Kimi K2 0905",
6502
+ api: "openai-completions",
6503
+ provider: "moonshotai-cn",
6504
+ baseUrl: "https://api.moonshot.cn/v1",
6505
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6506
+ reasoning: false,
6507
+ input: ["text"],
6508
+ cost: {
6509
+ input: 0.6,
6510
+ output: 2.5,
6511
+ cacheRead: 0.15,
6512
+ cacheWrite: 0,
6513
+ },
6514
+ contextWindow: 262144,
6515
+ maxTokens: 262144,
6516
+ },
6517
+ "kimi-k2-thinking": {
6518
+ id: "kimi-k2-thinking",
6519
+ name: "Kimi K2 Thinking",
6520
+ api: "openai-completions",
6521
+ provider: "moonshotai-cn",
6522
+ baseUrl: "https://api.moonshot.cn/v1",
6523
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6524
+ reasoning: true,
6525
+ input: ["text"],
6526
+ cost: {
6527
+ input: 0.6,
6528
+ output: 2.5,
6529
+ cacheRead: 0.15,
6530
+ cacheWrite: 0,
6531
+ },
6532
+ contextWindow: 262144,
6533
+ maxTokens: 262144,
6534
+ },
6535
+ "kimi-k2-thinking-turbo": {
6536
+ id: "kimi-k2-thinking-turbo",
6537
+ name: "Kimi K2 Thinking Turbo",
6538
+ api: "openai-completions",
6539
+ provider: "moonshotai-cn",
6540
+ baseUrl: "https://api.moonshot.cn/v1",
6541
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6542
+ reasoning: true,
6543
+ input: ["text"],
6544
+ cost: {
6545
+ input: 1.15,
6546
+ output: 8,
6547
+ cacheRead: 0.15,
6548
+ cacheWrite: 0,
6549
+ },
6550
+ contextWindow: 262144,
6551
+ maxTokens: 262144,
6552
+ },
6553
+ "kimi-k2-turbo-preview": {
6554
+ id: "kimi-k2-turbo-preview",
6555
+ name: "Kimi K2 Turbo",
6556
+ api: "openai-completions",
6557
+ provider: "moonshotai-cn",
6558
+ baseUrl: "https://api.moonshot.cn/v1",
6559
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6560
+ reasoning: false,
6561
+ input: ["text"],
6562
+ cost: {
6563
+ input: 2.4,
6564
+ output: 10,
6565
+ cacheRead: 0.6,
6566
+ cacheWrite: 0,
6567
+ },
6568
+ contextWindow: 262144,
6569
+ maxTokens: 262144,
6570
+ },
6571
+ "kimi-k2.5": {
6572
+ id: "kimi-k2.5",
6573
+ name: "Kimi K2.5",
6574
+ api: "openai-completions",
6575
+ provider: "moonshotai-cn",
6576
+ baseUrl: "https://api.moonshot.cn/v1",
6577
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6578
+ reasoning: true,
6579
+ input: ["text", "image"],
6580
+ cost: {
6581
+ input: 0.6,
6582
+ output: 3,
6583
+ cacheRead: 0.1,
6584
+ cacheWrite: 0,
6585
+ },
6586
+ contextWindow: 262144,
6587
+ maxTokens: 262144,
6588
+ },
6589
+ "kimi-k2.6": {
6590
+ id: "kimi-k2.6",
6591
+ name: "Kimi K2.6",
6592
+ api: "openai-completions",
6593
+ provider: "moonshotai-cn",
6594
+ baseUrl: "https://api.moonshot.cn/v1",
6595
+ compat: { "supportsStore": false, "supportsDeveloperRole": false, "supportsReasoningEffort": false, "maxTokensField": "max_tokens", "supportsStrictMode": false },
6596
+ reasoning: true,
6597
+ input: ["text", "image"],
6598
+ cost: {
6599
+ input: 0.95,
6600
+ output: 4,
6601
+ cacheRead: 0.16,
5807
6602
  cacheWrite: 0,
5808
6603
  },
5809
- contextWindow: 128000,
5810
- maxTokens: 128000,
6604
+ contextWindow: 262144,
6605
+ maxTokens: 262144,
5811
6606
  },
5812
6607
  },
5813
6608
  "openai": {
@@ -6372,6 +7167,23 @@ export const MODELS = {
6372
7167
  contextWindow: 272000,
6373
7168
  maxTokens: 128000,
6374
7169
  },
7170
+ "gpt-5.5-pro": {
7171
+ id: "gpt-5.5-pro",
7172
+ name: "GPT-5.5 Pro",
7173
+ api: "openai-responses",
7174
+ provider: "openai",
7175
+ baseUrl: "https://api.openai.com/v1",
7176
+ reasoning: true,
7177
+ input: ["text", "image"],
7178
+ cost: {
7179
+ input: 30,
7180
+ output: 180,
7181
+ cacheRead: 0,
7182
+ cacheWrite: 0,
7183
+ },
7184
+ contextWindow: 1050000,
7185
+ maxTokens: 128000,
7186
+ },
6375
7187
  "o1": {
6376
7188
  id: "o1",
6377
7189
  name: "o1",
@@ -7226,23 +8038,6 @@ export const MODELS = {
7226
8038
  contextWindow: 262144,
7227
8039
  maxTokens: 65536,
7228
8040
  },
7229
- "ling-2.6-flash-free": {
7230
- id: "ling-2.6-flash-free",
7231
- name: "Ling 2.6 Flash Free",
7232
- api: "openai-completions",
7233
- provider: "opencode",
7234
- baseUrl: "https://opencode.ai/zen/v1",
7235
- reasoning: false,
7236
- input: ["text"],
7237
- cost: {
7238
- input: 0,
7239
- output: 0,
7240
- cacheRead: 0,
7241
- cacheWrite: 0,
7242
- },
7243
- contextWindow: 262100,
7244
- maxTokens: 32800,
7245
- },
7246
8041
  "minimax-m2.5": {
7247
8042
  id: "minimax-m2.5",
7248
8043
  name: "MiniMax M2.5",
@@ -7353,7 +8148,7 @@ export const MODELS = {
7353
8148
  api: "openai-completions",
7354
8149
  provider: "opencode-go",
7355
8150
  baseUrl: "https://opencode.ai/zen/go/v1",
7356
- compat: { "requiresReasoningContentOnAssistantMessages": true },
8151
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
7357
8152
  reasoning: true,
7358
8153
  input: ["text"],
7359
8154
  cost: {
@@ -7371,7 +8166,7 @@ export const MODELS = {
7371
8166
  api: "openai-completions",
7372
8167
  provider: "opencode-go",
7373
8168
  baseUrl: "https://opencode.ai/zen/go/v1",
7374
- compat: { "requiresReasoningContentOnAssistantMessages": true },
8169
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
7375
8170
  reasoning: true,
7376
8171
  input: ["text"],
7377
8172
  cost: {
@@ -7397,8 +8192,8 @@ export const MODELS = {
7397
8192
  cacheRead: 0.2,
7398
8193
  cacheWrite: 0,
7399
8194
  },
7400
- contextWindow: 204800,
7401
- maxTokens: 131072,
8195
+ contextWindow: 202752,
8196
+ maxTokens: 32768,
7402
8197
  },
7403
8198
  "glm-5.1": {
7404
8199
  id: "glm-5.1",
@@ -7414,8 +8209,8 @@ export const MODELS = {
7414
8209
  cacheRead: 0.26,
7415
8210
  cacheWrite: 0,
7416
8211
  },
7417
- contextWindow: 204800,
7418
- maxTokens: 131072,
8212
+ contextWindow: 202752,
8213
+ maxTokens: 32768,
7419
8214
  },
7420
8215
  "kimi-k2.5": {
7421
8216
  id: "kimi-k2.5",
@@ -8326,7 +9121,7 @@ export const MODELS = {
8326
9121
  api: "openai-completions",
8327
9122
  provider: "openrouter",
8328
9123
  baseUrl: "https://openrouter.ai/api/v1",
8329
- compat: { "requiresReasoningContentOnAssistantMessages": true },
9124
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
8330
9125
  reasoning: true,
8331
9126
  input: ["text"],
8332
9127
  cost: {
@@ -8344,7 +9139,7 @@ export const MODELS = {
8344
9139
  api: "openai-completions",
8345
9140
  provider: "openrouter",
8346
9141
  baseUrl: "https://openrouter.ai/api/v1",
8347
- compat: { "requiresReasoningContentOnAssistantMessages": true },
9142
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
8348
9143
  reasoning: true,
8349
9144
  input: ["text"],
8350
9145
  cost: {
@@ -8679,6 +9474,23 @@ export const MODELS = {
8679
9474
  contextWindow: 262144,
8680
9475
  maxTokens: 32768,
8681
9476
  },
9477
+ "ibm-granite/granite-4.1-8b": {
9478
+ id: "ibm-granite/granite-4.1-8b",
9479
+ name: "IBM: Granite 4.1 8B",
9480
+ api: "openai-completions",
9481
+ provider: "openrouter",
9482
+ baseUrl: "https://openrouter.ai/api/v1",
9483
+ reasoning: false,
9484
+ input: ["text"],
9485
+ cost: {
9486
+ input: 0.049999999999999996,
9487
+ output: 0.09999999999999999,
9488
+ cacheRead: 0.049999999999999996,
9489
+ cacheWrite: 0,
9490
+ },
9491
+ contextWindow: 131072,
9492
+ maxTokens: 131072,
9493
+ },
8682
9494
  "inception/mercury-2": {
8683
9495
  id: "inception/mercury-2",
8684
9496
  name: "Inception: Mercury 2",
@@ -8713,18 +9525,18 @@ export const MODELS = {
8713
9525
  contextWindow: 262144,
8714
9526
  maxTokens: 32768,
8715
9527
  },
8716
- "inclusionai/ling-2.6-flash:free": {
8717
- id: "inclusionai/ling-2.6-flash:free",
8718
- name: "inclusionAI: Ling-2.6-flash (free)",
9528
+ "inclusionai/ling-2.6-flash": {
9529
+ id: "inclusionai/ling-2.6-flash",
9530
+ name: "inclusionAI: Ling-2.6-flash",
8719
9531
  api: "openai-completions",
8720
9532
  provider: "openrouter",
8721
9533
  baseUrl: "https://openrouter.ai/api/v1",
8722
9534
  reasoning: false,
8723
9535
  input: ["text"],
8724
9536
  cost: {
8725
- input: 0,
8726
- output: 0,
8727
- cacheRead: 0,
9537
+ input: 0.08,
9538
+ output: 0.24,
9539
+ cacheRead: 0.016,
8728
9540
  cacheWrite: 0,
8729
9541
  },
8730
9542
  contextWindow: 262144,
@@ -9182,12 +9994,12 @@ export const MODELS = {
9182
9994
  input: ["text"],
9183
9995
  cost: {
9184
9996
  input: 0.02,
9185
- output: 0.04,
9997
+ output: 0.03,
9186
9998
  cacheRead: 0,
9187
9999
  cacheWrite: 0,
9188
10000
  },
9189
10001
  contextWindow: 131072,
9190
- maxTokens: 16384,
10002
+ maxTokens: 4096,
9191
10003
  },
9192
10004
  "mistralai/mistral-saba": {
9193
10005
  id: "mistralai/mistral-saba",
@@ -9240,23 +10052,6 @@ export const MODELS = {
9240
10052
  contextWindow: 128000,
9241
10053
  maxTokens: 16384,
9242
10054
  },
9243
- "mistralai/mistral-small-creative": {
9244
- id: "mistralai/mistral-small-creative",
9245
- name: "Mistral: Mistral Small Creative",
9246
- api: "openai-completions",
9247
- provider: "openrouter",
9248
- baseUrl: "https://openrouter.ai/api/v1",
9249
- reasoning: false,
9250
- input: ["text"],
9251
- cost: {
9252
- input: 0.09999999999999999,
9253
- output: 0.3,
9254
- cacheRead: 0.01,
9255
- cacheWrite: 0,
9256
- },
9257
- contextWindow: 32768,
9258
- maxTokens: 4096,
9259
- },
9260
10055
  "mistralai/mixtral-8x22b-instruct": {
9261
10056
  id: "mistralai/mixtral-8x22b-instruct",
9262
10057
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -9402,13 +10197,13 @@ export const MODELS = {
9402
10197
  reasoning: true,
9403
10198
  input: ["text", "image"],
9404
10199
  cost: {
9405
- input: 0.7448,
9406
- output: 4.655,
9407
- cacheRead: 0.1463,
10200
+ input: 0.74,
10201
+ output: 3.49,
10202
+ cacheRead: 0.14,
9408
10203
  cacheWrite: 0,
9409
10204
  },
9410
- contextWindow: 256000,
9411
- maxTokens: 65536,
10205
+ contextWindow: 262142,
10206
+ maxTokens: 262142,
9412
10207
  },
9413
10208
  "nex-agi/deepseek-v3.1-nex-n1": {
9414
10209
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -9495,6 +10290,23 @@ export const MODELS = {
9495
10290
  contextWindow: 256000,
9496
10291
  maxTokens: 4096,
9497
10292
  },
10293
+ "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
10294
+ id: "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free",
10295
+ name: "NVIDIA: Nemotron 3 Nano Omni (free)",
10296
+ api: "openai-completions",
10297
+ provider: "openrouter",
10298
+ baseUrl: "https://openrouter.ai/api/v1",
10299
+ reasoning: true,
10300
+ input: ["text", "image"],
10301
+ cost: {
10302
+ input: 0,
10303
+ output: 0,
10304
+ cacheRead: 0,
10305
+ cacheWrite: 0,
10306
+ },
10307
+ contextWindow: 256000,
10308
+ maxTokens: 65536,
10309
+ },
9498
10310
  "nvidia/nemotron-3-super-120b-a12b": {
9499
10311
  id: "nvidia/nemotron-3-super-120b-a12b",
9500
10312
  name: "NVIDIA: Nemotron 3 Super",
@@ -10304,12 +11116,12 @@ export const MODELS = {
10304
11116
  input: ["text"],
10305
11117
  cost: {
10306
11118
  input: 0.039,
10307
- output: 0.19,
11119
+ output: 0.18,
10308
11120
  cacheRead: 0,
10309
11121
  cacheWrite: 0,
10310
11122
  },
10311
11123
  contextWindow: 131072,
10312
- maxTokens: 131072,
11124
+ maxTokens: 4096,
10313
11125
  },
10314
11126
  "openai/gpt-oss-120b:free": {
10315
11127
  id: "openai/gpt-oss-120b:free",
@@ -10566,6 +11378,57 @@ export const MODELS = {
10566
11378
  contextWindow: 200000,
10567
11379
  maxTokens: 4096,
10568
11380
  },
11381
+ "openrouter/owl-alpha": {
11382
+ id: "openrouter/owl-alpha",
11383
+ name: "Owl Alpha",
11384
+ api: "openai-completions",
11385
+ provider: "openrouter",
11386
+ baseUrl: "https://openrouter.ai/api/v1",
11387
+ reasoning: false,
11388
+ input: ["text"],
11389
+ cost: {
11390
+ input: 0,
11391
+ output: 0,
11392
+ cacheRead: 0,
11393
+ cacheWrite: 0,
11394
+ },
11395
+ contextWindow: 1048756,
11396
+ maxTokens: 262144,
11397
+ },
11398
+ "poolside/laguna-m.1:free": {
11399
+ id: "poolside/laguna-m.1:free",
11400
+ name: "Poolside: Laguna M.1 (free)",
11401
+ api: "openai-completions",
11402
+ provider: "openrouter",
11403
+ baseUrl: "https://openrouter.ai/api/v1",
11404
+ reasoning: true,
11405
+ input: ["text"],
11406
+ cost: {
11407
+ input: 0,
11408
+ output: 0,
11409
+ cacheRead: 0,
11410
+ cacheWrite: 0,
11411
+ },
11412
+ contextWindow: 131072,
11413
+ maxTokens: 8192,
11414
+ },
11415
+ "poolside/laguna-xs.2:free": {
11416
+ id: "poolside/laguna-xs.2:free",
11417
+ name: "Poolside: Laguna XS.2 (free)",
11418
+ api: "openai-completions",
11419
+ provider: "openrouter",
11420
+ baseUrl: "https://openrouter.ai/api/v1",
11421
+ reasoning: true,
11422
+ input: ["text"],
11423
+ cost: {
11424
+ input: 0,
11425
+ output: 0,
11426
+ cacheRead: 0,
11427
+ cacheWrite: 0,
11428
+ },
11429
+ contextWindow: 131072,
11430
+ maxTokens: 8192,
11431
+ },
10569
11432
  "prime-intellect/intellect-3": {
10570
11433
  id: "prime-intellect/intellect-3",
10571
11434
  name: "Prime Intellect: INTELLECT-3",
@@ -11323,12 +12186,12 @@ export const MODELS = {
11323
12186
  reasoning: true,
11324
12187
  input: ["text", "image"],
11325
12188
  cost: {
11326
- input: 0.5,
11327
- output: 2,
11328
- cacheRead: 0.25,
12189
+ input: 0.325,
12190
+ output: 3.25,
12191
+ cacheRead: 0,
11329
12192
  cacheWrite: 0,
11330
12193
  },
11331
- contextWindow: 262144,
12194
+ contextWindow: 256000,
11332
12195
  maxTokens: 65536,
11333
12196
  },
11334
12197
  "qwen/qwen3.6-flash": {
@@ -11357,10 +12220,10 @@ export const MODELS = {
11357
12220
  reasoning: true,
11358
12221
  input: ["text"],
11359
12222
  cost: {
11360
- input: 1.3,
11361
- output: 7.8,
12223
+ input: 1.04,
12224
+ output: 6.24,
11362
12225
  cacheRead: 0,
11363
- cacheWrite: 1.625,
12226
+ cacheWrite: 1.3,
11364
12227
  },
11365
12228
  contextWindow: 262144,
11366
12229
  maxTokens: 65536,
@@ -11382,23 +12245,6 @@ export const MODELS = {
11382
12245
  contextWindow: 1000000,
11383
12246
  maxTokens: 65536,
11384
12247
  },
11385
- "qwen/qwq-32b": {
11386
- id: "qwen/qwq-32b",
11387
- name: "Qwen: QwQ 32B",
11388
- api: "openai-completions",
11389
- provider: "openrouter",
11390
- baseUrl: "https://openrouter.ai/api/v1",
11391
- reasoning: true,
11392
- input: ["text"],
11393
- cost: {
11394
- input: 0.15,
11395
- output: 0.58,
11396
- cacheRead: 0,
11397
- cacheWrite: 0,
11398
- },
11399
- contextWindow: 131072,
11400
- maxTokens: 131072,
11401
- },
11402
12248
  "rekaai/reka-edge": {
11403
12249
  id: "rekaai/reka-edge",
11404
12250
  name: "Reka Edge",
@@ -12122,13 +12968,13 @@ export const MODELS = {
12122
12968
  reasoning: true,
12123
12969
  input: ["text", "image"],
12124
12970
  cost: {
12125
- input: 0.7448,
12126
- output: 4.655,
12127
- cacheRead: 0.1463,
12971
+ input: 0.74,
12972
+ output: 3.49,
12973
+ cacheRead: 0.14,
12128
12974
  cacheWrite: 0,
12129
12975
  },
12130
- contextWindow: 256000,
12131
- maxTokens: 65536,
12976
+ contextWindow: 262142,
12977
+ maxTokens: 262142,
12132
12978
  },
12133
12979
  "~openai/gpt-latest": {
12134
12980
  id: "~openai/gpt-latest",
@@ -12253,20 +13099,20 @@ export const MODELS = {
12253
13099
  },
12254
13100
  "alibaba/qwen3-235b-a22b-thinking": {
12255
13101
  id: "alibaba/qwen3-235b-a22b-thinking",
12256
- name: "Qwen3 235B A22B Thinking 2507",
13102
+ name: "Qwen3 VL 235B A22B Thinking",
12257
13103
  api: "anthropic-messages",
12258
13104
  provider: "vercel-ai-gateway",
12259
13105
  baseUrl: "https://ai-gateway.vercel.sh",
12260
13106
  reasoning: true,
12261
13107
  input: ["text", "image"],
12262
13108
  cost: {
12263
- input: 0.22999999999999998,
12264
- output: 2.3,
12265
- cacheRead: 0.19999999999999998,
13109
+ input: 0.39999999999999997,
13110
+ output: 4,
13111
+ cacheRead: 0,
12266
13112
  cacheWrite: 0,
12267
13113
  },
12268
- contextWindow: 262114,
12269
- maxTokens: 262114,
13114
+ contextWindow: 131072,
13115
+ maxTokens: 32768,
12270
13116
  },
12271
13117
  "alibaba/qwen3-coder": {
12272
13118
  id: "alibaba/qwen3-coder",
@@ -12438,6 +13284,23 @@ export const MODELS = {
12438
13284
  contextWindow: 1000000,
12439
13285
  maxTokens: 64000,
12440
13286
  },
13287
+ "alibaba/qwen3.6-27b": {
13288
+ id: "alibaba/qwen3.6-27b",
13289
+ name: "Qwen 3.6 27B",
13290
+ api: "anthropic-messages",
13291
+ provider: "vercel-ai-gateway",
13292
+ baseUrl: "https://ai-gateway.vercel.sh",
13293
+ reasoning: true,
13294
+ input: ["text", "image"],
13295
+ cost: {
13296
+ input: 0.6,
13297
+ output: 3.5999999999999996,
13298
+ cacheRead: 0,
13299
+ cacheWrite: 0,
13300
+ },
13301
+ contextWindow: 256000,
13302
+ maxTokens: 256000,
13303
+ },
12441
13304
  "alibaba/qwen3.6-plus": {
12442
13305
  id: "alibaba/qwen3.6-plus",
12443
13306
  name: "Qwen 3.6 Plus",
@@ -12818,16 +13681,16 @@ export const MODELS = {
12818
13681
  api: "anthropic-messages",
12819
13682
  provider: "vercel-ai-gateway",
12820
13683
  baseUrl: "https://ai-gateway.vercel.sh",
12821
- reasoning: true,
13684
+ reasoning: false,
12822
13685
  input: ["text"],
12823
13686
  cost: {
12824
- input: 0.28,
12825
- output: 0.42,
12826
- cacheRead: 0.028,
13687
+ input: 0.62,
13688
+ output: 1.85,
13689
+ cacheRead: 0,
12827
13690
  cacheWrite: 0,
12828
13691
  },
12829
13692
  contextWindow: 128000,
12830
- maxTokens: 64000,
13693
+ maxTokens: 8000,
12831
13694
  },
12832
13695
  "deepseek/deepseek-v4-flash": {
12833
13696
  id: "deepseek/deepseek-v4-flash",
@@ -12840,7 +13703,7 @@ export const MODELS = {
12840
13703
  cost: {
12841
13704
  input: 0.14,
12842
13705
  output: 0.28,
12843
- cacheRead: 0.028,
13706
+ cacheRead: 0.0028,
12844
13707
  cacheWrite: 0,
12845
13708
  },
12846
13709
  contextWindow: 1000000,
@@ -14665,6 +15528,40 @@ export const MODELS = {
14665
15528
  contextWindow: 1000000,
14666
15529
  maxTokens: 128000,
14667
15530
  },
15531
+ "xiaomi/mimo-v2.5": {
15532
+ id: "xiaomi/mimo-v2.5",
15533
+ name: "MiMo M2.5",
15534
+ api: "anthropic-messages",
15535
+ provider: "vercel-ai-gateway",
15536
+ baseUrl: "https://ai-gateway.vercel.sh",
15537
+ reasoning: true,
15538
+ input: ["text", "image"],
15539
+ cost: {
15540
+ input: 0.39999999999999997,
15541
+ output: 2,
15542
+ cacheRead: 0.08,
15543
+ cacheWrite: 0,
15544
+ },
15545
+ contextWindow: 1050000,
15546
+ maxTokens: 131100,
15547
+ },
15548
+ "xiaomi/mimo-v2.5-pro": {
15549
+ id: "xiaomi/mimo-v2.5-pro",
15550
+ name: "MiMo V2.5 Pro",
15551
+ api: "anthropic-messages",
15552
+ provider: "vercel-ai-gateway",
15553
+ baseUrl: "https://ai-gateway.vercel.sh",
15554
+ reasoning: true,
15555
+ input: ["text", "image"],
15556
+ cost: {
15557
+ input: 1,
15558
+ output: 3,
15559
+ cacheRead: 0.19999999999999998,
15560
+ cacheWrite: 0,
15561
+ },
15562
+ contextWindow: 1050000,
15563
+ maxTokens: 131000,
15564
+ },
14668
15565
  "zai/glm-4.5": {
14669
15566
  id: "zai/glm-4.5",
14670
15567
  name: "GLM-4.5",
@@ -15370,6 +16267,24 @@ export const MODELS = {
15370
16267
  contextWindow: 200000,
15371
16268
  maxTokens: 131072,
15372
16269
  },
16270
+ "glm-5v-turbo": {
16271
+ id: "glm-5v-turbo",
16272
+ name: "glm-5v-turbo",
16273
+ api: "openai-completions",
16274
+ provider: "zai",
16275
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
16276
+ compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
16277
+ reasoning: true,
16278
+ input: ["text", "image"],
16279
+ cost: {
16280
+ input: 0,
16281
+ output: 0,
16282
+ cacheRead: 0,
16283
+ cacheWrite: 0,
16284
+ },
16285
+ contextWindow: 200000,
16286
+ maxTokens: 131072,
16287
+ },
15373
16288
  },
15374
16289
  };
15375
16290
  //# sourceMappingURL=models.generated.js.map