@mariozechner/pi-ai 0.70.6 → 0.71.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (65) hide show
  1. package/README.md +3 -30
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +3 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +1 -1
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js.map +1 -1
  8. package/dist/models.d.ts +1 -1
  9. package/dist/models.d.ts.map +1 -1
  10. package/dist/models.generated.d.ts +1378 -551
  11. package/dist/models.generated.d.ts.map +1 -1
  12. package/dist/models.generated.js +1150 -449
  13. package/dist/models.generated.js.map +1 -1
  14. package/dist/models.js +3 -2
  15. package/dist/models.js.map +1 -1
  16. package/dist/providers/anthropic.d.ts.map +1 -1
  17. package/dist/providers/anthropic.js +38 -15
  18. package/dist/providers/anthropic.js.map +1 -1
  19. package/dist/providers/cloudflare.d.ts +9 -3
  20. package/dist/providers/cloudflare.d.ts.map +1 -1
  21. package/dist/providers/cloudflare.js +10 -3
  22. package/dist/providers/cloudflare.js.map +1 -1
  23. package/dist/providers/google-shared.d.ts +7 -2
  24. package/dist/providers/google-shared.d.ts.map +1 -1
  25. package/dist/providers/google-shared.js +4 -13
  26. package/dist/providers/google-shared.js.map +1 -1
  27. package/dist/providers/google-vertex.d.ts +1 -1
  28. package/dist/providers/google-vertex.d.ts.map +1 -1
  29. package/dist/providers/google-vertex.js.map +1 -1
  30. package/dist/providers/google.d.ts +1 -1
  31. package/dist/providers/google.d.ts.map +1 -1
  32. package/dist/providers/google.js.map +1 -1
  33. package/dist/providers/mistral.d.ts.map +1 -1
  34. package/dist/providers/mistral.js +1 -1
  35. package/dist/providers/mistral.js.map +1 -1
  36. package/dist/providers/openai-completions.d.ts.map +1 -1
  37. package/dist/providers/openai-completions.js +21 -7
  38. package/dist/providers/openai-completions.js.map +1 -1
  39. package/dist/providers/openai-responses.d.ts.map +1 -1
  40. package/dist/providers/openai-responses.js +10 -2
  41. package/dist/providers/openai-responses.js.map +1 -1
  42. package/dist/providers/register-builtins.d.ts +0 -3
  43. package/dist/providers/register-builtins.d.ts.map +1 -1
  44. package/dist/providers/register-builtins.js +0 -18
  45. package/dist/providers/register-builtins.js.map +1 -1
  46. package/dist/types.d.ts +3 -2
  47. package/dist/types.d.ts.map +1 -1
  48. package/dist/types.js.map +1 -1
  49. package/dist/utils/oauth/index.d.ts +0 -4
  50. package/dist/utils/oauth/index.d.ts.map +1 -1
  51. package/dist/utils/oauth/index.js +0 -10
  52. package/dist/utils/oauth/index.js.map +1 -1
  53. package/package.json +2 -6
  54. package/dist/providers/google-gemini-cli.d.ts +0 -74
  55. package/dist/providers/google-gemini-cli.d.ts.map +0 -1
  56. package/dist/providers/google-gemini-cli.js +0 -779
  57. package/dist/providers/google-gemini-cli.js.map +0 -1
  58. package/dist/utils/oauth/google-antigravity.d.ts +0 -26
  59. package/dist/utils/oauth/google-antigravity.d.ts.map +0 -1
  60. package/dist/utils/oauth/google-antigravity.js +0 -377
  61. package/dist/utils/oauth/google-antigravity.js.map +0 -1
  62. package/dist/utils/oauth/google-gemini-cli.d.ts +0 -26
  63. package/dist/utils/oauth/google-gemini-cli.d.ts.map +0 -1
  64. package/dist/utils/oauth/google-gemini-cli.js +0 -482
  65. package/dist/utils/oauth/google-gemini-cli.js.map +0 -1
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -2744,17 +2761,14 @@ export declare const MODELS: {
2744
2761
  maxTokens: number;
2745
2762
  };
2746
2763
  };
2747
- readonly "cloudflare-workers-ai": {
2748
- readonly "@cf/google/gemma-4-26b-a4b-it": {
2764
+ readonly "cloudflare-ai-gateway": {
2765
+ readonly "claude-3-5-haiku": {
2749
2766
  id: string;
2750
2767
  name: string;
2751
- api: "openai-completions";
2768
+ api: "anthropic-messages";
2752
2769
  provider: string;
2753
2770
  baseUrl: string;
2754
- compat: {
2755
- sendSessionAffinityHeaders: true;
2756
- };
2757
- reasoning: true;
2771
+ reasoning: false;
2758
2772
  input: ("image" | "text")[];
2759
2773
  cost: {
2760
2774
  input: number;
@@ -2765,15 +2779,12 @@ export declare const MODELS: {
2765
2779
  contextWindow: number;
2766
2780
  maxTokens: number;
2767
2781
  };
2768
- readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
2782
+ readonly "claude-3-haiku": {
2769
2783
  id: string;
2770
2784
  name: string;
2771
- api: "openai-completions";
2785
+ api: "anthropic-messages";
2772
2786
  provider: string;
2773
2787
  baseUrl: string;
2774
- compat: {
2775
- sendSessionAffinityHeaders: true;
2776
- };
2777
2788
  reasoning: false;
2778
2789
  input: ("image" | "text")[];
2779
2790
  cost: {
@@ -2785,16 +2796,13 @@ export declare const MODELS: {
2785
2796
  contextWindow: number;
2786
2797
  maxTokens: number;
2787
2798
  };
2788
- readonly "@cf/moonshotai/kimi-k2.5": {
2799
+ readonly "claude-3-opus": {
2789
2800
  id: string;
2790
2801
  name: string;
2791
- api: "openai-completions";
2802
+ api: "anthropic-messages";
2792
2803
  provider: string;
2793
2804
  baseUrl: string;
2794
- compat: {
2795
- sendSessionAffinityHeaders: true;
2796
- };
2797
- reasoning: true;
2805
+ reasoning: false;
2798
2806
  input: ("image" | "text")[];
2799
2807
  cost: {
2800
2808
  input: number;
@@ -2805,16 +2813,13 @@ export declare const MODELS: {
2805
2813
  contextWindow: number;
2806
2814
  maxTokens: number;
2807
2815
  };
2808
- readonly "@cf/moonshotai/kimi-k2.6": {
2816
+ readonly "claude-3-sonnet": {
2809
2817
  id: string;
2810
2818
  name: string;
2811
- api: "openai-completions";
2819
+ api: "anthropic-messages";
2812
2820
  provider: string;
2813
2821
  baseUrl: string;
2814
- compat: {
2815
- sendSessionAffinityHeaders: true;
2816
- };
2817
- reasoning: true;
2822
+ reasoning: false;
2818
2823
  input: ("image" | "text")[];
2819
2824
  cost: {
2820
2825
  input: number;
@@ -2825,17 +2830,14 @@ export declare const MODELS: {
2825
2830
  contextWindow: number;
2826
2831
  maxTokens: number;
2827
2832
  };
2828
- readonly "@cf/nvidia/nemotron-3-120b-a12b": {
2833
+ readonly "claude-3.5-haiku": {
2829
2834
  id: string;
2830
2835
  name: string;
2831
- api: "openai-completions";
2836
+ api: "anthropic-messages";
2832
2837
  provider: string;
2833
2838
  baseUrl: string;
2834
- compat: {
2835
- sendSessionAffinityHeaders: true;
2836
- };
2837
- reasoning: true;
2838
- input: "text"[];
2839
+ reasoning: false;
2840
+ input: ("image" | "text")[];
2839
2841
  cost: {
2840
2842
  input: number;
2841
2843
  output: number;
@@ -2845,17 +2847,14 @@ export declare const MODELS: {
2845
2847
  contextWindow: number;
2846
2848
  maxTokens: number;
2847
2849
  };
2848
- readonly "@cf/openai/gpt-oss-120b": {
2850
+ readonly "claude-3.5-sonnet": {
2849
2851
  id: string;
2850
2852
  name: string;
2851
- api: "openai-completions";
2853
+ api: "anthropic-messages";
2852
2854
  provider: string;
2853
2855
  baseUrl: string;
2854
- compat: {
2855
- sendSessionAffinityHeaders: true;
2856
- };
2857
- reasoning: true;
2858
- input: "text"[];
2856
+ reasoning: false;
2857
+ input: ("image" | "text")[];
2859
2858
  cost: {
2860
2859
  input: number;
2861
2860
  output: number;
@@ -2865,17 +2864,14 @@ export declare const MODELS: {
2865
2864
  contextWindow: number;
2866
2865
  maxTokens: number;
2867
2866
  };
2868
- readonly "@cf/openai/gpt-oss-20b": {
2867
+ readonly "claude-haiku-4-5": {
2869
2868
  id: string;
2870
2869
  name: string;
2871
- api: "openai-completions";
2870
+ api: "anthropic-messages";
2872
2871
  provider: string;
2873
2872
  baseUrl: string;
2874
- compat: {
2875
- sendSessionAffinityHeaders: true;
2876
- };
2877
2873
  reasoning: true;
2878
- input: "text"[];
2874
+ input: ("image" | "text")[];
2879
2875
  cost: {
2880
2876
  input: number;
2881
2877
  output: number;
@@ -2885,17 +2881,14 @@ export declare const MODELS: {
2885
2881
  contextWindow: number;
2886
2882
  maxTokens: number;
2887
2883
  };
2888
- readonly "@cf/zai-org/glm-4.7-flash": {
2884
+ readonly "claude-opus-4": {
2889
2885
  id: string;
2890
2886
  name: string;
2891
- api: "openai-completions";
2887
+ api: "anthropic-messages";
2892
2888
  provider: string;
2893
2889
  baseUrl: string;
2894
- compat: {
2895
- sendSessionAffinityHeaders: true;
2896
- };
2897
2890
  reasoning: true;
2898
- input: "text"[];
2891
+ input: ("image" | "text")[];
2899
2892
  cost: {
2900
2893
  input: number;
2901
2894
  output: number;
@@ -2905,27 +2898,14 @@ export declare const MODELS: {
2905
2898
  contextWindow: number;
2906
2899
  maxTokens: number;
2907
2900
  };
2908
- };
2909
- readonly deepseek: {
2910
- readonly "deepseek-v4-flash": {
2901
+ readonly "claude-opus-4-1": {
2911
2902
  id: string;
2912
2903
  name: string;
2913
- api: "openai-completions";
2904
+ api: "anthropic-messages";
2914
2905
  provider: string;
2915
2906
  baseUrl: string;
2916
- compat: {
2917
- requiresReasoningContentOnAssistantMessages: true;
2918
- thinkingFormat: "deepseek";
2919
- reasoningEffortMap: {
2920
- minimal: string;
2921
- low: string;
2922
- medium: string;
2923
- high: string;
2924
- xhigh: string;
2925
- };
2926
- };
2927
2907
  reasoning: true;
2928
- input: "text"[];
2908
+ input: ("image" | "text")[];
2929
2909
  cost: {
2930
2910
  input: number;
2931
2911
  output: number;
@@ -2935,25 +2915,14 @@ export declare const MODELS: {
2935
2915
  contextWindow: number;
2936
2916
  maxTokens: number;
2937
2917
  };
2938
- readonly "deepseek-v4-pro": {
2918
+ readonly "claude-opus-4-5": {
2939
2919
  id: string;
2940
2920
  name: string;
2941
- api: "openai-completions";
2921
+ api: "anthropic-messages";
2942
2922
  provider: string;
2943
2923
  baseUrl: string;
2944
- compat: {
2945
- requiresReasoningContentOnAssistantMessages: true;
2946
- thinkingFormat: "deepseek";
2947
- reasoningEffortMap: {
2948
- minimal: string;
2949
- low: string;
2950
- medium: string;
2951
- high: string;
2952
- xhigh: string;
2953
- };
2954
- };
2955
2924
  reasoning: true;
2956
- input: "text"[];
2925
+ input: ("image" | "text")[];
2957
2926
  cost: {
2958
2927
  input: number;
2959
2928
  output: number;
@@ -2963,16 +2932,14 @@ export declare const MODELS: {
2963
2932
  contextWindow: number;
2964
2933
  maxTokens: number;
2965
2934
  };
2966
- };
2967
- readonly fireworks: {
2968
- readonly "accounts/fireworks/models/deepseek-v3p1": {
2935
+ readonly "claude-opus-4-6": {
2969
2936
  id: string;
2970
2937
  name: string;
2971
2938
  api: "anthropic-messages";
2972
2939
  provider: string;
2973
2940
  baseUrl: string;
2974
2941
  reasoning: true;
2975
- input: "text"[];
2942
+ input: ("image" | "text")[];
2976
2943
  cost: {
2977
2944
  input: number;
2978
2945
  output: number;
@@ -2982,14 +2949,14 @@ export declare const MODELS: {
2982
2949
  contextWindow: number;
2983
2950
  maxTokens: number;
2984
2951
  };
2985
- readonly "accounts/fireworks/models/deepseek-v3p2": {
2952
+ readonly "claude-opus-4-7": {
2986
2953
  id: string;
2987
2954
  name: string;
2988
2955
  api: "anthropic-messages";
2989
2956
  provider: string;
2990
2957
  baseUrl: string;
2991
2958
  reasoning: true;
2992
- input: "text"[];
2959
+ input: ("image" | "text")[];
2993
2960
  cost: {
2994
2961
  input: number;
2995
2962
  output: number;
@@ -2999,14 +2966,14 @@ export declare const MODELS: {
2999
2966
  contextWindow: number;
3000
2967
  maxTokens: number;
3001
2968
  };
3002
- readonly "accounts/fireworks/models/deepseek-v4-pro": {
2969
+ readonly "claude-sonnet-4": {
3003
2970
  id: string;
3004
2971
  name: string;
3005
2972
  api: "anthropic-messages";
3006
2973
  provider: string;
3007
2974
  baseUrl: string;
3008
2975
  reasoning: true;
3009
- input: "text"[];
2976
+ input: ("image" | "text")[];
3010
2977
  cost: {
3011
2978
  input: number;
3012
2979
  output: number;
@@ -3016,14 +2983,14 @@ export declare const MODELS: {
3016
2983
  contextWindow: number;
3017
2984
  maxTokens: number;
3018
2985
  };
3019
- readonly "accounts/fireworks/models/glm-4p5": {
2986
+ readonly "claude-sonnet-4-5": {
3020
2987
  id: string;
3021
2988
  name: string;
3022
2989
  api: "anthropic-messages";
3023
2990
  provider: string;
3024
2991
  baseUrl: string;
3025
2992
  reasoning: true;
3026
- input: "text"[];
2993
+ input: ("image" | "text")[];
3027
2994
  cost: {
3028
2995
  input: number;
3029
2996
  output: number;
@@ -3033,14 +3000,14 @@ export declare const MODELS: {
3033
3000
  contextWindow: number;
3034
3001
  maxTokens: number;
3035
3002
  };
3036
- readonly "accounts/fireworks/models/glm-4p5-air": {
3003
+ readonly "claude-sonnet-4-6": {
3037
3004
  id: string;
3038
3005
  name: string;
3039
3006
  api: "anthropic-messages";
3040
3007
  provider: string;
3041
3008
  baseUrl: string;
3042
3009
  reasoning: true;
3043
- input: "text"[];
3010
+ input: ("image" | "text")[];
3044
3011
  cost: {
3045
3012
  input: number;
3046
3013
  output: number;
@@ -3050,13 +3017,13 @@ export declare const MODELS: {
3050
3017
  contextWindow: number;
3051
3018
  maxTokens: number;
3052
3019
  };
3053
- readonly "accounts/fireworks/models/glm-4p7": {
3020
+ readonly "gpt-4": {
3054
3021
  id: string;
3055
3022
  name: string;
3056
- api: "anthropic-messages";
3023
+ api: "openai-responses";
3057
3024
  provider: string;
3058
3025
  baseUrl: string;
3059
- reasoning: true;
3026
+ reasoning: false;
3060
3027
  input: "text"[];
3061
3028
  cost: {
3062
3029
  input: number;
@@ -3067,14 +3034,14 @@ export declare const MODELS: {
3067
3034
  contextWindow: number;
3068
3035
  maxTokens: number;
3069
3036
  };
3070
- readonly "accounts/fireworks/models/glm-5": {
3037
+ readonly "gpt-4-turbo": {
3071
3038
  id: string;
3072
3039
  name: string;
3073
- api: "anthropic-messages";
3040
+ api: "openai-responses";
3074
3041
  provider: string;
3075
3042
  baseUrl: string;
3076
- reasoning: true;
3077
- input: "text"[];
3043
+ reasoning: false;
3044
+ input: ("image" | "text")[];
3078
3045
  cost: {
3079
3046
  input: number;
3080
3047
  output: number;
@@ -3084,14 +3051,14 @@ export declare const MODELS: {
3084
3051
  contextWindow: number;
3085
3052
  maxTokens: number;
3086
3053
  };
3087
- readonly "accounts/fireworks/models/glm-5p1": {
3054
+ readonly "gpt-4o": {
3088
3055
  id: string;
3089
3056
  name: string;
3090
- api: "anthropic-messages";
3057
+ api: "openai-responses";
3091
3058
  provider: string;
3092
3059
  baseUrl: string;
3093
- reasoning: true;
3094
- input: "text"[];
3060
+ reasoning: false;
3061
+ input: ("image" | "text")[];
3095
3062
  cost: {
3096
3063
  input: number;
3097
3064
  output: number;
@@ -3101,14 +3068,14 @@ export declare const MODELS: {
3101
3068
  contextWindow: number;
3102
3069
  maxTokens: number;
3103
3070
  };
3104
- readonly "accounts/fireworks/models/gpt-oss-120b": {
3071
+ readonly "gpt-4o-mini": {
3105
3072
  id: string;
3106
3073
  name: string;
3107
- api: "anthropic-messages";
3074
+ api: "openai-responses";
3108
3075
  provider: string;
3109
3076
  baseUrl: string;
3110
- reasoning: true;
3111
- input: "text"[];
3077
+ reasoning: false;
3078
+ input: ("image" | "text")[];
3112
3079
  cost: {
3113
3080
  input: number;
3114
3081
  output: number;
@@ -3118,14 +3085,14 @@ export declare const MODELS: {
3118
3085
  contextWindow: number;
3119
3086
  maxTokens: number;
3120
3087
  };
3121
- readonly "accounts/fireworks/models/gpt-oss-20b": {
3088
+ readonly "gpt-5.1": {
3122
3089
  id: string;
3123
3090
  name: string;
3124
- api: "anthropic-messages";
3091
+ api: "openai-responses";
3125
3092
  provider: string;
3126
3093
  baseUrl: string;
3127
3094
  reasoning: true;
3128
- input: "text"[];
3095
+ input: ("image" | "text")[];
3129
3096
  cost: {
3130
3097
  input: number;
3131
3098
  output: number;
@@ -3135,14 +3102,14 @@ export declare const MODELS: {
3135
3102
  contextWindow: number;
3136
3103
  maxTokens: number;
3137
3104
  };
3138
- readonly "accounts/fireworks/models/kimi-k2-instruct": {
3105
+ readonly "gpt-5.1-codex": {
3139
3106
  id: string;
3140
3107
  name: string;
3141
- api: "anthropic-messages";
3108
+ api: "openai-responses";
3142
3109
  provider: string;
3143
3110
  baseUrl: string;
3144
- reasoning: false;
3145
- input: "text"[];
3111
+ reasoning: true;
3112
+ input: ("image" | "text")[];
3146
3113
  cost: {
3147
3114
  input: number;
3148
3115
  output: number;
@@ -3152,14 +3119,14 @@ export declare const MODELS: {
3152
3119
  contextWindow: number;
3153
3120
  maxTokens: number;
3154
3121
  };
3155
- readonly "accounts/fireworks/models/kimi-k2-thinking": {
3122
+ readonly "gpt-5.2": {
3156
3123
  id: string;
3157
3124
  name: string;
3158
- api: "anthropic-messages";
3125
+ api: "openai-responses";
3159
3126
  provider: string;
3160
3127
  baseUrl: string;
3161
3128
  reasoning: true;
3162
- input: "text"[];
3129
+ input: ("image" | "text")[];
3163
3130
  cost: {
3164
3131
  input: number;
3165
3132
  output: number;
@@ -3169,10 +3136,10 @@ export declare const MODELS: {
3169
3136
  contextWindow: number;
3170
3137
  maxTokens: number;
3171
3138
  };
3172
- readonly "accounts/fireworks/models/kimi-k2p5": {
3139
+ readonly "gpt-5.2-codex": {
3173
3140
  id: string;
3174
3141
  name: string;
3175
- api: "anthropic-messages";
3142
+ api: "openai-responses";
3176
3143
  provider: string;
3177
3144
  baseUrl: string;
3178
3145
  reasoning: true;
@@ -3186,10 +3153,10 @@ export declare const MODELS: {
3186
3153
  contextWindow: number;
3187
3154
  maxTokens: number;
3188
3155
  };
3189
- readonly "accounts/fireworks/models/kimi-k2p6": {
3156
+ readonly "gpt-5.3-codex": {
3190
3157
  id: string;
3191
3158
  name: string;
3192
- api: "anthropic-messages";
3159
+ api: "openai-responses";
3193
3160
  provider: string;
3194
3161
  baseUrl: string;
3195
3162
  reasoning: true;
@@ -3203,14 +3170,14 @@ export declare const MODELS: {
3203
3170
  contextWindow: number;
3204
3171
  maxTokens: number;
3205
3172
  };
3206
- readonly "accounts/fireworks/models/minimax-m2p1": {
3173
+ readonly "gpt-5.4": {
3207
3174
  id: string;
3208
3175
  name: string;
3209
- api: "anthropic-messages";
3176
+ api: "openai-responses";
3210
3177
  provider: string;
3211
3178
  baseUrl: string;
3212
3179
  reasoning: true;
3213
- input: "text"[];
3180
+ input: ("image" | "text")[];
3214
3181
  cost: {
3215
3182
  input: number;
3216
3183
  output: number;
@@ -3220,14 +3187,14 @@ export declare const MODELS: {
3220
3187
  contextWindow: number;
3221
3188
  maxTokens: number;
3222
3189
  };
3223
- readonly "accounts/fireworks/models/minimax-m2p5": {
3190
+ readonly "gpt-5.5": {
3224
3191
  id: string;
3225
3192
  name: string;
3226
- api: "anthropic-messages";
3193
+ api: "openai-responses";
3227
3194
  provider: string;
3228
3195
  baseUrl: string;
3229
3196
  reasoning: true;
3230
- input: "text"[];
3197
+ input: ("image" | "text")[];
3231
3198
  cost: {
3232
3199
  input: number;
3233
3200
  output: number;
@@ -3237,14 +3204,14 @@ export declare const MODELS: {
3237
3204
  contextWindow: number;
3238
3205
  maxTokens: number;
3239
3206
  };
3240
- readonly "accounts/fireworks/models/minimax-m2p7": {
3207
+ readonly o1: {
3241
3208
  id: string;
3242
3209
  name: string;
3243
- api: "anthropic-messages";
3210
+ api: "openai-responses";
3244
3211
  provider: string;
3245
3212
  baseUrl: string;
3246
3213
  reasoning: true;
3247
- input: "text"[];
3214
+ input: ("image" | "text")[];
3248
3215
  cost: {
3249
3216
  input: number;
3250
3217
  output: number;
@@ -3254,10 +3221,10 @@ export declare const MODELS: {
3254
3221
  contextWindow: number;
3255
3222
  maxTokens: number;
3256
3223
  };
3257
- readonly "accounts/fireworks/models/qwen3p6-plus": {
3224
+ readonly o3: {
3258
3225
  id: string;
3259
3226
  name: string;
3260
- api: "anthropic-messages";
3227
+ api: "openai-responses";
3261
3228
  provider: string;
3262
3229
  baseUrl: string;
3263
3230
  reasoning: true;
@@ -3271,14 +3238,14 @@ export declare const MODELS: {
3271
3238
  contextWindow: number;
3272
3239
  maxTokens: number;
3273
3240
  };
3274
- readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
3241
+ readonly "o3-mini": {
3275
3242
  id: string;
3276
3243
  name: string;
3277
- api: "anthropic-messages";
3244
+ api: "openai-responses";
3278
3245
  provider: string;
3279
3246
  baseUrl: string;
3280
3247
  reasoning: true;
3281
- input: ("image" | "text")[];
3248
+ input: "text"[];
3282
3249
  cost: {
3283
3250
  input: number;
3284
3251
  output: number;
@@ -3288,23 +3255,12 @@ export declare const MODELS: {
3288
3255
  contextWindow: number;
3289
3256
  maxTokens: number;
3290
3257
  };
3291
- };
3292
- readonly "github-copilot": {
3293
- readonly "claude-haiku-4.5": {
3258
+ readonly "o3-pro": {
3294
3259
  id: string;
3295
3260
  name: string;
3296
- api: "anthropic-messages";
3261
+ api: "openai-responses";
3297
3262
  provider: string;
3298
3263
  baseUrl: string;
3299
- headers: {
3300
- "User-Agent": string;
3301
- "Editor-Version": string;
3302
- "Editor-Plugin-Version": string;
3303
- "Copilot-Integration-Id": string;
3304
- };
3305
- compat: {
3306
- supportsEagerToolInputStreaming: false;
3307
- };
3308
3264
  reasoning: true;
3309
3265
  input: ("image" | "text")[];
3310
3266
  cost: {
@@ -3316,18 +3272,12 @@ export declare const MODELS: {
3316
3272
  contextWindow: number;
3317
3273
  maxTokens: number;
3318
3274
  };
3319
- readonly "claude-opus-4.5": {
3275
+ readonly "o4-mini": {
3320
3276
  id: string;
3321
3277
  name: string;
3322
- api: "anthropic-messages";
3278
+ api: "openai-responses";
3323
3279
  provider: string;
3324
3280
  baseUrl: string;
3325
- headers: {
3326
- "User-Agent": string;
3327
- "Editor-Version": string;
3328
- "Editor-Plugin-Version": string;
3329
- "Copilot-Integration-Id": string;
3330
- };
3331
3281
  reasoning: true;
3332
3282
  input: ("image" | "text")[];
3333
3283
  cost: {
@@ -3339,10 +3289,686 @@ export declare const MODELS: {
3339
3289
  contextWindow: number;
3340
3290
  maxTokens: number;
3341
3291
  };
3342
- readonly "claude-opus-4.6": {
3292
+ readonly "workers-ai/@cf/moonshotai/kimi-k2.5": {
3343
3293
  id: string;
3344
3294
  name: string;
3345
- api: "anthropic-messages";
3295
+ api: "openai-completions";
3296
+ provider: string;
3297
+ baseUrl: string;
3298
+ compat: {
3299
+ sendSessionAffinityHeaders: true;
3300
+ };
3301
+ reasoning: true;
3302
+ input: ("image" | "text")[];
3303
+ cost: {
3304
+ input: number;
3305
+ output: number;
3306
+ cacheRead: number;
3307
+ cacheWrite: number;
3308
+ };
3309
+ contextWindow: number;
3310
+ maxTokens: number;
3311
+ };
3312
+ readonly "workers-ai/@cf/moonshotai/kimi-k2.6": {
3313
+ id: string;
3314
+ name: string;
3315
+ api: "openai-completions";
3316
+ provider: string;
3317
+ baseUrl: string;
3318
+ compat: {
3319
+ sendSessionAffinityHeaders: true;
3320
+ };
3321
+ reasoning: true;
3322
+ input: ("image" | "text")[];
3323
+ cost: {
3324
+ input: number;
3325
+ output: number;
3326
+ cacheRead: number;
3327
+ cacheWrite: number;
3328
+ };
3329
+ contextWindow: number;
3330
+ maxTokens: number;
3331
+ };
3332
+ readonly "workers-ai/@cf/nvidia/nemotron-3-120b-a12b": {
3333
+ id: string;
3334
+ name: string;
3335
+ api: "openai-completions";
3336
+ provider: string;
3337
+ baseUrl: string;
3338
+ compat: {
3339
+ sendSessionAffinityHeaders: true;
3340
+ };
3341
+ reasoning: true;
3342
+ input: "text"[];
3343
+ cost: {
3344
+ input: number;
3345
+ output: number;
3346
+ cacheRead: number;
3347
+ cacheWrite: number;
3348
+ };
3349
+ contextWindow: number;
3350
+ maxTokens: number;
3351
+ };
3352
+ readonly "workers-ai/@cf/zai-org/glm-4.7-flash": {
3353
+ id: string;
3354
+ name: string;
3355
+ api: "openai-completions";
3356
+ provider: string;
3357
+ baseUrl: string;
3358
+ compat: {
3359
+ sendSessionAffinityHeaders: true;
3360
+ };
3361
+ reasoning: true;
3362
+ input: "text"[];
3363
+ cost: {
3364
+ input: number;
3365
+ output: number;
3366
+ cacheRead: number;
3367
+ cacheWrite: number;
3368
+ };
3369
+ contextWindow: number;
3370
+ maxTokens: number;
3371
+ };
3372
+ };
3373
+ readonly "cloudflare-workers-ai": {
3374
+ readonly "@cf/google/gemma-4-26b-a4b-it": {
3375
+ id: string;
3376
+ name: string;
3377
+ api: "openai-completions";
3378
+ provider: string;
3379
+ baseUrl: string;
3380
+ compat: {
3381
+ sendSessionAffinityHeaders: true;
3382
+ };
3383
+ reasoning: true;
3384
+ input: ("image" | "text")[];
3385
+ cost: {
3386
+ input: number;
3387
+ output: number;
3388
+ cacheRead: number;
3389
+ cacheWrite: number;
3390
+ };
3391
+ contextWindow: number;
3392
+ maxTokens: number;
3393
+ };
3394
+ readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
3395
+ id: string;
3396
+ name: string;
3397
+ api: "openai-completions";
3398
+ provider: string;
3399
+ baseUrl: string;
3400
+ compat: {
3401
+ sendSessionAffinityHeaders: true;
3402
+ };
3403
+ reasoning: false;
3404
+ input: ("image" | "text")[];
3405
+ cost: {
3406
+ input: number;
3407
+ output: number;
3408
+ cacheRead: number;
3409
+ cacheWrite: number;
3410
+ };
3411
+ contextWindow: number;
3412
+ maxTokens: number;
3413
+ };
3414
+ readonly "@cf/moonshotai/kimi-k2.5": {
3415
+ id: string;
3416
+ name: string;
3417
+ api: "openai-completions";
3418
+ provider: string;
3419
+ baseUrl: string;
3420
+ compat: {
3421
+ sendSessionAffinityHeaders: true;
3422
+ };
3423
+ reasoning: true;
3424
+ input: ("image" | "text")[];
3425
+ cost: {
3426
+ input: number;
3427
+ output: number;
3428
+ cacheRead: number;
3429
+ cacheWrite: number;
3430
+ };
3431
+ contextWindow: number;
3432
+ maxTokens: number;
3433
+ };
3434
+ readonly "@cf/moonshotai/kimi-k2.6": {
3435
+ id: string;
3436
+ name: string;
3437
+ api: "openai-completions";
3438
+ provider: string;
3439
+ baseUrl: string;
3440
+ compat: {
3441
+ sendSessionAffinityHeaders: true;
3442
+ };
3443
+ reasoning: true;
3444
+ input: ("image" | "text")[];
3445
+ cost: {
3446
+ input: number;
3447
+ output: number;
3448
+ cacheRead: number;
3449
+ cacheWrite: number;
3450
+ };
3451
+ contextWindow: number;
3452
+ maxTokens: number;
3453
+ };
3454
+ readonly "@cf/nvidia/nemotron-3-120b-a12b": {
3455
+ id: string;
3456
+ name: string;
3457
+ api: "openai-completions";
3458
+ provider: string;
3459
+ baseUrl: string;
3460
+ compat: {
3461
+ sendSessionAffinityHeaders: true;
3462
+ };
3463
+ reasoning: true;
3464
+ input: "text"[];
3465
+ cost: {
3466
+ input: number;
3467
+ output: number;
3468
+ cacheRead: number;
3469
+ cacheWrite: number;
3470
+ };
3471
+ contextWindow: number;
3472
+ maxTokens: number;
3473
+ };
3474
+ readonly "@cf/openai/gpt-oss-120b": {
3475
+ id: string;
3476
+ name: string;
3477
+ api: "openai-completions";
3478
+ provider: string;
3479
+ baseUrl: string;
3480
+ compat: {
3481
+ sendSessionAffinityHeaders: true;
3482
+ };
3483
+ reasoning: true;
3484
+ input: "text"[];
3485
+ cost: {
3486
+ input: number;
3487
+ output: number;
3488
+ cacheRead: number;
3489
+ cacheWrite: number;
3490
+ };
3491
+ contextWindow: number;
3492
+ maxTokens: number;
3493
+ };
3494
+ readonly "@cf/openai/gpt-oss-20b": {
3495
+ id: string;
3496
+ name: string;
3497
+ api: "openai-completions";
3498
+ provider: string;
3499
+ baseUrl: string;
3500
+ compat: {
3501
+ sendSessionAffinityHeaders: true;
3502
+ };
3503
+ reasoning: true;
3504
+ input: "text"[];
3505
+ cost: {
3506
+ input: number;
3507
+ output: number;
3508
+ cacheRead: number;
3509
+ cacheWrite: number;
3510
+ };
3511
+ contextWindow: number;
3512
+ maxTokens: number;
3513
+ };
3514
+ readonly "@cf/zai-org/glm-4.7-flash": {
3515
+ id: string;
3516
+ name: string;
3517
+ api: "openai-completions";
3518
+ provider: string;
3519
+ baseUrl: string;
3520
+ compat: {
3521
+ sendSessionAffinityHeaders: true;
3522
+ };
3523
+ reasoning: true;
3524
+ input: "text"[];
3525
+ cost: {
3526
+ input: number;
3527
+ output: number;
3528
+ cacheRead: number;
3529
+ cacheWrite: number;
3530
+ };
3531
+ contextWindow: number;
3532
+ maxTokens: number;
3533
+ };
3534
+ };
3535
+ readonly deepseek: {
3536
+ readonly "deepseek-v4-flash": {
3537
+ id: string;
3538
+ name: string;
3539
+ api: "openai-completions";
3540
+ provider: string;
3541
+ baseUrl: string;
3542
+ compat: {
3543
+ requiresReasoningContentOnAssistantMessages: true;
3544
+ thinkingFormat: "deepseek";
3545
+ reasoningEffortMap: {
3546
+ minimal: string;
3547
+ low: string;
3548
+ medium: string;
3549
+ high: string;
3550
+ xhigh: string;
3551
+ };
3552
+ };
3553
+ reasoning: true;
3554
+ input: "text"[];
3555
+ cost: {
3556
+ input: number;
3557
+ output: number;
3558
+ cacheRead: number;
3559
+ cacheWrite: number;
3560
+ };
3561
+ contextWindow: number;
3562
+ maxTokens: number;
3563
+ };
3564
+ readonly "deepseek-v4-pro": {
3565
+ id: string;
3566
+ name: string;
3567
+ api: "openai-completions";
3568
+ provider: string;
3569
+ baseUrl: string;
3570
+ compat: {
3571
+ requiresReasoningContentOnAssistantMessages: true;
3572
+ thinkingFormat: "deepseek";
3573
+ reasoningEffortMap: {
3574
+ minimal: string;
3575
+ low: string;
3576
+ medium: string;
3577
+ high: string;
3578
+ xhigh: string;
3579
+ };
3580
+ };
3581
+ reasoning: true;
3582
+ input: "text"[];
3583
+ cost: {
3584
+ input: number;
3585
+ output: number;
3586
+ cacheRead: number;
3587
+ cacheWrite: number;
3588
+ };
3589
+ contextWindow: number;
3590
+ maxTokens: number;
3591
+ };
3592
+ };
3593
+ readonly fireworks: {
3594
+ readonly "accounts/fireworks/models/deepseek-v3p1": {
3595
+ id: string;
3596
+ name: string;
3597
+ api: "anthropic-messages";
3598
+ provider: string;
3599
+ baseUrl: string;
3600
+ reasoning: true;
3601
+ input: "text"[];
3602
+ cost: {
3603
+ input: number;
3604
+ output: number;
3605
+ cacheRead: number;
3606
+ cacheWrite: number;
3607
+ };
3608
+ contextWindow: number;
3609
+ maxTokens: number;
3610
+ };
3611
+ readonly "accounts/fireworks/models/deepseek-v3p2": {
3612
+ id: string;
3613
+ name: string;
3614
+ api: "anthropic-messages";
3615
+ provider: string;
3616
+ baseUrl: string;
3617
+ reasoning: true;
3618
+ input: "text"[];
3619
+ cost: {
3620
+ input: number;
3621
+ output: number;
3622
+ cacheRead: number;
3623
+ cacheWrite: number;
3624
+ };
3625
+ contextWindow: number;
3626
+ maxTokens: number;
3627
+ };
3628
+ readonly "accounts/fireworks/models/deepseek-v4-pro": {
3629
+ id: string;
3630
+ name: string;
3631
+ api: "anthropic-messages";
3632
+ provider: string;
3633
+ baseUrl: string;
3634
+ reasoning: true;
3635
+ input: "text"[];
3636
+ cost: {
3637
+ input: number;
3638
+ output: number;
3639
+ cacheRead: number;
3640
+ cacheWrite: number;
3641
+ };
3642
+ contextWindow: number;
3643
+ maxTokens: number;
3644
+ };
3645
+ readonly "accounts/fireworks/models/glm-4p5": {
3646
+ id: string;
3647
+ name: string;
3648
+ api: "anthropic-messages";
3649
+ provider: string;
3650
+ baseUrl: string;
3651
+ reasoning: true;
3652
+ input: "text"[];
3653
+ cost: {
3654
+ input: number;
3655
+ output: number;
3656
+ cacheRead: number;
3657
+ cacheWrite: number;
3658
+ };
3659
+ contextWindow: number;
3660
+ maxTokens: number;
3661
+ };
3662
+ readonly "accounts/fireworks/models/glm-4p5-air": {
3663
+ id: string;
3664
+ name: string;
3665
+ api: "anthropic-messages";
3666
+ provider: string;
3667
+ baseUrl: string;
3668
+ reasoning: true;
3669
+ input: "text"[];
3670
+ cost: {
3671
+ input: number;
3672
+ output: number;
3673
+ cacheRead: number;
3674
+ cacheWrite: number;
3675
+ };
3676
+ contextWindow: number;
3677
+ maxTokens: number;
3678
+ };
3679
+ readonly "accounts/fireworks/models/glm-4p7": {
3680
+ id: string;
3681
+ name: string;
3682
+ api: "anthropic-messages";
3683
+ provider: string;
3684
+ baseUrl: string;
3685
+ reasoning: true;
3686
+ input: "text"[];
3687
+ cost: {
3688
+ input: number;
3689
+ output: number;
3690
+ cacheRead: number;
3691
+ cacheWrite: number;
3692
+ };
3693
+ contextWindow: number;
3694
+ maxTokens: number;
3695
+ };
3696
+ readonly "accounts/fireworks/models/glm-5": {
3697
+ id: string;
3698
+ name: string;
3699
+ api: "anthropic-messages";
3700
+ provider: string;
3701
+ baseUrl: string;
3702
+ reasoning: true;
3703
+ input: "text"[];
3704
+ cost: {
3705
+ input: number;
3706
+ output: number;
3707
+ cacheRead: number;
3708
+ cacheWrite: number;
3709
+ };
3710
+ contextWindow: number;
3711
+ maxTokens: number;
3712
+ };
3713
+ readonly "accounts/fireworks/models/glm-5p1": {
3714
+ id: string;
3715
+ name: string;
3716
+ api: "anthropic-messages";
3717
+ provider: string;
3718
+ baseUrl: string;
3719
+ reasoning: true;
3720
+ input: "text"[];
3721
+ cost: {
3722
+ input: number;
3723
+ output: number;
3724
+ cacheRead: number;
3725
+ cacheWrite: number;
3726
+ };
3727
+ contextWindow: number;
3728
+ maxTokens: number;
3729
+ };
3730
+ readonly "accounts/fireworks/models/gpt-oss-120b": {
3731
+ id: string;
3732
+ name: string;
3733
+ api: "anthropic-messages";
3734
+ provider: string;
3735
+ baseUrl: string;
3736
+ reasoning: true;
3737
+ input: "text"[];
3738
+ cost: {
3739
+ input: number;
3740
+ output: number;
3741
+ cacheRead: number;
3742
+ cacheWrite: number;
3743
+ };
3744
+ contextWindow: number;
3745
+ maxTokens: number;
3746
+ };
3747
+ readonly "accounts/fireworks/models/gpt-oss-20b": {
3748
+ id: string;
3749
+ name: string;
3750
+ api: "anthropic-messages";
3751
+ provider: string;
3752
+ baseUrl: string;
3753
+ reasoning: true;
3754
+ input: "text"[];
3755
+ cost: {
3756
+ input: number;
3757
+ output: number;
3758
+ cacheRead: number;
3759
+ cacheWrite: number;
3760
+ };
3761
+ contextWindow: number;
3762
+ maxTokens: number;
3763
+ };
3764
+ readonly "accounts/fireworks/models/kimi-k2-instruct": {
3765
+ id: string;
3766
+ name: string;
3767
+ api: "anthropic-messages";
3768
+ provider: string;
3769
+ baseUrl: string;
3770
+ reasoning: false;
3771
+ input: "text"[];
3772
+ cost: {
3773
+ input: number;
3774
+ output: number;
3775
+ cacheRead: number;
3776
+ cacheWrite: number;
3777
+ };
3778
+ contextWindow: number;
3779
+ maxTokens: number;
3780
+ };
3781
+ readonly "accounts/fireworks/models/kimi-k2-thinking": {
3782
+ id: string;
3783
+ name: string;
3784
+ api: "anthropic-messages";
3785
+ provider: string;
3786
+ baseUrl: string;
3787
+ reasoning: true;
3788
+ input: "text"[];
3789
+ cost: {
3790
+ input: number;
3791
+ output: number;
3792
+ cacheRead: number;
3793
+ cacheWrite: number;
3794
+ };
3795
+ contextWindow: number;
3796
+ maxTokens: number;
3797
+ };
3798
+ readonly "accounts/fireworks/models/kimi-k2p5": {
3799
+ id: string;
3800
+ name: string;
3801
+ api: "anthropic-messages";
3802
+ provider: string;
3803
+ baseUrl: string;
3804
+ reasoning: true;
3805
+ input: ("image" | "text")[];
3806
+ cost: {
3807
+ input: number;
3808
+ output: number;
3809
+ cacheRead: number;
3810
+ cacheWrite: number;
3811
+ };
3812
+ contextWindow: number;
3813
+ maxTokens: number;
3814
+ };
3815
+ readonly "accounts/fireworks/models/kimi-k2p6": {
3816
+ id: string;
3817
+ name: string;
3818
+ api: "anthropic-messages";
3819
+ provider: string;
3820
+ baseUrl: string;
3821
+ reasoning: true;
3822
+ input: ("image" | "text")[];
3823
+ cost: {
3824
+ input: number;
3825
+ output: number;
3826
+ cacheRead: number;
3827
+ cacheWrite: number;
3828
+ };
3829
+ contextWindow: number;
3830
+ maxTokens: number;
3831
+ };
3832
+ readonly "accounts/fireworks/models/minimax-m2p1": {
3833
+ id: string;
3834
+ name: string;
3835
+ api: "anthropic-messages";
3836
+ provider: string;
3837
+ baseUrl: string;
3838
+ reasoning: true;
3839
+ input: "text"[];
3840
+ cost: {
3841
+ input: number;
3842
+ output: number;
3843
+ cacheRead: number;
3844
+ cacheWrite: number;
3845
+ };
3846
+ contextWindow: number;
3847
+ maxTokens: number;
3848
+ };
3849
+ readonly "accounts/fireworks/models/minimax-m2p5": {
3850
+ id: string;
3851
+ name: string;
3852
+ api: "anthropic-messages";
3853
+ provider: string;
3854
+ baseUrl: string;
3855
+ reasoning: true;
3856
+ input: "text"[];
3857
+ cost: {
3858
+ input: number;
3859
+ output: number;
3860
+ cacheRead: number;
3861
+ cacheWrite: number;
3862
+ };
3863
+ contextWindow: number;
3864
+ maxTokens: number;
3865
+ };
3866
+ readonly "accounts/fireworks/models/minimax-m2p7": {
3867
+ id: string;
3868
+ name: string;
3869
+ api: "anthropic-messages";
3870
+ provider: string;
3871
+ baseUrl: string;
3872
+ reasoning: true;
3873
+ input: "text"[];
3874
+ cost: {
3875
+ input: number;
3876
+ output: number;
3877
+ cacheRead: number;
3878
+ cacheWrite: number;
3879
+ };
3880
+ contextWindow: number;
3881
+ maxTokens: number;
3882
+ };
3883
+ readonly "accounts/fireworks/models/qwen3p6-plus": {
3884
+ id: string;
3885
+ name: string;
3886
+ api: "anthropic-messages";
3887
+ provider: string;
3888
+ baseUrl: string;
3889
+ reasoning: true;
3890
+ input: ("image" | "text")[];
3891
+ cost: {
3892
+ input: number;
3893
+ output: number;
3894
+ cacheRead: number;
3895
+ cacheWrite: number;
3896
+ };
3897
+ contextWindow: number;
3898
+ maxTokens: number;
3899
+ };
3900
+ readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
3901
+ id: string;
3902
+ name: string;
3903
+ api: "anthropic-messages";
3904
+ provider: string;
3905
+ baseUrl: string;
3906
+ reasoning: true;
3907
+ input: ("image" | "text")[];
3908
+ cost: {
3909
+ input: number;
3910
+ output: number;
3911
+ cacheRead: number;
3912
+ cacheWrite: number;
3913
+ };
3914
+ contextWindow: number;
3915
+ maxTokens: number;
3916
+ };
3917
+ };
3918
+ readonly "github-copilot": {
3919
+ readonly "claude-haiku-4.5": {
3920
+ id: string;
3921
+ name: string;
3922
+ api: "anthropic-messages";
3923
+ provider: string;
3924
+ baseUrl: string;
3925
+ headers: {
3926
+ "User-Agent": string;
3927
+ "Editor-Version": string;
3928
+ "Editor-Plugin-Version": string;
3929
+ "Copilot-Integration-Id": string;
3930
+ };
3931
+ compat: {
3932
+ supportsEagerToolInputStreaming: false;
3933
+ };
3934
+ reasoning: true;
3935
+ input: ("image" | "text")[];
3936
+ cost: {
3937
+ input: number;
3938
+ output: number;
3939
+ cacheRead: number;
3940
+ cacheWrite: number;
3941
+ };
3942
+ contextWindow: number;
3943
+ maxTokens: number;
3944
+ };
3945
+ readonly "claude-opus-4.5": {
3946
+ id: string;
3947
+ name: string;
3948
+ api: "anthropic-messages";
3949
+ provider: string;
3950
+ baseUrl: string;
3951
+ headers: {
3952
+ "User-Agent": string;
3953
+ "Editor-Version": string;
3954
+ "Editor-Plugin-Version": string;
3955
+ "Copilot-Integration-Id": string;
3956
+ };
3957
+ reasoning: true;
3958
+ input: ("image" | "text")[];
3959
+ cost: {
3960
+ input: number;
3961
+ output: number;
3962
+ cacheRead: number;
3963
+ cacheWrite: number;
3964
+ };
3965
+ contextWindow: number;
3966
+ maxTokens: number;
3967
+ };
3968
+ readonly "claude-opus-4.6": {
3969
+ id: string;
3970
+ name: string;
3971
+ api: "anthropic-messages";
3346
3972
  provider: string;
3347
3973
  baseUrl: string;
3348
3974
  headers: {
@@ -4394,14 +5020,14 @@ export declare const MODELS: {
4394
5020
  maxTokens: number;
4395
5021
  };
4396
5022
  };
4397
- readonly "google-antigravity": {
4398
- readonly "claude-opus-4-5-thinking": {
5023
+ readonly "google-vertex": {
5024
+ readonly "gemini-1.5-flash": {
4399
5025
  id: string;
4400
5026
  name: string;
4401
- api: "google-gemini-cli";
5027
+ api: "google-vertex";
4402
5028
  provider: string;
4403
5029
  baseUrl: string;
4404
- reasoning: true;
5030
+ reasoning: false;
4405
5031
  input: ("image" | "text")[];
4406
5032
  cost: {
4407
5033
  input: number;
@@ -4412,13 +5038,13 @@ export declare const MODELS: {
4412
5038
  contextWindow: number;
4413
5039
  maxTokens: number;
4414
5040
  };
4415
- readonly "claude-opus-4-6-thinking": {
5041
+ readonly "gemini-1.5-flash-8b": {
4416
5042
  id: string;
4417
5043
  name: string;
4418
- api: "google-gemini-cli";
5044
+ api: "google-vertex";
4419
5045
  provider: string;
4420
5046
  baseUrl: string;
4421
- reasoning: true;
5047
+ reasoning: false;
4422
5048
  input: ("image" | "text")[];
4423
5049
  cost: {
4424
5050
  input: number;
@@ -4429,10 +5055,10 @@ export declare const MODELS: {
4429
5055
  contextWindow: number;
4430
5056
  maxTokens: number;
4431
5057
  };
4432
- readonly "claude-sonnet-4-5": {
5058
+ readonly "gemini-1.5-pro": {
4433
5059
  id: string;
4434
5060
  name: string;
4435
- api: "google-gemini-cli";
5061
+ api: "google-vertex";
4436
5062
  provider: string;
4437
5063
  baseUrl: string;
4438
5064
  reasoning: false;
@@ -4446,13 +5072,13 @@ export declare const MODELS: {
4446
5072
  contextWindow: number;
4447
5073
  maxTokens: number;
4448
5074
  };
4449
- readonly "claude-sonnet-4-5-thinking": {
5075
+ readonly "gemini-2.0-flash": {
4450
5076
  id: string;
4451
5077
  name: string;
4452
- api: "google-gemini-cli";
5078
+ api: "google-vertex";
4453
5079
  provider: string;
4454
5080
  baseUrl: string;
4455
- reasoning: true;
5081
+ reasoning: false;
4456
5082
  input: ("image" | "text")[];
4457
5083
  cost: {
4458
5084
  input: number;
@@ -4463,10 +5089,10 @@ export declare const MODELS: {
4463
5089
  contextWindow: number;
4464
5090
  maxTokens: number;
4465
5091
  };
4466
- readonly "claude-sonnet-4-6": {
5092
+ readonly "gemini-2.0-flash-lite": {
4467
5093
  id: string;
4468
5094
  name: string;
4469
- api: "google-gemini-cli";
5095
+ api: "google-vertex";
4470
5096
  provider: string;
4471
5097
  baseUrl: string;
4472
5098
  reasoning: true;
@@ -4480,10 +5106,10 @@ export declare const MODELS: {
4480
5106
  contextWindow: number;
4481
5107
  maxTokens: number;
4482
5108
  };
4483
- readonly "gemini-3-flash": {
5109
+ readonly "gemini-2.5-flash": {
4484
5110
  id: string;
4485
5111
  name: string;
4486
- api: "google-gemini-cli";
5112
+ api: "google-vertex";
4487
5113
  provider: string;
4488
5114
  baseUrl: string;
4489
5115
  reasoning: true;
@@ -4497,10 +5123,10 @@ export declare const MODELS: {
4497
5123
  contextWindow: number;
4498
5124
  maxTokens: number;
4499
5125
  };
4500
- readonly "gemini-3.1-pro-high": {
5126
+ readonly "gemini-2.5-flash-lite": {
4501
5127
  id: string;
4502
5128
  name: string;
4503
- api: "google-gemini-cli";
5129
+ api: "google-vertex";
4504
5130
  provider: string;
4505
5131
  baseUrl: string;
4506
5132
  reasoning: true;
@@ -4514,10 +5140,10 @@ export declare const MODELS: {
4514
5140
  contextWindow: number;
4515
5141
  maxTokens: number;
4516
5142
  };
4517
- readonly "gemini-3.1-pro-low": {
5143
+ readonly "gemini-2.5-flash-lite-preview-09-2025": {
4518
5144
  id: string;
4519
5145
  name: string;
4520
- api: "google-gemini-cli";
5146
+ api: "google-vertex";
4521
5147
  provider: string;
4522
5148
  baseUrl: string;
4523
5149
  reasoning: true;
@@ -4531,32 +5157,13 @@ export declare const MODELS: {
4531
5157
  contextWindow: number;
4532
5158
  maxTokens: number;
4533
5159
  };
4534
- readonly "gpt-oss-120b-medium": {
4535
- id: string;
4536
- name: string;
4537
- api: "google-gemini-cli";
4538
- provider: string;
4539
- baseUrl: string;
4540
- reasoning: false;
4541
- input: "text"[];
4542
- cost: {
4543
- input: number;
4544
- output: number;
4545
- cacheRead: number;
4546
- cacheWrite: number;
4547
- };
4548
- contextWindow: number;
4549
- maxTokens: number;
4550
- };
4551
- };
4552
- readonly "google-gemini-cli": {
4553
- readonly "gemini-2.0-flash": {
5160
+ readonly "gemini-2.5-pro": {
4554
5161
  id: string;
4555
5162
  name: string;
4556
- api: "google-gemini-cli";
5163
+ api: "google-vertex";
4557
5164
  provider: string;
4558
5165
  baseUrl: string;
4559
- reasoning: false;
5166
+ reasoning: true;
4560
5167
  input: ("image" | "text")[];
4561
5168
  cost: {
4562
5169
  input: number;
@@ -4567,10 +5174,10 @@ export declare const MODELS: {
4567
5174
  contextWindow: number;
4568
5175
  maxTokens: number;
4569
5176
  };
4570
- readonly "gemini-2.5-flash": {
5177
+ readonly "gemini-3-flash-preview": {
4571
5178
  id: string;
4572
5179
  name: string;
4573
- api: "google-gemini-cli";
5180
+ api: "google-vertex";
4574
5181
  provider: string;
4575
5182
  baseUrl: string;
4576
5183
  reasoning: true;
@@ -4584,10 +5191,10 @@ export declare const MODELS: {
4584
5191
  contextWindow: number;
4585
5192
  maxTokens: number;
4586
5193
  };
4587
- readonly "gemini-2.5-pro": {
5194
+ readonly "gemini-3-pro-preview": {
4588
5195
  id: string;
4589
5196
  name: string;
4590
- api: "google-gemini-cli";
5197
+ api: "google-vertex";
4591
5198
  provider: string;
4592
5199
  baseUrl: string;
4593
5200
  reasoning: true;
@@ -4601,10 +5208,10 @@ export declare const MODELS: {
4601
5208
  contextWindow: number;
4602
5209
  maxTokens: number;
4603
5210
  };
4604
- readonly "gemini-3-flash-preview": {
5211
+ readonly "gemini-3.1-pro-preview": {
4605
5212
  id: string;
4606
5213
  name: string;
4607
- api: "google-gemini-cli";
5214
+ api: "google-vertex";
4608
5215
  provider: string;
4609
5216
  baseUrl: string;
4610
5217
  reasoning: true;
@@ -4618,10 +5225,10 @@ export declare const MODELS: {
4618
5225
  contextWindow: number;
4619
5226
  maxTokens: number;
4620
5227
  };
4621
- readonly "gemini-3-pro-preview": {
5228
+ readonly "gemini-3.1-pro-preview-customtools": {
4622
5229
  id: string;
4623
5230
  name: string;
4624
- api: "google-gemini-cli";
5231
+ api: "google-vertex";
4625
5232
  provider: string;
4626
5233
  baseUrl: string;
4627
5234
  reasoning: true;
@@ -4635,14 +5242,16 @@ export declare const MODELS: {
4635
5242
  contextWindow: number;
4636
5243
  maxTokens: number;
4637
5244
  };
4638
- readonly "gemini-3.1-flash-lite-preview": {
5245
+ };
5246
+ readonly groq: {
5247
+ readonly "deepseek-r1-distill-llama-70b": {
4639
5248
  id: string;
4640
5249
  name: string;
4641
- api: "google-gemini-cli";
5250
+ api: "openai-completions";
4642
5251
  provider: string;
4643
5252
  baseUrl: string;
4644
5253
  reasoning: true;
4645
- input: ("image" | "text")[];
5254
+ input: "text"[];
4646
5255
  cost: {
4647
5256
  input: number;
4648
5257
  output: number;
@@ -4652,14 +5261,14 @@ export declare const MODELS: {
4652
5261
  contextWindow: number;
4653
5262
  maxTokens: number;
4654
5263
  };
4655
- readonly "gemini-3.1-pro-preview": {
5264
+ readonly "gemma2-9b-it": {
4656
5265
  id: string;
4657
5266
  name: string;
4658
- api: "google-gemini-cli";
5267
+ api: "openai-completions";
4659
5268
  provider: string;
4660
5269
  baseUrl: string;
4661
- reasoning: true;
4662
- input: ("image" | "text")[];
5270
+ reasoning: false;
5271
+ input: "text"[];
4663
5272
  cost: {
4664
5273
  input: number;
4665
5274
  output: number;
@@ -4669,16 +5278,14 @@ export declare const MODELS: {
4669
5278
  contextWindow: number;
4670
5279
  maxTokens: number;
4671
5280
  };
4672
- };
4673
- readonly "google-vertex": {
4674
- readonly "gemini-1.5-flash": {
5281
+ readonly "groq/compound": {
4675
5282
  id: string;
4676
5283
  name: string;
4677
- api: "google-vertex";
5284
+ api: "openai-completions";
4678
5285
  provider: string;
4679
5286
  baseUrl: string;
4680
- reasoning: false;
4681
- input: ("image" | "text")[];
5287
+ reasoning: true;
5288
+ input: "text"[];
4682
5289
  cost: {
4683
5290
  input: number;
4684
5291
  output: number;
@@ -4688,14 +5295,14 @@ export declare const MODELS: {
4688
5295
  contextWindow: number;
4689
5296
  maxTokens: number;
4690
5297
  };
4691
- readonly "gemini-1.5-flash-8b": {
5298
+ readonly "groq/compound-mini": {
4692
5299
  id: string;
4693
5300
  name: string;
4694
- api: "google-vertex";
5301
+ api: "openai-completions";
4695
5302
  provider: string;
4696
5303
  baseUrl: string;
4697
- reasoning: false;
4698
- input: ("image" | "text")[];
5304
+ reasoning: true;
5305
+ input: "text"[];
4699
5306
  cost: {
4700
5307
  input: number;
4701
5308
  output: number;
@@ -4705,14 +5312,14 @@ export declare const MODELS: {
4705
5312
  contextWindow: number;
4706
5313
  maxTokens: number;
4707
5314
  };
4708
- readonly "gemini-1.5-pro": {
5315
+ readonly "llama-3.1-8b-instant": {
4709
5316
  id: string;
4710
5317
  name: string;
4711
- api: "google-vertex";
5318
+ api: "openai-completions";
4712
5319
  provider: string;
4713
5320
  baseUrl: string;
4714
5321
  reasoning: false;
4715
- input: ("image" | "text")[];
5322
+ input: "text"[];
4716
5323
  cost: {
4717
5324
  input: number;
4718
5325
  output: number;
@@ -4722,14 +5329,14 @@ export declare const MODELS: {
4722
5329
  contextWindow: number;
4723
5330
  maxTokens: number;
4724
5331
  };
4725
- readonly "gemini-2.0-flash": {
5332
+ readonly "llama-3.3-70b-versatile": {
4726
5333
  id: string;
4727
5334
  name: string;
4728
- api: "google-vertex";
5335
+ api: "openai-completions";
4729
5336
  provider: string;
4730
5337
  baseUrl: string;
4731
5338
  reasoning: false;
4732
- input: ("image" | "text")[];
5339
+ input: "text"[];
4733
5340
  cost: {
4734
5341
  input: number;
4735
5342
  output: number;
@@ -4739,14 +5346,14 @@ export declare const MODELS: {
4739
5346
  contextWindow: number;
4740
5347
  maxTokens: number;
4741
5348
  };
4742
- readonly "gemini-2.0-flash-lite": {
5349
+ readonly "llama3-70b-8192": {
4743
5350
  id: string;
4744
5351
  name: string;
4745
- api: "google-vertex";
5352
+ api: "openai-completions";
4746
5353
  provider: string;
4747
5354
  baseUrl: string;
4748
- reasoning: true;
4749
- input: ("image" | "text")[];
5355
+ reasoning: false;
5356
+ input: "text"[];
4750
5357
  cost: {
4751
5358
  input: number;
4752
5359
  output: number;
@@ -4756,14 +5363,14 @@ export declare const MODELS: {
4756
5363
  contextWindow: number;
4757
5364
  maxTokens: number;
4758
5365
  };
4759
- readonly "gemini-2.5-flash": {
5366
+ readonly "llama3-8b-8192": {
4760
5367
  id: string;
4761
5368
  name: string;
4762
- api: "google-vertex";
5369
+ api: "openai-completions";
4763
5370
  provider: string;
4764
5371
  baseUrl: string;
4765
- reasoning: true;
4766
- input: ("image" | "text")[];
5372
+ reasoning: false;
5373
+ input: "text"[];
4767
5374
  cost: {
4768
5375
  input: number;
4769
5376
  output: number;
@@ -4773,13 +5380,13 @@ export declare const MODELS: {
4773
5380
  contextWindow: number;
4774
5381
  maxTokens: number;
4775
5382
  };
4776
- readonly "gemini-2.5-flash-lite": {
5383
+ readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
4777
5384
  id: string;
4778
5385
  name: string;
4779
- api: "google-vertex";
5386
+ api: "openai-completions";
4780
5387
  provider: string;
4781
5388
  baseUrl: string;
4782
- reasoning: true;
5389
+ reasoning: false;
4783
5390
  input: ("image" | "text")[];
4784
5391
  cost: {
4785
5392
  input: number;
@@ -4790,13 +5397,13 @@ export declare const MODELS: {
4790
5397
  contextWindow: number;
4791
5398
  maxTokens: number;
4792
5399
  };
4793
- readonly "gemini-2.5-flash-lite-preview-09-2025": {
5400
+ readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
4794
5401
  id: string;
4795
5402
  name: string;
4796
- api: "google-vertex";
5403
+ api: "openai-completions";
4797
5404
  provider: string;
4798
5405
  baseUrl: string;
4799
- reasoning: true;
5406
+ reasoning: false;
4800
5407
  input: ("image" | "text")[];
4801
5408
  cost: {
4802
5409
  input: number;
@@ -4807,14 +5414,14 @@ export declare const MODELS: {
4807
5414
  contextWindow: number;
4808
5415
  maxTokens: number;
4809
5416
  };
4810
- readonly "gemini-2.5-pro": {
5417
+ readonly "mistral-saba-24b": {
4811
5418
  id: string;
4812
5419
  name: string;
4813
- api: "google-vertex";
5420
+ api: "openai-completions";
4814
5421
  provider: string;
4815
5422
  baseUrl: string;
4816
- reasoning: true;
4817
- input: ("image" | "text")[];
5423
+ reasoning: false;
5424
+ input: "text"[];
4818
5425
  cost: {
4819
5426
  input: number;
4820
5427
  output: number;
@@ -4824,14 +5431,14 @@ export declare const MODELS: {
4824
5431
  contextWindow: number;
4825
5432
  maxTokens: number;
4826
5433
  };
4827
- readonly "gemini-3-flash-preview": {
5434
+ readonly "moonshotai/kimi-k2-instruct": {
4828
5435
  id: string;
4829
5436
  name: string;
4830
- api: "google-vertex";
5437
+ api: "openai-completions";
4831
5438
  provider: string;
4832
5439
  baseUrl: string;
4833
- reasoning: true;
4834
- input: ("image" | "text")[];
5440
+ reasoning: false;
5441
+ input: "text"[];
4835
5442
  cost: {
4836
5443
  input: number;
4837
5444
  output: number;
@@ -4841,14 +5448,14 @@ export declare const MODELS: {
4841
5448
  contextWindow: number;
4842
5449
  maxTokens: number;
4843
5450
  };
4844
- readonly "gemini-3-pro-preview": {
5451
+ readonly "moonshotai/kimi-k2-instruct-0905": {
4845
5452
  id: string;
4846
5453
  name: string;
4847
- api: "google-vertex";
5454
+ api: "openai-completions";
4848
5455
  provider: string;
4849
5456
  baseUrl: string;
4850
- reasoning: true;
4851
- input: ("image" | "text")[];
5457
+ reasoning: false;
5458
+ input: "text"[];
4852
5459
  cost: {
4853
5460
  input: number;
4854
5461
  output: number;
@@ -4858,14 +5465,14 @@ export declare const MODELS: {
4858
5465
  contextWindow: number;
4859
5466
  maxTokens: number;
4860
5467
  };
4861
- readonly "gemini-3.1-pro-preview": {
5468
+ readonly "openai/gpt-oss-120b": {
4862
5469
  id: string;
4863
5470
  name: string;
4864
- api: "google-vertex";
5471
+ api: "openai-completions";
4865
5472
  provider: string;
4866
5473
  baseUrl: string;
4867
5474
  reasoning: true;
4868
- input: ("image" | "text")[];
5475
+ input: "text"[];
4869
5476
  cost: {
4870
5477
  input: number;
4871
5478
  output: number;
@@ -4875,14 +5482,14 @@ export declare const MODELS: {
4875
5482
  contextWindow: number;
4876
5483
  maxTokens: number;
4877
5484
  };
4878
- readonly "gemini-3.1-pro-preview-customtools": {
5485
+ readonly "openai/gpt-oss-20b": {
4879
5486
  id: string;
4880
5487
  name: string;
4881
- api: "google-vertex";
5488
+ api: "openai-completions";
4882
5489
  provider: string;
4883
5490
  baseUrl: string;
4884
5491
  reasoning: true;
4885
- input: ("image" | "text")[];
5492
+ input: "text"[];
4886
5493
  cost: {
4887
5494
  input: number;
4888
5495
  output: number;
@@ -4892,9 +5499,7 @@ export declare const MODELS: {
4892
5499
  contextWindow: number;
4893
5500
  maxTokens: number;
4894
5501
  };
4895
- };
4896
- readonly groq: {
4897
- readonly "deepseek-r1-distill-llama-70b": {
5502
+ readonly "openai/gpt-oss-safeguard-20b": {
4898
5503
  id: string;
4899
5504
  name: string;
4900
5505
  api: "openai-completions";
@@ -4911,13 +5516,13 @@ export declare const MODELS: {
4911
5516
  contextWindow: number;
4912
5517
  maxTokens: number;
4913
5518
  };
4914
- readonly "gemma2-9b-it": {
5519
+ readonly "qwen-qwq-32b": {
4915
5520
  id: string;
4916
5521
  name: string;
4917
5522
  api: "openai-completions";
4918
5523
  provider: string;
4919
5524
  baseUrl: string;
4920
- reasoning: false;
5525
+ reasoning: true;
4921
5526
  input: "text"[];
4922
5527
  cost: {
4923
5528
  input: number;
@@ -4928,7 +5533,7 @@ export declare const MODELS: {
4928
5533
  contextWindow: number;
4929
5534
  maxTokens: number;
4930
5535
  };
4931
- readonly "groq/compound": {
5536
+ readonly "qwen/qwen3-32b": {
4932
5537
  id: string;
4933
5538
  name: string;
4934
5539
  api: "openai-completions";
@@ -4945,12 +5550,17 @@ export declare const MODELS: {
4945
5550
  contextWindow: number;
4946
5551
  maxTokens: number;
4947
5552
  };
4948
- readonly "groq/compound-mini": {
5553
+ };
5554
+ readonly huggingface: {
5555
+ readonly "MiniMaxAI/MiniMax-M2.1": {
4949
5556
  id: string;
4950
5557
  name: string;
4951
5558
  api: "openai-completions";
4952
5559
  provider: string;
4953
5560
  baseUrl: string;
5561
+ compat: {
5562
+ supportsDeveloperRole: false;
5563
+ };
4954
5564
  reasoning: true;
4955
5565
  input: "text"[];
4956
5566
  cost: {
@@ -4962,13 +5572,16 @@ export declare const MODELS: {
4962
5572
  contextWindow: number;
4963
5573
  maxTokens: number;
4964
5574
  };
4965
- readonly "llama-3.1-8b-instant": {
5575
+ readonly "MiniMaxAI/MiniMax-M2.5": {
4966
5576
  id: string;
4967
5577
  name: string;
4968
5578
  api: "openai-completions";
4969
5579
  provider: string;
4970
5580
  baseUrl: string;
4971
- reasoning: false;
5581
+ compat: {
5582
+ supportsDeveloperRole: false;
5583
+ };
5584
+ reasoning: true;
4972
5585
  input: "text"[];
4973
5586
  cost: {
4974
5587
  input: number;
@@ -4979,13 +5592,16 @@ export declare const MODELS: {
4979
5592
  contextWindow: number;
4980
5593
  maxTokens: number;
4981
5594
  };
4982
- readonly "llama-3.3-70b-versatile": {
5595
+ readonly "MiniMaxAI/MiniMax-M2.7": {
4983
5596
  id: string;
4984
5597
  name: string;
4985
5598
  api: "openai-completions";
4986
5599
  provider: string;
4987
5600
  baseUrl: string;
4988
- reasoning: false;
5601
+ compat: {
5602
+ supportsDeveloperRole: false;
5603
+ };
5604
+ reasoning: true;
4989
5605
  input: "text"[];
4990
5606
  cost: {
4991
5607
  input: number;
@@ -4996,13 +5612,16 @@ export declare const MODELS: {
4996
5612
  contextWindow: number;
4997
5613
  maxTokens: number;
4998
5614
  };
4999
- readonly "llama3-70b-8192": {
5615
+ readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
5000
5616
  id: string;
5001
5617
  name: string;
5002
5618
  api: "openai-completions";
5003
5619
  provider: string;
5004
5620
  baseUrl: string;
5005
- reasoning: false;
5621
+ compat: {
5622
+ supportsDeveloperRole: false;
5623
+ };
5624
+ reasoning: true;
5006
5625
  input: "text"[];
5007
5626
  cost: {
5008
5627
  input: number;
@@ -5013,12 +5632,15 @@ export declare const MODELS: {
5013
5632
  contextWindow: number;
5014
5633
  maxTokens: number;
5015
5634
  };
5016
- readonly "llama3-8b-8192": {
5635
+ readonly "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
5017
5636
  id: string;
5018
5637
  name: string;
5019
5638
  api: "openai-completions";
5020
5639
  provider: string;
5021
5640
  baseUrl: string;
5641
+ compat: {
5642
+ supportsDeveloperRole: false;
5643
+ };
5022
5644
  reasoning: false;
5023
5645
  input: "text"[];
5024
5646
  cost: {
@@ -5030,14 +5652,17 @@ export declare const MODELS: {
5030
5652
  contextWindow: number;
5031
5653
  maxTokens: number;
5032
5654
  };
5033
- readonly "meta-llama/llama-4-maverick-17b-128e-instruct": {
5655
+ readonly "Qwen/Qwen3-Coder-Next": {
5034
5656
  id: string;
5035
5657
  name: string;
5036
5658
  api: "openai-completions";
5037
5659
  provider: string;
5038
5660
  baseUrl: string;
5661
+ compat: {
5662
+ supportsDeveloperRole: false;
5663
+ };
5039
5664
  reasoning: false;
5040
- input: ("image" | "text")[];
5665
+ input: "text"[];
5041
5666
  cost: {
5042
5667
  input: number;
5043
5668
  output: number;
@@ -5047,14 +5672,17 @@ export declare const MODELS: {
5047
5672
  contextWindow: number;
5048
5673
  maxTokens: number;
5049
5674
  };
5050
- readonly "meta-llama/llama-4-scout-17b-16e-instruct": {
5675
+ readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
5051
5676
  id: string;
5052
5677
  name: string;
5053
5678
  api: "openai-completions";
5054
5679
  provider: string;
5055
5680
  baseUrl: string;
5681
+ compat: {
5682
+ supportsDeveloperRole: false;
5683
+ };
5056
5684
  reasoning: false;
5057
- input: ("image" | "text")[];
5685
+ input: "text"[];
5058
5686
  cost: {
5059
5687
  input: number;
5060
5688
  output: number;
@@ -5064,12 +5692,15 @@ export declare const MODELS: {
5064
5692
  contextWindow: number;
5065
5693
  maxTokens: number;
5066
5694
  };
5067
- readonly "mistral-saba-24b": {
5695
+ readonly "Qwen/Qwen3-Next-80B-A3B-Thinking": {
5068
5696
  id: string;
5069
5697
  name: string;
5070
5698
  api: "openai-completions";
5071
5699
  provider: string;
5072
5700
  baseUrl: string;
5701
+ compat: {
5702
+ supportsDeveloperRole: false;
5703
+ };
5073
5704
  reasoning: false;
5074
5705
  input: "text"[];
5075
5706
  cost: {
@@ -5081,14 +5712,17 @@ export declare const MODELS: {
5081
5712
  contextWindow: number;
5082
5713
  maxTokens: number;
5083
5714
  };
5084
- readonly "moonshotai/kimi-k2-instruct": {
5715
+ readonly "Qwen/Qwen3.5-397B-A17B": {
5085
5716
  id: string;
5086
5717
  name: string;
5087
5718
  api: "openai-completions";
5088
5719
  provider: string;
5089
5720
  baseUrl: string;
5090
- reasoning: false;
5091
- input: "text"[];
5721
+ compat: {
5722
+ supportsDeveloperRole: false;
5723
+ };
5724
+ reasoning: true;
5725
+ input: ("image" | "text")[];
5092
5726
  cost: {
5093
5727
  input: number;
5094
5728
  output: number;
@@ -5098,13 +5732,16 @@ export declare const MODELS: {
5098
5732
  contextWindow: number;
5099
5733
  maxTokens: number;
5100
5734
  };
5101
- readonly "moonshotai/kimi-k2-instruct-0905": {
5735
+ readonly "XiaomiMiMo/MiMo-V2-Flash": {
5102
5736
  id: string;
5103
5737
  name: string;
5104
5738
  api: "openai-completions";
5105
5739
  provider: string;
5106
5740
  baseUrl: string;
5107
- reasoning: false;
5741
+ compat: {
5742
+ supportsDeveloperRole: false;
5743
+ };
5744
+ reasoning: true;
5108
5745
  input: "text"[];
5109
5746
  cost: {
5110
5747
  input: number;
@@ -5115,12 +5752,15 @@ export declare const MODELS: {
5115
5752
  contextWindow: number;
5116
5753
  maxTokens: number;
5117
5754
  };
5118
- readonly "openai/gpt-oss-120b": {
5755
+ readonly "deepseek-ai/DeepSeek-R1-0528": {
5119
5756
  id: string;
5120
5757
  name: string;
5121
5758
  api: "openai-completions";
5122
5759
  provider: string;
5123
5760
  baseUrl: string;
5761
+ compat: {
5762
+ supportsDeveloperRole: false;
5763
+ };
5124
5764
  reasoning: true;
5125
5765
  input: "text"[];
5126
5766
  cost: {
@@ -5132,12 +5772,15 @@ export declare const MODELS: {
5132
5772
  contextWindow: number;
5133
5773
  maxTokens: number;
5134
5774
  };
5135
- readonly "openai/gpt-oss-20b": {
5775
+ readonly "deepseek-ai/DeepSeek-V3.2": {
5136
5776
  id: string;
5137
5777
  name: string;
5138
5778
  api: "openai-completions";
5139
5779
  provider: string;
5140
5780
  baseUrl: string;
5781
+ compat: {
5782
+ supportsDeveloperRole: false;
5783
+ };
5141
5784
  reasoning: true;
5142
5785
  input: "text"[];
5143
5786
  cost: {
@@ -5149,12 +5792,15 @@ export declare const MODELS: {
5149
5792
  contextWindow: number;
5150
5793
  maxTokens: number;
5151
5794
  };
5152
- readonly "openai/gpt-oss-safeguard-20b": {
5795
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
5153
5796
  id: string;
5154
5797
  name: string;
5155
5798
  api: "openai-completions";
5156
5799
  provider: string;
5157
5800
  baseUrl: string;
5801
+ compat: {
5802
+ supportsDeveloperRole: false;
5803
+ };
5158
5804
  reasoning: true;
5159
5805
  input: "text"[];
5160
5806
  cost: {
@@ -5166,13 +5812,16 @@ export declare const MODELS: {
5166
5812
  contextWindow: number;
5167
5813
  maxTokens: number;
5168
5814
  };
5169
- readonly "qwen-qwq-32b": {
5815
+ readonly "moonshotai/Kimi-K2-Instruct": {
5170
5816
  id: string;
5171
5817
  name: string;
5172
5818
  api: "openai-completions";
5173
5819
  provider: string;
5174
5820
  baseUrl: string;
5175
- reasoning: true;
5821
+ compat: {
5822
+ supportsDeveloperRole: false;
5823
+ };
5824
+ reasoning: false;
5176
5825
  input: "text"[];
5177
5826
  cost: {
5178
5827
  input: number;
@@ -5183,13 +5832,16 @@ export declare const MODELS: {
5183
5832
  contextWindow: number;
5184
5833
  maxTokens: number;
5185
5834
  };
5186
- readonly "qwen/qwen3-32b": {
5835
+ readonly "moonshotai/Kimi-K2-Instruct-0905": {
5187
5836
  id: string;
5188
5837
  name: string;
5189
5838
  api: "openai-completions";
5190
5839
  provider: string;
5191
5840
  baseUrl: string;
5192
- reasoning: true;
5841
+ compat: {
5842
+ supportsDeveloperRole: false;
5843
+ };
5844
+ reasoning: false;
5193
5845
  input: "text"[];
5194
5846
  cost: {
5195
5847
  input: number;
@@ -5200,9 +5852,7 @@ export declare const MODELS: {
5200
5852
  contextWindow: number;
5201
5853
  maxTokens: number;
5202
5854
  };
5203
- };
5204
- readonly huggingface: {
5205
- readonly "MiniMaxAI/MiniMax-M2.1": {
5855
+ readonly "moonshotai/Kimi-K2-Thinking": {
5206
5856
  id: string;
5207
5857
  name: string;
5208
5858
  api: "openai-completions";
@@ -5222,7 +5872,7 @@ export declare const MODELS: {
5222
5872
  contextWindow: number;
5223
5873
  maxTokens: number;
5224
5874
  };
5225
- readonly "MiniMaxAI/MiniMax-M2.5": {
5875
+ readonly "moonshotai/Kimi-K2.5": {
5226
5876
  id: string;
5227
5877
  name: string;
5228
5878
  api: "openai-completions";
@@ -5232,7 +5882,7 @@ export declare const MODELS: {
5232
5882
  supportsDeveloperRole: false;
5233
5883
  };
5234
5884
  reasoning: true;
5235
- input: "text"[];
5885
+ input: ("image" | "text")[];
5236
5886
  cost: {
5237
5887
  input: number;
5238
5888
  output: number;
@@ -5242,7 +5892,7 @@ export declare const MODELS: {
5242
5892
  contextWindow: number;
5243
5893
  maxTokens: number;
5244
5894
  };
5245
- readonly "MiniMaxAI/MiniMax-M2.7": {
5895
+ readonly "moonshotai/Kimi-K2.6": {
5246
5896
  id: string;
5247
5897
  name: string;
5248
5898
  api: "openai-completions";
@@ -5252,7 +5902,7 @@ export declare const MODELS: {
5252
5902
  supportsDeveloperRole: false;
5253
5903
  };
5254
5904
  reasoning: true;
5255
- input: "text"[];
5905
+ input: ("image" | "text")[];
5256
5906
  cost: {
5257
5907
  input: number;
5258
5908
  output: number;
@@ -5262,7 +5912,7 @@ export declare const MODELS: {
5262
5912
  contextWindow: number;
5263
5913
  maxTokens: number;
5264
5914
  };
5265
- readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
5915
+ readonly "zai-org/GLM-4.7": {
5266
5916
  id: string;
5267
5917
  name: string;
5268
5918
  api: "openai-completions";
@@ -5282,7 +5932,7 @@ export declare const MODELS: {
5282
5932
  contextWindow: number;
5283
5933
  maxTokens: number;
5284
5934
  };
5285
- readonly "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
5935
+ readonly "zai-org/GLM-4.7-Flash": {
5286
5936
  id: string;
5287
5937
  name: string;
5288
5938
  api: "openai-completions";
@@ -5291,7 +5941,7 @@ export declare const MODELS: {
5291
5941
  compat: {
5292
5942
  supportsDeveloperRole: false;
5293
5943
  };
5294
- reasoning: false;
5944
+ reasoning: true;
5295
5945
  input: "text"[];
5296
5946
  cost: {
5297
5947
  input: number;
@@ -5302,7 +5952,7 @@ export declare const MODELS: {
5302
5952
  contextWindow: number;
5303
5953
  maxTokens: number;
5304
5954
  };
5305
- readonly "Qwen/Qwen3-Coder-Next": {
5955
+ readonly "zai-org/GLM-5": {
5306
5956
  id: string;
5307
5957
  name: string;
5308
5958
  api: "openai-completions";
@@ -5311,7 +5961,7 @@ export declare const MODELS: {
5311
5961
  compat: {
5312
5962
  supportsDeveloperRole: false;
5313
5963
  };
5314
- reasoning: false;
5964
+ reasoning: true;
5315
5965
  input: "text"[];
5316
5966
  cost: {
5317
5967
  input: number;
@@ -5322,7 +5972,7 @@ export declare const MODELS: {
5322
5972
  contextWindow: number;
5323
5973
  maxTokens: number;
5324
5974
  };
5325
- readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
5975
+ readonly "zai-org/GLM-5.1": {
5326
5976
  id: string;
5327
5977
  name: string;
5328
5978
  api: "openai-completions";
@@ -5331,7 +5981,7 @@ export declare const MODELS: {
5331
5981
  compat: {
5332
5982
  supportsDeveloperRole: false;
5333
5983
  };
5334
- reasoning: false;
5984
+ reasoning: true;
5335
5985
  input: "text"[];
5336
5986
  cost: {
5337
5987
  input: number;
@@ -5342,17 +5992,19 @@ export declare const MODELS: {
5342
5992
  contextWindow: number;
5343
5993
  maxTokens: number;
5344
5994
  };
5345
- readonly "Qwen/Qwen3-Next-80B-A3B-Thinking": {
5995
+ };
5996
+ readonly "kimi-coding": {
5997
+ readonly k2p6: {
5346
5998
  id: string;
5347
5999
  name: string;
5348
- api: "openai-completions";
6000
+ api: "anthropic-messages";
5349
6001
  provider: string;
5350
6002
  baseUrl: string;
5351
- compat: {
5352
- supportsDeveloperRole: false;
6003
+ headers: {
6004
+ "User-Agent": string;
5353
6005
  };
5354
- reasoning: false;
5355
- input: "text"[];
6006
+ reasoning: true;
6007
+ input: ("image" | "text")[];
5356
6008
  cost: {
5357
6009
  input: number;
5358
6010
  output: number;
@@ -5362,14 +6014,14 @@ export declare const MODELS: {
5362
6014
  contextWindow: number;
5363
6015
  maxTokens: number;
5364
6016
  };
5365
- readonly "Qwen/Qwen3.5-397B-A17B": {
6017
+ readonly "kimi-for-coding": {
5366
6018
  id: string;
5367
6019
  name: string;
5368
- api: "openai-completions";
6020
+ api: "anthropic-messages";
5369
6021
  provider: string;
5370
6022
  baseUrl: string;
5371
- compat: {
5372
- supportsDeveloperRole: false;
6023
+ headers: {
6024
+ "User-Agent": string;
5373
6025
  };
5374
6026
  reasoning: true;
5375
6027
  input: ("image" | "text")[];
@@ -5382,14 +6034,14 @@ export declare const MODELS: {
5382
6034
  contextWindow: number;
5383
6035
  maxTokens: number;
5384
6036
  };
5385
- readonly "XiaomiMiMo/MiMo-V2-Flash": {
6037
+ readonly "kimi-k2-thinking": {
5386
6038
  id: string;
5387
6039
  name: string;
5388
- api: "openai-completions";
6040
+ api: "anthropic-messages";
5389
6041
  provider: string;
5390
6042
  baseUrl: string;
5391
- compat: {
5392
- supportsDeveloperRole: false;
6043
+ headers: {
6044
+ "User-Agent": string;
5393
6045
  };
5394
6046
  reasoning: true;
5395
6047
  input: "text"[];
@@ -5402,15 +6054,14 @@ export declare const MODELS: {
5402
6054
  contextWindow: number;
5403
6055
  maxTokens: number;
5404
6056
  };
5405
- readonly "deepseek-ai/DeepSeek-R1-0528": {
6057
+ };
6058
+ readonly minimax: {
6059
+ readonly "MiniMax-M2.7": {
5406
6060
  id: string;
5407
6061
  name: string;
5408
- api: "openai-completions";
6062
+ api: "anthropic-messages";
5409
6063
  provider: string;
5410
6064
  baseUrl: string;
5411
- compat: {
5412
- supportsDeveloperRole: false;
5413
- };
5414
6065
  reasoning: true;
5415
6066
  input: "text"[];
5416
6067
  cost: {
@@ -5422,15 +6073,12 @@ export declare const MODELS: {
5422
6073
  contextWindow: number;
5423
6074
  maxTokens: number;
5424
6075
  };
5425
- readonly "deepseek-ai/DeepSeek-V3.2": {
6076
+ readonly "MiniMax-M2.7-highspeed": {
5426
6077
  id: string;
5427
6078
  name: string;
5428
- api: "openai-completions";
6079
+ api: "anthropic-messages";
5429
6080
  provider: string;
5430
6081
  baseUrl: string;
5431
- compat: {
5432
- supportsDeveloperRole: false;
5433
- };
5434
6082
  reasoning: true;
5435
6083
  input: "text"[];
5436
6084
  cost: {
@@ -5442,15 +6090,14 @@ export declare const MODELS: {
5442
6090
  contextWindow: number;
5443
6091
  maxTokens: number;
5444
6092
  };
5445
- readonly "deepseek-ai/DeepSeek-V4-Pro": {
6093
+ };
6094
+ readonly "minimax-cn": {
6095
+ readonly "MiniMax-M2.7": {
5446
6096
  id: string;
5447
6097
  name: string;
5448
- api: "openai-completions";
6098
+ api: "anthropic-messages";
5449
6099
  provider: string;
5450
6100
  baseUrl: string;
5451
- compat: {
5452
- supportsDeveloperRole: false;
5453
- };
5454
6101
  reasoning: true;
5455
6102
  input: "text"[];
5456
6103
  cost: {
@@ -5462,16 +6109,13 @@ export declare const MODELS: {
5462
6109
  contextWindow: number;
5463
6110
  maxTokens: number;
5464
6111
  };
5465
- readonly "moonshotai/Kimi-K2-Instruct": {
6112
+ readonly "MiniMax-M2.7-highspeed": {
5466
6113
  id: string;
5467
6114
  name: string;
5468
- api: "openai-completions";
6115
+ api: "anthropic-messages";
5469
6116
  provider: string;
5470
6117
  baseUrl: string;
5471
- compat: {
5472
- supportsDeveloperRole: false;
5473
- };
5474
- reasoning: false;
6118
+ reasoning: true;
5475
6119
  input: "text"[];
5476
6120
  cost: {
5477
6121
  input: number;
@@ -5482,15 +6126,14 @@ export declare const MODELS: {
5482
6126
  contextWindow: number;
5483
6127
  maxTokens: number;
5484
6128
  };
5485
- readonly "moonshotai/Kimi-K2-Instruct-0905": {
6129
+ };
6130
+ readonly mistral: {
6131
+ readonly "codestral-latest": {
5486
6132
  id: string;
5487
6133
  name: string;
5488
- api: "openai-completions";
6134
+ api: "mistral-conversations";
5489
6135
  provider: string;
5490
6136
  baseUrl: string;
5491
- compat: {
5492
- supportsDeveloperRole: false;
5493
- };
5494
6137
  reasoning: false;
5495
6138
  input: "text"[];
5496
6139
  cost: {
@@ -5502,16 +6145,13 @@ export declare const MODELS: {
5502
6145
  contextWindow: number;
5503
6146
  maxTokens: number;
5504
6147
  };
5505
- readonly "moonshotai/Kimi-K2-Thinking": {
6148
+ readonly "devstral-2512": {
5506
6149
  id: string;
5507
6150
  name: string;
5508
- api: "openai-completions";
6151
+ api: "mistral-conversations";
5509
6152
  provider: string;
5510
6153
  baseUrl: string;
5511
- compat: {
5512
- supportsDeveloperRole: false;
5513
- };
5514
- reasoning: true;
6154
+ reasoning: false;
5515
6155
  input: "text"[];
5516
6156
  cost: {
5517
6157
  input: number;
@@ -5522,17 +6162,14 @@ export declare const MODELS: {
5522
6162
  contextWindow: number;
5523
6163
  maxTokens: number;
5524
6164
  };
5525
- readonly "moonshotai/Kimi-K2.5": {
6165
+ readonly "devstral-medium-2507": {
5526
6166
  id: string;
5527
6167
  name: string;
5528
- api: "openai-completions";
6168
+ api: "mistral-conversations";
5529
6169
  provider: string;
5530
6170
  baseUrl: string;
5531
- compat: {
5532
- supportsDeveloperRole: false;
5533
- };
5534
- reasoning: true;
5535
- input: ("image" | "text")[];
6171
+ reasoning: false;
6172
+ input: "text"[];
5536
6173
  cost: {
5537
6174
  input: number;
5538
6175
  output: number;
@@ -5542,17 +6179,14 @@ export declare const MODELS: {
5542
6179
  contextWindow: number;
5543
6180
  maxTokens: number;
5544
6181
  };
5545
- readonly "moonshotai/Kimi-K2.6": {
6182
+ readonly "devstral-medium-latest": {
5546
6183
  id: string;
5547
6184
  name: string;
5548
- api: "openai-completions";
6185
+ api: "mistral-conversations";
5549
6186
  provider: string;
5550
6187
  baseUrl: string;
5551
- compat: {
5552
- supportsDeveloperRole: false;
5553
- };
5554
- reasoning: true;
5555
- input: ("image" | "text")[];
6188
+ reasoning: false;
6189
+ input: "text"[];
5556
6190
  cost: {
5557
6191
  input: number;
5558
6192
  output: number;
@@ -5562,16 +6196,13 @@ export declare const MODELS: {
5562
6196
  contextWindow: number;
5563
6197
  maxTokens: number;
5564
6198
  };
5565
- readonly "zai-org/GLM-4.7": {
6199
+ readonly "devstral-small-2505": {
5566
6200
  id: string;
5567
6201
  name: string;
5568
- api: "openai-completions";
6202
+ api: "mistral-conversations";
5569
6203
  provider: string;
5570
6204
  baseUrl: string;
5571
- compat: {
5572
- supportsDeveloperRole: false;
5573
- };
5574
- reasoning: true;
6205
+ reasoning: false;
5575
6206
  input: "text"[];
5576
6207
  cost: {
5577
6208
  input: number;
@@ -5582,16 +6213,13 @@ export declare const MODELS: {
5582
6213
  contextWindow: number;
5583
6214
  maxTokens: number;
5584
6215
  };
5585
- readonly "zai-org/GLM-4.7-Flash": {
6216
+ readonly "devstral-small-2507": {
5586
6217
  id: string;
5587
6218
  name: string;
5588
- api: "openai-completions";
6219
+ api: "mistral-conversations";
5589
6220
  provider: string;
5590
6221
  baseUrl: string;
5591
- compat: {
5592
- supportsDeveloperRole: false;
5593
- };
5594
- reasoning: true;
6222
+ reasoning: false;
5595
6223
  input: "text"[];
5596
6224
  cost: {
5597
6225
  input: number;
@@ -5602,17 +6230,14 @@ export declare const MODELS: {
5602
6230
  contextWindow: number;
5603
6231
  maxTokens: number;
5604
6232
  };
5605
- readonly "zai-org/GLM-5": {
6233
+ readonly "labs-devstral-small-2512": {
5606
6234
  id: string;
5607
6235
  name: string;
5608
- api: "openai-completions";
6236
+ api: "mistral-conversations";
5609
6237
  provider: string;
5610
6238
  baseUrl: string;
5611
- compat: {
5612
- supportsDeveloperRole: false;
5613
- };
5614
- reasoning: true;
5615
- input: "text"[];
6239
+ reasoning: false;
6240
+ input: ("image" | "text")[];
5616
6241
  cost: {
5617
6242
  input: number;
5618
6243
  output: number;
@@ -5622,15 +6247,12 @@ export declare const MODELS: {
5622
6247
  contextWindow: number;
5623
6248
  maxTokens: number;
5624
6249
  };
5625
- readonly "zai-org/GLM-5.1": {
6250
+ readonly "magistral-medium-latest": {
5626
6251
  id: string;
5627
6252
  name: string;
5628
- api: "openai-completions";
6253
+ api: "mistral-conversations";
5629
6254
  provider: string;
5630
6255
  baseUrl: string;
5631
- compat: {
5632
- supportsDeveloperRole: false;
5633
- };
5634
6256
  reasoning: true;
5635
6257
  input: "text"[];
5636
6258
  cost: {
@@ -5642,19 +6264,14 @@ export declare const MODELS: {
5642
6264
  contextWindow: number;
5643
6265
  maxTokens: number;
5644
6266
  };
5645
- };
5646
- readonly "kimi-coding": {
5647
- readonly k2p6: {
6267
+ readonly "magistral-small": {
5648
6268
  id: string;
5649
6269
  name: string;
5650
- api: "anthropic-messages";
6270
+ api: "mistral-conversations";
5651
6271
  provider: string;
5652
6272
  baseUrl: string;
5653
- headers: {
5654
- "User-Agent": string;
5655
- };
5656
6273
  reasoning: true;
5657
- input: ("image" | "text")[];
6274
+ input: "text"[];
5658
6275
  cost: {
5659
6276
  input: number;
5660
6277
  output: number;
@@ -5664,17 +6281,14 @@ export declare const MODELS: {
5664
6281
  contextWindow: number;
5665
6282
  maxTokens: number;
5666
6283
  };
5667
- readonly "kimi-for-coding": {
6284
+ readonly "ministral-3b-latest": {
5668
6285
  id: string;
5669
6286
  name: string;
5670
- api: "anthropic-messages";
6287
+ api: "mistral-conversations";
5671
6288
  provider: string;
5672
6289
  baseUrl: string;
5673
- headers: {
5674
- "User-Agent": string;
5675
- };
5676
- reasoning: true;
5677
- input: ("image" | "text")[];
6290
+ reasoning: false;
6291
+ input: "text"[];
5678
6292
  cost: {
5679
6293
  input: number;
5680
6294
  output: number;
@@ -5684,16 +6298,13 @@ export declare const MODELS: {
5684
6298
  contextWindow: number;
5685
6299
  maxTokens: number;
5686
6300
  };
5687
- readonly "kimi-k2-thinking": {
6301
+ readonly "ministral-8b-latest": {
5688
6302
  id: string;
5689
6303
  name: string;
5690
- api: "anthropic-messages";
6304
+ api: "mistral-conversations";
5691
6305
  provider: string;
5692
6306
  baseUrl: string;
5693
- headers: {
5694
- "User-Agent": string;
5695
- };
5696
- reasoning: true;
6307
+ reasoning: false;
5697
6308
  input: "text"[];
5698
6309
  cost: {
5699
6310
  input: number;
@@ -5704,15 +6315,13 @@ export declare const MODELS: {
5704
6315
  contextWindow: number;
5705
6316
  maxTokens: number;
5706
6317
  };
5707
- };
5708
- readonly minimax: {
5709
- readonly "MiniMax-M2.7": {
6318
+ readonly "mistral-large-2411": {
5710
6319
  id: string;
5711
6320
  name: string;
5712
- api: "anthropic-messages";
6321
+ api: "mistral-conversations";
5713
6322
  provider: string;
5714
6323
  baseUrl: string;
5715
- reasoning: true;
6324
+ reasoning: false;
5716
6325
  input: "text"[];
5717
6326
  cost: {
5718
6327
  input: number;
@@ -5723,14 +6332,14 @@ export declare const MODELS: {
5723
6332
  contextWindow: number;
5724
6333
  maxTokens: number;
5725
6334
  };
5726
- readonly "MiniMax-M2.7-highspeed": {
6335
+ readonly "mistral-large-2512": {
5727
6336
  id: string;
5728
6337
  name: string;
5729
- api: "anthropic-messages";
6338
+ api: "mistral-conversations";
5730
6339
  provider: string;
5731
6340
  baseUrl: string;
5732
- reasoning: true;
5733
- input: "text"[];
6341
+ reasoning: false;
6342
+ input: ("image" | "text")[];
5734
6343
  cost: {
5735
6344
  input: number;
5736
6345
  output: number;
@@ -5740,16 +6349,14 @@ export declare const MODELS: {
5740
6349
  contextWindow: number;
5741
6350
  maxTokens: number;
5742
6351
  };
5743
- };
5744
- readonly "minimax-cn": {
5745
- readonly "MiniMax-M2.7": {
6352
+ readonly "mistral-large-latest": {
5746
6353
  id: string;
5747
6354
  name: string;
5748
- api: "anthropic-messages";
6355
+ api: "mistral-conversations";
5749
6356
  provider: string;
5750
6357
  baseUrl: string;
5751
- reasoning: true;
5752
- input: "text"[];
6358
+ reasoning: false;
6359
+ input: ("image" | "text")[];
5753
6360
  cost: {
5754
6361
  input: number;
5755
6362
  output: number;
@@ -5759,14 +6366,14 @@ export declare const MODELS: {
5759
6366
  contextWindow: number;
5760
6367
  maxTokens: number;
5761
6368
  };
5762
- readonly "MiniMax-M2.7-highspeed": {
6369
+ readonly "mistral-medium-2505": {
5763
6370
  id: string;
5764
6371
  name: string;
5765
- api: "anthropic-messages";
6372
+ api: "mistral-conversations";
5766
6373
  provider: string;
5767
6374
  baseUrl: string;
5768
- reasoning: true;
5769
- input: "text"[];
6375
+ reasoning: false;
6376
+ input: ("image" | "text")[];
5770
6377
  cost: {
5771
6378
  input: number;
5772
6379
  output: number;
@@ -5776,16 +6383,14 @@ export declare const MODELS: {
5776
6383
  contextWindow: number;
5777
6384
  maxTokens: number;
5778
6385
  };
5779
- };
5780
- readonly mistral: {
5781
- readonly "codestral-latest": {
6386
+ readonly "mistral-medium-2508": {
5782
6387
  id: string;
5783
6388
  name: string;
5784
6389
  api: "mistral-conversations";
5785
6390
  provider: string;
5786
6391
  baseUrl: string;
5787
6392
  reasoning: false;
5788
- input: "text"[];
6393
+ input: ("image" | "text")[];
5789
6394
  cost: {
5790
6395
  input: number;
5791
6396
  output: number;
@@ -5795,14 +6400,14 @@ export declare const MODELS: {
5795
6400
  contextWindow: number;
5796
6401
  maxTokens: number;
5797
6402
  };
5798
- readonly "devstral-2512": {
6403
+ readonly "mistral-medium-2604": {
5799
6404
  id: string;
5800
6405
  name: string;
5801
6406
  api: "mistral-conversations";
5802
6407
  provider: string;
5803
6408
  baseUrl: string;
5804
- reasoning: false;
5805
- input: "text"[];
6409
+ reasoning: true;
6410
+ input: ("image" | "text")[];
5806
6411
  cost: {
5807
6412
  input: number;
5808
6413
  output: number;
@@ -5812,14 +6417,14 @@ export declare const MODELS: {
5812
6417
  contextWindow: number;
5813
6418
  maxTokens: number;
5814
6419
  };
5815
- readonly "devstral-medium-2507": {
6420
+ readonly "mistral-medium-3.5": {
5816
6421
  id: string;
5817
6422
  name: string;
5818
6423
  api: "mistral-conversations";
5819
6424
  provider: string;
5820
6425
  baseUrl: string;
5821
- reasoning: false;
5822
- input: "text"[];
6426
+ reasoning: true;
6427
+ input: ("image" | "text")[];
5823
6428
  cost: {
5824
6429
  input: number;
5825
6430
  output: number;
@@ -5829,14 +6434,14 @@ export declare const MODELS: {
5829
6434
  contextWindow: number;
5830
6435
  maxTokens: number;
5831
6436
  };
5832
- readonly "devstral-medium-latest": {
6437
+ readonly "mistral-medium-latest": {
5833
6438
  id: string;
5834
6439
  name: string;
5835
6440
  api: "mistral-conversations";
5836
6441
  provider: string;
5837
6442
  baseUrl: string;
5838
- reasoning: false;
5839
- input: "text"[];
6443
+ reasoning: true;
6444
+ input: ("image" | "text")[];
5840
6445
  cost: {
5841
6446
  input: number;
5842
6447
  output: number;
@@ -5846,7 +6451,7 @@ export declare const MODELS: {
5846
6451
  contextWindow: number;
5847
6452
  maxTokens: number;
5848
6453
  };
5849
- readonly "devstral-small-2505": {
6454
+ readonly "mistral-nemo": {
5850
6455
  id: string;
5851
6456
  name: string;
5852
6457
  api: "mistral-conversations";
@@ -5863,14 +6468,14 @@ export declare const MODELS: {
5863
6468
  contextWindow: number;
5864
6469
  maxTokens: number;
5865
6470
  };
5866
- readonly "devstral-small-2507": {
6471
+ readonly "mistral-small-2506": {
5867
6472
  id: string;
5868
6473
  name: string;
5869
6474
  api: "mistral-conversations";
5870
6475
  provider: string;
5871
6476
  baseUrl: string;
5872
6477
  reasoning: false;
5873
- input: "text"[];
6478
+ input: ("image" | "text")[];
5874
6479
  cost: {
5875
6480
  input: number;
5876
6481
  output: number;
@@ -5880,13 +6485,13 @@ export declare const MODELS: {
5880
6485
  contextWindow: number;
5881
6486
  maxTokens: number;
5882
6487
  };
5883
- readonly "labs-devstral-small-2512": {
6488
+ readonly "mistral-small-2603": {
5884
6489
  id: string;
5885
6490
  name: string;
5886
6491
  api: "mistral-conversations";
5887
6492
  provider: string;
5888
6493
  baseUrl: string;
5889
- reasoning: false;
6494
+ reasoning: true;
5890
6495
  input: ("image" | "text")[];
5891
6496
  cost: {
5892
6497
  input: number;
@@ -5897,14 +6502,14 @@ export declare const MODELS: {
5897
6502
  contextWindow: number;
5898
6503
  maxTokens: number;
5899
6504
  };
5900
- readonly "magistral-medium-latest": {
6505
+ readonly "mistral-small-latest": {
5901
6506
  id: string;
5902
6507
  name: string;
5903
6508
  api: "mistral-conversations";
5904
6509
  provider: string;
5905
6510
  baseUrl: string;
5906
6511
  reasoning: true;
5907
- input: "text"[];
6512
+ input: ("image" | "text")[];
5908
6513
  cost: {
5909
6514
  input: number;
5910
6515
  output: number;
@@ -5914,13 +6519,13 @@ export declare const MODELS: {
5914
6519
  contextWindow: number;
5915
6520
  maxTokens: number;
5916
6521
  };
5917
- readonly "magistral-small": {
6522
+ readonly "open-mistral-7b": {
5918
6523
  id: string;
5919
6524
  name: string;
5920
6525
  api: "mistral-conversations";
5921
6526
  provider: string;
5922
6527
  baseUrl: string;
5923
- reasoning: true;
6528
+ reasoning: false;
5924
6529
  input: "text"[];
5925
6530
  cost: {
5926
6531
  input: number;
@@ -5931,7 +6536,7 @@ export declare const MODELS: {
5931
6536
  contextWindow: number;
5932
6537
  maxTokens: number;
5933
6538
  };
5934
- readonly "ministral-3b-latest": {
6539
+ readonly "open-mixtral-8x22b": {
5935
6540
  id: string;
5936
6541
  name: string;
5937
6542
  api: "mistral-conversations";
@@ -5948,7 +6553,7 @@ export declare const MODELS: {
5948
6553
  contextWindow: number;
5949
6554
  maxTokens: number;
5950
6555
  };
5951
- readonly "ministral-8b-latest": {
6556
+ readonly "open-mixtral-8x7b": {
5952
6557
  id: string;
5953
6558
  name: string;
5954
6559
  api: "mistral-conversations";
@@ -5965,14 +6570,14 @@ export declare const MODELS: {
5965
6570
  contextWindow: number;
5966
6571
  maxTokens: number;
5967
6572
  };
5968
- readonly "mistral-large-2411": {
6573
+ readonly "pixtral-12b": {
5969
6574
  id: string;
5970
6575
  name: string;
5971
6576
  api: "mistral-conversations";
5972
6577
  provider: string;
5973
6578
  baseUrl: string;
5974
6579
  reasoning: false;
5975
- input: "text"[];
6580
+ input: ("image" | "text")[];
5976
6581
  cost: {
5977
6582
  input: number;
5978
6583
  output: number;
@@ -5982,7 +6587,7 @@ export declare const MODELS: {
5982
6587
  contextWindow: number;
5983
6588
  maxTokens: number;
5984
6589
  };
5985
- readonly "mistral-large-2512": {
6590
+ readonly "pixtral-large-latest": {
5986
6591
  id: string;
5987
6592
  name: string;
5988
6593
  api: "mistral-conversations";
@@ -5999,14 +6604,23 @@ export declare const MODELS: {
5999
6604
  contextWindow: number;
6000
6605
  maxTokens: number;
6001
6606
  };
6002
- readonly "mistral-large-latest": {
6607
+ };
6608
+ readonly moonshotai: {
6609
+ readonly "kimi-k2-0711-preview": {
6003
6610
  id: string;
6004
6611
  name: string;
6005
- api: "mistral-conversations";
6612
+ api: "openai-completions";
6006
6613
  provider: string;
6007
6614
  baseUrl: string;
6615
+ compat: {
6616
+ supportsStore: false;
6617
+ supportsDeveloperRole: false;
6618
+ supportsReasoningEffort: false;
6619
+ maxTokensField: "max_tokens";
6620
+ supportsStrictMode: false;
6621
+ };
6008
6622
  reasoning: false;
6009
- input: ("image" | "text")[];
6623
+ input: "text"[];
6010
6624
  cost: {
6011
6625
  input: number;
6012
6626
  output: number;
@@ -6016,14 +6630,21 @@ export declare const MODELS: {
6016
6630
  contextWindow: number;
6017
6631
  maxTokens: number;
6018
6632
  };
6019
- readonly "mistral-medium-2505": {
6633
+ readonly "kimi-k2-0905-preview": {
6020
6634
  id: string;
6021
6635
  name: string;
6022
- api: "mistral-conversations";
6636
+ api: "openai-completions";
6023
6637
  provider: string;
6024
6638
  baseUrl: string;
6639
+ compat: {
6640
+ supportsStore: false;
6641
+ supportsDeveloperRole: false;
6642
+ supportsReasoningEffort: false;
6643
+ maxTokensField: "max_tokens";
6644
+ supportsStrictMode: false;
6645
+ };
6025
6646
  reasoning: false;
6026
- input: ("image" | "text")[];
6647
+ input: "text"[];
6027
6648
  cost: {
6028
6649
  input: number;
6029
6650
  output: number;
@@ -6033,14 +6654,21 @@ export declare const MODELS: {
6033
6654
  contextWindow: number;
6034
6655
  maxTokens: number;
6035
6656
  };
6036
- readonly "mistral-medium-2508": {
6657
+ readonly "kimi-k2-thinking": {
6037
6658
  id: string;
6038
6659
  name: string;
6039
- api: "mistral-conversations";
6660
+ api: "openai-completions";
6040
6661
  provider: string;
6041
6662
  baseUrl: string;
6042
- reasoning: false;
6043
- input: ("image" | "text")[];
6663
+ compat: {
6664
+ supportsStore: false;
6665
+ supportsDeveloperRole: false;
6666
+ supportsReasoningEffort: false;
6667
+ maxTokensField: "max_tokens";
6668
+ supportsStrictMode: false;
6669
+ };
6670
+ reasoning: true;
6671
+ input: "text"[];
6044
6672
  cost: {
6045
6673
  input: number;
6046
6674
  output: number;
@@ -6050,14 +6678,21 @@ export declare const MODELS: {
6050
6678
  contextWindow: number;
6051
6679
  maxTokens: number;
6052
6680
  };
6053
- readonly "mistral-medium-latest": {
6681
+ readonly "kimi-k2-thinking-turbo": {
6054
6682
  id: string;
6055
6683
  name: string;
6056
- api: "mistral-conversations";
6684
+ api: "openai-completions";
6057
6685
  provider: string;
6058
6686
  baseUrl: string;
6059
- reasoning: false;
6060
- input: ("image" | "text")[];
6687
+ compat: {
6688
+ supportsStore: false;
6689
+ supportsDeveloperRole: false;
6690
+ supportsReasoningEffort: false;
6691
+ maxTokensField: "max_tokens";
6692
+ supportsStrictMode: false;
6693
+ };
6694
+ reasoning: true;
6695
+ input: "text"[];
6061
6696
  cost: {
6062
6697
  input: number;
6063
6698
  output: number;
@@ -6067,12 +6702,19 @@ export declare const MODELS: {
6067
6702
  contextWindow: number;
6068
6703
  maxTokens: number;
6069
6704
  };
6070
- readonly "mistral-nemo": {
6705
+ readonly "kimi-k2-turbo-preview": {
6071
6706
  id: string;
6072
6707
  name: string;
6073
- api: "mistral-conversations";
6708
+ api: "openai-completions";
6074
6709
  provider: string;
6075
6710
  baseUrl: string;
6711
+ compat: {
6712
+ supportsStore: false;
6713
+ supportsDeveloperRole: false;
6714
+ supportsReasoningEffort: false;
6715
+ maxTokensField: "max_tokens";
6716
+ supportsStrictMode: false;
6717
+ };
6076
6718
  reasoning: false;
6077
6719
  input: "text"[];
6078
6720
  cost: {
@@ -6084,13 +6726,20 @@ export declare const MODELS: {
6084
6726
  contextWindow: number;
6085
6727
  maxTokens: number;
6086
6728
  };
6087
- readonly "mistral-small-2506": {
6729
+ readonly "kimi-k2.5": {
6088
6730
  id: string;
6089
6731
  name: string;
6090
- api: "mistral-conversations";
6732
+ api: "openai-completions";
6091
6733
  provider: string;
6092
6734
  baseUrl: string;
6093
- reasoning: false;
6735
+ compat: {
6736
+ supportsStore: false;
6737
+ supportsDeveloperRole: false;
6738
+ supportsReasoningEffort: false;
6739
+ maxTokensField: "max_tokens";
6740
+ supportsStrictMode: false;
6741
+ };
6742
+ reasoning: true;
6094
6743
  input: ("image" | "text")[];
6095
6744
  cost: {
6096
6745
  input: number;
@@ -6101,12 +6750,19 @@ export declare const MODELS: {
6101
6750
  contextWindow: number;
6102
6751
  maxTokens: number;
6103
6752
  };
6104
- readonly "mistral-small-2603": {
6753
+ readonly "kimi-k2.6": {
6105
6754
  id: string;
6106
6755
  name: string;
6107
- api: "mistral-conversations";
6756
+ api: "openai-completions";
6108
6757
  provider: string;
6109
6758
  baseUrl: string;
6759
+ compat: {
6760
+ supportsStore: false;
6761
+ supportsDeveloperRole: false;
6762
+ supportsReasoningEffort: false;
6763
+ maxTokensField: "max_tokens";
6764
+ supportsStrictMode: false;
6765
+ };
6110
6766
  reasoning: true;
6111
6767
  input: ("image" | "text")[];
6112
6768
  cost: {
@@ -6118,14 +6774,47 @@ export declare const MODELS: {
6118
6774
  contextWindow: number;
6119
6775
  maxTokens: number;
6120
6776
  };
6121
- readonly "mistral-small-latest": {
6777
+ };
6778
+ readonly "moonshotai-cn": {
6779
+ readonly "kimi-k2-0711-preview": {
6122
6780
  id: string;
6123
6781
  name: string;
6124
- api: "mistral-conversations";
6782
+ api: "openai-completions";
6125
6783
  provider: string;
6126
6784
  baseUrl: string;
6127
- reasoning: true;
6128
- input: ("image" | "text")[];
6785
+ compat: {
6786
+ supportsStore: false;
6787
+ supportsDeveloperRole: false;
6788
+ supportsReasoningEffort: false;
6789
+ maxTokensField: "max_tokens";
6790
+ supportsStrictMode: false;
6791
+ };
6792
+ reasoning: false;
6793
+ input: "text"[];
6794
+ cost: {
6795
+ input: number;
6796
+ output: number;
6797
+ cacheRead: number;
6798
+ cacheWrite: number;
6799
+ };
6800
+ contextWindow: number;
6801
+ maxTokens: number;
6802
+ };
6803
+ readonly "kimi-k2-0905-preview": {
6804
+ id: string;
6805
+ name: string;
6806
+ api: "openai-completions";
6807
+ provider: string;
6808
+ baseUrl: string;
6809
+ compat: {
6810
+ supportsStore: false;
6811
+ supportsDeveloperRole: false;
6812
+ supportsReasoningEffort: false;
6813
+ maxTokensField: "max_tokens";
6814
+ supportsStrictMode: false;
6815
+ };
6816
+ reasoning: false;
6817
+ input: "text"[];
6129
6818
  cost: {
6130
6819
  input: number;
6131
6820
  output: number;
@@ -6135,13 +6824,20 @@ export declare const MODELS: {
6135
6824
  contextWindow: number;
6136
6825
  maxTokens: number;
6137
6826
  };
6138
- readonly "open-mistral-7b": {
6827
+ readonly "kimi-k2-thinking": {
6139
6828
  id: string;
6140
6829
  name: string;
6141
- api: "mistral-conversations";
6830
+ api: "openai-completions";
6142
6831
  provider: string;
6143
6832
  baseUrl: string;
6144
- reasoning: false;
6833
+ compat: {
6834
+ supportsStore: false;
6835
+ supportsDeveloperRole: false;
6836
+ supportsReasoningEffort: false;
6837
+ maxTokensField: "max_tokens";
6838
+ supportsStrictMode: false;
6839
+ };
6840
+ reasoning: true;
6145
6841
  input: "text"[];
6146
6842
  cost: {
6147
6843
  input: number;
@@ -6152,13 +6848,20 @@ export declare const MODELS: {
6152
6848
  contextWindow: number;
6153
6849
  maxTokens: number;
6154
6850
  };
6155
- readonly "open-mixtral-8x22b": {
6851
+ readonly "kimi-k2-thinking-turbo": {
6156
6852
  id: string;
6157
6853
  name: string;
6158
- api: "mistral-conversations";
6854
+ api: "openai-completions";
6159
6855
  provider: string;
6160
6856
  baseUrl: string;
6161
- reasoning: false;
6857
+ compat: {
6858
+ supportsStore: false;
6859
+ supportsDeveloperRole: false;
6860
+ supportsReasoningEffort: false;
6861
+ maxTokensField: "max_tokens";
6862
+ supportsStrictMode: false;
6863
+ };
6864
+ reasoning: true;
6162
6865
  input: "text"[];
6163
6866
  cost: {
6164
6867
  input: number;
@@ -6169,12 +6872,19 @@ export declare const MODELS: {
6169
6872
  contextWindow: number;
6170
6873
  maxTokens: number;
6171
6874
  };
6172
- readonly "open-mixtral-8x7b": {
6875
+ readonly "kimi-k2-turbo-preview": {
6173
6876
  id: string;
6174
6877
  name: string;
6175
- api: "mistral-conversations";
6878
+ api: "openai-completions";
6176
6879
  provider: string;
6177
6880
  baseUrl: string;
6881
+ compat: {
6882
+ supportsStore: false;
6883
+ supportsDeveloperRole: false;
6884
+ supportsReasoningEffort: false;
6885
+ maxTokensField: "max_tokens";
6886
+ supportsStrictMode: false;
6887
+ };
6178
6888
  reasoning: false;
6179
6889
  input: "text"[];
6180
6890
  cost: {
@@ -6186,13 +6896,20 @@ export declare const MODELS: {
6186
6896
  contextWindow: number;
6187
6897
  maxTokens: number;
6188
6898
  };
6189
- readonly "pixtral-12b": {
6899
+ readonly "kimi-k2.5": {
6190
6900
  id: string;
6191
6901
  name: string;
6192
- api: "mistral-conversations";
6902
+ api: "openai-completions";
6193
6903
  provider: string;
6194
6904
  baseUrl: string;
6195
- reasoning: false;
6905
+ compat: {
6906
+ supportsStore: false;
6907
+ supportsDeveloperRole: false;
6908
+ supportsReasoningEffort: false;
6909
+ maxTokensField: "max_tokens";
6910
+ supportsStrictMode: false;
6911
+ };
6912
+ reasoning: true;
6196
6913
  input: ("image" | "text")[];
6197
6914
  cost: {
6198
6915
  input: number;
@@ -6203,13 +6920,20 @@ export declare const MODELS: {
6203
6920
  contextWindow: number;
6204
6921
  maxTokens: number;
6205
6922
  };
6206
- readonly "pixtral-large-latest": {
6923
+ readonly "kimi-k2.6": {
6207
6924
  id: string;
6208
6925
  name: string;
6209
- api: "mistral-conversations";
6926
+ api: "openai-completions";
6210
6927
  provider: string;
6211
6928
  baseUrl: string;
6212
- reasoning: false;
6929
+ compat: {
6930
+ supportsStore: false;
6931
+ supportsDeveloperRole: false;
6932
+ supportsReasoningEffort: false;
6933
+ maxTokensField: "max_tokens";
6934
+ supportsStrictMode: false;
6935
+ };
6936
+ reasoning: true;
6213
6937
  input: ("image" | "text")[];
6214
6938
  cost: {
6215
6939
  input: number;
@@ -6783,6 +7507,23 @@ export declare const MODELS: {
6783
7507
  contextWindow: number;
6784
7508
  maxTokens: number;
6785
7509
  };
7510
+ readonly "gpt-5.5-pro": {
7511
+ id: string;
7512
+ name: string;
7513
+ api: "openai-responses";
7514
+ provider: string;
7515
+ baseUrl: string;
7516
+ reasoning: true;
7517
+ input: ("image" | "text")[];
7518
+ cost: {
7519
+ input: number;
7520
+ output: number;
7521
+ cacheRead: number;
7522
+ cacheWrite: number;
7523
+ };
7524
+ contextWindow: number;
7525
+ maxTokens: number;
7526
+ };
6786
7527
  readonly o1: {
6787
7528
  id: string;
6788
7529
  name: string;
@@ -7637,23 +8378,6 @@ export declare const MODELS: {
7637
8378
  contextWindow: number;
7638
8379
  maxTokens: number;
7639
8380
  };
7640
- readonly "ling-2.6-flash-free": {
7641
- id: string;
7642
- name: string;
7643
- api: "openai-completions";
7644
- provider: string;
7645
- baseUrl: string;
7646
- reasoning: false;
7647
- input: "text"[];
7648
- cost: {
7649
- input: number;
7650
- output: number;
7651
- cacheRead: number;
7652
- cacheWrite: number;
7653
- };
7654
- contextWindow: number;
7655
- maxTokens: number;
7656
- };
7657
8381
  readonly "minimax-m2.5": {
7658
8382
  id: string;
7659
8383
  name: string;
@@ -7766,6 +8490,14 @@ export declare const MODELS: {
7766
8490
  baseUrl: string;
7767
8491
  compat: {
7768
8492
  requiresReasoningContentOnAssistantMessages: true;
8493
+ thinkingFormat: "deepseek";
8494
+ reasoningEffortMap: {
8495
+ minimal: string;
8496
+ low: string;
8497
+ medium: string;
8498
+ high: string;
8499
+ xhigh: string;
8500
+ };
7769
8501
  };
7770
8502
  reasoning: true;
7771
8503
  input: "text"[];
@@ -7786,6 +8518,14 @@ export declare const MODELS: {
7786
8518
  baseUrl: string;
7787
8519
  compat: {
7788
8520
  requiresReasoningContentOnAssistantMessages: true;
8521
+ thinkingFormat: "deepseek";
8522
+ reasoningEffortMap: {
8523
+ minimal: string;
8524
+ low: string;
8525
+ medium: string;
8526
+ high: string;
8527
+ xhigh: string;
8528
+ };
7789
8529
  };
7790
8530
  reasoning: true;
7791
8531
  input: "text"[];
@@ -8743,6 +9483,13 @@ export declare const MODELS: {
8743
9483
  baseUrl: string;
8744
9484
  compat: {
8745
9485
  requiresReasoningContentOnAssistantMessages: true;
9486
+ reasoningEffortMap: {
9487
+ minimal: string;
9488
+ low: string;
9489
+ medium: string;
9490
+ high: string;
9491
+ xhigh: string;
9492
+ };
8746
9493
  };
8747
9494
  reasoning: true;
8748
9495
  input: "text"[];
@@ -8763,6 +9510,13 @@ export declare const MODELS: {
8763
9510
  baseUrl: string;
8764
9511
  compat: {
8765
9512
  requiresReasoningContentOnAssistantMessages: true;
9513
+ reasoningEffortMap: {
9514
+ minimal: string;
9515
+ low: string;
9516
+ medium: string;
9517
+ high: string;
9518
+ xhigh: string;
9519
+ };
8766
9520
  };
8767
9521
  reasoning: true;
8768
9522
  input: "text"[];
@@ -9098,6 +9852,23 @@ export declare const MODELS: {
9098
9852
  contextWindow: number;
9099
9853
  maxTokens: number;
9100
9854
  };
9855
+ readonly "ibm-granite/granite-4.1-8b": {
9856
+ id: string;
9857
+ name: string;
9858
+ api: "openai-completions";
9859
+ provider: string;
9860
+ baseUrl: string;
9861
+ reasoning: false;
9862
+ input: "text"[];
9863
+ cost: {
9864
+ input: number;
9865
+ output: number;
9866
+ cacheRead: number;
9867
+ cacheWrite: number;
9868
+ };
9869
+ contextWindow: number;
9870
+ maxTokens: number;
9871
+ };
9101
9872
  readonly "inception/mercury-2": {
9102
9873
  id: string;
9103
9874
  name: string;
@@ -9132,7 +9903,7 @@ export declare const MODELS: {
9132
9903
  contextWindow: number;
9133
9904
  maxTokens: number;
9134
9905
  };
9135
- readonly "inclusionai/ling-2.6-flash:free": {
9906
+ readonly "inclusionai/ling-2.6-flash": {
9136
9907
  id: string;
9137
9908
  name: string;
9138
9909
  api: "openai-completions";
@@ -9659,23 +10430,6 @@ export declare const MODELS: {
9659
10430
  contextWindow: number;
9660
10431
  maxTokens: number;
9661
10432
  };
9662
- readonly "mistralai/mistral-small-creative": {
9663
- id: string;
9664
- name: string;
9665
- api: "openai-completions";
9666
- provider: string;
9667
- baseUrl: string;
9668
- reasoning: false;
9669
- input: "text"[];
9670
- cost: {
9671
- input: number;
9672
- output: number;
9673
- cacheRead: number;
9674
- cacheWrite: number;
9675
- };
9676
- contextWindow: number;
9677
- maxTokens: number;
9678
- };
9679
10433
  readonly "mistralai/mixtral-8x22b-instruct": {
9680
10434
  id: string;
9681
10435
  name: string;
@@ -11002,6 +11756,23 @@ export declare const MODELS: {
11002
11756
  contextWindow: number;
11003
11757
  maxTokens: number;
11004
11758
  };
11759
+ readonly "openrouter/owl-alpha": {
11760
+ id: string;
11761
+ name: string;
11762
+ api: "openai-completions";
11763
+ provider: string;
11764
+ baseUrl: string;
11765
+ reasoning: false;
11766
+ input: "text"[];
11767
+ cost: {
11768
+ input: number;
11769
+ output: number;
11770
+ cacheRead: number;
11771
+ cacheWrite: number;
11772
+ };
11773
+ contextWindow: number;
11774
+ maxTokens: number;
11775
+ };
11005
11776
  readonly "poolside/laguna-m.1:free": {
11006
11777
  id: string;
11007
11778
  name: string;
@@ -11852,23 +12623,6 @@ export declare const MODELS: {
11852
12623
  contextWindow: number;
11853
12624
  maxTokens: number;
11854
12625
  };
11855
- readonly "qwen/qwq-32b": {
11856
- id: string;
11857
- name: string;
11858
- api: "openai-completions";
11859
- provider: string;
11860
- baseUrl: string;
11861
- reasoning: true;
11862
- input: "text"[];
11863
- cost: {
11864
- input: number;
11865
- output: number;
11866
- cacheRead: number;
11867
- cacheWrite: number;
11868
- };
11869
- contextWindow: number;
11870
- maxTokens: number;
11871
- };
11872
12626
  readonly "rekaai/reka-edge": {
11873
12627
  id: string;
11874
12628
  name: string;
@@ -12908,6 +13662,23 @@ export declare const MODELS: {
12908
13662
  contextWindow: number;
12909
13663
  maxTokens: number;
12910
13664
  };
13665
+ readonly "alibaba/qwen3.6-27b": {
13666
+ id: string;
13667
+ name: string;
13668
+ api: "anthropic-messages";
13669
+ provider: string;
13670
+ baseUrl: string;
13671
+ reasoning: true;
13672
+ input: ("image" | "text")[];
13673
+ cost: {
13674
+ input: number;
13675
+ output: number;
13676
+ cacheRead: number;
13677
+ cacheWrite: number;
13678
+ };
13679
+ contextWindow: number;
13680
+ maxTokens: number;
13681
+ };
12911
13682
  readonly "alibaba/qwen3.6-plus": {
12912
13683
  id: string;
12913
13684
  name: string;
@@ -13288,7 +14059,7 @@ export declare const MODELS: {
13288
14059
  api: "anthropic-messages";
13289
14060
  provider: string;
13290
14061
  baseUrl: string;
13291
- reasoning: true;
14062
+ reasoning: false;
13292
14063
  input: "text"[];
13293
14064
  cost: {
13294
14065
  input: number;
@@ -15135,6 +15906,40 @@ export declare const MODELS: {
15135
15906
  contextWindow: number;
15136
15907
  maxTokens: number;
15137
15908
  };
15909
+ readonly "xiaomi/mimo-v2.5": {
15910
+ id: string;
15911
+ name: string;
15912
+ api: "anthropic-messages";
15913
+ provider: string;
15914
+ baseUrl: string;
15915
+ reasoning: true;
15916
+ input: ("image" | "text")[];
15917
+ cost: {
15918
+ input: number;
15919
+ output: number;
15920
+ cacheRead: number;
15921
+ cacheWrite: number;
15922
+ };
15923
+ contextWindow: number;
15924
+ maxTokens: number;
15925
+ };
15926
+ readonly "xiaomi/mimo-v2.5-pro": {
15927
+ id: string;
15928
+ name: string;
15929
+ api: "anthropic-messages";
15930
+ provider: string;
15931
+ baseUrl: string;
15932
+ reasoning: true;
15933
+ input: ("image" | "text")[];
15934
+ cost: {
15935
+ input: number;
15936
+ output: number;
15937
+ cacheRead: number;
15938
+ cacheWrite: number;
15939
+ };
15940
+ contextWindow: number;
15941
+ maxTokens: number;
15942
+ };
15138
15943
  readonly "zai/glm-4.5": {
15139
15944
  id: string;
15140
15945
  name: string;
@@ -15855,6 +16660,28 @@ export declare const MODELS: {
15855
16660
  contextWindow: number;
15856
16661
  maxTokens: number;
15857
16662
  };
16663
+ readonly "glm-5v-turbo": {
16664
+ id: string;
16665
+ name: string;
16666
+ api: "openai-completions";
16667
+ provider: string;
16668
+ baseUrl: string;
16669
+ compat: {
16670
+ supportsDeveloperRole: false;
16671
+ thinkingFormat: "zai";
16672
+ zaiToolStream: true;
16673
+ };
16674
+ reasoning: true;
16675
+ input: ("image" | "text")[];
16676
+ cost: {
16677
+ input: number;
16678
+ output: number;
16679
+ cacheRead: number;
16680
+ cacheWrite: number;
16681
+ };
16682
+ contextWindow: number;
16683
+ maxTokens: number;
16684
+ };
15858
16685
  };
15859
16686
  };
15860
16687
  //# sourceMappingURL=models.generated.d.ts.map