@mariozechner/pi-ai 0.70.6 → 0.71.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (70) hide show
  1. package/README.md +3 -30
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +3 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +3 -2
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js +1 -0
  8. package/dist/index.js.map +1 -1
  9. package/dist/models.d.ts +1 -1
  10. package/dist/models.d.ts.map +1 -1
  11. package/dist/models.generated.d.ts +1437 -576
  12. package/dist/models.generated.d.ts.map +1 -1
  13. package/dist/models.generated.js +1173 -438
  14. package/dist/models.generated.js.map +1 -1
  15. package/dist/models.js +3 -2
  16. package/dist/models.js.map +1 -1
  17. package/dist/providers/anthropic.d.ts.map +1 -1
  18. package/dist/providers/anthropic.js +38 -15
  19. package/dist/providers/anthropic.js.map +1 -1
  20. package/dist/providers/cloudflare.d.ts +9 -3
  21. package/dist/providers/cloudflare.d.ts.map +1 -1
  22. package/dist/providers/cloudflare.js +10 -3
  23. package/dist/providers/cloudflare.js.map +1 -1
  24. package/dist/providers/google-shared.d.ts +7 -2
  25. package/dist/providers/google-shared.d.ts.map +1 -1
  26. package/dist/providers/google-shared.js +4 -13
  27. package/dist/providers/google-shared.js.map +1 -1
  28. package/dist/providers/google-vertex.d.ts +1 -1
  29. package/dist/providers/google-vertex.d.ts.map +1 -1
  30. package/dist/providers/google-vertex.js.map +1 -1
  31. package/dist/providers/google.d.ts +1 -1
  32. package/dist/providers/google.d.ts.map +1 -1
  33. package/dist/providers/google.js.map +1 -1
  34. package/dist/providers/mistral.d.ts.map +1 -1
  35. package/dist/providers/mistral.js +1 -1
  36. package/dist/providers/mistral.js.map +1 -1
  37. package/dist/providers/openai-codex-responses.d.ts +15 -0
  38. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  39. package/dist/providers/openai-codex-responses.js +138 -3
  40. package/dist/providers/openai-codex-responses.js.map +1 -1
  41. package/dist/providers/openai-completions.d.ts.map +1 -1
  42. package/dist/providers/openai-completions.js +21 -7
  43. package/dist/providers/openai-completions.js.map +1 -1
  44. package/dist/providers/openai-responses.d.ts.map +1 -1
  45. package/dist/providers/openai-responses.js +10 -2
  46. package/dist/providers/openai-responses.js.map +1 -1
  47. package/dist/providers/register-builtins.d.ts +0 -3
  48. package/dist/providers/register-builtins.d.ts.map +1 -1
  49. package/dist/providers/register-builtins.js +0 -18
  50. package/dist/providers/register-builtins.js.map +1 -1
  51. package/dist/types.d.ts +4 -3
  52. package/dist/types.d.ts.map +1 -1
  53. package/dist/types.js.map +1 -1
  54. package/dist/utils/oauth/index.d.ts +0 -4
  55. package/dist/utils/oauth/index.d.ts.map +1 -1
  56. package/dist/utils/oauth/index.js +0 -10
  57. package/dist/utils/oauth/index.js.map +1 -1
  58. package/package.json +2 -6
  59. package/dist/providers/google-gemini-cli.d.ts +0 -74
  60. package/dist/providers/google-gemini-cli.d.ts.map +0 -1
  61. package/dist/providers/google-gemini-cli.js +0 -779
  62. package/dist/providers/google-gemini-cli.js.map +0 -1
  63. package/dist/utils/oauth/google-antigravity.d.ts +0 -26
  64. package/dist/utils/oauth/google-antigravity.d.ts.map +0 -1
  65. package/dist/utils/oauth/google-antigravity.js +0 -377
  66. package/dist/utils/oauth/google-antigravity.js.map +0 -1
  67. package/dist/utils/oauth/google-gemini-cli.d.ts +0 -26
  68. package/dist/utils/oauth/google-gemini-cli.d.ts.map +0 -1
  69. package/dist/utils/oauth/google-gemini-cli.js +0 -482
  70. package/dist/utils/oauth/google-gemini-cli.js.map +0 -1
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -2744,17 +2761,14 @@ export declare const MODELS: {
2744
2761
  maxTokens: number;
2745
2762
  };
2746
2763
  };
2747
- readonly "cloudflare-workers-ai": {
2748
- readonly "@cf/google/gemma-4-26b-a4b-it": {
2764
+ readonly "cloudflare-ai-gateway": {
2765
+ readonly "claude-3-5-haiku": {
2749
2766
  id: string;
2750
2767
  name: string;
2751
- api: "openai-completions";
2768
+ api: "anthropic-messages";
2752
2769
  provider: string;
2753
2770
  baseUrl: string;
2754
- compat: {
2755
- sendSessionAffinityHeaders: true;
2756
- };
2757
- reasoning: true;
2771
+ reasoning: false;
2758
2772
  input: ("image" | "text")[];
2759
2773
  cost: {
2760
2774
  input: number;
@@ -2765,15 +2779,12 @@ export declare const MODELS: {
2765
2779
  contextWindow: number;
2766
2780
  maxTokens: number;
2767
2781
  };
2768
- readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
2782
+ readonly "claude-3-haiku": {
2769
2783
  id: string;
2770
2784
  name: string;
2771
- api: "openai-completions";
2785
+ api: "anthropic-messages";
2772
2786
  provider: string;
2773
2787
  baseUrl: string;
2774
- compat: {
2775
- sendSessionAffinityHeaders: true;
2776
- };
2777
2788
  reasoning: false;
2778
2789
  input: ("image" | "text")[];
2779
2790
  cost: {
@@ -2785,16 +2796,13 @@ export declare const MODELS: {
2785
2796
  contextWindow: number;
2786
2797
  maxTokens: number;
2787
2798
  };
2788
- readonly "@cf/moonshotai/kimi-k2.5": {
2799
+ readonly "claude-3-opus": {
2789
2800
  id: string;
2790
2801
  name: string;
2791
- api: "openai-completions";
2802
+ api: "anthropic-messages";
2792
2803
  provider: string;
2793
2804
  baseUrl: string;
2794
- compat: {
2795
- sendSessionAffinityHeaders: true;
2796
- };
2797
- reasoning: true;
2805
+ reasoning: false;
2798
2806
  input: ("image" | "text")[];
2799
2807
  cost: {
2800
2808
  input: number;
@@ -2805,16 +2813,13 @@ export declare const MODELS: {
2805
2813
  contextWindow: number;
2806
2814
  maxTokens: number;
2807
2815
  };
2808
- readonly "@cf/moonshotai/kimi-k2.6": {
2816
+ readonly "claude-3-sonnet": {
2809
2817
  id: string;
2810
2818
  name: string;
2811
- api: "openai-completions";
2819
+ api: "anthropic-messages";
2812
2820
  provider: string;
2813
2821
  baseUrl: string;
2814
- compat: {
2815
- sendSessionAffinityHeaders: true;
2816
- };
2817
- reasoning: true;
2822
+ reasoning: false;
2818
2823
  input: ("image" | "text")[];
2819
2824
  cost: {
2820
2825
  input: number;
@@ -2825,17 +2830,14 @@ export declare const MODELS: {
2825
2830
  contextWindow: number;
2826
2831
  maxTokens: number;
2827
2832
  };
2828
- readonly "@cf/nvidia/nemotron-3-120b-a12b": {
2833
+ readonly "claude-3.5-haiku": {
2829
2834
  id: string;
2830
2835
  name: string;
2831
- api: "openai-completions";
2836
+ api: "anthropic-messages";
2832
2837
  provider: string;
2833
2838
  baseUrl: string;
2834
- compat: {
2835
- sendSessionAffinityHeaders: true;
2836
- };
2837
- reasoning: true;
2838
- input: "text"[];
2839
+ reasoning: false;
2840
+ input: ("image" | "text")[];
2839
2841
  cost: {
2840
2842
  input: number;
2841
2843
  output: number;
@@ -2845,17 +2847,14 @@ export declare const MODELS: {
2845
2847
  contextWindow: number;
2846
2848
  maxTokens: number;
2847
2849
  };
2848
- readonly "@cf/openai/gpt-oss-120b": {
2850
+ readonly "claude-3.5-sonnet": {
2849
2851
  id: string;
2850
2852
  name: string;
2851
- api: "openai-completions";
2853
+ api: "anthropic-messages";
2852
2854
  provider: string;
2853
2855
  baseUrl: string;
2854
- compat: {
2855
- sendSessionAffinityHeaders: true;
2856
- };
2857
- reasoning: true;
2858
- input: "text"[];
2856
+ reasoning: false;
2857
+ input: ("image" | "text")[];
2859
2858
  cost: {
2860
2859
  input: number;
2861
2860
  output: number;
@@ -2865,17 +2864,14 @@ export declare const MODELS: {
2865
2864
  contextWindow: number;
2866
2865
  maxTokens: number;
2867
2866
  };
2868
- readonly "@cf/openai/gpt-oss-20b": {
2867
+ readonly "claude-haiku-4-5": {
2869
2868
  id: string;
2870
2869
  name: string;
2871
- api: "openai-completions";
2870
+ api: "anthropic-messages";
2872
2871
  provider: string;
2873
2872
  baseUrl: string;
2874
- compat: {
2875
- sendSessionAffinityHeaders: true;
2876
- };
2877
2873
  reasoning: true;
2878
- input: "text"[];
2874
+ input: ("image" | "text")[];
2879
2875
  cost: {
2880
2876
  input: number;
2881
2877
  output: number;
@@ -2885,17 +2881,14 @@ export declare const MODELS: {
2885
2881
  contextWindow: number;
2886
2882
  maxTokens: number;
2887
2883
  };
2888
- readonly "@cf/zai-org/glm-4.7-flash": {
2884
+ readonly "claude-opus-4": {
2889
2885
  id: string;
2890
2886
  name: string;
2891
- api: "openai-completions";
2887
+ api: "anthropic-messages";
2892
2888
  provider: string;
2893
2889
  baseUrl: string;
2894
- compat: {
2895
- sendSessionAffinityHeaders: true;
2896
- };
2897
2890
  reasoning: true;
2898
- input: "text"[];
2891
+ input: ("image" | "text")[];
2899
2892
  cost: {
2900
2893
  input: number;
2901
2894
  output: number;
@@ -2905,27 +2898,14 @@ export declare const MODELS: {
2905
2898
  contextWindow: number;
2906
2899
  maxTokens: number;
2907
2900
  };
2908
- };
2909
- readonly deepseek: {
2910
- readonly "deepseek-v4-flash": {
2901
+ readonly "claude-opus-4-1": {
2911
2902
  id: string;
2912
2903
  name: string;
2913
- api: "openai-completions";
2904
+ api: "anthropic-messages";
2914
2905
  provider: string;
2915
2906
  baseUrl: string;
2916
- compat: {
2917
- requiresReasoningContentOnAssistantMessages: true;
2918
- thinkingFormat: "deepseek";
2919
- reasoningEffortMap: {
2920
- minimal: string;
2921
- low: string;
2922
- medium: string;
2923
- high: string;
2924
- xhigh: string;
2925
- };
2926
- };
2927
2907
  reasoning: true;
2928
- input: "text"[];
2908
+ input: ("image" | "text")[];
2929
2909
  cost: {
2930
2910
  input: number;
2931
2911
  output: number;
@@ -2935,25 +2915,14 @@ export declare const MODELS: {
2935
2915
  contextWindow: number;
2936
2916
  maxTokens: number;
2937
2917
  };
2938
- readonly "deepseek-v4-pro": {
2918
+ readonly "claude-opus-4-5": {
2939
2919
  id: string;
2940
2920
  name: string;
2941
- api: "openai-completions";
2921
+ api: "anthropic-messages";
2942
2922
  provider: string;
2943
2923
  baseUrl: string;
2944
- compat: {
2945
- requiresReasoningContentOnAssistantMessages: true;
2946
- thinkingFormat: "deepseek";
2947
- reasoningEffortMap: {
2948
- minimal: string;
2949
- low: string;
2950
- medium: string;
2951
- high: string;
2952
- xhigh: string;
2953
- };
2954
- };
2955
2924
  reasoning: true;
2956
- input: "text"[];
2925
+ input: ("image" | "text")[];
2957
2926
  cost: {
2958
2927
  input: number;
2959
2928
  output: number;
@@ -2963,16 +2932,14 @@ export declare const MODELS: {
2963
2932
  contextWindow: number;
2964
2933
  maxTokens: number;
2965
2934
  };
2966
- };
2967
- readonly fireworks: {
2968
- readonly "accounts/fireworks/models/deepseek-v3p1": {
2935
+ readonly "claude-opus-4-6": {
2969
2936
  id: string;
2970
2937
  name: string;
2971
2938
  api: "anthropic-messages";
2972
2939
  provider: string;
2973
2940
  baseUrl: string;
2974
2941
  reasoning: true;
2975
- input: "text"[];
2942
+ input: ("image" | "text")[];
2976
2943
  cost: {
2977
2944
  input: number;
2978
2945
  output: number;
@@ -2982,14 +2949,14 @@ export declare const MODELS: {
2982
2949
  contextWindow: number;
2983
2950
  maxTokens: number;
2984
2951
  };
2985
- readonly "accounts/fireworks/models/deepseek-v3p2": {
2952
+ readonly "claude-opus-4-7": {
2986
2953
  id: string;
2987
2954
  name: string;
2988
2955
  api: "anthropic-messages";
2989
2956
  provider: string;
2990
2957
  baseUrl: string;
2991
2958
  reasoning: true;
2992
- input: "text"[];
2959
+ input: ("image" | "text")[];
2993
2960
  cost: {
2994
2961
  input: number;
2995
2962
  output: number;
@@ -2999,14 +2966,14 @@ export declare const MODELS: {
2999
2966
  contextWindow: number;
3000
2967
  maxTokens: number;
3001
2968
  };
3002
- readonly "accounts/fireworks/models/deepseek-v4-pro": {
2969
+ readonly "claude-sonnet-4": {
3003
2970
  id: string;
3004
2971
  name: string;
3005
2972
  api: "anthropic-messages";
3006
2973
  provider: string;
3007
2974
  baseUrl: string;
3008
2975
  reasoning: true;
3009
- input: "text"[];
2976
+ input: ("image" | "text")[];
3010
2977
  cost: {
3011
2978
  input: number;
3012
2979
  output: number;
@@ -3016,14 +2983,14 @@ export declare const MODELS: {
3016
2983
  contextWindow: number;
3017
2984
  maxTokens: number;
3018
2985
  };
3019
- readonly "accounts/fireworks/models/glm-4p5": {
2986
+ readonly "claude-sonnet-4-5": {
3020
2987
  id: string;
3021
2988
  name: string;
3022
2989
  api: "anthropic-messages";
3023
2990
  provider: string;
3024
2991
  baseUrl: string;
3025
2992
  reasoning: true;
3026
- input: "text"[];
2993
+ input: ("image" | "text")[];
3027
2994
  cost: {
3028
2995
  input: number;
3029
2996
  output: number;
@@ -3033,14 +3000,14 @@ export declare const MODELS: {
3033
3000
  contextWindow: number;
3034
3001
  maxTokens: number;
3035
3002
  };
3036
- readonly "accounts/fireworks/models/glm-4p5-air": {
3003
+ readonly "claude-sonnet-4-6": {
3037
3004
  id: string;
3038
3005
  name: string;
3039
3006
  api: "anthropic-messages";
3040
3007
  provider: string;
3041
3008
  baseUrl: string;
3042
3009
  reasoning: true;
3043
- input: "text"[];
3010
+ input: ("image" | "text")[];
3044
3011
  cost: {
3045
3012
  input: number;
3046
3013
  output: number;
@@ -3050,13 +3017,13 @@ export declare const MODELS: {
3050
3017
  contextWindow: number;
3051
3018
  maxTokens: number;
3052
3019
  };
3053
- readonly "accounts/fireworks/models/glm-4p7": {
3020
+ readonly "gpt-4": {
3054
3021
  id: string;
3055
3022
  name: string;
3056
- api: "anthropic-messages";
3023
+ api: "openai-responses";
3057
3024
  provider: string;
3058
3025
  baseUrl: string;
3059
- reasoning: true;
3026
+ reasoning: false;
3060
3027
  input: "text"[];
3061
3028
  cost: {
3062
3029
  input: number;
@@ -3067,14 +3034,14 @@ export declare const MODELS: {
3067
3034
  contextWindow: number;
3068
3035
  maxTokens: number;
3069
3036
  };
3070
- readonly "accounts/fireworks/models/glm-5": {
3037
+ readonly "gpt-4-turbo": {
3071
3038
  id: string;
3072
3039
  name: string;
3073
- api: "anthropic-messages";
3040
+ api: "openai-responses";
3074
3041
  provider: string;
3075
3042
  baseUrl: string;
3076
- reasoning: true;
3077
- input: "text"[];
3043
+ reasoning: false;
3044
+ input: ("image" | "text")[];
3078
3045
  cost: {
3079
3046
  input: number;
3080
3047
  output: number;
@@ -3084,14 +3051,14 @@ export declare const MODELS: {
3084
3051
  contextWindow: number;
3085
3052
  maxTokens: number;
3086
3053
  };
3087
- readonly "accounts/fireworks/models/glm-5p1": {
3054
+ readonly "gpt-4o": {
3088
3055
  id: string;
3089
3056
  name: string;
3090
- api: "anthropic-messages";
3057
+ api: "openai-responses";
3091
3058
  provider: string;
3092
3059
  baseUrl: string;
3093
- reasoning: true;
3094
- input: "text"[];
3060
+ reasoning: false;
3061
+ input: ("image" | "text")[];
3095
3062
  cost: {
3096
3063
  input: number;
3097
3064
  output: number;
@@ -3101,14 +3068,14 @@ export declare const MODELS: {
3101
3068
  contextWindow: number;
3102
3069
  maxTokens: number;
3103
3070
  };
3104
- readonly "accounts/fireworks/models/gpt-oss-120b": {
3071
+ readonly "gpt-4o-mini": {
3105
3072
  id: string;
3106
3073
  name: string;
3107
- api: "anthropic-messages";
3074
+ api: "openai-responses";
3108
3075
  provider: string;
3109
3076
  baseUrl: string;
3110
- reasoning: true;
3111
- input: "text"[];
3077
+ reasoning: false;
3078
+ input: ("image" | "text")[];
3112
3079
  cost: {
3113
3080
  input: number;
3114
3081
  output: number;
@@ -3118,14 +3085,14 @@ export declare const MODELS: {
3118
3085
  contextWindow: number;
3119
3086
  maxTokens: number;
3120
3087
  };
3121
- readonly "accounts/fireworks/models/gpt-oss-20b": {
3088
+ readonly "gpt-5.1": {
3122
3089
  id: string;
3123
3090
  name: string;
3124
- api: "anthropic-messages";
3091
+ api: "openai-responses";
3125
3092
  provider: string;
3126
3093
  baseUrl: string;
3127
3094
  reasoning: true;
3128
- input: "text"[];
3095
+ input: ("image" | "text")[];
3129
3096
  cost: {
3130
3097
  input: number;
3131
3098
  output: number;
@@ -3135,14 +3102,14 @@ export declare const MODELS: {
3135
3102
  contextWindow: number;
3136
3103
  maxTokens: number;
3137
3104
  };
3138
- readonly "accounts/fireworks/models/kimi-k2-instruct": {
3105
+ readonly "gpt-5.1-codex": {
3139
3106
  id: string;
3140
3107
  name: string;
3141
- api: "anthropic-messages";
3108
+ api: "openai-responses";
3142
3109
  provider: string;
3143
3110
  baseUrl: string;
3144
- reasoning: false;
3145
- input: "text"[];
3111
+ reasoning: true;
3112
+ input: ("image" | "text")[];
3146
3113
  cost: {
3147
3114
  input: number;
3148
3115
  output: number;
@@ -3152,14 +3119,14 @@ export declare const MODELS: {
3152
3119
  contextWindow: number;
3153
3120
  maxTokens: number;
3154
3121
  };
3155
- readonly "accounts/fireworks/models/kimi-k2-thinking": {
3122
+ readonly "gpt-5.2": {
3156
3123
  id: string;
3157
3124
  name: string;
3158
- api: "anthropic-messages";
3125
+ api: "openai-responses";
3159
3126
  provider: string;
3160
3127
  baseUrl: string;
3161
3128
  reasoning: true;
3162
- input: "text"[];
3129
+ input: ("image" | "text")[];
3163
3130
  cost: {
3164
3131
  input: number;
3165
3132
  output: number;
@@ -3169,10 +3136,10 @@ export declare const MODELS: {
3169
3136
  contextWindow: number;
3170
3137
  maxTokens: number;
3171
3138
  };
3172
- readonly "accounts/fireworks/models/kimi-k2p5": {
3139
+ readonly "gpt-5.2-codex": {
3173
3140
  id: string;
3174
3141
  name: string;
3175
- api: "anthropic-messages";
3142
+ api: "openai-responses";
3176
3143
  provider: string;
3177
3144
  baseUrl: string;
3178
3145
  reasoning: true;
@@ -3186,10 +3153,10 @@ export declare const MODELS: {
3186
3153
  contextWindow: number;
3187
3154
  maxTokens: number;
3188
3155
  };
3189
- readonly "accounts/fireworks/models/kimi-k2p6": {
3156
+ readonly "gpt-5.3-codex": {
3190
3157
  id: string;
3191
3158
  name: string;
3192
- api: "anthropic-messages";
3159
+ api: "openai-responses";
3193
3160
  provider: string;
3194
3161
  baseUrl: string;
3195
3162
  reasoning: true;
@@ -3203,14 +3170,14 @@ export declare const MODELS: {
3203
3170
  contextWindow: number;
3204
3171
  maxTokens: number;
3205
3172
  };
3206
- readonly "accounts/fireworks/models/minimax-m2p1": {
3173
+ readonly "gpt-5.4": {
3207
3174
  id: string;
3208
3175
  name: string;
3209
- api: "anthropic-messages";
3176
+ api: "openai-responses";
3210
3177
  provider: string;
3211
3178
  baseUrl: string;
3212
3179
  reasoning: true;
3213
- input: "text"[];
3180
+ input: ("image" | "text")[];
3214
3181
  cost: {
3215
3182
  input: number;
3216
3183
  output: number;
@@ -3220,14 +3187,14 @@ export declare const MODELS: {
3220
3187
  contextWindow: number;
3221
3188
  maxTokens: number;
3222
3189
  };
3223
- readonly "accounts/fireworks/models/minimax-m2p5": {
3190
+ readonly "gpt-5.5": {
3224
3191
  id: string;
3225
3192
  name: string;
3226
- api: "anthropic-messages";
3193
+ api: "openai-responses";
3227
3194
  provider: string;
3228
3195
  baseUrl: string;
3229
3196
  reasoning: true;
3230
- input: "text"[];
3197
+ input: ("image" | "text")[];
3231
3198
  cost: {
3232
3199
  input: number;
3233
3200
  output: number;
@@ -3237,14 +3204,14 @@ export declare const MODELS: {
3237
3204
  contextWindow: number;
3238
3205
  maxTokens: number;
3239
3206
  };
3240
- readonly "accounts/fireworks/models/minimax-m2p7": {
3207
+ readonly o1: {
3241
3208
  id: string;
3242
3209
  name: string;
3243
- api: "anthropic-messages";
3210
+ api: "openai-responses";
3244
3211
  provider: string;
3245
3212
  baseUrl: string;
3246
3213
  reasoning: true;
3247
- input: "text"[];
3214
+ input: ("image" | "text")[];
3248
3215
  cost: {
3249
3216
  input: number;
3250
3217
  output: number;
@@ -3254,10 +3221,10 @@ export declare const MODELS: {
3254
3221
  contextWindow: number;
3255
3222
  maxTokens: number;
3256
3223
  };
3257
- readonly "accounts/fireworks/models/qwen3p6-plus": {
3224
+ readonly o3: {
3258
3225
  id: string;
3259
3226
  name: string;
3260
- api: "anthropic-messages";
3227
+ api: "openai-responses";
3261
3228
  provider: string;
3262
3229
  baseUrl: string;
3263
3230
  reasoning: true;
@@ -3271,14 +3238,14 @@ export declare const MODELS: {
3271
3238
  contextWindow: number;
3272
3239
  maxTokens: number;
3273
3240
  };
3274
- readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
3241
+ readonly "o3-mini": {
3275
3242
  id: string;
3276
3243
  name: string;
3277
- api: "anthropic-messages";
3244
+ api: "openai-responses";
3278
3245
  provider: string;
3279
3246
  baseUrl: string;
3280
3247
  reasoning: true;
3281
- input: ("image" | "text")[];
3248
+ input: "text"[];
3282
3249
  cost: {
3283
3250
  input: number;
3284
3251
  output: number;
@@ -3288,23 +3255,12 @@ export declare const MODELS: {
3288
3255
  contextWindow: number;
3289
3256
  maxTokens: number;
3290
3257
  };
3291
- };
3292
- readonly "github-copilot": {
3293
- readonly "claude-haiku-4.5": {
3258
+ readonly "o3-pro": {
3294
3259
  id: string;
3295
3260
  name: string;
3296
- api: "anthropic-messages";
3261
+ api: "openai-responses";
3297
3262
  provider: string;
3298
3263
  baseUrl: string;
3299
- headers: {
3300
- "User-Agent": string;
3301
- "Editor-Version": string;
3302
- "Editor-Plugin-Version": string;
3303
- "Copilot-Integration-Id": string;
3304
- };
3305
- compat: {
3306
- supportsEagerToolInputStreaming: false;
3307
- };
3308
3264
  reasoning: true;
3309
3265
  input: ("image" | "text")[];
3310
3266
  cost: {
@@ -3316,18 +3272,12 @@ export declare const MODELS: {
3316
3272
  contextWindow: number;
3317
3273
  maxTokens: number;
3318
3274
  };
3319
- readonly "claude-opus-4.5": {
3275
+ readonly "o4-mini": {
3320
3276
  id: string;
3321
3277
  name: string;
3322
- api: "anthropic-messages";
3278
+ api: "openai-responses";
3323
3279
  provider: string;
3324
3280
  baseUrl: string;
3325
- headers: {
3326
- "User-Agent": string;
3327
- "Editor-Version": string;
3328
- "Editor-Plugin-Version": string;
3329
- "Copilot-Integration-Id": string;
3330
- };
3331
3281
  reasoning: true;
3332
3282
  input: ("image" | "text")[];
3333
3283
  cost: {
@@ -3339,17 +3289,96 @@ export declare const MODELS: {
3339
3289
  contextWindow: number;
3340
3290
  maxTokens: number;
3341
3291
  };
3342
- readonly "claude-opus-4.6": {
3292
+ readonly "workers-ai/@cf/moonshotai/kimi-k2.5": {
3343
3293
  id: string;
3344
3294
  name: string;
3345
- api: "anthropic-messages";
3295
+ api: "openai-completions";
3346
3296
  provider: string;
3347
3297
  baseUrl: string;
3348
- headers: {
3349
- "User-Agent": string;
3350
- "Editor-Version": string;
3351
- "Editor-Plugin-Version": string;
3352
- "Copilot-Integration-Id": string;
3298
+ compat: {
3299
+ sendSessionAffinityHeaders: true;
3300
+ };
3301
+ reasoning: true;
3302
+ input: ("image" | "text")[];
3303
+ cost: {
3304
+ input: number;
3305
+ output: number;
3306
+ cacheRead: number;
3307
+ cacheWrite: number;
3308
+ };
3309
+ contextWindow: number;
3310
+ maxTokens: number;
3311
+ };
3312
+ readonly "workers-ai/@cf/moonshotai/kimi-k2.6": {
3313
+ id: string;
3314
+ name: string;
3315
+ api: "openai-completions";
3316
+ provider: string;
3317
+ baseUrl: string;
3318
+ compat: {
3319
+ sendSessionAffinityHeaders: true;
3320
+ };
3321
+ reasoning: true;
3322
+ input: ("image" | "text")[];
3323
+ cost: {
3324
+ input: number;
3325
+ output: number;
3326
+ cacheRead: number;
3327
+ cacheWrite: number;
3328
+ };
3329
+ contextWindow: number;
3330
+ maxTokens: number;
3331
+ };
3332
+ readonly "workers-ai/@cf/nvidia/nemotron-3-120b-a12b": {
3333
+ id: string;
3334
+ name: string;
3335
+ api: "openai-completions";
3336
+ provider: string;
3337
+ baseUrl: string;
3338
+ compat: {
3339
+ sendSessionAffinityHeaders: true;
3340
+ };
3341
+ reasoning: true;
3342
+ input: "text"[];
3343
+ cost: {
3344
+ input: number;
3345
+ output: number;
3346
+ cacheRead: number;
3347
+ cacheWrite: number;
3348
+ };
3349
+ contextWindow: number;
3350
+ maxTokens: number;
3351
+ };
3352
+ readonly "workers-ai/@cf/zai-org/glm-4.7-flash": {
3353
+ id: string;
3354
+ name: string;
3355
+ api: "openai-completions";
3356
+ provider: string;
3357
+ baseUrl: string;
3358
+ compat: {
3359
+ sendSessionAffinityHeaders: true;
3360
+ };
3361
+ reasoning: true;
3362
+ input: "text"[];
3363
+ cost: {
3364
+ input: number;
3365
+ output: number;
3366
+ cacheRead: number;
3367
+ cacheWrite: number;
3368
+ };
3369
+ contextWindow: number;
3370
+ maxTokens: number;
3371
+ };
3372
+ };
3373
+ readonly "cloudflare-workers-ai": {
3374
+ readonly "@cf/google/gemma-4-26b-a4b-it": {
3375
+ id: string;
3376
+ name: string;
3377
+ api: "openai-completions";
3378
+ provider: string;
3379
+ baseUrl: string;
3380
+ compat: {
3381
+ sendSessionAffinityHeaders: true;
3353
3382
  };
3354
3383
  reasoning: true;
3355
3384
  input: ("image" | "text")[];
@@ -3362,20 +3391,316 @@ export declare const MODELS: {
3362
3391
  contextWindow: number;
3363
3392
  maxTokens: number;
3364
3393
  };
3365
- readonly "claude-opus-4.7": {
3394
+ readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
3395
+ id: string;
3396
+ name: string;
3397
+ api: "openai-completions";
3398
+ provider: string;
3399
+ baseUrl: string;
3400
+ compat: {
3401
+ sendSessionAffinityHeaders: true;
3402
+ };
3403
+ reasoning: false;
3404
+ input: ("image" | "text")[];
3405
+ cost: {
3406
+ input: number;
3407
+ output: number;
3408
+ cacheRead: number;
3409
+ cacheWrite: number;
3410
+ };
3411
+ contextWindow: number;
3412
+ maxTokens: number;
3413
+ };
3414
+ readonly "@cf/moonshotai/kimi-k2.5": {
3415
+ id: string;
3416
+ name: string;
3417
+ api: "openai-completions";
3418
+ provider: string;
3419
+ baseUrl: string;
3420
+ compat: {
3421
+ sendSessionAffinityHeaders: true;
3422
+ };
3423
+ reasoning: true;
3424
+ input: ("image" | "text")[];
3425
+ cost: {
3426
+ input: number;
3427
+ output: number;
3428
+ cacheRead: number;
3429
+ cacheWrite: number;
3430
+ };
3431
+ contextWindow: number;
3432
+ maxTokens: number;
3433
+ };
3434
+ readonly "@cf/moonshotai/kimi-k2.6": {
3435
+ id: string;
3436
+ name: string;
3437
+ api: "openai-completions";
3438
+ provider: string;
3439
+ baseUrl: string;
3440
+ compat: {
3441
+ sendSessionAffinityHeaders: true;
3442
+ };
3443
+ reasoning: true;
3444
+ input: ("image" | "text")[];
3445
+ cost: {
3446
+ input: number;
3447
+ output: number;
3448
+ cacheRead: number;
3449
+ cacheWrite: number;
3450
+ };
3451
+ contextWindow: number;
3452
+ maxTokens: number;
3453
+ };
3454
+ readonly "@cf/nvidia/nemotron-3-120b-a12b": {
3455
+ id: string;
3456
+ name: string;
3457
+ api: "openai-completions";
3458
+ provider: string;
3459
+ baseUrl: string;
3460
+ compat: {
3461
+ sendSessionAffinityHeaders: true;
3462
+ };
3463
+ reasoning: true;
3464
+ input: "text"[];
3465
+ cost: {
3466
+ input: number;
3467
+ output: number;
3468
+ cacheRead: number;
3469
+ cacheWrite: number;
3470
+ };
3471
+ contextWindow: number;
3472
+ maxTokens: number;
3473
+ };
3474
+ readonly "@cf/openai/gpt-oss-120b": {
3475
+ id: string;
3476
+ name: string;
3477
+ api: "openai-completions";
3478
+ provider: string;
3479
+ baseUrl: string;
3480
+ compat: {
3481
+ sendSessionAffinityHeaders: true;
3482
+ };
3483
+ reasoning: true;
3484
+ input: "text"[];
3485
+ cost: {
3486
+ input: number;
3487
+ output: number;
3488
+ cacheRead: number;
3489
+ cacheWrite: number;
3490
+ };
3491
+ contextWindow: number;
3492
+ maxTokens: number;
3493
+ };
3494
+ readonly "@cf/openai/gpt-oss-20b": {
3495
+ id: string;
3496
+ name: string;
3497
+ api: "openai-completions";
3498
+ provider: string;
3499
+ baseUrl: string;
3500
+ compat: {
3501
+ sendSessionAffinityHeaders: true;
3502
+ };
3503
+ reasoning: true;
3504
+ input: "text"[];
3505
+ cost: {
3506
+ input: number;
3507
+ output: number;
3508
+ cacheRead: number;
3509
+ cacheWrite: number;
3510
+ };
3511
+ contextWindow: number;
3512
+ maxTokens: number;
3513
+ };
3514
+ readonly "@cf/zai-org/glm-4.7-flash": {
3515
+ id: string;
3516
+ name: string;
3517
+ api: "openai-completions";
3518
+ provider: string;
3519
+ baseUrl: string;
3520
+ compat: {
3521
+ sendSessionAffinityHeaders: true;
3522
+ };
3523
+ reasoning: true;
3524
+ input: "text"[];
3525
+ cost: {
3526
+ input: number;
3527
+ output: number;
3528
+ cacheRead: number;
3529
+ cacheWrite: number;
3530
+ };
3531
+ contextWindow: number;
3532
+ maxTokens: number;
3533
+ };
3534
+ };
3535
+ readonly deepseek: {
3536
+ readonly "deepseek-v4-flash": {
3537
+ id: string;
3538
+ name: string;
3539
+ api: "openai-completions";
3540
+ provider: string;
3541
+ baseUrl: string;
3542
+ compat: {
3543
+ requiresReasoningContentOnAssistantMessages: true;
3544
+ thinkingFormat: "deepseek";
3545
+ reasoningEffortMap: {
3546
+ minimal: string;
3547
+ low: string;
3548
+ medium: string;
3549
+ high: string;
3550
+ xhigh: string;
3551
+ };
3552
+ };
3553
+ reasoning: true;
3554
+ input: "text"[];
3555
+ cost: {
3556
+ input: number;
3557
+ output: number;
3558
+ cacheRead: number;
3559
+ cacheWrite: number;
3560
+ };
3561
+ contextWindow: number;
3562
+ maxTokens: number;
3563
+ };
3564
+ readonly "deepseek-v4-pro": {
3565
+ id: string;
3566
+ name: string;
3567
+ api: "openai-completions";
3568
+ provider: string;
3569
+ baseUrl: string;
3570
+ compat: {
3571
+ requiresReasoningContentOnAssistantMessages: true;
3572
+ thinkingFormat: "deepseek";
3573
+ reasoningEffortMap: {
3574
+ minimal: string;
3575
+ low: string;
3576
+ medium: string;
3577
+ high: string;
3578
+ xhigh: string;
3579
+ };
3580
+ };
3581
+ reasoning: true;
3582
+ input: "text"[];
3583
+ cost: {
3584
+ input: number;
3585
+ output: number;
3586
+ cacheRead: number;
3587
+ cacheWrite: number;
3588
+ };
3589
+ contextWindow: number;
3590
+ maxTokens: number;
3591
+ };
3592
+ };
3593
+ readonly fireworks: {
3594
+ readonly "accounts/fireworks/models/deepseek-v3p1": {
3595
+ id: string;
3596
+ name: string;
3597
+ api: "anthropic-messages";
3598
+ provider: string;
3599
+ baseUrl: string;
3600
+ reasoning: true;
3601
+ input: "text"[];
3602
+ cost: {
3603
+ input: number;
3604
+ output: number;
3605
+ cacheRead: number;
3606
+ cacheWrite: number;
3607
+ };
3608
+ contextWindow: number;
3609
+ maxTokens: number;
3610
+ };
3611
+ readonly "accounts/fireworks/models/deepseek-v3p2": {
3612
+ id: string;
3613
+ name: string;
3614
+ api: "anthropic-messages";
3615
+ provider: string;
3616
+ baseUrl: string;
3617
+ reasoning: true;
3618
+ input: "text"[];
3619
+ cost: {
3620
+ input: number;
3621
+ output: number;
3622
+ cacheRead: number;
3623
+ cacheWrite: number;
3624
+ };
3625
+ contextWindow: number;
3626
+ maxTokens: number;
3627
+ };
3628
+ readonly "accounts/fireworks/models/deepseek-v4-pro": {
3629
+ id: string;
3630
+ name: string;
3631
+ api: "anthropic-messages";
3632
+ provider: string;
3633
+ baseUrl: string;
3634
+ reasoning: true;
3635
+ input: "text"[];
3636
+ cost: {
3637
+ input: number;
3638
+ output: number;
3639
+ cacheRead: number;
3640
+ cacheWrite: number;
3641
+ };
3642
+ contextWindow: number;
3643
+ maxTokens: number;
3644
+ };
3645
+ readonly "accounts/fireworks/models/glm-4p5": {
3646
+ id: string;
3647
+ name: string;
3648
+ api: "anthropic-messages";
3649
+ provider: string;
3650
+ baseUrl: string;
3651
+ reasoning: true;
3652
+ input: "text"[];
3653
+ cost: {
3654
+ input: number;
3655
+ output: number;
3656
+ cacheRead: number;
3657
+ cacheWrite: number;
3658
+ };
3659
+ contextWindow: number;
3660
+ maxTokens: number;
3661
+ };
3662
+ readonly "accounts/fireworks/models/glm-4p5-air": {
3663
+ id: string;
3664
+ name: string;
3665
+ api: "anthropic-messages";
3666
+ provider: string;
3667
+ baseUrl: string;
3668
+ reasoning: true;
3669
+ input: "text"[];
3670
+ cost: {
3671
+ input: number;
3672
+ output: number;
3673
+ cacheRead: number;
3674
+ cacheWrite: number;
3675
+ };
3676
+ contextWindow: number;
3677
+ maxTokens: number;
3678
+ };
3679
+ readonly "accounts/fireworks/models/glm-4p7": {
3680
+ id: string;
3681
+ name: string;
3682
+ api: "anthropic-messages";
3683
+ provider: string;
3684
+ baseUrl: string;
3685
+ reasoning: true;
3686
+ input: "text"[];
3687
+ cost: {
3688
+ input: number;
3689
+ output: number;
3690
+ cacheRead: number;
3691
+ cacheWrite: number;
3692
+ };
3693
+ contextWindow: number;
3694
+ maxTokens: number;
3695
+ };
3696
+ readonly "accounts/fireworks/models/glm-5": {
3366
3697
  id: string;
3367
3698
  name: string;
3368
3699
  api: "anthropic-messages";
3369
3700
  provider: string;
3370
3701
  baseUrl: string;
3371
- headers: {
3372
- "User-Agent": string;
3373
- "Editor-Version": string;
3374
- "Editor-Plugin-Version": string;
3375
- "Copilot-Integration-Id": string;
3376
- };
3377
3702
  reasoning: true;
3378
- input: ("image" | "text")[];
3703
+ input: "text"[];
3379
3704
  cost: {
3380
3705
  input: number;
3381
3706
  output: number;
@@ -3385,23 +3710,14 @@ export declare const MODELS: {
3385
3710
  contextWindow: number;
3386
3711
  maxTokens: number;
3387
3712
  };
3388
- readonly "claude-sonnet-4": {
3713
+ readonly "accounts/fireworks/models/glm-5p1": {
3389
3714
  id: string;
3390
3715
  name: string;
3391
3716
  api: "anthropic-messages";
3392
3717
  provider: string;
3393
3718
  baseUrl: string;
3394
- headers: {
3395
- "User-Agent": string;
3396
- "Editor-Version": string;
3397
- "Editor-Plugin-Version": string;
3398
- "Copilot-Integration-Id": string;
3399
- };
3400
- compat: {
3401
- supportsEagerToolInputStreaming: false;
3402
- };
3403
3719
  reasoning: true;
3404
- input: ("image" | "text")[];
3720
+ input: "text"[];
3405
3721
  cost: {
3406
3722
  input: number;
3407
3723
  output: number;
@@ -3411,23 +3727,14 @@ export declare const MODELS: {
3411
3727
  contextWindow: number;
3412
3728
  maxTokens: number;
3413
3729
  };
3414
- readonly "claude-sonnet-4.5": {
3730
+ readonly "accounts/fireworks/models/gpt-oss-120b": {
3415
3731
  id: string;
3416
3732
  name: string;
3417
3733
  api: "anthropic-messages";
3418
3734
  provider: string;
3419
3735
  baseUrl: string;
3420
- headers: {
3421
- "User-Agent": string;
3422
- "Editor-Version": string;
3423
- "Editor-Plugin-Version": string;
3424
- "Copilot-Integration-Id": string;
3425
- };
3426
- compat: {
3427
- supportsEagerToolInputStreaming: false;
3428
- };
3429
3736
  reasoning: true;
3430
- input: ("image" | "text")[];
3737
+ input: "text"[];
3431
3738
  cost: {
3432
3739
  input: number;
3433
3740
  output: number;
@@ -3437,20 +3744,14 @@ export declare const MODELS: {
3437
3744
  contextWindow: number;
3438
3745
  maxTokens: number;
3439
3746
  };
3440
- readonly "claude-sonnet-4.6": {
3747
+ readonly "accounts/fireworks/models/gpt-oss-20b": {
3441
3748
  id: string;
3442
3749
  name: string;
3443
3750
  api: "anthropic-messages";
3444
3751
  provider: string;
3445
3752
  baseUrl: string;
3446
- headers: {
3447
- "User-Agent": string;
3448
- "Editor-Version": string;
3449
- "Editor-Plugin-Version": string;
3450
- "Copilot-Integration-Id": string;
3451
- };
3452
3753
  reasoning: true;
3453
- input: ("image" | "text")[];
3754
+ input: "text"[];
3454
3755
  cost: {
3455
3756
  input: number;
3456
3757
  output: number;
@@ -3460,25 +3761,14 @@ export declare const MODELS: {
3460
3761
  contextWindow: number;
3461
3762
  maxTokens: number;
3462
3763
  };
3463
- readonly "gemini-2.5-pro": {
3764
+ readonly "accounts/fireworks/models/kimi-k2-instruct": {
3464
3765
  id: string;
3465
3766
  name: string;
3466
- api: "openai-completions";
3767
+ api: "anthropic-messages";
3467
3768
  provider: string;
3468
3769
  baseUrl: string;
3469
- headers: {
3470
- "User-Agent": string;
3471
- "Editor-Version": string;
3472
- "Editor-Plugin-Version": string;
3473
- "Copilot-Integration-Id": string;
3474
- };
3475
- compat: {
3476
- supportsStore: false;
3477
- supportsDeveloperRole: false;
3478
- supportsReasoningEffort: false;
3479
- };
3480
3770
  reasoning: false;
3481
- input: ("image" | "text")[];
3771
+ input: "text"[];
3482
3772
  cost: {
3483
3773
  input: number;
3484
3774
  output: number;
@@ -3488,25 +3778,14 @@ export declare const MODELS: {
3488
3778
  contextWindow: number;
3489
3779
  maxTokens: number;
3490
3780
  };
3491
- readonly "gemini-3-flash-preview": {
3781
+ readonly "accounts/fireworks/models/kimi-k2-thinking": {
3492
3782
  id: string;
3493
3783
  name: string;
3494
- api: "openai-completions";
3784
+ api: "anthropic-messages";
3495
3785
  provider: string;
3496
3786
  baseUrl: string;
3497
- headers: {
3498
- "User-Agent": string;
3499
- "Editor-Version": string;
3500
- "Editor-Plugin-Version": string;
3501
- "Copilot-Integration-Id": string;
3502
- };
3503
- compat: {
3504
- supportsStore: false;
3505
- supportsDeveloperRole: false;
3506
- supportsReasoningEffort: false;
3507
- };
3508
3787
  reasoning: true;
3509
- input: ("image" | "text")[];
3788
+ input: "text"[];
3510
3789
  cost: {
3511
3790
  input: number;
3512
3791
  output: number;
@@ -3516,23 +3795,12 @@ export declare const MODELS: {
3516
3795
  contextWindow: number;
3517
3796
  maxTokens: number;
3518
3797
  };
3519
- readonly "gemini-3-pro-preview": {
3798
+ readonly "accounts/fireworks/models/kimi-k2p5": {
3520
3799
  id: string;
3521
3800
  name: string;
3522
- api: "openai-completions";
3801
+ api: "anthropic-messages";
3523
3802
  provider: string;
3524
3803
  baseUrl: string;
3525
- headers: {
3526
- "User-Agent": string;
3527
- "Editor-Version": string;
3528
- "Editor-Plugin-Version": string;
3529
- "Copilot-Integration-Id": string;
3530
- };
3531
- compat: {
3532
- supportsStore: false;
3533
- supportsDeveloperRole: false;
3534
- supportsReasoningEffort: false;
3535
- };
3536
3804
  reasoning: true;
3537
3805
  input: ("image" | "text")[];
3538
3806
  cost: {
@@ -3544,23 +3812,12 @@ export declare const MODELS: {
3544
3812
  contextWindow: number;
3545
3813
  maxTokens: number;
3546
3814
  };
3547
- readonly "gemini-3.1-pro-preview": {
3815
+ readonly "accounts/fireworks/models/kimi-k2p6": {
3548
3816
  id: string;
3549
3817
  name: string;
3550
- api: "openai-completions";
3818
+ api: "anthropic-messages";
3551
3819
  provider: string;
3552
3820
  baseUrl: string;
3553
- headers: {
3554
- "User-Agent": string;
3555
- "Editor-Version": string;
3556
- "Editor-Plugin-Version": string;
3557
- "Copilot-Integration-Id": string;
3558
- };
3559
- compat: {
3560
- supportsStore: false;
3561
- supportsDeveloperRole: false;
3562
- supportsReasoningEffort: false;
3563
- };
3564
3821
  reasoning: true;
3565
3822
  input: ("image" | "text")[];
3566
3823
  cost: {
@@ -3572,25 +3829,14 @@ export declare const MODELS: {
3572
3829
  contextWindow: number;
3573
3830
  maxTokens: number;
3574
3831
  };
3575
- readonly "gpt-4.1": {
3832
+ readonly "accounts/fireworks/models/minimax-m2p1": {
3576
3833
  id: string;
3577
3834
  name: string;
3578
- api: "openai-completions";
3835
+ api: "anthropic-messages";
3579
3836
  provider: string;
3580
3837
  baseUrl: string;
3581
- headers: {
3582
- "User-Agent": string;
3583
- "Editor-Version": string;
3584
- "Editor-Plugin-Version": string;
3585
- "Copilot-Integration-Id": string;
3586
- };
3587
- compat: {
3588
- supportsStore: false;
3589
- supportsDeveloperRole: false;
3590
- supportsReasoningEffort: false;
3591
- };
3592
- reasoning: false;
3593
- input: ("image" | "text")[];
3838
+ reasoning: true;
3839
+ input: "text"[];
3594
3840
  cost: {
3595
3841
  input: number;
3596
3842
  output: number;
@@ -3600,25 +3846,14 @@ export declare const MODELS: {
3600
3846
  contextWindow: number;
3601
3847
  maxTokens: number;
3602
3848
  };
3603
- readonly "gpt-4o": {
3849
+ readonly "accounts/fireworks/models/minimax-m2p5": {
3604
3850
  id: string;
3605
3851
  name: string;
3606
- api: "openai-completions";
3852
+ api: "anthropic-messages";
3607
3853
  provider: string;
3608
3854
  baseUrl: string;
3609
- headers: {
3610
- "User-Agent": string;
3611
- "Editor-Version": string;
3612
- "Editor-Plugin-Version": string;
3613
- "Copilot-Integration-Id": string;
3614
- };
3615
- compat: {
3616
- supportsStore: false;
3617
- supportsDeveloperRole: false;
3618
- supportsReasoningEffort: false;
3619
- };
3620
- reasoning: false;
3621
- input: ("image" | "text")[];
3855
+ reasoning: true;
3856
+ input: "text"[];
3622
3857
  cost: {
3623
3858
  input: number;
3624
3859
  output: number;
@@ -3628,20 +3863,14 @@ export declare const MODELS: {
3628
3863
  contextWindow: number;
3629
3864
  maxTokens: number;
3630
3865
  };
3631
- readonly "gpt-5": {
3866
+ readonly "accounts/fireworks/models/minimax-m2p7": {
3632
3867
  id: string;
3633
3868
  name: string;
3634
- api: "openai-responses";
3869
+ api: "anthropic-messages";
3635
3870
  provider: string;
3636
3871
  baseUrl: string;
3637
- headers: {
3638
- "User-Agent": string;
3639
- "Editor-Version": string;
3640
- "Editor-Plugin-Version": string;
3641
- "Copilot-Integration-Id": string;
3642
- };
3643
3872
  reasoning: true;
3644
- input: ("image" | "text")[];
3873
+ input: "text"[];
3645
3874
  cost: {
3646
3875
  input: number;
3647
3876
  output: number;
@@ -3651,18 +3880,12 @@ export declare const MODELS: {
3651
3880
  contextWindow: number;
3652
3881
  maxTokens: number;
3653
3882
  };
3654
- readonly "gpt-5-mini": {
3883
+ readonly "accounts/fireworks/models/qwen3p6-plus": {
3655
3884
  id: string;
3656
3885
  name: string;
3657
- api: "openai-responses";
3886
+ api: "anthropic-messages";
3658
3887
  provider: string;
3659
3888
  baseUrl: string;
3660
- headers: {
3661
- "User-Agent": string;
3662
- "Editor-Version": string;
3663
- "Editor-Plugin-Version": string;
3664
- "Copilot-Integration-Id": string;
3665
- };
3666
3889
  reasoning: true;
3667
3890
  input: ("image" | "text")[];
3668
3891
  cost: {
@@ -3674,18 +3897,12 @@ export declare const MODELS: {
3674
3897
  contextWindow: number;
3675
3898
  maxTokens: number;
3676
3899
  };
3677
- readonly "gpt-5.1": {
3900
+ readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
3678
3901
  id: string;
3679
3902
  name: string;
3680
- api: "openai-responses";
3903
+ api: "anthropic-messages";
3681
3904
  provider: string;
3682
3905
  baseUrl: string;
3683
- headers: {
3684
- "User-Agent": string;
3685
- "Editor-Version": string;
3686
- "Editor-Plugin-Version": string;
3687
- "Copilot-Integration-Id": string;
3688
- };
3689
3906
  reasoning: true;
3690
3907
  input: ("image" | "text")[];
3691
3908
  cost: {
@@ -3697,10 +3914,12 @@ export declare const MODELS: {
3697
3914
  contextWindow: number;
3698
3915
  maxTokens: number;
3699
3916
  };
3700
- readonly "gpt-5.1-codex": {
3917
+ };
3918
+ readonly "github-copilot": {
3919
+ readonly "claude-haiku-4.5": {
3701
3920
  id: string;
3702
3921
  name: string;
3703
- api: "openai-responses";
3922
+ api: "anthropic-messages";
3704
3923
  provider: string;
3705
3924
  baseUrl: string;
3706
3925
  headers: {
@@ -3709,6 +3928,9 @@ export declare const MODELS: {
3709
3928
  "Editor-Plugin-Version": string;
3710
3929
  "Copilot-Integration-Id": string;
3711
3930
  };
3931
+ compat: {
3932
+ supportsEagerToolInputStreaming: false;
3933
+ };
3712
3934
  reasoning: true;
3713
3935
  input: ("image" | "text")[];
3714
3936
  cost: {
@@ -3720,10 +3942,10 @@ export declare const MODELS: {
3720
3942
  contextWindow: number;
3721
3943
  maxTokens: number;
3722
3944
  };
3723
- readonly "gpt-5.1-codex-max": {
3945
+ readonly "claude-opus-4.5": {
3724
3946
  id: string;
3725
3947
  name: string;
3726
- api: "openai-responses";
3948
+ api: "anthropic-messages";
3727
3949
  provider: string;
3728
3950
  baseUrl: string;
3729
3951
  headers: {
@@ -3743,10 +3965,10 @@ export declare const MODELS: {
3743
3965
  contextWindow: number;
3744
3966
  maxTokens: number;
3745
3967
  };
3746
- readonly "gpt-5.1-codex-mini": {
3968
+ readonly "claude-opus-4.6": {
3747
3969
  id: string;
3748
3970
  name: string;
3749
- api: "openai-responses";
3971
+ api: "anthropic-messages";
3750
3972
  provider: string;
3751
3973
  baseUrl: string;
3752
3974
  headers: {
@@ -3766,10 +3988,10 @@ export declare const MODELS: {
3766
3988
  contextWindow: number;
3767
3989
  maxTokens: number;
3768
3990
  };
3769
- readonly "gpt-5.2": {
3991
+ readonly "claude-opus-4.7": {
3770
3992
  id: string;
3771
3993
  name: string;
3772
- api: "openai-responses";
3994
+ api: "anthropic-messages";
3773
3995
  provider: string;
3774
3996
  baseUrl: string;
3775
3997
  headers: {
@@ -3789,10 +4011,10 @@ export declare const MODELS: {
3789
4011
  contextWindow: number;
3790
4012
  maxTokens: number;
3791
4013
  };
3792
- readonly "gpt-5.2-codex": {
4014
+ readonly "claude-sonnet-4": {
3793
4015
  id: string;
3794
4016
  name: string;
3795
- api: "openai-responses";
4017
+ api: "anthropic-messages";
3796
4018
  provider: string;
3797
4019
  baseUrl: string;
3798
4020
  headers: {
@@ -3801,6 +4023,9 @@ export declare const MODELS: {
3801
4023
  "Editor-Plugin-Version": string;
3802
4024
  "Copilot-Integration-Id": string;
3803
4025
  };
4026
+ compat: {
4027
+ supportsEagerToolInputStreaming: false;
4028
+ };
3804
4029
  reasoning: true;
3805
4030
  input: ("image" | "text")[];
3806
4031
  cost: {
@@ -3812,10 +4037,10 @@ export declare const MODELS: {
3812
4037
  contextWindow: number;
3813
4038
  maxTokens: number;
3814
4039
  };
3815
- readonly "gpt-5.3-codex": {
4040
+ readonly "claude-sonnet-4.5": {
3816
4041
  id: string;
3817
4042
  name: string;
3818
- api: "openai-responses";
4043
+ api: "anthropic-messages";
3819
4044
  provider: string;
3820
4045
  baseUrl: string;
3821
4046
  headers: {
@@ -3824,6 +4049,9 @@ export declare const MODELS: {
3824
4049
  "Editor-Plugin-Version": string;
3825
4050
  "Copilot-Integration-Id": string;
3826
4051
  };
4052
+ compat: {
4053
+ supportsEagerToolInputStreaming: false;
4054
+ };
3827
4055
  reasoning: true;
3828
4056
  input: ("image" | "text")[];
3829
4057
  cost: {
@@ -3835,10 +4063,10 @@ export declare const MODELS: {
3835
4063
  contextWindow: number;
3836
4064
  maxTokens: number;
3837
4065
  };
3838
- readonly "gpt-5.4": {
4066
+ readonly "claude-sonnet-4.6": {
3839
4067
  id: string;
3840
4068
  name: string;
3841
- api: "openai-responses";
4069
+ api: "anthropic-messages";
3842
4070
  provider: string;
3843
4071
  baseUrl: string;
3844
4072
  headers: {
@@ -3858,10 +4086,10 @@ export declare const MODELS: {
3858
4086
  contextWindow: number;
3859
4087
  maxTokens: number;
3860
4088
  };
3861
- readonly "gpt-5.4-mini": {
4089
+ readonly "gemini-2.5-pro": {
3862
4090
  id: string;
3863
4091
  name: string;
3864
- api: "openai-responses";
4092
+ api: "openai-completions";
3865
4093
  provider: string;
3866
4094
  baseUrl: string;
3867
4095
  headers: {
@@ -3870,7 +4098,12 @@ export declare const MODELS: {
3870
4098
  "Editor-Plugin-Version": string;
3871
4099
  "Copilot-Integration-Id": string;
3872
4100
  };
3873
- reasoning: true;
4101
+ compat: {
4102
+ supportsStore: false;
4103
+ supportsDeveloperRole: false;
4104
+ supportsReasoningEffort: false;
4105
+ };
4106
+ reasoning: false;
3874
4107
  input: ("image" | "text")[];
3875
4108
  cost: {
3876
4109
  input: number;
@@ -3881,10 +4114,10 @@ export declare const MODELS: {
3881
4114
  contextWindow: number;
3882
4115
  maxTokens: number;
3883
4116
  };
3884
- readonly "gpt-5.5": {
4117
+ readonly "gemini-3-flash-preview": {
3885
4118
  id: string;
3886
4119
  name: string;
3887
- api: "openai-responses";
4120
+ api: "openai-completions";
3888
4121
  provider: string;
3889
4122
  baseUrl: string;
3890
4123
  headers: {
@@ -3893,6 +4126,11 @@ export declare const MODELS: {
3893
4126
  "Editor-Plugin-Version": string;
3894
4127
  "Copilot-Integration-Id": string;
3895
4128
  };
4129
+ compat: {
4130
+ supportsStore: false;
4131
+ supportsDeveloperRole: false;
4132
+ supportsReasoningEffort: false;
4133
+ };
3896
4134
  reasoning: true;
3897
4135
  input: ("image" | "text")[];
3898
4136
  cost: {
@@ -3904,7 +4142,7 @@ export declare const MODELS: {
3904
4142
  contextWindow: number;
3905
4143
  maxTokens: number;
3906
4144
  };
3907
- readonly "grok-code-fast-1": {
4145
+ readonly "gemini-3-pro-preview": {
3908
4146
  id: string;
3909
4147
  name: string;
3910
4148
  api: "openai-completions";
@@ -3922,7 +4160,7 @@ export declare const MODELS: {
3922
4160
  supportsReasoningEffort: false;
3923
4161
  };
3924
4162
  reasoning: true;
3925
- input: "text"[];
4163
+ input: ("image" | "text")[];
3926
4164
  cost: {
3927
4165
  input: number;
3928
4166
  output: number;
@@ -3932,15 +4170,24 @@ export declare const MODELS: {
3932
4170
  contextWindow: number;
3933
4171
  maxTokens: number;
3934
4172
  };
3935
- };
3936
- readonly google: {
3937
- readonly "gemini-1.5-flash": {
4173
+ readonly "gemini-3.1-pro-preview": {
3938
4174
  id: string;
3939
4175
  name: string;
3940
- api: "google-generative-ai";
4176
+ api: "openai-completions";
3941
4177
  provider: string;
3942
4178
  baseUrl: string;
3943
- reasoning: false;
4179
+ headers: {
4180
+ "User-Agent": string;
4181
+ "Editor-Version": string;
4182
+ "Editor-Plugin-Version": string;
4183
+ "Copilot-Integration-Id": string;
4184
+ };
4185
+ compat: {
4186
+ supportsStore: false;
4187
+ supportsDeveloperRole: false;
4188
+ supportsReasoningEffort: false;
4189
+ };
4190
+ reasoning: true;
3944
4191
  input: ("image" | "text")[];
3945
4192
  cost: {
3946
4193
  input: number;
@@ -3951,12 +4198,23 @@ export declare const MODELS: {
3951
4198
  contextWindow: number;
3952
4199
  maxTokens: number;
3953
4200
  };
3954
- readonly "gemini-1.5-flash-8b": {
4201
+ readonly "gpt-4.1": {
3955
4202
  id: string;
3956
4203
  name: string;
3957
- api: "google-generative-ai";
4204
+ api: "openai-completions";
3958
4205
  provider: string;
3959
4206
  baseUrl: string;
4207
+ headers: {
4208
+ "User-Agent": string;
4209
+ "Editor-Version": string;
4210
+ "Editor-Plugin-Version": string;
4211
+ "Copilot-Integration-Id": string;
4212
+ };
4213
+ compat: {
4214
+ supportsStore: false;
4215
+ supportsDeveloperRole: false;
4216
+ supportsReasoningEffort: false;
4217
+ };
3960
4218
  reasoning: false;
3961
4219
  input: ("image" | "text")[];
3962
4220
  cost: {
@@ -3968,12 +4226,23 @@ export declare const MODELS: {
3968
4226
  contextWindow: number;
3969
4227
  maxTokens: number;
3970
4228
  };
3971
- readonly "gemini-1.5-pro": {
4229
+ readonly "gpt-4o": {
3972
4230
  id: string;
3973
4231
  name: string;
3974
- api: "google-generative-ai";
4232
+ api: "openai-completions";
3975
4233
  provider: string;
3976
4234
  baseUrl: string;
4235
+ headers: {
4236
+ "User-Agent": string;
4237
+ "Editor-Version": string;
4238
+ "Editor-Plugin-Version": string;
4239
+ "Copilot-Integration-Id": string;
4240
+ };
4241
+ compat: {
4242
+ supportsStore: false;
4243
+ supportsDeveloperRole: false;
4244
+ supportsReasoningEffort: false;
4245
+ };
3977
4246
  reasoning: false;
3978
4247
  input: ("image" | "text")[];
3979
4248
  cost: {
@@ -3985,13 +4254,19 @@ export declare const MODELS: {
3985
4254
  contextWindow: number;
3986
4255
  maxTokens: number;
3987
4256
  };
3988
- readonly "gemini-2.0-flash": {
4257
+ readonly "gpt-5": {
3989
4258
  id: string;
3990
4259
  name: string;
3991
- api: "google-generative-ai";
4260
+ api: "openai-responses";
3992
4261
  provider: string;
3993
4262
  baseUrl: string;
3994
- reasoning: false;
4263
+ headers: {
4264
+ "User-Agent": string;
4265
+ "Editor-Version": string;
4266
+ "Editor-Plugin-Version": string;
4267
+ "Copilot-Integration-Id": string;
4268
+ };
4269
+ reasoning: true;
3995
4270
  input: ("image" | "text")[];
3996
4271
  cost: {
3997
4272
  input: number;
@@ -4002,13 +4277,19 @@ export declare const MODELS: {
4002
4277
  contextWindow: number;
4003
4278
  maxTokens: number;
4004
4279
  };
4005
- readonly "gemini-2.0-flash-lite": {
4280
+ readonly "gpt-5-mini": {
4006
4281
  id: string;
4007
4282
  name: string;
4008
- api: "google-generative-ai";
4283
+ api: "openai-responses";
4009
4284
  provider: string;
4010
4285
  baseUrl: string;
4011
- reasoning: false;
4286
+ headers: {
4287
+ "User-Agent": string;
4288
+ "Editor-Version": string;
4289
+ "Editor-Plugin-Version": string;
4290
+ "Copilot-Integration-Id": string;
4291
+ };
4292
+ reasoning: true;
4012
4293
  input: ("image" | "text")[];
4013
4294
  cost: {
4014
4295
  input: number;
@@ -4019,12 +4300,18 @@ export declare const MODELS: {
4019
4300
  contextWindow: number;
4020
4301
  maxTokens: number;
4021
4302
  };
4022
- readonly "gemini-2.5-flash": {
4303
+ readonly "gpt-5.1": {
4023
4304
  id: string;
4024
4305
  name: string;
4025
- api: "google-generative-ai";
4306
+ api: "openai-responses";
4026
4307
  provider: string;
4027
4308
  baseUrl: string;
4309
+ headers: {
4310
+ "User-Agent": string;
4311
+ "Editor-Version": string;
4312
+ "Editor-Plugin-Version": string;
4313
+ "Copilot-Integration-Id": string;
4314
+ };
4028
4315
  reasoning: true;
4029
4316
  input: ("image" | "text")[];
4030
4317
  cost: {
@@ -4036,12 +4323,18 @@ export declare const MODELS: {
4036
4323
  contextWindow: number;
4037
4324
  maxTokens: number;
4038
4325
  };
4039
- readonly "gemini-2.5-flash-lite": {
4326
+ readonly "gpt-5.1-codex": {
4040
4327
  id: string;
4041
4328
  name: string;
4042
- api: "google-generative-ai";
4329
+ api: "openai-responses";
4043
4330
  provider: string;
4044
4331
  baseUrl: string;
4332
+ headers: {
4333
+ "User-Agent": string;
4334
+ "Editor-Version": string;
4335
+ "Editor-Plugin-Version": string;
4336
+ "Copilot-Integration-Id": string;
4337
+ };
4045
4338
  reasoning: true;
4046
4339
  input: ("image" | "text")[];
4047
4340
  cost: {
@@ -4053,12 +4346,18 @@ export declare const MODELS: {
4053
4346
  contextWindow: number;
4054
4347
  maxTokens: number;
4055
4348
  };
4056
- readonly "gemini-2.5-flash-lite-preview-06-17": {
4349
+ readonly "gpt-5.1-codex-max": {
4057
4350
  id: string;
4058
4351
  name: string;
4059
- api: "google-generative-ai";
4352
+ api: "openai-responses";
4060
4353
  provider: string;
4061
4354
  baseUrl: string;
4355
+ headers: {
4356
+ "User-Agent": string;
4357
+ "Editor-Version": string;
4358
+ "Editor-Plugin-Version": string;
4359
+ "Copilot-Integration-Id": string;
4360
+ };
4062
4361
  reasoning: true;
4063
4362
  input: ("image" | "text")[];
4064
4363
  cost: {
@@ -4070,12 +4369,18 @@ export declare const MODELS: {
4070
4369
  contextWindow: number;
4071
4370
  maxTokens: number;
4072
4371
  };
4073
- readonly "gemini-2.5-flash-lite-preview-09-2025": {
4372
+ readonly "gpt-5.1-codex-mini": {
4074
4373
  id: string;
4075
4374
  name: string;
4076
- api: "google-generative-ai";
4375
+ api: "openai-responses";
4077
4376
  provider: string;
4078
4377
  baseUrl: string;
4378
+ headers: {
4379
+ "User-Agent": string;
4380
+ "Editor-Version": string;
4381
+ "Editor-Plugin-Version": string;
4382
+ "Copilot-Integration-Id": string;
4383
+ };
4079
4384
  reasoning: true;
4080
4385
  input: ("image" | "text")[];
4081
4386
  cost: {
@@ -4087,12 +4392,18 @@ export declare const MODELS: {
4087
4392
  contextWindow: number;
4088
4393
  maxTokens: number;
4089
4394
  };
4090
- readonly "gemini-2.5-flash-preview-04-17": {
4395
+ readonly "gpt-5.2": {
4091
4396
  id: string;
4092
4397
  name: string;
4093
- api: "google-generative-ai";
4398
+ api: "openai-responses";
4094
4399
  provider: string;
4095
4400
  baseUrl: string;
4401
+ headers: {
4402
+ "User-Agent": string;
4403
+ "Editor-Version": string;
4404
+ "Editor-Plugin-Version": string;
4405
+ "Copilot-Integration-Id": string;
4406
+ };
4096
4407
  reasoning: true;
4097
4408
  input: ("image" | "text")[];
4098
4409
  cost: {
@@ -4104,12 +4415,18 @@ export declare const MODELS: {
4104
4415
  contextWindow: number;
4105
4416
  maxTokens: number;
4106
4417
  };
4107
- readonly "gemini-2.5-flash-preview-05-20": {
4418
+ readonly "gpt-5.2-codex": {
4108
4419
  id: string;
4109
4420
  name: string;
4110
- api: "google-generative-ai";
4421
+ api: "openai-responses";
4111
4422
  provider: string;
4112
4423
  baseUrl: string;
4424
+ headers: {
4425
+ "User-Agent": string;
4426
+ "Editor-Version": string;
4427
+ "Editor-Plugin-Version": string;
4428
+ "Copilot-Integration-Id": string;
4429
+ };
4113
4430
  reasoning: true;
4114
4431
  input: ("image" | "text")[];
4115
4432
  cost: {
@@ -4121,12 +4438,18 @@ export declare const MODELS: {
4121
4438
  contextWindow: number;
4122
4439
  maxTokens: number;
4123
4440
  };
4124
- readonly "gemini-2.5-flash-preview-09-2025": {
4441
+ readonly "gpt-5.3-codex": {
4125
4442
  id: string;
4126
4443
  name: string;
4127
- api: "google-generative-ai";
4444
+ api: "openai-responses";
4128
4445
  provider: string;
4129
4446
  baseUrl: string;
4447
+ headers: {
4448
+ "User-Agent": string;
4449
+ "Editor-Version": string;
4450
+ "Editor-Plugin-Version": string;
4451
+ "Copilot-Integration-Id": string;
4452
+ };
4130
4453
  reasoning: true;
4131
4454
  input: ("image" | "text")[];
4132
4455
  cost: {
@@ -4138,12 +4461,18 @@ export declare const MODELS: {
4138
4461
  contextWindow: number;
4139
4462
  maxTokens: number;
4140
4463
  };
4141
- readonly "gemini-2.5-pro": {
4464
+ readonly "gpt-5.4": {
4142
4465
  id: string;
4143
4466
  name: string;
4144
- api: "google-generative-ai";
4467
+ api: "openai-responses";
4145
4468
  provider: string;
4146
4469
  baseUrl: string;
4470
+ headers: {
4471
+ "User-Agent": string;
4472
+ "Editor-Version": string;
4473
+ "Editor-Plugin-Version": string;
4474
+ "Copilot-Integration-Id": string;
4475
+ };
4147
4476
  reasoning: true;
4148
4477
  input: ("image" | "text")[];
4149
4478
  cost: {
@@ -4155,12 +4484,18 @@ export declare const MODELS: {
4155
4484
  contextWindow: number;
4156
4485
  maxTokens: number;
4157
4486
  };
4158
- readonly "gemini-2.5-pro-preview-05-06": {
4487
+ readonly "gpt-5.4-mini": {
4159
4488
  id: string;
4160
4489
  name: string;
4161
- api: "google-generative-ai";
4490
+ api: "openai-responses";
4162
4491
  provider: string;
4163
4492
  baseUrl: string;
4493
+ headers: {
4494
+ "User-Agent": string;
4495
+ "Editor-Version": string;
4496
+ "Editor-Plugin-Version": string;
4497
+ "Copilot-Integration-Id": string;
4498
+ };
4164
4499
  reasoning: true;
4165
4500
  input: ("image" | "text")[];
4166
4501
  cost: {
@@ -4172,12 +4507,18 @@ export declare const MODELS: {
4172
4507
  contextWindow: number;
4173
4508
  maxTokens: number;
4174
4509
  };
4175
- readonly "gemini-2.5-pro-preview-06-05": {
4510
+ readonly "gpt-5.5": {
4176
4511
  id: string;
4177
4512
  name: string;
4178
- api: "google-generative-ai";
4513
+ api: "openai-responses";
4179
4514
  provider: string;
4180
4515
  baseUrl: string;
4516
+ headers: {
4517
+ "User-Agent": string;
4518
+ "Editor-Version": string;
4519
+ "Editor-Plugin-Version": string;
4520
+ "Copilot-Integration-Id": string;
4521
+ };
4181
4522
  reasoning: true;
4182
4523
  input: ("image" | "text")[];
4183
4524
  cost: {
@@ -4189,14 +4530,25 @@ export declare const MODELS: {
4189
4530
  contextWindow: number;
4190
4531
  maxTokens: number;
4191
4532
  };
4192
- readonly "gemini-3-flash-preview": {
4533
+ readonly "grok-code-fast-1": {
4193
4534
  id: string;
4194
4535
  name: string;
4195
- api: "google-generative-ai";
4536
+ api: "openai-completions";
4196
4537
  provider: string;
4197
4538
  baseUrl: string;
4539
+ headers: {
4540
+ "User-Agent": string;
4541
+ "Editor-Version": string;
4542
+ "Editor-Plugin-Version": string;
4543
+ "Copilot-Integration-Id": string;
4544
+ };
4545
+ compat: {
4546
+ supportsStore: false;
4547
+ supportsDeveloperRole: false;
4548
+ supportsReasoningEffort: false;
4549
+ };
4198
4550
  reasoning: true;
4199
- input: ("image" | "text")[];
4551
+ input: "text"[];
4200
4552
  cost: {
4201
4553
  input: number;
4202
4554
  output: number;
@@ -4206,13 +4558,15 @@ export declare const MODELS: {
4206
4558
  contextWindow: number;
4207
4559
  maxTokens: number;
4208
4560
  };
4209
- readonly "gemini-3-pro-preview": {
4561
+ };
4562
+ readonly google: {
4563
+ readonly "gemini-1.5-flash": {
4210
4564
  id: string;
4211
4565
  name: string;
4212
4566
  api: "google-generative-ai";
4213
4567
  provider: string;
4214
4568
  baseUrl: string;
4215
- reasoning: true;
4569
+ reasoning: false;
4216
4570
  input: ("image" | "text")[];
4217
4571
  cost: {
4218
4572
  input: number;
@@ -4223,13 +4577,13 @@ export declare const MODELS: {
4223
4577
  contextWindow: number;
4224
4578
  maxTokens: number;
4225
4579
  };
4226
- readonly "gemini-3.1-flash-lite-preview": {
4580
+ readonly "gemini-1.5-flash-8b": {
4227
4581
  id: string;
4228
4582
  name: string;
4229
4583
  api: "google-generative-ai";
4230
4584
  provider: string;
4231
4585
  baseUrl: string;
4232
- reasoning: true;
4586
+ reasoning: false;
4233
4587
  input: ("image" | "text")[];
4234
4588
  cost: {
4235
4589
  input: number;
@@ -4240,13 +4594,13 @@ export declare const MODELS: {
4240
4594
  contextWindow: number;
4241
4595
  maxTokens: number;
4242
4596
  };
4243
- readonly "gemini-3.1-pro-preview": {
4597
+ readonly "gemini-1.5-pro": {
4244
4598
  id: string;
4245
4599
  name: string;
4246
4600
  api: "google-generative-ai";
4247
4601
  provider: string;
4248
4602
  baseUrl: string;
4249
- reasoning: true;
4603
+ reasoning: false;
4250
4604
  input: ("image" | "text")[];
4251
4605
  cost: {
4252
4606
  input: number;
@@ -4257,13 +4611,13 @@ export declare const MODELS: {
4257
4611
  contextWindow: number;
4258
4612
  maxTokens: number;
4259
4613
  };
4260
- readonly "gemini-3.1-pro-preview-customtools": {
4614
+ readonly "gemini-2.0-flash": {
4261
4615
  id: string;
4262
4616
  name: string;
4263
4617
  api: "google-generative-ai";
4264
4618
  provider: string;
4265
4619
  baseUrl: string;
4266
- reasoning: true;
4620
+ reasoning: false;
4267
4621
  input: ("image" | "text")[];
4268
4622
  cost: {
4269
4623
  input: number;
@@ -4274,13 +4628,13 @@ export declare const MODELS: {
4274
4628
  contextWindow: number;
4275
4629
  maxTokens: number;
4276
4630
  };
4277
- readonly "gemini-flash-latest": {
4631
+ readonly "gemini-2.0-flash-lite": {
4278
4632
  id: string;
4279
4633
  name: string;
4280
4634
  api: "google-generative-ai";
4281
4635
  provider: string;
4282
4636
  baseUrl: string;
4283
- reasoning: true;
4637
+ reasoning: false;
4284
4638
  input: ("image" | "text")[];
4285
4639
  cost: {
4286
4640
  input: number;
@@ -4291,7 +4645,7 @@ export declare const MODELS: {
4291
4645
  contextWindow: number;
4292
4646
  maxTokens: number;
4293
4647
  };
4294
- readonly "gemini-flash-lite-latest": {
4648
+ readonly "gemini-2.5-flash": {
4295
4649
  id: string;
4296
4650
  name: string;
4297
4651
  api: "google-generative-ai";
@@ -4308,7 +4662,7 @@ export declare const MODELS: {
4308
4662
  contextWindow: number;
4309
4663
  maxTokens: number;
4310
4664
  };
4311
- readonly "gemini-live-2.5-flash": {
4665
+ readonly "gemini-2.5-flash-lite": {
4312
4666
  id: string;
4313
4667
  name: string;
4314
4668
  api: "google-generative-ai";
@@ -4325,14 +4679,14 @@ export declare const MODELS: {
4325
4679
  contextWindow: number;
4326
4680
  maxTokens: number;
4327
4681
  };
4328
- readonly "gemini-live-2.5-flash-preview-native-audio": {
4682
+ readonly "gemini-2.5-flash-lite-preview-06-17": {
4329
4683
  id: string;
4330
4684
  name: string;
4331
4685
  api: "google-generative-ai";
4332
4686
  provider: string;
4333
4687
  baseUrl: string;
4334
4688
  reasoning: true;
4335
- input: "text"[];
4689
+ input: ("image" | "text")[];
4336
4690
  cost: {
4337
4691
  input: number;
4338
4692
  output: number;
@@ -4342,13 +4696,13 @@ export declare const MODELS: {
4342
4696
  contextWindow: number;
4343
4697
  maxTokens: number;
4344
4698
  };
4345
- readonly "gemma-3-27b-it": {
4699
+ readonly "gemini-2.5-flash-lite-preview-09-2025": {
4346
4700
  id: string;
4347
4701
  name: string;
4348
4702
  api: "google-generative-ai";
4349
4703
  provider: string;
4350
4704
  baseUrl: string;
4351
- reasoning: false;
4705
+ reasoning: true;
4352
4706
  input: ("image" | "text")[];
4353
4707
  cost: {
4354
4708
  input: number;
@@ -4359,7 +4713,7 @@ export declare const MODELS: {
4359
4713
  contextWindow: number;
4360
4714
  maxTokens: number;
4361
4715
  };
4362
- readonly "gemma-4-26b-a4b-it": {
4716
+ readonly "gemini-2.5-flash-preview-04-17": {
4363
4717
  id: string;
4364
4718
  name: string;
4365
4719
  api: "google-generative-ai";
@@ -4376,7 +4730,7 @@ export declare const MODELS: {
4376
4730
  contextWindow: number;
4377
4731
  maxTokens: number;
4378
4732
  };
4379
- readonly "gemma-4-31b-it": {
4733
+ readonly "gemini-2.5-flash-preview-05-20": {
4380
4734
  id: string;
4381
4735
  name: string;
4382
4736
  api: "google-generative-ai";
@@ -4393,12 +4747,10 @@ export declare const MODELS: {
4393
4747
  contextWindow: number;
4394
4748
  maxTokens: number;
4395
4749
  };
4396
- };
4397
- readonly "google-antigravity": {
4398
- readonly "claude-opus-4-5-thinking": {
4750
+ readonly "gemini-2.5-flash-preview-09-2025": {
4399
4751
  id: string;
4400
4752
  name: string;
4401
- api: "google-gemini-cli";
4753
+ api: "google-generative-ai";
4402
4754
  provider: string;
4403
4755
  baseUrl: string;
4404
4756
  reasoning: true;
@@ -4412,10 +4764,10 @@ export declare const MODELS: {
4412
4764
  contextWindow: number;
4413
4765
  maxTokens: number;
4414
4766
  };
4415
- readonly "claude-opus-4-6-thinking": {
4767
+ readonly "gemini-2.5-pro": {
4416
4768
  id: string;
4417
4769
  name: string;
4418
- api: "google-gemini-cli";
4770
+ api: "google-generative-ai";
4419
4771
  provider: string;
4420
4772
  baseUrl: string;
4421
4773
  reasoning: true;
@@ -4429,13 +4781,13 @@ export declare const MODELS: {
4429
4781
  contextWindow: number;
4430
4782
  maxTokens: number;
4431
4783
  };
4432
- readonly "claude-sonnet-4-5": {
4784
+ readonly "gemini-2.5-pro-preview-05-06": {
4433
4785
  id: string;
4434
4786
  name: string;
4435
- api: "google-gemini-cli";
4787
+ api: "google-generative-ai";
4436
4788
  provider: string;
4437
4789
  baseUrl: string;
4438
- reasoning: false;
4790
+ reasoning: true;
4439
4791
  input: ("image" | "text")[];
4440
4792
  cost: {
4441
4793
  input: number;
@@ -4446,10 +4798,10 @@ export declare const MODELS: {
4446
4798
  contextWindow: number;
4447
4799
  maxTokens: number;
4448
4800
  };
4449
- readonly "claude-sonnet-4-5-thinking": {
4801
+ readonly "gemini-2.5-pro-preview-06-05": {
4450
4802
  id: string;
4451
4803
  name: string;
4452
- api: "google-gemini-cli";
4804
+ api: "google-generative-ai";
4453
4805
  provider: string;
4454
4806
  baseUrl: string;
4455
4807
  reasoning: true;
@@ -4463,10 +4815,10 @@ export declare const MODELS: {
4463
4815
  contextWindow: number;
4464
4816
  maxTokens: number;
4465
4817
  };
4466
- readonly "claude-sonnet-4-6": {
4818
+ readonly "gemini-3-flash-preview": {
4467
4819
  id: string;
4468
4820
  name: string;
4469
- api: "google-gemini-cli";
4821
+ api: "google-generative-ai";
4470
4822
  provider: string;
4471
4823
  baseUrl: string;
4472
4824
  reasoning: true;
@@ -4480,10 +4832,10 @@ export declare const MODELS: {
4480
4832
  contextWindow: number;
4481
4833
  maxTokens: number;
4482
4834
  };
4483
- readonly "gemini-3-flash": {
4835
+ readonly "gemini-3-pro-preview": {
4484
4836
  id: string;
4485
4837
  name: string;
4486
- api: "google-gemini-cli";
4838
+ api: "google-generative-ai";
4487
4839
  provider: string;
4488
4840
  baseUrl: string;
4489
4841
  reasoning: true;
@@ -4497,10 +4849,10 @@ export declare const MODELS: {
4497
4849
  contextWindow: number;
4498
4850
  maxTokens: number;
4499
4851
  };
4500
- readonly "gemini-3.1-pro-high": {
4852
+ readonly "gemini-3.1-flash-lite-preview": {
4501
4853
  id: string;
4502
4854
  name: string;
4503
- api: "google-gemini-cli";
4855
+ api: "google-generative-ai";
4504
4856
  provider: string;
4505
4857
  baseUrl: string;
4506
4858
  reasoning: true;
@@ -4514,10 +4866,10 @@ export declare const MODELS: {
4514
4866
  contextWindow: number;
4515
4867
  maxTokens: number;
4516
4868
  };
4517
- readonly "gemini-3.1-pro-low": {
4869
+ readonly "gemini-3.1-pro-preview": {
4518
4870
  id: string;
4519
4871
  name: string;
4520
- api: "google-gemini-cli";
4872
+ api: "google-generative-ai";
4521
4873
  provider: string;
4522
4874
  baseUrl: string;
4523
4875
  reasoning: true;
@@ -4531,14 +4883,14 @@ export declare const MODELS: {
4531
4883
  contextWindow: number;
4532
4884
  maxTokens: number;
4533
4885
  };
4534
- readonly "gpt-oss-120b-medium": {
4886
+ readonly "gemini-3.1-pro-preview-customtools": {
4535
4887
  id: string;
4536
4888
  name: string;
4537
- api: "google-gemini-cli";
4889
+ api: "google-generative-ai";
4538
4890
  provider: string;
4539
4891
  baseUrl: string;
4540
- reasoning: false;
4541
- input: "text"[];
4892
+ reasoning: true;
4893
+ input: ("image" | "text")[];
4542
4894
  cost: {
4543
4895
  input: number;
4544
4896
  output: number;
@@ -4548,15 +4900,13 @@ export declare const MODELS: {
4548
4900
  contextWindow: number;
4549
4901
  maxTokens: number;
4550
4902
  };
4551
- };
4552
- readonly "google-gemini-cli": {
4553
- readonly "gemini-2.0-flash": {
4903
+ readonly "gemini-flash-latest": {
4554
4904
  id: string;
4555
4905
  name: string;
4556
- api: "google-gemini-cli";
4906
+ api: "google-generative-ai";
4557
4907
  provider: string;
4558
4908
  baseUrl: string;
4559
- reasoning: false;
4909
+ reasoning: true;
4560
4910
  input: ("image" | "text")[];
4561
4911
  cost: {
4562
4912
  input: number;
@@ -4567,10 +4917,10 @@ export declare const MODELS: {
4567
4917
  contextWindow: number;
4568
4918
  maxTokens: number;
4569
4919
  };
4570
- readonly "gemini-2.5-flash": {
4920
+ readonly "gemini-flash-lite-latest": {
4571
4921
  id: string;
4572
4922
  name: string;
4573
- api: "google-gemini-cli";
4923
+ api: "google-generative-ai";
4574
4924
  provider: string;
4575
4925
  baseUrl: string;
4576
4926
  reasoning: true;
@@ -4584,10 +4934,10 @@ export declare const MODELS: {
4584
4934
  contextWindow: number;
4585
4935
  maxTokens: number;
4586
4936
  };
4587
- readonly "gemini-2.5-pro": {
4937
+ readonly "gemini-live-2.5-flash": {
4588
4938
  id: string;
4589
4939
  name: string;
4590
- api: "google-gemini-cli";
4940
+ api: "google-generative-ai";
4591
4941
  provider: string;
4592
4942
  baseUrl: string;
4593
4943
  reasoning: true;
@@ -4601,14 +4951,14 @@ export declare const MODELS: {
4601
4951
  contextWindow: number;
4602
4952
  maxTokens: number;
4603
4953
  };
4604
- readonly "gemini-3-flash-preview": {
4954
+ readonly "gemini-live-2.5-flash-preview-native-audio": {
4605
4955
  id: string;
4606
4956
  name: string;
4607
- api: "google-gemini-cli";
4957
+ api: "google-generative-ai";
4608
4958
  provider: string;
4609
4959
  baseUrl: string;
4610
4960
  reasoning: true;
4611
- input: ("image" | "text")[];
4961
+ input: "text"[];
4612
4962
  cost: {
4613
4963
  input: number;
4614
4964
  output: number;
@@ -4618,13 +4968,13 @@ export declare const MODELS: {
4618
4968
  contextWindow: number;
4619
4969
  maxTokens: number;
4620
4970
  };
4621
- readonly "gemini-3-pro-preview": {
4971
+ readonly "gemma-3-27b-it": {
4622
4972
  id: string;
4623
4973
  name: string;
4624
- api: "google-gemini-cli";
4974
+ api: "google-generative-ai";
4625
4975
  provider: string;
4626
4976
  baseUrl: string;
4627
- reasoning: true;
4977
+ reasoning: false;
4628
4978
  input: ("image" | "text")[];
4629
4979
  cost: {
4630
4980
  input: number;
@@ -4635,10 +4985,10 @@ export declare const MODELS: {
4635
4985
  contextWindow: number;
4636
4986
  maxTokens: number;
4637
4987
  };
4638
- readonly "gemini-3.1-flash-lite-preview": {
4988
+ readonly "gemma-4-26b-a4b-it": {
4639
4989
  id: string;
4640
4990
  name: string;
4641
- api: "google-gemini-cli";
4991
+ api: "google-generative-ai";
4642
4992
  provider: string;
4643
4993
  baseUrl: string;
4644
4994
  reasoning: true;
@@ -4652,10 +5002,10 @@ export declare const MODELS: {
4652
5002
  contextWindow: number;
4653
5003
  maxTokens: number;
4654
5004
  };
4655
- readonly "gemini-3.1-pro-preview": {
5005
+ readonly "gemma-4-31b-it": {
4656
5006
  id: string;
4657
5007
  name: string;
4658
- api: "google-gemini-cli";
5008
+ api: "google-generative-ai";
4659
5009
  provider: string;
4660
5010
  baseUrl: string;
4661
5011
  reasoning: true;
@@ -5709,10 +6059,252 @@ export declare const MODELS: {
5709
6059
  readonly "MiniMax-M2.7": {
5710
6060
  id: string;
5711
6061
  name: string;
5712
- api: "anthropic-messages";
6062
+ api: "anthropic-messages";
6063
+ provider: string;
6064
+ baseUrl: string;
6065
+ reasoning: true;
6066
+ input: "text"[];
6067
+ cost: {
6068
+ input: number;
6069
+ output: number;
6070
+ cacheRead: number;
6071
+ cacheWrite: number;
6072
+ };
6073
+ contextWindow: number;
6074
+ maxTokens: number;
6075
+ };
6076
+ readonly "MiniMax-M2.7-highspeed": {
6077
+ id: string;
6078
+ name: string;
6079
+ api: "anthropic-messages";
6080
+ provider: string;
6081
+ baseUrl: string;
6082
+ reasoning: true;
6083
+ input: "text"[];
6084
+ cost: {
6085
+ input: number;
6086
+ output: number;
6087
+ cacheRead: number;
6088
+ cacheWrite: number;
6089
+ };
6090
+ contextWindow: number;
6091
+ maxTokens: number;
6092
+ };
6093
+ };
6094
+ readonly "minimax-cn": {
6095
+ readonly "MiniMax-M2.7": {
6096
+ id: string;
6097
+ name: string;
6098
+ api: "anthropic-messages";
6099
+ provider: string;
6100
+ baseUrl: string;
6101
+ reasoning: true;
6102
+ input: "text"[];
6103
+ cost: {
6104
+ input: number;
6105
+ output: number;
6106
+ cacheRead: number;
6107
+ cacheWrite: number;
6108
+ };
6109
+ contextWindow: number;
6110
+ maxTokens: number;
6111
+ };
6112
+ readonly "MiniMax-M2.7-highspeed": {
6113
+ id: string;
6114
+ name: string;
6115
+ api: "anthropic-messages";
6116
+ provider: string;
6117
+ baseUrl: string;
6118
+ reasoning: true;
6119
+ input: "text"[];
6120
+ cost: {
6121
+ input: number;
6122
+ output: number;
6123
+ cacheRead: number;
6124
+ cacheWrite: number;
6125
+ };
6126
+ contextWindow: number;
6127
+ maxTokens: number;
6128
+ };
6129
+ };
6130
+ readonly mistral: {
6131
+ readonly "codestral-latest": {
6132
+ id: string;
6133
+ name: string;
6134
+ api: "mistral-conversations";
6135
+ provider: string;
6136
+ baseUrl: string;
6137
+ reasoning: false;
6138
+ input: "text"[];
6139
+ cost: {
6140
+ input: number;
6141
+ output: number;
6142
+ cacheRead: number;
6143
+ cacheWrite: number;
6144
+ };
6145
+ contextWindow: number;
6146
+ maxTokens: number;
6147
+ };
6148
+ readonly "devstral-2512": {
6149
+ id: string;
6150
+ name: string;
6151
+ api: "mistral-conversations";
6152
+ provider: string;
6153
+ baseUrl: string;
6154
+ reasoning: false;
6155
+ input: "text"[];
6156
+ cost: {
6157
+ input: number;
6158
+ output: number;
6159
+ cacheRead: number;
6160
+ cacheWrite: number;
6161
+ };
6162
+ contextWindow: number;
6163
+ maxTokens: number;
6164
+ };
6165
+ readonly "devstral-medium-2507": {
6166
+ id: string;
6167
+ name: string;
6168
+ api: "mistral-conversations";
6169
+ provider: string;
6170
+ baseUrl: string;
6171
+ reasoning: false;
6172
+ input: "text"[];
6173
+ cost: {
6174
+ input: number;
6175
+ output: number;
6176
+ cacheRead: number;
6177
+ cacheWrite: number;
6178
+ };
6179
+ contextWindow: number;
6180
+ maxTokens: number;
6181
+ };
6182
+ readonly "devstral-medium-latest": {
6183
+ id: string;
6184
+ name: string;
6185
+ api: "mistral-conversations";
6186
+ provider: string;
6187
+ baseUrl: string;
6188
+ reasoning: false;
6189
+ input: "text"[];
6190
+ cost: {
6191
+ input: number;
6192
+ output: number;
6193
+ cacheRead: number;
6194
+ cacheWrite: number;
6195
+ };
6196
+ contextWindow: number;
6197
+ maxTokens: number;
6198
+ };
6199
+ readonly "devstral-small-2505": {
6200
+ id: string;
6201
+ name: string;
6202
+ api: "mistral-conversations";
6203
+ provider: string;
6204
+ baseUrl: string;
6205
+ reasoning: false;
6206
+ input: "text"[];
6207
+ cost: {
6208
+ input: number;
6209
+ output: number;
6210
+ cacheRead: number;
6211
+ cacheWrite: number;
6212
+ };
6213
+ contextWindow: number;
6214
+ maxTokens: number;
6215
+ };
6216
+ readonly "devstral-small-2507": {
6217
+ id: string;
6218
+ name: string;
6219
+ api: "mistral-conversations";
6220
+ provider: string;
6221
+ baseUrl: string;
6222
+ reasoning: false;
6223
+ input: "text"[];
6224
+ cost: {
6225
+ input: number;
6226
+ output: number;
6227
+ cacheRead: number;
6228
+ cacheWrite: number;
6229
+ };
6230
+ contextWindow: number;
6231
+ maxTokens: number;
6232
+ };
6233
+ readonly "labs-devstral-small-2512": {
6234
+ id: string;
6235
+ name: string;
6236
+ api: "mistral-conversations";
6237
+ provider: string;
6238
+ baseUrl: string;
6239
+ reasoning: false;
6240
+ input: ("image" | "text")[];
6241
+ cost: {
6242
+ input: number;
6243
+ output: number;
6244
+ cacheRead: number;
6245
+ cacheWrite: number;
6246
+ };
6247
+ contextWindow: number;
6248
+ maxTokens: number;
6249
+ };
6250
+ readonly "magistral-medium-latest": {
6251
+ id: string;
6252
+ name: string;
6253
+ api: "mistral-conversations";
6254
+ provider: string;
6255
+ baseUrl: string;
6256
+ reasoning: true;
6257
+ input: "text"[];
6258
+ cost: {
6259
+ input: number;
6260
+ output: number;
6261
+ cacheRead: number;
6262
+ cacheWrite: number;
6263
+ };
6264
+ contextWindow: number;
6265
+ maxTokens: number;
6266
+ };
6267
+ readonly "magistral-small": {
6268
+ id: string;
6269
+ name: string;
6270
+ api: "mistral-conversations";
6271
+ provider: string;
6272
+ baseUrl: string;
6273
+ reasoning: true;
6274
+ input: "text"[];
6275
+ cost: {
6276
+ input: number;
6277
+ output: number;
6278
+ cacheRead: number;
6279
+ cacheWrite: number;
6280
+ };
6281
+ contextWindow: number;
6282
+ maxTokens: number;
6283
+ };
6284
+ readonly "ministral-3b-latest": {
6285
+ id: string;
6286
+ name: string;
6287
+ api: "mistral-conversations";
6288
+ provider: string;
6289
+ baseUrl: string;
6290
+ reasoning: false;
6291
+ input: "text"[];
6292
+ cost: {
6293
+ input: number;
6294
+ output: number;
6295
+ cacheRead: number;
6296
+ cacheWrite: number;
6297
+ };
6298
+ contextWindow: number;
6299
+ maxTokens: number;
6300
+ };
6301
+ readonly "ministral-8b-latest": {
6302
+ id: string;
6303
+ name: string;
6304
+ api: "mistral-conversations";
5713
6305
  provider: string;
5714
6306
  baseUrl: string;
5715
- reasoning: true;
6307
+ reasoning: false;
5716
6308
  input: "text"[];
5717
6309
  cost: {
5718
6310
  input: number;
@@ -5723,13 +6315,13 @@ export declare const MODELS: {
5723
6315
  contextWindow: number;
5724
6316
  maxTokens: number;
5725
6317
  };
5726
- readonly "MiniMax-M2.7-highspeed": {
6318
+ readonly "mistral-large-2411": {
5727
6319
  id: string;
5728
6320
  name: string;
5729
- api: "anthropic-messages";
6321
+ api: "mistral-conversations";
5730
6322
  provider: string;
5731
6323
  baseUrl: string;
5732
- reasoning: true;
6324
+ reasoning: false;
5733
6325
  input: "text"[];
5734
6326
  cost: {
5735
6327
  input: number;
@@ -5740,16 +6332,14 @@ export declare const MODELS: {
5740
6332
  contextWindow: number;
5741
6333
  maxTokens: number;
5742
6334
  };
5743
- };
5744
- readonly "minimax-cn": {
5745
- readonly "MiniMax-M2.7": {
6335
+ readonly "mistral-large-2512": {
5746
6336
  id: string;
5747
6337
  name: string;
5748
- api: "anthropic-messages";
6338
+ api: "mistral-conversations";
5749
6339
  provider: string;
5750
6340
  baseUrl: string;
5751
- reasoning: true;
5752
- input: "text"[];
6341
+ reasoning: false;
6342
+ input: ("image" | "text")[];
5753
6343
  cost: {
5754
6344
  input: number;
5755
6345
  output: number;
@@ -5759,14 +6349,14 @@ export declare const MODELS: {
5759
6349
  contextWindow: number;
5760
6350
  maxTokens: number;
5761
6351
  };
5762
- readonly "MiniMax-M2.7-highspeed": {
6352
+ readonly "mistral-large-latest": {
5763
6353
  id: string;
5764
6354
  name: string;
5765
- api: "anthropic-messages";
6355
+ api: "mistral-conversations";
5766
6356
  provider: string;
5767
6357
  baseUrl: string;
5768
- reasoning: true;
5769
- input: "text"[];
6358
+ reasoning: false;
6359
+ input: ("image" | "text")[];
5770
6360
  cost: {
5771
6361
  input: number;
5772
6362
  output: number;
@@ -5776,16 +6366,14 @@ export declare const MODELS: {
5776
6366
  contextWindow: number;
5777
6367
  maxTokens: number;
5778
6368
  };
5779
- };
5780
- readonly mistral: {
5781
- readonly "codestral-latest": {
6369
+ readonly "mistral-medium-2505": {
5782
6370
  id: string;
5783
6371
  name: string;
5784
6372
  api: "mistral-conversations";
5785
6373
  provider: string;
5786
6374
  baseUrl: string;
5787
6375
  reasoning: false;
5788
- input: "text"[];
6376
+ input: ("image" | "text")[];
5789
6377
  cost: {
5790
6378
  input: number;
5791
6379
  output: number;
@@ -5795,14 +6383,14 @@ export declare const MODELS: {
5795
6383
  contextWindow: number;
5796
6384
  maxTokens: number;
5797
6385
  };
5798
- readonly "devstral-2512": {
6386
+ readonly "mistral-medium-2508": {
5799
6387
  id: string;
5800
6388
  name: string;
5801
6389
  api: "mistral-conversations";
5802
6390
  provider: string;
5803
6391
  baseUrl: string;
5804
6392
  reasoning: false;
5805
- input: "text"[];
6393
+ input: ("image" | "text")[];
5806
6394
  cost: {
5807
6395
  input: number;
5808
6396
  output: number;
@@ -5812,14 +6400,14 @@ export declare const MODELS: {
5812
6400
  contextWindow: number;
5813
6401
  maxTokens: number;
5814
6402
  };
5815
- readonly "devstral-medium-2507": {
6403
+ readonly "mistral-medium-2604": {
5816
6404
  id: string;
5817
6405
  name: string;
5818
6406
  api: "mistral-conversations";
5819
6407
  provider: string;
5820
6408
  baseUrl: string;
5821
- reasoning: false;
5822
- input: "text"[];
6409
+ reasoning: true;
6410
+ input: ("image" | "text")[];
5823
6411
  cost: {
5824
6412
  input: number;
5825
6413
  output: number;
@@ -5829,14 +6417,14 @@ export declare const MODELS: {
5829
6417
  contextWindow: number;
5830
6418
  maxTokens: number;
5831
6419
  };
5832
- readonly "devstral-medium-latest": {
6420
+ readonly "mistral-medium-3.5": {
5833
6421
  id: string;
5834
6422
  name: string;
5835
6423
  api: "mistral-conversations";
5836
6424
  provider: string;
5837
6425
  baseUrl: string;
5838
- reasoning: false;
5839
- input: "text"[];
6426
+ reasoning: true;
6427
+ input: ("image" | "text")[];
5840
6428
  cost: {
5841
6429
  input: number;
5842
6430
  output: number;
@@ -5846,14 +6434,14 @@ export declare const MODELS: {
5846
6434
  contextWindow: number;
5847
6435
  maxTokens: number;
5848
6436
  };
5849
- readonly "devstral-small-2505": {
6437
+ readonly "mistral-medium-latest": {
5850
6438
  id: string;
5851
6439
  name: string;
5852
6440
  api: "mistral-conversations";
5853
6441
  provider: string;
5854
6442
  baseUrl: string;
5855
- reasoning: false;
5856
- input: "text"[];
6443
+ reasoning: true;
6444
+ input: ("image" | "text")[];
5857
6445
  cost: {
5858
6446
  input: number;
5859
6447
  output: number;
@@ -5863,7 +6451,7 @@ export declare const MODELS: {
5863
6451
  contextWindow: number;
5864
6452
  maxTokens: number;
5865
6453
  };
5866
- readonly "devstral-small-2507": {
6454
+ readonly "mistral-nemo": {
5867
6455
  id: string;
5868
6456
  name: string;
5869
6457
  api: "mistral-conversations";
@@ -5880,7 +6468,7 @@ export declare const MODELS: {
5880
6468
  contextWindow: number;
5881
6469
  maxTokens: number;
5882
6470
  };
5883
- readonly "labs-devstral-small-2512": {
6471
+ readonly "mistral-small-2506": {
5884
6472
  id: string;
5885
6473
  name: string;
5886
6474
  api: "mistral-conversations";
@@ -5897,14 +6485,14 @@ export declare const MODELS: {
5897
6485
  contextWindow: number;
5898
6486
  maxTokens: number;
5899
6487
  };
5900
- readonly "magistral-medium-latest": {
6488
+ readonly "mistral-small-2603": {
5901
6489
  id: string;
5902
6490
  name: string;
5903
6491
  api: "mistral-conversations";
5904
6492
  provider: string;
5905
6493
  baseUrl: string;
5906
6494
  reasoning: true;
5907
- input: "text"[];
6495
+ input: ("image" | "text")[];
5908
6496
  cost: {
5909
6497
  input: number;
5910
6498
  output: number;
@@ -5914,14 +6502,14 @@ export declare const MODELS: {
5914
6502
  contextWindow: number;
5915
6503
  maxTokens: number;
5916
6504
  };
5917
- readonly "magistral-small": {
6505
+ readonly "mistral-small-latest": {
5918
6506
  id: string;
5919
6507
  name: string;
5920
6508
  api: "mistral-conversations";
5921
6509
  provider: string;
5922
6510
  baseUrl: string;
5923
6511
  reasoning: true;
5924
- input: "text"[];
6512
+ input: ("image" | "text")[];
5925
6513
  cost: {
5926
6514
  input: number;
5927
6515
  output: number;
@@ -5931,7 +6519,7 @@ export declare const MODELS: {
5931
6519
  contextWindow: number;
5932
6520
  maxTokens: number;
5933
6521
  };
5934
- readonly "ministral-3b-latest": {
6522
+ readonly "open-mistral-7b": {
5935
6523
  id: string;
5936
6524
  name: string;
5937
6525
  api: "mistral-conversations";
@@ -5948,7 +6536,7 @@ export declare const MODELS: {
5948
6536
  contextWindow: number;
5949
6537
  maxTokens: number;
5950
6538
  };
5951
- readonly "ministral-8b-latest": {
6539
+ readonly "open-mixtral-8x22b": {
5952
6540
  id: string;
5953
6541
  name: string;
5954
6542
  api: "mistral-conversations";
@@ -5965,7 +6553,7 @@ export declare const MODELS: {
5965
6553
  contextWindow: number;
5966
6554
  maxTokens: number;
5967
6555
  };
5968
- readonly "mistral-large-2411": {
6556
+ readonly "open-mixtral-8x7b": {
5969
6557
  id: string;
5970
6558
  name: string;
5971
6559
  api: "mistral-conversations";
@@ -5982,7 +6570,7 @@ export declare const MODELS: {
5982
6570
  contextWindow: number;
5983
6571
  maxTokens: number;
5984
6572
  };
5985
- readonly "mistral-large-2512": {
6573
+ readonly "pixtral-12b": {
5986
6574
  id: string;
5987
6575
  name: string;
5988
6576
  api: "mistral-conversations";
@@ -5999,7 +6587,7 @@ export declare const MODELS: {
5999
6587
  contextWindow: number;
6000
6588
  maxTokens: number;
6001
6589
  };
6002
- readonly "mistral-large-latest": {
6590
+ readonly "pixtral-large-latest": {
6003
6591
  id: string;
6004
6592
  name: string;
6005
6593
  api: "mistral-conversations";
@@ -6016,14 +6604,23 @@ export declare const MODELS: {
6016
6604
  contextWindow: number;
6017
6605
  maxTokens: number;
6018
6606
  };
6019
- readonly "mistral-medium-2505": {
6607
+ };
6608
+ readonly moonshotai: {
6609
+ readonly "kimi-k2-0711-preview": {
6020
6610
  id: string;
6021
6611
  name: string;
6022
- api: "mistral-conversations";
6612
+ api: "openai-completions";
6023
6613
  provider: string;
6024
6614
  baseUrl: string;
6615
+ compat: {
6616
+ supportsStore: false;
6617
+ supportsDeveloperRole: false;
6618
+ supportsReasoningEffort: false;
6619
+ maxTokensField: "max_tokens";
6620
+ supportsStrictMode: false;
6621
+ };
6025
6622
  reasoning: false;
6026
- input: ("image" | "text")[];
6623
+ input: "text"[];
6027
6624
  cost: {
6028
6625
  input: number;
6029
6626
  output: number;
@@ -6033,14 +6630,21 @@ export declare const MODELS: {
6033
6630
  contextWindow: number;
6034
6631
  maxTokens: number;
6035
6632
  };
6036
- readonly "mistral-medium-2508": {
6633
+ readonly "kimi-k2-0905-preview": {
6037
6634
  id: string;
6038
6635
  name: string;
6039
- api: "mistral-conversations";
6636
+ api: "openai-completions";
6040
6637
  provider: string;
6041
6638
  baseUrl: string;
6639
+ compat: {
6640
+ supportsStore: false;
6641
+ supportsDeveloperRole: false;
6642
+ supportsReasoningEffort: false;
6643
+ maxTokensField: "max_tokens";
6644
+ supportsStrictMode: false;
6645
+ };
6042
6646
  reasoning: false;
6043
- input: ("image" | "text")[];
6647
+ input: "text"[];
6044
6648
  cost: {
6045
6649
  input: number;
6046
6650
  output: number;
@@ -6050,14 +6654,21 @@ export declare const MODELS: {
6050
6654
  contextWindow: number;
6051
6655
  maxTokens: number;
6052
6656
  };
6053
- readonly "mistral-medium-latest": {
6657
+ readonly "kimi-k2-thinking": {
6054
6658
  id: string;
6055
6659
  name: string;
6056
- api: "mistral-conversations";
6660
+ api: "openai-completions";
6057
6661
  provider: string;
6058
6662
  baseUrl: string;
6059
- reasoning: false;
6060
- input: ("image" | "text")[];
6663
+ compat: {
6664
+ supportsStore: false;
6665
+ supportsDeveloperRole: false;
6666
+ supportsReasoningEffort: false;
6667
+ maxTokensField: "max_tokens";
6668
+ supportsStrictMode: false;
6669
+ };
6670
+ reasoning: true;
6671
+ input: "text"[];
6061
6672
  cost: {
6062
6673
  input: number;
6063
6674
  output: number;
@@ -6067,13 +6678,20 @@ export declare const MODELS: {
6067
6678
  contextWindow: number;
6068
6679
  maxTokens: number;
6069
6680
  };
6070
- readonly "mistral-nemo": {
6681
+ readonly "kimi-k2-thinking-turbo": {
6071
6682
  id: string;
6072
6683
  name: string;
6073
- api: "mistral-conversations";
6684
+ api: "openai-completions";
6074
6685
  provider: string;
6075
6686
  baseUrl: string;
6076
- reasoning: false;
6687
+ compat: {
6688
+ supportsStore: false;
6689
+ supportsDeveloperRole: false;
6690
+ supportsReasoningEffort: false;
6691
+ maxTokensField: "max_tokens";
6692
+ supportsStrictMode: false;
6693
+ };
6694
+ reasoning: true;
6077
6695
  input: "text"[];
6078
6696
  cost: {
6079
6697
  input: number;
@@ -6084,14 +6702,21 @@ export declare const MODELS: {
6084
6702
  contextWindow: number;
6085
6703
  maxTokens: number;
6086
6704
  };
6087
- readonly "mistral-small-2506": {
6705
+ readonly "kimi-k2-turbo-preview": {
6088
6706
  id: string;
6089
6707
  name: string;
6090
- api: "mistral-conversations";
6708
+ api: "openai-completions";
6091
6709
  provider: string;
6092
6710
  baseUrl: string;
6711
+ compat: {
6712
+ supportsStore: false;
6713
+ supportsDeveloperRole: false;
6714
+ supportsReasoningEffort: false;
6715
+ maxTokensField: "max_tokens";
6716
+ supportsStrictMode: false;
6717
+ };
6093
6718
  reasoning: false;
6094
- input: ("image" | "text")[];
6719
+ input: "text"[];
6095
6720
  cost: {
6096
6721
  input: number;
6097
6722
  output: number;
@@ -6101,12 +6726,19 @@ export declare const MODELS: {
6101
6726
  contextWindow: number;
6102
6727
  maxTokens: number;
6103
6728
  };
6104
- readonly "mistral-small-2603": {
6729
+ readonly "kimi-k2.5": {
6105
6730
  id: string;
6106
6731
  name: string;
6107
- api: "mistral-conversations";
6732
+ api: "openai-completions";
6108
6733
  provider: string;
6109
6734
  baseUrl: string;
6735
+ compat: {
6736
+ supportsStore: false;
6737
+ supportsDeveloperRole: false;
6738
+ supportsReasoningEffort: false;
6739
+ maxTokensField: "max_tokens";
6740
+ supportsStrictMode: false;
6741
+ };
6110
6742
  reasoning: true;
6111
6743
  input: ("image" | "text")[];
6112
6744
  cost: {
@@ -6118,12 +6750,19 @@ export declare const MODELS: {
6118
6750
  contextWindow: number;
6119
6751
  maxTokens: number;
6120
6752
  };
6121
- readonly "mistral-small-latest": {
6753
+ readonly "kimi-k2.6": {
6122
6754
  id: string;
6123
6755
  name: string;
6124
- api: "mistral-conversations";
6756
+ api: "openai-completions";
6125
6757
  provider: string;
6126
6758
  baseUrl: string;
6759
+ compat: {
6760
+ supportsStore: false;
6761
+ supportsDeveloperRole: false;
6762
+ supportsReasoningEffort: false;
6763
+ maxTokensField: "max_tokens";
6764
+ supportsStrictMode: false;
6765
+ };
6127
6766
  reasoning: true;
6128
6767
  input: ("image" | "text")[];
6129
6768
  cost: {
@@ -6135,12 +6774,21 @@ export declare const MODELS: {
6135
6774
  contextWindow: number;
6136
6775
  maxTokens: number;
6137
6776
  };
6138
- readonly "open-mistral-7b": {
6777
+ };
6778
+ readonly "moonshotai-cn": {
6779
+ readonly "kimi-k2-0711-preview": {
6139
6780
  id: string;
6140
6781
  name: string;
6141
- api: "mistral-conversations";
6782
+ api: "openai-completions";
6142
6783
  provider: string;
6143
6784
  baseUrl: string;
6785
+ compat: {
6786
+ supportsStore: false;
6787
+ supportsDeveloperRole: false;
6788
+ supportsReasoningEffort: false;
6789
+ maxTokensField: "max_tokens";
6790
+ supportsStrictMode: false;
6791
+ };
6144
6792
  reasoning: false;
6145
6793
  input: "text"[];
6146
6794
  cost: {
@@ -6152,12 +6800,19 @@ export declare const MODELS: {
6152
6800
  contextWindow: number;
6153
6801
  maxTokens: number;
6154
6802
  };
6155
- readonly "open-mixtral-8x22b": {
6803
+ readonly "kimi-k2-0905-preview": {
6156
6804
  id: string;
6157
6805
  name: string;
6158
- api: "mistral-conversations";
6806
+ api: "openai-completions";
6159
6807
  provider: string;
6160
6808
  baseUrl: string;
6809
+ compat: {
6810
+ supportsStore: false;
6811
+ supportsDeveloperRole: false;
6812
+ supportsReasoningEffort: false;
6813
+ maxTokensField: "max_tokens";
6814
+ supportsStrictMode: false;
6815
+ };
6161
6816
  reasoning: false;
6162
6817
  input: "text"[];
6163
6818
  cost: {
@@ -6169,13 +6824,20 @@ export declare const MODELS: {
6169
6824
  contextWindow: number;
6170
6825
  maxTokens: number;
6171
6826
  };
6172
- readonly "open-mixtral-8x7b": {
6827
+ readonly "kimi-k2-thinking": {
6173
6828
  id: string;
6174
6829
  name: string;
6175
- api: "mistral-conversations";
6830
+ api: "openai-completions";
6176
6831
  provider: string;
6177
6832
  baseUrl: string;
6178
- reasoning: false;
6833
+ compat: {
6834
+ supportsStore: false;
6835
+ supportsDeveloperRole: false;
6836
+ supportsReasoningEffort: false;
6837
+ maxTokensField: "max_tokens";
6838
+ supportsStrictMode: false;
6839
+ };
6840
+ reasoning: true;
6179
6841
  input: "text"[];
6180
6842
  cost: {
6181
6843
  input: number;
@@ -6186,13 +6848,68 @@ export declare const MODELS: {
6186
6848
  contextWindow: number;
6187
6849
  maxTokens: number;
6188
6850
  };
6189
- readonly "pixtral-12b": {
6851
+ readonly "kimi-k2-thinking-turbo": {
6190
6852
  id: string;
6191
6853
  name: string;
6192
- api: "mistral-conversations";
6854
+ api: "openai-completions";
6855
+ provider: string;
6856
+ baseUrl: string;
6857
+ compat: {
6858
+ supportsStore: false;
6859
+ supportsDeveloperRole: false;
6860
+ supportsReasoningEffort: false;
6861
+ maxTokensField: "max_tokens";
6862
+ supportsStrictMode: false;
6863
+ };
6864
+ reasoning: true;
6865
+ input: "text"[];
6866
+ cost: {
6867
+ input: number;
6868
+ output: number;
6869
+ cacheRead: number;
6870
+ cacheWrite: number;
6871
+ };
6872
+ contextWindow: number;
6873
+ maxTokens: number;
6874
+ };
6875
+ readonly "kimi-k2-turbo-preview": {
6876
+ id: string;
6877
+ name: string;
6878
+ api: "openai-completions";
6193
6879
  provider: string;
6194
6880
  baseUrl: string;
6881
+ compat: {
6882
+ supportsStore: false;
6883
+ supportsDeveloperRole: false;
6884
+ supportsReasoningEffort: false;
6885
+ maxTokensField: "max_tokens";
6886
+ supportsStrictMode: false;
6887
+ };
6195
6888
  reasoning: false;
6889
+ input: "text"[];
6890
+ cost: {
6891
+ input: number;
6892
+ output: number;
6893
+ cacheRead: number;
6894
+ cacheWrite: number;
6895
+ };
6896
+ contextWindow: number;
6897
+ maxTokens: number;
6898
+ };
6899
+ readonly "kimi-k2.5": {
6900
+ id: string;
6901
+ name: string;
6902
+ api: "openai-completions";
6903
+ provider: string;
6904
+ baseUrl: string;
6905
+ compat: {
6906
+ supportsStore: false;
6907
+ supportsDeveloperRole: false;
6908
+ supportsReasoningEffort: false;
6909
+ maxTokensField: "max_tokens";
6910
+ supportsStrictMode: false;
6911
+ };
6912
+ reasoning: true;
6196
6913
  input: ("image" | "text")[];
6197
6914
  cost: {
6198
6915
  input: number;
@@ -6203,13 +6920,20 @@ export declare const MODELS: {
6203
6920
  contextWindow: number;
6204
6921
  maxTokens: number;
6205
6922
  };
6206
- readonly "pixtral-large-latest": {
6923
+ readonly "kimi-k2.6": {
6207
6924
  id: string;
6208
6925
  name: string;
6209
- api: "mistral-conversations";
6926
+ api: "openai-completions";
6210
6927
  provider: string;
6211
6928
  baseUrl: string;
6212
- reasoning: false;
6929
+ compat: {
6930
+ supportsStore: false;
6931
+ supportsDeveloperRole: false;
6932
+ supportsReasoningEffort: false;
6933
+ maxTokensField: "max_tokens";
6934
+ supportsStrictMode: false;
6935
+ };
6936
+ reasoning: true;
6213
6937
  input: ("image" | "text")[];
6214
6938
  cost: {
6215
6939
  input: number;
@@ -6715,7 +7439,24 @@ export declare const MODELS: {
6715
7439
  contextWindow: number;
6716
7440
  maxTokens: number;
6717
7441
  };
6718
- readonly "gpt-5.4-mini": {
7442
+ readonly "gpt-5.4-mini": {
7443
+ id: string;
7444
+ name: string;
7445
+ api: "openai-responses";
7446
+ provider: string;
7447
+ baseUrl: string;
7448
+ reasoning: true;
7449
+ input: ("image" | "text")[];
7450
+ cost: {
7451
+ input: number;
7452
+ output: number;
7453
+ cacheRead: number;
7454
+ cacheWrite: number;
7455
+ };
7456
+ contextWindow: number;
7457
+ maxTokens: number;
7458
+ };
7459
+ readonly "gpt-5.4-nano": {
6719
7460
  id: string;
6720
7461
  name: string;
6721
7462
  api: "openai-responses";
@@ -6732,7 +7473,7 @@ export declare const MODELS: {
6732
7473
  contextWindow: number;
6733
7474
  maxTokens: number;
6734
7475
  };
6735
- readonly "gpt-5.4-nano": {
7476
+ readonly "gpt-5.4-pro": {
6736
7477
  id: string;
6737
7478
  name: string;
6738
7479
  api: "openai-responses";
@@ -6749,7 +7490,7 @@ export declare const MODELS: {
6749
7490
  contextWindow: number;
6750
7491
  maxTokens: number;
6751
7492
  };
6752
- readonly "gpt-5.4-pro": {
7493
+ readonly "gpt-5.5": {
6753
7494
  id: string;
6754
7495
  name: string;
6755
7496
  api: "openai-responses";
@@ -6766,7 +7507,7 @@ export declare const MODELS: {
6766
7507
  contextWindow: number;
6767
7508
  maxTokens: number;
6768
7509
  };
6769
- readonly "gpt-5.5": {
7510
+ readonly "gpt-5.5-pro": {
6770
7511
  id: string;
6771
7512
  name: string;
6772
7513
  api: "openai-responses";
@@ -7637,23 +8378,6 @@ export declare const MODELS: {
7637
8378
  contextWindow: number;
7638
8379
  maxTokens: number;
7639
8380
  };
7640
- readonly "ling-2.6-flash-free": {
7641
- id: string;
7642
- name: string;
7643
- api: "openai-completions";
7644
- provider: string;
7645
- baseUrl: string;
7646
- reasoning: false;
7647
- input: "text"[];
7648
- cost: {
7649
- input: number;
7650
- output: number;
7651
- cacheRead: number;
7652
- cacheWrite: number;
7653
- };
7654
- contextWindow: number;
7655
- maxTokens: number;
7656
- };
7657
8381
  readonly "minimax-m2.5": {
7658
8382
  id: string;
7659
8383
  name: string;
@@ -7766,6 +8490,14 @@ export declare const MODELS: {
7766
8490
  baseUrl: string;
7767
8491
  compat: {
7768
8492
  requiresReasoningContentOnAssistantMessages: true;
8493
+ thinkingFormat: "deepseek";
8494
+ reasoningEffortMap: {
8495
+ minimal: string;
8496
+ low: string;
8497
+ medium: string;
8498
+ high: string;
8499
+ xhigh: string;
8500
+ };
7769
8501
  };
7770
8502
  reasoning: true;
7771
8503
  input: "text"[];
@@ -7786,6 +8518,14 @@ export declare const MODELS: {
7786
8518
  baseUrl: string;
7787
8519
  compat: {
7788
8520
  requiresReasoningContentOnAssistantMessages: true;
8521
+ thinkingFormat: "deepseek";
8522
+ reasoningEffortMap: {
8523
+ minimal: string;
8524
+ low: string;
8525
+ medium: string;
8526
+ high: string;
8527
+ xhigh: string;
8528
+ };
7789
8529
  };
7790
8530
  reasoning: true;
7791
8531
  input: "text"[];
@@ -8743,6 +9483,13 @@ export declare const MODELS: {
8743
9483
  baseUrl: string;
8744
9484
  compat: {
8745
9485
  requiresReasoningContentOnAssistantMessages: true;
9486
+ reasoningEffortMap: {
9487
+ minimal: string;
9488
+ low: string;
9489
+ medium: string;
9490
+ high: string;
9491
+ xhigh: string;
9492
+ };
8746
9493
  };
8747
9494
  reasoning: true;
8748
9495
  input: "text"[];
@@ -8763,6 +9510,13 @@ export declare const MODELS: {
8763
9510
  baseUrl: string;
8764
9511
  compat: {
8765
9512
  requiresReasoningContentOnAssistantMessages: true;
9513
+ reasoningEffortMap: {
9514
+ minimal: string;
9515
+ low: string;
9516
+ medium: string;
9517
+ high: string;
9518
+ xhigh: string;
9519
+ };
8766
9520
  };
8767
9521
  reasoning: true;
8768
9522
  input: "text"[];
@@ -9098,6 +9852,23 @@ export declare const MODELS: {
9098
9852
  contextWindow: number;
9099
9853
  maxTokens: number;
9100
9854
  };
9855
+ readonly "ibm-granite/granite-4.1-8b": {
9856
+ id: string;
9857
+ name: string;
9858
+ api: "openai-completions";
9859
+ provider: string;
9860
+ baseUrl: string;
9861
+ reasoning: false;
9862
+ input: "text"[];
9863
+ cost: {
9864
+ input: number;
9865
+ output: number;
9866
+ cacheRead: number;
9867
+ cacheWrite: number;
9868
+ };
9869
+ contextWindow: number;
9870
+ maxTokens: number;
9871
+ };
9101
9872
  readonly "inception/mercury-2": {
9102
9873
  id: string;
9103
9874
  name: string;
@@ -9132,7 +9903,7 @@ export declare const MODELS: {
9132
9903
  contextWindow: number;
9133
9904
  maxTokens: number;
9134
9905
  };
9135
- readonly "inclusionai/ling-2.6-flash:free": {
9906
+ readonly "inclusionai/ling-2.6-flash": {
9136
9907
  id: string;
9137
9908
  name: string;
9138
9909
  api: "openai-completions";
@@ -9659,23 +10430,6 @@ export declare const MODELS: {
9659
10430
  contextWindow: number;
9660
10431
  maxTokens: number;
9661
10432
  };
9662
- readonly "mistralai/mistral-small-creative": {
9663
- id: string;
9664
- name: string;
9665
- api: "openai-completions";
9666
- provider: string;
9667
- baseUrl: string;
9668
- reasoning: false;
9669
- input: "text"[];
9670
- cost: {
9671
- input: number;
9672
- output: number;
9673
- cacheRead: number;
9674
- cacheWrite: number;
9675
- };
9676
- contextWindow: number;
9677
- maxTokens: number;
9678
- };
9679
10433
  readonly "mistralai/mixtral-8x22b-instruct": {
9680
10434
  id: string;
9681
10435
  name: string;
@@ -11002,6 +11756,23 @@ export declare const MODELS: {
11002
11756
  contextWindow: number;
11003
11757
  maxTokens: number;
11004
11758
  };
11759
+ readonly "openrouter/owl-alpha": {
11760
+ id: string;
11761
+ name: string;
11762
+ api: "openai-completions";
11763
+ provider: string;
11764
+ baseUrl: string;
11765
+ reasoning: false;
11766
+ input: "text"[];
11767
+ cost: {
11768
+ input: number;
11769
+ output: number;
11770
+ cacheRead: number;
11771
+ cacheWrite: number;
11772
+ };
11773
+ contextWindow: number;
11774
+ maxTokens: number;
11775
+ };
11005
11776
  readonly "poolside/laguna-m.1:free": {
11006
11777
  id: string;
11007
11778
  name: string;
@@ -11852,23 +12623,6 @@ export declare const MODELS: {
11852
12623
  contextWindow: number;
11853
12624
  maxTokens: number;
11854
12625
  };
11855
- readonly "qwen/qwq-32b": {
11856
- id: string;
11857
- name: string;
11858
- api: "openai-completions";
11859
- provider: string;
11860
- baseUrl: string;
11861
- reasoning: true;
11862
- input: "text"[];
11863
- cost: {
11864
- input: number;
11865
- output: number;
11866
- cacheRead: number;
11867
- cacheWrite: number;
11868
- };
11869
- contextWindow: number;
11870
- maxTokens: number;
11871
- };
11872
12626
  readonly "rekaai/reka-edge": {
11873
12627
  id: string;
11874
12628
  name: string;
@@ -12175,6 +12929,23 @@ export declare const MODELS: {
12175
12929
  contextWindow: number;
12176
12930
  maxTokens: number;
12177
12931
  };
12932
+ readonly "x-ai/grok-4.3": {
12933
+ id: string;
12934
+ name: string;
12935
+ api: "openai-completions";
12936
+ provider: string;
12937
+ baseUrl: string;
12938
+ reasoning: true;
12939
+ input: ("image" | "text")[];
12940
+ cost: {
12941
+ input: number;
12942
+ output: number;
12943
+ cacheRead: number;
12944
+ cacheWrite: number;
12945
+ };
12946
+ contextWindow: number;
12947
+ maxTokens: number;
12948
+ };
12178
12949
  readonly "x-ai/grok-code-fast-1": {
12179
12950
  id: string;
12180
12951
  name: string;
@@ -12908,6 +13679,23 @@ export declare const MODELS: {
12908
13679
  contextWindow: number;
12909
13680
  maxTokens: number;
12910
13681
  };
13682
+ readonly "alibaba/qwen3.6-27b": {
13683
+ id: string;
13684
+ name: string;
13685
+ api: "anthropic-messages";
13686
+ provider: string;
13687
+ baseUrl: string;
13688
+ reasoning: true;
13689
+ input: ("image" | "text")[];
13690
+ cost: {
13691
+ input: number;
13692
+ output: number;
13693
+ cacheRead: number;
13694
+ cacheWrite: number;
13695
+ };
13696
+ contextWindow: number;
13697
+ maxTokens: number;
13698
+ };
12911
13699
  readonly "alibaba/qwen3.6-plus": {
12912
13700
  id: string;
12913
13701
  name: string;
@@ -13288,7 +14076,7 @@ export declare const MODELS: {
13288
14076
  api: "anthropic-messages";
13289
14077
  provider: string;
13290
14078
  baseUrl: string;
13291
- reasoning: true;
14079
+ reasoning: false;
13292
14080
  input: "text"[];
13293
14081
  cost: {
13294
14082
  input: number;
@@ -14812,23 +15600,6 @@ export declare const MODELS: {
14812
15600
  contextWindow: number;
14813
15601
  maxTokens: number;
14814
15602
  };
14815
- readonly "prime-intellect/intellect-3": {
14816
- id: string;
14817
- name: string;
14818
- api: "anthropic-messages";
14819
- provider: string;
14820
- baseUrl: string;
14821
- reasoning: true;
14822
- input: "text"[];
14823
- cost: {
14824
- input: number;
14825
- output: number;
14826
- cacheRead: number;
14827
- cacheWrite: number;
14828
- };
14829
- contextWindow: number;
14830
- maxTokens: number;
14831
- };
14832
15603
  readonly "xai/grok-3": {
14833
15604
  id: string;
14834
15605
  name: string;
@@ -15084,6 +15855,23 @@ export declare const MODELS: {
15084
15855
  contextWindow: number;
15085
15856
  maxTokens: number;
15086
15857
  };
15858
+ readonly "xai/grok-4.3": {
15859
+ id: string;
15860
+ name: string;
15861
+ api: "anthropic-messages";
15862
+ provider: string;
15863
+ baseUrl: string;
15864
+ reasoning: true;
15865
+ input: ("image" | "text")[];
15866
+ cost: {
15867
+ input: number;
15868
+ output: number;
15869
+ cacheRead: number;
15870
+ cacheWrite: number;
15871
+ };
15872
+ contextWindow: number;
15873
+ maxTokens: number;
15874
+ };
15087
15875
  readonly "xai/grok-code-fast-1": {
15088
15876
  id: string;
15089
15877
  name: string;
@@ -15135,6 +15923,40 @@ export declare const MODELS: {
15135
15923
  contextWindow: number;
15136
15924
  maxTokens: number;
15137
15925
  };
15926
+ readonly "xiaomi/mimo-v2.5": {
15927
+ id: string;
15928
+ name: string;
15929
+ api: "anthropic-messages";
15930
+ provider: string;
15931
+ baseUrl: string;
15932
+ reasoning: true;
15933
+ input: ("image" | "text")[];
15934
+ cost: {
15935
+ input: number;
15936
+ output: number;
15937
+ cacheRead: number;
15938
+ cacheWrite: number;
15939
+ };
15940
+ contextWindow: number;
15941
+ maxTokens: number;
15942
+ };
15943
+ readonly "xiaomi/mimo-v2.5-pro": {
15944
+ id: string;
15945
+ name: string;
15946
+ api: "anthropic-messages";
15947
+ provider: string;
15948
+ baseUrl: string;
15949
+ reasoning: true;
15950
+ input: ("image" | "text")[];
15951
+ cost: {
15952
+ input: number;
15953
+ output: number;
15954
+ cacheRead: number;
15955
+ cacheWrite: number;
15956
+ };
15957
+ contextWindow: number;
15958
+ maxTokens: number;
15959
+ };
15138
15960
  readonly "zai/glm-4.5": {
15139
15961
  id: string;
15140
15962
  name: string;
@@ -15715,6 +16537,23 @@ export declare const MODELS: {
15715
16537
  contextWindow: number;
15716
16538
  maxTokens: number;
15717
16539
  };
16540
+ readonly "grok-4.3": {
16541
+ id: string;
16542
+ name: string;
16543
+ api: "openai-completions";
16544
+ provider: string;
16545
+ baseUrl: string;
16546
+ reasoning: true;
16547
+ input: ("image" | "text")[];
16548
+ cost: {
16549
+ input: number;
16550
+ output: number;
16551
+ cacheRead: number;
16552
+ cacheWrite: number;
16553
+ };
16554
+ contextWindow: number;
16555
+ maxTokens: number;
16556
+ };
15718
16557
  readonly "grok-beta": {
15719
16558
  id: string;
15720
16559
  name: string;
@@ -15855,6 +16694,28 @@ export declare const MODELS: {
15855
16694
  contextWindow: number;
15856
16695
  maxTokens: number;
15857
16696
  };
16697
+ readonly "glm-5v-turbo": {
16698
+ id: string;
16699
+ name: string;
16700
+ api: "openai-completions";
16701
+ provider: string;
16702
+ baseUrl: string;
16703
+ compat: {
16704
+ supportsDeveloperRole: false;
16705
+ thinkingFormat: "zai";
16706
+ zaiToolStream: true;
16707
+ };
16708
+ reasoning: true;
16709
+ input: ("image" | "text")[];
16710
+ cost: {
16711
+ input: number;
16712
+ output: number;
16713
+ cacheRead: number;
16714
+ cacheWrite: number;
16715
+ };
16716
+ contextWindow: number;
16717
+ maxTokens: number;
16718
+ };
15858
16719
  };
15859
16720
  };
15860
16721
  //# sourceMappingURL=models.generated.d.ts.map