@mariozechner/pi-ai 0.70.5 → 0.71.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (68) hide show
  1. package/README.md +6 -31
  2. package/dist/env-api-keys.d.ts.map +1 -1
  3. package/dist/env-api-keys.js +4 -0
  4. package/dist/env-api-keys.js.map +1 -1
  5. package/dist/index.d.ts +1 -1
  6. package/dist/index.d.ts.map +1 -1
  7. package/dist/index.js.map +1 -1
  8. package/dist/models.d.ts +1 -1
  9. package/dist/models.d.ts.map +1 -1
  10. package/dist/models.generated.d.ts +1653 -596
  11. package/dist/models.generated.d.ts.map +1 -1
  12. package/dist/models.generated.js +1337 -422
  13. package/dist/models.generated.js.map +1 -1
  14. package/dist/models.js +3 -2
  15. package/dist/models.js.map +1 -1
  16. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  17. package/dist/providers/amazon-bedrock.js +14 -15
  18. package/dist/providers/amazon-bedrock.js.map +1 -1
  19. package/dist/providers/anthropic.d.ts.map +1 -1
  20. package/dist/providers/anthropic.js +38 -15
  21. package/dist/providers/anthropic.js.map +1 -1
  22. package/dist/providers/cloudflare.d.ts +13 -0
  23. package/dist/providers/cloudflare.d.ts.map +1 -0
  24. package/dist/providers/cloudflare.js +26 -0
  25. package/dist/providers/cloudflare.js.map +1 -0
  26. package/dist/providers/google-shared.d.ts +7 -2
  27. package/dist/providers/google-shared.d.ts.map +1 -1
  28. package/dist/providers/google-shared.js +4 -13
  29. package/dist/providers/google-shared.js.map +1 -1
  30. package/dist/providers/google-vertex.d.ts +1 -1
  31. package/dist/providers/google-vertex.d.ts.map +1 -1
  32. package/dist/providers/google-vertex.js.map +1 -1
  33. package/dist/providers/google.d.ts +1 -1
  34. package/dist/providers/google.d.ts.map +1 -1
  35. package/dist/providers/google.js.map +1 -1
  36. package/dist/providers/mistral.d.ts.map +1 -1
  37. package/dist/providers/mistral.js +1 -1
  38. package/dist/providers/mistral.js.map +1 -1
  39. package/dist/providers/openai-completions.d.ts.map +1 -1
  40. package/dist/providers/openai-completions.js +25 -8
  41. package/dist/providers/openai-completions.js.map +1 -1
  42. package/dist/providers/openai-responses.d.ts.map +1 -1
  43. package/dist/providers/openai-responses.js +10 -2
  44. package/dist/providers/openai-responses.js.map +1 -1
  45. package/dist/providers/register-builtins.d.ts +0 -3
  46. package/dist/providers/register-builtins.d.ts.map +1 -1
  47. package/dist/providers/register-builtins.js +0 -18
  48. package/dist/providers/register-builtins.js.map +1 -1
  49. package/dist/types.d.ts +3 -2
  50. package/dist/types.d.ts.map +1 -1
  51. package/dist/types.js.map +1 -1
  52. package/dist/utils/oauth/index.d.ts +0 -4
  53. package/dist/utils/oauth/index.d.ts.map +1 -1
  54. package/dist/utils/oauth/index.js +0 -10
  55. package/dist/utils/oauth/index.js.map +1 -1
  56. package/package.json +2 -6
  57. package/dist/providers/google-gemini-cli.d.ts +0 -74
  58. package/dist/providers/google-gemini-cli.d.ts.map +0 -1
  59. package/dist/providers/google-gemini-cli.js +0 -779
  60. package/dist/providers/google-gemini-cli.js.map +0 -1
  61. package/dist/utils/oauth/google-antigravity.d.ts +0 -26
  62. package/dist/utils/oauth/google-antigravity.d.ts.map +0 -1
  63. package/dist/utils/oauth/google-antigravity.js +0 -377
  64. package/dist/utils/oauth/google-antigravity.js.map +0 -1
  65. package/dist/utils/oauth/google-gemini-cli.d.ts +0 -26
  66. package/dist/utils/oauth/google-gemini-cli.d.ts.map +0 -1
  67. package/dist/utils/oauth/google-gemini-cli.js +0 -482
  68. package/dist/utils/oauth/google-gemini-cli.js.map +0 -1
@@ -2537,6 +2537,23 @@ export declare const MODELS: {
2537
2537
  contextWindow: number;
2538
2538
  maxTokens: number;
2539
2539
  };
2540
+ readonly "gpt-5.5-pro": {
2541
+ id: string;
2542
+ name: string;
2543
+ api: "azure-openai-responses";
2544
+ provider: string;
2545
+ baseUrl: string;
2546
+ reasoning: true;
2547
+ input: ("image" | "text")[];
2548
+ cost: {
2549
+ input: number;
2550
+ output: number;
2551
+ cacheRead: number;
2552
+ cacheWrite: number;
2553
+ };
2554
+ contextWindow: number;
2555
+ maxTokens: number;
2556
+ };
2540
2557
  readonly o1: {
2541
2558
  id: string;
2542
2559
  name: string;
@@ -2744,26 +2761,15 @@ export declare const MODELS: {
2744
2761
  maxTokens: number;
2745
2762
  };
2746
2763
  };
2747
- readonly deepseek: {
2748
- readonly "deepseek-v4-flash": {
2764
+ readonly "cloudflare-ai-gateway": {
2765
+ readonly "claude-3-5-haiku": {
2749
2766
  id: string;
2750
2767
  name: string;
2751
- api: "openai-completions";
2768
+ api: "anthropic-messages";
2752
2769
  provider: string;
2753
2770
  baseUrl: string;
2754
- compat: {
2755
- requiresReasoningContentOnAssistantMessages: true;
2756
- thinkingFormat: "deepseek";
2757
- reasoningEffortMap: {
2758
- minimal: string;
2759
- low: string;
2760
- medium: string;
2761
- high: string;
2762
- xhigh: string;
2763
- };
2764
- };
2765
- reasoning: true;
2766
- input: "text"[];
2771
+ reasoning: false;
2772
+ input: ("image" | "text")[];
2767
2773
  cost: {
2768
2774
  input: number;
2769
2775
  output: number;
@@ -2773,25 +2779,14 @@ export declare const MODELS: {
2773
2779
  contextWindow: number;
2774
2780
  maxTokens: number;
2775
2781
  };
2776
- readonly "deepseek-v4-pro": {
2782
+ readonly "claude-3-haiku": {
2777
2783
  id: string;
2778
2784
  name: string;
2779
- api: "openai-completions";
2785
+ api: "anthropic-messages";
2780
2786
  provider: string;
2781
2787
  baseUrl: string;
2782
- compat: {
2783
- requiresReasoningContentOnAssistantMessages: true;
2784
- thinkingFormat: "deepseek";
2785
- reasoningEffortMap: {
2786
- minimal: string;
2787
- low: string;
2788
- medium: string;
2789
- high: string;
2790
- xhigh: string;
2791
- };
2792
- };
2793
- reasoning: true;
2794
- input: "text"[];
2788
+ reasoning: false;
2789
+ input: ("image" | "text")[];
2795
2790
  cost: {
2796
2791
  input: number;
2797
2792
  output: number;
@@ -2801,16 +2796,14 @@ export declare const MODELS: {
2801
2796
  contextWindow: number;
2802
2797
  maxTokens: number;
2803
2798
  };
2804
- };
2805
- readonly fireworks: {
2806
- readonly "accounts/fireworks/models/deepseek-v3p1": {
2799
+ readonly "claude-3-opus": {
2807
2800
  id: string;
2808
2801
  name: string;
2809
2802
  api: "anthropic-messages";
2810
2803
  provider: string;
2811
2804
  baseUrl: string;
2812
- reasoning: true;
2813
- input: "text"[];
2805
+ reasoning: false;
2806
+ input: ("image" | "text")[];
2814
2807
  cost: {
2815
2808
  input: number;
2816
2809
  output: number;
@@ -2820,14 +2813,14 @@ export declare const MODELS: {
2820
2813
  contextWindow: number;
2821
2814
  maxTokens: number;
2822
2815
  };
2823
- readonly "accounts/fireworks/models/deepseek-v3p2": {
2816
+ readonly "claude-3-sonnet": {
2824
2817
  id: string;
2825
2818
  name: string;
2826
2819
  api: "anthropic-messages";
2827
2820
  provider: string;
2828
2821
  baseUrl: string;
2829
- reasoning: true;
2830
- input: "text"[];
2822
+ reasoning: false;
2823
+ input: ("image" | "text")[];
2831
2824
  cost: {
2832
2825
  input: number;
2833
2826
  output: number;
@@ -2837,14 +2830,14 @@ export declare const MODELS: {
2837
2830
  contextWindow: number;
2838
2831
  maxTokens: number;
2839
2832
  };
2840
- readonly "accounts/fireworks/models/glm-4p5": {
2833
+ readonly "claude-3.5-haiku": {
2841
2834
  id: string;
2842
2835
  name: string;
2843
2836
  api: "anthropic-messages";
2844
2837
  provider: string;
2845
2838
  baseUrl: string;
2846
- reasoning: true;
2847
- input: "text"[];
2839
+ reasoning: false;
2840
+ input: ("image" | "text")[];
2848
2841
  cost: {
2849
2842
  input: number;
2850
2843
  output: number;
@@ -2854,14 +2847,14 @@ export declare const MODELS: {
2854
2847
  contextWindow: number;
2855
2848
  maxTokens: number;
2856
2849
  };
2857
- readonly "accounts/fireworks/models/glm-4p5-air": {
2850
+ readonly "claude-3.5-sonnet": {
2858
2851
  id: string;
2859
2852
  name: string;
2860
2853
  api: "anthropic-messages";
2861
2854
  provider: string;
2862
2855
  baseUrl: string;
2863
- reasoning: true;
2864
- input: "text"[];
2856
+ reasoning: false;
2857
+ input: ("image" | "text")[];
2865
2858
  cost: {
2866
2859
  input: number;
2867
2860
  output: number;
@@ -2871,14 +2864,14 @@ export declare const MODELS: {
2871
2864
  contextWindow: number;
2872
2865
  maxTokens: number;
2873
2866
  };
2874
- readonly "accounts/fireworks/models/glm-4p7": {
2867
+ readonly "claude-haiku-4-5": {
2875
2868
  id: string;
2876
2869
  name: string;
2877
2870
  api: "anthropic-messages";
2878
2871
  provider: string;
2879
2872
  baseUrl: string;
2880
2873
  reasoning: true;
2881
- input: "text"[];
2874
+ input: ("image" | "text")[];
2882
2875
  cost: {
2883
2876
  input: number;
2884
2877
  output: number;
@@ -2888,14 +2881,14 @@ export declare const MODELS: {
2888
2881
  contextWindow: number;
2889
2882
  maxTokens: number;
2890
2883
  };
2891
- readonly "accounts/fireworks/models/glm-5": {
2884
+ readonly "claude-opus-4": {
2892
2885
  id: string;
2893
2886
  name: string;
2894
2887
  api: "anthropic-messages";
2895
2888
  provider: string;
2896
2889
  baseUrl: string;
2897
2890
  reasoning: true;
2898
- input: "text"[];
2891
+ input: ("image" | "text")[];
2899
2892
  cost: {
2900
2893
  input: number;
2901
2894
  output: number;
@@ -2905,14 +2898,14 @@ export declare const MODELS: {
2905
2898
  contextWindow: number;
2906
2899
  maxTokens: number;
2907
2900
  };
2908
- readonly "accounts/fireworks/models/glm-5p1": {
2901
+ readonly "claude-opus-4-1": {
2909
2902
  id: string;
2910
2903
  name: string;
2911
2904
  api: "anthropic-messages";
2912
2905
  provider: string;
2913
2906
  baseUrl: string;
2914
2907
  reasoning: true;
2915
- input: "text"[];
2908
+ input: ("image" | "text")[];
2916
2909
  cost: {
2917
2910
  input: number;
2918
2911
  output: number;
@@ -2922,14 +2915,14 @@ export declare const MODELS: {
2922
2915
  contextWindow: number;
2923
2916
  maxTokens: number;
2924
2917
  };
2925
- readonly "accounts/fireworks/models/gpt-oss-120b": {
2918
+ readonly "claude-opus-4-5": {
2926
2919
  id: string;
2927
2920
  name: string;
2928
2921
  api: "anthropic-messages";
2929
2922
  provider: string;
2930
2923
  baseUrl: string;
2931
2924
  reasoning: true;
2932
- input: "text"[];
2925
+ input: ("image" | "text")[];
2933
2926
  cost: {
2934
2927
  input: number;
2935
2928
  output: number;
@@ -2939,14 +2932,14 @@ export declare const MODELS: {
2939
2932
  contextWindow: number;
2940
2933
  maxTokens: number;
2941
2934
  };
2942
- readonly "accounts/fireworks/models/gpt-oss-20b": {
2935
+ readonly "claude-opus-4-6": {
2943
2936
  id: string;
2944
2937
  name: string;
2945
2938
  api: "anthropic-messages";
2946
2939
  provider: string;
2947
2940
  baseUrl: string;
2948
2941
  reasoning: true;
2949
- input: "text"[];
2942
+ input: ("image" | "text")[];
2950
2943
  cost: {
2951
2944
  input: number;
2952
2945
  output: number;
@@ -2956,14 +2949,14 @@ export declare const MODELS: {
2956
2949
  contextWindow: number;
2957
2950
  maxTokens: number;
2958
2951
  };
2959
- readonly "accounts/fireworks/models/kimi-k2-instruct": {
2952
+ readonly "claude-opus-4-7": {
2960
2953
  id: string;
2961
2954
  name: string;
2962
2955
  api: "anthropic-messages";
2963
2956
  provider: string;
2964
2957
  baseUrl: string;
2965
- reasoning: false;
2966
- input: "text"[];
2958
+ reasoning: true;
2959
+ input: ("image" | "text")[];
2967
2960
  cost: {
2968
2961
  input: number;
2969
2962
  output: number;
@@ -2973,14 +2966,14 @@ export declare const MODELS: {
2973
2966
  contextWindow: number;
2974
2967
  maxTokens: number;
2975
2968
  };
2976
- readonly "accounts/fireworks/models/kimi-k2-thinking": {
2969
+ readonly "claude-sonnet-4": {
2977
2970
  id: string;
2978
2971
  name: string;
2979
2972
  api: "anthropic-messages";
2980
2973
  provider: string;
2981
2974
  baseUrl: string;
2982
2975
  reasoning: true;
2983
- input: "text"[];
2976
+ input: ("image" | "text")[];
2984
2977
  cost: {
2985
2978
  input: number;
2986
2979
  output: number;
@@ -2990,7 +2983,7 @@ export declare const MODELS: {
2990
2983
  contextWindow: number;
2991
2984
  maxTokens: number;
2992
2985
  };
2993
- readonly "accounts/fireworks/models/kimi-k2p5": {
2986
+ readonly "claude-sonnet-4-5": {
2994
2987
  id: string;
2995
2988
  name: string;
2996
2989
  api: "anthropic-messages";
@@ -3007,7 +3000,7 @@ export declare const MODELS: {
3007
3000
  contextWindow: number;
3008
3001
  maxTokens: number;
3009
3002
  };
3010
- readonly "accounts/fireworks/models/kimi-k2p6": {
3003
+ readonly "claude-sonnet-4-6": {
3011
3004
  id: string;
3012
3005
  name: string;
3013
3006
  api: "anthropic-messages";
@@ -3024,13 +3017,13 @@ export declare const MODELS: {
3024
3017
  contextWindow: number;
3025
3018
  maxTokens: number;
3026
3019
  };
3027
- readonly "accounts/fireworks/models/minimax-m2p1": {
3020
+ readonly "gpt-4": {
3028
3021
  id: string;
3029
3022
  name: string;
3030
- api: "anthropic-messages";
3023
+ api: "openai-responses";
3031
3024
  provider: string;
3032
3025
  baseUrl: string;
3033
- reasoning: true;
3026
+ reasoning: false;
3034
3027
  input: "text"[];
3035
3028
  cost: {
3036
3029
  input: number;
@@ -3041,14 +3034,14 @@ export declare const MODELS: {
3041
3034
  contextWindow: number;
3042
3035
  maxTokens: number;
3043
3036
  };
3044
- readonly "accounts/fireworks/models/minimax-m2p5": {
3037
+ readonly "gpt-4-turbo": {
3045
3038
  id: string;
3046
3039
  name: string;
3047
- api: "anthropic-messages";
3040
+ api: "openai-responses";
3048
3041
  provider: string;
3049
3042
  baseUrl: string;
3050
- reasoning: true;
3051
- input: "text"[];
3043
+ reasoning: false;
3044
+ input: ("image" | "text")[];
3052
3045
  cost: {
3053
3046
  input: number;
3054
3047
  output: number;
@@ -3058,14 +3051,14 @@ export declare const MODELS: {
3058
3051
  contextWindow: number;
3059
3052
  maxTokens: number;
3060
3053
  };
3061
- readonly "accounts/fireworks/models/minimax-m2p7": {
3054
+ readonly "gpt-4o": {
3062
3055
  id: string;
3063
3056
  name: string;
3064
- api: "anthropic-messages";
3057
+ api: "openai-responses";
3065
3058
  provider: string;
3066
3059
  baseUrl: string;
3067
- reasoning: true;
3068
- input: "text"[];
3060
+ reasoning: false;
3061
+ input: ("image" | "text")[];
3069
3062
  cost: {
3070
3063
  input: number;
3071
3064
  output: number;
@@ -3075,13 +3068,13 @@ export declare const MODELS: {
3075
3068
  contextWindow: number;
3076
3069
  maxTokens: number;
3077
3070
  };
3078
- readonly "accounts/fireworks/models/qwen3p6-plus": {
3071
+ readonly "gpt-4o-mini": {
3079
3072
  id: string;
3080
3073
  name: string;
3081
- api: "anthropic-messages";
3074
+ api: "openai-responses";
3082
3075
  provider: string;
3083
3076
  baseUrl: string;
3084
- reasoning: true;
3077
+ reasoning: false;
3085
3078
  input: ("image" | "text")[];
3086
3079
  cost: {
3087
3080
  input: number;
@@ -3092,10 +3085,10 @@ export declare const MODELS: {
3092
3085
  contextWindow: number;
3093
3086
  maxTokens: number;
3094
3087
  };
3095
- readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
3088
+ readonly "gpt-5.1": {
3096
3089
  id: string;
3097
3090
  name: string;
3098
- api: "anthropic-messages";
3091
+ api: "openai-responses";
3099
3092
  provider: string;
3100
3093
  baseUrl: string;
3101
3094
  reasoning: true;
@@ -3109,23 +3102,12 @@ export declare const MODELS: {
3109
3102
  contextWindow: number;
3110
3103
  maxTokens: number;
3111
3104
  };
3112
- };
3113
- readonly "github-copilot": {
3114
- readonly "claude-haiku-4.5": {
3105
+ readonly "gpt-5.1-codex": {
3115
3106
  id: string;
3116
3107
  name: string;
3117
- api: "anthropic-messages";
3108
+ api: "openai-responses";
3118
3109
  provider: string;
3119
3110
  baseUrl: string;
3120
- headers: {
3121
- "User-Agent": string;
3122
- "Editor-Version": string;
3123
- "Editor-Plugin-Version": string;
3124
- "Copilot-Integration-Id": string;
3125
- };
3126
- compat: {
3127
- supportsEagerToolInputStreaming: false;
3128
- };
3129
3111
  reasoning: true;
3130
3112
  input: ("image" | "text")[];
3131
3113
  cost: {
@@ -3137,18 +3119,12 @@ export declare const MODELS: {
3137
3119
  contextWindow: number;
3138
3120
  maxTokens: number;
3139
3121
  };
3140
- readonly "claude-opus-4.5": {
3122
+ readonly "gpt-5.2": {
3141
3123
  id: string;
3142
3124
  name: string;
3143
- api: "anthropic-messages";
3125
+ api: "openai-responses";
3144
3126
  provider: string;
3145
3127
  baseUrl: string;
3146
- headers: {
3147
- "User-Agent": string;
3148
- "Editor-Version": string;
3149
- "Editor-Plugin-Version": string;
3150
- "Copilot-Integration-Id": string;
3151
- };
3152
3128
  reasoning: true;
3153
3129
  input: ("image" | "text")[];
3154
3130
  cost: {
@@ -3160,18 +3136,12 @@ export declare const MODELS: {
3160
3136
  contextWindow: number;
3161
3137
  maxTokens: number;
3162
3138
  };
3163
- readonly "claude-opus-4.6": {
3139
+ readonly "gpt-5.2-codex": {
3164
3140
  id: string;
3165
3141
  name: string;
3166
- api: "anthropic-messages";
3142
+ api: "openai-responses";
3167
3143
  provider: string;
3168
3144
  baseUrl: string;
3169
- headers: {
3170
- "User-Agent": string;
3171
- "Editor-Version": string;
3172
- "Editor-Plugin-Version": string;
3173
- "Copilot-Integration-Id": string;
3174
- };
3175
3145
  reasoning: true;
3176
3146
  input: ("image" | "text")[];
3177
3147
  cost: {
@@ -3183,18 +3153,12 @@ export declare const MODELS: {
3183
3153
  contextWindow: number;
3184
3154
  maxTokens: number;
3185
3155
  };
3186
- readonly "claude-opus-4.7": {
3156
+ readonly "gpt-5.3-codex": {
3187
3157
  id: string;
3188
3158
  name: string;
3189
- api: "anthropic-messages";
3159
+ api: "openai-responses";
3190
3160
  provider: string;
3191
3161
  baseUrl: string;
3192
- headers: {
3193
- "User-Agent": string;
3194
- "Editor-Version": string;
3195
- "Editor-Plugin-Version": string;
3196
- "Copilot-Integration-Id": string;
3197
- };
3198
3162
  reasoning: true;
3199
3163
  input: ("image" | "text")[];
3200
3164
  cost: {
@@ -3206,21 +3170,12 @@ export declare const MODELS: {
3206
3170
  contextWindow: number;
3207
3171
  maxTokens: number;
3208
3172
  };
3209
- readonly "claude-sonnet-4": {
3173
+ readonly "gpt-5.4": {
3210
3174
  id: string;
3211
3175
  name: string;
3212
- api: "anthropic-messages";
3176
+ api: "openai-responses";
3213
3177
  provider: string;
3214
3178
  baseUrl: string;
3215
- headers: {
3216
- "User-Agent": string;
3217
- "Editor-Version": string;
3218
- "Editor-Plugin-Version": string;
3219
- "Copilot-Integration-Id": string;
3220
- };
3221
- compat: {
3222
- supportsEagerToolInputStreaming: false;
3223
- };
3224
3179
  reasoning: true;
3225
3180
  input: ("image" | "text")[];
3226
3181
  cost: {
@@ -3232,21 +3187,12 @@ export declare const MODELS: {
3232
3187
  contextWindow: number;
3233
3188
  maxTokens: number;
3234
3189
  };
3235
- readonly "claude-sonnet-4.5": {
3190
+ readonly "gpt-5.5": {
3236
3191
  id: string;
3237
3192
  name: string;
3238
- api: "anthropic-messages";
3193
+ api: "openai-responses";
3239
3194
  provider: string;
3240
3195
  baseUrl: string;
3241
- headers: {
3242
- "User-Agent": string;
3243
- "Editor-Version": string;
3244
- "Editor-Plugin-Version": string;
3245
- "Copilot-Integration-Id": string;
3246
- };
3247
- compat: {
3248
- supportsEagerToolInputStreaming: false;
3249
- };
3250
3196
  reasoning: true;
3251
3197
  input: ("image" | "text")[];
3252
3198
  cost: {
@@ -3258,18 +3204,12 @@ export declare const MODELS: {
3258
3204
  contextWindow: number;
3259
3205
  maxTokens: number;
3260
3206
  };
3261
- readonly "claude-sonnet-4.6": {
3207
+ readonly o1: {
3262
3208
  id: string;
3263
3209
  name: string;
3264
- api: "anthropic-messages";
3210
+ api: "openai-responses";
3265
3211
  provider: string;
3266
3212
  baseUrl: string;
3267
- headers: {
3268
- "User-Agent": string;
3269
- "Editor-Version": string;
3270
- "Editor-Plugin-Version": string;
3271
- "Copilot-Integration-Id": string;
3272
- };
3273
3213
  reasoning: true;
3274
3214
  input: ("image" | "text")[];
3275
3215
  cost: {
@@ -3281,24 +3221,13 @@ export declare const MODELS: {
3281
3221
  contextWindow: number;
3282
3222
  maxTokens: number;
3283
3223
  };
3284
- readonly "gemini-2.5-pro": {
3224
+ readonly o3: {
3285
3225
  id: string;
3286
3226
  name: string;
3287
- api: "openai-completions";
3227
+ api: "openai-responses";
3288
3228
  provider: string;
3289
3229
  baseUrl: string;
3290
- headers: {
3291
- "User-Agent": string;
3292
- "Editor-Version": string;
3293
- "Editor-Plugin-Version": string;
3294
- "Copilot-Integration-Id": string;
3295
- };
3296
- compat: {
3297
- supportsStore: false;
3298
- supportsDeveloperRole: false;
3299
- supportsReasoningEffort: false;
3300
- };
3301
- reasoning: false;
3230
+ reasoning: true;
3302
3231
  input: ("image" | "text")[];
3303
3232
  cost: {
3304
3233
  input: number;
@@ -3309,25 +3238,901 @@ export declare const MODELS: {
3309
3238
  contextWindow: number;
3310
3239
  maxTokens: number;
3311
3240
  };
3312
- readonly "gemini-3-flash-preview": {
3241
+ readonly "o3-mini": {
3313
3242
  id: string;
3314
3243
  name: string;
3315
- api: "openai-completions";
3244
+ api: "openai-responses";
3316
3245
  provider: string;
3317
3246
  baseUrl: string;
3318
- headers: {
3319
- "User-Agent": string;
3320
- "Editor-Version": string;
3321
- "Editor-Plugin-Version": string;
3322
- "Copilot-Integration-Id": string;
3323
- };
3324
- compat: {
3325
- supportsStore: false;
3326
- supportsDeveloperRole: false;
3327
- supportsReasoningEffort: false;
3328
- };
3329
3247
  reasoning: true;
3330
- input: ("image" | "text")[];
3248
+ input: "text"[];
3249
+ cost: {
3250
+ input: number;
3251
+ output: number;
3252
+ cacheRead: number;
3253
+ cacheWrite: number;
3254
+ };
3255
+ contextWindow: number;
3256
+ maxTokens: number;
3257
+ };
3258
+ readonly "o3-pro": {
3259
+ id: string;
3260
+ name: string;
3261
+ api: "openai-responses";
3262
+ provider: string;
3263
+ baseUrl: string;
3264
+ reasoning: true;
3265
+ input: ("image" | "text")[];
3266
+ cost: {
3267
+ input: number;
3268
+ output: number;
3269
+ cacheRead: number;
3270
+ cacheWrite: number;
3271
+ };
3272
+ contextWindow: number;
3273
+ maxTokens: number;
3274
+ };
3275
+ readonly "o4-mini": {
3276
+ id: string;
3277
+ name: string;
3278
+ api: "openai-responses";
3279
+ provider: string;
3280
+ baseUrl: string;
3281
+ reasoning: true;
3282
+ input: ("image" | "text")[];
3283
+ cost: {
3284
+ input: number;
3285
+ output: number;
3286
+ cacheRead: number;
3287
+ cacheWrite: number;
3288
+ };
3289
+ contextWindow: number;
3290
+ maxTokens: number;
3291
+ };
3292
+ readonly "workers-ai/@cf/moonshotai/kimi-k2.5": {
3293
+ id: string;
3294
+ name: string;
3295
+ api: "openai-completions";
3296
+ provider: string;
3297
+ baseUrl: string;
3298
+ compat: {
3299
+ sendSessionAffinityHeaders: true;
3300
+ };
3301
+ reasoning: true;
3302
+ input: ("image" | "text")[];
3303
+ cost: {
3304
+ input: number;
3305
+ output: number;
3306
+ cacheRead: number;
3307
+ cacheWrite: number;
3308
+ };
3309
+ contextWindow: number;
3310
+ maxTokens: number;
3311
+ };
3312
+ readonly "workers-ai/@cf/moonshotai/kimi-k2.6": {
3313
+ id: string;
3314
+ name: string;
3315
+ api: "openai-completions";
3316
+ provider: string;
3317
+ baseUrl: string;
3318
+ compat: {
3319
+ sendSessionAffinityHeaders: true;
3320
+ };
3321
+ reasoning: true;
3322
+ input: ("image" | "text")[];
3323
+ cost: {
3324
+ input: number;
3325
+ output: number;
3326
+ cacheRead: number;
3327
+ cacheWrite: number;
3328
+ };
3329
+ contextWindow: number;
3330
+ maxTokens: number;
3331
+ };
3332
+ readonly "workers-ai/@cf/nvidia/nemotron-3-120b-a12b": {
3333
+ id: string;
3334
+ name: string;
3335
+ api: "openai-completions";
3336
+ provider: string;
3337
+ baseUrl: string;
3338
+ compat: {
3339
+ sendSessionAffinityHeaders: true;
3340
+ };
3341
+ reasoning: true;
3342
+ input: "text"[];
3343
+ cost: {
3344
+ input: number;
3345
+ output: number;
3346
+ cacheRead: number;
3347
+ cacheWrite: number;
3348
+ };
3349
+ contextWindow: number;
3350
+ maxTokens: number;
3351
+ };
3352
+ readonly "workers-ai/@cf/zai-org/glm-4.7-flash": {
3353
+ id: string;
3354
+ name: string;
3355
+ api: "openai-completions";
3356
+ provider: string;
3357
+ baseUrl: string;
3358
+ compat: {
3359
+ sendSessionAffinityHeaders: true;
3360
+ };
3361
+ reasoning: true;
3362
+ input: "text"[];
3363
+ cost: {
3364
+ input: number;
3365
+ output: number;
3366
+ cacheRead: number;
3367
+ cacheWrite: number;
3368
+ };
3369
+ contextWindow: number;
3370
+ maxTokens: number;
3371
+ };
3372
+ };
3373
+ readonly "cloudflare-workers-ai": {
3374
+ readonly "@cf/google/gemma-4-26b-a4b-it": {
3375
+ id: string;
3376
+ name: string;
3377
+ api: "openai-completions";
3378
+ provider: string;
3379
+ baseUrl: string;
3380
+ compat: {
3381
+ sendSessionAffinityHeaders: true;
3382
+ };
3383
+ reasoning: true;
3384
+ input: ("image" | "text")[];
3385
+ cost: {
3386
+ input: number;
3387
+ output: number;
3388
+ cacheRead: number;
3389
+ cacheWrite: number;
3390
+ };
3391
+ contextWindow: number;
3392
+ maxTokens: number;
3393
+ };
3394
+ readonly "@cf/meta/llama-4-scout-17b-16e-instruct": {
3395
+ id: string;
3396
+ name: string;
3397
+ api: "openai-completions";
3398
+ provider: string;
3399
+ baseUrl: string;
3400
+ compat: {
3401
+ sendSessionAffinityHeaders: true;
3402
+ };
3403
+ reasoning: false;
3404
+ input: ("image" | "text")[];
3405
+ cost: {
3406
+ input: number;
3407
+ output: number;
3408
+ cacheRead: number;
3409
+ cacheWrite: number;
3410
+ };
3411
+ contextWindow: number;
3412
+ maxTokens: number;
3413
+ };
3414
+ readonly "@cf/moonshotai/kimi-k2.5": {
3415
+ id: string;
3416
+ name: string;
3417
+ api: "openai-completions";
3418
+ provider: string;
3419
+ baseUrl: string;
3420
+ compat: {
3421
+ sendSessionAffinityHeaders: true;
3422
+ };
3423
+ reasoning: true;
3424
+ input: ("image" | "text")[];
3425
+ cost: {
3426
+ input: number;
3427
+ output: number;
3428
+ cacheRead: number;
3429
+ cacheWrite: number;
3430
+ };
3431
+ contextWindow: number;
3432
+ maxTokens: number;
3433
+ };
3434
+ readonly "@cf/moonshotai/kimi-k2.6": {
3435
+ id: string;
3436
+ name: string;
3437
+ api: "openai-completions";
3438
+ provider: string;
3439
+ baseUrl: string;
3440
+ compat: {
3441
+ sendSessionAffinityHeaders: true;
3442
+ };
3443
+ reasoning: true;
3444
+ input: ("image" | "text")[];
3445
+ cost: {
3446
+ input: number;
3447
+ output: number;
3448
+ cacheRead: number;
3449
+ cacheWrite: number;
3450
+ };
3451
+ contextWindow: number;
3452
+ maxTokens: number;
3453
+ };
3454
+ readonly "@cf/nvidia/nemotron-3-120b-a12b": {
3455
+ id: string;
3456
+ name: string;
3457
+ api: "openai-completions";
3458
+ provider: string;
3459
+ baseUrl: string;
3460
+ compat: {
3461
+ sendSessionAffinityHeaders: true;
3462
+ };
3463
+ reasoning: true;
3464
+ input: "text"[];
3465
+ cost: {
3466
+ input: number;
3467
+ output: number;
3468
+ cacheRead: number;
3469
+ cacheWrite: number;
3470
+ };
3471
+ contextWindow: number;
3472
+ maxTokens: number;
3473
+ };
3474
+ readonly "@cf/openai/gpt-oss-120b": {
3475
+ id: string;
3476
+ name: string;
3477
+ api: "openai-completions";
3478
+ provider: string;
3479
+ baseUrl: string;
3480
+ compat: {
3481
+ sendSessionAffinityHeaders: true;
3482
+ };
3483
+ reasoning: true;
3484
+ input: "text"[];
3485
+ cost: {
3486
+ input: number;
3487
+ output: number;
3488
+ cacheRead: number;
3489
+ cacheWrite: number;
3490
+ };
3491
+ contextWindow: number;
3492
+ maxTokens: number;
3493
+ };
3494
+ readonly "@cf/openai/gpt-oss-20b": {
3495
+ id: string;
3496
+ name: string;
3497
+ api: "openai-completions";
3498
+ provider: string;
3499
+ baseUrl: string;
3500
+ compat: {
3501
+ sendSessionAffinityHeaders: true;
3502
+ };
3503
+ reasoning: true;
3504
+ input: "text"[];
3505
+ cost: {
3506
+ input: number;
3507
+ output: number;
3508
+ cacheRead: number;
3509
+ cacheWrite: number;
3510
+ };
3511
+ contextWindow: number;
3512
+ maxTokens: number;
3513
+ };
3514
+ readonly "@cf/zai-org/glm-4.7-flash": {
3515
+ id: string;
3516
+ name: string;
3517
+ api: "openai-completions";
3518
+ provider: string;
3519
+ baseUrl: string;
3520
+ compat: {
3521
+ sendSessionAffinityHeaders: true;
3522
+ };
3523
+ reasoning: true;
3524
+ input: "text"[];
3525
+ cost: {
3526
+ input: number;
3527
+ output: number;
3528
+ cacheRead: number;
3529
+ cacheWrite: number;
3530
+ };
3531
+ contextWindow: number;
3532
+ maxTokens: number;
3533
+ };
3534
+ };
3535
+ readonly deepseek: {
3536
+ readonly "deepseek-v4-flash": {
3537
+ id: string;
3538
+ name: string;
3539
+ api: "openai-completions";
3540
+ provider: string;
3541
+ baseUrl: string;
3542
+ compat: {
3543
+ requiresReasoningContentOnAssistantMessages: true;
3544
+ thinkingFormat: "deepseek";
3545
+ reasoningEffortMap: {
3546
+ minimal: string;
3547
+ low: string;
3548
+ medium: string;
3549
+ high: string;
3550
+ xhigh: string;
3551
+ };
3552
+ };
3553
+ reasoning: true;
3554
+ input: "text"[];
3555
+ cost: {
3556
+ input: number;
3557
+ output: number;
3558
+ cacheRead: number;
3559
+ cacheWrite: number;
3560
+ };
3561
+ contextWindow: number;
3562
+ maxTokens: number;
3563
+ };
3564
+ readonly "deepseek-v4-pro": {
3565
+ id: string;
3566
+ name: string;
3567
+ api: "openai-completions";
3568
+ provider: string;
3569
+ baseUrl: string;
3570
+ compat: {
3571
+ requiresReasoningContentOnAssistantMessages: true;
3572
+ thinkingFormat: "deepseek";
3573
+ reasoningEffortMap: {
3574
+ minimal: string;
3575
+ low: string;
3576
+ medium: string;
3577
+ high: string;
3578
+ xhigh: string;
3579
+ };
3580
+ };
3581
+ reasoning: true;
3582
+ input: "text"[];
3583
+ cost: {
3584
+ input: number;
3585
+ output: number;
3586
+ cacheRead: number;
3587
+ cacheWrite: number;
3588
+ };
3589
+ contextWindow: number;
3590
+ maxTokens: number;
3591
+ };
3592
+ };
3593
+ readonly fireworks: {
3594
+ readonly "accounts/fireworks/models/deepseek-v3p1": {
3595
+ id: string;
3596
+ name: string;
3597
+ api: "anthropic-messages";
3598
+ provider: string;
3599
+ baseUrl: string;
3600
+ reasoning: true;
3601
+ input: "text"[];
3602
+ cost: {
3603
+ input: number;
3604
+ output: number;
3605
+ cacheRead: number;
3606
+ cacheWrite: number;
3607
+ };
3608
+ contextWindow: number;
3609
+ maxTokens: number;
3610
+ };
3611
+ readonly "accounts/fireworks/models/deepseek-v3p2": {
3612
+ id: string;
3613
+ name: string;
3614
+ api: "anthropic-messages";
3615
+ provider: string;
3616
+ baseUrl: string;
3617
+ reasoning: true;
3618
+ input: "text"[];
3619
+ cost: {
3620
+ input: number;
3621
+ output: number;
3622
+ cacheRead: number;
3623
+ cacheWrite: number;
3624
+ };
3625
+ contextWindow: number;
3626
+ maxTokens: number;
3627
+ };
3628
+ readonly "accounts/fireworks/models/deepseek-v4-pro": {
3629
+ id: string;
3630
+ name: string;
3631
+ api: "anthropic-messages";
3632
+ provider: string;
3633
+ baseUrl: string;
3634
+ reasoning: true;
3635
+ input: "text"[];
3636
+ cost: {
3637
+ input: number;
3638
+ output: number;
3639
+ cacheRead: number;
3640
+ cacheWrite: number;
3641
+ };
3642
+ contextWindow: number;
3643
+ maxTokens: number;
3644
+ };
3645
+ readonly "accounts/fireworks/models/glm-4p5": {
3646
+ id: string;
3647
+ name: string;
3648
+ api: "anthropic-messages";
3649
+ provider: string;
3650
+ baseUrl: string;
3651
+ reasoning: true;
3652
+ input: "text"[];
3653
+ cost: {
3654
+ input: number;
3655
+ output: number;
3656
+ cacheRead: number;
3657
+ cacheWrite: number;
3658
+ };
3659
+ contextWindow: number;
3660
+ maxTokens: number;
3661
+ };
3662
+ readonly "accounts/fireworks/models/glm-4p5-air": {
3663
+ id: string;
3664
+ name: string;
3665
+ api: "anthropic-messages";
3666
+ provider: string;
3667
+ baseUrl: string;
3668
+ reasoning: true;
3669
+ input: "text"[];
3670
+ cost: {
3671
+ input: number;
3672
+ output: number;
3673
+ cacheRead: number;
3674
+ cacheWrite: number;
3675
+ };
3676
+ contextWindow: number;
3677
+ maxTokens: number;
3678
+ };
3679
+ readonly "accounts/fireworks/models/glm-4p7": {
3680
+ id: string;
3681
+ name: string;
3682
+ api: "anthropic-messages";
3683
+ provider: string;
3684
+ baseUrl: string;
3685
+ reasoning: true;
3686
+ input: "text"[];
3687
+ cost: {
3688
+ input: number;
3689
+ output: number;
3690
+ cacheRead: number;
3691
+ cacheWrite: number;
3692
+ };
3693
+ contextWindow: number;
3694
+ maxTokens: number;
3695
+ };
3696
+ readonly "accounts/fireworks/models/glm-5": {
3697
+ id: string;
3698
+ name: string;
3699
+ api: "anthropic-messages";
3700
+ provider: string;
3701
+ baseUrl: string;
3702
+ reasoning: true;
3703
+ input: "text"[];
3704
+ cost: {
3705
+ input: number;
3706
+ output: number;
3707
+ cacheRead: number;
3708
+ cacheWrite: number;
3709
+ };
3710
+ contextWindow: number;
3711
+ maxTokens: number;
3712
+ };
3713
+ readonly "accounts/fireworks/models/glm-5p1": {
3714
+ id: string;
3715
+ name: string;
3716
+ api: "anthropic-messages";
3717
+ provider: string;
3718
+ baseUrl: string;
3719
+ reasoning: true;
3720
+ input: "text"[];
3721
+ cost: {
3722
+ input: number;
3723
+ output: number;
3724
+ cacheRead: number;
3725
+ cacheWrite: number;
3726
+ };
3727
+ contextWindow: number;
3728
+ maxTokens: number;
3729
+ };
3730
+ readonly "accounts/fireworks/models/gpt-oss-120b": {
3731
+ id: string;
3732
+ name: string;
3733
+ api: "anthropic-messages";
3734
+ provider: string;
3735
+ baseUrl: string;
3736
+ reasoning: true;
3737
+ input: "text"[];
3738
+ cost: {
3739
+ input: number;
3740
+ output: number;
3741
+ cacheRead: number;
3742
+ cacheWrite: number;
3743
+ };
3744
+ contextWindow: number;
3745
+ maxTokens: number;
3746
+ };
3747
+ readonly "accounts/fireworks/models/gpt-oss-20b": {
3748
+ id: string;
3749
+ name: string;
3750
+ api: "anthropic-messages";
3751
+ provider: string;
3752
+ baseUrl: string;
3753
+ reasoning: true;
3754
+ input: "text"[];
3755
+ cost: {
3756
+ input: number;
3757
+ output: number;
3758
+ cacheRead: number;
3759
+ cacheWrite: number;
3760
+ };
3761
+ contextWindow: number;
3762
+ maxTokens: number;
3763
+ };
3764
+ readonly "accounts/fireworks/models/kimi-k2-instruct": {
3765
+ id: string;
3766
+ name: string;
3767
+ api: "anthropic-messages";
3768
+ provider: string;
3769
+ baseUrl: string;
3770
+ reasoning: false;
3771
+ input: "text"[];
3772
+ cost: {
3773
+ input: number;
3774
+ output: number;
3775
+ cacheRead: number;
3776
+ cacheWrite: number;
3777
+ };
3778
+ contextWindow: number;
3779
+ maxTokens: number;
3780
+ };
3781
+ readonly "accounts/fireworks/models/kimi-k2-thinking": {
3782
+ id: string;
3783
+ name: string;
3784
+ api: "anthropic-messages";
3785
+ provider: string;
3786
+ baseUrl: string;
3787
+ reasoning: true;
3788
+ input: "text"[];
3789
+ cost: {
3790
+ input: number;
3791
+ output: number;
3792
+ cacheRead: number;
3793
+ cacheWrite: number;
3794
+ };
3795
+ contextWindow: number;
3796
+ maxTokens: number;
3797
+ };
3798
+ readonly "accounts/fireworks/models/kimi-k2p5": {
3799
+ id: string;
3800
+ name: string;
3801
+ api: "anthropic-messages";
3802
+ provider: string;
3803
+ baseUrl: string;
3804
+ reasoning: true;
3805
+ input: ("image" | "text")[];
3806
+ cost: {
3807
+ input: number;
3808
+ output: number;
3809
+ cacheRead: number;
3810
+ cacheWrite: number;
3811
+ };
3812
+ contextWindow: number;
3813
+ maxTokens: number;
3814
+ };
3815
+ readonly "accounts/fireworks/models/kimi-k2p6": {
3816
+ id: string;
3817
+ name: string;
3818
+ api: "anthropic-messages";
3819
+ provider: string;
3820
+ baseUrl: string;
3821
+ reasoning: true;
3822
+ input: ("image" | "text")[];
3823
+ cost: {
3824
+ input: number;
3825
+ output: number;
3826
+ cacheRead: number;
3827
+ cacheWrite: number;
3828
+ };
3829
+ contextWindow: number;
3830
+ maxTokens: number;
3831
+ };
3832
+ readonly "accounts/fireworks/models/minimax-m2p1": {
3833
+ id: string;
3834
+ name: string;
3835
+ api: "anthropic-messages";
3836
+ provider: string;
3837
+ baseUrl: string;
3838
+ reasoning: true;
3839
+ input: "text"[];
3840
+ cost: {
3841
+ input: number;
3842
+ output: number;
3843
+ cacheRead: number;
3844
+ cacheWrite: number;
3845
+ };
3846
+ contextWindow: number;
3847
+ maxTokens: number;
3848
+ };
3849
+ readonly "accounts/fireworks/models/minimax-m2p5": {
3850
+ id: string;
3851
+ name: string;
3852
+ api: "anthropic-messages";
3853
+ provider: string;
3854
+ baseUrl: string;
3855
+ reasoning: true;
3856
+ input: "text"[];
3857
+ cost: {
3858
+ input: number;
3859
+ output: number;
3860
+ cacheRead: number;
3861
+ cacheWrite: number;
3862
+ };
3863
+ contextWindow: number;
3864
+ maxTokens: number;
3865
+ };
3866
+ readonly "accounts/fireworks/models/minimax-m2p7": {
3867
+ id: string;
3868
+ name: string;
3869
+ api: "anthropic-messages";
3870
+ provider: string;
3871
+ baseUrl: string;
3872
+ reasoning: true;
3873
+ input: "text"[];
3874
+ cost: {
3875
+ input: number;
3876
+ output: number;
3877
+ cacheRead: number;
3878
+ cacheWrite: number;
3879
+ };
3880
+ contextWindow: number;
3881
+ maxTokens: number;
3882
+ };
3883
+ readonly "accounts/fireworks/models/qwen3p6-plus": {
3884
+ id: string;
3885
+ name: string;
3886
+ api: "anthropic-messages";
3887
+ provider: string;
3888
+ baseUrl: string;
3889
+ reasoning: true;
3890
+ input: ("image" | "text")[];
3891
+ cost: {
3892
+ input: number;
3893
+ output: number;
3894
+ cacheRead: number;
3895
+ cacheWrite: number;
3896
+ };
3897
+ contextWindow: number;
3898
+ maxTokens: number;
3899
+ };
3900
+ readonly "accounts/fireworks/routers/kimi-k2p5-turbo": {
3901
+ id: string;
3902
+ name: string;
3903
+ api: "anthropic-messages";
3904
+ provider: string;
3905
+ baseUrl: string;
3906
+ reasoning: true;
3907
+ input: ("image" | "text")[];
3908
+ cost: {
3909
+ input: number;
3910
+ output: number;
3911
+ cacheRead: number;
3912
+ cacheWrite: number;
3913
+ };
3914
+ contextWindow: number;
3915
+ maxTokens: number;
3916
+ };
3917
+ };
3918
+ readonly "github-copilot": {
3919
+ readonly "claude-haiku-4.5": {
3920
+ id: string;
3921
+ name: string;
3922
+ api: "anthropic-messages";
3923
+ provider: string;
3924
+ baseUrl: string;
3925
+ headers: {
3926
+ "User-Agent": string;
3927
+ "Editor-Version": string;
3928
+ "Editor-Plugin-Version": string;
3929
+ "Copilot-Integration-Id": string;
3930
+ };
3931
+ compat: {
3932
+ supportsEagerToolInputStreaming: false;
3933
+ };
3934
+ reasoning: true;
3935
+ input: ("image" | "text")[];
3936
+ cost: {
3937
+ input: number;
3938
+ output: number;
3939
+ cacheRead: number;
3940
+ cacheWrite: number;
3941
+ };
3942
+ contextWindow: number;
3943
+ maxTokens: number;
3944
+ };
3945
+ readonly "claude-opus-4.5": {
3946
+ id: string;
3947
+ name: string;
3948
+ api: "anthropic-messages";
3949
+ provider: string;
3950
+ baseUrl: string;
3951
+ headers: {
3952
+ "User-Agent": string;
3953
+ "Editor-Version": string;
3954
+ "Editor-Plugin-Version": string;
3955
+ "Copilot-Integration-Id": string;
3956
+ };
3957
+ reasoning: true;
3958
+ input: ("image" | "text")[];
3959
+ cost: {
3960
+ input: number;
3961
+ output: number;
3962
+ cacheRead: number;
3963
+ cacheWrite: number;
3964
+ };
3965
+ contextWindow: number;
3966
+ maxTokens: number;
3967
+ };
3968
+ readonly "claude-opus-4.6": {
3969
+ id: string;
3970
+ name: string;
3971
+ api: "anthropic-messages";
3972
+ provider: string;
3973
+ baseUrl: string;
3974
+ headers: {
3975
+ "User-Agent": string;
3976
+ "Editor-Version": string;
3977
+ "Editor-Plugin-Version": string;
3978
+ "Copilot-Integration-Id": string;
3979
+ };
3980
+ reasoning: true;
3981
+ input: ("image" | "text")[];
3982
+ cost: {
3983
+ input: number;
3984
+ output: number;
3985
+ cacheRead: number;
3986
+ cacheWrite: number;
3987
+ };
3988
+ contextWindow: number;
3989
+ maxTokens: number;
3990
+ };
3991
+ readonly "claude-opus-4.7": {
3992
+ id: string;
3993
+ name: string;
3994
+ api: "anthropic-messages";
3995
+ provider: string;
3996
+ baseUrl: string;
3997
+ headers: {
3998
+ "User-Agent": string;
3999
+ "Editor-Version": string;
4000
+ "Editor-Plugin-Version": string;
4001
+ "Copilot-Integration-Id": string;
4002
+ };
4003
+ reasoning: true;
4004
+ input: ("image" | "text")[];
4005
+ cost: {
4006
+ input: number;
4007
+ output: number;
4008
+ cacheRead: number;
4009
+ cacheWrite: number;
4010
+ };
4011
+ contextWindow: number;
4012
+ maxTokens: number;
4013
+ };
4014
+ readonly "claude-sonnet-4": {
4015
+ id: string;
4016
+ name: string;
4017
+ api: "anthropic-messages";
4018
+ provider: string;
4019
+ baseUrl: string;
4020
+ headers: {
4021
+ "User-Agent": string;
4022
+ "Editor-Version": string;
4023
+ "Editor-Plugin-Version": string;
4024
+ "Copilot-Integration-Id": string;
4025
+ };
4026
+ compat: {
4027
+ supportsEagerToolInputStreaming: false;
4028
+ };
4029
+ reasoning: true;
4030
+ input: ("image" | "text")[];
4031
+ cost: {
4032
+ input: number;
4033
+ output: number;
4034
+ cacheRead: number;
4035
+ cacheWrite: number;
4036
+ };
4037
+ contextWindow: number;
4038
+ maxTokens: number;
4039
+ };
4040
+ readonly "claude-sonnet-4.5": {
4041
+ id: string;
4042
+ name: string;
4043
+ api: "anthropic-messages";
4044
+ provider: string;
4045
+ baseUrl: string;
4046
+ headers: {
4047
+ "User-Agent": string;
4048
+ "Editor-Version": string;
4049
+ "Editor-Plugin-Version": string;
4050
+ "Copilot-Integration-Id": string;
4051
+ };
4052
+ compat: {
4053
+ supportsEagerToolInputStreaming: false;
4054
+ };
4055
+ reasoning: true;
4056
+ input: ("image" | "text")[];
4057
+ cost: {
4058
+ input: number;
4059
+ output: number;
4060
+ cacheRead: number;
4061
+ cacheWrite: number;
4062
+ };
4063
+ contextWindow: number;
4064
+ maxTokens: number;
4065
+ };
4066
+ readonly "claude-sonnet-4.6": {
4067
+ id: string;
4068
+ name: string;
4069
+ api: "anthropic-messages";
4070
+ provider: string;
4071
+ baseUrl: string;
4072
+ headers: {
4073
+ "User-Agent": string;
4074
+ "Editor-Version": string;
4075
+ "Editor-Plugin-Version": string;
4076
+ "Copilot-Integration-Id": string;
4077
+ };
4078
+ reasoning: true;
4079
+ input: ("image" | "text")[];
4080
+ cost: {
4081
+ input: number;
4082
+ output: number;
4083
+ cacheRead: number;
4084
+ cacheWrite: number;
4085
+ };
4086
+ contextWindow: number;
4087
+ maxTokens: number;
4088
+ };
4089
+ readonly "gemini-2.5-pro": {
4090
+ id: string;
4091
+ name: string;
4092
+ api: "openai-completions";
4093
+ provider: string;
4094
+ baseUrl: string;
4095
+ headers: {
4096
+ "User-Agent": string;
4097
+ "Editor-Version": string;
4098
+ "Editor-Plugin-Version": string;
4099
+ "Copilot-Integration-Id": string;
4100
+ };
4101
+ compat: {
4102
+ supportsStore: false;
4103
+ supportsDeveloperRole: false;
4104
+ supportsReasoningEffort: false;
4105
+ };
4106
+ reasoning: false;
4107
+ input: ("image" | "text")[];
4108
+ cost: {
4109
+ input: number;
4110
+ output: number;
4111
+ cacheRead: number;
4112
+ cacheWrite: number;
4113
+ };
4114
+ contextWindow: number;
4115
+ maxTokens: number;
4116
+ };
4117
+ readonly "gemini-3-flash-preview": {
4118
+ id: string;
4119
+ name: string;
4120
+ api: "openai-completions";
4121
+ provider: string;
4122
+ baseUrl: string;
4123
+ headers: {
4124
+ "User-Agent": string;
4125
+ "Editor-Version": string;
4126
+ "Editor-Plugin-Version": string;
4127
+ "Copilot-Integration-Id": string;
4128
+ };
4129
+ compat: {
4130
+ supportsStore: false;
4131
+ supportsDeveloperRole: false;
4132
+ supportsReasoningEffort: false;
4133
+ };
4134
+ reasoning: true;
4135
+ input: ("image" | "text")[];
3331
4136
  cost: {
3332
4137
  input: number;
3333
4138
  output: number;
@@ -4169,283 +4974,7 @@ export declare const MODELS: {
4169
4974
  api: "google-generative-ai";
4170
4975
  provider: string;
4171
4976
  baseUrl: string;
4172
- reasoning: false;
4173
- input: ("image" | "text")[];
4174
- cost: {
4175
- input: number;
4176
- output: number;
4177
- cacheRead: number;
4178
- cacheWrite: number;
4179
- };
4180
- contextWindow: number;
4181
- maxTokens: number;
4182
- };
4183
- readonly "gemma-4-26b-a4b-it": {
4184
- id: string;
4185
- name: string;
4186
- api: "google-generative-ai";
4187
- provider: string;
4188
- baseUrl: string;
4189
- reasoning: true;
4190
- input: ("image" | "text")[];
4191
- cost: {
4192
- input: number;
4193
- output: number;
4194
- cacheRead: number;
4195
- cacheWrite: number;
4196
- };
4197
- contextWindow: number;
4198
- maxTokens: number;
4199
- };
4200
- readonly "gemma-4-31b-it": {
4201
- id: string;
4202
- name: string;
4203
- api: "google-generative-ai";
4204
- provider: string;
4205
- baseUrl: string;
4206
- reasoning: true;
4207
- input: ("image" | "text")[];
4208
- cost: {
4209
- input: number;
4210
- output: number;
4211
- cacheRead: number;
4212
- cacheWrite: number;
4213
- };
4214
- contextWindow: number;
4215
- maxTokens: number;
4216
- };
4217
- };
4218
- readonly "google-antigravity": {
4219
- readonly "claude-opus-4-5-thinking": {
4220
- id: string;
4221
- name: string;
4222
- api: "google-gemini-cli";
4223
- provider: string;
4224
- baseUrl: string;
4225
- reasoning: true;
4226
- input: ("image" | "text")[];
4227
- cost: {
4228
- input: number;
4229
- output: number;
4230
- cacheRead: number;
4231
- cacheWrite: number;
4232
- };
4233
- contextWindow: number;
4234
- maxTokens: number;
4235
- };
4236
- readonly "claude-opus-4-6-thinking": {
4237
- id: string;
4238
- name: string;
4239
- api: "google-gemini-cli";
4240
- provider: string;
4241
- baseUrl: string;
4242
- reasoning: true;
4243
- input: ("image" | "text")[];
4244
- cost: {
4245
- input: number;
4246
- output: number;
4247
- cacheRead: number;
4248
- cacheWrite: number;
4249
- };
4250
- contextWindow: number;
4251
- maxTokens: number;
4252
- };
4253
- readonly "claude-sonnet-4-5": {
4254
- id: string;
4255
- name: string;
4256
- api: "google-gemini-cli";
4257
- provider: string;
4258
- baseUrl: string;
4259
- reasoning: false;
4260
- input: ("image" | "text")[];
4261
- cost: {
4262
- input: number;
4263
- output: number;
4264
- cacheRead: number;
4265
- cacheWrite: number;
4266
- };
4267
- contextWindow: number;
4268
- maxTokens: number;
4269
- };
4270
- readonly "claude-sonnet-4-5-thinking": {
4271
- id: string;
4272
- name: string;
4273
- api: "google-gemini-cli";
4274
- provider: string;
4275
- baseUrl: string;
4276
- reasoning: true;
4277
- input: ("image" | "text")[];
4278
- cost: {
4279
- input: number;
4280
- output: number;
4281
- cacheRead: number;
4282
- cacheWrite: number;
4283
- };
4284
- contextWindow: number;
4285
- maxTokens: number;
4286
- };
4287
- readonly "claude-sonnet-4-6": {
4288
- id: string;
4289
- name: string;
4290
- api: "google-gemini-cli";
4291
- provider: string;
4292
- baseUrl: string;
4293
- reasoning: true;
4294
- input: ("image" | "text")[];
4295
- cost: {
4296
- input: number;
4297
- output: number;
4298
- cacheRead: number;
4299
- cacheWrite: number;
4300
- };
4301
- contextWindow: number;
4302
- maxTokens: number;
4303
- };
4304
- readonly "gemini-3-flash": {
4305
- id: string;
4306
- name: string;
4307
- api: "google-gemini-cli";
4308
- provider: string;
4309
- baseUrl: string;
4310
- reasoning: true;
4311
- input: ("image" | "text")[];
4312
- cost: {
4313
- input: number;
4314
- output: number;
4315
- cacheRead: number;
4316
- cacheWrite: number;
4317
- };
4318
- contextWindow: number;
4319
- maxTokens: number;
4320
- };
4321
- readonly "gemini-3.1-pro-high": {
4322
- id: string;
4323
- name: string;
4324
- api: "google-gemini-cli";
4325
- provider: string;
4326
- baseUrl: string;
4327
- reasoning: true;
4328
- input: ("image" | "text")[];
4329
- cost: {
4330
- input: number;
4331
- output: number;
4332
- cacheRead: number;
4333
- cacheWrite: number;
4334
- };
4335
- contextWindow: number;
4336
- maxTokens: number;
4337
- };
4338
- readonly "gemini-3.1-pro-low": {
4339
- id: string;
4340
- name: string;
4341
- api: "google-gemini-cli";
4342
- provider: string;
4343
- baseUrl: string;
4344
- reasoning: true;
4345
- input: ("image" | "text")[];
4346
- cost: {
4347
- input: number;
4348
- output: number;
4349
- cacheRead: number;
4350
- cacheWrite: number;
4351
- };
4352
- contextWindow: number;
4353
- maxTokens: number;
4354
- };
4355
- readonly "gpt-oss-120b-medium": {
4356
- id: string;
4357
- name: string;
4358
- api: "google-gemini-cli";
4359
- provider: string;
4360
- baseUrl: string;
4361
- reasoning: false;
4362
- input: "text"[];
4363
- cost: {
4364
- input: number;
4365
- output: number;
4366
- cacheRead: number;
4367
- cacheWrite: number;
4368
- };
4369
- contextWindow: number;
4370
- maxTokens: number;
4371
- };
4372
- };
4373
- readonly "google-gemini-cli": {
4374
- readonly "gemini-2.0-flash": {
4375
- id: string;
4376
- name: string;
4377
- api: "google-gemini-cli";
4378
- provider: string;
4379
- baseUrl: string;
4380
- reasoning: false;
4381
- input: ("image" | "text")[];
4382
- cost: {
4383
- input: number;
4384
- output: number;
4385
- cacheRead: number;
4386
- cacheWrite: number;
4387
- };
4388
- contextWindow: number;
4389
- maxTokens: number;
4390
- };
4391
- readonly "gemini-2.5-flash": {
4392
- id: string;
4393
- name: string;
4394
- api: "google-gemini-cli";
4395
- provider: string;
4396
- baseUrl: string;
4397
- reasoning: true;
4398
- input: ("image" | "text")[];
4399
- cost: {
4400
- input: number;
4401
- output: number;
4402
- cacheRead: number;
4403
- cacheWrite: number;
4404
- };
4405
- contextWindow: number;
4406
- maxTokens: number;
4407
- };
4408
- readonly "gemini-2.5-pro": {
4409
- id: string;
4410
- name: string;
4411
- api: "google-gemini-cli";
4412
- provider: string;
4413
- baseUrl: string;
4414
- reasoning: true;
4415
- input: ("image" | "text")[];
4416
- cost: {
4417
- input: number;
4418
- output: number;
4419
- cacheRead: number;
4420
- cacheWrite: number;
4421
- };
4422
- contextWindow: number;
4423
- maxTokens: number;
4424
- };
4425
- readonly "gemini-3-flash-preview": {
4426
- id: string;
4427
- name: string;
4428
- api: "google-gemini-cli";
4429
- provider: string;
4430
- baseUrl: string;
4431
- reasoning: true;
4432
- input: ("image" | "text")[];
4433
- cost: {
4434
- input: number;
4435
- output: number;
4436
- cacheRead: number;
4437
- cacheWrite: number;
4438
- };
4439
- contextWindow: number;
4440
- maxTokens: number;
4441
- };
4442
- readonly "gemini-3-pro-preview": {
4443
- id: string;
4444
- name: string;
4445
- api: "google-gemini-cli";
4446
- provider: string;
4447
- baseUrl: string;
4448
- reasoning: true;
4977
+ reasoning: false;
4449
4978
  input: ("image" | "text")[];
4450
4979
  cost: {
4451
4980
  input: number;
@@ -4456,10 +4985,10 @@ export declare const MODELS: {
4456
4985
  contextWindow: number;
4457
4986
  maxTokens: number;
4458
4987
  };
4459
- readonly "gemini-3.1-flash-lite-preview": {
4988
+ readonly "gemma-4-26b-a4b-it": {
4460
4989
  id: string;
4461
4990
  name: string;
4462
- api: "google-gemini-cli";
4991
+ api: "google-generative-ai";
4463
4992
  provider: string;
4464
4993
  baseUrl: string;
4465
4994
  reasoning: true;
@@ -4473,10 +5002,10 @@ export declare const MODELS: {
4473
5002
  contextWindow: number;
4474
5003
  maxTokens: number;
4475
5004
  };
4476
- readonly "gemini-3.1-pro-preview": {
5005
+ readonly "gemma-4-31b-it": {
4477
5006
  id: string;
4478
5007
  name: string;
4479
- api: "google-gemini-cli";
5008
+ api: "google-generative-ai";
4480
5009
  provider: string;
4481
5010
  baseUrl: string;
4482
5011
  reasoning: true;
@@ -5475,7 +6004,257 @@ export declare const MODELS: {
5475
6004
  "User-Agent": string;
5476
6005
  };
5477
6006
  reasoning: true;
5478
- input: ("image" | "text")[];
6007
+ input: ("image" | "text")[];
6008
+ cost: {
6009
+ input: number;
6010
+ output: number;
6011
+ cacheRead: number;
6012
+ cacheWrite: number;
6013
+ };
6014
+ contextWindow: number;
6015
+ maxTokens: number;
6016
+ };
6017
+ readonly "kimi-for-coding": {
6018
+ id: string;
6019
+ name: string;
6020
+ api: "anthropic-messages";
6021
+ provider: string;
6022
+ baseUrl: string;
6023
+ headers: {
6024
+ "User-Agent": string;
6025
+ };
6026
+ reasoning: true;
6027
+ input: ("image" | "text")[];
6028
+ cost: {
6029
+ input: number;
6030
+ output: number;
6031
+ cacheRead: number;
6032
+ cacheWrite: number;
6033
+ };
6034
+ contextWindow: number;
6035
+ maxTokens: number;
6036
+ };
6037
+ readonly "kimi-k2-thinking": {
6038
+ id: string;
6039
+ name: string;
6040
+ api: "anthropic-messages";
6041
+ provider: string;
6042
+ baseUrl: string;
6043
+ headers: {
6044
+ "User-Agent": string;
6045
+ };
6046
+ reasoning: true;
6047
+ input: "text"[];
6048
+ cost: {
6049
+ input: number;
6050
+ output: number;
6051
+ cacheRead: number;
6052
+ cacheWrite: number;
6053
+ };
6054
+ contextWindow: number;
6055
+ maxTokens: number;
6056
+ };
6057
+ };
6058
+ readonly minimax: {
6059
+ readonly "MiniMax-M2.7": {
6060
+ id: string;
6061
+ name: string;
6062
+ api: "anthropic-messages";
6063
+ provider: string;
6064
+ baseUrl: string;
6065
+ reasoning: true;
6066
+ input: "text"[];
6067
+ cost: {
6068
+ input: number;
6069
+ output: number;
6070
+ cacheRead: number;
6071
+ cacheWrite: number;
6072
+ };
6073
+ contextWindow: number;
6074
+ maxTokens: number;
6075
+ };
6076
+ readonly "MiniMax-M2.7-highspeed": {
6077
+ id: string;
6078
+ name: string;
6079
+ api: "anthropic-messages";
6080
+ provider: string;
6081
+ baseUrl: string;
6082
+ reasoning: true;
6083
+ input: "text"[];
6084
+ cost: {
6085
+ input: number;
6086
+ output: number;
6087
+ cacheRead: number;
6088
+ cacheWrite: number;
6089
+ };
6090
+ contextWindow: number;
6091
+ maxTokens: number;
6092
+ };
6093
+ };
6094
+ readonly "minimax-cn": {
6095
+ readonly "MiniMax-M2.7": {
6096
+ id: string;
6097
+ name: string;
6098
+ api: "anthropic-messages";
6099
+ provider: string;
6100
+ baseUrl: string;
6101
+ reasoning: true;
6102
+ input: "text"[];
6103
+ cost: {
6104
+ input: number;
6105
+ output: number;
6106
+ cacheRead: number;
6107
+ cacheWrite: number;
6108
+ };
6109
+ contextWindow: number;
6110
+ maxTokens: number;
6111
+ };
6112
+ readonly "MiniMax-M2.7-highspeed": {
6113
+ id: string;
6114
+ name: string;
6115
+ api: "anthropic-messages";
6116
+ provider: string;
6117
+ baseUrl: string;
6118
+ reasoning: true;
6119
+ input: "text"[];
6120
+ cost: {
6121
+ input: number;
6122
+ output: number;
6123
+ cacheRead: number;
6124
+ cacheWrite: number;
6125
+ };
6126
+ contextWindow: number;
6127
+ maxTokens: number;
6128
+ };
6129
+ };
6130
+ readonly mistral: {
6131
+ readonly "codestral-latest": {
6132
+ id: string;
6133
+ name: string;
6134
+ api: "mistral-conversations";
6135
+ provider: string;
6136
+ baseUrl: string;
6137
+ reasoning: false;
6138
+ input: "text"[];
6139
+ cost: {
6140
+ input: number;
6141
+ output: number;
6142
+ cacheRead: number;
6143
+ cacheWrite: number;
6144
+ };
6145
+ contextWindow: number;
6146
+ maxTokens: number;
6147
+ };
6148
+ readonly "devstral-2512": {
6149
+ id: string;
6150
+ name: string;
6151
+ api: "mistral-conversations";
6152
+ provider: string;
6153
+ baseUrl: string;
6154
+ reasoning: false;
6155
+ input: "text"[];
6156
+ cost: {
6157
+ input: number;
6158
+ output: number;
6159
+ cacheRead: number;
6160
+ cacheWrite: number;
6161
+ };
6162
+ contextWindow: number;
6163
+ maxTokens: number;
6164
+ };
6165
+ readonly "devstral-medium-2507": {
6166
+ id: string;
6167
+ name: string;
6168
+ api: "mistral-conversations";
6169
+ provider: string;
6170
+ baseUrl: string;
6171
+ reasoning: false;
6172
+ input: "text"[];
6173
+ cost: {
6174
+ input: number;
6175
+ output: number;
6176
+ cacheRead: number;
6177
+ cacheWrite: number;
6178
+ };
6179
+ contextWindow: number;
6180
+ maxTokens: number;
6181
+ };
6182
+ readonly "devstral-medium-latest": {
6183
+ id: string;
6184
+ name: string;
6185
+ api: "mistral-conversations";
6186
+ provider: string;
6187
+ baseUrl: string;
6188
+ reasoning: false;
6189
+ input: "text"[];
6190
+ cost: {
6191
+ input: number;
6192
+ output: number;
6193
+ cacheRead: number;
6194
+ cacheWrite: number;
6195
+ };
6196
+ contextWindow: number;
6197
+ maxTokens: number;
6198
+ };
6199
+ readonly "devstral-small-2505": {
6200
+ id: string;
6201
+ name: string;
6202
+ api: "mistral-conversations";
6203
+ provider: string;
6204
+ baseUrl: string;
6205
+ reasoning: false;
6206
+ input: "text"[];
6207
+ cost: {
6208
+ input: number;
6209
+ output: number;
6210
+ cacheRead: number;
6211
+ cacheWrite: number;
6212
+ };
6213
+ contextWindow: number;
6214
+ maxTokens: number;
6215
+ };
6216
+ readonly "devstral-small-2507": {
6217
+ id: string;
6218
+ name: string;
6219
+ api: "mistral-conversations";
6220
+ provider: string;
6221
+ baseUrl: string;
6222
+ reasoning: false;
6223
+ input: "text"[];
6224
+ cost: {
6225
+ input: number;
6226
+ output: number;
6227
+ cacheRead: number;
6228
+ cacheWrite: number;
6229
+ };
6230
+ contextWindow: number;
6231
+ maxTokens: number;
6232
+ };
6233
+ readonly "labs-devstral-small-2512": {
6234
+ id: string;
6235
+ name: string;
6236
+ api: "mistral-conversations";
6237
+ provider: string;
6238
+ baseUrl: string;
6239
+ reasoning: false;
6240
+ input: ("image" | "text")[];
6241
+ cost: {
6242
+ input: number;
6243
+ output: number;
6244
+ cacheRead: number;
6245
+ cacheWrite: number;
6246
+ };
6247
+ contextWindow: number;
6248
+ maxTokens: number;
6249
+ };
6250
+ readonly "magistral-medium-latest": {
6251
+ id: string;
6252
+ name: string;
6253
+ api: "mistral-conversations";
6254
+ provider: string;
6255
+ baseUrl: string;
6256
+ reasoning: true;
6257
+ input: "text"[];
5479
6258
  cost: {
5480
6259
  input: number;
5481
6260
  output: number;
@@ -5485,17 +6264,14 @@ export declare const MODELS: {
5485
6264
  contextWindow: number;
5486
6265
  maxTokens: number;
5487
6266
  };
5488
- readonly "kimi-for-coding": {
6267
+ readonly "magistral-small": {
5489
6268
  id: string;
5490
6269
  name: string;
5491
- api: "anthropic-messages";
6270
+ api: "mistral-conversations";
5492
6271
  provider: string;
5493
6272
  baseUrl: string;
5494
- headers: {
5495
- "User-Agent": string;
5496
- };
5497
6273
  reasoning: true;
5498
- input: ("image" | "text")[];
6274
+ input: "text"[];
5499
6275
  cost: {
5500
6276
  input: number;
5501
6277
  output: number;
@@ -5505,16 +6281,13 @@ export declare const MODELS: {
5505
6281
  contextWindow: number;
5506
6282
  maxTokens: number;
5507
6283
  };
5508
- readonly "kimi-k2-thinking": {
6284
+ readonly "ministral-3b-latest": {
5509
6285
  id: string;
5510
6286
  name: string;
5511
- api: "anthropic-messages";
6287
+ api: "mistral-conversations";
5512
6288
  provider: string;
5513
6289
  baseUrl: string;
5514
- headers: {
5515
- "User-Agent": string;
5516
- };
5517
- reasoning: true;
6290
+ reasoning: false;
5518
6291
  input: "text"[];
5519
6292
  cost: {
5520
6293
  input: number;
@@ -5525,15 +6298,13 @@ export declare const MODELS: {
5525
6298
  contextWindow: number;
5526
6299
  maxTokens: number;
5527
6300
  };
5528
- };
5529
- readonly minimax: {
5530
- readonly "MiniMax-M2.7": {
6301
+ readonly "ministral-8b-latest": {
5531
6302
  id: string;
5532
6303
  name: string;
5533
- api: "anthropic-messages";
6304
+ api: "mistral-conversations";
5534
6305
  provider: string;
5535
6306
  baseUrl: string;
5536
- reasoning: true;
6307
+ reasoning: false;
5537
6308
  input: "text"[];
5538
6309
  cost: {
5539
6310
  input: number;
@@ -5544,13 +6315,13 @@ export declare const MODELS: {
5544
6315
  contextWindow: number;
5545
6316
  maxTokens: number;
5546
6317
  };
5547
- readonly "MiniMax-M2.7-highspeed": {
6318
+ readonly "mistral-large-2411": {
5548
6319
  id: string;
5549
6320
  name: string;
5550
- api: "anthropic-messages";
6321
+ api: "mistral-conversations";
5551
6322
  provider: string;
5552
6323
  baseUrl: string;
5553
- reasoning: true;
6324
+ reasoning: false;
5554
6325
  input: "text"[];
5555
6326
  cost: {
5556
6327
  input: number;
@@ -5561,16 +6332,14 @@ export declare const MODELS: {
5561
6332
  contextWindow: number;
5562
6333
  maxTokens: number;
5563
6334
  };
5564
- };
5565
- readonly "minimax-cn": {
5566
- readonly "MiniMax-M2.7": {
6335
+ readonly "mistral-large-2512": {
5567
6336
  id: string;
5568
6337
  name: string;
5569
- api: "anthropic-messages";
6338
+ api: "mistral-conversations";
5570
6339
  provider: string;
5571
6340
  baseUrl: string;
5572
- reasoning: true;
5573
- input: "text"[];
6341
+ reasoning: false;
6342
+ input: ("image" | "text")[];
5574
6343
  cost: {
5575
6344
  input: number;
5576
6345
  output: number;
@@ -5580,14 +6349,14 @@ export declare const MODELS: {
5580
6349
  contextWindow: number;
5581
6350
  maxTokens: number;
5582
6351
  };
5583
- readonly "MiniMax-M2.7-highspeed": {
6352
+ readonly "mistral-large-latest": {
5584
6353
  id: string;
5585
6354
  name: string;
5586
- api: "anthropic-messages";
6355
+ api: "mistral-conversations";
5587
6356
  provider: string;
5588
6357
  baseUrl: string;
5589
- reasoning: true;
5590
- input: "text"[];
6358
+ reasoning: false;
6359
+ input: ("image" | "text")[];
5591
6360
  cost: {
5592
6361
  input: number;
5593
6362
  output: number;
@@ -5597,16 +6366,14 @@ export declare const MODELS: {
5597
6366
  contextWindow: number;
5598
6367
  maxTokens: number;
5599
6368
  };
5600
- };
5601
- readonly mistral: {
5602
- readonly "codestral-latest": {
6369
+ readonly "mistral-medium-2505": {
5603
6370
  id: string;
5604
6371
  name: string;
5605
6372
  api: "mistral-conversations";
5606
6373
  provider: string;
5607
6374
  baseUrl: string;
5608
6375
  reasoning: false;
5609
- input: "text"[];
6376
+ input: ("image" | "text")[];
5610
6377
  cost: {
5611
6378
  input: number;
5612
6379
  output: number;
@@ -5616,14 +6383,14 @@ export declare const MODELS: {
5616
6383
  contextWindow: number;
5617
6384
  maxTokens: number;
5618
6385
  };
5619
- readonly "devstral-2512": {
6386
+ readonly "mistral-medium-2508": {
5620
6387
  id: string;
5621
6388
  name: string;
5622
6389
  api: "mistral-conversations";
5623
6390
  provider: string;
5624
6391
  baseUrl: string;
5625
6392
  reasoning: false;
5626
- input: "text"[];
6393
+ input: ("image" | "text")[];
5627
6394
  cost: {
5628
6395
  input: number;
5629
6396
  output: number;
@@ -5633,14 +6400,14 @@ export declare const MODELS: {
5633
6400
  contextWindow: number;
5634
6401
  maxTokens: number;
5635
6402
  };
5636
- readonly "devstral-medium-2507": {
6403
+ readonly "mistral-medium-2604": {
5637
6404
  id: string;
5638
6405
  name: string;
5639
6406
  api: "mistral-conversations";
5640
6407
  provider: string;
5641
6408
  baseUrl: string;
5642
- reasoning: false;
5643
- input: "text"[];
6409
+ reasoning: true;
6410
+ input: ("image" | "text")[];
5644
6411
  cost: {
5645
6412
  input: number;
5646
6413
  output: number;
@@ -5650,14 +6417,14 @@ export declare const MODELS: {
5650
6417
  contextWindow: number;
5651
6418
  maxTokens: number;
5652
6419
  };
5653
- readonly "devstral-medium-latest": {
6420
+ readonly "mistral-medium-3.5": {
5654
6421
  id: string;
5655
6422
  name: string;
5656
6423
  api: "mistral-conversations";
5657
6424
  provider: string;
5658
6425
  baseUrl: string;
5659
- reasoning: false;
5660
- input: "text"[];
6426
+ reasoning: true;
6427
+ input: ("image" | "text")[];
5661
6428
  cost: {
5662
6429
  input: number;
5663
6430
  output: number;
@@ -5667,14 +6434,14 @@ export declare const MODELS: {
5667
6434
  contextWindow: number;
5668
6435
  maxTokens: number;
5669
6436
  };
5670
- readonly "devstral-small-2505": {
6437
+ readonly "mistral-medium-latest": {
5671
6438
  id: string;
5672
6439
  name: string;
5673
6440
  api: "mistral-conversations";
5674
6441
  provider: string;
5675
6442
  baseUrl: string;
5676
- reasoning: false;
5677
- input: "text"[];
6443
+ reasoning: true;
6444
+ input: ("image" | "text")[];
5678
6445
  cost: {
5679
6446
  input: number;
5680
6447
  output: number;
@@ -5684,7 +6451,7 @@ export declare const MODELS: {
5684
6451
  contextWindow: number;
5685
6452
  maxTokens: number;
5686
6453
  };
5687
- readonly "devstral-small-2507": {
6454
+ readonly "mistral-nemo": {
5688
6455
  id: string;
5689
6456
  name: string;
5690
6457
  api: "mistral-conversations";
@@ -5701,7 +6468,7 @@ export declare const MODELS: {
5701
6468
  contextWindow: number;
5702
6469
  maxTokens: number;
5703
6470
  };
5704
- readonly "labs-devstral-small-2512": {
6471
+ readonly "mistral-small-2506": {
5705
6472
  id: string;
5706
6473
  name: string;
5707
6474
  api: "mistral-conversations";
@@ -5718,14 +6485,14 @@ export declare const MODELS: {
5718
6485
  contextWindow: number;
5719
6486
  maxTokens: number;
5720
6487
  };
5721
- readonly "magistral-medium-latest": {
6488
+ readonly "mistral-small-2603": {
5722
6489
  id: string;
5723
6490
  name: string;
5724
6491
  api: "mistral-conversations";
5725
6492
  provider: string;
5726
6493
  baseUrl: string;
5727
6494
  reasoning: true;
5728
- input: "text"[];
6495
+ input: ("image" | "text")[];
5729
6496
  cost: {
5730
6497
  input: number;
5731
6498
  output: number;
@@ -5735,14 +6502,14 @@ export declare const MODELS: {
5735
6502
  contextWindow: number;
5736
6503
  maxTokens: number;
5737
6504
  };
5738
- readonly "magistral-small": {
6505
+ readonly "mistral-small-latest": {
5739
6506
  id: string;
5740
6507
  name: string;
5741
6508
  api: "mistral-conversations";
5742
6509
  provider: string;
5743
6510
  baseUrl: string;
5744
6511
  reasoning: true;
5745
- input: "text"[];
6512
+ input: ("image" | "text")[];
5746
6513
  cost: {
5747
6514
  input: number;
5748
6515
  output: number;
@@ -5752,7 +6519,7 @@ export declare const MODELS: {
5752
6519
  contextWindow: number;
5753
6520
  maxTokens: number;
5754
6521
  };
5755
- readonly "ministral-3b-latest": {
6522
+ readonly "open-mistral-7b": {
5756
6523
  id: string;
5757
6524
  name: string;
5758
6525
  api: "mistral-conversations";
@@ -5769,7 +6536,7 @@ export declare const MODELS: {
5769
6536
  contextWindow: number;
5770
6537
  maxTokens: number;
5771
6538
  };
5772
- readonly "ministral-8b-latest": {
6539
+ readonly "open-mixtral-8x22b": {
5773
6540
  id: string;
5774
6541
  name: string;
5775
6542
  api: "mistral-conversations";
@@ -5786,7 +6553,7 @@ export declare const MODELS: {
5786
6553
  contextWindow: number;
5787
6554
  maxTokens: number;
5788
6555
  };
5789
- readonly "mistral-large-2411": {
6556
+ readonly "open-mixtral-8x7b": {
5790
6557
  id: string;
5791
6558
  name: string;
5792
6559
  api: "mistral-conversations";
@@ -5803,7 +6570,7 @@ export declare const MODELS: {
5803
6570
  contextWindow: number;
5804
6571
  maxTokens: number;
5805
6572
  };
5806
- readonly "mistral-large-2512": {
6573
+ readonly "pixtral-12b": {
5807
6574
  id: string;
5808
6575
  name: string;
5809
6576
  api: "mistral-conversations";
@@ -5820,7 +6587,7 @@ export declare const MODELS: {
5820
6587
  contextWindow: number;
5821
6588
  maxTokens: number;
5822
6589
  };
5823
- readonly "mistral-large-latest": {
6590
+ readonly "pixtral-large-latest": {
5824
6591
  id: string;
5825
6592
  name: string;
5826
6593
  api: "mistral-conversations";
@@ -5837,14 +6604,23 @@ export declare const MODELS: {
5837
6604
  contextWindow: number;
5838
6605
  maxTokens: number;
5839
6606
  };
5840
- readonly "mistral-medium-2505": {
6607
+ };
6608
+ readonly moonshotai: {
6609
+ readonly "kimi-k2-0711-preview": {
5841
6610
  id: string;
5842
6611
  name: string;
5843
- api: "mistral-conversations";
6612
+ api: "openai-completions";
5844
6613
  provider: string;
5845
6614
  baseUrl: string;
6615
+ compat: {
6616
+ supportsStore: false;
6617
+ supportsDeveloperRole: false;
6618
+ supportsReasoningEffort: false;
6619
+ maxTokensField: "max_tokens";
6620
+ supportsStrictMode: false;
6621
+ };
5846
6622
  reasoning: false;
5847
- input: ("image" | "text")[];
6623
+ input: "text"[];
5848
6624
  cost: {
5849
6625
  input: number;
5850
6626
  output: number;
@@ -5854,14 +6630,21 @@ export declare const MODELS: {
5854
6630
  contextWindow: number;
5855
6631
  maxTokens: number;
5856
6632
  };
5857
- readonly "mistral-medium-2508": {
6633
+ readonly "kimi-k2-0905-preview": {
5858
6634
  id: string;
5859
6635
  name: string;
5860
- api: "mistral-conversations";
6636
+ api: "openai-completions";
5861
6637
  provider: string;
5862
6638
  baseUrl: string;
6639
+ compat: {
6640
+ supportsStore: false;
6641
+ supportsDeveloperRole: false;
6642
+ supportsReasoningEffort: false;
6643
+ maxTokensField: "max_tokens";
6644
+ supportsStrictMode: false;
6645
+ };
5863
6646
  reasoning: false;
5864
- input: ("image" | "text")[];
6647
+ input: "text"[];
5865
6648
  cost: {
5866
6649
  input: number;
5867
6650
  output: number;
@@ -5871,14 +6654,21 @@ export declare const MODELS: {
5871
6654
  contextWindow: number;
5872
6655
  maxTokens: number;
5873
6656
  };
5874
- readonly "mistral-medium-latest": {
6657
+ readonly "kimi-k2-thinking": {
5875
6658
  id: string;
5876
6659
  name: string;
5877
- api: "mistral-conversations";
6660
+ api: "openai-completions";
5878
6661
  provider: string;
5879
6662
  baseUrl: string;
5880
- reasoning: false;
5881
- input: ("image" | "text")[];
6663
+ compat: {
6664
+ supportsStore: false;
6665
+ supportsDeveloperRole: false;
6666
+ supportsReasoningEffort: false;
6667
+ maxTokensField: "max_tokens";
6668
+ supportsStrictMode: false;
6669
+ };
6670
+ reasoning: true;
6671
+ input: "text"[];
5882
6672
  cost: {
5883
6673
  input: number;
5884
6674
  output: number;
@@ -5888,13 +6678,20 @@ export declare const MODELS: {
5888
6678
  contextWindow: number;
5889
6679
  maxTokens: number;
5890
6680
  };
5891
- readonly "mistral-nemo": {
6681
+ readonly "kimi-k2-thinking-turbo": {
5892
6682
  id: string;
5893
6683
  name: string;
5894
- api: "mistral-conversations";
6684
+ api: "openai-completions";
5895
6685
  provider: string;
5896
6686
  baseUrl: string;
5897
- reasoning: false;
6687
+ compat: {
6688
+ supportsStore: false;
6689
+ supportsDeveloperRole: false;
6690
+ supportsReasoningEffort: false;
6691
+ maxTokensField: "max_tokens";
6692
+ supportsStrictMode: false;
6693
+ };
6694
+ reasoning: true;
5898
6695
  input: "text"[];
5899
6696
  cost: {
5900
6697
  input: number;
@@ -5905,14 +6702,21 @@ export declare const MODELS: {
5905
6702
  contextWindow: number;
5906
6703
  maxTokens: number;
5907
6704
  };
5908
- readonly "mistral-small-2506": {
6705
+ readonly "kimi-k2-turbo-preview": {
5909
6706
  id: string;
5910
6707
  name: string;
5911
- api: "mistral-conversations";
6708
+ api: "openai-completions";
5912
6709
  provider: string;
5913
6710
  baseUrl: string;
6711
+ compat: {
6712
+ supportsStore: false;
6713
+ supportsDeveloperRole: false;
6714
+ supportsReasoningEffort: false;
6715
+ maxTokensField: "max_tokens";
6716
+ supportsStrictMode: false;
6717
+ };
5914
6718
  reasoning: false;
5915
- input: ("image" | "text")[];
6719
+ input: "text"[];
5916
6720
  cost: {
5917
6721
  input: number;
5918
6722
  output: number;
@@ -5922,12 +6726,19 @@ export declare const MODELS: {
5922
6726
  contextWindow: number;
5923
6727
  maxTokens: number;
5924
6728
  };
5925
- readonly "mistral-small-2603": {
6729
+ readonly "kimi-k2.5": {
5926
6730
  id: string;
5927
6731
  name: string;
5928
- api: "mistral-conversations";
6732
+ api: "openai-completions";
5929
6733
  provider: string;
5930
6734
  baseUrl: string;
6735
+ compat: {
6736
+ supportsStore: false;
6737
+ supportsDeveloperRole: false;
6738
+ supportsReasoningEffort: false;
6739
+ maxTokensField: "max_tokens";
6740
+ supportsStrictMode: false;
6741
+ };
5931
6742
  reasoning: true;
5932
6743
  input: ("image" | "text")[];
5933
6744
  cost: {
@@ -5939,12 +6750,19 @@ export declare const MODELS: {
5939
6750
  contextWindow: number;
5940
6751
  maxTokens: number;
5941
6752
  };
5942
- readonly "mistral-small-latest": {
6753
+ readonly "kimi-k2.6": {
5943
6754
  id: string;
5944
6755
  name: string;
5945
- api: "mistral-conversations";
6756
+ api: "openai-completions";
5946
6757
  provider: string;
5947
6758
  baseUrl: string;
6759
+ compat: {
6760
+ supportsStore: false;
6761
+ supportsDeveloperRole: false;
6762
+ supportsReasoningEffort: false;
6763
+ maxTokensField: "max_tokens";
6764
+ supportsStrictMode: false;
6765
+ };
5948
6766
  reasoning: true;
5949
6767
  input: ("image" | "text")[];
5950
6768
  cost: {
@@ -5956,12 +6774,21 @@ export declare const MODELS: {
5956
6774
  contextWindow: number;
5957
6775
  maxTokens: number;
5958
6776
  };
5959
- readonly "open-mistral-7b": {
6777
+ };
6778
+ readonly "moonshotai-cn": {
6779
+ readonly "kimi-k2-0711-preview": {
5960
6780
  id: string;
5961
6781
  name: string;
5962
- api: "mistral-conversations";
6782
+ api: "openai-completions";
5963
6783
  provider: string;
5964
6784
  baseUrl: string;
6785
+ compat: {
6786
+ supportsStore: false;
6787
+ supportsDeveloperRole: false;
6788
+ supportsReasoningEffort: false;
6789
+ maxTokensField: "max_tokens";
6790
+ supportsStrictMode: false;
6791
+ };
5965
6792
  reasoning: false;
5966
6793
  input: "text"[];
5967
6794
  cost: {
@@ -5973,12 +6800,19 @@ export declare const MODELS: {
5973
6800
  contextWindow: number;
5974
6801
  maxTokens: number;
5975
6802
  };
5976
- readonly "open-mixtral-8x22b": {
6803
+ readonly "kimi-k2-0905-preview": {
5977
6804
  id: string;
5978
6805
  name: string;
5979
- api: "mistral-conversations";
6806
+ api: "openai-completions";
5980
6807
  provider: string;
5981
6808
  baseUrl: string;
6809
+ compat: {
6810
+ supportsStore: false;
6811
+ supportsDeveloperRole: false;
6812
+ supportsReasoningEffort: false;
6813
+ maxTokensField: "max_tokens";
6814
+ supportsStrictMode: false;
6815
+ };
5982
6816
  reasoning: false;
5983
6817
  input: "text"[];
5984
6818
  cost: {
@@ -5990,13 +6824,20 @@ export declare const MODELS: {
5990
6824
  contextWindow: number;
5991
6825
  maxTokens: number;
5992
6826
  };
5993
- readonly "open-mixtral-8x7b": {
6827
+ readonly "kimi-k2-thinking": {
5994
6828
  id: string;
5995
6829
  name: string;
5996
- api: "mistral-conversations";
6830
+ api: "openai-completions";
5997
6831
  provider: string;
5998
6832
  baseUrl: string;
5999
- reasoning: false;
6833
+ compat: {
6834
+ supportsStore: false;
6835
+ supportsDeveloperRole: false;
6836
+ supportsReasoningEffort: false;
6837
+ maxTokensField: "max_tokens";
6838
+ supportsStrictMode: false;
6839
+ };
6840
+ reasoning: true;
6000
6841
  input: "text"[];
6001
6842
  cost: {
6002
6843
  input: number;
@@ -6007,13 +6848,68 @@ export declare const MODELS: {
6007
6848
  contextWindow: number;
6008
6849
  maxTokens: number;
6009
6850
  };
6010
- readonly "pixtral-12b": {
6851
+ readonly "kimi-k2-thinking-turbo": {
6011
6852
  id: string;
6012
6853
  name: string;
6013
- api: "mistral-conversations";
6854
+ api: "openai-completions";
6855
+ provider: string;
6856
+ baseUrl: string;
6857
+ compat: {
6858
+ supportsStore: false;
6859
+ supportsDeveloperRole: false;
6860
+ supportsReasoningEffort: false;
6861
+ maxTokensField: "max_tokens";
6862
+ supportsStrictMode: false;
6863
+ };
6864
+ reasoning: true;
6865
+ input: "text"[];
6866
+ cost: {
6867
+ input: number;
6868
+ output: number;
6869
+ cacheRead: number;
6870
+ cacheWrite: number;
6871
+ };
6872
+ contextWindow: number;
6873
+ maxTokens: number;
6874
+ };
6875
+ readonly "kimi-k2-turbo-preview": {
6876
+ id: string;
6877
+ name: string;
6878
+ api: "openai-completions";
6014
6879
  provider: string;
6015
6880
  baseUrl: string;
6881
+ compat: {
6882
+ supportsStore: false;
6883
+ supportsDeveloperRole: false;
6884
+ supportsReasoningEffort: false;
6885
+ maxTokensField: "max_tokens";
6886
+ supportsStrictMode: false;
6887
+ };
6016
6888
  reasoning: false;
6889
+ input: "text"[];
6890
+ cost: {
6891
+ input: number;
6892
+ output: number;
6893
+ cacheRead: number;
6894
+ cacheWrite: number;
6895
+ };
6896
+ contextWindow: number;
6897
+ maxTokens: number;
6898
+ };
6899
+ readonly "kimi-k2.5": {
6900
+ id: string;
6901
+ name: string;
6902
+ api: "openai-completions";
6903
+ provider: string;
6904
+ baseUrl: string;
6905
+ compat: {
6906
+ supportsStore: false;
6907
+ supportsDeveloperRole: false;
6908
+ supportsReasoningEffort: false;
6909
+ maxTokensField: "max_tokens";
6910
+ supportsStrictMode: false;
6911
+ };
6912
+ reasoning: true;
6017
6913
  input: ("image" | "text")[];
6018
6914
  cost: {
6019
6915
  input: number;
@@ -6024,13 +6920,20 @@ export declare const MODELS: {
6024
6920
  contextWindow: number;
6025
6921
  maxTokens: number;
6026
6922
  };
6027
- readonly "pixtral-large-latest": {
6923
+ readonly "kimi-k2.6": {
6028
6924
  id: string;
6029
6925
  name: string;
6030
- api: "mistral-conversations";
6926
+ api: "openai-completions";
6031
6927
  provider: string;
6032
6928
  baseUrl: string;
6033
- reasoning: false;
6929
+ compat: {
6930
+ supportsStore: false;
6931
+ supportsDeveloperRole: false;
6932
+ supportsReasoningEffort: false;
6933
+ maxTokensField: "max_tokens";
6934
+ supportsStrictMode: false;
6935
+ };
6936
+ reasoning: true;
6034
6937
  input: ("image" | "text")[];
6035
6938
  cost: {
6036
6939
  input: number;
@@ -6604,6 +7507,23 @@ export declare const MODELS: {
6604
7507
  contextWindow: number;
6605
7508
  maxTokens: number;
6606
7509
  };
7510
+ readonly "gpt-5.5-pro": {
7511
+ id: string;
7512
+ name: string;
7513
+ api: "openai-responses";
7514
+ provider: string;
7515
+ baseUrl: string;
7516
+ reasoning: true;
7517
+ input: ("image" | "text")[];
7518
+ cost: {
7519
+ input: number;
7520
+ output: number;
7521
+ cacheRead: number;
7522
+ cacheWrite: number;
7523
+ };
7524
+ contextWindow: number;
7525
+ maxTokens: number;
7526
+ };
6607
7527
  readonly o1: {
6608
7528
  id: string;
6609
7529
  name: string;
@@ -7458,23 +8378,6 @@ export declare const MODELS: {
7458
8378
  contextWindow: number;
7459
8379
  maxTokens: number;
7460
8380
  };
7461
- readonly "ling-2.6-flash-free": {
7462
- id: string;
7463
- name: string;
7464
- api: "openai-completions";
7465
- provider: string;
7466
- baseUrl: string;
7467
- reasoning: false;
7468
- input: "text"[];
7469
- cost: {
7470
- input: number;
7471
- output: number;
7472
- cacheRead: number;
7473
- cacheWrite: number;
7474
- };
7475
- contextWindow: number;
7476
- maxTokens: number;
7477
- };
7478
8381
  readonly "minimax-m2.5": {
7479
8382
  id: string;
7480
8383
  name: string;
@@ -7587,6 +8490,14 @@ export declare const MODELS: {
7587
8490
  baseUrl: string;
7588
8491
  compat: {
7589
8492
  requiresReasoningContentOnAssistantMessages: true;
8493
+ thinkingFormat: "deepseek";
8494
+ reasoningEffortMap: {
8495
+ minimal: string;
8496
+ low: string;
8497
+ medium: string;
8498
+ high: string;
8499
+ xhigh: string;
8500
+ };
7590
8501
  };
7591
8502
  reasoning: true;
7592
8503
  input: "text"[];
@@ -7607,6 +8518,14 @@ export declare const MODELS: {
7607
8518
  baseUrl: string;
7608
8519
  compat: {
7609
8520
  requiresReasoningContentOnAssistantMessages: true;
8521
+ thinkingFormat: "deepseek";
8522
+ reasoningEffortMap: {
8523
+ minimal: string;
8524
+ low: string;
8525
+ medium: string;
8526
+ high: string;
8527
+ xhigh: string;
8528
+ };
7610
8529
  };
7611
8530
  reasoning: true;
7612
8531
  input: "text"[];
@@ -8564,6 +9483,13 @@ export declare const MODELS: {
8564
9483
  baseUrl: string;
8565
9484
  compat: {
8566
9485
  requiresReasoningContentOnAssistantMessages: true;
9486
+ reasoningEffortMap: {
9487
+ minimal: string;
9488
+ low: string;
9489
+ medium: string;
9490
+ high: string;
9491
+ xhigh: string;
9492
+ };
8567
9493
  };
8568
9494
  reasoning: true;
8569
9495
  input: "text"[];
@@ -8584,6 +9510,13 @@ export declare const MODELS: {
8584
9510
  baseUrl: string;
8585
9511
  compat: {
8586
9512
  requiresReasoningContentOnAssistantMessages: true;
9513
+ reasoningEffortMap: {
9514
+ minimal: string;
9515
+ low: string;
9516
+ medium: string;
9517
+ high: string;
9518
+ xhigh: string;
9519
+ };
8587
9520
  };
8588
9521
  reasoning: true;
8589
9522
  input: "text"[];
@@ -8919,6 +9852,23 @@ export declare const MODELS: {
8919
9852
  contextWindow: number;
8920
9853
  maxTokens: number;
8921
9854
  };
9855
+ readonly "ibm-granite/granite-4.1-8b": {
9856
+ id: string;
9857
+ name: string;
9858
+ api: "openai-completions";
9859
+ provider: string;
9860
+ baseUrl: string;
9861
+ reasoning: false;
9862
+ input: "text"[];
9863
+ cost: {
9864
+ input: number;
9865
+ output: number;
9866
+ cacheRead: number;
9867
+ cacheWrite: number;
9868
+ };
9869
+ contextWindow: number;
9870
+ maxTokens: number;
9871
+ };
8922
9872
  readonly "inception/mercury-2": {
8923
9873
  id: string;
8924
9874
  name: string;
@@ -8953,7 +9903,7 @@ export declare const MODELS: {
8953
9903
  contextWindow: number;
8954
9904
  maxTokens: number;
8955
9905
  };
8956
- readonly "inclusionai/ling-2.6-flash:free": {
9906
+ readonly "inclusionai/ling-2.6-flash": {
8957
9907
  id: string;
8958
9908
  name: string;
8959
9909
  api: "openai-completions";
@@ -9480,23 +10430,6 @@ export declare const MODELS: {
9480
10430
  contextWindow: number;
9481
10431
  maxTokens: number;
9482
10432
  };
9483
- readonly "mistralai/mistral-small-creative": {
9484
- id: string;
9485
- name: string;
9486
- api: "openai-completions";
9487
- provider: string;
9488
- baseUrl: string;
9489
- reasoning: false;
9490
- input: "text"[];
9491
- cost: {
9492
- input: number;
9493
- output: number;
9494
- cacheRead: number;
9495
- cacheWrite: number;
9496
- };
9497
- contextWindow: number;
9498
- maxTokens: number;
9499
- };
9500
10433
  readonly "mistralai/mixtral-8x22b-instruct": {
9501
10434
  id: string;
9502
10435
  name: string;
@@ -9735,6 +10668,23 @@ export declare const MODELS: {
9735
10668
  contextWindow: number;
9736
10669
  maxTokens: number;
9737
10670
  };
10671
+ readonly "nvidia/nemotron-3-nano-omni-30b-a3b-reasoning:free": {
10672
+ id: string;
10673
+ name: string;
10674
+ api: "openai-completions";
10675
+ provider: string;
10676
+ baseUrl: string;
10677
+ reasoning: true;
10678
+ input: ("image" | "text")[];
10679
+ cost: {
10680
+ input: number;
10681
+ output: number;
10682
+ cacheRead: number;
10683
+ cacheWrite: number;
10684
+ };
10685
+ contextWindow: number;
10686
+ maxTokens: number;
10687
+ };
9738
10688
  readonly "nvidia/nemotron-3-super-120b-a12b": {
9739
10689
  id: string;
9740
10690
  name: string;
@@ -10806,6 +11756,57 @@ export declare const MODELS: {
10806
11756
  contextWindow: number;
10807
11757
  maxTokens: number;
10808
11758
  };
11759
+ readonly "openrouter/owl-alpha": {
11760
+ id: string;
11761
+ name: string;
11762
+ api: "openai-completions";
11763
+ provider: string;
11764
+ baseUrl: string;
11765
+ reasoning: false;
11766
+ input: "text"[];
11767
+ cost: {
11768
+ input: number;
11769
+ output: number;
11770
+ cacheRead: number;
11771
+ cacheWrite: number;
11772
+ };
11773
+ contextWindow: number;
11774
+ maxTokens: number;
11775
+ };
11776
+ readonly "poolside/laguna-m.1:free": {
11777
+ id: string;
11778
+ name: string;
11779
+ api: "openai-completions";
11780
+ provider: string;
11781
+ baseUrl: string;
11782
+ reasoning: true;
11783
+ input: "text"[];
11784
+ cost: {
11785
+ input: number;
11786
+ output: number;
11787
+ cacheRead: number;
11788
+ cacheWrite: number;
11789
+ };
11790
+ contextWindow: number;
11791
+ maxTokens: number;
11792
+ };
11793
+ readonly "poolside/laguna-xs.2:free": {
11794
+ id: string;
11795
+ name: string;
11796
+ api: "openai-completions";
11797
+ provider: string;
11798
+ baseUrl: string;
11799
+ reasoning: true;
11800
+ input: "text"[];
11801
+ cost: {
11802
+ input: number;
11803
+ output: number;
11804
+ cacheRead: number;
11805
+ cacheWrite: number;
11806
+ };
11807
+ contextWindow: number;
11808
+ maxTokens: number;
11809
+ };
10809
11810
  readonly "prime-intellect/intellect-3": {
10810
11811
  id: string;
10811
11812
  name: string;
@@ -11622,23 +12623,6 @@ export declare const MODELS: {
11622
12623
  contextWindow: number;
11623
12624
  maxTokens: number;
11624
12625
  };
11625
- readonly "qwen/qwq-32b": {
11626
- id: string;
11627
- name: string;
11628
- api: "openai-completions";
11629
- provider: string;
11630
- baseUrl: string;
11631
- reasoning: true;
11632
- input: "text"[];
11633
- cost: {
11634
- input: number;
11635
- output: number;
11636
- cacheRead: number;
11637
- cacheWrite: number;
11638
- };
11639
- contextWindow: number;
11640
- maxTokens: number;
11641
- };
11642
12626
  readonly "rekaai/reka-edge": {
11643
12627
  id: string;
11644
12628
  name: string;
@@ -12678,6 +13662,23 @@ export declare const MODELS: {
12678
13662
  contextWindow: number;
12679
13663
  maxTokens: number;
12680
13664
  };
13665
+ readonly "alibaba/qwen3.6-27b": {
13666
+ id: string;
13667
+ name: string;
13668
+ api: "anthropic-messages";
13669
+ provider: string;
13670
+ baseUrl: string;
13671
+ reasoning: true;
13672
+ input: ("image" | "text")[];
13673
+ cost: {
13674
+ input: number;
13675
+ output: number;
13676
+ cacheRead: number;
13677
+ cacheWrite: number;
13678
+ };
13679
+ contextWindow: number;
13680
+ maxTokens: number;
13681
+ };
12681
13682
  readonly "alibaba/qwen3.6-plus": {
12682
13683
  id: string;
12683
13684
  name: string;
@@ -13058,7 +14059,7 @@ export declare const MODELS: {
13058
14059
  api: "anthropic-messages";
13059
14060
  provider: string;
13060
14061
  baseUrl: string;
13061
- reasoning: true;
14062
+ reasoning: false;
13062
14063
  input: "text"[];
13063
14064
  cost: {
13064
14065
  input: number;
@@ -14905,6 +15906,40 @@ export declare const MODELS: {
14905
15906
  contextWindow: number;
14906
15907
  maxTokens: number;
14907
15908
  };
15909
+ readonly "xiaomi/mimo-v2.5": {
15910
+ id: string;
15911
+ name: string;
15912
+ api: "anthropic-messages";
15913
+ provider: string;
15914
+ baseUrl: string;
15915
+ reasoning: true;
15916
+ input: ("image" | "text")[];
15917
+ cost: {
15918
+ input: number;
15919
+ output: number;
15920
+ cacheRead: number;
15921
+ cacheWrite: number;
15922
+ };
15923
+ contextWindow: number;
15924
+ maxTokens: number;
15925
+ };
15926
+ readonly "xiaomi/mimo-v2.5-pro": {
15927
+ id: string;
15928
+ name: string;
15929
+ api: "anthropic-messages";
15930
+ provider: string;
15931
+ baseUrl: string;
15932
+ reasoning: true;
15933
+ input: ("image" | "text")[];
15934
+ cost: {
15935
+ input: number;
15936
+ output: number;
15937
+ cacheRead: number;
15938
+ cacheWrite: number;
15939
+ };
15940
+ contextWindow: number;
15941
+ maxTokens: number;
15942
+ };
14908
15943
  readonly "zai/glm-4.5": {
14909
15944
  id: string;
14910
15945
  name: string;
@@ -15625,6 +16660,28 @@ export declare const MODELS: {
15625
16660
  contextWindow: number;
15626
16661
  maxTokens: number;
15627
16662
  };
16663
+ readonly "glm-5v-turbo": {
16664
+ id: string;
16665
+ name: string;
16666
+ api: "openai-completions";
16667
+ provider: string;
16668
+ baseUrl: string;
16669
+ compat: {
16670
+ supportsDeveloperRole: false;
16671
+ thinkingFormat: "zai";
16672
+ zaiToolStream: true;
16673
+ };
16674
+ reasoning: true;
16675
+ input: ("image" | "text")[];
16676
+ cost: {
16677
+ input: number;
16678
+ output: number;
16679
+ cacheRead: number;
16680
+ cacheWrite: number;
16681
+ };
16682
+ contextWindow: number;
16683
+ maxTokens: number;
16684
+ };
15628
16685
  };
15629
16686
  };
15630
16687
  //# sourceMappingURL=models.generated.d.ts.map