@roo-code/types 1.86.0 → 1.87.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -1927,133 +1927,39 @@ var fireworksModels = {
1927
1927
  };
1928
1928
 
1929
1929
  // src/providers/gemini.ts
1930
- var geminiDefaultModelId = "gemini-2.0-flash-001";
1930
+ var geminiDefaultModelId = "gemini-2.5-pro";
1931
1931
  var geminiModels = {
1932
- // Latest models (pointing to the most recent stable versions)
1933
- "gemini-flash-latest": {
1934
- maxTokens: 65536,
1935
- contextWindow: 1048576,
1936
- supportsImages: true,
1937
- supportsPromptCache: true,
1938
- inputPrice: 0.3,
1939
- outputPrice: 2.5,
1940
- cacheReadsPrice: 0.075,
1941
- cacheWritesPrice: 1,
1942
- maxThinkingTokens: 24576,
1943
- supportsReasoningBudget: true
1944
- },
1945
- "gemini-flash-lite-latest": {
1946
- maxTokens: 65536,
1947
- contextWindow: 1048576,
1948
- supportsImages: true,
1949
- supportsPromptCache: true,
1950
- inputPrice: 0.1,
1951
- outputPrice: 0.4,
1952
- cacheReadsPrice: 0.025,
1953
- cacheWritesPrice: 1,
1954
- supportsReasoningBudget: true,
1955
- maxThinkingTokens: 24576
1956
- },
1957
- // 2.5 Flash models (09-2025 versions - most recent)
1958
- "gemini-2.5-flash-preview-09-2025": {
1959
- maxTokens: 65536,
1960
- contextWindow: 1048576,
1961
- supportsImages: true,
1962
- supportsPromptCache: true,
1963
- inputPrice: 0.3,
1964
- outputPrice: 2.5,
1965
- cacheReadsPrice: 0.075,
1966
- cacheWritesPrice: 1,
1967
- maxThinkingTokens: 24576,
1968
- supportsReasoningBudget: true
1969
- },
1970
- "gemini-2.5-flash-lite-preview-09-2025": {
1932
+ "gemini-3-pro-preview": {
1971
1933
  maxTokens: 65536,
1972
1934
  contextWindow: 1048576,
1973
1935
  supportsImages: true,
1936
+ supportsNativeTools: true,
1974
1937
  supportsPromptCache: true,
1975
- inputPrice: 0.1,
1976
- outputPrice: 0.4,
1977
- cacheReadsPrice: 0.025,
1978
- cacheWritesPrice: 1,
1979
- supportsReasoningBudget: true,
1980
- maxThinkingTokens: 24576
1981
- },
1982
- // 2.5 Flash models (06-17 version)
1983
- "gemini-2.5-flash-lite-preview-06-17": {
1984
- maxTokens: 64e3,
1985
- contextWindow: 1048576,
1986
- supportsImages: true,
1987
- supportsPromptCache: true,
1988
- inputPrice: 0.1,
1989
- outputPrice: 0.4,
1990
- cacheReadsPrice: 0.025,
1991
- cacheWritesPrice: 1,
1992
- supportsReasoningBudget: true,
1993
- maxThinkingTokens: 24576
1994
- },
1995
- // 2.5 Flash models (05-20 versions)
1996
- "gemini-2.5-flash-preview-05-20:thinking": {
1997
- maxTokens: 65535,
1998
- contextWindow: 1048576,
1999
- supportsImages: true,
2000
- supportsPromptCache: true,
2001
- inputPrice: 0.15,
2002
- outputPrice: 3.5,
2003
- cacheReadsPrice: 0.0375,
2004
- cacheWritesPrice: 1,
2005
- maxThinkingTokens: 24576,
2006
- supportsReasoningBudget: true,
2007
- requiredReasoningBudget: true
2008
- },
2009
- "gemini-2.5-flash-preview-05-20": {
2010
- maxTokens: 65535,
2011
- contextWindow: 1048576,
2012
- supportsImages: true,
2013
- supportsPromptCache: true,
2014
- inputPrice: 0.15,
2015
- outputPrice: 0.6,
2016
- cacheReadsPrice: 0.0375,
2017
- cacheWritesPrice: 1
2018
- },
2019
- // 2.5 Flash models (04-17 versions)
2020
- "gemini-2.5-flash-preview-04-17:thinking": {
2021
- maxTokens: 65535,
2022
- contextWindow: 1048576,
2023
- supportsImages: true,
2024
- supportsPromptCache: false,
2025
- inputPrice: 0.15,
2026
- outputPrice: 3.5,
2027
- maxThinkingTokens: 24576,
2028
- supportsReasoningBudget: true,
2029
- requiredReasoningBudget: true
2030
- },
2031
- "gemini-2.5-flash-preview-04-17": {
2032
- maxTokens: 65535,
2033
- contextWindow: 1048576,
2034
- supportsImages: true,
2035
- supportsPromptCache: false,
2036
- inputPrice: 0.15,
2037
- outputPrice: 0.6
2038
- },
2039
- // 2.5 Flash stable
2040
- "gemini-2.5-flash": {
2041
- maxTokens: 64e3,
2042
- contextWindow: 1048576,
2043
- supportsImages: true,
2044
- supportsPromptCache: true,
2045
- inputPrice: 0.3,
2046
- outputPrice: 2.5,
2047
- cacheReadsPrice: 0.075,
2048
- cacheWritesPrice: 1,
2049
- maxThinkingTokens: 24576,
2050
- supportsReasoningBudget: true
1938
+ supportsReasoningEffort: ["low", "high"],
1939
+ reasoningEffort: "low",
1940
+ supportsTemperature: true,
1941
+ defaultTemperature: 1,
1942
+ inputPrice: 4,
1943
+ outputPrice: 18,
1944
+ tiers: [
1945
+ {
1946
+ contextWindow: 2e5,
1947
+ inputPrice: 2,
1948
+ outputPrice: 12
1949
+ },
1950
+ {
1951
+ contextWindow: Infinity,
1952
+ inputPrice: 4,
1953
+ outputPrice: 18
1954
+ }
1955
+ ]
2051
1956
  },
2052
1957
  // 2.5 Pro models
2053
- "gemini-2.5-pro-preview-06-05": {
2054
- maxTokens: 65535,
1958
+ "gemini-2.5-pro": {
1959
+ maxTokens: 64e3,
2055
1960
  contextWindow: 1048576,
2056
1961
  supportsImages: true,
1962
+ supportsNativeTools: true,
2057
1963
  supportsPromptCache: true,
2058
1964
  inputPrice: 2.5,
2059
1965
  // This is the pricing for prompts above 200k tokens.
@@ -2062,6 +1968,7 @@ var geminiModels = {
2062
1968
  cacheWritesPrice: 4.5,
2063
1969
  maxThinkingTokens: 32768,
2064
1970
  supportsReasoningBudget: true,
1971
+ requiredReasoningBudget: true,
2065
1972
  tiers: [
2066
1973
  {
2067
1974
  contextWindow: 2e5,
@@ -2077,16 +1984,19 @@ var geminiModels = {
2077
1984
  }
2078
1985
  ]
2079
1986
  },
2080
- "gemini-2.5-pro-preview-05-06": {
1987
+ "gemini-2.5-pro-preview-06-05": {
2081
1988
  maxTokens: 65535,
2082
1989
  contextWindow: 1048576,
2083
1990
  supportsImages: true,
1991
+ supportsNativeTools: true,
2084
1992
  supportsPromptCache: true,
2085
1993
  inputPrice: 2.5,
2086
1994
  // This is the pricing for prompts above 200k tokens.
2087
1995
  outputPrice: 15,
2088
1996
  cacheReadsPrice: 0.625,
2089
1997
  cacheWritesPrice: 4.5,
1998
+ maxThinkingTokens: 32768,
1999
+ supportsReasoningBudget: true,
2090
2000
  tiers: [
2091
2001
  {
2092
2002
  contextWindow: 2e5,
@@ -2102,18 +2012,17 @@ var geminiModels = {
2102
2012
  }
2103
2013
  ]
2104
2014
  },
2105
- "gemini-2.5-pro-preview-03-25": {
2015
+ "gemini-2.5-pro-preview-05-06": {
2106
2016
  maxTokens: 65535,
2107
2017
  contextWindow: 1048576,
2108
2018
  supportsImages: true,
2019
+ supportsNativeTools: true,
2109
2020
  supportsPromptCache: true,
2110
2021
  inputPrice: 2.5,
2111
2022
  // This is the pricing for prompts above 200k tokens.
2112
2023
  outputPrice: 15,
2113
2024
  cacheReadsPrice: 0.625,
2114
2025
  cacheWritesPrice: 4.5,
2115
- maxThinkingTokens: 32768,
2116
- supportsReasoningBudget: true,
2117
2026
  tiers: [
2118
2027
  {
2119
2028
  contextWindow: 2e5,
@@ -2129,18 +2038,11 @@ var geminiModels = {
2129
2038
  }
2130
2039
  ]
2131
2040
  },
2132
- "gemini-2.5-pro-exp-03-25": {
2041
+ "gemini-2.5-pro-preview-03-25": {
2133
2042
  maxTokens: 65535,
2134
2043
  contextWindow: 1048576,
2135
2044
  supportsImages: true,
2136
- supportsPromptCache: false,
2137
- inputPrice: 0,
2138
- outputPrice: 0
2139
- },
2140
- "gemini-2.5-pro": {
2141
- maxTokens: 64e3,
2142
- contextWindow: 1048576,
2143
- supportsImages: true,
2045
+ supportsNativeTools: true,
2144
2046
  supportsPromptCache: true,
2145
2047
  inputPrice: 2.5,
2146
2048
  // This is the pricing for prompts above 200k tokens.
@@ -2149,7 +2051,6 @@ var geminiModels = {
2149
2051
  cacheWritesPrice: 4.5,
2150
2052
  maxThinkingTokens: 32768,
2151
2053
  supportsReasoningBudget: true,
2152
- requiredReasoningBudget: true,
2153
2054
  tiers: [
2154
2055
  {
2155
2056
  contextWindow: 2e5,
@@ -2165,125 +2066,72 @@ var geminiModels = {
2165
2066
  }
2166
2067
  ]
2167
2068
  },
2168
- // 2.0 Flash models
2169
- "gemini-2.0-flash-lite-preview-02-05": {
2170
- maxTokens: 8192,
2069
+ // 2.5 Flash models
2070
+ "gemini-flash-latest": {
2071
+ maxTokens: 65536,
2171
2072
  contextWindow: 1048576,
2172
2073
  supportsImages: true,
2173
- supportsPromptCache: false,
2174
- inputPrice: 0,
2175
- outputPrice: 0
2074
+ supportsNativeTools: true,
2075
+ supportsPromptCache: true,
2076
+ inputPrice: 0.3,
2077
+ outputPrice: 2.5,
2078
+ cacheReadsPrice: 0.075,
2079
+ cacheWritesPrice: 1,
2080
+ maxThinkingTokens: 24576,
2081
+ supportsReasoningBudget: true
2176
2082
  },
2177
- "gemini-2.0-flash-thinking-exp-01-21": {
2083
+ "gemini-2.5-flash-preview-09-2025": {
2178
2084
  maxTokens: 65536,
2179
2085
  contextWindow: 1048576,
2180
2086
  supportsImages: true,
2181
- supportsPromptCache: false,
2182
- inputPrice: 0,
2183
- outputPrice: 0
2184
- },
2185
- "gemini-2.0-flash-thinking-exp-1219": {
2186
- maxTokens: 8192,
2187
- contextWindow: 32767,
2188
- supportsImages: true,
2189
- supportsPromptCache: false,
2190
- inputPrice: 0,
2191
- outputPrice: 0
2087
+ supportsNativeTools: true,
2088
+ supportsPromptCache: true,
2089
+ inputPrice: 0.3,
2090
+ outputPrice: 2.5,
2091
+ cacheReadsPrice: 0.075,
2092
+ cacheWritesPrice: 1,
2093
+ maxThinkingTokens: 24576,
2094
+ supportsReasoningBudget: true
2192
2095
  },
2193
- "gemini-2.0-flash-exp": {
2194
- maxTokens: 8192,
2096
+ "gemini-2.5-flash": {
2097
+ maxTokens: 64e3,
2195
2098
  contextWindow: 1048576,
2196
2099
  supportsImages: true,
2197
- supportsPromptCache: false,
2198
- inputPrice: 0,
2199
- outputPrice: 0
2100
+ supportsNativeTools: true,
2101
+ supportsPromptCache: true,
2102
+ inputPrice: 0.3,
2103
+ outputPrice: 2.5,
2104
+ cacheReadsPrice: 0.075,
2105
+ cacheWritesPrice: 1,
2106
+ maxThinkingTokens: 24576,
2107
+ supportsReasoningBudget: true
2200
2108
  },
2201
- "gemini-2.0-flash-001": {
2202
- maxTokens: 8192,
2109
+ // 2.5 Flash Lite models
2110
+ "gemini-flash-lite-latest": {
2111
+ maxTokens: 65536,
2203
2112
  contextWindow: 1048576,
2204
2113
  supportsImages: true,
2114
+ supportsNativeTools: true,
2205
2115
  supportsPromptCache: true,
2206
2116
  inputPrice: 0.1,
2207
2117
  outputPrice: 0.4,
2208
2118
  cacheReadsPrice: 0.025,
2209
- cacheWritesPrice: 1
2210
- },
2211
- // 2.0 Pro models
2212
- "gemini-2.0-pro-exp-02-05": {
2213
- maxTokens: 8192,
2214
- contextWindow: 2097152,
2215
- supportsImages: true,
2216
- supportsPromptCache: false,
2217
- inputPrice: 0,
2218
- outputPrice: 0
2119
+ cacheWritesPrice: 1,
2120
+ supportsReasoningBudget: true,
2121
+ maxThinkingTokens: 24576
2219
2122
  },
2220
- // 1.5 Flash models
2221
- "gemini-1.5-flash-002": {
2222
- maxTokens: 8192,
2123
+ "gemini-2.5-flash-lite-preview-09-2025": {
2124
+ maxTokens: 65536,
2223
2125
  contextWindow: 1048576,
2224
2126
  supportsImages: true,
2127
+ supportsNativeTools: true,
2225
2128
  supportsPromptCache: true,
2226
- inputPrice: 0.15,
2227
- // This is the pricing for prompts above 128k tokens.
2228
- outputPrice: 0.6,
2229
- cacheReadsPrice: 0.0375,
2129
+ inputPrice: 0.1,
2130
+ outputPrice: 0.4,
2131
+ cacheReadsPrice: 0.025,
2230
2132
  cacheWritesPrice: 1,
2231
- tiers: [
2232
- {
2233
- contextWindow: 128e3,
2234
- inputPrice: 0.075,
2235
- outputPrice: 0.3,
2236
- cacheReadsPrice: 0.01875
2237
- },
2238
- {
2239
- contextWindow: Infinity,
2240
- inputPrice: 0.15,
2241
- outputPrice: 0.6,
2242
- cacheReadsPrice: 0.0375
2243
- }
2244
- ]
2245
- },
2246
- "gemini-1.5-flash-exp-0827": {
2247
- maxTokens: 8192,
2248
- contextWindow: 1048576,
2249
- supportsImages: true,
2250
- supportsPromptCache: false,
2251
- inputPrice: 0,
2252
- outputPrice: 0
2253
- },
2254
- "gemini-1.5-flash-8b-exp-0827": {
2255
- maxTokens: 8192,
2256
- contextWindow: 1048576,
2257
- supportsImages: true,
2258
- supportsPromptCache: false,
2259
- inputPrice: 0,
2260
- outputPrice: 0
2261
- },
2262
- // 1.5 Pro models
2263
- "gemini-1.5-pro-002": {
2264
- maxTokens: 8192,
2265
- contextWindow: 2097152,
2266
- supportsImages: true,
2267
- supportsPromptCache: false,
2268
- inputPrice: 0,
2269
- outputPrice: 0
2270
- },
2271
- "gemini-1.5-pro-exp-0827": {
2272
- maxTokens: 8192,
2273
- contextWindow: 2097152,
2274
- supportsImages: true,
2275
- supportsPromptCache: false,
2276
- inputPrice: 0,
2277
- outputPrice: 0
2278
- },
2279
- // Experimental models
2280
- "gemini-exp-1206": {
2281
- maxTokens: 8192,
2282
- contextWindow: 2097152,
2283
- supportsImages: true,
2284
- supportsPromptCache: false,
2285
- inputPrice: 0,
2286
- outputPrice: 0
2133
+ supportsReasoningBudget: true,
2134
+ maxThinkingTokens: 24576
2287
2135
  }
2288
2136
  };
2289
2137
 
@@ -2658,6 +2506,7 @@ var openAiNativeModels = {
2658
2506
  "gpt-5.1": {
2659
2507
  maxTokens: 128e3,
2660
2508
  contextWindow: 4e5,
2509
+ supportsNativeTools: true,
2661
2510
  supportsImages: true,
2662
2511
  supportsPromptCache: true,
2663
2512
  promptCacheRetention: "24h",
@@ -2677,6 +2526,7 @@ var openAiNativeModels = {
2677
2526
  "gpt-5.1-codex": {
2678
2527
  maxTokens: 128e3,
2679
2528
  contextWindow: 4e5,
2529
+ supportsNativeTools: true,
2680
2530
  supportsImages: true,
2681
2531
  supportsPromptCache: true,
2682
2532
  promptCacheRetention: "24h",
@@ -2692,6 +2542,7 @@ var openAiNativeModels = {
2692
2542
  "gpt-5.1-codex-mini": {
2693
2543
  maxTokens: 128e3,
2694
2544
  contextWindow: 4e5,
2545
+ supportsNativeTools: true,
2695
2546
  supportsImages: true,
2696
2547
  supportsPromptCache: true,
2697
2548
  promptCacheRetention: "24h",
@@ -2706,6 +2557,7 @@ var openAiNativeModels = {
2706
2557
  "gpt-5": {
2707
2558
  maxTokens: 128e3,
2708
2559
  contextWindow: 4e5,
2560
+ supportsNativeTools: true,
2709
2561
  supportsImages: true,
2710
2562
  supportsPromptCache: true,
2711
2563
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -2724,6 +2576,7 @@ var openAiNativeModels = {
2724
2576
  "gpt-5-mini": {
2725
2577
  maxTokens: 128e3,
2726
2578
  contextWindow: 4e5,
2579
+ supportsNativeTools: true,
2727
2580
  supportsImages: true,
2728
2581
  supportsPromptCache: true,
2729
2582
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -2742,6 +2595,7 @@ var openAiNativeModels = {
2742
2595
  "gpt-5-codex": {
2743
2596
  maxTokens: 128e3,
2744
2597
  contextWindow: 4e5,
2598
+ supportsNativeTools: true,
2745
2599
  supportsImages: true,
2746
2600
  supportsPromptCache: true,
2747
2601
  supportsReasoningEffort: ["low", "medium", "high"],
@@ -2756,6 +2610,7 @@ var openAiNativeModels = {
2756
2610
  "gpt-5-nano": {
2757
2611
  maxTokens: 128e3,
2758
2612
  contextWindow: 4e5,
2613
+ supportsNativeTools: true,
2759
2614
  supportsImages: true,
2760
2615
  supportsPromptCache: true,
2761
2616
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -2771,6 +2626,7 @@ var openAiNativeModels = {
2771
2626
  "gpt-5-chat-latest": {
2772
2627
  maxTokens: 128e3,
2773
2628
  contextWindow: 4e5,
2629
+ supportsNativeTools: true,
2774
2630
  supportsImages: true,
2775
2631
  supportsPromptCache: true,
2776
2632
  inputPrice: 1.25,
@@ -2781,6 +2637,7 @@ var openAiNativeModels = {
2781
2637
  "gpt-4.1": {
2782
2638
  maxTokens: 32768,
2783
2639
  contextWindow: 1047576,
2640
+ supportsNativeTools: true,
2784
2641
  supportsImages: true,
2785
2642
  supportsPromptCache: true,
2786
2643
  inputPrice: 2,
@@ -2794,6 +2651,7 @@ var openAiNativeModels = {
2794
2651
  "gpt-4.1-mini": {
2795
2652
  maxTokens: 32768,
2796
2653
  contextWindow: 1047576,
2654
+ supportsNativeTools: true,
2797
2655
  supportsImages: true,
2798
2656
  supportsPromptCache: true,
2799
2657
  inputPrice: 0.4,
@@ -2807,6 +2665,7 @@ var openAiNativeModels = {
2807
2665
  "gpt-4.1-nano": {
2808
2666
  maxTokens: 32768,
2809
2667
  contextWindow: 1047576,
2668
+ supportsNativeTools: true,
2810
2669
  supportsImages: true,
2811
2670
  supportsPromptCache: true,
2812
2671
  inputPrice: 0.1,
@@ -2820,6 +2679,7 @@ var openAiNativeModels = {
2820
2679
  o3: {
2821
2680
  maxTokens: 1e5,
2822
2681
  contextWindow: 2e5,
2682
+ supportsNativeTools: true,
2823
2683
  supportsImages: true,
2824
2684
  supportsPromptCache: true,
2825
2685
  inputPrice: 2,
@@ -2836,6 +2696,7 @@ var openAiNativeModels = {
2836
2696
  "o3-high": {
2837
2697
  maxTokens: 1e5,
2838
2698
  contextWindow: 2e5,
2699
+ supportsNativeTools: true,
2839
2700
  supportsImages: true,
2840
2701
  supportsPromptCache: true,
2841
2702
  inputPrice: 2,
@@ -2847,6 +2708,7 @@ var openAiNativeModels = {
2847
2708
  "o3-low": {
2848
2709
  maxTokens: 1e5,
2849
2710
  contextWindow: 2e5,
2711
+ supportsNativeTools: true,
2850
2712
  supportsImages: true,
2851
2713
  supportsPromptCache: true,
2852
2714
  inputPrice: 2,
@@ -2858,6 +2720,7 @@ var openAiNativeModels = {
2858
2720
  "o4-mini": {
2859
2721
  maxTokens: 1e5,
2860
2722
  contextWindow: 2e5,
2723
+ supportsNativeTools: true,
2861
2724
  supportsImages: true,
2862
2725
  supportsPromptCache: true,
2863
2726
  inputPrice: 1.1,
@@ -2874,6 +2737,7 @@ var openAiNativeModels = {
2874
2737
  "o4-mini-high": {
2875
2738
  maxTokens: 1e5,
2876
2739
  contextWindow: 2e5,
2740
+ supportsNativeTools: true,
2877
2741
  supportsImages: true,
2878
2742
  supportsPromptCache: true,
2879
2743
  inputPrice: 1.1,
@@ -2885,6 +2749,7 @@ var openAiNativeModels = {
2885
2749
  "o4-mini-low": {
2886
2750
  maxTokens: 1e5,
2887
2751
  contextWindow: 2e5,
2752
+ supportsNativeTools: true,
2888
2753
  supportsImages: true,
2889
2754
  supportsPromptCache: true,
2890
2755
  inputPrice: 1.1,
@@ -2896,6 +2761,7 @@ var openAiNativeModels = {
2896
2761
  "o3-mini": {
2897
2762
  maxTokens: 1e5,
2898
2763
  contextWindow: 2e5,
2764
+ supportsNativeTools: true,
2899
2765
  supportsImages: false,
2900
2766
  supportsPromptCache: true,
2901
2767
  inputPrice: 1.1,
@@ -2908,6 +2774,7 @@ var openAiNativeModels = {
2908
2774
  "o3-mini-high": {
2909
2775
  maxTokens: 1e5,
2910
2776
  contextWindow: 2e5,
2777
+ supportsNativeTools: true,
2911
2778
  supportsImages: false,
2912
2779
  supportsPromptCache: true,
2913
2780
  inputPrice: 1.1,
@@ -2919,6 +2786,7 @@ var openAiNativeModels = {
2919
2786
  "o3-mini-low": {
2920
2787
  maxTokens: 1e5,
2921
2788
  contextWindow: 2e5,
2789
+ supportsNativeTools: true,
2922
2790
  supportsImages: false,
2923
2791
  supportsPromptCache: true,
2924
2792
  inputPrice: 1.1,
@@ -2930,6 +2798,7 @@ var openAiNativeModels = {
2930
2798
  o1: {
2931
2799
  maxTokens: 1e5,
2932
2800
  contextWindow: 2e5,
2801
+ supportsNativeTools: true,
2933
2802
  supportsImages: true,
2934
2803
  supportsPromptCache: true,
2935
2804
  inputPrice: 15,
@@ -2940,6 +2809,7 @@ var openAiNativeModels = {
2940
2809
  "o1-preview": {
2941
2810
  maxTokens: 32768,
2942
2811
  contextWindow: 128e3,
2812
+ supportsNativeTools: true,
2943
2813
  supportsImages: true,
2944
2814
  supportsPromptCache: true,
2945
2815
  inputPrice: 15,
@@ -2950,6 +2820,7 @@ var openAiNativeModels = {
2950
2820
  "o1-mini": {
2951
2821
  maxTokens: 65536,
2952
2822
  contextWindow: 128e3,
2823
+ supportsNativeTools: true,
2953
2824
  supportsImages: true,
2954
2825
  supportsPromptCache: true,
2955
2826
  inputPrice: 1.1,
@@ -2960,6 +2831,7 @@ var openAiNativeModels = {
2960
2831
  "gpt-4o": {
2961
2832
  maxTokens: 16384,
2962
2833
  contextWindow: 128e3,
2834
+ supportsNativeTools: true,
2963
2835
  supportsImages: true,
2964
2836
  supportsPromptCache: true,
2965
2837
  inputPrice: 2.5,
@@ -2973,6 +2845,7 @@ var openAiNativeModels = {
2973
2845
  "gpt-4o-mini": {
2974
2846
  maxTokens: 16384,
2975
2847
  contextWindow: 128e3,
2848
+ supportsNativeTools: true,
2976
2849
  supportsImages: true,
2977
2850
  supportsPromptCache: true,
2978
2851
  inputPrice: 0.15,
@@ -2986,6 +2859,7 @@ var openAiNativeModels = {
2986
2859
  "codex-mini-latest": {
2987
2860
  maxTokens: 16384,
2988
2861
  contextWindow: 2e5,
2862
+ supportsNativeTools: true,
2989
2863
  supportsImages: false,
2990
2864
  supportsPromptCache: false,
2991
2865
  inputPrice: 1.5,
@@ -2998,6 +2872,7 @@ var openAiNativeModels = {
2998
2872
  "gpt-5-2025-08-07": {
2999
2873
  maxTokens: 128e3,
3000
2874
  contextWindow: 4e5,
2875
+ supportsNativeTools: true,
3001
2876
  supportsImages: true,
3002
2877
  supportsPromptCache: true,
3003
2878
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3016,6 +2891,7 @@ var openAiNativeModels = {
3016
2891
  "gpt-5-mini-2025-08-07": {
3017
2892
  maxTokens: 128e3,
3018
2893
  contextWindow: 4e5,
2894
+ supportsNativeTools: true,
3019
2895
  supportsImages: true,
3020
2896
  supportsPromptCache: true,
3021
2897
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3034,6 +2910,7 @@ var openAiNativeModels = {
3034
2910
  "gpt-5-nano-2025-08-07": {
3035
2911
  maxTokens: 128e3,
3036
2912
  contextWindow: 4e5,
2913
+ supportsNativeTools: true,
3037
2914
  supportsImages: true,
3038
2915
  supportsPromptCache: true,
3039
2916
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3053,7 +2930,8 @@ var openAiModelInfoSaneDefaults = {
3053
2930
  supportsImages: true,
3054
2931
  supportsPromptCache: false,
3055
2932
  inputPrice: 0,
3056
- outputPrice: 0
2933
+ outputPrice: 0,
2934
+ supportsNativeTools: true
3057
2935
  };
3058
2936
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3059
2937
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3175,16 +3053,6 @@ var requestyDefaultModelInfo = {
3175
3053
  import { z as z7 } from "zod";
3176
3054
  var rooDefaultModelId = "xai/grok-code-fast-1";
3177
3055
  var rooModels = {};
3178
- var rooModelDefaults = {
3179
- // Add model-specific defaults below.
3180
- // You can configure defaultToolProtocol and other ModelInfo fields for specific model IDs.
3181
- "anthropic/claude-haiku-4.5": {
3182
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3183
- },
3184
- "minimax/minimax-m2:free": {
3185
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3186
- }
3187
- };
3188
3056
  var RooPricingSchema = z7.object({
3189
3057
  input: z7.string(),
3190
3058
  output: z7.string(),
@@ -3322,6 +3190,30 @@ var unboundDefaultModelInfo = {
3322
3190
  // src/providers/vertex.ts
3323
3191
  var vertexDefaultModelId = "claude-sonnet-4-5@20250929";
3324
3192
  var vertexModels = {
3193
+ "gemini-3-pro-preview": {
3194
+ maxTokens: 65536,
3195
+ contextWindow: 1048576,
3196
+ supportsImages: true,
3197
+ supportsPromptCache: true,
3198
+ supportsReasoningEffort: ["low", "high"],
3199
+ reasoningEffort: "low",
3200
+ supportsTemperature: true,
3201
+ defaultTemperature: 1,
3202
+ inputPrice: 4,
3203
+ outputPrice: 18,
3204
+ tiers: [
3205
+ {
3206
+ contextWindow: 2e5,
3207
+ inputPrice: 2,
3208
+ outputPrice: 12
3209
+ },
3210
+ {
3211
+ contextWindow: Infinity,
3212
+ inputPrice: 4,
3213
+ outputPrice: 18
3214
+ }
3215
+ ]
3216
+ },
3325
3217
  "gemini-2.5-flash-preview-05-20:thinking": {
3326
3218
  maxTokens: 65535,
3327
3219
  contextWindow: 1048576,
@@ -4297,11 +4189,12 @@ var minimaxModels = {
4297
4189
  contextWindow: 192e3,
4298
4190
  supportsImages: false,
4299
4191
  supportsPromptCache: true,
4192
+ supportsNativeTools: true,
4193
+ preserveReasoning: true,
4300
4194
  inputPrice: 0.3,
4301
4195
  outputPrice: 1.2,
4302
4196
  cacheWritesPrice: 0.375,
4303
4197
  cacheReadsPrice: 0.03,
4304
- preserveReasoning: true,
4305
4198
  description: "MiniMax M2, a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4306
4199
  },
4307
4200
  "MiniMax-M2-Stable": {
@@ -4309,14 +4202,17 @@ var minimaxModels = {
4309
4202
  contextWindow: 192e3,
4310
4203
  supportsImages: false,
4311
4204
  supportsPromptCache: true,
4205
+ supportsNativeTools: true,
4206
+ preserveReasoning: true,
4312
4207
  inputPrice: 0.3,
4313
4208
  outputPrice: 1.2,
4314
4209
  cacheWritesPrice: 0.375,
4315
4210
  cacheReadsPrice: 0.03,
4316
- preserveReasoning: true,
4317
4211
  description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4318
4212
  }
4319
4213
  };
4214
+ var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
4215
+ var MINIMAX_DEFAULT_MAX_TOKENS = 16384;
4320
4216
  var MINIMAX_DEFAULT_TEMPERATURE = 1;
4321
4217
 
4322
4218
  // src/providers/index.ts
@@ -4817,7 +4713,7 @@ var modelIdKeysByProvider = {
4817
4713
  roo: "apiModelId",
4818
4714
  "vercel-ai-gateway": "vercelAiGatewayModelId"
4819
4715
  };
4820
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock"];
4716
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
4821
4717
  var getApiProtocol = (provider, modelId) => {
4822
4718
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
4823
4719
  return "anthropic";
@@ -4954,8 +4850,7 @@ var experimentIds = [
4954
4850
  "multiFileApplyDiff",
4955
4851
  "preventFocusDisruption",
4956
4852
  "imageGeneration",
4957
- "runSlashCommand",
4958
- "nativeToolCalling"
4853
+ "runSlashCommand"
4959
4854
  ];
4960
4855
  var experimentIdsSchema = z10.enum(experimentIds);
4961
4856
  var experimentsSchema = z10.object({
@@ -4963,8 +4858,7 @@ var experimentsSchema = z10.object({
4963
4858
  multiFileApplyDiff: z10.boolean().optional(),
4964
4859
  preventFocusDisruption: z10.boolean().optional(),
4965
4860
  imageGeneration: z10.boolean().optional(),
4966
- runSlashCommand: z10.boolean().optional(),
4967
- nativeToolCalling: z10.boolean().optional()
4861
+ runSlashCommand: z10.boolean().optional()
4968
4862
  });
4969
4863
 
4970
4864
  // src/telemetry.ts
@@ -5994,6 +5888,15 @@ var followUpDataSchema = z17.object({
5994
5888
  suggest: z17.array(suggestionItemSchema).optional()
5995
5889
  });
5996
5890
 
5891
+ // src/image-generation.ts
5892
+ var IMAGE_GENERATION_MODELS = [
5893
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image" },
5894
+ { value: "google/gemini-3-pro-image-preview", label: "Gemini 3 Pro Image Preview" },
5895
+ { value: "openai/gpt-5-image", label: "GPT-5 Image" },
5896
+ { value: "openai/gpt-5-image-mini", label: "GPT-5 Image Mini" }
5897
+ ];
5898
+ var IMAGE_GENERATION_MODEL_IDS = IMAGE_GENERATION_MODELS.map((m) => m.value);
5899
+
5997
5900
  // src/ipc.ts
5998
5901
  import { z as z18 } from "zod";
5999
5902
  var IpcMessageType = /* @__PURE__ */ ((IpcMessageType2) => {
@@ -6181,12 +6084,15 @@ export {
6181
6084
  HUGGINGFACE_SLIDER_MIN,
6182
6085
  HUGGINGFACE_SLIDER_STEP,
6183
6086
  HUGGINGFACE_TEMPERATURE_MAX_VALUE,
6087
+ IMAGE_GENERATION_MODELS,
6088
+ IMAGE_GENERATION_MODEL_IDS,
6184
6089
  INSTANCE_TTL_SECONDS,
6185
6090
  IO_INTELLIGENCE_CACHE_DURATION,
6186
6091
  IpcMessageType,
6187
6092
  IpcOrigin,
6188
6093
  LMSTUDIO_DEFAULT_TEMPERATURE,
6189
6094
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6095
+ MINIMAX_DEFAULT_MAX_TOKENS,
6190
6096
  MINIMAX_DEFAULT_TEMPERATURE,
6191
6097
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
6192
6098
  MISTRAL_DEFAULT_TEMPERATURE,
@@ -6332,6 +6238,7 @@ export {
6332
6238
  mcpMarketplaceItemSchema,
6333
6239
  mcpParameterSchema,
6334
6240
  minimaxDefaultModelId,
6241
+ minimaxDefaultModelInfo,
6335
6242
  minimaxModels,
6336
6243
  mistralDefaultModelId,
6337
6244
  mistralModels,
@@ -6381,7 +6288,6 @@ export {
6381
6288
  rooCodeSettingsSchema,
6382
6289
  rooCodeTelemetryEventSchema,
6383
6290
  rooDefaultModelId,
6384
- rooModelDefaults,
6385
6291
  rooModels,
6386
6292
  sambaNovaDefaultModelId,
6387
6293
  sambaNovaModels,