@roo-code/types 1.86.0 → 1.87.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -60,12 +60,15 @@ __export(index_exports, {
60
60
  HUGGINGFACE_SLIDER_MIN: () => HUGGINGFACE_SLIDER_MIN,
61
61
  HUGGINGFACE_SLIDER_STEP: () => HUGGINGFACE_SLIDER_STEP,
62
62
  HUGGINGFACE_TEMPERATURE_MAX_VALUE: () => HUGGINGFACE_TEMPERATURE_MAX_VALUE,
63
+ IMAGE_GENERATION_MODELS: () => IMAGE_GENERATION_MODELS,
64
+ IMAGE_GENERATION_MODEL_IDS: () => IMAGE_GENERATION_MODEL_IDS,
63
65
  INSTANCE_TTL_SECONDS: () => INSTANCE_TTL_SECONDS,
64
66
  IO_INTELLIGENCE_CACHE_DURATION: () => IO_INTELLIGENCE_CACHE_DURATION,
65
67
  IpcMessageType: () => IpcMessageType,
66
68
  IpcOrigin: () => IpcOrigin,
67
69
  LMSTUDIO_DEFAULT_TEMPERATURE: () => LMSTUDIO_DEFAULT_TEMPERATURE,
68
70
  MAX_CHECKPOINT_TIMEOUT_SECONDS: () => MAX_CHECKPOINT_TIMEOUT_SECONDS,
71
+ MINIMAX_DEFAULT_MAX_TOKENS: () => MINIMAX_DEFAULT_MAX_TOKENS,
69
72
  MINIMAX_DEFAULT_TEMPERATURE: () => MINIMAX_DEFAULT_TEMPERATURE,
70
73
  MIN_CHECKPOINT_TIMEOUT_SECONDS: () => MIN_CHECKPOINT_TIMEOUT_SECONDS,
71
74
  MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
@@ -211,6 +214,7 @@ __export(index_exports, {
211
214
  mcpMarketplaceItemSchema: () => mcpMarketplaceItemSchema,
212
215
  mcpParameterSchema: () => mcpParameterSchema,
213
216
  minimaxDefaultModelId: () => minimaxDefaultModelId,
217
+ minimaxDefaultModelInfo: () => minimaxDefaultModelInfo,
214
218
  minimaxModels: () => minimaxModels,
215
219
  mistralDefaultModelId: () => mistralDefaultModelId,
216
220
  mistralModels: () => mistralModels,
@@ -260,7 +264,6 @@ __export(index_exports, {
260
264
  rooCodeSettingsSchema: () => rooCodeSettingsSchema,
261
265
  rooCodeTelemetryEventSchema: () => rooCodeTelemetryEventSchema,
262
266
  rooDefaultModelId: () => rooDefaultModelId,
263
- rooModelDefaults: () => rooModelDefaults,
264
267
  rooModels: () => rooModels,
265
268
  sambaNovaDefaultModelId: () => sambaNovaDefaultModelId,
266
269
  sambaNovaModels: () => sambaNovaModels,
@@ -2239,133 +2242,39 @@ var fireworksModels = {
2239
2242
  };
2240
2243
 
2241
2244
  // src/providers/gemini.ts
2242
- var geminiDefaultModelId = "gemini-2.0-flash-001";
2245
+ var geminiDefaultModelId = "gemini-2.5-pro";
2243
2246
  var geminiModels = {
2244
- // Latest models (pointing to the most recent stable versions)
2245
- "gemini-flash-latest": {
2246
- maxTokens: 65536,
2247
- contextWindow: 1048576,
2248
- supportsImages: true,
2249
- supportsPromptCache: true,
2250
- inputPrice: 0.3,
2251
- outputPrice: 2.5,
2252
- cacheReadsPrice: 0.075,
2253
- cacheWritesPrice: 1,
2254
- maxThinkingTokens: 24576,
2255
- supportsReasoningBudget: true
2256
- },
2257
- "gemini-flash-lite-latest": {
2258
- maxTokens: 65536,
2259
- contextWindow: 1048576,
2260
- supportsImages: true,
2261
- supportsPromptCache: true,
2262
- inputPrice: 0.1,
2263
- outputPrice: 0.4,
2264
- cacheReadsPrice: 0.025,
2265
- cacheWritesPrice: 1,
2266
- supportsReasoningBudget: true,
2267
- maxThinkingTokens: 24576
2268
- },
2269
- // 2.5 Flash models (09-2025 versions - most recent)
2270
- "gemini-2.5-flash-preview-09-2025": {
2271
- maxTokens: 65536,
2272
- contextWindow: 1048576,
2273
- supportsImages: true,
2274
- supportsPromptCache: true,
2275
- inputPrice: 0.3,
2276
- outputPrice: 2.5,
2277
- cacheReadsPrice: 0.075,
2278
- cacheWritesPrice: 1,
2279
- maxThinkingTokens: 24576,
2280
- supportsReasoningBudget: true
2281
- },
2282
- "gemini-2.5-flash-lite-preview-09-2025": {
2247
+ "gemini-3-pro-preview": {
2283
2248
  maxTokens: 65536,
2284
2249
  contextWindow: 1048576,
2285
2250
  supportsImages: true,
2251
+ supportsNativeTools: true,
2286
2252
  supportsPromptCache: true,
2287
- inputPrice: 0.1,
2288
- outputPrice: 0.4,
2289
- cacheReadsPrice: 0.025,
2290
- cacheWritesPrice: 1,
2291
- supportsReasoningBudget: true,
2292
- maxThinkingTokens: 24576
2293
- },
2294
- // 2.5 Flash models (06-17 version)
2295
- "gemini-2.5-flash-lite-preview-06-17": {
2296
- maxTokens: 64e3,
2297
- contextWindow: 1048576,
2298
- supportsImages: true,
2299
- supportsPromptCache: true,
2300
- inputPrice: 0.1,
2301
- outputPrice: 0.4,
2302
- cacheReadsPrice: 0.025,
2303
- cacheWritesPrice: 1,
2304
- supportsReasoningBudget: true,
2305
- maxThinkingTokens: 24576
2306
- },
2307
- // 2.5 Flash models (05-20 versions)
2308
- "gemini-2.5-flash-preview-05-20:thinking": {
2309
- maxTokens: 65535,
2310
- contextWindow: 1048576,
2311
- supportsImages: true,
2312
- supportsPromptCache: true,
2313
- inputPrice: 0.15,
2314
- outputPrice: 3.5,
2315
- cacheReadsPrice: 0.0375,
2316
- cacheWritesPrice: 1,
2317
- maxThinkingTokens: 24576,
2318
- supportsReasoningBudget: true,
2319
- requiredReasoningBudget: true
2320
- },
2321
- "gemini-2.5-flash-preview-05-20": {
2322
- maxTokens: 65535,
2323
- contextWindow: 1048576,
2324
- supportsImages: true,
2325
- supportsPromptCache: true,
2326
- inputPrice: 0.15,
2327
- outputPrice: 0.6,
2328
- cacheReadsPrice: 0.0375,
2329
- cacheWritesPrice: 1
2330
- },
2331
- // 2.5 Flash models (04-17 versions)
2332
- "gemini-2.5-flash-preview-04-17:thinking": {
2333
- maxTokens: 65535,
2334
- contextWindow: 1048576,
2335
- supportsImages: true,
2336
- supportsPromptCache: false,
2337
- inputPrice: 0.15,
2338
- outputPrice: 3.5,
2339
- maxThinkingTokens: 24576,
2340
- supportsReasoningBudget: true,
2341
- requiredReasoningBudget: true
2342
- },
2343
- "gemini-2.5-flash-preview-04-17": {
2344
- maxTokens: 65535,
2345
- contextWindow: 1048576,
2346
- supportsImages: true,
2347
- supportsPromptCache: false,
2348
- inputPrice: 0.15,
2349
- outputPrice: 0.6
2350
- },
2351
- // 2.5 Flash stable
2352
- "gemini-2.5-flash": {
2353
- maxTokens: 64e3,
2354
- contextWindow: 1048576,
2355
- supportsImages: true,
2356
- supportsPromptCache: true,
2357
- inputPrice: 0.3,
2358
- outputPrice: 2.5,
2359
- cacheReadsPrice: 0.075,
2360
- cacheWritesPrice: 1,
2361
- maxThinkingTokens: 24576,
2362
- supportsReasoningBudget: true
2253
+ supportsReasoningEffort: ["low", "high"],
2254
+ reasoningEffort: "low",
2255
+ supportsTemperature: true,
2256
+ defaultTemperature: 1,
2257
+ inputPrice: 4,
2258
+ outputPrice: 18,
2259
+ tiers: [
2260
+ {
2261
+ contextWindow: 2e5,
2262
+ inputPrice: 2,
2263
+ outputPrice: 12
2264
+ },
2265
+ {
2266
+ contextWindow: Infinity,
2267
+ inputPrice: 4,
2268
+ outputPrice: 18
2269
+ }
2270
+ ]
2363
2271
  },
2364
2272
  // 2.5 Pro models
2365
- "gemini-2.5-pro-preview-06-05": {
2366
- maxTokens: 65535,
2273
+ "gemini-2.5-pro": {
2274
+ maxTokens: 64e3,
2367
2275
  contextWindow: 1048576,
2368
2276
  supportsImages: true,
2277
+ supportsNativeTools: true,
2369
2278
  supportsPromptCache: true,
2370
2279
  inputPrice: 2.5,
2371
2280
  // This is the pricing for prompts above 200k tokens.
@@ -2374,6 +2283,7 @@ var geminiModels = {
2374
2283
  cacheWritesPrice: 4.5,
2375
2284
  maxThinkingTokens: 32768,
2376
2285
  supportsReasoningBudget: true,
2286
+ requiredReasoningBudget: true,
2377
2287
  tiers: [
2378
2288
  {
2379
2289
  contextWindow: 2e5,
@@ -2389,16 +2299,19 @@ var geminiModels = {
2389
2299
  }
2390
2300
  ]
2391
2301
  },
2392
- "gemini-2.5-pro-preview-05-06": {
2302
+ "gemini-2.5-pro-preview-06-05": {
2393
2303
  maxTokens: 65535,
2394
2304
  contextWindow: 1048576,
2395
2305
  supportsImages: true,
2306
+ supportsNativeTools: true,
2396
2307
  supportsPromptCache: true,
2397
2308
  inputPrice: 2.5,
2398
2309
  // This is the pricing for prompts above 200k tokens.
2399
2310
  outputPrice: 15,
2400
2311
  cacheReadsPrice: 0.625,
2401
2312
  cacheWritesPrice: 4.5,
2313
+ maxThinkingTokens: 32768,
2314
+ supportsReasoningBudget: true,
2402
2315
  tiers: [
2403
2316
  {
2404
2317
  contextWindow: 2e5,
@@ -2414,18 +2327,17 @@ var geminiModels = {
2414
2327
  }
2415
2328
  ]
2416
2329
  },
2417
- "gemini-2.5-pro-preview-03-25": {
2330
+ "gemini-2.5-pro-preview-05-06": {
2418
2331
  maxTokens: 65535,
2419
2332
  contextWindow: 1048576,
2420
2333
  supportsImages: true,
2334
+ supportsNativeTools: true,
2421
2335
  supportsPromptCache: true,
2422
2336
  inputPrice: 2.5,
2423
2337
  // This is the pricing for prompts above 200k tokens.
2424
2338
  outputPrice: 15,
2425
2339
  cacheReadsPrice: 0.625,
2426
2340
  cacheWritesPrice: 4.5,
2427
- maxThinkingTokens: 32768,
2428
- supportsReasoningBudget: true,
2429
2341
  tiers: [
2430
2342
  {
2431
2343
  contextWindow: 2e5,
@@ -2441,18 +2353,11 @@ var geminiModels = {
2441
2353
  }
2442
2354
  ]
2443
2355
  },
2444
- "gemini-2.5-pro-exp-03-25": {
2356
+ "gemini-2.5-pro-preview-03-25": {
2445
2357
  maxTokens: 65535,
2446
2358
  contextWindow: 1048576,
2447
2359
  supportsImages: true,
2448
- supportsPromptCache: false,
2449
- inputPrice: 0,
2450
- outputPrice: 0
2451
- },
2452
- "gemini-2.5-pro": {
2453
- maxTokens: 64e3,
2454
- contextWindow: 1048576,
2455
- supportsImages: true,
2360
+ supportsNativeTools: true,
2456
2361
  supportsPromptCache: true,
2457
2362
  inputPrice: 2.5,
2458
2363
  // This is the pricing for prompts above 200k tokens.
@@ -2461,7 +2366,6 @@ var geminiModels = {
2461
2366
  cacheWritesPrice: 4.5,
2462
2367
  maxThinkingTokens: 32768,
2463
2368
  supportsReasoningBudget: true,
2464
- requiredReasoningBudget: true,
2465
2369
  tiers: [
2466
2370
  {
2467
2371
  contextWindow: 2e5,
@@ -2477,125 +2381,72 @@ var geminiModels = {
2477
2381
  }
2478
2382
  ]
2479
2383
  },
2480
- // 2.0 Flash models
2481
- "gemini-2.0-flash-lite-preview-02-05": {
2482
- maxTokens: 8192,
2384
+ // 2.5 Flash models
2385
+ "gemini-flash-latest": {
2386
+ maxTokens: 65536,
2483
2387
  contextWindow: 1048576,
2484
2388
  supportsImages: true,
2485
- supportsPromptCache: false,
2486
- inputPrice: 0,
2487
- outputPrice: 0
2389
+ supportsNativeTools: true,
2390
+ supportsPromptCache: true,
2391
+ inputPrice: 0.3,
2392
+ outputPrice: 2.5,
2393
+ cacheReadsPrice: 0.075,
2394
+ cacheWritesPrice: 1,
2395
+ maxThinkingTokens: 24576,
2396
+ supportsReasoningBudget: true
2488
2397
  },
2489
- "gemini-2.0-flash-thinking-exp-01-21": {
2398
+ "gemini-2.5-flash-preview-09-2025": {
2490
2399
  maxTokens: 65536,
2491
2400
  contextWindow: 1048576,
2492
2401
  supportsImages: true,
2493
- supportsPromptCache: false,
2494
- inputPrice: 0,
2495
- outputPrice: 0
2496
- },
2497
- "gemini-2.0-flash-thinking-exp-1219": {
2498
- maxTokens: 8192,
2499
- contextWindow: 32767,
2500
- supportsImages: true,
2501
- supportsPromptCache: false,
2502
- inputPrice: 0,
2503
- outputPrice: 0
2402
+ supportsNativeTools: true,
2403
+ supportsPromptCache: true,
2404
+ inputPrice: 0.3,
2405
+ outputPrice: 2.5,
2406
+ cacheReadsPrice: 0.075,
2407
+ cacheWritesPrice: 1,
2408
+ maxThinkingTokens: 24576,
2409
+ supportsReasoningBudget: true
2504
2410
  },
2505
- "gemini-2.0-flash-exp": {
2506
- maxTokens: 8192,
2411
+ "gemini-2.5-flash": {
2412
+ maxTokens: 64e3,
2507
2413
  contextWindow: 1048576,
2508
2414
  supportsImages: true,
2509
- supportsPromptCache: false,
2510
- inputPrice: 0,
2511
- outputPrice: 0
2415
+ supportsNativeTools: true,
2416
+ supportsPromptCache: true,
2417
+ inputPrice: 0.3,
2418
+ outputPrice: 2.5,
2419
+ cacheReadsPrice: 0.075,
2420
+ cacheWritesPrice: 1,
2421
+ maxThinkingTokens: 24576,
2422
+ supportsReasoningBudget: true
2512
2423
  },
2513
- "gemini-2.0-flash-001": {
2514
- maxTokens: 8192,
2424
+ // 2.5 Flash Lite models
2425
+ "gemini-flash-lite-latest": {
2426
+ maxTokens: 65536,
2515
2427
  contextWindow: 1048576,
2516
2428
  supportsImages: true,
2429
+ supportsNativeTools: true,
2517
2430
  supportsPromptCache: true,
2518
2431
  inputPrice: 0.1,
2519
2432
  outputPrice: 0.4,
2520
2433
  cacheReadsPrice: 0.025,
2521
- cacheWritesPrice: 1
2522
- },
2523
- // 2.0 Pro models
2524
- "gemini-2.0-pro-exp-02-05": {
2525
- maxTokens: 8192,
2526
- contextWindow: 2097152,
2527
- supportsImages: true,
2528
- supportsPromptCache: false,
2529
- inputPrice: 0,
2530
- outputPrice: 0
2434
+ cacheWritesPrice: 1,
2435
+ supportsReasoningBudget: true,
2436
+ maxThinkingTokens: 24576
2531
2437
  },
2532
- // 1.5 Flash models
2533
- "gemini-1.5-flash-002": {
2534
- maxTokens: 8192,
2438
+ "gemini-2.5-flash-lite-preview-09-2025": {
2439
+ maxTokens: 65536,
2535
2440
  contextWindow: 1048576,
2536
2441
  supportsImages: true,
2442
+ supportsNativeTools: true,
2537
2443
  supportsPromptCache: true,
2538
- inputPrice: 0.15,
2539
- // This is the pricing for prompts above 128k tokens.
2540
- outputPrice: 0.6,
2541
- cacheReadsPrice: 0.0375,
2444
+ inputPrice: 0.1,
2445
+ outputPrice: 0.4,
2446
+ cacheReadsPrice: 0.025,
2542
2447
  cacheWritesPrice: 1,
2543
- tiers: [
2544
- {
2545
- contextWindow: 128e3,
2546
- inputPrice: 0.075,
2547
- outputPrice: 0.3,
2548
- cacheReadsPrice: 0.01875
2549
- },
2550
- {
2551
- contextWindow: Infinity,
2552
- inputPrice: 0.15,
2553
- outputPrice: 0.6,
2554
- cacheReadsPrice: 0.0375
2555
- }
2556
- ]
2557
- },
2558
- "gemini-1.5-flash-exp-0827": {
2559
- maxTokens: 8192,
2560
- contextWindow: 1048576,
2561
- supportsImages: true,
2562
- supportsPromptCache: false,
2563
- inputPrice: 0,
2564
- outputPrice: 0
2565
- },
2566
- "gemini-1.5-flash-8b-exp-0827": {
2567
- maxTokens: 8192,
2568
- contextWindow: 1048576,
2569
- supportsImages: true,
2570
- supportsPromptCache: false,
2571
- inputPrice: 0,
2572
- outputPrice: 0
2573
- },
2574
- // 1.5 Pro models
2575
- "gemini-1.5-pro-002": {
2576
- maxTokens: 8192,
2577
- contextWindow: 2097152,
2578
- supportsImages: true,
2579
- supportsPromptCache: false,
2580
- inputPrice: 0,
2581
- outputPrice: 0
2582
- },
2583
- "gemini-1.5-pro-exp-0827": {
2584
- maxTokens: 8192,
2585
- contextWindow: 2097152,
2586
- supportsImages: true,
2587
- supportsPromptCache: false,
2588
- inputPrice: 0,
2589
- outputPrice: 0
2590
- },
2591
- // Experimental models
2592
- "gemini-exp-1206": {
2593
- maxTokens: 8192,
2594
- contextWindow: 2097152,
2595
- supportsImages: true,
2596
- supportsPromptCache: false,
2597
- inputPrice: 0,
2598
- outputPrice: 0
2448
+ supportsReasoningBudget: true,
2449
+ maxThinkingTokens: 24576
2599
2450
  }
2600
2451
  };
2601
2452
 
@@ -2970,6 +2821,7 @@ var openAiNativeModels = {
2970
2821
  "gpt-5.1": {
2971
2822
  maxTokens: 128e3,
2972
2823
  contextWindow: 4e5,
2824
+ supportsNativeTools: true,
2973
2825
  supportsImages: true,
2974
2826
  supportsPromptCache: true,
2975
2827
  promptCacheRetention: "24h",
@@ -2989,6 +2841,7 @@ var openAiNativeModels = {
2989
2841
  "gpt-5.1-codex": {
2990
2842
  maxTokens: 128e3,
2991
2843
  contextWindow: 4e5,
2844
+ supportsNativeTools: true,
2992
2845
  supportsImages: true,
2993
2846
  supportsPromptCache: true,
2994
2847
  promptCacheRetention: "24h",
@@ -3004,6 +2857,7 @@ var openAiNativeModels = {
3004
2857
  "gpt-5.1-codex-mini": {
3005
2858
  maxTokens: 128e3,
3006
2859
  contextWindow: 4e5,
2860
+ supportsNativeTools: true,
3007
2861
  supportsImages: true,
3008
2862
  supportsPromptCache: true,
3009
2863
  promptCacheRetention: "24h",
@@ -3018,6 +2872,7 @@ var openAiNativeModels = {
3018
2872
  "gpt-5": {
3019
2873
  maxTokens: 128e3,
3020
2874
  contextWindow: 4e5,
2875
+ supportsNativeTools: true,
3021
2876
  supportsImages: true,
3022
2877
  supportsPromptCache: true,
3023
2878
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3036,6 +2891,7 @@ var openAiNativeModels = {
3036
2891
  "gpt-5-mini": {
3037
2892
  maxTokens: 128e3,
3038
2893
  contextWindow: 4e5,
2894
+ supportsNativeTools: true,
3039
2895
  supportsImages: true,
3040
2896
  supportsPromptCache: true,
3041
2897
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3054,6 +2910,7 @@ var openAiNativeModels = {
3054
2910
  "gpt-5-codex": {
3055
2911
  maxTokens: 128e3,
3056
2912
  contextWindow: 4e5,
2913
+ supportsNativeTools: true,
3057
2914
  supportsImages: true,
3058
2915
  supportsPromptCache: true,
3059
2916
  supportsReasoningEffort: ["low", "medium", "high"],
@@ -3068,6 +2925,7 @@ var openAiNativeModels = {
3068
2925
  "gpt-5-nano": {
3069
2926
  maxTokens: 128e3,
3070
2927
  contextWindow: 4e5,
2928
+ supportsNativeTools: true,
3071
2929
  supportsImages: true,
3072
2930
  supportsPromptCache: true,
3073
2931
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3083,6 +2941,7 @@ var openAiNativeModels = {
3083
2941
  "gpt-5-chat-latest": {
3084
2942
  maxTokens: 128e3,
3085
2943
  contextWindow: 4e5,
2944
+ supportsNativeTools: true,
3086
2945
  supportsImages: true,
3087
2946
  supportsPromptCache: true,
3088
2947
  inputPrice: 1.25,
@@ -3093,6 +2952,7 @@ var openAiNativeModels = {
3093
2952
  "gpt-4.1": {
3094
2953
  maxTokens: 32768,
3095
2954
  contextWindow: 1047576,
2955
+ supportsNativeTools: true,
3096
2956
  supportsImages: true,
3097
2957
  supportsPromptCache: true,
3098
2958
  inputPrice: 2,
@@ -3106,6 +2966,7 @@ var openAiNativeModels = {
3106
2966
  "gpt-4.1-mini": {
3107
2967
  maxTokens: 32768,
3108
2968
  contextWindow: 1047576,
2969
+ supportsNativeTools: true,
3109
2970
  supportsImages: true,
3110
2971
  supportsPromptCache: true,
3111
2972
  inputPrice: 0.4,
@@ -3119,6 +2980,7 @@ var openAiNativeModels = {
3119
2980
  "gpt-4.1-nano": {
3120
2981
  maxTokens: 32768,
3121
2982
  contextWindow: 1047576,
2983
+ supportsNativeTools: true,
3122
2984
  supportsImages: true,
3123
2985
  supportsPromptCache: true,
3124
2986
  inputPrice: 0.1,
@@ -3132,6 +2994,7 @@ var openAiNativeModels = {
3132
2994
  o3: {
3133
2995
  maxTokens: 1e5,
3134
2996
  contextWindow: 2e5,
2997
+ supportsNativeTools: true,
3135
2998
  supportsImages: true,
3136
2999
  supportsPromptCache: true,
3137
3000
  inputPrice: 2,
@@ -3148,6 +3011,7 @@ var openAiNativeModels = {
3148
3011
  "o3-high": {
3149
3012
  maxTokens: 1e5,
3150
3013
  contextWindow: 2e5,
3014
+ supportsNativeTools: true,
3151
3015
  supportsImages: true,
3152
3016
  supportsPromptCache: true,
3153
3017
  inputPrice: 2,
@@ -3159,6 +3023,7 @@ var openAiNativeModels = {
3159
3023
  "o3-low": {
3160
3024
  maxTokens: 1e5,
3161
3025
  contextWindow: 2e5,
3026
+ supportsNativeTools: true,
3162
3027
  supportsImages: true,
3163
3028
  supportsPromptCache: true,
3164
3029
  inputPrice: 2,
@@ -3170,6 +3035,7 @@ var openAiNativeModels = {
3170
3035
  "o4-mini": {
3171
3036
  maxTokens: 1e5,
3172
3037
  contextWindow: 2e5,
3038
+ supportsNativeTools: true,
3173
3039
  supportsImages: true,
3174
3040
  supportsPromptCache: true,
3175
3041
  inputPrice: 1.1,
@@ -3186,6 +3052,7 @@ var openAiNativeModels = {
3186
3052
  "o4-mini-high": {
3187
3053
  maxTokens: 1e5,
3188
3054
  contextWindow: 2e5,
3055
+ supportsNativeTools: true,
3189
3056
  supportsImages: true,
3190
3057
  supportsPromptCache: true,
3191
3058
  inputPrice: 1.1,
@@ -3197,6 +3064,7 @@ var openAiNativeModels = {
3197
3064
  "o4-mini-low": {
3198
3065
  maxTokens: 1e5,
3199
3066
  contextWindow: 2e5,
3067
+ supportsNativeTools: true,
3200
3068
  supportsImages: true,
3201
3069
  supportsPromptCache: true,
3202
3070
  inputPrice: 1.1,
@@ -3208,6 +3076,7 @@ var openAiNativeModels = {
3208
3076
  "o3-mini": {
3209
3077
  maxTokens: 1e5,
3210
3078
  contextWindow: 2e5,
3079
+ supportsNativeTools: true,
3211
3080
  supportsImages: false,
3212
3081
  supportsPromptCache: true,
3213
3082
  inputPrice: 1.1,
@@ -3220,6 +3089,7 @@ var openAiNativeModels = {
3220
3089
  "o3-mini-high": {
3221
3090
  maxTokens: 1e5,
3222
3091
  contextWindow: 2e5,
3092
+ supportsNativeTools: true,
3223
3093
  supportsImages: false,
3224
3094
  supportsPromptCache: true,
3225
3095
  inputPrice: 1.1,
@@ -3231,6 +3101,7 @@ var openAiNativeModels = {
3231
3101
  "o3-mini-low": {
3232
3102
  maxTokens: 1e5,
3233
3103
  contextWindow: 2e5,
3104
+ supportsNativeTools: true,
3234
3105
  supportsImages: false,
3235
3106
  supportsPromptCache: true,
3236
3107
  inputPrice: 1.1,
@@ -3242,6 +3113,7 @@ var openAiNativeModels = {
3242
3113
  o1: {
3243
3114
  maxTokens: 1e5,
3244
3115
  contextWindow: 2e5,
3116
+ supportsNativeTools: true,
3245
3117
  supportsImages: true,
3246
3118
  supportsPromptCache: true,
3247
3119
  inputPrice: 15,
@@ -3252,6 +3124,7 @@ var openAiNativeModels = {
3252
3124
  "o1-preview": {
3253
3125
  maxTokens: 32768,
3254
3126
  contextWindow: 128e3,
3127
+ supportsNativeTools: true,
3255
3128
  supportsImages: true,
3256
3129
  supportsPromptCache: true,
3257
3130
  inputPrice: 15,
@@ -3262,6 +3135,7 @@ var openAiNativeModels = {
3262
3135
  "o1-mini": {
3263
3136
  maxTokens: 65536,
3264
3137
  contextWindow: 128e3,
3138
+ supportsNativeTools: true,
3265
3139
  supportsImages: true,
3266
3140
  supportsPromptCache: true,
3267
3141
  inputPrice: 1.1,
@@ -3272,6 +3146,7 @@ var openAiNativeModels = {
3272
3146
  "gpt-4o": {
3273
3147
  maxTokens: 16384,
3274
3148
  contextWindow: 128e3,
3149
+ supportsNativeTools: true,
3275
3150
  supportsImages: true,
3276
3151
  supportsPromptCache: true,
3277
3152
  inputPrice: 2.5,
@@ -3285,6 +3160,7 @@ var openAiNativeModels = {
3285
3160
  "gpt-4o-mini": {
3286
3161
  maxTokens: 16384,
3287
3162
  contextWindow: 128e3,
3163
+ supportsNativeTools: true,
3288
3164
  supportsImages: true,
3289
3165
  supportsPromptCache: true,
3290
3166
  inputPrice: 0.15,
@@ -3298,6 +3174,7 @@ var openAiNativeModels = {
3298
3174
  "codex-mini-latest": {
3299
3175
  maxTokens: 16384,
3300
3176
  contextWindow: 2e5,
3177
+ supportsNativeTools: true,
3301
3178
  supportsImages: false,
3302
3179
  supportsPromptCache: false,
3303
3180
  inputPrice: 1.5,
@@ -3310,6 +3187,7 @@ var openAiNativeModels = {
3310
3187
  "gpt-5-2025-08-07": {
3311
3188
  maxTokens: 128e3,
3312
3189
  contextWindow: 4e5,
3190
+ supportsNativeTools: true,
3313
3191
  supportsImages: true,
3314
3192
  supportsPromptCache: true,
3315
3193
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3328,6 +3206,7 @@ var openAiNativeModels = {
3328
3206
  "gpt-5-mini-2025-08-07": {
3329
3207
  maxTokens: 128e3,
3330
3208
  contextWindow: 4e5,
3209
+ supportsNativeTools: true,
3331
3210
  supportsImages: true,
3332
3211
  supportsPromptCache: true,
3333
3212
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3346,6 +3225,7 @@ var openAiNativeModels = {
3346
3225
  "gpt-5-nano-2025-08-07": {
3347
3226
  maxTokens: 128e3,
3348
3227
  contextWindow: 4e5,
3228
+ supportsNativeTools: true,
3349
3229
  supportsImages: true,
3350
3230
  supportsPromptCache: true,
3351
3231
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3365,7 +3245,8 @@ var openAiModelInfoSaneDefaults = {
3365
3245
  supportsImages: true,
3366
3246
  supportsPromptCache: false,
3367
3247
  inputPrice: 0,
3368
- outputPrice: 0
3248
+ outputPrice: 0,
3249
+ supportsNativeTools: true
3369
3250
  };
3370
3251
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3371
3252
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3487,16 +3368,6 @@ var requestyDefaultModelInfo = {
3487
3368
  var import_zod7 = require("zod");
3488
3369
  var rooDefaultModelId = "xai/grok-code-fast-1";
3489
3370
  var rooModels = {};
3490
- var rooModelDefaults = {
3491
- // Add model-specific defaults below.
3492
- // You can configure defaultToolProtocol and other ModelInfo fields for specific model IDs.
3493
- "anthropic/claude-haiku-4.5": {
3494
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3495
- },
3496
- "minimax/minimax-m2:free": {
3497
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3498
- }
3499
- };
3500
3371
  var RooPricingSchema = import_zod7.z.object({
3501
3372
  input: import_zod7.z.string(),
3502
3373
  output: import_zod7.z.string(),
@@ -3634,6 +3505,30 @@ var unboundDefaultModelInfo = {
3634
3505
  // src/providers/vertex.ts
3635
3506
  var vertexDefaultModelId = "claude-sonnet-4-5@20250929";
3636
3507
  var vertexModels = {
3508
+ "gemini-3-pro-preview": {
3509
+ maxTokens: 65536,
3510
+ contextWindow: 1048576,
3511
+ supportsImages: true,
3512
+ supportsPromptCache: true,
3513
+ supportsReasoningEffort: ["low", "high"],
3514
+ reasoningEffort: "low",
3515
+ supportsTemperature: true,
3516
+ defaultTemperature: 1,
3517
+ inputPrice: 4,
3518
+ outputPrice: 18,
3519
+ tiers: [
3520
+ {
3521
+ contextWindow: 2e5,
3522
+ inputPrice: 2,
3523
+ outputPrice: 12
3524
+ },
3525
+ {
3526
+ contextWindow: Infinity,
3527
+ inputPrice: 4,
3528
+ outputPrice: 18
3529
+ }
3530
+ ]
3531
+ },
3637
3532
  "gemini-2.5-flash-preview-05-20:thinking": {
3638
3533
  maxTokens: 65535,
3639
3534
  contextWindow: 1048576,
@@ -4609,11 +4504,12 @@ var minimaxModels = {
4609
4504
  contextWindow: 192e3,
4610
4505
  supportsImages: false,
4611
4506
  supportsPromptCache: true,
4507
+ supportsNativeTools: true,
4508
+ preserveReasoning: true,
4612
4509
  inputPrice: 0.3,
4613
4510
  outputPrice: 1.2,
4614
4511
  cacheWritesPrice: 0.375,
4615
4512
  cacheReadsPrice: 0.03,
4616
- preserveReasoning: true,
4617
4513
  description: "MiniMax M2, a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4618
4514
  },
4619
4515
  "MiniMax-M2-Stable": {
@@ -4621,14 +4517,17 @@ var minimaxModels = {
4621
4517
  contextWindow: 192e3,
4622
4518
  supportsImages: false,
4623
4519
  supportsPromptCache: true,
4520
+ supportsNativeTools: true,
4521
+ preserveReasoning: true,
4624
4522
  inputPrice: 0.3,
4625
4523
  outputPrice: 1.2,
4626
4524
  cacheWritesPrice: 0.375,
4627
4525
  cacheReadsPrice: 0.03,
4628
- preserveReasoning: true,
4629
4526
  description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4630
4527
  }
4631
4528
  };
4529
+ var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
4530
+ var MINIMAX_DEFAULT_MAX_TOKENS = 16384;
4632
4531
  var MINIMAX_DEFAULT_TEMPERATURE = 1;
4633
4532
 
4634
4533
  // src/providers/index.ts
@@ -5129,7 +5028,7 @@ var modelIdKeysByProvider = {
5129
5028
  roo: "apiModelId",
5130
5029
  "vercel-ai-gateway": "vercelAiGatewayModelId"
5131
5030
  };
5132
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock"];
5031
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
5133
5032
  var getApiProtocol = (provider, modelId) => {
5134
5033
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
5135
5034
  return "anthropic";
@@ -5266,8 +5165,7 @@ var experimentIds = [
5266
5165
  "multiFileApplyDiff",
5267
5166
  "preventFocusDisruption",
5268
5167
  "imageGeneration",
5269
- "runSlashCommand",
5270
- "nativeToolCalling"
5168
+ "runSlashCommand"
5271
5169
  ];
5272
5170
  var experimentIdsSchema = import_zod10.z.enum(experimentIds);
5273
5171
  var experimentsSchema = import_zod10.z.object({
@@ -5275,8 +5173,7 @@ var experimentsSchema = import_zod10.z.object({
5275
5173
  multiFileApplyDiff: import_zod10.z.boolean().optional(),
5276
5174
  preventFocusDisruption: import_zod10.z.boolean().optional(),
5277
5175
  imageGeneration: import_zod10.z.boolean().optional(),
5278
- runSlashCommand: import_zod10.z.boolean().optional(),
5279
- nativeToolCalling: import_zod10.z.boolean().optional()
5176
+ runSlashCommand: import_zod10.z.boolean().optional()
5280
5177
  });
5281
5178
 
5282
5179
  // src/telemetry.ts
@@ -6306,6 +6203,15 @@ var followUpDataSchema = import_zod17.z.object({
6306
6203
  suggest: import_zod17.z.array(suggestionItemSchema).optional()
6307
6204
  });
6308
6205
 
6206
+ // src/image-generation.ts
6207
+ var IMAGE_GENERATION_MODELS = [
6208
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image" },
6209
+ { value: "google/gemini-3-pro-image-preview", label: "Gemini 3 Pro Image Preview" },
6210
+ { value: "openai/gpt-5-image", label: "GPT-5 Image" },
6211
+ { value: "openai/gpt-5-image-mini", label: "GPT-5 Image Mini" }
6212
+ ];
6213
+ var IMAGE_GENERATION_MODEL_IDS = IMAGE_GENERATION_MODELS.map((m) => m.value);
6214
+
6309
6215
  // src/ipc.ts
6310
6216
  var import_zod18 = require("zod");
6311
6217
  var IpcMessageType = /* @__PURE__ */ ((IpcMessageType2) => {
@@ -6494,12 +6400,15 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6494
6400
  HUGGINGFACE_SLIDER_MIN,
6495
6401
  HUGGINGFACE_SLIDER_STEP,
6496
6402
  HUGGINGFACE_TEMPERATURE_MAX_VALUE,
6403
+ IMAGE_GENERATION_MODELS,
6404
+ IMAGE_GENERATION_MODEL_IDS,
6497
6405
  INSTANCE_TTL_SECONDS,
6498
6406
  IO_INTELLIGENCE_CACHE_DURATION,
6499
6407
  IpcMessageType,
6500
6408
  IpcOrigin,
6501
6409
  LMSTUDIO_DEFAULT_TEMPERATURE,
6502
6410
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6411
+ MINIMAX_DEFAULT_MAX_TOKENS,
6503
6412
  MINIMAX_DEFAULT_TEMPERATURE,
6504
6413
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
6505
6414
  MISTRAL_DEFAULT_TEMPERATURE,
@@ -6645,6 +6554,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6645
6554
  mcpMarketplaceItemSchema,
6646
6555
  mcpParameterSchema,
6647
6556
  minimaxDefaultModelId,
6557
+ minimaxDefaultModelInfo,
6648
6558
  minimaxModels,
6649
6559
  mistralDefaultModelId,
6650
6560
  mistralModels,
@@ -6694,7 +6604,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6694
6604
  rooCodeSettingsSchema,
6695
6605
  rooCodeTelemetryEventSchema,
6696
6606
  rooDefaultModelId,
6697
- rooModelDefaults,
6698
6607
  rooModels,
6699
6608
  sambaNovaDefaultModelId,
6700
6609
  sambaNovaModels,