@roo-code/types 1.62.0 → 1.63.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -178,7 +178,8 @@ var toolNames = [
178
178
  "new_task",
179
179
  "fetch_instructions",
180
180
  "codebase_search",
181
- "update_todo_list"
181
+ "update_todo_list",
182
+ "generate_image"
182
183
  ];
183
184
  var toolNamesSchema = z3.enum(toolNames);
184
185
  var toolUsageSchema = z3.record(
@@ -357,12 +358,18 @@ var taskEventSchema = z4.discriminatedUnion("eventName", [
357
358
 
358
359
  // src/experiment.ts
359
360
  import { z as z5 } from "zod";
360
- var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption"];
361
+ var experimentIds = [
362
+ "powerSteering",
363
+ "multiFileApplyDiff",
364
+ "preventFocusDisruption",
365
+ "imageGeneration"
366
+ ];
361
367
  var experimentIdsSchema = z5.enum(experimentIds);
362
368
  var experimentsSchema = z5.object({
363
369
  powerSteering: z5.boolean().optional(),
364
370
  multiFileApplyDiff: z5.boolean().optional(),
365
- preventFocusDisruption: z5.boolean().optional()
371
+ preventFocusDisruption: z5.boolean().optional(),
372
+ imageGeneration: z5.boolean().optional()
366
373
  });
367
374
 
368
375
  // src/followup.ts
@@ -402,6 +409,8 @@ var modelInfoSchema = z7.object({
402
409
  // Capability flag to indicate whether the model supports an output verbosity parameter
403
410
  supportsVerbosity: z7.boolean().optional(),
404
411
  supportsReasoningBudget: z7.boolean().optional(),
412
+ // Capability flag to indicate whether the model supports temperature parameter
413
+ supportsTemperature: z7.boolean().optional(),
405
414
  requiredReasoningBudget: z7.boolean().optional(),
406
415
  supportsReasoningEffort: z7.boolean().optional(),
407
416
  supportedParameters: z7.array(modelParametersSchema).optional(),
@@ -2313,7 +2322,8 @@ var openAiNativeModels = {
2313
2322
  cacheReadsPrice: 0.13,
2314
2323
  description: "GPT-5: The best model for coding and agentic tasks across domains",
2315
2324
  // supportsVerbosity is a new capability; ensure ModelInfo includes it
2316
- supportsVerbosity: true
2325
+ supportsVerbosity: true,
2326
+ supportsTemperature: false
2317
2327
  },
2318
2328
  "gpt-5-mini-2025-08-07": {
2319
2329
  maxTokens: 128e3,
@@ -2326,7 +2336,8 @@ var openAiNativeModels = {
2326
2336
  outputPrice: 2,
2327
2337
  cacheReadsPrice: 0.03,
2328
2338
  description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
2329
- supportsVerbosity: true
2339
+ supportsVerbosity: true,
2340
+ supportsTemperature: false
2330
2341
  },
2331
2342
  "gpt-5-nano-2025-08-07": {
2332
2343
  maxTokens: 128e3,
@@ -2339,7 +2350,8 @@ var openAiNativeModels = {
2339
2350
  outputPrice: 0.4,
2340
2351
  cacheReadsPrice: 0.01,
2341
2352
  description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
2342
- supportsVerbosity: true
2353
+ supportsVerbosity: true,
2354
+ supportsTemperature: false
2343
2355
  },
2344
2356
  "gpt-4.1": {
2345
2357
  maxTokens: 32768,
@@ -2348,7 +2360,8 @@ var openAiNativeModels = {
2348
2360
  supportsPromptCache: true,
2349
2361
  inputPrice: 2,
2350
2362
  outputPrice: 8,
2351
- cacheReadsPrice: 0.5
2363
+ cacheReadsPrice: 0.5,
2364
+ supportsTemperature: true
2352
2365
  },
2353
2366
  "gpt-4.1-mini": {
2354
2367
  maxTokens: 32768,
@@ -2357,7 +2370,8 @@ var openAiNativeModels = {
2357
2370
  supportsPromptCache: true,
2358
2371
  inputPrice: 0.4,
2359
2372
  outputPrice: 1.6,
2360
- cacheReadsPrice: 0.1
2373
+ cacheReadsPrice: 0.1,
2374
+ supportsTemperature: true
2361
2375
  },
2362
2376
  "gpt-4.1-nano": {
2363
2377
  maxTokens: 32768,
@@ -2366,7 +2380,8 @@ var openAiNativeModels = {
2366
2380
  supportsPromptCache: true,
2367
2381
  inputPrice: 0.1,
2368
2382
  outputPrice: 0.4,
2369
- cacheReadsPrice: 0.025
2383
+ cacheReadsPrice: 0.025,
2384
+ supportsTemperature: true
2370
2385
  },
2371
2386
  o3: {
2372
2387
  maxTokens: 1e5,
@@ -2377,7 +2392,8 @@ var openAiNativeModels = {
2377
2392
  outputPrice: 8,
2378
2393
  cacheReadsPrice: 0.5,
2379
2394
  supportsReasoningEffort: true,
2380
- reasoningEffort: "medium"
2395
+ reasoningEffort: "medium",
2396
+ supportsTemperature: false
2381
2397
  },
2382
2398
  "o3-high": {
2383
2399
  maxTokens: 1e5,
@@ -2387,7 +2403,8 @@ var openAiNativeModels = {
2387
2403
  inputPrice: 2,
2388
2404
  outputPrice: 8,
2389
2405
  cacheReadsPrice: 0.5,
2390
- reasoningEffort: "high"
2406
+ reasoningEffort: "high",
2407
+ supportsTemperature: false
2391
2408
  },
2392
2409
  "o3-low": {
2393
2410
  maxTokens: 1e5,
@@ -2397,7 +2414,8 @@ var openAiNativeModels = {
2397
2414
  inputPrice: 2,
2398
2415
  outputPrice: 8,
2399
2416
  cacheReadsPrice: 0.5,
2400
- reasoningEffort: "low"
2417
+ reasoningEffort: "low",
2418
+ supportsTemperature: false
2401
2419
  },
2402
2420
  "o4-mini": {
2403
2421
  maxTokens: 1e5,
@@ -2408,7 +2426,8 @@ var openAiNativeModels = {
2408
2426
  outputPrice: 4.4,
2409
2427
  cacheReadsPrice: 0.275,
2410
2428
  supportsReasoningEffort: true,
2411
- reasoningEffort: "medium"
2429
+ reasoningEffort: "medium",
2430
+ supportsTemperature: false
2412
2431
  },
2413
2432
  "o4-mini-high": {
2414
2433
  maxTokens: 1e5,
@@ -2418,7 +2437,8 @@ var openAiNativeModels = {
2418
2437
  inputPrice: 1.1,
2419
2438
  outputPrice: 4.4,
2420
2439
  cacheReadsPrice: 0.275,
2421
- reasoningEffort: "high"
2440
+ reasoningEffort: "high",
2441
+ supportsTemperature: false
2422
2442
  },
2423
2443
  "o4-mini-low": {
2424
2444
  maxTokens: 1e5,
@@ -2428,7 +2448,8 @@ var openAiNativeModels = {
2428
2448
  inputPrice: 1.1,
2429
2449
  outputPrice: 4.4,
2430
2450
  cacheReadsPrice: 0.275,
2431
- reasoningEffort: "low"
2451
+ reasoningEffort: "low",
2452
+ supportsTemperature: false
2432
2453
  },
2433
2454
  "o3-mini": {
2434
2455
  maxTokens: 1e5,
@@ -2439,7 +2460,8 @@ var openAiNativeModels = {
2439
2460
  outputPrice: 4.4,
2440
2461
  cacheReadsPrice: 0.55,
2441
2462
  supportsReasoningEffort: true,
2442
- reasoningEffort: "medium"
2463
+ reasoningEffort: "medium",
2464
+ supportsTemperature: false
2443
2465
  },
2444
2466
  "o3-mini-high": {
2445
2467
  maxTokens: 1e5,
@@ -2449,7 +2471,8 @@ var openAiNativeModels = {
2449
2471
  inputPrice: 1.1,
2450
2472
  outputPrice: 4.4,
2451
2473
  cacheReadsPrice: 0.55,
2452
- reasoningEffort: "high"
2474
+ reasoningEffort: "high",
2475
+ supportsTemperature: false
2453
2476
  },
2454
2477
  "o3-mini-low": {
2455
2478
  maxTokens: 1e5,
@@ -2459,7 +2482,8 @@ var openAiNativeModels = {
2459
2482
  inputPrice: 1.1,
2460
2483
  outputPrice: 4.4,
2461
2484
  cacheReadsPrice: 0.55,
2462
- reasoningEffort: "low"
2485
+ reasoningEffort: "low",
2486
+ supportsTemperature: false
2463
2487
  },
2464
2488
  o1: {
2465
2489
  maxTokens: 1e5,
@@ -2468,7 +2492,8 @@ var openAiNativeModels = {
2468
2492
  supportsPromptCache: true,
2469
2493
  inputPrice: 15,
2470
2494
  outputPrice: 60,
2471
- cacheReadsPrice: 7.5
2495
+ cacheReadsPrice: 7.5,
2496
+ supportsTemperature: false
2472
2497
  },
2473
2498
  "o1-preview": {
2474
2499
  maxTokens: 32768,
@@ -2477,7 +2502,8 @@ var openAiNativeModels = {
2477
2502
  supportsPromptCache: true,
2478
2503
  inputPrice: 15,
2479
2504
  outputPrice: 60,
2480
- cacheReadsPrice: 7.5
2505
+ cacheReadsPrice: 7.5,
2506
+ supportsTemperature: false
2481
2507
  },
2482
2508
  "o1-mini": {
2483
2509
  maxTokens: 65536,
@@ -2486,7 +2512,8 @@ var openAiNativeModels = {
2486
2512
  supportsPromptCache: true,
2487
2513
  inputPrice: 1.1,
2488
2514
  outputPrice: 4.4,
2489
- cacheReadsPrice: 0.55
2515
+ cacheReadsPrice: 0.55,
2516
+ supportsTemperature: false
2490
2517
  },
2491
2518
  "gpt-4o": {
2492
2519
  maxTokens: 16384,
@@ -2495,7 +2522,8 @@ var openAiNativeModels = {
2495
2522
  supportsPromptCache: true,
2496
2523
  inputPrice: 2.5,
2497
2524
  outputPrice: 10,
2498
- cacheReadsPrice: 1.25
2525
+ cacheReadsPrice: 1.25,
2526
+ supportsTemperature: true
2499
2527
  },
2500
2528
  "gpt-4o-mini": {
2501
2529
  maxTokens: 16384,
@@ -2504,7 +2532,8 @@ var openAiNativeModels = {
2504
2532
  supportsPromptCache: true,
2505
2533
  inputPrice: 0.15,
2506
2534
  outputPrice: 0.6,
2507
- cacheReadsPrice: 0.075
2535
+ cacheReadsPrice: 0.075,
2536
+ supportsTemperature: true
2508
2537
  },
2509
2538
  "codex-mini-latest": {
2510
2539
  maxTokens: 16384,
@@ -2514,6 +2543,7 @@ var openAiNativeModels = {
2514
2543
  inputPrice: 1.5,
2515
2544
  outputPrice: 6,
2516
2545
  cacheReadsPrice: 0,
2546
+ supportsTemperature: false,
2517
2547
  description: "Codex Mini: Cloud-based software engineering agent powered by codex-1, a version of o3 optimized for coding tasks. Trained with reinforcement learning to generate human-style code, adhere to instructions, and iteratively run tests."
2518
2548
  }
2519
2549
  };
@@ -3630,7 +3660,12 @@ var openRouterSchema = baseProviderSettingsSchema.extend({
3630
3660
  openRouterModelId: z8.string().optional(),
3631
3661
  openRouterBaseUrl: z8.string().optional(),
3632
3662
  openRouterSpecificProvider: z8.string().optional(),
3633
- openRouterUseMiddleOutTransform: z8.boolean().optional()
3663
+ openRouterUseMiddleOutTransform: z8.boolean().optional(),
3664
+ // Image generation settings (experimental)
3665
+ openRouterImageGenerationSettings: z8.object({
3666
+ openRouterApiKey: z8.string().optional(),
3667
+ selectedModel: z8.string().optional()
3668
+ }).optional()
3634
3669
  });
3635
3670
  var bedrockSchema = apiModelIdProviderModelSchema.extend({
3636
3671
  awsAccessKey: z8.string().optional(),