@roo-code/types 1.61.0 → 1.63.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -259,7 +259,7 @@ var CODEBASE_INDEX_DEFAULTS = {
259
259
  var codebaseIndexConfigSchema = import_zod.z.object({
260
260
  codebaseIndexEnabled: import_zod.z.boolean().optional(),
261
261
  codebaseIndexQdrantUrl: import_zod.z.string().optional(),
262
- codebaseIndexEmbedderProvider: import_zod.z.enum(["openai", "ollama", "openai-compatible", "gemini", "mistral"]).optional(),
262
+ codebaseIndexEmbedderProvider: import_zod.z.enum(["openai", "ollama", "openai-compatible", "gemini", "mistral", "vercel-ai-gateway"]).optional(),
263
263
  codebaseIndexEmbedderBaseUrl: import_zod.z.string().optional(),
264
264
  codebaseIndexEmbedderModelId: import_zod.z.string().optional(),
265
265
  codebaseIndexEmbedderModelDimension: import_zod.z.number().optional(),
@@ -274,7 +274,8 @@ var codebaseIndexModelsSchema = import_zod.z.object({
274
274
  ollama: import_zod.z.record(import_zod.z.string(), import_zod.z.object({ dimension: import_zod.z.number() })).optional(),
275
275
  "openai-compatible": import_zod.z.record(import_zod.z.string(), import_zod.z.object({ dimension: import_zod.z.number() })).optional(),
276
276
  gemini: import_zod.z.record(import_zod.z.string(), import_zod.z.object({ dimension: import_zod.z.number() })).optional(),
277
- mistral: import_zod.z.record(import_zod.z.string(), import_zod.z.object({ dimension: import_zod.z.number() })).optional()
277
+ mistral: import_zod.z.record(import_zod.z.string(), import_zod.z.object({ dimension: import_zod.z.number() })).optional(),
278
+ "vercel-ai-gateway": import_zod.z.record(import_zod.z.string(), import_zod.z.object({ dimension: import_zod.z.number() })).optional()
278
279
  });
279
280
  var codebaseIndexProviderSchema = import_zod.z.object({
280
281
  codeIndexOpenAiKey: import_zod.z.string().optional(),
@@ -283,7 +284,8 @@ var codebaseIndexProviderSchema = import_zod.z.object({
283
284
  codebaseIndexOpenAiCompatibleApiKey: import_zod.z.string().optional(),
284
285
  codebaseIndexOpenAiCompatibleModelDimension: import_zod.z.number().optional(),
285
286
  codebaseIndexGeminiApiKey: import_zod.z.string().optional(),
286
- codebaseIndexMistralApiKey: import_zod.z.string().optional()
287
+ codebaseIndexMistralApiKey: import_zod.z.string().optional(),
288
+ codebaseIndexVercelAiGatewayApiKey: import_zod.z.string().optional()
287
289
  });
288
290
 
289
291
  // src/events.ts
@@ -422,7 +424,8 @@ var toolNames = [
422
424
  "new_task",
423
425
  "fetch_instructions",
424
426
  "codebase_search",
425
- "update_todo_list"
427
+ "update_todo_list",
428
+ "generate_image"
426
429
  ];
427
430
  var toolNamesSchema = import_zod3.z.enum(toolNames);
428
431
  var toolUsageSchema = import_zod3.z.record(
@@ -601,12 +604,18 @@ var taskEventSchema = import_zod4.z.discriminatedUnion("eventName", [
601
604
 
602
605
  // src/experiment.ts
603
606
  var import_zod5 = require("zod");
604
- var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption"];
607
+ var experimentIds = [
608
+ "powerSteering",
609
+ "multiFileApplyDiff",
610
+ "preventFocusDisruption",
611
+ "imageGeneration"
612
+ ];
605
613
  var experimentIdsSchema = import_zod5.z.enum(experimentIds);
606
614
  var experimentsSchema = import_zod5.z.object({
607
615
  powerSteering: import_zod5.z.boolean().optional(),
608
616
  multiFileApplyDiff: import_zod5.z.boolean().optional(),
609
- preventFocusDisruption: import_zod5.z.boolean().optional()
617
+ preventFocusDisruption: import_zod5.z.boolean().optional(),
618
+ imageGeneration: import_zod5.z.boolean().optional()
610
619
  });
611
620
 
612
621
  // src/followup.ts
@@ -646,6 +655,8 @@ var modelInfoSchema = import_zod7.z.object({
646
655
  // Capability flag to indicate whether the model supports an output verbosity parameter
647
656
  supportsVerbosity: import_zod7.z.boolean().optional(),
648
657
  supportsReasoningBudget: import_zod7.z.boolean().optional(),
658
+ // Capability flag to indicate whether the model supports temperature parameter
659
+ supportsTemperature: import_zod7.z.boolean().optional(),
649
660
  requiredReasoningBudget: import_zod7.z.boolean().optional(),
650
661
  supportsReasoningEffort: import_zod7.z.boolean().optional(),
651
662
  supportedParameters: import_zod7.z.array(modelParametersSchema).optional(),
@@ -2557,7 +2568,8 @@ var openAiNativeModels = {
2557
2568
  cacheReadsPrice: 0.13,
2558
2569
  description: "GPT-5: The best model for coding and agentic tasks across domains",
2559
2570
  // supportsVerbosity is a new capability; ensure ModelInfo includes it
2560
- supportsVerbosity: true
2571
+ supportsVerbosity: true,
2572
+ supportsTemperature: false
2561
2573
  },
2562
2574
  "gpt-5-mini-2025-08-07": {
2563
2575
  maxTokens: 128e3,
@@ -2570,7 +2582,8 @@ var openAiNativeModels = {
2570
2582
  outputPrice: 2,
2571
2583
  cacheReadsPrice: 0.03,
2572
2584
  description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
2573
- supportsVerbosity: true
2585
+ supportsVerbosity: true,
2586
+ supportsTemperature: false
2574
2587
  },
2575
2588
  "gpt-5-nano-2025-08-07": {
2576
2589
  maxTokens: 128e3,
@@ -2583,7 +2596,8 @@ var openAiNativeModels = {
2583
2596
  outputPrice: 0.4,
2584
2597
  cacheReadsPrice: 0.01,
2585
2598
  description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
2586
- supportsVerbosity: true
2599
+ supportsVerbosity: true,
2600
+ supportsTemperature: false
2587
2601
  },
2588
2602
  "gpt-4.1": {
2589
2603
  maxTokens: 32768,
@@ -2592,7 +2606,8 @@ var openAiNativeModels = {
2592
2606
  supportsPromptCache: true,
2593
2607
  inputPrice: 2,
2594
2608
  outputPrice: 8,
2595
- cacheReadsPrice: 0.5
2609
+ cacheReadsPrice: 0.5,
2610
+ supportsTemperature: true
2596
2611
  },
2597
2612
  "gpt-4.1-mini": {
2598
2613
  maxTokens: 32768,
@@ -2601,7 +2616,8 @@ var openAiNativeModels = {
2601
2616
  supportsPromptCache: true,
2602
2617
  inputPrice: 0.4,
2603
2618
  outputPrice: 1.6,
2604
- cacheReadsPrice: 0.1
2619
+ cacheReadsPrice: 0.1,
2620
+ supportsTemperature: true
2605
2621
  },
2606
2622
  "gpt-4.1-nano": {
2607
2623
  maxTokens: 32768,
@@ -2610,7 +2626,8 @@ var openAiNativeModels = {
2610
2626
  supportsPromptCache: true,
2611
2627
  inputPrice: 0.1,
2612
2628
  outputPrice: 0.4,
2613
- cacheReadsPrice: 0.025
2629
+ cacheReadsPrice: 0.025,
2630
+ supportsTemperature: true
2614
2631
  },
2615
2632
  o3: {
2616
2633
  maxTokens: 1e5,
@@ -2621,7 +2638,8 @@ var openAiNativeModels = {
2621
2638
  outputPrice: 8,
2622
2639
  cacheReadsPrice: 0.5,
2623
2640
  supportsReasoningEffort: true,
2624
- reasoningEffort: "medium"
2641
+ reasoningEffort: "medium",
2642
+ supportsTemperature: false
2625
2643
  },
2626
2644
  "o3-high": {
2627
2645
  maxTokens: 1e5,
@@ -2631,7 +2649,8 @@ var openAiNativeModels = {
2631
2649
  inputPrice: 2,
2632
2650
  outputPrice: 8,
2633
2651
  cacheReadsPrice: 0.5,
2634
- reasoningEffort: "high"
2652
+ reasoningEffort: "high",
2653
+ supportsTemperature: false
2635
2654
  },
2636
2655
  "o3-low": {
2637
2656
  maxTokens: 1e5,
@@ -2641,7 +2660,8 @@ var openAiNativeModels = {
2641
2660
  inputPrice: 2,
2642
2661
  outputPrice: 8,
2643
2662
  cacheReadsPrice: 0.5,
2644
- reasoningEffort: "low"
2663
+ reasoningEffort: "low",
2664
+ supportsTemperature: false
2645
2665
  },
2646
2666
  "o4-mini": {
2647
2667
  maxTokens: 1e5,
@@ -2652,7 +2672,8 @@ var openAiNativeModels = {
2652
2672
  outputPrice: 4.4,
2653
2673
  cacheReadsPrice: 0.275,
2654
2674
  supportsReasoningEffort: true,
2655
- reasoningEffort: "medium"
2675
+ reasoningEffort: "medium",
2676
+ supportsTemperature: false
2656
2677
  },
2657
2678
  "o4-mini-high": {
2658
2679
  maxTokens: 1e5,
@@ -2662,7 +2683,8 @@ var openAiNativeModels = {
2662
2683
  inputPrice: 1.1,
2663
2684
  outputPrice: 4.4,
2664
2685
  cacheReadsPrice: 0.275,
2665
- reasoningEffort: "high"
2686
+ reasoningEffort: "high",
2687
+ supportsTemperature: false
2666
2688
  },
2667
2689
  "o4-mini-low": {
2668
2690
  maxTokens: 1e5,
@@ -2672,7 +2694,8 @@ var openAiNativeModels = {
2672
2694
  inputPrice: 1.1,
2673
2695
  outputPrice: 4.4,
2674
2696
  cacheReadsPrice: 0.275,
2675
- reasoningEffort: "low"
2697
+ reasoningEffort: "low",
2698
+ supportsTemperature: false
2676
2699
  },
2677
2700
  "o3-mini": {
2678
2701
  maxTokens: 1e5,
@@ -2683,7 +2706,8 @@ var openAiNativeModels = {
2683
2706
  outputPrice: 4.4,
2684
2707
  cacheReadsPrice: 0.55,
2685
2708
  supportsReasoningEffort: true,
2686
- reasoningEffort: "medium"
2709
+ reasoningEffort: "medium",
2710
+ supportsTemperature: false
2687
2711
  },
2688
2712
  "o3-mini-high": {
2689
2713
  maxTokens: 1e5,
@@ -2693,7 +2717,8 @@ var openAiNativeModels = {
2693
2717
  inputPrice: 1.1,
2694
2718
  outputPrice: 4.4,
2695
2719
  cacheReadsPrice: 0.55,
2696
- reasoningEffort: "high"
2720
+ reasoningEffort: "high",
2721
+ supportsTemperature: false
2697
2722
  },
2698
2723
  "o3-mini-low": {
2699
2724
  maxTokens: 1e5,
@@ -2703,7 +2728,8 @@ var openAiNativeModels = {
2703
2728
  inputPrice: 1.1,
2704
2729
  outputPrice: 4.4,
2705
2730
  cacheReadsPrice: 0.55,
2706
- reasoningEffort: "low"
2731
+ reasoningEffort: "low",
2732
+ supportsTemperature: false
2707
2733
  },
2708
2734
  o1: {
2709
2735
  maxTokens: 1e5,
@@ -2712,7 +2738,8 @@ var openAiNativeModels = {
2712
2738
  supportsPromptCache: true,
2713
2739
  inputPrice: 15,
2714
2740
  outputPrice: 60,
2715
- cacheReadsPrice: 7.5
2741
+ cacheReadsPrice: 7.5,
2742
+ supportsTemperature: false
2716
2743
  },
2717
2744
  "o1-preview": {
2718
2745
  maxTokens: 32768,
@@ -2721,7 +2748,8 @@ var openAiNativeModels = {
2721
2748
  supportsPromptCache: true,
2722
2749
  inputPrice: 15,
2723
2750
  outputPrice: 60,
2724
- cacheReadsPrice: 7.5
2751
+ cacheReadsPrice: 7.5,
2752
+ supportsTemperature: false
2725
2753
  },
2726
2754
  "o1-mini": {
2727
2755
  maxTokens: 65536,
@@ -2730,7 +2758,8 @@ var openAiNativeModels = {
2730
2758
  supportsPromptCache: true,
2731
2759
  inputPrice: 1.1,
2732
2760
  outputPrice: 4.4,
2733
- cacheReadsPrice: 0.55
2761
+ cacheReadsPrice: 0.55,
2762
+ supportsTemperature: false
2734
2763
  },
2735
2764
  "gpt-4o": {
2736
2765
  maxTokens: 16384,
@@ -2739,7 +2768,8 @@ var openAiNativeModels = {
2739
2768
  supportsPromptCache: true,
2740
2769
  inputPrice: 2.5,
2741
2770
  outputPrice: 10,
2742
- cacheReadsPrice: 1.25
2771
+ cacheReadsPrice: 1.25,
2772
+ supportsTemperature: true
2743
2773
  },
2744
2774
  "gpt-4o-mini": {
2745
2775
  maxTokens: 16384,
@@ -2748,7 +2778,8 @@ var openAiNativeModels = {
2748
2778
  supportsPromptCache: true,
2749
2779
  inputPrice: 0.15,
2750
2780
  outputPrice: 0.6,
2751
- cacheReadsPrice: 0.075
2781
+ cacheReadsPrice: 0.075,
2782
+ supportsTemperature: true
2752
2783
  },
2753
2784
  "codex-mini-latest": {
2754
2785
  maxTokens: 16384,
@@ -2758,6 +2789,7 @@ var openAiNativeModels = {
2758
2789
  inputPrice: 1.5,
2759
2790
  outputPrice: 6,
2760
2791
  cacheReadsPrice: 0,
2792
+ supportsTemperature: false,
2761
2793
  description: "Codex Mini: Cloud-based software engineering agent powered by codex-1, a version of o3 optimized for coding tasks. Trained with reinforcement learning to generate human-style code, adhere to instructions, and iteratively run tests."
2762
2794
  }
2763
2795
  };
@@ -3831,7 +3863,8 @@ var providerNamesSchema = import_zod8.z.enum(providerNames);
3831
3863
  var providerSettingsEntrySchema = import_zod8.z.object({
3832
3864
  id: import_zod8.z.string(),
3833
3865
  name: import_zod8.z.string(),
3834
- apiProvider: providerNamesSchema.optional()
3866
+ apiProvider: providerNamesSchema.optional(),
3867
+ modelId: import_zod8.z.string().optional()
3835
3868
  });
3836
3869
  var DEFAULT_CONSECUTIVE_MISTAKE_LIMIT = 3;
3837
3870
  var baseProviderSettingsSchema = import_zod8.z.object({
@@ -3873,7 +3906,12 @@ var openRouterSchema = baseProviderSettingsSchema.extend({
3873
3906
  openRouterModelId: import_zod8.z.string().optional(),
3874
3907
  openRouterBaseUrl: import_zod8.z.string().optional(),
3875
3908
  openRouterSpecificProvider: import_zod8.z.string().optional(),
3876
- openRouterUseMiddleOutTransform: import_zod8.z.boolean().optional()
3909
+ openRouterUseMiddleOutTransform: import_zod8.z.boolean().optional(),
3910
+ // Image generation settings (experimental)
3911
+ openRouterImageGenerationSettings: import_zod8.z.object({
3912
+ openRouterApiKey: import_zod8.z.string().optional(),
3913
+ selectedModel: import_zod8.z.string().optional()
3914
+ }).optional()
3877
3915
  });
3878
3916
  var bedrockSchema = apiModelIdProviderModelSchema.extend({
3879
3917
  awsAccessKey: import_zod8.z.string().optional(),
@@ -4138,6 +4176,9 @@ var getApiProtocol = (provider, modelId) => {
4138
4176
  if (provider && provider === "vertex" && modelId && modelId.toLowerCase().includes("claude")) {
4139
4177
  return "anthropic";
4140
4178
  }
4179
+ if (provider && provider === "vercel-ai-gateway" && modelId && modelId.toLowerCase().startsWith("anthropic/")) {
4180
+ return "anthropic";
4181
+ }
4141
4182
  return "openai";
4142
4183
  };
4143
4184
  var MODELS_BY_PROVIDER = {
@@ -4706,6 +4747,7 @@ var SECRET_STATE_KEYS = [
4706
4747
  "codebaseIndexOpenAiCompatibleApiKey",
4707
4748
  "codebaseIndexGeminiApiKey",
4708
4749
  "codebaseIndexMistralApiKey",
4750
+ "codebaseIndexVercelAiGatewayApiKey",
4709
4751
  "huggingFaceApiKey",
4710
4752
  "sambaNovaApiKey",
4711
4753
  "zaiApiKey",