@roo-code/types 1.27.0 → 1.29.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -737,6 +737,36 @@ var chutesModels = {
737
737
  }
738
738
  };
739
739
 
740
+ // src/providers/claude-code.ts
741
+ var claudeCodeDefaultModelId = "claude-sonnet-4-20250514";
742
+ var claudeCodeModels = {
743
+ "claude-sonnet-4-20250514": {
744
+ ...anthropicModels["claude-sonnet-4-20250514"],
745
+ supportsImages: false,
746
+ supportsPromptCache: false
747
+ },
748
+ "claude-opus-4-20250514": {
749
+ ...anthropicModels["claude-opus-4-20250514"],
750
+ supportsImages: false,
751
+ supportsPromptCache: false
752
+ },
753
+ "claude-3-7-sonnet-20250219": {
754
+ ...anthropicModels["claude-3-7-sonnet-20250219"],
755
+ supportsImages: false,
756
+ supportsPromptCache: false
757
+ },
758
+ "claude-3-5-sonnet-20241022": {
759
+ ...anthropicModels["claude-3-5-sonnet-20241022"],
760
+ supportsImages: false,
761
+ supportsPromptCache: false
762
+ },
763
+ "claude-3-5-haiku-20241022": {
764
+ ...anthropicModels["claude-3-5-haiku-20241022"],
765
+ supportsImages: false,
766
+ supportsPromptCache: false
767
+ }
768
+ };
769
+
740
770
  // src/providers/deepseek.ts
741
771
  var deepSeekDefaultModelId = "deepseek-chat";
742
772
  var deepSeekModels = {
@@ -818,6 +848,18 @@ var geminiModels = {
818
848
  cacheReadsPrice: 0.0375,
819
849
  cacheWritesPrice: 1
820
850
  },
851
+ "gemini-2.5-flash": {
852
+ maxTokens: 64e3,
853
+ contextWindow: 1048576,
854
+ supportsImages: true,
855
+ supportsPromptCache: true,
856
+ inputPrice: 0.3,
857
+ outputPrice: 2.5,
858
+ cacheReadsPrice: 0.075,
859
+ cacheWritesPrice: 1,
860
+ maxThinkingTokens: 24576,
861
+ supportsReasoningBudget: true
862
+ },
821
863
  "gemini-2.5-pro-exp-03-25": {
822
864
  maxTokens: 65535,
823
865
  contextWindow: 1048576,
@@ -903,6 +945,34 @@ var geminiModels = {
903
945
  }
904
946
  ]
905
947
  },
948
+ "gemini-2.5-pro": {
949
+ maxTokens: 64e3,
950
+ contextWindow: 1048576,
951
+ supportsImages: true,
952
+ supportsPromptCache: true,
953
+ inputPrice: 2.5,
954
+ // This is the pricing for prompts above 200k tokens.
955
+ outputPrice: 15,
956
+ cacheReadsPrice: 0.625,
957
+ cacheWritesPrice: 4.5,
958
+ maxThinkingTokens: 32768,
959
+ supportsReasoningBudget: true,
960
+ requiredReasoningBudget: true,
961
+ tiers: [
962
+ {
963
+ contextWindow: 2e5,
964
+ inputPrice: 1.25,
965
+ outputPrice: 10,
966
+ cacheReadsPrice: 0.31
967
+ },
968
+ {
969
+ contextWindow: Infinity,
970
+ inputPrice: 2.5,
971
+ outputPrice: 15,
972
+ cacheReadsPrice: 0.625
973
+ }
974
+ ]
975
+ },
906
976
  "gemini-2.0-flash-001": {
907
977
  maxTokens: 8192,
908
978
  contextWindow: 1048576,
@@ -1017,6 +1087,18 @@ var geminiModels = {
1017
1087
  supportsPromptCache: false,
1018
1088
  inputPrice: 0,
1019
1089
  outputPrice: 0
1090
+ },
1091
+ "gemini-2.5-flash-lite-preview-06-17": {
1092
+ maxTokens: 64e3,
1093
+ contextWindow: 1048576,
1094
+ supportsImages: true,
1095
+ supportsPromptCache: true,
1096
+ inputPrice: 0.1,
1097
+ outputPrice: 0.4,
1098
+ cacheReadsPrice: 0.025,
1099
+ cacheWritesPrice: 1,
1100
+ supportsReasoningBudget: true,
1101
+ maxThinkingTokens: 24576
1020
1102
  }
1021
1103
  };
1022
1104
 
@@ -1045,8 +1127,8 @@ var groqModels = {
1045
1127
  contextWindow: 131072,
1046
1128
  supportsImages: false,
1047
1129
  supportsPromptCache: false,
1048
- inputPrice: 0,
1049
- outputPrice: 0,
1130
+ inputPrice: 0.05,
1131
+ outputPrice: 0.08,
1050
1132
  description: "Meta Llama 3.1 8B Instant model, 128K context."
1051
1133
  },
1052
1134
  "llama-3.3-70b-versatile": {
@@ -1054,8 +1136,8 @@ var groqModels = {
1054
1136
  contextWindow: 131072,
1055
1137
  supportsImages: false,
1056
1138
  supportsPromptCache: false,
1057
- inputPrice: 0,
1058
- outputPrice: 0,
1139
+ inputPrice: 0.59,
1140
+ outputPrice: 0.79,
1059
1141
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
1060
1142
  },
1061
1143
  "meta-llama/llama-4-scout-17b-16e-instruct": {
@@ -1063,8 +1145,8 @@ var groqModels = {
1063
1145
  contextWindow: 131072,
1064
1146
  supportsImages: false,
1065
1147
  supportsPromptCache: false,
1066
- inputPrice: 0,
1067
- outputPrice: 0,
1148
+ inputPrice: 0.11,
1149
+ outputPrice: 0.34,
1068
1150
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
1069
1151
  },
1070
1152
  "meta-llama/llama-4-maverick-17b-128e-instruct": {
@@ -1072,8 +1154,8 @@ var groqModels = {
1072
1154
  contextWindow: 131072,
1073
1155
  supportsImages: false,
1074
1156
  supportsPromptCache: false,
1075
- inputPrice: 0,
1076
- outputPrice: 0,
1157
+ inputPrice: 0.2,
1158
+ outputPrice: 0.6,
1077
1159
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
1078
1160
  },
1079
1161
  "mistral-saba-24b": {
@@ -1081,8 +1163,8 @@ var groqModels = {
1081
1163
  contextWindow: 32768,
1082
1164
  supportsImages: false,
1083
1165
  supportsPromptCache: false,
1084
- inputPrice: 0,
1085
- outputPrice: 0,
1166
+ inputPrice: 0.79,
1167
+ outputPrice: 0.79,
1086
1168
  description: "Mistral Saba 24B model, 32K context."
1087
1169
  },
1088
1170
  "qwen-qwq-32b": {
@@ -1090,17 +1172,26 @@ var groqModels = {
1090
1172
  contextWindow: 131072,
1091
1173
  supportsImages: false,
1092
1174
  supportsPromptCache: false,
1093
- inputPrice: 0,
1094
- outputPrice: 0,
1175
+ inputPrice: 0.29,
1176
+ outputPrice: 0.39,
1095
1177
  description: "Alibaba Qwen QwQ 32B model, 128K context."
1096
1178
  },
1179
+ "qwen/qwen3-32b": {
1180
+ maxTokens: 131072,
1181
+ contextWindow: 131072,
1182
+ supportsImages: false,
1183
+ supportsPromptCache: false,
1184
+ inputPrice: 0.29,
1185
+ outputPrice: 0.59,
1186
+ description: "Alibaba Qwen 3 32B model, 128K context."
1187
+ },
1097
1188
  "deepseek-r1-distill-llama-70b": {
1098
1189
  maxTokens: 131072,
1099
1190
  contextWindow: 131072,
1100
1191
  supportsImages: false,
1101
1192
  supportsPromptCache: false,
1102
- inputPrice: 0,
1103
- outputPrice: 0,
1193
+ inputPrice: 0.75,
1194
+ outputPrice: 0.99,
1104
1195
  description: "DeepSeek R1 Distill Llama 70B model, 128K context."
1105
1196
  }
1106
1197
  };
@@ -1152,10 +1243,39 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1152
1243
 
1153
1244
  // src/providers/lm-studio.ts
1154
1245
  var LMSTUDIO_DEFAULT_TEMPERATURE = 0;
1246
+ var lMStudioDefaultModelId = "mistralai/devstral-small-2505";
1247
+ var lMStudioDefaultModelInfo = {
1248
+ maxTokens: 8192,
1249
+ contextWindow: 2e5,
1250
+ supportsImages: true,
1251
+ supportsComputerUse: true,
1252
+ supportsPromptCache: true,
1253
+ inputPrice: 0,
1254
+ outputPrice: 0,
1255
+ cacheWritesPrice: 0,
1256
+ cacheReadsPrice: 0,
1257
+ description: "LM Studio hosted models"
1258
+ };
1155
1259
 
1156
1260
  // src/providers/mistral.ts
1157
1261
  var mistralDefaultModelId = "codestral-latest";
1158
1262
  var mistralModels = {
1263
+ "magistral-medium-latest": {
1264
+ maxTokens: 41e3,
1265
+ contextWindow: 41e3,
1266
+ supportsImages: false,
1267
+ supportsPromptCache: false,
1268
+ inputPrice: 2,
1269
+ outputPrice: 5
1270
+ },
1271
+ "mistral-medium-latest": {
1272
+ maxTokens: 131e3,
1273
+ contextWindow: 131e3,
1274
+ supportsImages: true,
1275
+ supportsPromptCache: false,
1276
+ inputPrice: 0.4,
1277
+ outputPrice: 2
1278
+ },
1159
1279
  "codestral-latest": {
1160
1280
  maxTokens: 256e3,
1161
1281
  contextWindow: 256e3,
@@ -1207,6 +1327,21 @@ var mistralModels = {
1207
1327
  };
1208
1328
  var MISTRAL_DEFAULT_TEMPERATURE = 0;
1209
1329
 
1330
+ // src/providers/ollama.ts
1331
+ var ollamaDefaultModelId = "devstral:24b";
1332
+ var ollamaDefaultModelInfo = {
1333
+ maxTokens: 4096,
1334
+ contextWindow: 2e5,
1335
+ supportsImages: true,
1336
+ supportsComputerUse: true,
1337
+ supportsPromptCache: true,
1338
+ inputPrice: 0,
1339
+ outputPrice: 0,
1340
+ cacheWritesPrice: 0,
1341
+ cacheReadsPrice: 0,
1342
+ description: "Ollama hosted models"
1343
+ };
1344
+
1210
1345
  // src/providers/openai.ts
1211
1346
  var openAiNativeDefaultModelId = "gpt-4.1";
1212
1347
  var openAiNativeModels = {
@@ -1436,6 +1571,8 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
1436
1571
  "google/gemini-2.5-flash-preview:thinking",
1437
1572
  "google/gemini-2.5-flash-preview-05-20",
1438
1573
  "google/gemini-2.5-flash-preview-05-20:thinking",
1574
+ "google/gemini-2.5-flash",
1575
+ "google/gemini-2.5-flash-lite-preview-06-17",
1439
1576
  "google/gemini-2.0-flash-001",
1440
1577
  "google/gemini-flash-1.5",
1441
1578
  "google/gemini-flash-1.5-8b"
@@ -1451,6 +1588,7 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1451
1588
  ]);
1452
1589
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1453
1590
  "anthropic/claude-3.7-sonnet:thinking",
1591
+ "google/gemini-2.5-pro",
1454
1592
  "google/gemini-2.5-flash-preview-05-20:thinking"
1455
1593
  ]);
1456
1594
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
@@ -1458,7 +1596,10 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1458
1596
  "anthropic/claude-opus-4",
1459
1597
  "anthropic/claude-sonnet-4",
1460
1598
  "google/gemini-2.5-pro-preview",
1599
+ "google/gemini-2.5-pro",
1461
1600
  "google/gemini-2.5-flash-preview-05-20",
1601
+ "google/gemini-2.5-flash",
1602
+ "google/gemini-2.5-flash-lite-preview-06-17",
1462
1603
  // Also include the models that require the reasoning budget to be enabled
1463
1604
  // even though `OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS` takes precedence.
1464
1605
  "anthropic/claude-3.7-sonnet:thinking",
@@ -1515,6 +1656,18 @@ var vertexModels = {
1515
1656
  inputPrice: 0.15,
1516
1657
  outputPrice: 0.6
1517
1658
  },
1659
+ "gemini-2.5-flash": {
1660
+ maxTokens: 64e3,
1661
+ contextWindow: 1048576,
1662
+ supportsImages: true,
1663
+ supportsPromptCache: true,
1664
+ inputPrice: 0.3,
1665
+ outputPrice: 2.5,
1666
+ cacheReadsPrice: 0.075,
1667
+ cacheWritesPrice: 1,
1668
+ maxThinkingTokens: 24576,
1669
+ supportsReasoningBudget: true
1670
+ },
1518
1671
  "gemini-2.5-flash-preview-04-17:thinking": {
1519
1672
  maxTokens: 65535,
1520
1673
  contextWindow: 1048576,
@@ -1560,6 +1713,31 @@ var vertexModels = {
1560
1713
  maxThinkingTokens: 32768,
1561
1714
  supportsReasoningBudget: true
1562
1715
  },
1716
+ "gemini-2.5-pro": {
1717
+ maxTokens: 64e3,
1718
+ contextWindow: 1048576,
1719
+ supportsImages: true,
1720
+ supportsPromptCache: true,
1721
+ inputPrice: 2.5,
1722
+ outputPrice: 15,
1723
+ maxThinkingTokens: 32768,
1724
+ supportsReasoningBudget: true,
1725
+ requiredReasoningBudget: true,
1726
+ tiers: [
1727
+ {
1728
+ contextWindow: 2e5,
1729
+ inputPrice: 1.25,
1730
+ outputPrice: 10,
1731
+ cacheReadsPrice: 0.31
1732
+ },
1733
+ {
1734
+ contextWindow: Infinity,
1735
+ inputPrice: 2.5,
1736
+ outputPrice: 15,
1737
+ cacheReadsPrice: 0.625
1738
+ }
1739
+ ]
1740
+ },
1563
1741
  "gemini-2.5-pro-exp-03-25": {
1564
1742
  maxTokens: 65535,
1565
1743
  contextWindow: 1048576,
@@ -1713,6 +1891,18 @@ var vertexModels = {
1713
1891
  outputPrice: 1.25,
1714
1892
  cacheWritesPrice: 0.3,
1715
1893
  cacheReadsPrice: 0.03
1894
+ },
1895
+ "gemini-2.5-flash-lite-preview-06-17": {
1896
+ maxTokens: 64e3,
1897
+ contextWindow: 1048576,
1898
+ supportsImages: true,
1899
+ supportsPromptCache: true,
1900
+ inputPrice: 0.1,
1901
+ outputPrice: 0.4,
1902
+ cacheReadsPrice: 0.025,
1903
+ cacheWritesPrice: 1,
1904
+ maxThinkingTokens: 24576,
1905
+ supportsReasoningBudget: true
1716
1906
  }
1717
1907
  };
1718
1908
  var VERTEX_REGIONS = [
@@ -2014,6 +2204,7 @@ var modelInfoSchema = z2.object({
2014
2204
  // src/provider-settings.ts
2015
2205
  var providerNames = [
2016
2206
  "anthropic",
2207
+ "claude-code",
2017
2208
  "glama",
2018
2209
  "openrouter",
2019
2210
  "bedrock",
@@ -2061,6 +2252,9 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
2061
2252
  anthropicBaseUrl: z3.string().optional(),
2062
2253
  anthropicUseAuthToken: z3.boolean().optional()
2063
2254
  });
2255
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
2256
+ claudeCodePath: z3.string().optional()
2257
+ });
2064
2258
  var glamaSchema = baseProviderSettingsSchema.extend({
2065
2259
  glamaModelId: z3.string().optional(),
2066
2260
  glamaApiKey: z3.string().optional()
@@ -2171,6 +2365,7 @@ var defaultSchema = z3.object({
2171
2365
  });
2172
2366
  var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
2173
2367
  anthropicSchema.merge(z3.object({ apiProvider: z3.literal("anthropic") })),
2368
+ claudeCodeSchema.merge(z3.object({ apiProvider: z3.literal("claude-code") })),
2174
2369
  glamaSchema.merge(z3.object({ apiProvider: z3.literal("glama") })),
2175
2370
  openRouterSchema.merge(z3.object({ apiProvider: z3.literal("openrouter") })),
2176
2371
  bedrockSchema.merge(z3.object({ apiProvider: z3.literal("bedrock") })),
@@ -2196,6 +2391,7 @@ var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
2196
2391
  var providerSettingsSchema = z3.object({
2197
2392
  apiProvider: providerNamesSchema.optional(),
2198
2393
  ...anthropicSchema.shape,
2394
+ ...claudeCodeSchema.shape,
2199
2395
  ...glamaSchema.shape,
2200
2396
  ...openRouterSchema.shape,
2201
2397
  ...bedrockSchema.shape,
@@ -2254,12 +2450,10 @@ var historyItemSchema = z4.object({
2254
2450
 
2255
2451
  // src/experiment.ts
2256
2452
  import { z as z5 } from "zod";
2257
- var experimentIds = ["powerSteering", "disableCompletionCommand", "marketplace", "multiFileApplyDiff"];
2453
+ var experimentIds = ["powerSteering", "multiFileApplyDiff"];
2258
2454
  var experimentIdsSchema = z5.enum(experimentIds);
2259
2455
  var experimentsSchema = z5.object({
2260
2456
  powerSteering: z5.boolean().optional(),
2261
- disableCompletionCommand: z5.boolean().optional(),
2262
- marketplace: z5.boolean().optional(),
2263
2457
  multiFileApplyDiff: z5.boolean().optional()
2264
2458
  });
2265
2459
 
@@ -2359,6 +2553,9 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
2359
2553
  TelemetryEventName2["CHECKPOINT_CREATED"] = "Checkpoint Created";
2360
2554
  TelemetryEventName2["CHECKPOINT_RESTORED"] = "Checkpoint Restored";
2361
2555
  TelemetryEventName2["CHECKPOINT_DIFFED"] = "Checkpoint Diffed";
2556
+ TelemetryEventName2["TAB_SHOWN"] = "Tab Shown";
2557
+ TelemetryEventName2["MODE_SETTINGS_CHANGED"] = "Mode Setting Changed";
2558
+ TelemetryEventName2["CUSTOM_MODE_CREATED"] = "Custom Mode Created";
2362
2559
  TelemetryEventName2["CONTEXT_CONDENSED"] = "Context Condensed";
2363
2560
  TelemetryEventName2["SLIDING_WINDOW_TRUNCATION"] = "Sliding Window Truncation";
2364
2561
  TelemetryEventName2["CODE_ACTION_USED"] = "Code Action Used";
@@ -2391,9 +2588,18 @@ var taskPropertiesSchema = z7.object({
2391
2588
  diffStrategy: z7.string().optional(),
2392
2589
  isSubtask: z7.boolean().optional()
2393
2590
  });
2591
+ var gitPropertiesSchema = z7.object({
2592
+ repositoryUrl: z7.string().optional(),
2593
+ repositoryName: z7.string().optional(),
2594
+ defaultBranch: z7.string().optional()
2595
+ });
2394
2596
  var telemetryPropertiesSchema = z7.object({
2395
2597
  ...appPropertiesSchema.shape,
2396
- ...taskPropertiesSchema.shape
2598
+ ...taskPropertiesSchema.shape,
2599
+ ...gitPropertiesSchema.shape
2600
+ });
2601
+ var cloudTelemetryPropertiesSchema = z7.object({
2602
+ ...telemetryPropertiesSchema.shape
2397
2603
  });
2398
2604
  var rooCodeTelemetryEventSchema = z7.discriminatedUnion("type", [
2399
2605
  z7.object({
@@ -2418,14 +2624,17 @@ var rooCodeTelemetryEventSchema = z7.discriminatedUnion("type", [
2418
2624
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
2419
2625
  "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */,
2420
2626
  "Context Condensed" /* CONTEXT_CONDENSED */,
2421
- "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */
2627
+ "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
2628
+ "Tab Shown" /* TAB_SHOWN */,
2629
+ "Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
2630
+ "Custom Mode Created" /* CUSTOM_MODE_CREATED */
2422
2631
  ]),
2423
- properties: telemetryPropertiesSchema
2632
+ properties: cloudTelemetryPropertiesSchema
2424
2633
  }),
2425
2634
  z7.object({
2426
2635
  type: z7.literal("Task Message" /* TASK_MESSAGE */),
2427
2636
  properties: z7.object({
2428
- ...telemetryPropertiesSchema.shape,
2637
+ ...cloudTelemetryPropertiesSchema.shape,
2429
2638
  taskId: z7.string(),
2430
2639
  message: clineMessageSchema
2431
2640
  })
@@ -2433,7 +2642,7 @@ var rooCodeTelemetryEventSchema = z7.discriminatedUnion("type", [
2433
2642
  z7.object({
2434
2643
  type: z7.literal("LLM Completion" /* LLM_COMPLETION */),
2435
2644
  properties: z7.object({
2436
- ...telemetryPropertiesSchema.shape,
2645
+ ...cloudTelemetryPropertiesSchema.shape,
2437
2646
  inputTokens: z7.number(),
2438
2647
  outputTokens: z7.number(),
2439
2648
  cacheReadTokens: z7.number().optional(),
@@ -2517,6 +2726,7 @@ var modeConfigSchema = z9.object({
2517
2726
  name: z9.string().min(1, "Name is required"),
2518
2727
  roleDefinition: z9.string().min(1, "Role definition is required"),
2519
2728
  whenToUse: z9.string().optional(),
2729
+ description: z9.string().optional(),
2520
2730
  customInstructions: z9.string().optional(),
2521
2731
  groups: groupEntryArraySchema,
2522
2732
  source: z9.enum(["global", "project"]).optional()
@@ -2541,6 +2751,7 @@ var customModesSettingsSchema = z9.object({
2541
2751
  var promptComponentSchema = z9.object({
2542
2752
  roleDefinition: z9.string().optional(),
2543
2753
  whenToUse: z9.string().optional(),
2754
+ description: z9.string().optional(),
2544
2755
  customInstructions: z9.string().optional()
2545
2756
  });
2546
2757
  var customModePromptsSchema = z9.record(z9.string(), promptComponentSchema.optional());
@@ -2568,7 +2779,8 @@ var commandIds = [
2568
2779
  "newTask",
2569
2780
  "setCustomStoragePath",
2570
2781
  "focusInput",
2571
- "acceptInput"
2782
+ "acceptInput",
2783
+ "focusPanel"
2572
2784
  ];
2573
2785
  var languages = [
2574
2786
  "ca",
@@ -2663,7 +2875,9 @@ var globalSettingsSchema = z11.object({
2663
2875
  customModePrompts: customModePromptsSchema.optional(),
2664
2876
  customSupportPrompts: customSupportPromptsSchema.optional(),
2665
2877
  enhancementApiConfigId: z11.string().optional(),
2666
- historyPreviewCollapsed: z11.boolean().optional()
2878
+ historyPreviewCollapsed: z11.boolean().optional(),
2879
+ profileThresholds: z11.record(z11.string(), z11.number()).optional(),
2880
+ hasOpenedModeSelector: z11.boolean().optional()
2667
2881
  });
2668
2882
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
2669
2883
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -2800,6 +3014,7 @@ var ORGANIZATION_ALLOW_ALL = {
2800
3014
  var ORGANIZATION_DEFAULT = {
2801
3015
  version: 0,
2802
3016
  cloudSettings: {
3017
+ recordTaskMessages: true,
2803
3018
  enableTaskSharing: true,
2804
3019
  taskShareExpirationDays: 30
2805
3020
  },
@@ -2809,11 +3024,16 @@ var ORGANIZATION_DEFAULT = {
2809
3024
  var shareResponseSchema = z12.object({
2810
3025
  success: z12.boolean(),
2811
3026
  shareUrl: z12.string().optional(),
2812
- error: z12.string().optional()
3027
+ error: z12.string().optional(),
3028
+ isNewShare: z12.boolean().optional(),
3029
+ manageUrl: z12.string().optional()
2813
3030
  });
2814
3031
 
2815
3032
  // src/ipc.ts
2816
3033
  import { z as z13 } from "zod";
3034
+ var isSubtaskSchema = z13.object({
3035
+ isSubtask: z13.boolean()
3036
+ });
2817
3037
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
2818
3038
  RooCodeEventName2["Message"] = "message";
2819
3039
  RooCodeEventName2["TaskCreated"] = "taskCreated";
@@ -2847,7 +3067,7 @@ var rooCodeEventsSchema = z13.object({
2847
3067
  ["taskAskResponded" /* TaskAskResponded */]: z13.tuple([z13.string()]),
2848
3068
  ["taskAborted" /* TaskAborted */]: z13.tuple([z13.string()]),
2849
3069
  ["taskSpawned" /* TaskSpawned */]: z13.tuple([z13.string(), z13.string()]),
2850
- ["taskCompleted" /* TaskCompleted */]: z13.tuple([z13.string(), tokenUsageSchema, toolUsageSchema]),
3070
+ ["taskCompleted" /* TaskCompleted */]: z13.tuple([z13.string(), tokenUsageSchema, toolUsageSchema, isSubtaskSchema]),
2851
3071
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z13.tuple([z13.string(), tokenUsageSchema]),
2852
3072
  ["taskToolFailed" /* TaskToolFailed */]: z13.tuple([z13.string(), toolNamesSchema, z13.string()])
2853
3073
  });
@@ -3130,11 +3350,14 @@ export {
3130
3350
  bedrockModels,
3131
3351
  chutesDefaultModelId,
3132
3352
  chutesModels,
3353
+ claudeCodeDefaultModelId,
3354
+ claudeCodeModels,
3133
3355
  clineAskSchema,
3134
3356
  clineAsks,
3135
3357
  clineMessageSchema,
3136
3358
  clineSaySchema,
3137
3359
  clineSays,
3360
+ cloudTelemetryPropertiesSchema,
3138
3361
  codeActionIds,
3139
3362
  codebaseIndexConfigSchema,
3140
3363
  codebaseIndexModelsSchema,
@@ -3153,6 +3376,7 @@ export {
3153
3376
  geminiDefaultModelId,
3154
3377
  geminiModels,
3155
3378
  getModelId,
3379
+ gitPropertiesSchema,
3156
3380
  glamaDefaultModelId,
3157
3381
  glamaDefaultModelInfo,
3158
3382
  globalSettingsSchema,
@@ -3167,6 +3391,9 @@ export {
3167
3391
  isLanguage,
3168
3392
  isModelParameter,
3169
3393
  isSecretStateKey,
3394
+ isSubtaskSchema,
3395
+ lMStudioDefaultModelId,
3396
+ lMStudioDefaultModelInfo,
3170
3397
  languages,
3171
3398
  languagesSchema,
3172
3399
  litellmDefaultModelId,
@@ -3184,6 +3411,8 @@ export {
3184
3411
  modelInfoSchema,
3185
3412
  modelParameters,
3186
3413
  modelParametersSchema,
3414
+ ollamaDefaultModelId,
3415
+ ollamaDefaultModelInfo,
3187
3416
  openAiModelInfoSaneDefaults,
3188
3417
  openAiNativeDefaultModelId,
3189
3418
  openAiNativeModels,