@roo-code/types 1.27.0 → 1.28.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -737,6 +737,16 @@ var chutesModels = {
737
737
  }
738
738
  };
739
739
 
740
+ // src/providers/claude-code.ts
741
+ var claudeCodeDefaultModelId = "claude-sonnet-4-20250514";
742
+ var claudeCodeModels = {
743
+ "claude-sonnet-4-20250514": anthropicModels["claude-sonnet-4-20250514"],
744
+ "claude-opus-4-20250514": anthropicModels["claude-opus-4-20250514"],
745
+ "claude-3-7-sonnet-20250219": anthropicModels["claude-3-7-sonnet-20250219"],
746
+ "claude-3-5-sonnet-20241022": anthropicModels["claude-3-5-sonnet-20241022"],
747
+ "claude-3-5-haiku-20241022": anthropicModels["claude-3-5-haiku-20241022"]
748
+ };
749
+
740
750
  // src/providers/deepseek.ts
741
751
  var deepSeekDefaultModelId = "deepseek-chat";
742
752
  var deepSeekModels = {
@@ -818,6 +828,18 @@ var geminiModels = {
818
828
  cacheReadsPrice: 0.0375,
819
829
  cacheWritesPrice: 1
820
830
  },
831
+ "gemini-2.5-flash": {
832
+ maxTokens: 64e3,
833
+ contextWindow: 1048576,
834
+ supportsImages: true,
835
+ supportsPromptCache: true,
836
+ inputPrice: 0.3,
837
+ outputPrice: 2.5,
838
+ cacheReadsPrice: 0.075,
839
+ cacheWritesPrice: 1,
840
+ maxThinkingTokens: 24576,
841
+ supportsReasoningBudget: true
842
+ },
821
843
  "gemini-2.5-pro-exp-03-25": {
822
844
  maxTokens: 65535,
823
845
  contextWindow: 1048576,
@@ -903,6 +925,34 @@ var geminiModels = {
903
925
  }
904
926
  ]
905
927
  },
928
+ "gemini-2.5-pro": {
929
+ maxTokens: 64e3,
930
+ contextWindow: 1048576,
931
+ supportsImages: true,
932
+ supportsPromptCache: true,
933
+ inputPrice: 2.5,
934
+ // This is the pricing for prompts above 200k tokens.
935
+ outputPrice: 15,
936
+ cacheReadsPrice: 0.625,
937
+ cacheWritesPrice: 4.5,
938
+ maxThinkingTokens: 32768,
939
+ supportsReasoningBudget: true,
940
+ requiredReasoningBudget: true,
941
+ tiers: [
942
+ {
943
+ contextWindow: 2e5,
944
+ inputPrice: 1.25,
945
+ outputPrice: 10,
946
+ cacheReadsPrice: 0.31
947
+ },
948
+ {
949
+ contextWindow: Infinity,
950
+ inputPrice: 2.5,
951
+ outputPrice: 15,
952
+ cacheReadsPrice: 0.625
953
+ }
954
+ ]
955
+ },
906
956
  "gemini-2.0-flash-001": {
907
957
  maxTokens: 8192,
908
958
  contextWindow: 1048576,
@@ -1017,6 +1067,18 @@ var geminiModels = {
1017
1067
  supportsPromptCache: false,
1018
1068
  inputPrice: 0,
1019
1069
  outputPrice: 0
1070
+ },
1071
+ "gemini-2.5-flash-lite-preview-06-17": {
1072
+ maxTokens: 64e3,
1073
+ contextWindow: 1048576,
1074
+ supportsImages: true,
1075
+ supportsPromptCache: true,
1076
+ inputPrice: 0.1,
1077
+ outputPrice: 0.4,
1078
+ cacheReadsPrice: 0.025,
1079
+ cacheWritesPrice: 1,
1080
+ supportsReasoningBudget: true,
1081
+ maxThinkingTokens: 24576
1020
1082
  }
1021
1083
  };
1022
1084
 
@@ -1045,8 +1107,8 @@ var groqModels = {
1045
1107
  contextWindow: 131072,
1046
1108
  supportsImages: false,
1047
1109
  supportsPromptCache: false,
1048
- inputPrice: 0,
1049
- outputPrice: 0,
1110
+ inputPrice: 0.05,
1111
+ outputPrice: 0.08,
1050
1112
  description: "Meta Llama 3.1 8B Instant model, 128K context."
1051
1113
  },
1052
1114
  "llama-3.3-70b-versatile": {
@@ -1054,8 +1116,8 @@ var groqModels = {
1054
1116
  contextWindow: 131072,
1055
1117
  supportsImages: false,
1056
1118
  supportsPromptCache: false,
1057
- inputPrice: 0,
1058
- outputPrice: 0,
1119
+ inputPrice: 0.59,
1120
+ outputPrice: 0.79,
1059
1121
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
1060
1122
  },
1061
1123
  "meta-llama/llama-4-scout-17b-16e-instruct": {
@@ -1063,8 +1125,8 @@ var groqModels = {
1063
1125
  contextWindow: 131072,
1064
1126
  supportsImages: false,
1065
1127
  supportsPromptCache: false,
1066
- inputPrice: 0,
1067
- outputPrice: 0,
1128
+ inputPrice: 0.11,
1129
+ outputPrice: 0.34,
1068
1130
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
1069
1131
  },
1070
1132
  "meta-llama/llama-4-maverick-17b-128e-instruct": {
@@ -1072,8 +1134,8 @@ var groqModels = {
1072
1134
  contextWindow: 131072,
1073
1135
  supportsImages: false,
1074
1136
  supportsPromptCache: false,
1075
- inputPrice: 0,
1076
- outputPrice: 0,
1137
+ inputPrice: 0.2,
1138
+ outputPrice: 0.6,
1077
1139
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
1078
1140
  },
1079
1141
  "mistral-saba-24b": {
@@ -1081,8 +1143,8 @@ var groqModels = {
1081
1143
  contextWindow: 32768,
1082
1144
  supportsImages: false,
1083
1145
  supportsPromptCache: false,
1084
- inputPrice: 0,
1085
- outputPrice: 0,
1146
+ inputPrice: 0.79,
1147
+ outputPrice: 0.79,
1086
1148
  description: "Mistral Saba 24B model, 32K context."
1087
1149
  },
1088
1150
  "qwen-qwq-32b": {
@@ -1090,17 +1152,26 @@ var groqModels = {
1090
1152
  contextWindow: 131072,
1091
1153
  supportsImages: false,
1092
1154
  supportsPromptCache: false,
1093
- inputPrice: 0,
1094
- outputPrice: 0,
1155
+ inputPrice: 0.29,
1156
+ outputPrice: 0.39,
1095
1157
  description: "Alibaba Qwen QwQ 32B model, 128K context."
1096
1158
  },
1159
+ "qwen/qwen3-32b": {
1160
+ maxTokens: 131072,
1161
+ contextWindow: 131072,
1162
+ supportsImages: false,
1163
+ supportsPromptCache: false,
1164
+ inputPrice: 0.29,
1165
+ outputPrice: 0.59,
1166
+ description: "Alibaba Qwen 3 32B model, 128K context."
1167
+ },
1097
1168
  "deepseek-r1-distill-llama-70b": {
1098
1169
  maxTokens: 131072,
1099
1170
  contextWindow: 131072,
1100
1171
  supportsImages: false,
1101
1172
  supportsPromptCache: false,
1102
- inputPrice: 0,
1103
- outputPrice: 0,
1173
+ inputPrice: 0.75,
1174
+ outputPrice: 0.99,
1104
1175
  description: "DeepSeek R1 Distill Llama 70B model, 128K context."
1105
1176
  }
1106
1177
  };
@@ -1152,10 +1223,39 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1152
1223
 
1153
1224
  // src/providers/lm-studio.ts
1154
1225
  var LMSTUDIO_DEFAULT_TEMPERATURE = 0;
1226
+ var lMStudioDefaultModelId = "mistralai/devstral-small-2505";
1227
+ var lMStudioDefaultModelInfo = {
1228
+ maxTokens: 8192,
1229
+ contextWindow: 2e5,
1230
+ supportsImages: true,
1231
+ supportsComputerUse: true,
1232
+ supportsPromptCache: true,
1233
+ inputPrice: 0,
1234
+ outputPrice: 0,
1235
+ cacheWritesPrice: 0,
1236
+ cacheReadsPrice: 0,
1237
+ description: "LM Studio hosted models"
1238
+ };
1155
1239
 
1156
1240
  // src/providers/mistral.ts
1157
1241
  var mistralDefaultModelId = "codestral-latest";
1158
1242
  var mistralModels = {
1243
+ "magistral-medium-latest": {
1244
+ maxTokens: 41e3,
1245
+ contextWindow: 41e3,
1246
+ supportsImages: false,
1247
+ supportsPromptCache: false,
1248
+ inputPrice: 2,
1249
+ outputPrice: 5
1250
+ },
1251
+ "mistral-medium-latest": {
1252
+ maxTokens: 131e3,
1253
+ contextWindow: 131e3,
1254
+ supportsImages: true,
1255
+ supportsPromptCache: false,
1256
+ inputPrice: 0.4,
1257
+ outputPrice: 2
1258
+ },
1159
1259
  "codestral-latest": {
1160
1260
  maxTokens: 256e3,
1161
1261
  contextWindow: 256e3,
@@ -1207,6 +1307,21 @@ var mistralModels = {
1207
1307
  };
1208
1308
  var MISTRAL_DEFAULT_TEMPERATURE = 0;
1209
1309
 
1310
+ // src/providers/ollama.ts
1311
+ var ollamaDefaultModelId = "devstral:24b";
1312
+ var ollamaDefaultModelInfo = {
1313
+ maxTokens: 4096,
1314
+ contextWindow: 2e5,
1315
+ supportsImages: true,
1316
+ supportsComputerUse: true,
1317
+ supportsPromptCache: true,
1318
+ inputPrice: 0,
1319
+ outputPrice: 0,
1320
+ cacheWritesPrice: 0,
1321
+ cacheReadsPrice: 0,
1322
+ description: "Ollama hosted models"
1323
+ };
1324
+
1210
1325
  // src/providers/openai.ts
1211
1326
  var openAiNativeDefaultModelId = "gpt-4.1";
1212
1327
  var openAiNativeModels = {
@@ -1436,6 +1551,8 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
1436
1551
  "google/gemini-2.5-flash-preview:thinking",
1437
1552
  "google/gemini-2.5-flash-preview-05-20",
1438
1553
  "google/gemini-2.5-flash-preview-05-20:thinking",
1554
+ "google/gemini-2.5-flash",
1555
+ "google/gemini-2.5-flash-lite-preview-06-17",
1439
1556
  "google/gemini-2.0-flash-001",
1440
1557
  "google/gemini-flash-1.5",
1441
1558
  "google/gemini-flash-1.5-8b"
@@ -1451,6 +1568,7 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1451
1568
  ]);
1452
1569
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1453
1570
  "anthropic/claude-3.7-sonnet:thinking",
1571
+ "google/gemini-2.5-pro",
1454
1572
  "google/gemini-2.5-flash-preview-05-20:thinking"
1455
1573
  ]);
1456
1574
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
@@ -1458,7 +1576,10 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1458
1576
  "anthropic/claude-opus-4",
1459
1577
  "anthropic/claude-sonnet-4",
1460
1578
  "google/gemini-2.5-pro-preview",
1579
+ "google/gemini-2.5-pro",
1461
1580
  "google/gemini-2.5-flash-preview-05-20",
1581
+ "google/gemini-2.5-flash",
1582
+ "google/gemini-2.5-flash-lite-preview-06-17",
1462
1583
  // Also include the models that require the reasoning budget to be enabled
1463
1584
  // even though `OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS` takes precedence.
1464
1585
  "anthropic/claude-3.7-sonnet:thinking",
@@ -1515,6 +1636,18 @@ var vertexModels = {
1515
1636
  inputPrice: 0.15,
1516
1637
  outputPrice: 0.6
1517
1638
  },
1639
+ "gemini-2.5-flash": {
1640
+ maxTokens: 64e3,
1641
+ contextWindow: 1048576,
1642
+ supportsImages: true,
1643
+ supportsPromptCache: true,
1644
+ inputPrice: 0.3,
1645
+ outputPrice: 2.5,
1646
+ cacheReadsPrice: 0.075,
1647
+ cacheWritesPrice: 1,
1648
+ maxThinkingTokens: 24576,
1649
+ supportsReasoningBudget: true
1650
+ },
1518
1651
  "gemini-2.5-flash-preview-04-17:thinking": {
1519
1652
  maxTokens: 65535,
1520
1653
  contextWindow: 1048576,
@@ -1560,6 +1693,31 @@ var vertexModels = {
1560
1693
  maxThinkingTokens: 32768,
1561
1694
  supportsReasoningBudget: true
1562
1695
  },
1696
+ "gemini-2.5-pro": {
1697
+ maxTokens: 64e3,
1698
+ contextWindow: 1048576,
1699
+ supportsImages: true,
1700
+ supportsPromptCache: true,
1701
+ inputPrice: 2.5,
1702
+ outputPrice: 15,
1703
+ maxThinkingTokens: 32768,
1704
+ supportsReasoningBudget: true,
1705
+ requiredReasoningBudget: true,
1706
+ tiers: [
1707
+ {
1708
+ contextWindow: 2e5,
1709
+ inputPrice: 1.25,
1710
+ outputPrice: 10,
1711
+ cacheReadsPrice: 0.31
1712
+ },
1713
+ {
1714
+ contextWindow: Infinity,
1715
+ inputPrice: 2.5,
1716
+ outputPrice: 15,
1717
+ cacheReadsPrice: 0.625
1718
+ }
1719
+ ]
1720
+ },
1563
1721
  "gemini-2.5-pro-exp-03-25": {
1564
1722
  maxTokens: 65535,
1565
1723
  contextWindow: 1048576,
@@ -1713,6 +1871,18 @@ var vertexModels = {
1713
1871
  outputPrice: 1.25,
1714
1872
  cacheWritesPrice: 0.3,
1715
1873
  cacheReadsPrice: 0.03
1874
+ },
1875
+ "gemini-2.5-flash-lite-preview-06-17": {
1876
+ maxTokens: 64e3,
1877
+ contextWindow: 1048576,
1878
+ supportsImages: true,
1879
+ supportsPromptCache: true,
1880
+ inputPrice: 0.1,
1881
+ outputPrice: 0.4,
1882
+ cacheReadsPrice: 0.025,
1883
+ cacheWritesPrice: 1,
1884
+ maxThinkingTokens: 24576,
1885
+ supportsReasoningBudget: true
1716
1886
  }
1717
1887
  };
1718
1888
  var VERTEX_REGIONS = [
@@ -2014,6 +2184,7 @@ var modelInfoSchema = z2.object({
2014
2184
  // src/provider-settings.ts
2015
2185
  var providerNames = [
2016
2186
  "anthropic",
2187
+ "claude-code",
2017
2188
  "glama",
2018
2189
  "openrouter",
2019
2190
  "bedrock",
@@ -2061,6 +2232,9 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
2061
2232
  anthropicBaseUrl: z3.string().optional(),
2062
2233
  anthropicUseAuthToken: z3.boolean().optional()
2063
2234
  });
2235
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
2236
+ claudeCodePath: z3.string().optional()
2237
+ });
2064
2238
  var glamaSchema = baseProviderSettingsSchema.extend({
2065
2239
  glamaModelId: z3.string().optional(),
2066
2240
  glamaApiKey: z3.string().optional()
@@ -2171,6 +2345,7 @@ var defaultSchema = z3.object({
2171
2345
  });
2172
2346
  var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
2173
2347
  anthropicSchema.merge(z3.object({ apiProvider: z3.literal("anthropic") })),
2348
+ claudeCodeSchema.merge(z3.object({ apiProvider: z3.literal("claude-code") })),
2174
2349
  glamaSchema.merge(z3.object({ apiProvider: z3.literal("glama") })),
2175
2350
  openRouterSchema.merge(z3.object({ apiProvider: z3.literal("openrouter") })),
2176
2351
  bedrockSchema.merge(z3.object({ apiProvider: z3.literal("bedrock") })),
@@ -2196,6 +2371,7 @@ var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
2196
2371
  var providerSettingsSchema = z3.object({
2197
2372
  apiProvider: providerNamesSchema.optional(),
2198
2373
  ...anthropicSchema.shape,
2374
+ ...claudeCodeSchema.shape,
2199
2375
  ...glamaSchema.shape,
2200
2376
  ...openRouterSchema.shape,
2201
2377
  ...bedrockSchema.shape,
@@ -2254,12 +2430,10 @@ var historyItemSchema = z4.object({
2254
2430
 
2255
2431
  // src/experiment.ts
2256
2432
  import { z as z5 } from "zod";
2257
- var experimentIds = ["powerSteering", "disableCompletionCommand", "marketplace", "multiFileApplyDiff"];
2433
+ var experimentIds = ["powerSteering", "multiFileApplyDiff"];
2258
2434
  var experimentIdsSchema = z5.enum(experimentIds);
2259
2435
  var experimentsSchema = z5.object({
2260
2436
  powerSteering: z5.boolean().optional(),
2261
- disableCompletionCommand: z5.boolean().optional(),
2262
- marketplace: z5.boolean().optional(),
2263
2437
  multiFileApplyDiff: z5.boolean().optional()
2264
2438
  });
2265
2439
 
@@ -2568,7 +2742,8 @@ var commandIds = [
2568
2742
  "newTask",
2569
2743
  "setCustomStoragePath",
2570
2744
  "focusInput",
2571
- "acceptInput"
2745
+ "acceptInput",
2746
+ "focusPanel"
2572
2747
  ];
2573
2748
  var languages = [
2574
2749
  "ca",
@@ -2663,7 +2838,8 @@ var globalSettingsSchema = z11.object({
2663
2838
  customModePrompts: customModePromptsSchema.optional(),
2664
2839
  customSupportPrompts: customSupportPromptsSchema.optional(),
2665
2840
  enhancementApiConfigId: z11.string().optional(),
2666
- historyPreviewCollapsed: z11.boolean().optional()
2841
+ historyPreviewCollapsed: z11.boolean().optional(),
2842
+ profileThresholds: z11.record(z11.string(), z11.number()).optional()
2667
2843
  });
2668
2844
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
2669
2845
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -2800,6 +2976,7 @@ var ORGANIZATION_ALLOW_ALL = {
2800
2976
  var ORGANIZATION_DEFAULT = {
2801
2977
  version: 0,
2802
2978
  cloudSettings: {
2979
+ recordTaskMessages: true,
2803
2980
  enableTaskSharing: true,
2804
2981
  taskShareExpirationDays: 30
2805
2982
  },
@@ -2809,11 +2986,16 @@ var ORGANIZATION_DEFAULT = {
2809
2986
  var shareResponseSchema = z12.object({
2810
2987
  success: z12.boolean(),
2811
2988
  shareUrl: z12.string().optional(),
2812
- error: z12.string().optional()
2989
+ error: z12.string().optional(),
2990
+ isNewShare: z12.boolean().optional(),
2991
+ manageUrl: z12.string().optional()
2813
2992
  });
2814
2993
 
2815
2994
  // src/ipc.ts
2816
2995
  import { z as z13 } from "zod";
2996
+ var isSubtaskSchema = z13.object({
2997
+ isSubtask: z13.boolean()
2998
+ });
2817
2999
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
2818
3000
  RooCodeEventName2["Message"] = "message";
2819
3001
  RooCodeEventName2["TaskCreated"] = "taskCreated";
@@ -2847,7 +3029,7 @@ var rooCodeEventsSchema = z13.object({
2847
3029
  ["taskAskResponded" /* TaskAskResponded */]: z13.tuple([z13.string()]),
2848
3030
  ["taskAborted" /* TaskAborted */]: z13.tuple([z13.string()]),
2849
3031
  ["taskSpawned" /* TaskSpawned */]: z13.tuple([z13.string(), z13.string()]),
2850
- ["taskCompleted" /* TaskCompleted */]: z13.tuple([z13.string(), tokenUsageSchema, toolUsageSchema]),
3032
+ ["taskCompleted" /* TaskCompleted */]: z13.tuple([z13.string(), tokenUsageSchema, toolUsageSchema, isSubtaskSchema]),
2851
3033
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z13.tuple([z13.string(), tokenUsageSchema]),
2852
3034
  ["taskToolFailed" /* TaskToolFailed */]: z13.tuple([z13.string(), toolNamesSchema, z13.string()])
2853
3035
  });
@@ -3130,6 +3312,8 @@ export {
3130
3312
  bedrockModels,
3131
3313
  chutesDefaultModelId,
3132
3314
  chutesModels,
3315
+ claudeCodeDefaultModelId,
3316
+ claudeCodeModels,
3133
3317
  clineAskSchema,
3134
3318
  clineAsks,
3135
3319
  clineMessageSchema,
@@ -3167,6 +3351,9 @@ export {
3167
3351
  isLanguage,
3168
3352
  isModelParameter,
3169
3353
  isSecretStateKey,
3354
+ isSubtaskSchema,
3355
+ lMStudioDefaultModelId,
3356
+ lMStudioDefaultModelInfo,
3170
3357
  languages,
3171
3358
  languagesSchema,
3172
3359
  litellmDefaultModelId,
@@ -3184,6 +3371,8 @@ export {
3184
3371
  modelInfoSchema,
3185
3372
  modelParameters,
3186
3373
  modelParametersSchema,
3374
+ ollamaDefaultModelId,
3375
+ ollamaDefaultModelInfo,
3187
3376
  openAiModelInfoSaneDefaults,
3188
3377
  openAiNativeDefaultModelId,
3189
3378
  openAiNativeModels,