@roo-code/types 1.27.0 → 1.28.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -63,6 +63,8 @@ __export(index_exports, {
63
63
  bedrockModels: () => bedrockModels,
64
64
  chutesDefaultModelId: () => chutesDefaultModelId,
65
65
  chutesModels: () => chutesModels,
66
+ claudeCodeDefaultModelId: () => claudeCodeDefaultModelId,
67
+ claudeCodeModels: () => claudeCodeModels,
66
68
  clineAskSchema: () => clineAskSchema,
67
69
  clineAsks: () => clineAsks,
68
70
  clineMessageSchema: () => clineMessageSchema,
@@ -100,6 +102,9 @@ __export(index_exports, {
100
102
  isLanguage: () => isLanguage,
101
103
  isModelParameter: () => isModelParameter,
102
104
  isSecretStateKey: () => isSecretStateKey,
105
+ isSubtaskSchema: () => isSubtaskSchema,
106
+ lMStudioDefaultModelId: () => lMStudioDefaultModelId,
107
+ lMStudioDefaultModelInfo: () => lMStudioDefaultModelInfo,
103
108
  languages: () => languages,
104
109
  languagesSchema: () => languagesSchema,
105
110
  litellmDefaultModelId: () => litellmDefaultModelId,
@@ -117,6 +122,8 @@ __export(index_exports, {
117
122
  modelInfoSchema: () => modelInfoSchema,
118
123
  modelParameters: () => modelParameters,
119
124
  modelParametersSchema: () => modelParametersSchema,
125
+ ollamaDefaultModelId: () => ollamaDefaultModelId,
126
+ ollamaDefaultModelInfo: () => ollamaDefaultModelInfo,
120
127
  openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
121
128
  openAiNativeDefaultModelId: () => openAiNativeDefaultModelId,
122
129
  openAiNativeModels: () => openAiNativeModels,
@@ -904,6 +911,16 @@ var chutesModels = {
904
911
  }
905
912
  };
906
913
 
914
+ // src/providers/claude-code.ts
915
+ var claudeCodeDefaultModelId = "claude-sonnet-4-20250514";
916
+ var claudeCodeModels = {
917
+ "claude-sonnet-4-20250514": anthropicModels["claude-sonnet-4-20250514"],
918
+ "claude-opus-4-20250514": anthropicModels["claude-opus-4-20250514"],
919
+ "claude-3-7-sonnet-20250219": anthropicModels["claude-3-7-sonnet-20250219"],
920
+ "claude-3-5-sonnet-20241022": anthropicModels["claude-3-5-sonnet-20241022"],
921
+ "claude-3-5-haiku-20241022": anthropicModels["claude-3-5-haiku-20241022"]
922
+ };
923
+
907
924
  // src/providers/deepseek.ts
908
925
  var deepSeekDefaultModelId = "deepseek-chat";
909
926
  var deepSeekModels = {
@@ -985,6 +1002,18 @@ var geminiModels = {
985
1002
  cacheReadsPrice: 0.0375,
986
1003
  cacheWritesPrice: 1
987
1004
  },
1005
+ "gemini-2.5-flash": {
1006
+ maxTokens: 64e3,
1007
+ contextWindow: 1048576,
1008
+ supportsImages: true,
1009
+ supportsPromptCache: true,
1010
+ inputPrice: 0.3,
1011
+ outputPrice: 2.5,
1012
+ cacheReadsPrice: 0.075,
1013
+ cacheWritesPrice: 1,
1014
+ maxThinkingTokens: 24576,
1015
+ supportsReasoningBudget: true
1016
+ },
988
1017
  "gemini-2.5-pro-exp-03-25": {
989
1018
  maxTokens: 65535,
990
1019
  contextWindow: 1048576,
@@ -1070,6 +1099,34 @@ var geminiModels = {
1070
1099
  }
1071
1100
  ]
1072
1101
  },
1102
+ "gemini-2.5-pro": {
1103
+ maxTokens: 64e3,
1104
+ contextWindow: 1048576,
1105
+ supportsImages: true,
1106
+ supportsPromptCache: true,
1107
+ inputPrice: 2.5,
1108
+ // This is the pricing for prompts above 200k tokens.
1109
+ outputPrice: 15,
1110
+ cacheReadsPrice: 0.625,
1111
+ cacheWritesPrice: 4.5,
1112
+ maxThinkingTokens: 32768,
1113
+ supportsReasoningBudget: true,
1114
+ requiredReasoningBudget: true,
1115
+ tiers: [
1116
+ {
1117
+ contextWindow: 2e5,
1118
+ inputPrice: 1.25,
1119
+ outputPrice: 10,
1120
+ cacheReadsPrice: 0.31
1121
+ },
1122
+ {
1123
+ contextWindow: Infinity,
1124
+ inputPrice: 2.5,
1125
+ outputPrice: 15,
1126
+ cacheReadsPrice: 0.625
1127
+ }
1128
+ ]
1129
+ },
1073
1130
  "gemini-2.0-flash-001": {
1074
1131
  maxTokens: 8192,
1075
1132
  contextWindow: 1048576,
@@ -1184,6 +1241,18 @@ var geminiModels = {
1184
1241
  supportsPromptCache: false,
1185
1242
  inputPrice: 0,
1186
1243
  outputPrice: 0
1244
+ },
1245
+ "gemini-2.5-flash-lite-preview-06-17": {
1246
+ maxTokens: 64e3,
1247
+ contextWindow: 1048576,
1248
+ supportsImages: true,
1249
+ supportsPromptCache: true,
1250
+ inputPrice: 0.1,
1251
+ outputPrice: 0.4,
1252
+ cacheReadsPrice: 0.025,
1253
+ cacheWritesPrice: 1,
1254
+ supportsReasoningBudget: true,
1255
+ maxThinkingTokens: 24576
1187
1256
  }
1188
1257
  };
1189
1258
 
@@ -1212,8 +1281,8 @@ var groqModels = {
1212
1281
  contextWindow: 131072,
1213
1282
  supportsImages: false,
1214
1283
  supportsPromptCache: false,
1215
- inputPrice: 0,
1216
- outputPrice: 0,
1284
+ inputPrice: 0.05,
1285
+ outputPrice: 0.08,
1217
1286
  description: "Meta Llama 3.1 8B Instant model, 128K context."
1218
1287
  },
1219
1288
  "llama-3.3-70b-versatile": {
@@ -1221,8 +1290,8 @@ var groqModels = {
1221
1290
  contextWindow: 131072,
1222
1291
  supportsImages: false,
1223
1292
  supportsPromptCache: false,
1224
- inputPrice: 0,
1225
- outputPrice: 0,
1293
+ inputPrice: 0.59,
1294
+ outputPrice: 0.79,
1226
1295
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
1227
1296
  },
1228
1297
  "meta-llama/llama-4-scout-17b-16e-instruct": {
@@ -1230,8 +1299,8 @@ var groqModels = {
1230
1299
  contextWindow: 131072,
1231
1300
  supportsImages: false,
1232
1301
  supportsPromptCache: false,
1233
- inputPrice: 0,
1234
- outputPrice: 0,
1302
+ inputPrice: 0.11,
1303
+ outputPrice: 0.34,
1235
1304
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
1236
1305
  },
1237
1306
  "meta-llama/llama-4-maverick-17b-128e-instruct": {
@@ -1239,8 +1308,8 @@ var groqModels = {
1239
1308
  contextWindow: 131072,
1240
1309
  supportsImages: false,
1241
1310
  supportsPromptCache: false,
1242
- inputPrice: 0,
1243
- outputPrice: 0,
1311
+ inputPrice: 0.2,
1312
+ outputPrice: 0.6,
1244
1313
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
1245
1314
  },
1246
1315
  "mistral-saba-24b": {
@@ -1248,8 +1317,8 @@ var groqModels = {
1248
1317
  contextWindow: 32768,
1249
1318
  supportsImages: false,
1250
1319
  supportsPromptCache: false,
1251
- inputPrice: 0,
1252
- outputPrice: 0,
1320
+ inputPrice: 0.79,
1321
+ outputPrice: 0.79,
1253
1322
  description: "Mistral Saba 24B model, 32K context."
1254
1323
  },
1255
1324
  "qwen-qwq-32b": {
@@ -1257,17 +1326,26 @@ var groqModels = {
1257
1326
  contextWindow: 131072,
1258
1327
  supportsImages: false,
1259
1328
  supportsPromptCache: false,
1260
- inputPrice: 0,
1261
- outputPrice: 0,
1329
+ inputPrice: 0.29,
1330
+ outputPrice: 0.39,
1262
1331
  description: "Alibaba Qwen QwQ 32B model, 128K context."
1263
1332
  },
1333
+ "qwen/qwen3-32b": {
1334
+ maxTokens: 131072,
1335
+ contextWindow: 131072,
1336
+ supportsImages: false,
1337
+ supportsPromptCache: false,
1338
+ inputPrice: 0.29,
1339
+ outputPrice: 0.59,
1340
+ description: "Alibaba Qwen 3 32B model, 128K context."
1341
+ },
1264
1342
  "deepseek-r1-distill-llama-70b": {
1265
1343
  maxTokens: 131072,
1266
1344
  contextWindow: 131072,
1267
1345
  supportsImages: false,
1268
1346
  supportsPromptCache: false,
1269
- inputPrice: 0,
1270
- outputPrice: 0,
1347
+ inputPrice: 0.75,
1348
+ outputPrice: 0.99,
1271
1349
  description: "DeepSeek R1 Distill Llama 70B model, 128K context."
1272
1350
  }
1273
1351
  };
@@ -1319,10 +1397,39 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1319
1397
 
1320
1398
  // src/providers/lm-studio.ts
1321
1399
  var LMSTUDIO_DEFAULT_TEMPERATURE = 0;
1400
+ var lMStudioDefaultModelId = "mistralai/devstral-small-2505";
1401
+ var lMStudioDefaultModelInfo = {
1402
+ maxTokens: 8192,
1403
+ contextWindow: 2e5,
1404
+ supportsImages: true,
1405
+ supportsComputerUse: true,
1406
+ supportsPromptCache: true,
1407
+ inputPrice: 0,
1408
+ outputPrice: 0,
1409
+ cacheWritesPrice: 0,
1410
+ cacheReadsPrice: 0,
1411
+ description: "LM Studio hosted models"
1412
+ };
1322
1413
 
1323
1414
  // src/providers/mistral.ts
1324
1415
  var mistralDefaultModelId = "codestral-latest";
1325
1416
  var mistralModels = {
1417
+ "magistral-medium-latest": {
1418
+ maxTokens: 41e3,
1419
+ contextWindow: 41e3,
1420
+ supportsImages: false,
1421
+ supportsPromptCache: false,
1422
+ inputPrice: 2,
1423
+ outputPrice: 5
1424
+ },
1425
+ "mistral-medium-latest": {
1426
+ maxTokens: 131e3,
1427
+ contextWindow: 131e3,
1428
+ supportsImages: true,
1429
+ supportsPromptCache: false,
1430
+ inputPrice: 0.4,
1431
+ outputPrice: 2
1432
+ },
1326
1433
  "codestral-latest": {
1327
1434
  maxTokens: 256e3,
1328
1435
  contextWindow: 256e3,
@@ -1374,6 +1481,21 @@ var mistralModels = {
1374
1481
  };
1375
1482
  var MISTRAL_DEFAULT_TEMPERATURE = 0;
1376
1483
 
1484
+ // src/providers/ollama.ts
1485
+ var ollamaDefaultModelId = "devstral:24b";
1486
+ var ollamaDefaultModelInfo = {
1487
+ maxTokens: 4096,
1488
+ contextWindow: 2e5,
1489
+ supportsImages: true,
1490
+ supportsComputerUse: true,
1491
+ supportsPromptCache: true,
1492
+ inputPrice: 0,
1493
+ outputPrice: 0,
1494
+ cacheWritesPrice: 0,
1495
+ cacheReadsPrice: 0,
1496
+ description: "Ollama hosted models"
1497
+ };
1498
+
1377
1499
  // src/providers/openai.ts
1378
1500
  var openAiNativeDefaultModelId = "gpt-4.1";
1379
1501
  var openAiNativeModels = {
@@ -1603,6 +1725,8 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
1603
1725
  "google/gemini-2.5-flash-preview:thinking",
1604
1726
  "google/gemini-2.5-flash-preview-05-20",
1605
1727
  "google/gemini-2.5-flash-preview-05-20:thinking",
1728
+ "google/gemini-2.5-flash",
1729
+ "google/gemini-2.5-flash-lite-preview-06-17",
1606
1730
  "google/gemini-2.0-flash-001",
1607
1731
  "google/gemini-flash-1.5",
1608
1732
  "google/gemini-flash-1.5-8b"
@@ -1618,6 +1742,7 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1618
1742
  ]);
1619
1743
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1620
1744
  "anthropic/claude-3.7-sonnet:thinking",
1745
+ "google/gemini-2.5-pro",
1621
1746
  "google/gemini-2.5-flash-preview-05-20:thinking"
1622
1747
  ]);
1623
1748
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
@@ -1625,7 +1750,10 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1625
1750
  "anthropic/claude-opus-4",
1626
1751
  "anthropic/claude-sonnet-4",
1627
1752
  "google/gemini-2.5-pro-preview",
1753
+ "google/gemini-2.5-pro",
1628
1754
  "google/gemini-2.5-flash-preview-05-20",
1755
+ "google/gemini-2.5-flash",
1756
+ "google/gemini-2.5-flash-lite-preview-06-17",
1629
1757
  // Also include the models that require the reasoning budget to be enabled
1630
1758
  // even though `OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS` takes precedence.
1631
1759
  "anthropic/claude-3.7-sonnet:thinking",
@@ -1682,6 +1810,18 @@ var vertexModels = {
1682
1810
  inputPrice: 0.15,
1683
1811
  outputPrice: 0.6
1684
1812
  },
1813
+ "gemini-2.5-flash": {
1814
+ maxTokens: 64e3,
1815
+ contextWindow: 1048576,
1816
+ supportsImages: true,
1817
+ supportsPromptCache: true,
1818
+ inputPrice: 0.3,
1819
+ outputPrice: 2.5,
1820
+ cacheReadsPrice: 0.075,
1821
+ cacheWritesPrice: 1,
1822
+ maxThinkingTokens: 24576,
1823
+ supportsReasoningBudget: true
1824
+ },
1685
1825
  "gemini-2.5-flash-preview-04-17:thinking": {
1686
1826
  maxTokens: 65535,
1687
1827
  contextWindow: 1048576,
@@ -1727,6 +1867,31 @@ var vertexModels = {
1727
1867
  maxThinkingTokens: 32768,
1728
1868
  supportsReasoningBudget: true
1729
1869
  },
1870
+ "gemini-2.5-pro": {
1871
+ maxTokens: 64e3,
1872
+ contextWindow: 1048576,
1873
+ supportsImages: true,
1874
+ supportsPromptCache: true,
1875
+ inputPrice: 2.5,
1876
+ outputPrice: 15,
1877
+ maxThinkingTokens: 32768,
1878
+ supportsReasoningBudget: true,
1879
+ requiredReasoningBudget: true,
1880
+ tiers: [
1881
+ {
1882
+ contextWindow: 2e5,
1883
+ inputPrice: 1.25,
1884
+ outputPrice: 10,
1885
+ cacheReadsPrice: 0.31
1886
+ },
1887
+ {
1888
+ contextWindow: Infinity,
1889
+ inputPrice: 2.5,
1890
+ outputPrice: 15,
1891
+ cacheReadsPrice: 0.625
1892
+ }
1893
+ ]
1894
+ },
1730
1895
  "gemini-2.5-pro-exp-03-25": {
1731
1896
  maxTokens: 65535,
1732
1897
  contextWindow: 1048576,
@@ -1880,6 +2045,18 @@ var vertexModels = {
1880
2045
  outputPrice: 1.25,
1881
2046
  cacheWritesPrice: 0.3,
1882
2047
  cacheReadsPrice: 0.03
2048
+ },
2049
+ "gemini-2.5-flash-lite-preview-06-17": {
2050
+ maxTokens: 64e3,
2051
+ contextWindow: 1048576,
2052
+ supportsImages: true,
2053
+ supportsPromptCache: true,
2054
+ inputPrice: 0.1,
2055
+ outputPrice: 0.4,
2056
+ cacheReadsPrice: 0.025,
2057
+ cacheWritesPrice: 1,
2058
+ maxThinkingTokens: 24576,
2059
+ supportsReasoningBudget: true
1883
2060
  }
1884
2061
  };
1885
2062
  var VERTEX_REGIONS = [
@@ -2181,6 +2358,7 @@ var modelInfoSchema = import_zod2.z.object({
2181
2358
  // src/provider-settings.ts
2182
2359
  var providerNames = [
2183
2360
  "anthropic",
2361
+ "claude-code",
2184
2362
  "glama",
2185
2363
  "openrouter",
2186
2364
  "bedrock",
@@ -2228,6 +2406,9 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
2228
2406
  anthropicBaseUrl: import_zod3.z.string().optional(),
2229
2407
  anthropicUseAuthToken: import_zod3.z.boolean().optional()
2230
2408
  });
2409
+ var claudeCodeSchema = apiModelIdProviderModelSchema.extend({
2410
+ claudeCodePath: import_zod3.z.string().optional()
2411
+ });
2231
2412
  var glamaSchema = baseProviderSettingsSchema.extend({
2232
2413
  glamaModelId: import_zod3.z.string().optional(),
2233
2414
  glamaApiKey: import_zod3.z.string().optional()
@@ -2338,6 +2519,7 @@ var defaultSchema = import_zod3.z.object({
2338
2519
  });
2339
2520
  var providerSettingsSchemaDiscriminated = import_zod3.z.discriminatedUnion("apiProvider", [
2340
2521
  anthropicSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("anthropic") })),
2522
+ claudeCodeSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("claude-code") })),
2341
2523
  glamaSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("glama") })),
2342
2524
  openRouterSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("openrouter") })),
2343
2525
  bedrockSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("bedrock") })),
@@ -2363,6 +2545,7 @@ var providerSettingsSchemaDiscriminated = import_zod3.z.discriminatedUnion("apiP
2363
2545
  var providerSettingsSchema = import_zod3.z.object({
2364
2546
  apiProvider: providerNamesSchema.optional(),
2365
2547
  ...anthropicSchema.shape,
2548
+ ...claudeCodeSchema.shape,
2366
2549
  ...glamaSchema.shape,
2367
2550
  ...openRouterSchema.shape,
2368
2551
  ...bedrockSchema.shape,
@@ -2421,12 +2604,10 @@ var historyItemSchema = import_zod4.z.object({
2421
2604
 
2422
2605
  // src/experiment.ts
2423
2606
  var import_zod5 = require("zod");
2424
- var experimentIds = ["powerSteering", "disableCompletionCommand", "marketplace", "multiFileApplyDiff"];
2607
+ var experimentIds = ["powerSteering", "multiFileApplyDiff"];
2425
2608
  var experimentIdsSchema = import_zod5.z.enum(experimentIds);
2426
2609
  var experimentsSchema = import_zod5.z.object({
2427
2610
  powerSteering: import_zod5.z.boolean().optional(),
2428
- disableCompletionCommand: import_zod5.z.boolean().optional(),
2429
- marketplace: import_zod5.z.boolean().optional(),
2430
2611
  multiFileApplyDiff: import_zod5.z.boolean().optional()
2431
2612
  });
2432
2613
 
@@ -2735,7 +2916,8 @@ var commandIds = [
2735
2916
  "newTask",
2736
2917
  "setCustomStoragePath",
2737
2918
  "focusInput",
2738
- "acceptInput"
2919
+ "acceptInput",
2920
+ "focusPanel"
2739
2921
  ];
2740
2922
  var languages = [
2741
2923
  "ca",
@@ -2830,7 +3012,8 @@ var globalSettingsSchema = import_zod11.z.object({
2830
3012
  customModePrompts: customModePromptsSchema.optional(),
2831
3013
  customSupportPrompts: customSupportPromptsSchema.optional(),
2832
3014
  enhancementApiConfigId: import_zod11.z.string().optional(),
2833
- historyPreviewCollapsed: import_zod11.z.boolean().optional()
3015
+ historyPreviewCollapsed: import_zod11.z.boolean().optional(),
3016
+ profileThresholds: import_zod11.z.record(import_zod11.z.string(), import_zod11.z.number()).optional()
2834
3017
  });
2835
3018
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
2836
3019
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -2967,6 +3150,7 @@ var ORGANIZATION_ALLOW_ALL = {
2967
3150
  var ORGANIZATION_DEFAULT = {
2968
3151
  version: 0,
2969
3152
  cloudSettings: {
3153
+ recordTaskMessages: true,
2970
3154
  enableTaskSharing: true,
2971
3155
  taskShareExpirationDays: 30
2972
3156
  },
@@ -2976,11 +3160,16 @@ var ORGANIZATION_DEFAULT = {
2976
3160
  var shareResponseSchema = import_zod12.z.object({
2977
3161
  success: import_zod12.z.boolean(),
2978
3162
  shareUrl: import_zod12.z.string().optional(),
2979
- error: import_zod12.z.string().optional()
3163
+ error: import_zod12.z.string().optional(),
3164
+ isNewShare: import_zod12.z.boolean().optional(),
3165
+ manageUrl: import_zod12.z.string().optional()
2980
3166
  });
2981
3167
 
2982
3168
  // src/ipc.ts
2983
3169
  var import_zod13 = require("zod");
3170
+ var isSubtaskSchema = import_zod13.z.object({
3171
+ isSubtask: import_zod13.z.boolean()
3172
+ });
2984
3173
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
2985
3174
  RooCodeEventName2["Message"] = "message";
2986
3175
  RooCodeEventName2["TaskCreated"] = "taskCreated";
@@ -3014,7 +3203,7 @@ var rooCodeEventsSchema = import_zod13.z.object({
3014
3203
  ["taskAskResponded" /* TaskAskResponded */]: import_zod13.z.tuple([import_zod13.z.string()]),
3015
3204
  ["taskAborted" /* TaskAborted */]: import_zod13.z.tuple([import_zod13.z.string()]),
3016
3205
  ["taskSpawned" /* TaskSpawned */]: import_zod13.z.tuple([import_zod13.z.string(), import_zod13.z.string()]),
3017
- ["taskCompleted" /* TaskCompleted */]: import_zod13.z.tuple([import_zod13.z.string(), tokenUsageSchema, toolUsageSchema]),
3206
+ ["taskCompleted" /* TaskCompleted */]: import_zod13.z.tuple([import_zod13.z.string(), tokenUsageSchema, toolUsageSchema, isSubtaskSchema]),
3018
3207
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: import_zod13.z.tuple([import_zod13.z.string(), tokenUsageSchema]),
3019
3208
  ["taskToolFailed" /* TaskToolFailed */]: import_zod13.z.tuple([import_zod13.z.string(), toolNamesSchema, import_zod13.z.string()])
3020
3209
  });
@@ -3298,6 +3487,8 @@ var commandExecutionStatusSchema = import_zod16.z.discriminatedUnion("status", [
3298
3487
  bedrockModels,
3299
3488
  chutesDefaultModelId,
3300
3489
  chutesModels,
3490
+ claudeCodeDefaultModelId,
3491
+ claudeCodeModels,
3301
3492
  clineAskSchema,
3302
3493
  clineAsks,
3303
3494
  clineMessageSchema,
@@ -3335,6 +3526,9 @@ var commandExecutionStatusSchema = import_zod16.z.discriminatedUnion("status", [
3335
3526
  isLanguage,
3336
3527
  isModelParameter,
3337
3528
  isSecretStateKey,
3529
+ isSubtaskSchema,
3530
+ lMStudioDefaultModelId,
3531
+ lMStudioDefaultModelInfo,
3338
3532
  languages,
3339
3533
  languagesSchema,
3340
3534
  litellmDefaultModelId,
@@ -3352,6 +3546,8 @@ var commandExecutionStatusSchema = import_zod16.z.discriminatedUnion("status", [
3352
3546
  modelInfoSchema,
3353
3547
  modelParameters,
3354
3548
  modelParametersSchema,
3549
+ ollamaDefaultModelId,
3550
+ ollamaDefaultModelInfo,
3355
3551
  openAiModelInfoSaneDefaults,
3356
3552
  openAiNativeDefaultModelId,
3357
3553
  openAiNativeModels,