@roo-code/types 1.80.0 → 1.82.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -388,6 +388,8 @@ var modelInfoSchema = z5.object({
388
388
  minTokensPerCachePoint: z5.number().optional(),
389
389
  maxCachePoints: z5.number().optional(),
390
390
  cachableFields: z5.array(z5.string()).optional(),
391
+ // Flag to indicate if the model is deprecated and should not be used
392
+ deprecated: z5.boolean().optional(),
391
393
  /**
392
394
  * Service tiers with pricing information.
393
395
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -1192,6 +1194,33 @@ var chutesModels = {
1192
1194
  outputPrice: 0,
1193
1195
  description: "DeepSeek V3.1 model."
1194
1196
  },
1197
+ "deepseek-ai/DeepSeek-V3.1-Terminus": {
1198
+ maxTokens: 163840,
1199
+ contextWindow: 163840,
1200
+ supportsImages: false,
1201
+ supportsPromptCache: false,
1202
+ inputPrice: 0.23,
1203
+ outputPrice: 0.9,
1204
+ description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
1205
+ },
1206
+ "deepseek-ai/DeepSeek-V3.1-turbo": {
1207
+ maxTokens: 32768,
1208
+ contextWindow: 163840,
1209
+ supportsImages: false,
1210
+ supportsPromptCache: false,
1211
+ inputPrice: 1,
1212
+ outputPrice: 3,
1213
+ description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
1214
+ },
1215
+ "deepseek-ai/DeepSeek-V3.2-Exp": {
1216
+ maxTokens: 163840,
1217
+ contextWindow: 163840,
1218
+ supportsImages: false,
1219
+ supportsPromptCache: false,
1220
+ inputPrice: 0.25,
1221
+ outputPrice: 0.35,
1222
+ description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
1223
+ },
1195
1224
  "unsloth/Llama-3.3-70B-Instruct": {
1196
1225
  maxTokens: 32768,
1197
1226
  // From Groq
@@ -1401,6 +1430,15 @@ var chutesModels = {
1401
1430
  outputPrice: 3,
1402
1431
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
1403
1432
  },
1433
+ "zai-org/GLM-4.6-FP8": {
1434
+ maxTokens: 32768,
1435
+ contextWindow: 202752,
1436
+ supportsImages: false,
1437
+ supportsPromptCache: false,
1438
+ inputPrice: 0,
1439
+ outputPrice: 0,
1440
+ description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
1441
+ },
1404
1442
  "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": {
1405
1443
  maxTokens: 32768,
1406
1444
  contextWindow: 262144,
@@ -1454,6 +1492,15 @@ var chutesModels = {
1454
1492
  inputPrice: 0,
1455
1493
  outputPrice: 0,
1456
1494
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
1495
+ },
1496
+ "Qwen/Qwen3-VL-235B-A22B-Thinking": {
1497
+ maxTokens: 262144,
1498
+ contextWindow: 262144,
1499
+ supportsImages: true,
1500
+ supportsPromptCache: false,
1501
+ inputPrice: 0.16,
1502
+ outputPrice: 0.65,
1503
+ description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
1457
1504
  }
1458
1505
  };
1459
1506
 
@@ -2923,7 +2970,8 @@ var rooModels = {
2923
2970
  supportsPromptCache: false,
2924
2971
  inputPrice: 0,
2925
2972
  outputPrice: 0,
2926
- description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)"
2973
+ description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)",
2974
+ deprecated: true
2927
2975
  },
2928
2976
  "deepseek/deepseek-chat-v3.1": {
2929
2977
  maxTokens: 16384,
@@ -4472,7 +4520,7 @@ var MODELS_BY_PROVIDER = {
4472
4520
  },
4473
4521
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
4474
4522
  zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
4475
- // Dynamic providers; models pulled from the respective APIs.
4523
+ // Dynamic providers; models pulled from remote APIs.
4476
4524
  glama: { id: "glama", label: "Glama", models: [] },
4477
4525
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
4478
4526
  litellm: { id: "litellm", label: "LiteLLM", models: [] },
@@ -4480,7 +4528,10 @@ var MODELS_BY_PROVIDER = {
4480
4528
  requesty: { id: "requesty", label: "Requesty", models: [] },
4481
4529
  unbound: { id: "unbound", label: "Unbound", models: [] },
4482
4530
  deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
4483
- "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
4531
+ "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] },
4532
+ // Local providers; models discovered from localhost endpoints.
4533
+ lmstudio: { id: "lmstudio", label: "LM Studio", models: [] },
4534
+ ollama: { id: "ollama", label: "Ollama", models: [] }
4484
4535
  };
4485
4536
 
4486
4537
  // src/history.ts
@@ -4591,6 +4642,7 @@ var appPropertiesSchema = z10.object({
4591
4642
  });
4592
4643
  var taskPropertiesSchema = z10.object({
4593
4644
  taskId: z10.string().optional(),
4645
+ parentTaskId: z10.string().optional(),
4594
4646
  apiProvider: z10.enum(providerNames).optional(),
4595
4647
  modelId: z10.string().optional(),
4596
4648
  diffStrategy: z10.string().optional(),