@roo-code/types 1.80.0 → 1.81.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -681,6 +681,8 @@ var modelInfoSchema = import_zod5.z.object({
681
681
  minTokensPerCachePoint: import_zod5.z.number().optional(),
682
682
  maxCachePoints: import_zod5.z.number().optional(),
683
683
  cachableFields: import_zod5.z.array(import_zod5.z.string()).optional(),
684
+ // Flag to indicate if the model is deprecated and should not be used
685
+ deprecated: import_zod5.z.boolean().optional(),
684
686
  /**
685
687
  * Service tiers with pricing information.
686
688
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -1485,6 +1487,33 @@ var chutesModels = {
1485
1487
  outputPrice: 0,
1486
1488
  description: "DeepSeek V3.1 model."
1487
1489
  },
1490
+ "deepseek-ai/DeepSeek-V3.1-Terminus": {
1491
+ maxTokens: 163840,
1492
+ contextWindow: 163840,
1493
+ supportsImages: false,
1494
+ supportsPromptCache: false,
1495
+ inputPrice: 0.23,
1496
+ outputPrice: 0.9,
1497
+ description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
1498
+ },
1499
+ "deepseek-ai/DeepSeek-V3.1-turbo": {
1500
+ maxTokens: 32768,
1501
+ contextWindow: 163840,
1502
+ supportsImages: false,
1503
+ supportsPromptCache: false,
1504
+ inputPrice: 1,
1505
+ outputPrice: 3,
1506
+ description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
1507
+ },
1508
+ "deepseek-ai/DeepSeek-V3.2-Exp": {
1509
+ maxTokens: 163840,
1510
+ contextWindow: 163840,
1511
+ supportsImages: false,
1512
+ supportsPromptCache: false,
1513
+ inputPrice: 0.25,
1514
+ outputPrice: 0.35,
1515
+ description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
1516
+ },
1488
1517
  "unsloth/Llama-3.3-70B-Instruct": {
1489
1518
  maxTokens: 32768,
1490
1519
  // From Groq
@@ -1694,6 +1723,15 @@ var chutesModels = {
1694
1723
  outputPrice: 3,
1695
1724
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
1696
1725
  },
1726
+ "zai-org/GLM-4.6-FP8": {
1727
+ maxTokens: 32768,
1728
+ contextWindow: 202752,
1729
+ supportsImages: false,
1730
+ supportsPromptCache: false,
1731
+ inputPrice: 0,
1732
+ outputPrice: 0,
1733
+ description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
1734
+ },
1697
1735
  "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": {
1698
1736
  maxTokens: 32768,
1699
1737
  contextWindow: 262144,
@@ -1747,6 +1785,15 @@ var chutesModels = {
1747
1785
  inputPrice: 0,
1748
1786
  outputPrice: 0,
1749
1787
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
1788
+ },
1789
+ "Qwen/Qwen3-VL-235B-A22B-Thinking": {
1790
+ maxTokens: 262144,
1791
+ contextWindow: 262144,
1792
+ supportsImages: true,
1793
+ supportsPromptCache: false,
1794
+ inputPrice: 0.16,
1795
+ outputPrice: 0.65,
1796
+ description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
1750
1797
  }
1751
1798
  };
1752
1799
 
@@ -3216,7 +3263,8 @@ var rooModels = {
3216
3263
  supportsPromptCache: false,
3217
3264
  inputPrice: 0,
3218
3265
  outputPrice: 0,
3219
- description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)"
3266
+ description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)",
3267
+ deprecated: true
3220
3268
  },
3221
3269
  "deepseek/deepseek-chat-v3.1": {
3222
3270
  maxTokens: 16384,
@@ -4765,7 +4813,7 @@ var MODELS_BY_PROVIDER = {
4765
4813
  },
4766
4814
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
4767
4815
  zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
4768
- // Dynamic providers; models pulled from the respective APIs.
4816
+ // Dynamic providers; models pulled from remote APIs.
4769
4817
  glama: { id: "glama", label: "Glama", models: [] },
4770
4818
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
4771
4819
  litellm: { id: "litellm", label: "LiteLLM", models: [] },
@@ -4773,7 +4821,10 @@ var MODELS_BY_PROVIDER = {
4773
4821
  requesty: { id: "requesty", label: "Requesty", models: [] },
4774
4822
  unbound: { id: "unbound", label: "Unbound", models: [] },
4775
4823
  deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
4776
- "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
4824
+ "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] },
4825
+ // Local providers; models discovered from localhost endpoints.
4826
+ lmstudio: { id: "lmstudio", label: "LM Studio", models: [] },
4827
+ ollama: { id: "ollama", label: "Ollama", models: [] }
4777
4828
  };
4778
4829
 
4779
4830
  // src/history.ts