@roo-code/types 1.80.0 → 1.81.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +54 -3
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +103 -2
- package/dist/index.d.ts +103 -2
- package/dist/index.js +54 -3
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.cjs
CHANGED
|
@@ -681,6 +681,8 @@ var modelInfoSchema = import_zod5.z.object({
|
|
|
681
681
|
minTokensPerCachePoint: import_zod5.z.number().optional(),
|
|
682
682
|
maxCachePoints: import_zod5.z.number().optional(),
|
|
683
683
|
cachableFields: import_zod5.z.array(import_zod5.z.string()).optional(),
|
|
684
|
+
// Flag to indicate if the model is deprecated and should not be used
|
|
685
|
+
deprecated: import_zod5.z.boolean().optional(),
|
|
684
686
|
/**
|
|
685
687
|
* Service tiers with pricing information.
|
|
686
688
|
* Each tier can have a name (for OpenAI service tiers) and pricing overrides.
|
|
@@ -1485,6 +1487,33 @@ var chutesModels = {
|
|
|
1485
1487
|
outputPrice: 0,
|
|
1486
1488
|
description: "DeepSeek V3.1 model."
|
|
1487
1489
|
},
|
|
1490
|
+
"deepseek-ai/DeepSeek-V3.1-Terminus": {
|
|
1491
|
+
maxTokens: 163840,
|
|
1492
|
+
contextWindow: 163840,
|
|
1493
|
+
supportsImages: false,
|
|
1494
|
+
supportsPromptCache: false,
|
|
1495
|
+
inputPrice: 0.23,
|
|
1496
|
+
outputPrice: 0.9,
|
|
1497
|
+
description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
|
|
1498
|
+
},
|
|
1499
|
+
"deepseek-ai/DeepSeek-V3.1-turbo": {
|
|
1500
|
+
maxTokens: 32768,
|
|
1501
|
+
contextWindow: 163840,
|
|
1502
|
+
supportsImages: false,
|
|
1503
|
+
supportsPromptCache: false,
|
|
1504
|
+
inputPrice: 1,
|
|
1505
|
+
outputPrice: 3,
|
|
1506
|
+
description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
|
|
1507
|
+
},
|
|
1508
|
+
"deepseek-ai/DeepSeek-V3.2-Exp": {
|
|
1509
|
+
maxTokens: 163840,
|
|
1510
|
+
contextWindow: 163840,
|
|
1511
|
+
supportsImages: false,
|
|
1512
|
+
supportsPromptCache: false,
|
|
1513
|
+
inputPrice: 0.25,
|
|
1514
|
+
outputPrice: 0.35,
|
|
1515
|
+
description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
|
|
1516
|
+
},
|
|
1488
1517
|
"unsloth/Llama-3.3-70B-Instruct": {
|
|
1489
1518
|
maxTokens: 32768,
|
|
1490
1519
|
// From Groq
|
|
@@ -1694,6 +1723,15 @@ var chutesModels = {
|
|
|
1694
1723
|
outputPrice: 3,
|
|
1695
1724
|
description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
|
|
1696
1725
|
},
|
|
1726
|
+
"zai-org/GLM-4.6-FP8": {
|
|
1727
|
+
maxTokens: 32768,
|
|
1728
|
+
contextWindow: 202752,
|
|
1729
|
+
supportsImages: false,
|
|
1730
|
+
supportsPromptCache: false,
|
|
1731
|
+
inputPrice: 0,
|
|
1732
|
+
outputPrice: 0,
|
|
1733
|
+
description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
|
|
1734
|
+
},
|
|
1697
1735
|
"Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": {
|
|
1698
1736
|
maxTokens: 32768,
|
|
1699
1737
|
contextWindow: 262144,
|
|
@@ -1747,6 +1785,15 @@ var chutesModels = {
|
|
|
1747
1785
|
inputPrice: 0,
|
|
1748
1786
|
outputPrice: 0,
|
|
1749
1787
|
description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
|
|
1788
|
+
},
|
|
1789
|
+
"Qwen/Qwen3-VL-235B-A22B-Thinking": {
|
|
1790
|
+
maxTokens: 262144,
|
|
1791
|
+
contextWindow: 262144,
|
|
1792
|
+
supportsImages: true,
|
|
1793
|
+
supportsPromptCache: false,
|
|
1794
|
+
inputPrice: 0.16,
|
|
1795
|
+
outputPrice: 0.65,
|
|
1796
|
+
description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
|
|
1750
1797
|
}
|
|
1751
1798
|
};
|
|
1752
1799
|
|
|
@@ -3216,7 +3263,8 @@ var rooModels = {
|
|
|
3216
3263
|
supportsPromptCache: false,
|
|
3217
3264
|
inputPrice: 0,
|
|
3218
3265
|
outputPrice: 0,
|
|
3219
|
-
description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)"
|
|
3266
|
+
description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)",
|
|
3267
|
+
deprecated: true
|
|
3220
3268
|
},
|
|
3221
3269
|
"deepseek/deepseek-chat-v3.1": {
|
|
3222
3270
|
maxTokens: 16384,
|
|
@@ -4765,7 +4813,7 @@ var MODELS_BY_PROVIDER = {
|
|
|
4765
4813
|
},
|
|
4766
4814
|
xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
|
|
4767
4815
|
zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
|
|
4768
|
-
// Dynamic providers; models pulled from
|
|
4816
|
+
// Dynamic providers; models pulled from remote APIs.
|
|
4769
4817
|
glama: { id: "glama", label: "Glama", models: [] },
|
|
4770
4818
|
huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
|
|
4771
4819
|
litellm: { id: "litellm", label: "LiteLLM", models: [] },
|
|
@@ -4773,7 +4821,10 @@ var MODELS_BY_PROVIDER = {
|
|
|
4773
4821
|
requesty: { id: "requesty", label: "Requesty", models: [] },
|
|
4774
4822
|
unbound: { id: "unbound", label: "Unbound", models: [] },
|
|
4775
4823
|
deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
|
|
4776
|
-
"vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
|
|
4824
|
+
"vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] },
|
|
4825
|
+
// Local providers; models discovered from localhost endpoints.
|
|
4826
|
+
lmstudio: { id: "lmstudio", label: "LM Studio", models: [] },
|
|
4827
|
+
ollama: { id: "ollama", label: "Ollama", models: [] }
|
|
4777
4828
|
};
|
|
4778
4829
|
|
|
4779
4830
|
// src/history.ts
|