@roo-code/types 1.80.0 → 1.82.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +55 -3
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +129 -2
- package/dist/index.d.ts +129 -2
- package/dist/index.js +55 -3
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.js
CHANGED
|
@@ -388,6 +388,8 @@ var modelInfoSchema = z5.object({
|
|
|
388
388
|
minTokensPerCachePoint: z5.number().optional(),
|
|
389
389
|
maxCachePoints: z5.number().optional(),
|
|
390
390
|
cachableFields: z5.array(z5.string()).optional(),
|
|
391
|
+
// Flag to indicate if the model is deprecated and should not be used
|
|
392
|
+
deprecated: z5.boolean().optional(),
|
|
391
393
|
/**
|
|
392
394
|
* Service tiers with pricing information.
|
|
393
395
|
* Each tier can have a name (for OpenAI service tiers) and pricing overrides.
|
|
@@ -1192,6 +1194,33 @@ var chutesModels = {
|
|
|
1192
1194
|
outputPrice: 0,
|
|
1193
1195
|
description: "DeepSeek V3.1 model."
|
|
1194
1196
|
},
|
|
1197
|
+
"deepseek-ai/DeepSeek-V3.1-Terminus": {
|
|
1198
|
+
maxTokens: 163840,
|
|
1199
|
+
contextWindow: 163840,
|
|
1200
|
+
supportsImages: false,
|
|
1201
|
+
supportsPromptCache: false,
|
|
1202
|
+
inputPrice: 0.23,
|
|
1203
|
+
outputPrice: 0.9,
|
|
1204
|
+
description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
|
|
1205
|
+
},
|
|
1206
|
+
"deepseek-ai/DeepSeek-V3.1-turbo": {
|
|
1207
|
+
maxTokens: 32768,
|
|
1208
|
+
contextWindow: 163840,
|
|
1209
|
+
supportsImages: false,
|
|
1210
|
+
supportsPromptCache: false,
|
|
1211
|
+
inputPrice: 1,
|
|
1212
|
+
outputPrice: 3,
|
|
1213
|
+
description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
|
|
1214
|
+
},
|
|
1215
|
+
"deepseek-ai/DeepSeek-V3.2-Exp": {
|
|
1216
|
+
maxTokens: 163840,
|
|
1217
|
+
contextWindow: 163840,
|
|
1218
|
+
supportsImages: false,
|
|
1219
|
+
supportsPromptCache: false,
|
|
1220
|
+
inputPrice: 0.25,
|
|
1221
|
+
outputPrice: 0.35,
|
|
1222
|
+
description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
|
|
1223
|
+
},
|
|
1195
1224
|
"unsloth/Llama-3.3-70B-Instruct": {
|
|
1196
1225
|
maxTokens: 32768,
|
|
1197
1226
|
// From Groq
|
|
@@ -1401,6 +1430,15 @@ var chutesModels = {
|
|
|
1401
1430
|
outputPrice: 3,
|
|
1402
1431
|
description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
|
|
1403
1432
|
},
|
|
1433
|
+
"zai-org/GLM-4.6-FP8": {
|
|
1434
|
+
maxTokens: 32768,
|
|
1435
|
+
contextWindow: 202752,
|
|
1436
|
+
supportsImages: false,
|
|
1437
|
+
supportsPromptCache: false,
|
|
1438
|
+
inputPrice: 0,
|
|
1439
|
+
outputPrice: 0,
|
|
1440
|
+
description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
|
|
1441
|
+
},
|
|
1404
1442
|
"Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": {
|
|
1405
1443
|
maxTokens: 32768,
|
|
1406
1444
|
contextWindow: 262144,
|
|
@@ -1454,6 +1492,15 @@ var chutesModels = {
|
|
|
1454
1492
|
inputPrice: 0,
|
|
1455
1493
|
outputPrice: 0,
|
|
1456
1494
|
description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
|
|
1495
|
+
},
|
|
1496
|
+
"Qwen/Qwen3-VL-235B-A22B-Thinking": {
|
|
1497
|
+
maxTokens: 262144,
|
|
1498
|
+
contextWindow: 262144,
|
|
1499
|
+
supportsImages: true,
|
|
1500
|
+
supportsPromptCache: false,
|
|
1501
|
+
inputPrice: 0.16,
|
|
1502
|
+
outputPrice: 0.65,
|
|
1503
|
+
description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
|
|
1457
1504
|
}
|
|
1458
1505
|
};
|
|
1459
1506
|
|
|
@@ -2923,7 +2970,8 @@ var rooModels = {
|
|
|
2923
2970
|
supportsPromptCache: false,
|
|
2924
2971
|
inputPrice: 0,
|
|
2925
2972
|
outputPrice: 0,
|
|
2926
|
-
description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)"
|
|
2973
|
+
description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)",
|
|
2974
|
+
deprecated: true
|
|
2927
2975
|
},
|
|
2928
2976
|
"deepseek/deepseek-chat-v3.1": {
|
|
2929
2977
|
maxTokens: 16384,
|
|
@@ -4472,7 +4520,7 @@ var MODELS_BY_PROVIDER = {
|
|
|
4472
4520
|
},
|
|
4473
4521
|
xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
|
|
4474
4522
|
zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
|
|
4475
|
-
// Dynamic providers; models pulled from
|
|
4523
|
+
// Dynamic providers; models pulled from remote APIs.
|
|
4476
4524
|
glama: { id: "glama", label: "Glama", models: [] },
|
|
4477
4525
|
huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
|
|
4478
4526
|
litellm: { id: "litellm", label: "LiteLLM", models: [] },
|
|
@@ -4480,7 +4528,10 @@ var MODELS_BY_PROVIDER = {
|
|
|
4480
4528
|
requesty: { id: "requesty", label: "Requesty", models: [] },
|
|
4481
4529
|
unbound: { id: "unbound", label: "Unbound", models: [] },
|
|
4482
4530
|
deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
|
|
4483
|
-
"vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
|
|
4531
|
+
"vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] },
|
|
4532
|
+
// Local providers; models discovered from localhost endpoints.
|
|
4533
|
+
lmstudio: { id: "lmstudio", label: "LM Studio", models: [] },
|
|
4534
|
+
ollama: { id: "ollama", label: "Ollama", models: [] }
|
|
4484
4535
|
};
|
|
4485
4536
|
|
|
4486
4537
|
// src/history.ts
|
|
@@ -4591,6 +4642,7 @@ var appPropertiesSchema = z10.object({
|
|
|
4591
4642
|
});
|
|
4592
4643
|
var taskPropertiesSchema = z10.object({
|
|
4593
4644
|
taskId: z10.string().optional(),
|
|
4645
|
+
parentTaskId: z10.string().optional(),
|
|
4594
4646
|
apiProvider: z10.enum(providerNames).optional(),
|
|
4595
4647
|
modelId: z10.string().optional(),
|
|
4596
4648
|
diffStrategy: z10.string().optional(),
|