@roo-code/types 1.99.0 → 1.102.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -1200,39 +1200,6 @@ var bedrockModels = {
1200
1200
  inputPrice: 0.25,
1201
1201
  outputPrice: 1.25
1202
1202
  },
1203
- "anthropic.claude-2-1-v1:0": {
1204
- maxTokens: 4096,
1205
- contextWindow: 1e5,
1206
- supportsImages: false,
1207
- supportsPromptCache: false,
1208
- supportsNativeTools: true,
1209
- defaultToolProtocol: "native",
1210
- inputPrice: 8,
1211
- outputPrice: 24,
1212
- description: "Claude 2.1"
1213
- },
1214
- "anthropic.claude-2-0-v1:0": {
1215
- maxTokens: 4096,
1216
- contextWindow: 1e5,
1217
- supportsImages: false,
1218
- supportsPromptCache: false,
1219
- supportsNativeTools: true,
1220
- defaultToolProtocol: "native",
1221
- inputPrice: 8,
1222
- outputPrice: 24,
1223
- description: "Claude 2.0"
1224
- },
1225
- "anthropic.claude-instant-v1:0": {
1226
- maxTokens: 4096,
1227
- contextWindow: 1e5,
1228
- supportsImages: false,
1229
- supportsPromptCache: false,
1230
- supportsNativeTools: true,
1231
- defaultToolProtocol: "native",
1232
- inputPrice: 0.8,
1233
- outputPrice: 2.4,
1234
- description: "Claude Instant"
1235
- },
1236
1203
  "deepseek.r1-v1:0": {
1237
1204
  maxTokens: 32768,
1238
1205
  contextWindow: 128e3,
@@ -1532,7 +1499,19 @@ var cerebrasModels = {
1532
1499
  defaultToolProtocol: "native",
1533
1500
  inputPrice: 0,
1534
1501
  outputPrice: 0,
1535
- description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
1502
+ description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
1503
+ },
1504
+ "zai-glm-4.7": {
1505
+ maxTokens: 16384,
1506
+ // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1507
+ contextWindow: 131072,
1508
+ supportsImages: false,
1509
+ supportsPromptCache: false,
1510
+ supportsNativeTools: true,
1511
+ defaultToolProtocol: "native",
1512
+ inputPrice: 0,
1513
+ outputPrice: 0,
1514
+ description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
1536
1515
  },
1537
1516
  "qwen-3-235b-a22b-instruct-2507": {
1538
1517
  maxTokens: 16384,
@@ -2283,6 +2262,20 @@ var fireworksModels = {
2283
2262
  outputPrice: 2.5,
2284
2263
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
2285
2264
  },
2265
+ "accounts/fireworks/models/kimi-k2-thinking": {
2266
+ maxTokens: 16e3,
2267
+ contextWindow: 256e3,
2268
+ supportsImages: false,
2269
+ supportsPromptCache: true,
2270
+ supportsNativeTools: true,
2271
+ supportsTemperature: true,
2272
+ preserveReasoning: true,
2273
+ defaultTemperature: 1,
2274
+ inputPrice: 0.6,
2275
+ outputPrice: 2.5,
2276
+ cacheReadsPrice: 0.15,
2277
+ description: "The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems."
2278
+ },
2286
2279
  "accounts/fireworks/models/minimax-m2": {
2287
2280
  maxTokens: 4096,
2288
2281
  contextWindow: 204800,
@@ -2418,8 +2411,6 @@ var geminiModels = {
2418
2411
  supportsPromptCache: true,
2419
2412
  supportsReasoningEffort: ["low", "high"],
2420
2413
  reasoningEffort: "low",
2421
- includedTools: ["write_file", "edit_file"],
2422
- excludedTools: ["apply_diff"],
2423
2414
  supportsTemperature: true,
2424
2415
  defaultTemperature: 1,
2425
2416
  inputPrice: 4,
@@ -2446,8 +2437,6 @@ var geminiModels = {
2446
2437
  supportsPromptCache: true,
2447
2438
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2448
2439
  reasoningEffort: "medium",
2449
- includedTools: ["write_file", "edit_file"],
2450
- excludedTools: ["apply_diff"],
2451
2440
  supportsTemperature: true,
2452
2441
  defaultTemperature: 1,
2453
2442
  inputPrice: 0.3,
@@ -2463,8 +2452,6 @@ var geminiModels = {
2463
2452
  supportsNativeTools: true,
2464
2453
  defaultToolProtocol: "native",
2465
2454
  supportsPromptCache: true,
2466
- includedTools: ["write_file", "edit_file"],
2467
- excludedTools: ["apply_diff"],
2468
2455
  inputPrice: 2.5,
2469
2456
  // This is the pricing for prompts above 200k tokens.
2470
2457
  outputPrice: 15,
@@ -2495,8 +2482,6 @@ var geminiModels = {
2495
2482
  supportsNativeTools: true,
2496
2483
  defaultToolProtocol: "native",
2497
2484
  supportsPromptCache: true,
2498
- includedTools: ["write_file", "edit_file"],
2499
- excludedTools: ["apply_diff"],
2500
2485
  inputPrice: 2.5,
2501
2486
  // This is the pricing for prompts above 200k tokens.
2502
2487
  outputPrice: 15,
@@ -2526,8 +2511,6 @@ var geminiModels = {
2526
2511
  supportsNativeTools: true,
2527
2512
  defaultToolProtocol: "native",
2528
2513
  supportsPromptCache: true,
2529
- includedTools: ["write_file", "edit_file"],
2530
- excludedTools: ["apply_diff"],
2531
2514
  inputPrice: 2.5,
2532
2515
  // This is the pricing for prompts above 200k tokens.
2533
2516
  outputPrice: 15,
@@ -2555,8 +2538,6 @@ var geminiModels = {
2555
2538
  supportsNativeTools: true,
2556
2539
  defaultToolProtocol: "native",
2557
2540
  supportsPromptCache: true,
2558
- includedTools: ["write_file", "edit_file"],
2559
- excludedTools: ["apply_diff"],
2560
2541
  inputPrice: 2.5,
2561
2542
  // This is the pricing for prompts above 200k tokens.
2562
2543
  outputPrice: 15,
@@ -2587,8 +2568,6 @@ var geminiModels = {
2587
2568
  supportsNativeTools: true,
2588
2569
  defaultToolProtocol: "native",
2589
2570
  supportsPromptCache: true,
2590
- includedTools: ["write_file", "edit_file"],
2591
- excludedTools: ["apply_diff"],
2592
2571
  inputPrice: 0.3,
2593
2572
  outputPrice: 2.5,
2594
2573
  cacheReadsPrice: 0.075,
@@ -2603,8 +2582,6 @@ var geminiModels = {
2603
2582
  supportsNativeTools: true,
2604
2583
  defaultToolProtocol: "native",
2605
2584
  supportsPromptCache: true,
2606
- includedTools: ["write_file", "edit_file"],
2607
- excludedTools: ["apply_diff"],
2608
2585
  inputPrice: 0.3,
2609
2586
  outputPrice: 2.5,
2610
2587
  cacheReadsPrice: 0.075,
@@ -2619,8 +2596,6 @@ var geminiModels = {
2619
2596
  supportsNativeTools: true,
2620
2597
  defaultToolProtocol: "native",
2621
2598
  supportsPromptCache: true,
2622
- includedTools: ["write_file", "edit_file"],
2623
- excludedTools: ["apply_diff"],
2624
2599
  inputPrice: 0.3,
2625
2600
  outputPrice: 2.5,
2626
2601
  cacheReadsPrice: 0.075,
@@ -2636,8 +2611,6 @@ var geminiModels = {
2636
2611
  supportsNativeTools: true,
2637
2612
  defaultToolProtocol: "native",
2638
2613
  supportsPromptCache: true,
2639
- includedTools: ["write_file", "edit_file"],
2640
- excludedTools: ["apply_diff"],
2641
2614
  inputPrice: 0.1,
2642
2615
  outputPrice: 0.4,
2643
2616
  cacheReadsPrice: 0.025,
@@ -2652,8 +2625,6 @@ var geminiModels = {
2652
2625
  supportsNativeTools: true,
2653
2626
  defaultToolProtocol: "native",
2654
2627
  supportsPromptCache: true,
2655
- includedTools: ["write_file", "edit_file"],
2656
- excludedTools: ["apply_diff"],
2657
2628
  inputPrice: 0.1,
2658
2629
  outputPrice: 0.4,
2659
2630
  cacheReadsPrice: 0.025,
@@ -3846,8 +3817,6 @@ var vertexModels = {
3846
3817
  supportsPromptCache: true,
3847
3818
  supportsReasoningEffort: ["low", "high"],
3848
3819
  reasoningEffort: "low",
3849
- includedTools: ["write_file", "edit_file"],
3850
- excludedTools: ["apply_diff"],
3851
3820
  supportsTemperature: true,
3852
3821
  defaultTemperature: 1,
3853
3822
  inputPrice: 4,
@@ -3874,8 +3843,6 @@ var vertexModels = {
3874
3843
  supportsPromptCache: true,
3875
3844
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3876
3845
  reasoningEffort: "medium",
3877
- includedTools: ["write_file", "edit_file"],
3878
- excludedTools: ["apply_diff"],
3879
3846
  supportsTemperature: true,
3880
3847
  defaultTemperature: 1,
3881
3848
  inputPrice: 0.3,
@@ -3890,8 +3857,6 @@ var vertexModels = {
3890
3857
  supportsNativeTools: true,
3891
3858
  defaultToolProtocol: "native",
3892
3859
  supportsPromptCache: true,
3893
- includedTools: ["write_file", "edit_file"],
3894
- excludedTools: ["apply_diff"],
3895
3860
  inputPrice: 0.15,
3896
3861
  outputPrice: 3.5,
3897
3862
  maxThinkingTokens: 24576,
@@ -3905,8 +3870,6 @@ var vertexModels = {
3905
3870
  supportsNativeTools: true,
3906
3871
  defaultToolProtocol: "native",
3907
3872
  supportsPromptCache: true,
3908
- includedTools: ["write_file", "edit_file"],
3909
- excludedTools: ["apply_diff"],
3910
3873
  inputPrice: 0.15,
3911
3874
  outputPrice: 0.6
3912
3875
  },
@@ -3917,8 +3880,6 @@ var vertexModels = {
3917
3880
  supportsNativeTools: true,
3918
3881
  defaultToolProtocol: "native",
3919
3882
  supportsPromptCache: true,
3920
- includedTools: ["write_file", "edit_file"],
3921
- excludedTools: ["apply_diff"],
3922
3883
  inputPrice: 0.3,
3923
3884
  outputPrice: 2.5,
3924
3885
  cacheReadsPrice: 0.075,
@@ -3933,8 +3894,6 @@ var vertexModels = {
3933
3894
  supportsNativeTools: true,
3934
3895
  defaultToolProtocol: "native",
3935
3896
  supportsPromptCache: false,
3936
- includedTools: ["write_file", "edit_file"],
3937
- excludedTools: ["apply_diff"],
3938
3897
  inputPrice: 0.15,
3939
3898
  outputPrice: 3.5,
3940
3899
  maxThinkingTokens: 24576,
@@ -3948,8 +3907,6 @@ var vertexModels = {
3948
3907
  supportsNativeTools: true,
3949
3908
  defaultToolProtocol: "native",
3950
3909
  supportsPromptCache: false,
3951
- includedTools: ["write_file", "edit_file"],
3952
- excludedTools: ["apply_diff"],
3953
3910
  inputPrice: 0.15,
3954
3911
  outputPrice: 0.6
3955
3912
  },
@@ -3960,8 +3917,6 @@ var vertexModels = {
3960
3917
  supportsNativeTools: true,
3961
3918
  defaultToolProtocol: "native",
3962
3919
  supportsPromptCache: true,
3963
- includedTools: ["write_file", "edit_file"],
3964
- excludedTools: ["apply_diff"],
3965
3920
  inputPrice: 2.5,
3966
3921
  outputPrice: 15
3967
3922
  },
@@ -3972,8 +3927,6 @@ var vertexModels = {
3972
3927
  supportsNativeTools: true,
3973
3928
  defaultToolProtocol: "native",
3974
3929
  supportsPromptCache: true,
3975
- includedTools: ["write_file", "edit_file"],
3976
- excludedTools: ["apply_diff"],
3977
3930
  inputPrice: 2.5,
3978
3931
  outputPrice: 15
3979
3932
  },
@@ -3984,8 +3937,6 @@ var vertexModels = {
3984
3937
  supportsNativeTools: true,
3985
3938
  defaultToolProtocol: "native",
3986
3939
  supportsPromptCache: true,
3987
- includedTools: ["write_file", "edit_file"],
3988
- excludedTools: ["apply_diff"],
3989
3940
  inputPrice: 2.5,
3990
3941
  outputPrice: 15,
3991
3942
  maxThinkingTokens: 32768,
@@ -3998,8 +3949,6 @@ var vertexModels = {
3998
3949
  supportsNativeTools: true,
3999
3950
  defaultToolProtocol: "native",
4000
3951
  supportsPromptCache: true,
4001
- includedTools: ["write_file", "edit_file"],
4002
- excludedTools: ["apply_diff"],
4003
3952
  inputPrice: 2.5,
4004
3953
  outputPrice: 15,
4005
3954
  maxThinkingTokens: 32768,
@@ -4027,8 +3976,6 @@ var vertexModels = {
4027
3976
  supportsNativeTools: true,
4028
3977
  defaultToolProtocol: "native",
4029
3978
  supportsPromptCache: false,
4030
- includedTools: ["write_file", "edit_file"],
4031
- excludedTools: ["apply_diff"],
4032
3979
  inputPrice: 0,
4033
3980
  outputPrice: 0
4034
3981
  },
@@ -4039,8 +3986,6 @@ var vertexModels = {
4039
3986
  supportsNativeTools: true,
4040
3987
  defaultToolProtocol: "native",
4041
3988
  supportsPromptCache: false,
4042
- includedTools: ["write_file", "edit_file"],
4043
- excludedTools: ["apply_diff"],
4044
3989
  inputPrice: 0,
4045
3990
  outputPrice: 0
4046
3991
  },
@@ -4051,8 +3996,6 @@ var vertexModels = {
4051
3996
  supportsNativeTools: true,
4052
3997
  defaultToolProtocol: "native",
4053
3998
  supportsPromptCache: true,
4054
- includedTools: ["write_file", "edit_file"],
4055
- excludedTools: ["apply_diff"],
4056
3999
  inputPrice: 0.15,
4057
4000
  outputPrice: 0.6
4058
4001
  },
@@ -4063,8 +4006,6 @@ var vertexModels = {
4063
4006
  supportsNativeTools: true,
4064
4007
  defaultToolProtocol: "native",
4065
4008
  supportsPromptCache: false,
4066
- includedTools: ["write_file", "edit_file"],
4067
- excludedTools: ["apply_diff"],
4068
4009
  inputPrice: 0.075,
4069
4010
  outputPrice: 0.3
4070
4011
  },
@@ -4075,8 +4016,6 @@ var vertexModels = {
4075
4016
  supportsNativeTools: true,
4076
4017
  defaultToolProtocol: "native",
4077
4018
  supportsPromptCache: false,
4078
- includedTools: ["write_file", "edit_file"],
4079
- excludedTools: ["apply_diff"],
4080
4019
  inputPrice: 0,
4081
4020
  outputPrice: 0
4082
4021
  },
@@ -4087,8 +4026,6 @@ var vertexModels = {
4087
4026
  supportsNativeTools: true,
4088
4027
  defaultToolProtocol: "native",
4089
4028
  supportsPromptCache: true,
4090
- includedTools: ["write_file", "edit_file"],
4091
- excludedTools: ["apply_diff"],
4092
4029
  inputPrice: 0.075,
4093
4030
  outputPrice: 0.3
4094
4031
  },
@@ -4099,8 +4036,6 @@ var vertexModels = {
4099
4036
  supportsNativeTools: true,
4100
4037
  defaultToolProtocol: "native",
4101
4038
  supportsPromptCache: false,
4102
- includedTools: ["write_file", "edit_file"],
4103
- excludedTools: ["apply_diff"],
4104
4039
  inputPrice: 1.25,
4105
4040
  outputPrice: 5
4106
4041
  },
@@ -4314,8 +4249,6 @@ var vertexModels = {
4314
4249
  supportsNativeTools: true,
4315
4250
  defaultToolProtocol: "native",
4316
4251
  supportsPromptCache: true,
4317
- includedTools: ["write_file", "edit_file"],
4318
- excludedTools: ["apply_diff"],
4319
4252
  inputPrice: 0.1,
4320
4253
  outputPrice: 0.4,
4321
4254
  cacheReadsPrice: 0.025,
@@ -5376,7 +5309,6 @@ var vertexSchema = apiModelIdProviderModelSchema.extend({
5376
5309
  var openAiSchema = baseProviderSettingsSchema.extend({
5377
5310
  openAiBaseUrl: z8.string().optional(),
5378
5311
  openAiApiKey: z8.string().optional(),
5379
- openAiLegacyFormat: z8.boolean().optional(),
5380
5312
  openAiR1FormatEnabled: z8.boolean().optional(),
5381
5313
  openAiModelId: z8.string().optional(),
5382
5314
  openAiCustomModelInfo: modelInfoSchema.nullish(),
@@ -5506,7 +5438,8 @@ var qwenCodeSchema = apiModelIdProviderModelSchema.extend({
5506
5438
  qwenCodeOauthPath: z8.string().optional()
5507
5439
  });
5508
5440
  var rooSchema = apiModelIdProviderModelSchema.extend({
5509
- // No additional fields needed - uses cloud authentication.
5441
+ // Can use cloud authentication or provide an API key (cli).
5442
+ rooApiKey: z8.string().optional()
5510
5443
  });
5511
5444
  var vercelAiGatewaySchema = baseProviderSettingsSchema.extend({
5512
5445
  vercelAiGatewayApiKey: z8.string().optional(),
@@ -5795,6 +5728,8 @@ var historyItemSchema = z9.object({
5795
5728
  * This ensures task resumption works correctly even when NTC settings change.
5796
5729
  */
5797
5730
  toolProtocol: z9.enum(["xml", "native"]).optional(),
5731
+ apiConfigName: z9.string().optional(),
5732
+ // Provider profile name for sticky profile feature
5798
5733
  status: z9.enum(["active", "completed", "delegated"]).optional(),
5799
5734
  delegatedToId: z9.string().optional(),
5800
5735
  // Last child this parent delegated to
@@ -7179,6 +7114,36 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
7179
7114
  status: z22.literal("timeout")
7180
7115
  })
7181
7116
  ]);
7117
+
7118
+ // src/vscode-extension-host.ts
7119
+ import { z as z23 } from "zod";
7120
+ var checkoutDiffPayloadSchema = z23.object({
7121
+ ts: z23.number().optional(),
7122
+ previousCommitHash: z23.string().optional(),
7123
+ commitHash: z23.string(),
7124
+ mode: z23.enum(["full", "checkpoint", "from-init", "to-current"])
7125
+ });
7126
+ var checkoutRestorePayloadSchema = z23.object({
7127
+ ts: z23.number(),
7128
+ commitHash: z23.string(),
7129
+ mode: z23.enum(["preview", "restore"])
7130
+ });
7131
+ var installMarketplaceItemWithParametersPayloadSchema = z23.object({
7132
+ item: marketplaceItemSchema,
7133
+ parameters: z23.record(z23.string(), z23.any())
7134
+ });
7135
+ var browserActions = [
7136
+ "launch",
7137
+ "click",
7138
+ "hover",
7139
+ "type",
7140
+ "press",
7141
+ "scroll_down",
7142
+ "scroll_up",
7143
+ "resize",
7144
+ "close",
7145
+ "screenshot"
7146
+ ];
7182
7147
  export {
7183
7148
  ANTHROPIC_DEFAULT_MAX_TOKENS,
7184
7149
  ANTHROPIC_STYLE_PROVIDERS,
@@ -7277,8 +7242,11 @@ export {
7277
7242
  bedrockDefaultModelId,
7278
7243
  bedrockDefaultPromptRouterModelId,
7279
7244
  bedrockModels,
7245
+ browserActions,
7280
7246
  cerebrasDefaultModelId,
7281
7247
  cerebrasModels,
7248
+ checkoutDiffPayloadSchema,
7249
+ checkoutRestorePayloadSchema,
7282
7250
  chutesDefaultModelId,
7283
7251
  chutesDefaultModelInfo,
7284
7252
  chutesModels,
@@ -7347,6 +7315,7 @@ export {
7347
7315
  historyItemSchema,
7348
7316
  idleAsks,
7349
7317
  installMarketplaceItemOptionsSchema,
7318
+ installMarketplaceItemWithParametersPayloadSchema,
7350
7319
  interactiveAsks,
7351
7320
  internalProviders,
7352
7321
  internationalZAiDefaultModelId,