@roo-code/types 1.22.0 → 1.24.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -20,18 +20,44 @@ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: tru
20
20
  // src/index.ts
21
21
  var index_exports = {};
22
22
  __export(index_exports, {
23
+ ANTHROPIC_DEFAULT_MAX_TOKENS: () => ANTHROPIC_DEFAULT_MAX_TOKENS,
24
+ BEDROCK_DEFAULT_TEMPERATURE: () => BEDROCK_DEFAULT_TEMPERATURE,
25
+ BEDROCK_MAX_TOKENS: () => BEDROCK_MAX_TOKENS,
26
+ BEDROCK_REGIONS: () => BEDROCK_REGIONS,
27
+ BEDROCK_REGION_INFO: () => BEDROCK_REGION_INFO,
28
+ DEEP_SEEK_DEFAULT_TEMPERATURE: () => DEEP_SEEK_DEFAULT_TEMPERATURE,
29
+ GLAMA_DEFAULT_TEMPERATURE: () => GLAMA_DEFAULT_TEMPERATURE,
23
30
  GLOBAL_SETTINGS_KEYS: () => GLOBAL_SETTINGS_KEYS,
24
31
  GLOBAL_STATE_KEYS: () => GLOBAL_STATE_KEYS,
25
32
  IpcMessageType: () => IpcMessageType,
26
33
  IpcOrigin: () => IpcOrigin,
34
+ LITELLM_COMPUTER_USE_MODELS: () => LITELLM_COMPUTER_USE_MODELS,
35
+ LMSTUDIO_DEFAULT_TEMPERATURE: () => LMSTUDIO_DEFAULT_TEMPERATURE,
36
+ MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
37
+ OPENAI_AZURE_AI_INFERENCE_PATH: () => OPENAI_AZURE_AI_INFERENCE_PATH,
38
+ OPENAI_NATIVE_DEFAULT_TEMPERATURE: () => OPENAI_NATIVE_DEFAULT_TEMPERATURE,
39
+ OPENROUTER_DEFAULT_PROVIDER_NAME: () => OPENROUTER_DEFAULT_PROVIDER_NAME,
40
+ OPEN_ROUTER_COMPUTER_USE_MODELS: () => OPEN_ROUTER_COMPUTER_USE_MODELS,
41
+ OPEN_ROUTER_PROMPT_CACHING_MODELS: () => OPEN_ROUTER_PROMPT_CACHING_MODELS,
42
+ OPEN_ROUTER_REASONING_BUDGET_MODELS: () => OPEN_ROUTER_REASONING_BUDGET_MODELS,
43
+ OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS: () => OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS,
27
44
  ORGANIZATION_ALLOW_ALL: () => ORGANIZATION_ALLOW_ALL,
28
45
  PROVIDER_SETTINGS_KEYS: () => PROVIDER_SETTINGS_KEYS,
29
46
  RooCodeEventName: () => RooCodeEventName,
30
47
  SECRET_STATE_KEYS: () => SECRET_STATE_KEYS,
31
48
  TaskCommandName: () => TaskCommandName,
32
49
  TelemetryEventName: () => TelemetryEventName,
50
+ VERTEX_REGIONS: () => VERTEX_REGIONS,
33
51
  ackSchema: () => ackSchema,
52
+ anthropicDefaultModelId: () => anthropicDefaultModelId,
53
+ anthropicModels: () => anthropicModels,
34
54
  appPropertiesSchema: () => appPropertiesSchema,
55
+ azureOpenAiDefaultApiVersion: () => azureOpenAiDefaultApiVersion,
56
+ bedrockDefaultModelId: () => bedrockDefaultModelId,
57
+ bedrockDefaultPromptRouterModelId: () => bedrockDefaultPromptRouterModelId,
58
+ bedrockModels: () => bedrockModels,
59
+ chutesDefaultModelId: () => chutesDefaultModelId,
60
+ chutesModels: () => chutesModels,
35
61
  clineAskSchema: () => clineAskSchema,
36
62
  clineAsks: () => clineAsks,
37
63
  clineMessageSchema: () => clineMessageSchema,
@@ -47,10 +73,18 @@ __export(index_exports, {
47
73
  customModePromptsSchema: () => customModePromptsSchema,
48
74
  customModesSettingsSchema: () => customModesSettingsSchema,
49
75
  customSupportPromptsSchema: () => customSupportPromptsSchema,
76
+ deepSeekDefaultModelId: () => deepSeekDefaultModelId,
77
+ deepSeekModels: () => deepSeekModels,
50
78
  experimentIds: () => experimentIds,
51
79
  experimentIdsSchema: () => experimentIdsSchema,
52
80
  experimentsSchema: () => experimentsSchema,
81
+ geminiDefaultModelId: () => geminiDefaultModelId,
82
+ geminiModels: () => geminiModels,
83
+ glamaDefaultModelId: () => glamaDefaultModelId,
84
+ glamaDefaultModelInfo: () => glamaDefaultModelInfo,
53
85
  globalSettingsSchema: () => globalSettingsSchema,
86
+ groqDefaultModelId: () => groqDefaultModelId,
87
+ groqModels: () => groqModels,
54
88
  groupEntrySchema: () => groupEntrySchema,
55
89
  groupOptionsSchema: () => groupOptionsSchema,
56
90
  historyItemSchema: () => historyItemSchema,
@@ -62,10 +96,19 @@ __export(index_exports, {
62
96
  keysOf: () => keysOf,
63
97
  languages: () => languages,
64
98
  languagesSchema: () => languagesSchema,
99
+ litellmDefaultModelId: () => litellmDefaultModelId,
100
+ litellmDefaultModelInfo: () => litellmDefaultModelInfo,
101
+ mistralDefaultModelId: () => mistralDefaultModelId,
102
+ mistralModels: () => mistralModels,
65
103
  modeConfigSchema: () => modeConfigSchema,
66
104
  modelInfoSchema: () => modelInfoSchema,
67
105
  modelParameters: () => modelParameters,
68
106
  modelParametersSchema: () => modelParametersSchema,
107
+ openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
108
+ openAiNativeDefaultModelId: () => openAiNativeDefaultModelId,
109
+ openAiNativeModels: () => openAiNativeModels,
110
+ openRouterDefaultModelId: () => openRouterDefaultModelId,
111
+ openRouterDefaultModelInfo: () => openRouterDefaultModelInfo,
69
112
  organizationAllowListSchema: () => organizationAllowListSchema,
70
113
  organizationSettingsSchema: () => organizationSettingsSchema,
71
114
  promptComponentSchema: () => promptComponentSchema,
@@ -76,6 +119,8 @@ __export(index_exports, {
76
119
  providerSettingsSchemaDiscriminated: () => providerSettingsSchemaDiscriminated,
77
120
  reasoningEfforts: () => reasoningEfforts,
78
121
  reasoningEffortsSchema: () => reasoningEffortsSchema,
122
+ requestyDefaultModelId: () => requestyDefaultModelId,
123
+ requestyDefaultModelInfo: () => requestyDefaultModelInfo,
79
124
  rooCodeEventsSchema: () => rooCodeEventsSchema,
80
125
  rooCodeSettingsSchema: () => rooCodeSettingsSchema,
81
126
  rooCodeTelemetryEventSchema: () => rooCodeTelemetryEventSchema,
@@ -92,10 +137,2002 @@ __export(index_exports, {
92
137
  toolNames: () => toolNames,
93
138
  toolNamesSchema: () => toolNamesSchema,
94
139
  toolProgressStatusSchema: () => toolProgressStatusSchema,
95
- toolUsageSchema: () => toolUsageSchema
140
+ toolUsageSchema: () => toolUsageSchema,
141
+ unboundDefaultModelId: () => unboundDefaultModelId,
142
+ unboundDefaultModelInfo: () => unboundDefaultModelInfo,
143
+ vertexDefaultModelId: () => vertexDefaultModelId,
144
+ vertexModels: () => vertexModels,
145
+ vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
146
+ vscodeLlmModels: () => vscodeLlmModels,
147
+ xaiDefaultModelId: () => xaiDefaultModelId,
148
+ xaiModels: () => xaiModels
96
149
  });
97
150
  module.exports = __toCommonJS(index_exports);
98
151
 
152
+ // src/providers/anthropic.ts
153
+ var anthropicDefaultModelId = "claude-sonnet-4-20250514";
154
+ var anthropicModels = {
155
+ "claude-sonnet-4-20250514": {
156
+ maxTokens: 64e3,
157
+ // Overridden to 8k if `enableReasoningEffort` is false.
158
+ contextWindow: 2e5,
159
+ supportsImages: true,
160
+ supportsComputerUse: true,
161
+ supportsPromptCache: true,
162
+ inputPrice: 3,
163
+ // $3 per million input tokens
164
+ outputPrice: 15,
165
+ // $15 per million output tokens
166
+ cacheWritesPrice: 3.75,
167
+ // $3.75 per million tokens
168
+ cacheReadsPrice: 0.3,
169
+ // $0.30 per million tokens
170
+ supportsReasoningBudget: true
171
+ },
172
+ "claude-opus-4-20250514": {
173
+ maxTokens: 32e3,
174
+ // Overridden to 8k if `enableReasoningEffort` is false.
175
+ contextWindow: 2e5,
176
+ supportsImages: true,
177
+ supportsComputerUse: true,
178
+ supportsPromptCache: true,
179
+ inputPrice: 15,
180
+ // $15 per million input tokens
181
+ outputPrice: 75,
182
+ // $75 per million output tokens
183
+ cacheWritesPrice: 18.75,
184
+ // $18.75 per million tokens
185
+ cacheReadsPrice: 1.5,
186
+ // $1.50 per million tokens
187
+ supportsReasoningBudget: true
188
+ },
189
+ "claude-3-7-sonnet-20250219:thinking": {
190
+ maxTokens: 128e3,
191
+ // Unlocked by passing `beta` flag to the model. Otherwise, it's 64k.
192
+ contextWindow: 2e5,
193
+ supportsImages: true,
194
+ supportsComputerUse: true,
195
+ supportsPromptCache: true,
196
+ inputPrice: 3,
197
+ // $3 per million input tokens
198
+ outputPrice: 15,
199
+ // $15 per million output tokens
200
+ cacheWritesPrice: 3.75,
201
+ // $3.75 per million tokens
202
+ cacheReadsPrice: 0.3,
203
+ // $0.30 per million tokens
204
+ supportsReasoningBudget: true,
205
+ requiredReasoningBudget: true
206
+ },
207
+ "claude-3-7-sonnet-20250219": {
208
+ maxTokens: 8192,
209
+ // Since we already have a `:thinking` virtual model we aren't setting `supportsReasoningBudget: true` here.
210
+ contextWindow: 2e5,
211
+ supportsImages: true,
212
+ supportsComputerUse: true,
213
+ supportsPromptCache: true,
214
+ inputPrice: 3,
215
+ // $3 per million input tokens
216
+ outputPrice: 15,
217
+ // $15 per million output tokens
218
+ cacheWritesPrice: 3.75,
219
+ // $3.75 per million tokens
220
+ cacheReadsPrice: 0.3
221
+ // $0.30 per million tokens
222
+ },
223
+ "claude-3-5-sonnet-20241022": {
224
+ maxTokens: 8192,
225
+ contextWindow: 2e5,
226
+ supportsImages: true,
227
+ supportsComputerUse: true,
228
+ supportsPromptCache: true,
229
+ inputPrice: 3,
230
+ // $3 per million input tokens
231
+ outputPrice: 15,
232
+ // $15 per million output tokens
233
+ cacheWritesPrice: 3.75,
234
+ // $3.75 per million tokens
235
+ cacheReadsPrice: 0.3
236
+ // $0.30 per million tokens
237
+ },
238
+ "claude-3-5-haiku-20241022": {
239
+ maxTokens: 8192,
240
+ contextWindow: 2e5,
241
+ supportsImages: false,
242
+ supportsPromptCache: true,
243
+ inputPrice: 1,
244
+ outputPrice: 5,
245
+ cacheWritesPrice: 1.25,
246
+ cacheReadsPrice: 0.1
247
+ },
248
+ "claude-3-opus-20240229": {
249
+ maxTokens: 4096,
250
+ contextWindow: 2e5,
251
+ supportsImages: true,
252
+ supportsPromptCache: true,
253
+ inputPrice: 15,
254
+ outputPrice: 75,
255
+ cacheWritesPrice: 18.75,
256
+ cacheReadsPrice: 1.5
257
+ },
258
+ "claude-3-haiku-20240307": {
259
+ maxTokens: 4096,
260
+ contextWindow: 2e5,
261
+ supportsImages: true,
262
+ supportsPromptCache: true,
263
+ inputPrice: 0.25,
264
+ outputPrice: 1.25,
265
+ cacheWritesPrice: 0.3,
266
+ cacheReadsPrice: 0.03
267
+ }
268
+ };
269
+ var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
270
+
271
+ // src/providers/bedrock.ts
272
+ var bedrockDefaultModelId = "anthropic.claude-sonnet-4-20250514-v1:0";
273
+ var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
274
+ var bedrockModels = {
275
+ "amazon.nova-pro-v1:0": {
276
+ maxTokens: 5e3,
277
+ contextWindow: 3e5,
278
+ supportsImages: true,
279
+ supportsComputerUse: false,
280
+ supportsPromptCache: true,
281
+ inputPrice: 0.8,
282
+ outputPrice: 3.2,
283
+ cacheWritesPrice: 0.8,
284
+ // per million tokens
285
+ cacheReadsPrice: 0.2,
286
+ // per million tokens
287
+ minTokensPerCachePoint: 1,
288
+ maxCachePoints: 1,
289
+ cachableFields: ["system"]
290
+ },
291
+ "amazon.nova-pro-latency-optimized-v1:0": {
292
+ maxTokens: 5e3,
293
+ contextWindow: 3e5,
294
+ supportsImages: true,
295
+ supportsComputerUse: false,
296
+ supportsPromptCache: false,
297
+ inputPrice: 1,
298
+ outputPrice: 4,
299
+ cacheWritesPrice: 1,
300
+ // per million tokens
301
+ cacheReadsPrice: 0.25,
302
+ // per million tokens
303
+ description: "Amazon Nova Pro with latency optimized inference"
304
+ },
305
+ "amazon.nova-lite-v1:0": {
306
+ maxTokens: 5e3,
307
+ contextWindow: 3e5,
308
+ supportsImages: true,
309
+ supportsComputerUse: false,
310
+ supportsPromptCache: true,
311
+ inputPrice: 0.06,
312
+ outputPrice: 0.24,
313
+ cacheWritesPrice: 0.06,
314
+ // per million tokens
315
+ cacheReadsPrice: 0.015,
316
+ // per million tokens
317
+ minTokensPerCachePoint: 1,
318
+ maxCachePoints: 1,
319
+ cachableFields: ["system"]
320
+ },
321
+ "amazon.nova-micro-v1:0": {
322
+ maxTokens: 5e3,
323
+ contextWindow: 128e3,
324
+ supportsImages: false,
325
+ supportsComputerUse: false,
326
+ supportsPromptCache: true,
327
+ inputPrice: 0.035,
328
+ outputPrice: 0.14,
329
+ cacheWritesPrice: 0.035,
330
+ // per million tokens
331
+ cacheReadsPrice: 875e-5,
332
+ // per million tokens
333
+ minTokensPerCachePoint: 1,
334
+ maxCachePoints: 1,
335
+ cachableFields: ["system"]
336
+ },
337
+ "anthropic.claude-sonnet-4-20250514-v1:0": {
338
+ maxTokens: 8192,
339
+ contextWindow: 2e5,
340
+ supportsImages: true,
341
+ supportsComputerUse: true,
342
+ supportsPromptCache: true,
343
+ inputPrice: 3,
344
+ outputPrice: 15,
345
+ cacheWritesPrice: 3.75,
346
+ cacheReadsPrice: 0.3,
347
+ minTokensPerCachePoint: 1024,
348
+ maxCachePoints: 4,
349
+ cachableFields: ["system", "messages", "tools"]
350
+ },
351
+ "anthropic.claude-opus-4-20250514-v1:0": {
352
+ maxTokens: 8192,
353
+ contextWindow: 2e5,
354
+ supportsImages: true,
355
+ supportsComputerUse: true,
356
+ supportsPromptCache: true,
357
+ inputPrice: 15,
358
+ outputPrice: 75,
359
+ cacheWritesPrice: 18.75,
360
+ cacheReadsPrice: 1.5,
361
+ minTokensPerCachePoint: 1024,
362
+ maxCachePoints: 4,
363
+ cachableFields: ["system", "messages", "tools"]
364
+ },
365
+ "anthropic.claude-3-7-sonnet-20250219-v1:0": {
366
+ maxTokens: 8192,
367
+ contextWindow: 2e5,
368
+ supportsImages: true,
369
+ supportsComputerUse: true,
370
+ supportsPromptCache: true,
371
+ inputPrice: 3,
372
+ outputPrice: 15,
373
+ cacheWritesPrice: 3.75,
374
+ cacheReadsPrice: 0.3,
375
+ minTokensPerCachePoint: 1024,
376
+ maxCachePoints: 4,
377
+ cachableFields: ["system", "messages", "tools"]
378
+ },
379
+ "anthropic.claude-3-5-sonnet-20241022-v2:0": {
380
+ maxTokens: 8192,
381
+ contextWindow: 2e5,
382
+ supportsImages: true,
383
+ supportsComputerUse: true,
384
+ supportsPromptCache: true,
385
+ inputPrice: 3,
386
+ outputPrice: 15,
387
+ cacheWritesPrice: 3.75,
388
+ cacheReadsPrice: 0.3,
389
+ minTokensPerCachePoint: 1024,
390
+ maxCachePoints: 4,
391
+ cachableFields: ["system", "messages", "tools"]
392
+ },
393
+ "anthropic.claude-3-5-haiku-20241022-v1:0": {
394
+ maxTokens: 8192,
395
+ contextWindow: 2e5,
396
+ supportsImages: false,
397
+ supportsPromptCache: true,
398
+ inputPrice: 0.8,
399
+ outputPrice: 4,
400
+ cacheWritesPrice: 1,
401
+ cacheReadsPrice: 0.08,
402
+ minTokensPerCachePoint: 2048,
403
+ maxCachePoints: 4,
404
+ cachableFields: ["system", "messages", "tools"]
405
+ },
406
+ "anthropic.claude-3-5-sonnet-20240620-v1:0": {
407
+ maxTokens: 8192,
408
+ contextWindow: 2e5,
409
+ supportsImages: true,
410
+ supportsPromptCache: false,
411
+ inputPrice: 3,
412
+ outputPrice: 15
413
+ },
414
+ "anthropic.claude-3-opus-20240229-v1:0": {
415
+ maxTokens: 4096,
416
+ contextWindow: 2e5,
417
+ supportsImages: true,
418
+ supportsPromptCache: false,
419
+ inputPrice: 15,
420
+ outputPrice: 75
421
+ },
422
+ "anthropic.claude-3-sonnet-20240229-v1:0": {
423
+ maxTokens: 4096,
424
+ contextWindow: 2e5,
425
+ supportsImages: true,
426
+ supportsPromptCache: false,
427
+ inputPrice: 3,
428
+ outputPrice: 15
429
+ },
430
+ "anthropic.claude-3-haiku-20240307-v1:0": {
431
+ maxTokens: 4096,
432
+ contextWindow: 2e5,
433
+ supportsImages: true,
434
+ supportsPromptCache: false,
435
+ inputPrice: 0.25,
436
+ outputPrice: 1.25
437
+ },
438
+ "anthropic.claude-2-1-v1:0": {
439
+ maxTokens: 4096,
440
+ contextWindow: 1e5,
441
+ supportsImages: false,
442
+ supportsPromptCache: false,
443
+ inputPrice: 8,
444
+ outputPrice: 24,
445
+ description: "Claude 2.1"
446
+ },
447
+ "anthropic.claude-2-0-v1:0": {
448
+ maxTokens: 4096,
449
+ contextWindow: 1e5,
450
+ supportsImages: false,
451
+ supportsPromptCache: false,
452
+ inputPrice: 8,
453
+ outputPrice: 24,
454
+ description: "Claude 2.0"
455
+ },
456
+ "anthropic.claude-instant-v1:0": {
457
+ maxTokens: 4096,
458
+ contextWindow: 1e5,
459
+ supportsImages: false,
460
+ supportsPromptCache: false,
461
+ inputPrice: 0.8,
462
+ outputPrice: 2.4,
463
+ description: "Claude Instant"
464
+ },
465
+ "deepseek.r1-v1:0": {
466
+ maxTokens: 32768,
467
+ contextWindow: 128e3,
468
+ supportsImages: false,
469
+ supportsPromptCache: false,
470
+ inputPrice: 1.35,
471
+ outputPrice: 5.4
472
+ },
473
+ "meta.llama3-3-70b-instruct-v1:0": {
474
+ maxTokens: 8192,
475
+ contextWindow: 128e3,
476
+ supportsImages: false,
477
+ supportsComputerUse: false,
478
+ supportsPromptCache: false,
479
+ inputPrice: 0.72,
480
+ outputPrice: 0.72,
481
+ description: "Llama 3.3 Instruct (70B)"
482
+ },
483
+ "meta.llama3-2-90b-instruct-v1:0": {
484
+ maxTokens: 8192,
485
+ contextWindow: 128e3,
486
+ supportsImages: true,
487
+ supportsComputerUse: false,
488
+ supportsPromptCache: false,
489
+ inputPrice: 0.72,
490
+ outputPrice: 0.72,
491
+ description: "Llama 3.2 Instruct (90B)"
492
+ },
493
+ "meta.llama3-2-11b-instruct-v1:0": {
494
+ maxTokens: 8192,
495
+ contextWindow: 128e3,
496
+ supportsImages: true,
497
+ supportsComputerUse: false,
498
+ supportsPromptCache: false,
499
+ inputPrice: 0.16,
500
+ outputPrice: 0.16,
501
+ description: "Llama 3.2 Instruct (11B)"
502
+ },
503
+ "meta.llama3-2-3b-instruct-v1:0": {
504
+ maxTokens: 8192,
505
+ contextWindow: 128e3,
506
+ supportsImages: false,
507
+ supportsComputerUse: false,
508
+ supportsPromptCache: false,
509
+ inputPrice: 0.15,
510
+ outputPrice: 0.15,
511
+ description: "Llama 3.2 Instruct (3B)"
512
+ },
513
+ "meta.llama3-2-1b-instruct-v1:0": {
514
+ maxTokens: 8192,
515
+ contextWindow: 128e3,
516
+ supportsImages: false,
517
+ supportsComputerUse: false,
518
+ supportsPromptCache: false,
519
+ inputPrice: 0.1,
520
+ outputPrice: 0.1,
521
+ description: "Llama 3.2 Instruct (1B)"
522
+ },
523
+ "meta.llama3-1-405b-instruct-v1:0": {
524
+ maxTokens: 8192,
525
+ contextWindow: 128e3,
526
+ supportsImages: false,
527
+ supportsComputerUse: false,
528
+ supportsPromptCache: false,
529
+ inputPrice: 2.4,
530
+ outputPrice: 2.4,
531
+ description: "Llama 3.1 Instruct (405B)"
532
+ },
533
+ "meta.llama3-1-70b-instruct-v1:0": {
534
+ maxTokens: 8192,
535
+ contextWindow: 128e3,
536
+ supportsImages: false,
537
+ supportsComputerUse: false,
538
+ supportsPromptCache: false,
539
+ inputPrice: 0.72,
540
+ outputPrice: 0.72,
541
+ description: "Llama 3.1 Instruct (70B)"
542
+ },
543
+ "meta.llama3-1-70b-instruct-latency-optimized-v1:0": {
544
+ maxTokens: 8192,
545
+ contextWindow: 128e3,
546
+ supportsImages: false,
547
+ supportsComputerUse: false,
548
+ supportsPromptCache: false,
549
+ inputPrice: 0.9,
550
+ outputPrice: 0.9,
551
+ description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
552
+ },
553
+ "meta.llama3-1-8b-instruct-v1:0": {
554
+ maxTokens: 8192,
555
+ contextWindow: 8e3,
556
+ supportsImages: false,
557
+ supportsComputerUse: false,
558
+ supportsPromptCache: false,
559
+ inputPrice: 0.22,
560
+ outputPrice: 0.22,
561
+ description: "Llama 3.1 Instruct (8B)"
562
+ },
563
+ "meta.llama3-70b-instruct-v1:0": {
564
+ maxTokens: 2048,
565
+ contextWindow: 8e3,
566
+ supportsImages: false,
567
+ supportsComputerUse: false,
568
+ supportsPromptCache: false,
569
+ inputPrice: 2.65,
570
+ outputPrice: 3.5
571
+ },
572
+ "meta.llama3-8b-instruct-v1:0": {
573
+ maxTokens: 2048,
574
+ contextWindow: 4e3,
575
+ supportsImages: false,
576
+ supportsComputerUse: false,
577
+ supportsPromptCache: false,
578
+ inputPrice: 0.3,
579
+ outputPrice: 0.6
580
+ },
581
+ "amazon.titan-text-lite-v1:0": {
582
+ maxTokens: 4096,
583
+ contextWindow: 8e3,
584
+ supportsImages: false,
585
+ supportsComputerUse: false,
586
+ supportsPromptCache: false,
587
+ inputPrice: 0.15,
588
+ outputPrice: 0.2,
589
+ description: "Amazon Titan Text Lite"
590
+ },
591
+ "amazon.titan-text-express-v1:0": {
592
+ maxTokens: 4096,
593
+ contextWindow: 8e3,
594
+ supportsImages: false,
595
+ supportsComputerUse: false,
596
+ supportsPromptCache: false,
597
+ inputPrice: 0.2,
598
+ outputPrice: 0.6,
599
+ description: "Amazon Titan Text Express"
600
+ },
601
+ "amazon.titan-text-embeddings-v1:0": {
602
+ maxTokens: 8192,
603
+ contextWindow: 8e3,
604
+ supportsImages: false,
605
+ supportsComputerUse: false,
606
+ supportsPromptCache: false,
607
+ inputPrice: 0.1,
608
+ description: "Amazon Titan Text Embeddings"
609
+ },
610
+ "amazon.titan-text-embeddings-v2:0": {
611
+ maxTokens: 8192,
612
+ contextWindow: 8e3,
613
+ supportsImages: false,
614
+ supportsComputerUse: false,
615
+ supportsPromptCache: false,
616
+ inputPrice: 0.02,
617
+ description: "Amazon Titan Text Embeddings V2"
618
+ }
619
+ };
620
+ var BEDROCK_DEFAULT_TEMPERATURE = 0.3;
621
+ var BEDROCK_MAX_TOKENS = 4096;
622
+ var BEDROCK_REGION_INFO = {
623
+ /*
624
+ * This JSON generated by AWS's AI assistant - Amazon Q on March 29, 2025
625
+ *
626
+ * - Africa (Cape Town) region does not appear to support Amazon Bedrock at this time.
627
+ * - Some Asia Pacific regions, such as Asia Pacific (Hong Kong) and Asia Pacific (Jakarta), are not listed among the supported regions for Bedrock services.
628
+ * - Middle East regions, including Middle East (Bahrain) and Middle East (UAE), are not mentioned in the list of supported regions for Bedrock. [3]
629
+ * - China regions (Beijing and Ningxia) are not listed as supported for Amazon Bedrock.
630
+ * - Some newer or specialized AWS regions may not have Bedrock support yet.
631
+ */
632
+ "us.": { regionId: "us-east-1", description: "US East (N. Virginia)", pattern: "us-", multiRegion: true },
633
+ "use.": { regionId: "us-east-1", description: "US East (N. Virginia)" },
634
+ "use1.": { regionId: "us-east-1", description: "US East (N. Virginia)" },
635
+ "use2.": { regionId: "us-east-2", description: "US East (Ohio)" },
636
+ "usw.": { regionId: "us-west-2", description: "US West (Oregon)" },
637
+ "usw2.": { regionId: "us-west-2", description: "US West (Oregon)" },
638
+ "ug.": {
639
+ regionId: "us-gov-west-1",
640
+ description: "AWS GovCloud (US-West)",
641
+ pattern: "us-gov-",
642
+ multiRegion: true
643
+ },
644
+ "uge1.": { regionId: "us-gov-east-1", description: "AWS GovCloud (US-East)" },
645
+ "ugw1.": { regionId: "us-gov-west-1", description: "AWS GovCloud (US-West)" },
646
+ "eu.": { regionId: "eu-west-1", description: "Europe (Ireland)", pattern: "eu-", multiRegion: true },
647
+ "euw1.": { regionId: "eu-west-1", description: "Europe (Ireland)" },
648
+ "euw2.": { regionId: "eu-west-2", description: "Europe (London)" },
649
+ "euw3.": { regionId: "eu-west-3", description: "Europe (Paris)" },
650
+ "euc1.": { regionId: "eu-central-1", description: "Europe (Frankfurt)" },
651
+ "euc2.": { regionId: "eu-central-2", description: "Europe (Zurich)" },
652
+ "eun1.": { regionId: "eu-north-1", description: "Europe (Stockholm)" },
653
+ "eus1.": { regionId: "eu-south-1", description: "Europe (Milan)" },
654
+ "eus2.": { regionId: "eu-south-2", description: "Europe (Spain)" },
655
+ "ap.": {
656
+ regionId: "ap-southeast-1",
657
+ description: "Asia Pacific (Singapore)",
658
+ pattern: "ap-",
659
+ multiRegion: true
660
+ },
661
+ "ape1.": { regionId: "ap-east-1", description: "Asia Pacific (Hong Kong)" },
662
+ "apne1.": { regionId: "ap-northeast-1", description: "Asia Pacific (Tokyo)" },
663
+ "apne2.": { regionId: "ap-northeast-2", description: "Asia Pacific (Seoul)" },
664
+ "apne3.": { regionId: "ap-northeast-3", description: "Asia Pacific (Osaka)" },
665
+ "aps1.": { regionId: "ap-south-1", description: "Asia Pacific (Mumbai)" },
666
+ "aps2.": { regionId: "ap-south-2", description: "Asia Pacific (Hyderabad)" },
667
+ "apse1.": { regionId: "ap-southeast-1", description: "Asia Pacific (Singapore)" },
668
+ "apse2.": { regionId: "ap-southeast-2", description: "Asia Pacific (Sydney)" },
669
+ "ca.": { regionId: "ca-central-1", description: "Canada (Central)", pattern: "ca-", multiRegion: true },
670
+ "cac1.": { regionId: "ca-central-1", description: "Canada (Central)" },
671
+ "sa.": { regionId: "sa-east-1", description: "South America (S\xE3o Paulo)", pattern: "sa-", multiRegion: true },
672
+ "sae1.": { regionId: "sa-east-1", description: "South America (S\xE3o Paulo)" },
673
+ // These are not official - they weren't generated by Amazon Q nor were
674
+ // found in the AWS documentation but another Roo contributor found apac.
675
+ // Was needed so I've added the pattern of the other geo zones.
676
+ "apac.": { regionId: "ap-southeast-1", description: "Default APAC region", pattern: "ap-", multiRegion: true },
677
+ "emea.": { regionId: "eu-west-1", description: "Default EMEA region", pattern: "eu-", multiRegion: true },
678
+ "amer.": { regionId: "us-east-1", description: "Default Americas region", pattern: "us-", multiRegion: true }
679
+ };
680
+ var BEDROCK_REGIONS = Object.values(BEDROCK_REGION_INFO).map((info) => ({ value: info.regionId, label: info.regionId })).filter((region, index, self) => index === self.findIndex((r) => r.value === region.value)).sort((a, b) => a.value.localeCompare(b.value));
681
+
682
+ // src/providers/chutes.ts
683
+ var chutesDefaultModelId = "deepseek-ai/DeepSeek-R1-0528";
684
+ var chutesModels = {
685
+ "deepseek-ai/DeepSeek-R1-0528": {
686
+ maxTokens: 32768,
687
+ contextWindow: 163840,
688
+ supportsImages: false,
689
+ supportsPromptCache: false,
690
+ inputPrice: 0,
691
+ outputPrice: 0,
692
+ description: "DeepSeek R1 0528 model."
693
+ },
694
+ "deepseek-ai/DeepSeek-R1": {
695
+ maxTokens: 32768,
696
+ contextWindow: 163840,
697
+ supportsImages: false,
698
+ supportsPromptCache: false,
699
+ inputPrice: 0,
700
+ outputPrice: 0,
701
+ description: "DeepSeek R1 model."
702
+ },
703
+ "deepseek-ai/DeepSeek-V3": {
704
+ maxTokens: 32768,
705
+ contextWindow: 163840,
706
+ supportsImages: false,
707
+ supportsPromptCache: false,
708
+ inputPrice: 0,
709
+ outputPrice: 0,
710
+ description: "DeepSeek V3 model."
711
+ },
712
+ "unsloth/Llama-3.3-70B-Instruct": {
713
+ maxTokens: 32768,
714
+ // From Groq
715
+ contextWindow: 131072,
716
+ // From Groq
717
+ supportsImages: false,
718
+ supportsPromptCache: false,
719
+ inputPrice: 0,
720
+ outputPrice: 0,
721
+ description: "Unsloth Llama 3.3 70B Instruct model."
722
+ },
723
+ "chutesai/Llama-4-Scout-17B-16E-Instruct": {
724
+ maxTokens: 32768,
725
+ contextWindow: 512e3,
726
+ supportsImages: false,
727
+ supportsPromptCache: false,
728
+ inputPrice: 0,
729
+ outputPrice: 0,
730
+ description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
731
+ },
732
+ "unsloth/Mistral-Nemo-Instruct-2407": {
733
+ maxTokens: 32768,
734
+ contextWindow: 128e3,
735
+ supportsImages: false,
736
+ supportsPromptCache: false,
737
+ inputPrice: 0,
738
+ outputPrice: 0,
739
+ description: "Unsloth Mistral Nemo Instruct model."
740
+ },
741
+ "unsloth/gemma-3-12b-it": {
742
+ maxTokens: 32768,
743
+ contextWindow: 131072,
744
+ supportsImages: false,
745
+ supportsPromptCache: false,
746
+ inputPrice: 0,
747
+ outputPrice: 0,
748
+ description: "Unsloth Gemma 3 12B IT model."
749
+ },
750
+ "NousResearch/DeepHermes-3-Llama-3-8B-Preview": {
751
+ maxTokens: 32768,
752
+ contextWindow: 131072,
753
+ supportsImages: false,
754
+ supportsPromptCache: false,
755
+ inputPrice: 0,
756
+ outputPrice: 0,
757
+ description: "Nous DeepHermes 3 Llama 3 8B Preview model."
758
+ },
759
+ "unsloth/gemma-3-4b-it": {
760
+ maxTokens: 32768,
761
+ contextWindow: 131072,
762
+ supportsImages: false,
763
+ supportsPromptCache: false,
764
+ inputPrice: 0,
765
+ outputPrice: 0,
766
+ description: "Unsloth Gemma 3 4B IT model."
767
+ },
768
+ "nvidia/Llama-3_3-Nemotron-Super-49B-v1": {
769
+ maxTokens: 32768,
770
+ contextWindow: 131072,
771
+ supportsImages: false,
772
+ supportsPromptCache: false,
773
+ inputPrice: 0,
774
+ outputPrice: 0,
775
+ description: "Nvidia Llama 3.3 Nemotron Super 49B model."
776
+ },
777
+ "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1": {
778
+ maxTokens: 32768,
779
+ contextWindow: 131072,
780
+ supportsImages: false,
781
+ supportsPromptCache: false,
782
+ inputPrice: 0,
783
+ outputPrice: 0,
784
+ description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
785
+ },
786
+ "chutesai/Llama-4-Maverick-17B-128E-Instruct-FP8": {
787
+ maxTokens: 32768,
788
+ contextWindow: 256e3,
789
+ supportsImages: false,
790
+ supportsPromptCache: false,
791
+ inputPrice: 0,
792
+ outputPrice: 0,
793
+ description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
794
+ },
795
+ "deepseek-ai/DeepSeek-V3-Base": {
796
+ maxTokens: 32768,
797
+ contextWindow: 163840,
798
+ supportsImages: false,
799
+ supportsPromptCache: false,
800
+ inputPrice: 0,
801
+ outputPrice: 0,
802
+ description: "DeepSeek V3 Base model."
803
+ },
804
+ "deepseek-ai/DeepSeek-R1-Zero": {
805
+ maxTokens: 32768,
806
+ contextWindow: 163840,
807
+ supportsImages: false,
808
+ supportsPromptCache: false,
809
+ inputPrice: 0,
810
+ outputPrice: 0,
811
+ description: "DeepSeek R1 Zero model."
812
+ },
813
+ "deepseek-ai/DeepSeek-V3-0324": {
814
+ maxTokens: 32768,
815
+ contextWindow: 163840,
816
+ supportsImages: false,
817
+ supportsPromptCache: false,
818
+ inputPrice: 0,
819
+ outputPrice: 0,
820
+ description: "DeepSeek V3 (0324) model."
821
+ },
822
+ "Qwen/Qwen3-235B-A22B": {
823
+ maxTokens: 32768,
824
+ contextWindow: 40960,
825
+ supportsImages: false,
826
+ supportsPromptCache: false,
827
+ inputPrice: 0,
828
+ outputPrice: 0,
829
+ description: "Qwen3 235B A22B model."
830
+ },
831
+ "Qwen/Qwen3-32B": {
832
+ maxTokens: 32768,
833
+ contextWindow: 40960,
834
+ supportsImages: false,
835
+ supportsPromptCache: false,
836
+ inputPrice: 0,
837
+ outputPrice: 0,
838
+ description: "Qwen3 32B model."
839
+ },
840
+ "Qwen/Qwen3-30B-A3B": {
841
+ maxTokens: 32768,
842
+ contextWindow: 40960,
843
+ supportsImages: false,
844
+ supportsPromptCache: false,
845
+ inputPrice: 0,
846
+ outputPrice: 0,
847
+ description: "Qwen3 30B A3B model."
848
+ },
849
+ "Qwen/Qwen3-14B": {
850
+ maxTokens: 32768,
851
+ contextWindow: 40960,
852
+ supportsImages: false,
853
+ supportsPromptCache: false,
854
+ inputPrice: 0,
855
+ outputPrice: 0,
856
+ description: "Qwen3 14B model."
857
+ },
858
+ "Qwen/Qwen3-8B": {
859
+ maxTokens: 32768,
860
+ contextWindow: 40960,
861
+ supportsImages: false,
862
+ supportsPromptCache: false,
863
+ inputPrice: 0,
864
+ outputPrice: 0,
865
+ description: "Qwen3 8B model."
866
+ },
867
+ "microsoft/MAI-DS-R1-FP8": {
868
+ maxTokens: 32768,
869
+ contextWindow: 163840,
870
+ supportsImages: false,
871
+ supportsPromptCache: false,
872
+ inputPrice: 0,
873
+ outputPrice: 0,
874
+ description: "Microsoft MAI-DS-R1 FP8 model."
875
+ },
876
+ "tngtech/DeepSeek-R1T-Chimera": {
877
+ maxTokens: 32768,
878
+ contextWindow: 163840,
879
+ supportsImages: false,
880
+ supportsPromptCache: false,
881
+ inputPrice: 0,
882
+ outputPrice: 0,
883
+ description: "TNGTech DeepSeek R1T Chimera model."
884
+ }
885
+ };
886
+
887
+ // src/providers/deepseek.ts
888
+ var deepSeekDefaultModelId = "deepseek-chat";
889
+ var deepSeekModels = {
890
+ "deepseek-chat": {
891
+ maxTokens: 8192,
892
+ contextWindow: 64e3,
893
+ supportsImages: false,
894
+ supportsPromptCache: true,
895
+ inputPrice: 0.27,
896
+ // $0.27 per million tokens (cache miss)
897
+ outputPrice: 1.1,
898
+ // $1.10 per million tokens
899
+ cacheWritesPrice: 0.27,
900
+ // $0.27 per million tokens (cache miss)
901
+ cacheReadsPrice: 0.07,
902
+ // $0.07 per million tokens (cache hit).
903
+ description: `DeepSeek-V3 achieves a significant breakthrough in inference speed over previous models. It tops the leaderboard among open-source models and rivals the most advanced closed-source models globally.`
904
+ },
905
+ "deepseek-reasoner": {
906
+ maxTokens: 8192,
907
+ contextWindow: 64e3,
908
+ supportsImages: false,
909
+ supportsPromptCache: true,
910
+ inputPrice: 0.55,
911
+ // $0.55 per million tokens (cache miss)
912
+ outputPrice: 2.19,
913
+ // $2.19 per million tokens
914
+ cacheWritesPrice: 0.55,
915
+ // $0.55 per million tokens (cache miss)
916
+ cacheReadsPrice: 0.14,
917
+ // $0.14 per million tokens (cache hit)
918
+ description: `DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 32K tokens.`
919
+ }
920
+ };
921
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.6;
922
+
923
+ // src/providers/gemini.ts
924
+ var geminiDefaultModelId = "gemini-2.0-flash-001";
925
+ var geminiModels = {
926
+ "gemini-2.5-flash-preview-04-17:thinking": {
927
+ maxTokens: 65535,
928
+ contextWindow: 1048576,
929
+ supportsImages: true,
930
+ supportsPromptCache: false,
931
+ inputPrice: 0.15,
932
+ outputPrice: 3.5,
933
+ maxThinkingTokens: 24576,
934
+ supportsReasoningBudget: true,
935
+ requiredReasoningBudget: true
936
+ },
937
+ "gemini-2.5-flash-preview-04-17": {
938
+ maxTokens: 65535,
939
+ contextWindow: 1048576,
940
+ supportsImages: true,
941
+ supportsPromptCache: false,
942
+ inputPrice: 0.15,
943
+ outputPrice: 0.6
944
+ },
945
+ "gemini-2.5-flash-preview-05-20:thinking": {
946
+ maxTokens: 65535,
947
+ contextWindow: 1048576,
948
+ supportsImages: true,
949
+ supportsPromptCache: true,
950
+ inputPrice: 0.15,
951
+ outputPrice: 3.5,
952
+ cacheReadsPrice: 0.0375,
953
+ cacheWritesPrice: 1,
954
+ maxThinkingTokens: 24576,
955
+ supportsReasoningBudget: true,
956
+ requiredReasoningBudget: true
957
+ },
958
+ "gemini-2.5-flash-preview-05-20": {
959
+ maxTokens: 65535,
960
+ contextWindow: 1048576,
961
+ supportsImages: true,
962
+ supportsPromptCache: true,
963
+ inputPrice: 0.15,
964
+ outputPrice: 0.6,
965
+ cacheReadsPrice: 0.0375,
966
+ cacheWritesPrice: 1
967
+ },
968
+ "gemini-2.5-pro-exp-03-25": {
969
+ maxTokens: 65535,
970
+ contextWindow: 1048576,
971
+ supportsImages: true,
972
+ supportsPromptCache: false,
973
+ inputPrice: 0,
974
+ outputPrice: 0
975
+ },
976
+ "gemini-2.5-pro-preview-03-25": {
977
+ maxTokens: 65535,
978
+ contextWindow: 1048576,
979
+ supportsImages: true,
980
+ supportsPromptCache: true,
981
+ inputPrice: 2.5,
982
+ // This is the pricing for prompts above 200k tokens.
983
+ outputPrice: 15,
984
+ cacheReadsPrice: 0.625,
985
+ cacheWritesPrice: 4.5,
986
+ tiers: [
987
+ {
988
+ contextWindow: 2e5,
989
+ inputPrice: 1.25,
990
+ outputPrice: 10,
991
+ cacheReadsPrice: 0.31
992
+ },
993
+ {
994
+ contextWindow: Infinity,
995
+ inputPrice: 2.5,
996
+ outputPrice: 15,
997
+ cacheReadsPrice: 0.625
998
+ }
999
+ ]
1000
+ },
1001
+ "gemini-2.5-pro-preview-05-06": {
1002
+ maxTokens: 65535,
1003
+ contextWindow: 1048576,
1004
+ supportsImages: true,
1005
+ supportsPromptCache: true,
1006
+ inputPrice: 2.5,
1007
+ // This is the pricing for prompts above 200k tokens.
1008
+ outputPrice: 15,
1009
+ cacheReadsPrice: 0.625,
1010
+ cacheWritesPrice: 4.5,
1011
+ tiers: [
1012
+ {
1013
+ contextWindow: 2e5,
1014
+ inputPrice: 1.25,
1015
+ outputPrice: 10,
1016
+ cacheReadsPrice: 0.31
1017
+ },
1018
+ {
1019
+ contextWindow: Infinity,
1020
+ inputPrice: 2.5,
1021
+ outputPrice: 15,
1022
+ cacheReadsPrice: 0.625
1023
+ }
1024
+ ]
1025
+ },
1026
+ "gemini-2.0-flash-001": {
1027
+ maxTokens: 8192,
1028
+ contextWindow: 1048576,
1029
+ supportsImages: true,
1030
+ supportsPromptCache: true,
1031
+ inputPrice: 0.1,
1032
+ outputPrice: 0.4,
1033
+ cacheReadsPrice: 0.025,
1034
+ cacheWritesPrice: 1
1035
+ },
1036
+ "gemini-2.0-flash-lite-preview-02-05": {
1037
+ maxTokens: 8192,
1038
+ contextWindow: 1048576,
1039
+ supportsImages: true,
1040
+ supportsPromptCache: false,
1041
+ inputPrice: 0,
1042
+ outputPrice: 0
1043
+ },
1044
+ "gemini-2.0-pro-exp-02-05": {
1045
+ maxTokens: 8192,
1046
+ contextWindow: 2097152,
1047
+ supportsImages: true,
1048
+ supportsPromptCache: false,
1049
+ inputPrice: 0,
1050
+ outputPrice: 0
1051
+ },
1052
+ "gemini-2.0-flash-thinking-exp-01-21": {
1053
+ maxTokens: 65536,
1054
+ contextWindow: 1048576,
1055
+ supportsImages: true,
1056
+ supportsPromptCache: false,
1057
+ inputPrice: 0,
1058
+ outputPrice: 0
1059
+ },
1060
+ "gemini-2.0-flash-thinking-exp-1219": {
1061
+ maxTokens: 8192,
1062
+ contextWindow: 32767,
1063
+ supportsImages: true,
1064
+ supportsPromptCache: false,
1065
+ inputPrice: 0,
1066
+ outputPrice: 0
1067
+ },
1068
+ "gemini-2.0-flash-exp": {
1069
+ maxTokens: 8192,
1070
+ contextWindow: 1048576,
1071
+ supportsImages: true,
1072
+ supportsPromptCache: false,
1073
+ inputPrice: 0,
1074
+ outputPrice: 0
1075
+ },
1076
+ "gemini-1.5-flash-002": {
1077
+ maxTokens: 8192,
1078
+ contextWindow: 1048576,
1079
+ supportsImages: true,
1080
+ supportsPromptCache: true,
1081
+ inputPrice: 0.15,
1082
+ // This is the pricing for prompts above 128k tokens.
1083
+ outputPrice: 0.6,
1084
+ cacheReadsPrice: 0.0375,
1085
+ cacheWritesPrice: 1,
1086
+ tiers: [
1087
+ {
1088
+ contextWindow: 128e3,
1089
+ inputPrice: 0.075,
1090
+ outputPrice: 0.3,
1091
+ cacheReadsPrice: 0.01875
1092
+ },
1093
+ {
1094
+ contextWindow: Infinity,
1095
+ inputPrice: 0.15,
1096
+ outputPrice: 0.6,
1097
+ cacheReadsPrice: 0.0375
1098
+ }
1099
+ ]
1100
+ },
1101
+ "gemini-1.5-flash-exp-0827": {
1102
+ maxTokens: 8192,
1103
+ contextWindow: 1048576,
1104
+ supportsImages: true,
1105
+ supportsPromptCache: false,
1106
+ inputPrice: 0,
1107
+ outputPrice: 0
1108
+ },
1109
+ "gemini-1.5-flash-8b-exp-0827": {
1110
+ maxTokens: 8192,
1111
+ contextWindow: 1048576,
1112
+ supportsImages: true,
1113
+ supportsPromptCache: false,
1114
+ inputPrice: 0,
1115
+ outputPrice: 0
1116
+ },
1117
+ "gemini-1.5-pro-002": {
1118
+ maxTokens: 8192,
1119
+ contextWindow: 2097152,
1120
+ supportsImages: true,
1121
+ supportsPromptCache: false,
1122
+ inputPrice: 0,
1123
+ outputPrice: 0
1124
+ },
1125
+ "gemini-1.5-pro-exp-0827": {
1126
+ maxTokens: 8192,
1127
+ contextWindow: 2097152,
1128
+ supportsImages: true,
1129
+ supportsPromptCache: false,
1130
+ inputPrice: 0,
1131
+ outputPrice: 0
1132
+ },
1133
+ "gemini-exp-1206": {
1134
+ maxTokens: 8192,
1135
+ contextWindow: 2097152,
1136
+ supportsImages: true,
1137
+ supportsPromptCache: false,
1138
+ inputPrice: 0,
1139
+ outputPrice: 0
1140
+ }
1141
+ };
1142
+
1143
+ // src/providers/glama.ts
1144
+ var glamaDefaultModelId = "anthropic/claude-3-7-sonnet";
1145
+ var glamaDefaultModelInfo = {
1146
+ maxTokens: 8192,
1147
+ contextWindow: 2e5,
1148
+ supportsImages: true,
1149
+ supportsComputerUse: true,
1150
+ supportsPromptCache: true,
1151
+ inputPrice: 3,
1152
+ outputPrice: 15,
1153
+ cacheWritesPrice: 3.75,
1154
+ cacheReadsPrice: 0.3,
1155
+ description: "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. Claude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks. Read more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)"
1156
+ };
1157
+ var GLAMA_DEFAULT_TEMPERATURE = 0;
1158
+
1159
+ // src/providers/groq.ts
1160
+ var groqDefaultModelId = "llama-3.3-70b-versatile";
1161
+ var groqModels = {
1162
+ // Models based on API response: https://api.groq.com/openai/v1/models
1163
+ "llama-3.1-8b-instant": {
1164
+ maxTokens: 131072,
1165
+ contextWindow: 131072,
1166
+ supportsImages: false,
1167
+ supportsPromptCache: false,
1168
+ inputPrice: 0,
1169
+ outputPrice: 0,
1170
+ description: "Meta Llama 3.1 8B Instant model, 128K context."
1171
+ },
1172
+ "llama-3.3-70b-versatile": {
1173
+ maxTokens: 32768,
1174
+ contextWindow: 131072,
1175
+ supportsImages: false,
1176
+ supportsPromptCache: false,
1177
+ inputPrice: 0,
1178
+ outputPrice: 0,
1179
+ description: "Meta Llama 3.3 70B Versatile model, 128K context."
1180
+ },
1181
+ "meta-llama/llama-4-scout-17b-16e-instruct": {
1182
+ maxTokens: 8192,
1183
+ contextWindow: 131072,
1184
+ supportsImages: false,
1185
+ supportsPromptCache: false,
1186
+ inputPrice: 0,
1187
+ outputPrice: 0,
1188
+ description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
1189
+ },
1190
+ "meta-llama/llama-4-maverick-17b-128e-instruct": {
1191
+ maxTokens: 8192,
1192
+ contextWindow: 131072,
1193
+ supportsImages: false,
1194
+ supportsPromptCache: false,
1195
+ inputPrice: 0,
1196
+ outputPrice: 0,
1197
+ description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
1198
+ },
1199
+ "mistral-saba-24b": {
1200
+ maxTokens: 32768,
1201
+ contextWindow: 32768,
1202
+ supportsImages: false,
1203
+ supportsPromptCache: false,
1204
+ inputPrice: 0,
1205
+ outputPrice: 0,
1206
+ description: "Mistral Saba 24B model, 32K context."
1207
+ },
1208
+ "qwen-qwq-32b": {
1209
+ maxTokens: 131072,
1210
+ contextWindow: 131072,
1211
+ supportsImages: false,
1212
+ supportsPromptCache: false,
1213
+ inputPrice: 0,
1214
+ outputPrice: 0,
1215
+ description: "Alibaba Qwen QwQ 32B model, 128K context."
1216
+ },
1217
+ "deepseek-r1-distill-llama-70b": {
1218
+ maxTokens: 131072,
1219
+ contextWindow: 131072,
1220
+ supportsImages: false,
1221
+ supportsPromptCache: false,
1222
+ inputPrice: 0,
1223
+ outputPrice: 0,
1224
+ description: "DeepSeek R1 Distill Llama 70B model, 128K context."
1225
+ }
1226
+ };
1227
+
1228
+ // src/providers/lite-llm.ts
1229
+ var litellmDefaultModelId = "claude-3-7-sonnet-20250219";
1230
+ var litellmDefaultModelInfo = {
1231
+ maxTokens: 8192,
1232
+ contextWindow: 2e5,
1233
+ supportsImages: true,
1234
+ supportsComputerUse: true,
1235
+ supportsPromptCache: true,
1236
+ inputPrice: 3,
1237
+ outputPrice: 15,
1238
+ cacheWritesPrice: 3.75,
1239
+ cacheReadsPrice: 0.3
1240
+ };
1241
+ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1242
+ "claude-3-5-sonnet-latest",
1243
+ "claude-opus-4-20250514",
1244
+ "claude-sonnet-4-20250514",
1245
+ "claude-3-7-sonnet-latest",
1246
+ "claude-3-7-sonnet-20250219",
1247
+ "claude-3-5-sonnet-20241022",
1248
+ "vertex_ai/claude-3-5-sonnet",
1249
+ "vertex_ai/claude-3-5-sonnet-v2",
1250
+ "vertex_ai/claude-3-5-sonnet-v2@20241022",
1251
+ "vertex_ai/claude-3-7-sonnet@20250219",
1252
+ "vertex_ai/claude-opus-4@20250514",
1253
+ "vertex_ai/claude-sonnet-4@20250514",
1254
+ "openrouter/anthropic/claude-3.5-sonnet",
1255
+ "openrouter/anthropic/claude-3.5-sonnet:beta",
1256
+ "openrouter/anthropic/claude-3.7-sonnet",
1257
+ "openrouter/anthropic/claude-3.7-sonnet:beta",
1258
+ "anthropic.claude-opus-4-20250514-v1:0",
1259
+ "anthropic.claude-sonnet-4-20250514-v1:0",
1260
+ "anthropic.claude-3-7-sonnet-20250219-v1:0",
1261
+ "anthropic.claude-3-5-sonnet-20241022-v2:0",
1262
+ "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
1263
+ "us.anthropic.claude-3-7-sonnet-20250219-v1:0",
1264
+ "us.anthropic.claude-opus-4-20250514-v1:0",
1265
+ "us.anthropic.claude-sonnet-4-20250514-v1:0",
1266
+ "eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
1267
+ "eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
1268
+ "eu.anthropic.claude-opus-4-20250514-v1:0",
1269
+ "eu.anthropic.claude-sonnet-4-20250514-v1:0",
1270
+ "snowflake/claude-3-5-sonnet"
1271
+ ]);
1272
+
1273
+ // src/providers/lm-studio.ts
1274
+ var LMSTUDIO_DEFAULT_TEMPERATURE = 0;
1275
+
1276
+ // src/providers/mistral.ts
1277
+ var mistralDefaultModelId = "codestral-latest";
1278
+ var mistralModels = {
1279
+ "codestral-latest": {
1280
+ maxTokens: 256e3,
1281
+ contextWindow: 256e3,
1282
+ supportsImages: false,
1283
+ supportsPromptCache: false,
1284
+ inputPrice: 0.3,
1285
+ outputPrice: 0.9
1286
+ },
1287
+ "mistral-large-latest": {
1288
+ maxTokens: 131e3,
1289
+ contextWindow: 131e3,
1290
+ supportsImages: false,
1291
+ supportsPromptCache: false,
1292
+ inputPrice: 2,
1293
+ outputPrice: 6
1294
+ },
1295
+ "ministral-8b-latest": {
1296
+ maxTokens: 131e3,
1297
+ contextWindow: 131e3,
1298
+ supportsImages: false,
1299
+ supportsPromptCache: false,
1300
+ inputPrice: 0.1,
1301
+ outputPrice: 0.1
1302
+ },
1303
+ "ministral-3b-latest": {
1304
+ maxTokens: 131e3,
1305
+ contextWindow: 131e3,
1306
+ supportsImages: false,
1307
+ supportsPromptCache: false,
1308
+ inputPrice: 0.04,
1309
+ outputPrice: 0.04
1310
+ },
1311
+ "mistral-small-latest": {
1312
+ maxTokens: 32e3,
1313
+ contextWindow: 32e3,
1314
+ supportsImages: false,
1315
+ supportsPromptCache: false,
1316
+ inputPrice: 0.2,
1317
+ outputPrice: 0.6
1318
+ },
1319
+ "pixtral-large-latest": {
1320
+ maxTokens: 131e3,
1321
+ contextWindow: 131e3,
1322
+ supportsImages: true,
1323
+ supportsPromptCache: false,
1324
+ inputPrice: 2,
1325
+ outputPrice: 6
1326
+ }
1327
+ };
1328
+ var MISTRAL_DEFAULT_TEMPERATURE = 0;
1329
+
1330
+ // src/providers/openai.ts
1331
+ var openAiNativeDefaultModelId = "gpt-4.1";
1332
+ var openAiNativeModels = {
1333
+ "gpt-4.1": {
1334
+ maxTokens: 32768,
1335
+ contextWindow: 1047576,
1336
+ supportsImages: true,
1337
+ supportsPromptCache: true,
1338
+ inputPrice: 2,
1339
+ outputPrice: 8,
1340
+ cacheReadsPrice: 0.5
1341
+ },
1342
+ "gpt-4.1-mini": {
1343
+ maxTokens: 32768,
1344
+ contextWindow: 1047576,
1345
+ supportsImages: true,
1346
+ supportsPromptCache: true,
1347
+ inputPrice: 0.4,
1348
+ outputPrice: 1.6,
1349
+ cacheReadsPrice: 0.1
1350
+ },
1351
+ "gpt-4.1-nano": {
1352
+ maxTokens: 32768,
1353
+ contextWindow: 1047576,
1354
+ supportsImages: true,
1355
+ supportsPromptCache: true,
1356
+ inputPrice: 0.1,
1357
+ outputPrice: 0.4,
1358
+ cacheReadsPrice: 0.025
1359
+ },
1360
+ o3: {
1361
+ maxTokens: 1e5,
1362
+ contextWindow: 2e5,
1363
+ supportsImages: true,
1364
+ supportsPromptCache: true,
1365
+ inputPrice: 10,
1366
+ outputPrice: 40,
1367
+ cacheReadsPrice: 2.5,
1368
+ supportsReasoningEffort: true,
1369
+ reasoningEffort: "medium"
1370
+ },
1371
+ "o3-high": {
1372
+ maxTokens: 1e5,
1373
+ contextWindow: 2e5,
1374
+ supportsImages: true,
1375
+ supportsPromptCache: true,
1376
+ inputPrice: 10,
1377
+ outputPrice: 40,
1378
+ cacheReadsPrice: 2.5,
1379
+ reasoningEffort: "high"
1380
+ },
1381
+ "o3-low": {
1382
+ maxTokens: 1e5,
1383
+ contextWindow: 2e5,
1384
+ supportsImages: true,
1385
+ supportsPromptCache: true,
1386
+ inputPrice: 10,
1387
+ outputPrice: 40,
1388
+ cacheReadsPrice: 2.5,
1389
+ reasoningEffort: "low"
1390
+ },
1391
+ "o4-mini": {
1392
+ maxTokens: 1e5,
1393
+ contextWindow: 2e5,
1394
+ supportsImages: true,
1395
+ supportsPromptCache: true,
1396
+ inputPrice: 1.1,
1397
+ outputPrice: 4.4,
1398
+ cacheReadsPrice: 0.275,
1399
+ supportsReasoningEffort: true,
1400
+ reasoningEffort: "medium"
1401
+ },
1402
+ "o4-mini-high": {
1403
+ maxTokens: 1e5,
1404
+ contextWindow: 2e5,
1405
+ supportsImages: true,
1406
+ supportsPromptCache: true,
1407
+ inputPrice: 1.1,
1408
+ outputPrice: 4.4,
1409
+ cacheReadsPrice: 0.275,
1410
+ reasoningEffort: "high"
1411
+ },
1412
+ "o4-mini-low": {
1413
+ maxTokens: 1e5,
1414
+ contextWindow: 2e5,
1415
+ supportsImages: true,
1416
+ supportsPromptCache: true,
1417
+ inputPrice: 1.1,
1418
+ outputPrice: 4.4,
1419
+ cacheReadsPrice: 0.275,
1420
+ reasoningEffort: "low"
1421
+ },
1422
+ "o3-mini": {
1423
+ maxTokens: 1e5,
1424
+ contextWindow: 2e5,
1425
+ supportsImages: false,
1426
+ supportsPromptCache: true,
1427
+ inputPrice: 1.1,
1428
+ outputPrice: 4.4,
1429
+ cacheReadsPrice: 0.55,
1430
+ supportsReasoningEffort: true,
1431
+ reasoningEffort: "medium"
1432
+ },
1433
+ "o3-mini-high": {
1434
+ maxTokens: 1e5,
1435
+ contextWindow: 2e5,
1436
+ supportsImages: false,
1437
+ supportsPromptCache: true,
1438
+ inputPrice: 1.1,
1439
+ outputPrice: 4.4,
1440
+ cacheReadsPrice: 0.55,
1441
+ reasoningEffort: "high"
1442
+ },
1443
+ "o3-mini-low": {
1444
+ maxTokens: 1e5,
1445
+ contextWindow: 2e5,
1446
+ supportsImages: false,
1447
+ supportsPromptCache: true,
1448
+ inputPrice: 1.1,
1449
+ outputPrice: 4.4,
1450
+ cacheReadsPrice: 0.55,
1451
+ reasoningEffort: "low"
1452
+ },
1453
+ o1: {
1454
+ maxTokens: 1e5,
1455
+ contextWindow: 2e5,
1456
+ supportsImages: true,
1457
+ supportsPromptCache: true,
1458
+ inputPrice: 15,
1459
+ outputPrice: 60,
1460
+ cacheReadsPrice: 7.5
1461
+ },
1462
+ "o1-preview": {
1463
+ maxTokens: 32768,
1464
+ contextWindow: 128e3,
1465
+ supportsImages: true,
1466
+ supportsPromptCache: true,
1467
+ inputPrice: 15,
1468
+ outputPrice: 60,
1469
+ cacheReadsPrice: 7.5
1470
+ },
1471
+ "o1-mini": {
1472
+ maxTokens: 65536,
1473
+ contextWindow: 128e3,
1474
+ supportsImages: true,
1475
+ supportsPromptCache: true,
1476
+ inputPrice: 1.1,
1477
+ outputPrice: 4.4,
1478
+ cacheReadsPrice: 0.55
1479
+ },
1480
+ "gpt-4.5-preview": {
1481
+ maxTokens: 16384,
1482
+ contextWindow: 128e3,
1483
+ supportsImages: true,
1484
+ supportsPromptCache: true,
1485
+ inputPrice: 75,
1486
+ outputPrice: 150,
1487
+ cacheReadsPrice: 37.5
1488
+ },
1489
+ "gpt-4o": {
1490
+ maxTokens: 16384,
1491
+ contextWindow: 128e3,
1492
+ supportsImages: true,
1493
+ supportsPromptCache: true,
1494
+ inputPrice: 2.5,
1495
+ outputPrice: 10,
1496
+ cacheReadsPrice: 1.25
1497
+ },
1498
+ "gpt-4o-mini": {
1499
+ maxTokens: 16384,
1500
+ contextWindow: 128e3,
1501
+ supportsImages: true,
1502
+ supportsPromptCache: true,
1503
+ inputPrice: 0.15,
1504
+ outputPrice: 0.6,
1505
+ cacheReadsPrice: 0.075
1506
+ }
1507
+ };
1508
+ var openAiModelInfoSaneDefaults = {
1509
+ maxTokens: -1,
1510
+ contextWindow: 128e3,
1511
+ supportsImages: true,
1512
+ supportsPromptCache: false,
1513
+ inputPrice: 0,
1514
+ outputPrice: 0
1515
+ };
1516
+ var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
1517
+ var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
1518
+ var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
1519
+
1520
+ // src/providers/openrouter.ts
1521
+ var openRouterDefaultModelId = "anthropic/claude-sonnet-4";
1522
+ var openRouterDefaultModelInfo = {
1523
+ maxTokens: 8192,
1524
+ contextWindow: 2e5,
1525
+ supportsImages: true,
1526
+ supportsComputerUse: true,
1527
+ supportsPromptCache: true,
1528
+ inputPrice: 3,
1529
+ outputPrice: 15,
1530
+ cacheWritesPrice: 3.75,
1531
+ cacheReadsPrice: 0.3,
1532
+ description: "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. Claude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks. Read more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)"
1533
+ };
1534
+ var OPENROUTER_DEFAULT_PROVIDER_NAME = "[default]";
1535
+ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
1536
+ "anthropic/claude-3-haiku",
1537
+ "anthropic/claude-3-haiku:beta",
1538
+ "anthropic/claude-3-opus",
1539
+ "anthropic/claude-3-opus:beta",
1540
+ "anthropic/claude-3-sonnet",
1541
+ "anthropic/claude-3-sonnet:beta",
1542
+ "anthropic/claude-3.5-haiku",
1543
+ "anthropic/claude-3.5-haiku-20241022",
1544
+ "anthropic/claude-3.5-haiku-20241022:beta",
1545
+ "anthropic/claude-3.5-haiku:beta",
1546
+ "anthropic/claude-3.5-sonnet",
1547
+ "anthropic/claude-3.5-sonnet-20240620",
1548
+ "anthropic/claude-3.5-sonnet-20240620:beta",
1549
+ "anthropic/claude-3.5-sonnet:beta",
1550
+ "anthropic/claude-3.7-sonnet",
1551
+ "anthropic/claude-3.7-sonnet:beta",
1552
+ "anthropic/claude-3.7-sonnet:thinking",
1553
+ "anthropic/claude-sonnet-4",
1554
+ "anthropic/claude-opus-4",
1555
+ "google/gemini-2.5-pro-preview",
1556
+ "google/gemini-2.5-flash-preview",
1557
+ "google/gemini-2.5-flash-preview:thinking",
1558
+ "google/gemini-2.5-flash-preview-05-20",
1559
+ "google/gemini-2.5-flash-preview-05-20:thinking",
1560
+ "google/gemini-2.0-flash-001",
1561
+ "google/gemini-flash-1.5",
1562
+ "google/gemini-flash-1.5-8b"
1563
+ ]);
1564
+ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1565
+ "anthropic/claude-3.5-sonnet",
1566
+ "anthropic/claude-3.5-sonnet:beta",
1567
+ "anthropic/claude-3.7-sonnet",
1568
+ "anthropic/claude-3.7-sonnet:beta",
1569
+ "anthropic/claude-3.7-sonnet:thinking",
1570
+ "anthropic/claude-sonnet-4",
1571
+ "anthropic/claude-opus-4"
1572
+ ]);
1573
+ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1574
+ "anthropic/claude-3.7-sonnet:beta",
1575
+ "anthropic/claude-3.7-sonnet:thinking",
1576
+ "anthropic/claude-opus-4",
1577
+ "anthropic/claude-sonnet-4",
1578
+ "google/gemini-2.5-flash-preview-05-20",
1579
+ "google/gemini-2.5-flash-preview-05-20:thinking"
1580
+ ]);
1581
+ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1582
+ "anthropic/claude-3.7-sonnet:thinking",
1583
+ "google/gemini-2.5-flash-preview-05-20:thinking"
1584
+ ]);
1585
+
1586
+ // src/providers/requesty.ts
1587
+ var requestyDefaultModelId = "coding/claude-4-sonnet";
1588
+ var requestyDefaultModelInfo = {
1589
+ maxTokens: 8192,
1590
+ contextWindow: 2e5,
1591
+ supportsImages: true,
1592
+ supportsComputerUse: true,
1593
+ supportsPromptCache: true,
1594
+ inputPrice: 3,
1595
+ outputPrice: 15,
1596
+ cacheWritesPrice: 3.75,
1597
+ cacheReadsPrice: 0.3,
1598
+ description: "The best coding model, optimized by Requesty, and automatically routed to the fastest provider. Claude 4 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities."
1599
+ };
1600
+
1601
+ // src/providers/unbound.ts
1602
+ var unboundDefaultModelId = "anthropic/claude-3-7-sonnet-20250219";
1603
+ var unboundDefaultModelInfo = {
1604
+ maxTokens: 8192,
1605
+ contextWindow: 2e5,
1606
+ supportsImages: true,
1607
+ supportsPromptCache: true,
1608
+ inputPrice: 3,
1609
+ outputPrice: 15,
1610
+ cacheWritesPrice: 3.75,
1611
+ cacheReadsPrice: 0.3
1612
+ };
1613
+
1614
+ // src/providers/vertex.ts
1615
+ var vertexDefaultModelId = "claude-sonnet-4@20250514";
1616
+ var vertexModels = {
1617
+ "gemini-2.5-flash-preview-05-20:thinking": {
1618
+ maxTokens: 65535,
1619
+ contextWindow: 1048576,
1620
+ supportsImages: true,
1621
+ supportsPromptCache: true,
1622
+ inputPrice: 0.15,
1623
+ outputPrice: 3.5,
1624
+ maxThinkingTokens: 24576,
1625
+ supportsReasoningBudget: true,
1626
+ requiredReasoningBudget: true
1627
+ },
1628
+ "gemini-2.5-flash-preview-05-20": {
1629
+ maxTokens: 65535,
1630
+ contextWindow: 1048576,
1631
+ supportsImages: true,
1632
+ supportsPromptCache: true,
1633
+ inputPrice: 0.15,
1634
+ outputPrice: 0.6
1635
+ },
1636
+ "gemini-2.5-flash-preview-04-17:thinking": {
1637
+ maxTokens: 65535,
1638
+ contextWindow: 1048576,
1639
+ supportsImages: true,
1640
+ supportsPromptCache: false,
1641
+ inputPrice: 0.15,
1642
+ outputPrice: 3.5,
1643
+ maxThinkingTokens: 24576,
1644
+ supportsReasoningBudget: true,
1645
+ requiredReasoningBudget: true
1646
+ },
1647
+ "gemini-2.5-flash-preview-04-17": {
1648
+ maxTokens: 65535,
1649
+ contextWindow: 1048576,
1650
+ supportsImages: true,
1651
+ supportsPromptCache: false,
1652
+ inputPrice: 0.15,
1653
+ outputPrice: 0.6
1654
+ },
1655
+ "gemini-2.5-pro-preview-03-25": {
1656
+ maxTokens: 65535,
1657
+ contextWindow: 1048576,
1658
+ supportsImages: true,
1659
+ supportsPromptCache: true,
1660
+ inputPrice: 2.5,
1661
+ outputPrice: 15
1662
+ },
1663
+ "gemini-2.5-pro-preview-05-06": {
1664
+ maxTokens: 65535,
1665
+ contextWindow: 1048576,
1666
+ supportsImages: true,
1667
+ supportsPromptCache: true,
1668
+ inputPrice: 2.5,
1669
+ outputPrice: 15
1670
+ },
1671
+ "gemini-2.5-pro-exp-03-25": {
1672
+ maxTokens: 65535,
1673
+ contextWindow: 1048576,
1674
+ supportsImages: true,
1675
+ supportsPromptCache: false,
1676
+ inputPrice: 0,
1677
+ outputPrice: 0
1678
+ },
1679
+ "gemini-2.0-pro-exp-02-05": {
1680
+ maxTokens: 8192,
1681
+ contextWindow: 2097152,
1682
+ supportsImages: true,
1683
+ supportsPromptCache: false,
1684
+ inputPrice: 0,
1685
+ outputPrice: 0
1686
+ },
1687
+ "gemini-2.0-flash-001": {
1688
+ maxTokens: 8192,
1689
+ contextWindow: 1048576,
1690
+ supportsImages: true,
1691
+ supportsPromptCache: true,
1692
+ inputPrice: 0.15,
1693
+ outputPrice: 0.6
1694
+ },
1695
+ "gemini-2.0-flash-lite-001": {
1696
+ maxTokens: 8192,
1697
+ contextWindow: 1048576,
1698
+ supportsImages: true,
1699
+ supportsPromptCache: false,
1700
+ inputPrice: 0.075,
1701
+ outputPrice: 0.3
1702
+ },
1703
+ "gemini-2.0-flash-thinking-exp-01-21": {
1704
+ maxTokens: 8192,
1705
+ contextWindow: 32768,
1706
+ supportsImages: true,
1707
+ supportsPromptCache: false,
1708
+ inputPrice: 0,
1709
+ outputPrice: 0
1710
+ },
1711
+ "gemini-1.5-flash-002": {
1712
+ maxTokens: 8192,
1713
+ contextWindow: 1048576,
1714
+ supportsImages: true,
1715
+ supportsPromptCache: true,
1716
+ inputPrice: 0.075,
1717
+ outputPrice: 0.3
1718
+ },
1719
+ "gemini-1.5-pro-002": {
1720
+ maxTokens: 8192,
1721
+ contextWindow: 2097152,
1722
+ supportsImages: true,
1723
+ supportsPromptCache: false,
1724
+ inputPrice: 1.25,
1725
+ outputPrice: 5
1726
+ },
1727
+ "claude-sonnet-4@20250514": {
1728
+ maxTokens: 8192,
1729
+ contextWindow: 2e5,
1730
+ supportsImages: true,
1731
+ supportsComputerUse: true,
1732
+ supportsPromptCache: true,
1733
+ inputPrice: 3,
1734
+ outputPrice: 15,
1735
+ cacheWritesPrice: 3.75,
1736
+ cacheReadsPrice: 0.3,
1737
+ supportsReasoningBudget: true
1738
+ },
1739
+ "claude-opus-4@20250514": {
1740
+ maxTokens: 8192,
1741
+ contextWindow: 2e5,
1742
+ supportsImages: true,
1743
+ supportsComputerUse: true,
1744
+ supportsPromptCache: true,
1745
+ inputPrice: 15,
1746
+ outputPrice: 75,
1747
+ cacheWritesPrice: 18.75,
1748
+ cacheReadsPrice: 1.5
1749
+ },
1750
+ "claude-3-7-sonnet@20250219:thinking": {
1751
+ maxTokens: 64e3,
1752
+ contextWindow: 2e5,
1753
+ supportsImages: true,
1754
+ supportsComputerUse: true,
1755
+ supportsPromptCache: true,
1756
+ inputPrice: 3,
1757
+ outputPrice: 15,
1758
+ cacheWritesPrice: 3.75,
1759
+ cacheReadsPrice: 0.3,
1760
+ supportsReasoningBudget: true,
1761
+ requiredReasoningBudget: true
1762
+ },
1763
+ "claude-3-7-sonnet@20250219": {
1764
+ maxTokens: 8192,
1765
+ contextWindow: 2e5,
1766
+ supportsImages: true,
1767
+ supportsComputerUse: true,
1768
+ supportsPromptCache: true,
1769
+ inputPrice: 3,
1770
+ outputPrice: 15,
1771
+ cacheWritesPrice: 3.75,
1772
+ cacheReadsPrice: 0.3
1773
+ },
1774
+ "claude-3-5-sonnet-v2@20241022": {
1775
+ maxTokens: 8192,
1776
+ contextWindow: 2e5,
1777
+ supportsImages: true,
1778
+ supportsComputerUse: true,
1779
+ supportsPromptCache: true,
1780
+ inputPrice: 3,
1781
+ outputPrice: 15,
1782
+ cacheWritesPrice: 3.75,
1783
+ cacheReadsPrice: 0.3
1784
+ },
1785
+ "claude-3-5-sonnet@20240620": {
1786
+ maxTokens: 8192,
1787
+ contextWindow: 2e5,
1788
+ supportsImages: true,
1789
+ supportsPromptCache: true,
1790
+ inputPrice: 3,
1791
+ outputPrice: 15,
1792
+ cacheWritesPrice: 3.75,
1793
+ cacheReadsPrice: 0.3
1794
+ },
1795
+ "claude-3-5-haiku@20241022": {
1796
+ maxTokens: 8192,
1797
+ contextWindow: 2e5,
1798
+ supportsImages: false,
1799
+ supportsPromptCache: true,
1800
+ inputPrice: 1,
1801
+ outputPrice: 5,
1802
+ cacheWritesPrice: 1.25,
1803
+ cacheReadsPrice: 0.1
1804
+ },
1805
+ "claude-3-opus@20240229": {
1806
+ maxTokens: 4096,
1807
+ contextWindow: 2e5,
1808
+ supportsImages: true,
1809
+ supportsPromptCache: true,
1810
+ inputPrice: 15,
1811
+ outputPrice: 75,
1812
+ cacheWritesPrice: 18.75,
1813
+ cacheReadsPrice: 1.5
1814
+ },
1815
+ "claude-3-haiku@20240307": {
1816
+ maxTokens: 4096,
1817
+ contextWindow: 2e5,
1818
+ supportsImages: true,
1819
+ supportsPromptCache: true,
1820
+ inputPrice: 0.25,
1821
+ outputPrice: 1.25,
1822
+ cacheWritesPrice: 0.3,
1823
+ cacheReadsPrice: 0.03
1824
+ }
1825
+ };
1826
+ var VERTEX_REGIONS = [
1827
+ { value: "us-east5", label: "us-east5" },
1828
+ { value: "us-central1", label: "us-central1" },
1829
+ { value: "europe-west1", label: "europe-west1" },
1830
+ { value: "europe-west4", label: "europe-west4" },
1831
+ { value: "asia-southeast1", label: "asia-southeast1" }
1832
+ ];
1833
+
1834
+ // src/providers/vscode-llm.ts
1835
+ var vscodeLlmDefaultModelId = "claude-3.5-sonnet";
1836
+ var vscodeLlmModels = {
1837
+ "gpt-3.5-turbo": {
1838
+ contextWindow: 12114,
1839
+ supportsImages: false,
1840
+ supportsPromptCache: false,
1841
+ inputPrice: 0,
1842
+ outputPrice: 0,
1843
+ family: "gpt-3.5-turbo",
1844
+ version: "gpt-3.5-turbo-0613",
1845
+ name: "GPT 3.5 Turbo",
1846
+ supportsToolCalling: true,
1847
+ maxInputTokens: 12114
1848
+ },
1849
+ "gpt-4o-mini": {
1850
+ contextWindow: 12115,
1851
+ supportsImages: false,
1852
+ supportsPromptCache: false,
1853
+ inputPrice: 0,
1854
+ outputPrice: 0,
1855
+ family: "gpt-4o-mini",
1856
+ version: "gpt-4o-mini-2024-07-18",
1857
+ name: "GPT-4o mini",
1858
+ supportsToolCalling: true,
1859
+ maxInputTokens: 12115
1860
+ },
1861
+ "gpt-4": {
1862
+ contextWindow: 28501,
1863
+ supportsImages: false,
1864
+ supportsPromptCache: false,
1865
+ inputPrice: 0,
1866
+ outputPrice: 0,
1867
+ family: "gpt-4",
1868
+ version: "gpt-4-0613",
1869
+ name: "GPT 4",
1870
+ supportsToolCalling: true,
1871
+ maxInputTokens: 28501
1872
+ },
1873
+ "gpt-4-0125-preview": {
1874
+ contextWindow: 63826,
1875
+ supportsImages: false,
1876
+ supportsPromptCache: false,
1877
+ inputPrice: 0,
1878
+ outputPrice: 0,
1879
+ family: "gpt-4-turbo",
1880
+ version: "gpt-4-0125-preview",
1881
+ name: "GPT 4 Turbo",
1882
+ supportsToolCalling: true,
1883
+ maxInputTokens: 63826
1884
+ },
1885
+ "gpt-4o": {
1886
+ contextWindow: 63827,
1887
+ supportsImages: true,
1888
+ supportsPromptCache: false,
1889
+ inputPrice: 0,
1890
+ outputPrice: 0,
1891
+ family: "gpt-4o",
1892
+ version: "gpt-4o-2024-11-20",
1893
+ name: "GPT-4o",
1894
+ supportsToolCalling: true,
1895
+ maxInputTokens: 63827
1896
+ },
1897
+ o1: {
1898
+ contextWindow: 19827,
1899
+ supportsImages: false,
1900
+ supportsPromptCache: false,
1901
+ inputPrice: 0,
1902
+ outputPrice: 0,
1903
+ family: "o1-ga",
1904
+ version: "o1-2024-12-17",
1905
+ name: "o1 (Preview)",
1906
+ supportsToolCalling: true,
1907
+ maxInputTokens: 19827
1908
+ },
1909
+ "o3-mini": {
1910
+ contextWindow: 63827,
1911
+ supportsImages: false,
1912
+ supportsPromptCache: false,
1913
+ inputPrice: 0,
1914
+ outputPrice: 0,
1915
+ family: "o3-mini",
1916
+ version: "o3-mini-2025-01-31",
1917
+ name: "o3-mini",
1918
+ supportsToolCalling: true,
1919
+ maxInputTokens: 63827
1920
+ },
1921
+ "claude-3.5-sonnet": {
1922
+ contextWindow: 81638,
1923
+ supportsImages: true,
1924
+ supportsPromptCache: false,
1925
+ inputPrice: 0,
1926
+ outputPrice: 0,
1927
+ family: "claude-3.5-sonnet",
1928
+ version: "claude-3.5-sonnet",
1929
+ name: "Claude 3.5 Sonnet",
1930
+ supportsToolCalling: true,
1931
+ maxInputTokens: 81638
1932
+ },
1933
+ "gemini-2.0-flash-001": {
1934
+ contextWindow: 127827,
1935
+ supportsImages: true,
1936
+ supportsPromptCache: false,
1937
+ inputPrice: 0,
1938
+ outputPrice: 0,
1939
+ family: "gemini-2.0-flash",
1940
+ version: "gemini-2.0-flash-001",
1941
+ name: "Gemini 2.0 Flash",
1942
+ supportsToolCalling: false,
1943
+ maxInputTokens: 127827
1944
+ },
1945
+ "gemini-2.5-pro": {
1946
+ contextWindow: 63830,
1947
+ supportsImages: true,
1948
+ supportsPromptCache: false,
1949
+ inputPrice: 0,
1950
+ outputPrice: 0,
1951
+ family: "gemini-2.5-pro",
1952
+ version: "gemini-2.5-pro-preview-03-25",
1953
+ name: "Gemini 2.5 Pro (Preview)",
1954
+ supportsToolCalling: true,
1955
+ maxInputTokens: 63830
1956
+ },
1957
+ "o4-mini": {
1958
+ contextWindow: 111446,
1959
+ supportsImages: false,
1960
+ supportsPromptCache: false,
1961
+ inputPrice: 0,
1962
+ outputPrice: 0,
1963
+ family: "o4-mini",
1964
+ version: "o4-mini-2025-04-16",
1965
+ name: "o4-mini (Preview)",
1966
+ supportsToolCalling: true,
1967
+ maxInputTokens: 111446
1968
+ },
1969
+ "gpt-4.1": {
1970
+ contextWindow: 111446,
1971
+ supportsImages: true,
1972
+ supportsPromptCache: false,
1973
+ inputPrice: 0,
1974
+ outputPrice: 0,
1975
+ family: "gpt-4.1",
1976
+ version: "gpt-4.1-2025-04-14",
1977
+ name: "GPT-4.1 (Preview)",
1978
+ supportsToolCalling: true,
1979
+ maxInputTokens: 111446
1980
+ }
1981
+ };
1982
+
1983
+ // src/providers/xai.ts
1984
+ var xaiDefaultModelId = "grok-3";
1985
+ var xaiModels = {
1986
+ "grok-3-beta": {
1987
+ maxTokens: 8192,
1988
+ contextWindow: 131072,
1989
+ supportsImages: false,
1990
+ supportsPromptCache: false,
1991
+ inputPrice: 3,
1992
+ outputPrice: 15,
1993
+ description: "xAI's Grok-3 beta model with 131K context window"
1994
+ },
1995
+ "grok-3-fast-beta": {
1996
+ maxTokens: 8192,
1997
+ contextWindow: 131072,
1998
+ supportsImages: false,
1999
+ supportsPromptCache: false,
2000
+ inputPrice: 5,
2001
+ outputPrice: 25,
2002
+ description: "xAI's Grok-3 fast beta model with 131K context window"
2003
+ },
2004
+ "grok-3-mini-beta": {
2005
+ maxTokens: 8192,
2006
+ contextWindow: 131072,
2007
+ supportsImages: false,
2008
+ supportsPromptCache: false,
2009
+ inputPrice: 0.3,
2010
+ outputPrice: 0.5,
2011
+ description: "xAI's Grok-3 mini beta model with 131K context window",
2012
+ supportsReasoningEffort: true
2013
+ },
2014
+ "grok-3-mini-fast-beta": {
2015
+ maxTokens: 8192,
2016
+ contextWindow: 131072,
2017
+ supportsImages: false,
2018
+ supportsPromptCache: false,
2019
+ inputPrice: 0.6,
2020
+ outputPrice: 4,
2021
+ description: "xAI's Grok-3 mini fast beta model with 131K context window",
2022
+ supportsReasoningEffort: true
2023
+ },
2024
+ "grok-3": {
2025
+ maxTokens: 8192,
2026
+ contextWindow: 131072,
2027
+ supportsImages: false,
2028
+ supportsPromptCache: false,
2029
+ inputPrice: 3,
2030
+ outputPrice: 15,
2031
+ description: "xAI's Grok-3 model with 131K context window"
2032
+ },
2033
+ "grok-3-fast": {
2034
+ maxTokens: 8192,
2035
+ contextWindow: 131072,
2036
+ supportsImages: false,
2037
+ supportsPromptCache: false,
2038
+ inputPrice: 5,
2039
+ outputPrice: 25,
2040
+ description: "xAI's Grok-3 fast model with 131K context window"
2041
+ },
2042
+ "grok-3-mini": {
2043
+ maxTokens: 8192,
2044
+ contextWindow: 131072,
2045
+ supportsImages: false,
2046
+ supportsPromptCache: false,
2047
+ inputPrice: 0.3,
2048
+ outputPrice: 0.5,
2049
+ description: "xAI's Grok-3 mini model with 131K context window",
2050
+ supportsReasoningEffort: true
2051
+ },
2052
+ "grok-3-mini-fast": {
2053
+ maxTokens: 8192,
2054
+ contextWindow: 131072,
2055
+ supportsImages: false,
2056
+ supportsPromptCache: false,
2057
+ inputPrice: 0.6,
2058
+ outputPrice: 4,
2059
+ description: "xAI's Grok-3 mini fast model with 131K context window",
2060
+ supportsReasoningEffort: true
2061
+ },
2062
+ "grok-2-latest": {
2063
+ maxTokens: 8192,
2064
+ contextWindow: 131072,
2065
+ supportsImages: false,
2066
+ supportsPromptCache: false,
2067
+ inputPrice: 2,
2068
+ outputPrice: 10,
2069
+ description: "xAI's Grok-2 model - latest version with 131K context window"
2070
+ },
2071
+ "grok-2": {
2072
+ maxTokens: 8192,
2073
+ contextWindow: 131072,
2074
+ supportsImages: false,
2075
+ supportsPromptCache: false,
2076
+ inputPrice: 2,
2077
+ outputPrice: 10,
2078
+ description: "xAI's Grok-2 model with 131K context window"
2079
+ },
2080
+ "grok-2-1212": {
2081
+ maxTokens: 8192,
2082
+ contextWindow: 131072,
2083
+ supportsImages: false,
2084
+ supportsPromptCache: false,
2085
+ inputPrice: 2,
2086
+ outputPrice: 10,
2087
+ description: "xAI's Grok-2 model (version 1212) with 131K context window"
2088
+ },
2089
+ "grok-2-vision-latest": {
2090
+ maxTokens: 8192,
2091
+ contextWindow: 32768,
2092
+ supportsImages: true,
2093
+ supportsPromptCache: false,
2094
+ inputPrice: 2,
2095
+ outputPrice: 10,
2096
+ description: "xAI's Grok-2 Vision model - latest version with image support and 32K context window"
2097
+ },
2098
+ "grok-2-vision": {
2099
+ maxTokens: 8192,
2100
+ contextWindow: 32768,
2101
+ supportsImages: true,
2102
+ supportsPromptCache: false,
2103
+ inputPrice: 2,
2104
+ outputPrice: 10,
2105
+ description: "xAI's Grok-2 Vision model with image support and 32K context window"
2106
+ },
2107
+ "grok-2-vision-1212": {
2108
+ maxTokens: 8192,
2109
+ contextWindow: 32768,
2110
+ supportsImages: true,
2111
+ supportsPromptCache: false,
2112
+ inputPrice: 2,
2113
+ outputPrice: 10,
2114
+ description: "xAI's Grok-2 Vision model (version 1212) with image support and 32K context window"
2115
+ },
2116
+ "grok-vision-beta": {
2117
+ maxTokens: 8192,
2118
+ contextWindow: 8192,
2119
+ supportsImages: true,
2120
+ supportsPromptCache: false,
2121
+ inputPrice: 5,
2122
+ outputPrice: 15,
2123
+ description: "xAI's Grok Vision Beta model with image support and 8K context window"
2124
+ },
2125
+ "grok-beta": {
2126
+ maxTokens: 8192,
2127
+ contextWindow: 131072,
2128
+ supportsImages: false,
2129
+ supportsPromptCache: false,
2130
+ inputPrice: 5,
2131
+ outputPrice: 15,
2132
+ description: "xAI's Grok Beta model (legacy) with 131K context window"
2133
+ }
2134
+ };
2135
+
99
2136
  // src/codebase-index.ts
100
2137
  var import_zod = require("zod");
101
2138
  var codebaseIndexConfigSchema = import_zod.z.object({
@@ -139,16 +2176,19 @@ var organizationSettingsSchema = import_zod2.z.object({
139
2176
  maxReadFileLine: import_zod2.z.number().optional(),
140
2177
  fuzzyMatchThreshold: import_zod2.z.number().optional()
141
2178
  }).optional(),
2179
+ cloudSettings: import_zod2.z.object({
2180
+ recordTaskMessages: import_zod2.z.boolean().optional()
2181
+ }).optional(),
142
2182
  allowList: organizationAllowListSchema
143
2183
  });
144
2184
 
145
2185
  // src/experiment.ts
146
2186
  var import_zod3 = require("zod");
147
- var experimentIds = ["autoCondenseContext", "powerSteering"];
2187
+ var experimentIds = ["powerSteering", "concurrentFileReads"];
148
2188
  var experimentIdsSchema = import_zod3.z.enum(experimentIds);
149
2189
  var experimentsSchema = import_zod3.z.object({
150
- autoCondenseContext: import_zod3.z.boolean(),
151
- powerSteering: import_zod3.z.boolean()
2190
+ powerSteering: import_zod3.z.boolean(),
2191
+ concurrentFileReads: import_zod3.z.boolean()
152
2192
  });
153
2193
 
154
2194
  // src/global-settings.ts
@@ -270,7 +2310,9 @@ var bedrockSchema = apiModelIdProviderModelSchema.extend({
270
2310
  awsUsePromptCache: import_zod5.z.boolean().optional(),
271
2311
  awsProfile: import_zod5.z.string().optional(),
272
2312
  awsUseProfile: import_zod5.z.boolean().optional(),
273
- awsCustomArn: import_zod5.z.string().optional()
2313
+ awsCustomArn: import_zod5.z.string().optional(),
2314
+ awsBedrockEndpointEnabled: import_zod5.z.boolean().optional(),
2315
+ awsBedrockEndpoint: import_zod5.z.string().optional()
274
2316
  });
275
2317
  var vertexSchema = apiModelIdProviderModelSchema.extend({
276
2318
  vertexKeyFile: import_zod5.z.string().optional(),
@@ -430,6 +2472,8 @@ var PROVIDER_SETTINGS_KEYS = keysOf()([
430
2472
  "awsProfile",
431
2473
  "awsUseProfile",
432
2474
  "awsCustomArn",
2475
+ "awsBedrockEndpointEnabled",
2476
+ "awsBedrockEndpoint",
433
2477
  // Google Vertex
434
2478
  "vertexKeyFile",
435
2479
  "vertexJsonCredentials",
@@ -562,6 +2606,7 @@ var clineSays = [
562
2606
  "rooignore_error",
563
2607
  "diff_error",
564
2608
  "condense_context",
2609
+ "condense_context_error",
565
2610
  "codebase_search_result"
566
2611
  ];
567
2612
  var clineSaySchema = import_zod7.z.enum(clineSays);
@@ -626,6 +2671,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
626
2671
  return TelemetryEventName2;
627
2672
  })(TelemetryEventName || {});
628
2673
  var appPropertiesSchema = import_zod8.z.object({
2674
+ appName: import_zod8.z.string(),
629
2675
  appVersion: import_zod8.z.string(),
630
2676
  vscodeVersion: import_zod8.z.string(),
631
2677
  platform: import_zod8.z.string(),
@@ -663,13 +2709,16 @@ var rooCodeTelemetryEventSchema = import_zod8.z.discriminatedUnion("type", [
663
2709
  "Schema Validation Error" /* SCHEMA_VALIDATION_ERROR */,
664
2710
  "Diff Application Error" /* DIFF_APPLICATION_ERROR */,
665
2711
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
666
- "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */
2712
+ "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */,
2713
+ "Context Condensed" /* CONTEXT_CONDENSED */,
2714
+ "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */
667
2715
  ]),
668
2716
  properties: telemetryPropertiesSchema
669
2717
  }),
670
2718
  import_zod8.z.object({
671
2719
  type: import_zod8.z.literal("Task Message" /* TASK_MESSAGE */),
672
2720
  properties: import_zod8.z.object({
2721
+ ...telemetryPropertiesSchema.shape,
673
2722
  taskId: import_zod8.z.string(),
674
2723
  message: clineMessageSchema
675
2724
  })
@@ -860,7 +2909,9 @@ var globalSettingsSchema = import_zod12.z.object({
860
2909
  alwaysAllowExecute: import_zod12.z.boolean().optional(),
861
2910
  allowedCommands: import_zod12.z.array(import_zod12.z.string()).optional(),
862
2911
  allowedMaxRequests: import_zod12.z.number().nullish(),
2912
+ autoCondenseContext: import_zod12.z.boolean().optional(),
863
2913
  autoCondenseContextPercent: import_zod12.z.number().optional(),
2914
+ maxConcurrentFileReads: import_zod12.z.number().optional(),
864
2915
  browserToolEnabled: import_zod12.z.boolean().optional(),
865
2916
  browserViewportSize: import_zod12.z.string().optional(),
866
2917
  screenshotQuality: import_zod12.z.number().optional(),
@@ -928,7 +2979,9 @@ var GLOBAL_SETTINGS_KEYS = keysOf()([
928
2979
  "alwaysAllowExecute",
929
2980
  "allowedCommands",
930
2981
  "allowedMaxRequests",
2982
+ "autoCondenseContext",
931
2983
  "autoCondenseContextPercent",
2984
+ "maxConcurrentFileReads",
932
2985
  "browserToolEnabled",
933
2986
  "browserViewportSize",
934
2987
  "screenshotQuality",
@@ -1172,18 +3225,44 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1172
3225
  ]);
1173
3226
  // Annotate the CommonJS export names for ESM import in node:
1174
3227
  0 && (module.exports = {
3228
+ ANTHROPIC_DEFAULT_MAX_TOKENS,
3229
+ BEDROCK_DEFAULT_TEMPERATURE,
3230
+ BEDROCK_MAX_TOKENS,
3231
+ BEDROCK_REGIONS,
3232
+ BEDROCK_REGION_INFO,
3233
+ DEEP_SEEK_DEFAULT_TEMPERATURE,
3234
+ GLAMA_DEFAULT_TEMPERATURE,
1175
3235
  GLOBAL_SETTINGS_KEYS,
1176
3236
  GLOBAL_STATE_KEYS,
1177
3237
  IpcMessageType,
1178
3238
  IpcOrigin,
3239
+ LITELLM_COMPUTER_USE_MODELS,
3240
+ LMSTUDIO_DEFAULT_TEMPERATURE,
3241
+ MISTRAL_DEFAULT_TEMPERATURE,
3242
+ OPENAI_AZURE_AI_INFERENCE_PATH,
3243
+ OPENAI_NATIVE_DEFAULT_TEMPERATURE,
3244
+ OPENROUTER_DEFAULT_PROVIDER_NAME,
3245
+ OPEN_ROUTER_COMPUTER_USE_MODELS,
3246
+ OPEN_ROUTER_PROMPT_CACHING_MODELS,
3247
+ OPEN_ROUTER_REASONING_BUDGET_MODELS,
3248
+ OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS,
1179
3249
  ORGANIZATION_ALLOW_ALL,
1180
3250
  PROVIDER_SETTINGS_KEYS,
1181
3251
  RooCodeEventName,
1182
3252
  SECRET_STATE_KEYS,
1183
3253
  TaskCommandName,
1184
3254
  TelemetryEventName,
3255
+ VERTEX_REGIONS,
1185
3256
  ackSchema,
3257
+ anthropicDefaultModelId,
3258
+ anthropicModels,
1186
3259
  appPropertiesSchema,
3260
+ azureOpenAiDefaultApiVersion,
3261
+ bedrockDefaultModelId,
3262
+ bedrockDefaultPromptRouterModelId,
3263
+ bedrockModels,
3264
+ chutesDefaultModelId,
3265
+ chutesModels,
1187
3266
  clineAskSchema,
1188
3267
  clineAsks,
1189
3268
  clineMessageSchema,
@@ -1199,10 +3278,18 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1199
3278
  customModePromptsSchema,
1200
3279
  customModesSettingsSchema,
1201
3280
  customSupportPromptsSchema,
3281
+ deepSeekDefaultModelId,
3282
+ deepSeekModels,
1202
3283
  experimentIds,
1203
3284
  experimentIdsSchema,
1204
3285
  experimentsSchema,
3286
+ geminiDefaultModelId,
3287
+ geminiModels,
3288
+ glamaDefaultModelId,
3289
+ glamaDefaultModelInfo,
1205
3290
  globalSettingsSchema,
3291
+ groqDefaultModelId,
3292
+ groqModels,
1206
3293
  groupEntrySchema,
1207
3294
  groupOptionsSchema,
1208
3295
  historyItemSchema,
@@ -1214,10 +3301,19 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1214
3301
  keysOf,
1215
3302
  languages,
1216
3303
  languagesSchema,
3304
+ litellmDefaultModelId,
3305
+ litellmDefaultModelInfo,
3306
+ mistralDefaultModelId,
3307
+ mistralModels,
1217
3308
  modeConfigSchema,
1218
3309
  modelInfoSchema,
1219
3310
  modelParameters,
1220
3311
  modelParametersSchema,
3312
+ openAiModelInfoSaneDefaults,
3313
+ openAiNativeDefaultModelId,
3314
+ openAiNativeModels,
3315
+ openRouterDefaultModelId,
3316
+ openRouterDefaultModelInfo,
1221
3317
  organizationAllowListSchema,
1222
3318
  organizationSettingsSchema,
1223
3319
  promptComponentSchema,
@@ -1228,6 +3324,8 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1228
3324
  providerSettingsSchemaDiscriminated,
1229
3325
  reasoningEfforts,
1230
3326
  reasoningEffortsSchema,
3327
+ requestyDefaultModelId,
3328
+ requestyDefaultModelInfo,
1231
3329
  rooCodeEventsSchema,
1232
3330
  rooCodeSettingsSchema,
1233
3331
  rooCodeTelemetryEventSchema,
@@ -1244,6 +3342,14 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1244
3342
  toolNames,
1245
3343
  toolNamesSchema,
1246
3344
  toolProgressStatusSchema,
1247
- toolUsageSchema
3345
+ toolUsageSchema,
3346
+ unboundDefaultModelId,
3347
+ unboundDefaultModelInfo,
3348
+ vertexDefaultModelId,
3349
+ vertexModels,
3350
+ vscodeLlmDefaultModelId,
3351
+ vscodeLlmModels,
3352
+ xaiDefaultModelId,
3353
+ xaiModels
1248
3354
  });
1249
3355
  //# sourceMappingURL=index.cjs.map