@roo-code/types 1.21.0 → 1.23.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -20,18 +20,44 @@ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: tru
20
20
  // src/index.ts
21
21
  var index_exports = {};
22
22
  __export(index_exports, {
23
+ ANTHROPIC_DEFAULT_MAX_TOKENS: () => ANTHROPIC_DEFAULT_MAX_TOKENS,
24
+ BEDROCK_DEFAULT_TEMPERATURE: () => BEDROCK_DEFAULT_TEMPERATURE,
25
+ BEDROCK_MAX_TOKENS: () => BEDROCK_MAX_TOKENS,
26
+ BEDROCK_REGIONS: () => BEDROCK_REGIONS,
27
+ BEDROCK_REGION_INFO: () => BEDROCK_REGION_INFO,
28
+ DEEP_SEEK_DEFAULT_TEMPERATURE: () => DEEP_SEEK_DEFAULT_TEMPERATURE,
29
+ GLAMA_DEFAULT_TEMPERATURE: () => GLAMA_DEFAULT_TEMPERATURE,
23
30
  GLOBAL_SETTINGS_KEYS: () => GLOBAL_SETTINGS_KEYS,
24
31
  GLOBAL_STATE_KEYS: () => GLOBAL_STATE_KEYS,
25
32
  IpcMessageType: () => IpcMessageType,
26
33
  IpcOrigin: () => IpcOrigin,
34
+ LITELLM_COMPUTER_USE_MODELS: () => LITELLM_COMPUTER_USE_MODELS,
35
+ LMSTUDIO_DEFAULT_TEMPERATURE: () => LMSTUDIO_DEFAULT_TEMPERATURE,
36
+ MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
37
+ OPENAI_AZURE_AI_INFERENCE_PATH: () => OPENAI_AZURE_AI_INFERENCE_PATH,
38
+ OPENAI_NATIVE_DEFAULT_TEMPERATURE: () => OPENAI_NATIVE_DEFAULT_TEMPERATURE,
39
+ OPENROUTER_DEFAULT_PROVIDER_NAME: () => OPENROUTER_DEFAULT_PROVIDER_NAME,
40
+ OPEN_ROUTER_COMPUTER_USE_MODELS: () => OPEN_ROUTER_COMPUTER_USE_MODELS,
41
+ OPEN_ROUTER_PROMPT_CACHING_MODELS: () => OPEN_ROUTER_PROMPT_CACHING_MODELS,
42
+ OPEN_ROUTER_REASONING_BUDGET_MODELS: () => OPEN_ROUTER_REASONING_BUDGET_MODELS,
43
+ OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS: () => OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS,
27
44
  ORGANIZATION_ALLOW_ALL: () => ORGANIZATION_ALLOW_ALL,
28
45
  PROVIDER_SETTINGS_KEYS: () => PROVIDER_SETTINGS_KEYS,
29
46
  RooCodeEventName: () => RooCodeEventName,
30
47
  SECRET_STATE_KEYS: () => SECRET_STATE_KEYS,
31
48
  TaskCommandName: () => TaskCommandName,
32
49
  TelemetryEventName: () => TelemetryEventName,
50
+ VERTEX_REGIONS: () => VERTEX_REGIONS,
33
51
  ackSchema: () => ackSchema,
52
+ anthropicDefaultModelId: () => anthropicDefaultModelId,
53
+ anthropicModels: () => anthropicModels,
34
54
  appPropertiesSchema: () => appPropertiesSchema,
55
+ azureOpenAiDefaultApiVersion: () => azureOpenAiDefaultApiVersion,
56
+ bedrockDefaultModelId: () => bedrockDefaultModelId,
57
+ bedrockDefaultPromptRouterModelId: () => bedrockDefaultPromptRouterModelId,
58
+ bedrockModels: () => bedrockModels,
59
+ chutesDefaultModelId: () => chutesDefaultModelId,
60
+ chutesModels: () => chutesModels,
35
61
  clineAskSchema: () => clineAskSchema,
36
62
  clineAsks: () => clineAsks,
37
63
  clineMessageSchema: () => clineMessageSchema,
@@ -47,10 +73,18 @@ __export(index_exports, {
47
73
  customModePromptsSchema: () => customModePromptsSchema,
48
74
  customModesSettingsSchema: () => customModesSettingsSchema,
49
75
  customSupportPromptsSchema: () => customSupportPromptsSchema,
76
+ deepSeekDefaultModelId: () => deepSeekDefaultModelId,
77
+ deepSeekModels: () => deepSeekModels,
50
78
  experimentIds: () => experimentIds,
51
79
  experimentIdsSchema: () => experimentIdsSchema,
52
80
  experimentsSchema: () => experimentsSchema,
81
+ geminiDefaultModelId: () => geminiDefaultModelId,
82
+ geminiModels: () => geminiModels,
83
+ glamaDefaultModelId: () => glamaDefaultModelId,
84
+ glamaDefaultModelInfo: () => glamaDefaultModelInfo,
53
85
  globalSettingsSchema: () => globalSettingsSchema,
86
+ groqDefaultModelId: () => groqDefaultModelId,
87
+ groqModels: () => groqModels,
54
88
  groupEntrySchema: () => groupEntrySchema,
55
89
  groupOptionsSchema: () => groupOptionsSchema,
56
90
  historyItemSchema: () => historyItemSchema,
@@ -62,10 +96,19 @@ __export(index_exports, {
62
96
  keysOf: () => keysOf,
63
97
  languages: () => languages,
64
98
  languagesSchema: () => languagesSchema,
99
+ litellmDefaultModelId: () => litellmDefaultModelId,
100
+ litellmDefaultModelInfo: () => litellmDefaultModelInfo,
101
+ mistralDefaultModelId: () => mistralDefaultModelId,
102
+ mistralModels: () => mistralModels,
65
103
  modeConfigSchema: () => modeConfigSchema,
66
104
  modelInfoSchema: () => modelInfoSchema,
67
105
  modelParameters: () => modelParameters,
68
106
  modelParametersSchema: () => modelParametersSchema,
107
+ openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
108
+ openAiNativeDefaultModelId: () => openAiNativeDefaultModelId,
109
+ openAiNativeModels: () => openAiNativeModels,
110
+ openRouterDefaultModelId: () => openRouterDefaultModelId,
111
+ openRouterDefaultModelInfo: () => openRouterDefaultModelInfo,
69
112
  organizationAllowListSchema: () => organizationAllowListSchema,
70
113
  organizationSettingsSchema: () => organizationSettingsSchema,
71
114
  promptComponentSchema: () => promptComponentSchema,
@@ -76,6 +119,8 @@ __export(index_exports, {
76
119
  providerSettingsSchemaDiscriminated: () => providerSettingsSchemaDiscriminated,
77
120
  reasoningEfforts: () => reasoningEfforts,
78
121
  reasoningEffortsSchema: () => reasoningEffortsSchema,
122
+ requestyDefaultModelId: () => requestyDefaultModelId,
123
+ requestyDefaultModelInfo: () => requestyDefaultModelInfo,
79
124
  rooCodeEventsSchema: () => rooCodeEventsSchema,
80
125
  rooCodeSettingsSchema: () => rooCodeSettingsSchema,
81
126
  rooCodeTelemetryEventSchema: () => rooCodeTelemetryEventSchema,
@@ -92,10 +137,2002 @@ __export(index_exports, {
92
137
  toolNames: () => toolNames,
93
138
  toolNamesSchema: () => toolNamesSchema,
94
139
  toolProgressStatusSchema: () => toolProgressStatusSchema,
95
- toolUsageSchema: () => toolUsageSchema
140
+ toolUsageSchema: () => toolUsageSchema,
141
+ unboundDefaultModelId: () => unboundDefaultModelId,
142
+ unboundDefaultModelInfo: () => unboundDefaultModelInfo,
143
+ vertexDefaultModelId: () => vertexDefaultModelId,
144
+ vertexModels: () => vertexModels,
145
+ vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
146
+ vscodeLlmModels: () => vscodeLlmModels,
147
+ xaiDefaultModelId: () => xaiDefaultModelId,
148
+ xaiModels: () => xaiModels
96
149
  });
97
150
  module.exports = __toCommonJS(index_exports);
98
151
 
152
+ // src/providers/anthropic.ts
153
+ var anthropicDefaultModelId = "claude-sonnet-4-20250514";
154
+ var anthropicModels = {
155
+ "claude-sonnet-4-20250514": {
156
+ maxTokens: 64e3,
157
+ // Overridden to 8k if `enableReasoningEffort` is false.
158
+ contextWindow: 2e5,
159
+ supportsImages: true,
160
+ supportsComputerUse: true,
161
+ supportsPromptCache: true,
162
+ inputPrice: 3,
163
+ // $3 per million input tokens
164
+ outputPrice: 15,
165
+ // $15 per million output tokens
166
+ cacheWritesPrice: 3.75,
167
+ // $3.75 per million tokens
168
+ cacheReadsPrice: 0.3,
169
+ // $0.30 per million tokens
170
+ supportsReasoningBudget: true
171
+ },
172
+ "claude-opus-4-20250514": {
173
+ maxTokens: 32e3,
174
+ // Overridden to 8k if `enableReasoningEffort` is false.
175
+ contextWindow: 2e5,
176
+ supportsImages: true,
177
+ supportsComputerUse: true,
178
+ supportsPromptCache: true,
179
+ inputPrice: 15,
180
+ // $15 per million input tokens
181
+ outputPrice: 75,
182
+ // $75 per million output tokens
183
+ cacheWritesPrice: 18.75,
184
+ // $18.75 per million tokens
185
+ cacheReadsPrice: 1.5,
186
+ // $1.50 per million tokens
187
+ supportsReasoningBudget: true
188
+ },
189
+ "claude-3-7-sonnet-20250219:thinking": {
190
+ maxTokens: 128e3,
191
+ // Unlocked by passing `beta` flag to the model. Otherwise, it's 64k.
192
+ contextWindow: 2e5,
193
+ supportsImages: true,
194
+ supportsComputerUse: true,
195
+ supportsPromptCache: true,
196
+ inputPrice: 3,
197
+ // $3 per million input tokens
198
+ outputPrice: 15,
199
+ // $15 per million output tokens
200
+ cacheWritesPrice: 3.75,
201
+ // $3.75 per million tokens
202
+ cacheReadsPrice: 0.3,
203
+ // $0.30 per million tokens
204
+ supportsReasoningBudget: true,
205
+ requiredReasoningBudget: true
206
+ },
207
+ "claude-3-7-sonnet-20250219": {
208
+ maxTokens: 8192,
209
+ // Since we already have a `:thinking` virtual model we aren't setting `supportsReasoningBudget: true` here.
210
+ contextWindow: 2e5,
211
+ supportsImages: true,
212
+ supportsComputerUse: true,
213
+ supportsPromptCache: true,
214
+ inputPrice: 3,
215
+ // $3 per million input tokens
216
+ outputPrice: 15,
217
+ // $15 per million output tokens
218
+ cacheWritesPrice: 3.75,
219
+ // $3.75 per million tokens
220
+ cacheReadsPrice: 0.3
221
+ // $0.30 per million tokens
222
+ },
223
+ "claude-3-5-sonnet-20241022": {
224
+ maxTokens: 8192,
225
+ contextWindow: 2e5,
226
+ supportsImages: true,
227
+ supportsComputerUse: true,
228
+ supportsPromptCache: true,
229
+ inputPrice: 3,
230
+ // $3 per million input tokens
231
+ outputPrice: 15,
232
+ // $15 per million output tokens
233
+ cacheWritesPrice: 3.75,
234
+ // $3.75 per million tokens
235
+ cacheReadsPrice: 0.3
236
+ // $0.30 per million tokens
237
+ },
238
+ "claude-3-5-haiku-20241022": {
239
+ maxTokens: 8192,
240
+ contextWindow: 2e5,
241
+ supportsImages: false,
242
+ supportsPromptCache: true,
243
+ inputPrice: 1,
244
+ outputPrice: 5,
245
+ cacheWritesPrice: 1.25,
246
+ cacheReadsPrice: 0.1
247
+ },
248
+ "claude-3-opus-20240229": {
249
+ maxTokens: 4096,
250
+ contextWindow: 2e5,
251
+ supportsImages: true,
252
+ supportsPromptCache: true,
253
+ inputPrice: 15,
254
+ outputPrice: 75,
255
+ cacheWritesPrice: 18.75,
256
+ cacheReadsPrice: 1.5
257
+ },
258
+ "claude-3-haiku-20240307": {
259
+ maxTokens: 4096,
260
+ contextWindow: 2e5,
261
+ supportsImages: true,
262
+ supportsPromptCache: true,
263
+ inputPrice: 0.25,
264
+ outputPrice: 1.25,
265
+ cacheWritesPrice: 0.3,
266
+ cacheReadsPrice: 0.03
267
+ }
268
+ };
269
+ var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
270
+
271
+ // src/providers/bedrock.ts
272
+ var bedrockDefaultModelId = "anthropic.claude-sonnet-4-20250514-v1:0";
273
+ var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
274
+ var bedrockModels = {
275
+ "amazon.nova-pro-v1:0": {
276
+ maxTokens: 5e3,
277
+ contextWindow: 3e5,
278
+ supportsImages: true,
279
+ supportsComputerUse: false,
280
+ supportsPromptCache: true,
281
+ inputPrice: 0.8,
282
+ outputPrice: 3.2,
283
+ cacheWritesPrice: 0.8,
284
+ // per million tokens
285
+ cacheReadsPrice: 0.2,
286
+ // per million tokens
287
+ minTokensPerCachePoint: 1,
288
+ maxCachePoints: 1,
289
+ cachableFields: ["system"]
290
+ },
291
+ "amazon.nova-pro-latency-optimized-v1:0": {
292
+ maxTokens: 5e3,
293
+ contextWindow: 3e5,
294
+ supportsImages: true,
295
+ supportsComputerUse: false,
296
+ supportsPromptCache: false,
297
+ inputPrice: 1,
298
+ outputPrice: 4,
299
+ cacheWritesPrice: 1,
300
+ // per million tokens
301
+ cacheReadsPrice: 0.25,
302
+ // per million tokens
303
+ description: "Amazon Nova Pro with latency optimized inference"
304
+ },
305
+ "amazon.nova-lite-v1:0": {
306
+ maxTokens: 5e3,
307
+ contextWindow: 3e5,
308
+ supportsImages: true,
309
+ supportsComputerUse: false,
310
+ supportsPromptCache: true,
311
+ inputPrice: 0.06,
312
+ outputPrice: 0.24,
313
+ cacheWritesPrice: 0.06,
314
+ // per million tokens
315
+ cacheReadsPrice: 0.015,
316
+ // per million tokens
317
+ minTokensPerCachePoint: 1,
318
+ maxCachePoints: 1,
319
+ cachableFields: ["system"]
320
+ },
321
+ "amazon.nova-micro-v1:0": {
322
+ maxTokens: 5e3,
323
+ contextWindow: 128e3,
324
+ supportsImages: false,
325
+ supportsComputerUse: false,
326
+ supportsPromptCache: true,
327
+ inputPrice: 0.035,
328
+ outputPrice: 0.14,
329
+ cacheWritesPrice: 0.035,
330
+ // per million tokens
331
+ cacheReadsPrice: 875e-5,
332
+ // per million tokens
333
+ minTokensPerCachePoint: 1,
334
+ maxCachePoints: 1,
335
+ cachableFields: ["system"]
336
+ },
337
+ "anthropic.claude-sonnet-4-20250514-v1:0": {
338
+ maxTokens: 8192,
339
+ contextWindow: 2e5,
340
+ supportsImages: true,
341
+ supportsComputerUse: true,
342
+ supportsPromptCache: true,
343
+ inputPrice: 3,
344
+ outputPrice: 15,
345
+ cacheWritesPrice: 3.75,
346
+ cacheReadsPrice: 0.3,
347
+ minTokensPerCachePoint: 1024,
348
+ maxCachePoints: 4,
349
+ cachableFields: ["system", "messages", "tools"]
350
+ },
351
+ "anthropic.claude-opus-4-20250514-v1:0": {
352
+ maxTokens: 8192,
353
+ contextWindow: 2e5,
354
+ supportsImages: true,
355
+ supportsComputerUse: true,
356
+ supportsPromptCache: true,
357
+ inputPrice: 15,
358
+ outputPrice: 75,
359
+ cacheWritesPrice: 18.75,
360
+ cacheReadsPrice: 1.5,
361
+ minTokensPerCachePoint: 1024,
362
+ maxCachePoints: 4,
363
+ cachableFields: ["system", "messages", "tools"]
364
+ },
365
+ "anthropic.claude-3-7-sonnet-20250219-v1:0": {
366
+ maxTokens: 8192,
367
+ contextWindow: 2e5,
368
+ supportsImages: true,
369
+ supportsComputerUse: true,
370
+ supportsPromptCache: true,
371
+ inputPrice: 3,
372
+ outputPrice: 15,
373
+ cacheWritesPrice: 3.75,
374
+ cacheReadsPrice: 0.3,
375
+ minTokensPerCachePoint: 1024,
376
+ maxCachePoints: 4,
377
+ cachableFields: ["system", "messages", "tools"]
378
+ },
379
+ "anthropic.claude-3-5-sonnet-20241022-v2:0": {
380
+ maxTokens: 8192,
381
+ contextWindow: 2e5,
382
+ supportsImages: true,
383
+ supportsComputerUse: true,
384
+ supportsPromptCache: true,
385
+ inputPrice: 3,
386
+ outputPrice: 15,
387
+ cacheWritesPrice: 3.75,
388
+ cacheReadsPrice: 0.3,
389
+ minTokensPerCachePoint: 1024,
390
+ maxCachePoints: 4,
391
+ cachableFields: ["system", "messages", "tools"]
392
+ },
393
+ "anthropic.claude-3-5-haiku-20241022-v1:0": {
394
+ maxTokens: 8192,
395
+ contextWindow: 2e5,
396
+ supportsImages: false,
397
+ supportsPromptCache: true,
398
+ inputPrice: 0.8,
399
+ outputPrice: 4,
400
+ cacheWritesPrice: 1,
401
+ cacheReadsPrice: 0.08,
402
+ minTokensPerCachePoint: 2048,
403
+ maxCachePoints: 4,
404
+ cachableFields: ["system", "messages", "tools"]
405
+ },
406
+ "anthropic.claude-3-5-sonnet-20240620-v1:0": {
407
+ maxTokens: 8192,
408
+ contextWindow: 2e5,
409
+ supportsImages: true,
410
+ supportsPromptCache: false,
411
+ inputPrice: 3,
412
+ outputPrice: 15
413
+ },
414
+ "anthropic.claude-3-opus-20240229-v1:0": {
415
+ maxTokens: 4096,
416
+ contextWindow: 2e5,
417
+ supportsImages: true,
418
+ supportsPromptCache: false,
419
+ inputPrice: 15,
420
+ outputPrice: 75
421
+ },
422
+ "anthropic.claude-3-sonnet-20240229-v1:0": {
423
+ maxTokens: 4096,
424
+ contextWindow: 2e5,
425
+ supportsImages: true,
426
+ supportsPromptCache: false,
427
+ inputPrice: 3,
428
+ outputPrice: 15
429
+ },
430
+ "anthropic.claude-3-haiku-20240307-v1:0": {
431
+ maxTokens: 4096,
432
+ contextWindow: 2e5,
433
+ supportsImages: true,
434
+ supportsPromptCache: false,
435
+ inputPrice: 0.25,
436
+ outputPrice: 1.25
437
+ },
438
+ "anthropic.claude-2-1-v1:0": {
439
+ maxTokens: 4096,
440
+ contextWindow: 1e5,
441
+ supportsImages: false,
442
+ supportsPromptCache: false,
443
+ inputPrice: 8,
444
+ outputPrice: 24,
445
+ description: "Claude 2.1"
446
+ },
447
+ "anthropic.claude-2-0-v1:0": {
448
+ maxTokens: 4096,
449
+ contextWindow: 1e5,
450
+ supportsImages: false,
451
+ supportsPromptCache: false,
452
+ inputPrice: 8,
453
+ outputPrice: 24,
454
+ description: "Claude 2.0"
455
+ },
456
+ "anthropic.claude-instant-v1:0": {
457
+ maxTokens: 4096,
458
+ contextWindow: 1e5,
459
+ supportsImages: false,
460
+ supportsPromptCache: false,
461
+ inputPrice: 0.8,
462
+ outputPrice: 2.4,
463
+ description: "Claude Instant"
464
+ },
465
+ "deepseek.r1-v1:0": {
466
+ maxTokens: 32768,
467
+ contextWindow: 128e3,
468
+ supportsImages: false,
469
+ supportsPromptCache: false,
470
+ inputPrice: 1.35,
471
+ outputPrice: 5.4
472
+ },
473
+ "meta.llama3-3-70b-instruct-v1:0": {
474
+ maxTokens: 8192,
475
+ contextWindow: 128e3,
476
+ supportsImages: false,
477
+ supportsComputerUse: false,
478
+ supportsPromptCache: false,
479
+ inputPrice: 0.72,
480
+ outputPrice: 0.72,
481
+ description: "Llama 3.3 Instruct (70B)"
482
+ },
483
+ "meta.llama3-2-90b-instruct-v1:0": {
484
+ maxTokens: 8192,
485
+ contextWindow: 128e3,
486
+ supportsImages: true,
487
+ supportsComputerUse: false,
488
+ supportsPromptCache: false,
489
+ inputPrice: 0.72,
490
+ outputPrice: 0.72,
491
+ description: "Llama 3.2 Instruct (90B)"
492
+ },
493
+ "meta.llama3-2-11b-instruct-v1:0": {
494
+ maxTokens: 8192,
495
+ contextWindow: 128e3,
496
+ supportsImages: true,
497
+ supportsComputerUse: false,
498
+ supportsPromptCache: false,
499
+ inputPrice: 0.16,
500
+ outputPrice: 0.16,
501
+ description: "Llama 3.2 Instruct (11B)"
502
+ },
503
+ "meta.llama3-2-3b-instruct-v1:0": {
504
+ maxTokens: 8192,
505
+ contextWindow: 128e3,
506
+ supportsImages: false,
507
+ supportsComputerUse: false,
508
+ supportsPromptCache: false,
509
+ inputPrice: 0.15,
510
+ outputPrice: 0.15,
511
+ description: "Llama 3.2 Instruct (3B)"
512
+ },
513
+ "meta.llama3-2-1b-instruct-v1:0": {
514
+ maxTokens: 8192,
515
+ contextWindow: 128e3,
516
+ supportsImages: false,
517
+ supportsComputerUse: false,
518
+ supportsPromptCache: false,
519
+ inputPrice: 0.1,
520
+ outputPrice: 0.1,
521
+ description: "Llama 3.2 Instruct (1B)"
522
+ },
523
+ "meta.llama3-1-405b-instruct-v1:0": {
524
+ maxTokens: 8192,
525
+ contextWindow: 128e3,
526
+ supportsImages: false,
527
+ supportsComputerUse: false,
528
+ supportsPromptCache: false,
529
+ inputPrice: 2.4,
530
+ outputPrice: 2.4,
531
+ description: "Llama 3.1 Instruct (405B)"
532
+ },
533
+ "meta.llama3-1-70b-instruct-v1:0": {
534
+ maxTokens: 8192,
535
+ contextWindow: 128e3,
536
+ supportsImages: false,
537
+ supportsComputerUse: false,
538
+ supportsPromptCache: false,
539
+ inputPrice: 0.72,
540
+ outputPrice: 0.72,
541
+ description: "Llama 3.1 Instruct (70B)"
542
+ },
543
+ "meta.llama3-1-70b-instruct-latency-optimized-v1:0": {
544
+ maxTokens: 8192,
545
+ contextWindow: 128e3,
546
+ supportsImages: false,
547
+ supportsComputerUse: false,
548
+ supportsPromptCache: false,
549
+ inputPrice: 0.9,
550
+ outputPrice: 0.9,
551
+ description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
552
+ },
553
+ "meta.llama3-1-8b-instruct-v1:0": {
554
+ maxTokens: 8192,
555
+ contextWindow: 8e3,
556
+ supportsImages: false,
557
+ supportsComputerUse: false,
558
+ supportsPromptCache: false,
559
+ inputPrice: 0.22,
560
+ outputPrice: 0.22,
561
+ description: "Llama 3.1 Instruct (8B)"
562
+ },
563
+ "meta.llama3-70b-instruct-v1:0": {
564
+ maxTokens: 2048,
565
+ contextWindow: 8e3,
566
+ supportsImages: false,
567
+ supportsComputerUse: false,
568
+ supportsPromptCache: false,
569
+ inputPrice: 2.65,
570
+ outputPrice: 3.5
571
+ },
572
+ "meta.llama3-8b-instruct-v1:0": {
573
+ maxTokens: 2048,
574
+ contextWindow: 4e3,
575
+ supportsImages: false,
576
+ supportsComputerUse: false,
577
+ supportsPromptCache: false,
578
+ inputPrice: 0.3,
579
+ outputPrice: 0.6
580
+ },
581
+ "amazon.titan-text-lite-v1:0": {
582
+ maxTokens: 4096,
583
+ contextWindow: 8e3,
584
+ supportsImages: false,
585
+ supportsComputerUse: false,
586
+ supportsPromptCache: false,
587
+ inputPrice: 0.15,
588
+ outputPrice: 0.2,
589
+ description: "Amazon Titan Text Lite"
590
+ },
591
+ "amazon.titan-text-express-v1:0": {
592
+ maxTokens: 4096,
593
+ contextWindow: 8e3,
594
+ supportsImages: false,
595
+ supportsComputerUse: false,
596
+ supportsPromptCache: false,
597
+ inputPrice: 0.2,
598
+ outputPrice: 0.6,
599
+ description: "Amazon Titan Text Express"
600
+ },
601
+ "amazon.titan-text-embeddings-v1:0": {
602
+ maxTokens: 8192,
603
+ contextWindow: 8e3,
604
+ supportsImages: false,
605
+ supportsComputerUse: false,
606
+ supportsPromptCache: false,
607
+ inputPrice: 0.1,
608
+ description: "Amazon Titan Text Embeddings"
609
+ },
610
+ "amazon.titan-text-embeddings-v2:0": {
611
+ maxTokens: 8192,
612
+ contextWindow: 8e3,
613
+ supportsImages: false,
614
+ supportsComputerUse: false,
615
+ supportsPromptCache: false,
616
+ inputPrice: 0.02,
617
+ description: "Amazon Titan Text Embeddings V2"
618
+ }
619
+ };
620
+ var BEDROCK_DEFAULT_TEMPERATURE = 0.3;
621
+ var BEDROCK_MAX_TOKENS = 4096;
622
+ var BEDROCK_REGION_INFO = {
623
+ /*
624
+ * This JSON generated by AWS's AI assistant - Amazon Q on March 29, 2025
625
+ *
626
+ * - Africa (Cape Town) region does not appear to support Amazon Bedrock at this time.
627
+ * - Some Asia Pacific regions, such as Asia Pacific (Hong Kong) and Asia Pacific (Jakarta), are not listed among the supported regions for Bedrock services.
628
+ * - Middle East regions, including Middle East (Bahrain) and Middle East (UAE), are not mentioned in the list of supported regions for Bedrock. [3]
629
+ * - China regions (Beijing and Ningxia) are not listed as supported for Amazon Bedrock.
630
+ * - Some newer or specialized AWS regions may not have Bedrock support yet.
631
+ */
632
+ "us.": { regionId: "us-east-1", description: "US East (N. Virginia)", pattern: "us-", multiRegion: true },
633
+ "use.": { regionId: "us-east-1", description: "US East (N. Virginia)" },
634
+ "use1.": { regionId: "us-east-1", description: "US East (N. Virginia)" },
635
+ "use2.": { regionId: "us-east-2", description: "US East (Ohio)" },
636
+ "usw.": { regionId: "us-west-2", description: "US West (Oregon)" },
637
+ "usw2.": { regionId: "us-west-2", description: "US West (Oregon)" },
638
+ "ug.": {
639
+ regionId: "us-gov-west-1",
640
+ description: "AWS GovCloud (US-West)",
641
+ pattern: "us-gov-",
642
+ multiRegion: true
643
+ },
644
+ "uge1.": { regionId: "us-gov-east-1", description: "AWS GovCloud (US-East)" },
645
+ "ugw1.": { regionId: "us-gov-west-1", description: "AWS GovCloud (US-West)" },
646
+ "eu.": { regionId: "eu-west-1", description: "Europe (Ireland)", pattern: "eu-", multiRegion: true },
647
+ "euw1.": { regionId: "eu-west-1", description: "Europe (Ireland)" },
648
+ "euw2.": { regionId: "eu-west-2", description: "Europe (London)" },
649
+ "euw3.": { regionId: "eu-west-3", description: "Europe (Paris)" },
650
+ "euc1.": { regionId: "eu-central-1", description: "Europe (Frankfurt)" },
651
+ "euc2.": { regionId: "eu-central-2", description: "Europe (Zurich)" },
652
+ "eun1.": { regionId: "eu-north-1", description: "Europe (Stockholm)" },
653
+ "eus1.": { regionId: "eu-south-1", description: "Europe (Milan)" },
654
+ "eus2.": { regionId: "eu-south-2", description: "Europe (Spain)" },
655
+ "ap.": {
656
+ regionId: "ap-southeast-1",
657
+ description: "Asia Pacific (Singapore)",
658
+ pattern: "ap-",
659
+ multiRegion: true
660
+ },
661
+ "ape1.": { regionId: "ap-east-1", description: "Asia Pacific (Hong Kong)" },
662
+ "apne1.": { regionId: "ap-northeast-1", description: "Asia Pacific (Tokyo)" },
663
+ "apne2.": { regionId: "ap-northeast-2", description: "Asia Pacific (Seoul)" },
664
+ "apne3.": { regionId: "ap-northeast-3", description: "Asia Pacific (Osaka)" },
665
+ "aps1.": { regionId: "ap-south-1", description: "Asia Pacific (Mumbai)" },
666
+ "aps2.": { regionId: "ap-south-2", description: "Asia Pacific (Hyderabad)" },
667
+ "apse1.": { regionId: "ap-southeast-1", description: "Asia Pacific (Singapore)" },
668
+ "apse2.": { regionId: "ap-southeast-2", description: "Asia Pacific (Sydney)" },
669
+ "ca.": { regionId: "ca-central-1", description: "Canada (Central)", pattern: "ca-", multiRegion: true },
670
+ "cac1.": { regionId: "ca-central-1", description: "Canada (Central)" },
671
+ "sa.": { regionId: "sa-east-1", description: "South America (S\xE3o Paulo)", pattern: "sa-", multiRegion: true },
672
+ "sae1.": { regionId: "sa-east-1", description: "South America (S\xE3o Paulo)" },
673
+ // These are not official - they weren't generated by Amazon Q nor were
674
+ // found in the AWS documentation but another Roo contributor found apac.
675
+ // Was needed so I've added the pattern of the other geo zones.
676
+ "apac.": { regionId: "ap-southeast-1", description: "Default APAC region", pattern: "ap-", multiRegion: true },
677
+ "emea.": { regionId: "eu-west-1", description: "Default EMEA region", pattern: "eu-", multiRegion: true },
678
+ "amer.": { regionId: "us-east-1", description: "Default Americas region", pattern: "us-", multiRegion: true }
679
+ };
680
+ var BEDROCK_REGIONS = Object.values(BEDROCK_REGION_INFO).map((info) => ({ value: info.regionId, label: info.regionId })).filter((region, index, self) => index === self.findIndex((r) => r.value === region.value)).sort((a, b) => a.value.localeCompare(b.value));
681
+
682
+ // src/providers/chutes.ts
683
+ var chutesDefaultModelId = "deepseek-ai/DeepSeek-R1-0528";
684
+ var chutesModels = {
685
+ "deepseek-ai/DeepSeek-R1-0528": {
686
+ maxTokens: 32768,
687
+ contextWindow: 163840,
688
+ supportsImages: false,
689
+ supportsPromptCache: false,
690
+ inputPrice: 0,
691
+ outputPrice: 0,
692
+ description: "DeepSeek R1 0528 model."
693
+ },
694
+ "deepseek-ai/DeepSeek-R1": {
695
+ maxTokens: 32768,
696
+ contextWindow: 163840,
697
+ supportsImages: false,
698
+ supportsPromptCache: false,
699
+ inputPrice: 0,
700
+ outputPrice: 0,
701
+ description: "DeepSeek R1 model."
702
+ },
703
+ "deepseek-ai/DeepSeek-V3": {
704
+ maxTokens: 32768,
705
+ contextWindow: 163840,
706
+ supportsImages: false,
707
+ supportsPromptCache: false,
708
+ inputPrice: 0,
709
+ outputPrice: 0,
710
+ description: "DeepSeek V3 model."
711
+ },
712
+ "unsloth/Llama-3.3-70B-Instruct": {
713
+ maxTokens: 32768,
714
+ // From Groq
715
+ contextWindow: 131072,
716
+ // From Groq
717
+ supportsImages: false,
718
+ supportsPromptCache: false,
719
+ inputPrice: 0,
720
+ outputPrice: 0,
721
+ description: "Unsloth Llama 3.3 70B Instruct model."
722
+ },
723
+ "chutesai/Llama-4-Scout-17B-16E-Instruct": {
724
+ maxTokens: 32768,
725
+ contextWindow: 512e3,
726
+ supportsImages: false,
727
+ supportsPromptCache: false,
728
+ inputPrice: 0,
729
+ outputPrice: 0,
730
+ description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
731
+ },
732
+ "unsloth/Mistral-Nemo-Instruct-2407": {
733
+ maxTokens: 32768,
734
+ contextWindow: 128e3,
735
+ supportsImages: false,
736
+ supportsPromptCache: false,
737
+ inputPrice: 0,
738
+ outputPrice: 0,
739
+ description: "Unsloth Mistral Nemo Instruct model."
740
+ },
741
+ "unsloth/gemma-3-12b-it": {
742
+ maxTokens: 32768,
743
+ contextWindow: 131072,
744
+ supportsImages: false,
745
+ supportsPromptCache: false,
746
+ inputPrice: 0,
747
+ outputPrice: 0,
748
+ description: "Unsloth Gemma 3 12B IT model."
749
+ },
750
+ "NousResearch/DeepHermes-3-Llama-3-8B-Preview": {
751
+ maxTokens: 32768,
752
+ contextWindow: 131072,
753
+ supportsImages: false,
754
+ supportsPromptCache: false,
755
+ inputPrice: 0,
756
+ outputPrice: 0,
757
+ description: "Nous DeepHermes 3 Llama 3 8B Preview model."
758
+ },
759
+ "unsloth/gemma-3-4b-it": {
760
+ maxTokens: 32768,
761
+ contextWindow: 131072,
762
+ supportsImages: false,
763
+ supportsPromptCache: false,
764
+ inputPrice: 0,
765
+ outputPrice: 0,
766
+ description: "Unsloth Gemma 3 4B IT model."
767
+ },
768
+ "nvidia/Llama-3_3-Nemotron-Super-49B-v1": {
769
+ maxTokens: 32768,
770
+ contextWindow: 131072,
771
+ supportsImages: false,
772
+ supportsPromptCache: false,
773
+ inputPrice: 0,
774
+ outputPrice: 0,
775
+ description: "Nvidia Llama 3.3 Nemotron Super 49B model."
776
+ },
777
+ "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1": {
778
+ maxTokens: 32768,
779
+ contextWindow: 131072,
780
+ supportsImages: false,
781
+ supportsPromptCache: false,
782
+ inputPrice: 0,
783
+ outputPrice: 0,
784
+ description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
785
+ },
786
+ "chutesai/Llama-4-Maverick-17B-128E-Instruct-FP8": {
787
+ maxTokens: 32768,
788
+ contextWindow: 256e3,
789
+ supportsImages: false,
790
+ supportsPromptCache: false,
791
+ inputPrice: 0,
792
+ outputPrice: 0,
793
+ description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
794
+ },
795
+ "deepseek-ai/DeepSeek-V3-Base": {
796
+ maxTokens: 32768,
797
+ contextWindow: 163840,
798
+ supportsImages: false,
799
+ supportsPromptCache: false,
800
+ inputPrice: 0,
801
+ outputPrice: 0,
802
+ description: "DeepSeek V3 Base model."
803
+ },
804
+ "deepseek-ai/DeepSeek-R1-Zero": {
805
+ maxTokens: 32768,
806
+ contextWindow: 163840,
807
+ supportsImages: false,
808
+ supportsPromptCache: false,
809
+ inputPrice: 0,
810
+ outputPrice: 0,
811
+ description: "DeepSeek R1 Zero model."
812
+ },
813
+ "deepseek-ai/DeepSeek-V3-0324": {
814
+ maxTokens: 32768,
815
+ contextWindow: 163840,
816
+ supportsImages: false,
817
+ supportsPromptCache: false,
818
+ inputPrice: 0,
819
+ outputPrice: 0,
820
+ description: "DeepSeek V3 (0324) model."
821
+ },
822
+ "Qwen/Qwen3-235B-A22B": {
823
+ maxTokens: 32768,
824
+ contextWindow: 40960,
825
+ supportsImages: false,
826
+ supportsPromptCache: false,
827
+ inputPrice: 0,
828
+ outputPrice: 0,
829
+ description: "Qwen3 235B A22B model."
830
+ },
831
+ "Qwen/Qwen3-32B": {
832
+ maxTokens: 32768,
833
+ contextWindow: 40960,
834
+ supportsImages: false,
835
+ supportsPromptCache: false,
836
+ inputPrice: 0,
837
+ outputPrice: 0,
838
+ description: "Qwen3 32B model."
839
+ },
840
+ "Qwen/Qwen3-30B-A3B": {
841
+ maxTokens: 32768,
842
+ contextWindow: 40960,
843
+ supportsImages: false,
844
+ supportsPromptCache: false,
845
+ inputPrice: 0,
846
+ outputPrice: 0,
847
+ description: "Qwen3 30B A3B model."
848
+ },
849
+ "Qwen/Qwen3-14B": {
850
+ maxTokens: 32768,
851
+ contextWindow: 40960,
852
+ supportsImages: false,
853
+ supportsPromptCache: false,
854
+ inputPrice: 0,
855
+ outputPrice: 0,
856
+ description: "Qwen3 14B model."
857
+ },
858
+ "Qwen/Qwen3-8B": {
859
+ maxTokens: 32768,
860
+ contextWindow: 40960,
861
+ supportsImages: false,
862
+ supportsPromptCache: false,
863
+ inputPrice: 0,
864
+ outputPrice: 0,
865
+ description: "Qwen3 8B model."
866
+ },
867
+ "microsoft/MAI-DS-R1-FP8": {
868
+ maxTokens: 32768,
869
+ contextWindow: 163840,
870
+ supportsImages: false,
871
+ supportsPromptCache: false,
872
+ inputPrice: 0,
873
+ outputPrice: 0,
874
+ description: "Microsoft MAI-DS-R1 FP8 model."
875
+ },
876
+ "tngtech/DeepSeek-R1T-Chimera": {
877
+ maxTokens: 32768,
878
+ contextWindow: 163840,
879
+ supportsImages: false,
880
+ supportsPromptCache: false,
881
+ inputPrice: 0,
882
+ outputPrice: 0,
883
+ description: "TNGTech DeepSeek R1T Chimera model."
884
+ }
885
+ };
886
+
887
+ // src/providers/deepseek.ts
888
+ var deepSeekDefaultModelId = "deepseek-chat";
889
+ var deepSeekModels = {
890
+ "deepseek-chat": {
891
+ maxTokens: 8192,
892
+ contextWindow: 64e3,
893
+ supportsImages: false,
894
+ supportsPromptCache: true,
895
+ inputPrice: 0.27,
896
+ // $0.27 per million tokens (cache miss)
897
+ outputPrice: 1.1,
898
+ // $1.10 per million tokens
899
+ cacheWritesPrice: 0.27,
900
+ // $0.27 per million tokens (cache miss)
901
+ cacheReadsPrice: 0.07,
902
+ // $0.07 per million tokens (cache hit).
903
+ description: `DeepSeek-V3 achieves a significant breakthrough in inference speed over previous models. It tops the leaderboard among open-source models and rivals the most advanced closed-source models globally.`
904
+ },
905
+ "deepseek-reasoner": {
906
+ maxTokens: 8192,
907
+ contextWindow: 64e3,
908
+ supportsImages: false,
909
+ supportsPromptCache: true,
910
+ inputPrice: 0.55,
911
+ // $0.55 per million tokens (cache miss)
912
+ outputPrice: 2.19,
913
+ // $2.19 per million tokens
914
+ cacheWritesPrice: 0.55,
915
+ // $0.55 per million tokens (cache miss)
916
+ cacheReadsPrice: 0.14,
917
+ // $0.14 per million tokens (cache hit)
918
+ description: `DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 32K tokens.`
919
+ }
920
+ };
921
+ var DEEP_SEEK_DEFAULT_TEMPERATURE = 0.6;
922
+
923
+ // src/providers/gemini.ts
924
+ var geminiDefaultModelId = "gemini-2.0-flash-001";
925
+ var geminiModels = {
926
+ "gemini-2.5-flash-preview-04-17:thinking": {
927
+ maxTokens: 65535,
928
+ contextWindow: 1048576,
929
+ supportsImages: true,
930
+ supportsPromptCache: false,
931
+ inputPrice: 0.15,
932
+ outputPrice: 3.5,
933
+ maxThinkingTokens: 24576,
934
+ supportsReasoningBudget: true,
935
+ requiredReasoningBudget: true
936
+ },
937
+ "gemini-2.5-flash-preview-04-17": {
938
+ maxTokens: 65535,
939
+ contextWindow: 1048576,
940
+ supportsImages: true,
941
+ supportsPromptCache: false,
942
+ inputPrice: 0.15,
943
+ outputPrice: 0.6
944
+ },
945
+ "gemini-2.5-flash-preview-05-20:thinking": {
946
+ maxTokens: 65535,
947
+ contextWindow: 1048576,
948
+ supportsImages: true,
949
+ supportsPromptCache: true,
950
+ inputPrice: 0.15,
951
+ outputPrice: 3.5,
952
+ cacheReadsPrice: 0.0375,
953
+ cacheWritesPrice: 1,
954
+ maxThinkingTokens: 24576,
955
+ supportsReasoningBudget: true,
956
+ requiredReasoningBudget: true
957
+ },
958
+ "gemini-2.5-flash-preview-05-20": {
959
+ maxTokens: 65535,
960
+ contextWindow: 1048576,
961
+ supportsImages: true,
962
+ supportsPromptCache: true,
963
+ inputPrice: 0.15,
964
+ outputPrice: 0.6,
965
+ cacheReadsPrice: 0.0375,
966
+ cacheWritesPrice: 1
967
+ },
968
+ "gemini-2.5-pro-exp-03-25": {
969
+ maxTokens: 65535,
970
+ contextWindow: 1048576,
971
+ supportsImages: true,
972
+ supportsPromptCache: false,
973
+ inputPrice: 0,
974
+ outputPrice: 0
975
+ },
976
+ "gemini-2.5-pro-preview-03-25": {
977
+ maxTokens: 65535,
978
+ contextWindow: 1048576,
979
+ supportsImages: true,
980
+ supportsPromptCache: true,
981
+ inputPrice: 2.5,
982
+ // This is the pricing for prompts above 200k tokens.
983
+ outputPrice: 15,
984
+ cacheReadsPrice: 0.625,
985
+ cacheWritesPrice: 4.5,
986
+ tiers: [
987
+ {
988
+ contextWindow: 2e5,
989
+ inputPrice: 1.25,
990
+ outputPrice: 10,
991
+ cacheReadsPrice: 0.31
992
+ },
993
+ {
994
+ contextWindow: Infinity,
995
+ inputPrice: 2.5,
996
+ outputPrice: 15,
997
+ cacheReadsPrice: 0.625
998
+ }
999
+ ]
1000
+ },
1001
+ "gemini-2.5-pro-preview-05-06": {
1002
+ maxTokens: 65535,
1003
+ contextWindow: 1048576,
1004
+ supportsImages: true,
1005
+ supportsPromptCache: true,
1006
+ inputPrice: 2.5,
1007
+ // This is the pricing for prompts above 200k tokens.
1008
+ outputPrice: 15,
1009
+ cacheReadsPrice: 0.625,
1010
+ cacheWritesPrice: 4.5,
1011
+ tiers: [
1012
+ {
1013
+ contextWindow: 2e5,
1014
+ inputPrice: 1.25,
1015
+ outputPrice: 10,
1016
+ cacheReadsPrice: 0.31
1017
+ },
1018
+ {
1019
+ contextWindow: Infinity,
1020
+ inputPrice: 2.5,
1021
+ outputPrice: 15,
1022
+ cacheReadsPrice: 0.625
1023
+ }
1024
+ ]
1025
+ },
1026
+ "gemini-2.0-flash-001": {
1027
+ maxTokens: 8192,
1028
+ contextWindow: 1048576,
1029
+ supportsImages: true,
1030
+ supportsPromptCache: true,
1031
+ inputPrice: 0.1,
1032
+ outputPrice: 0.4,
1033
+ cacheReadsPrice: 0.025,
1034
+ cacheWritesPrice: 1
1035
+ },
1036
+ "gemini-2.0-flash-lite-preview-02-05": {
1037
+ maxTokens: 8192,
1038
+ contextWindow: 1048576,
1039
+ supportsImages: true,
1040
+ supportsPromptCache: false,
1041
+ inputPrice: 0,
1042
+ outputPrice: 0
1043
+ },
1044
+ "gemini-2.0-pro-exp-02-05": {
1045
+ maxTokens: 8192,
1046
+ contextWindow: 2097152,
1047
+ supportsImages: true,
1048
+ supportsPromptCache: false,
1049
+ inputPrice: 0,
1050
+ outputPrice: 0
1051
+ },
1052
+ "gemini-2.0-flash-thinking-exp-01-21": {
1053
+ maxTokens: 65536,
1054
+ contextWindow: 1048576,
1055
+ supportsImages: true,
1056
+ supportsPromptCache: false,
1057
+ inputPrice: 0,
1058
+ outputPrice: 0
1059
+ },
1060
+ "gemini-2.0-flash-thinking-exp-1219": {
1061
+ maxTokens: 8192,
1062
+ contextWindow: 32767,
1063
+ supportsImages: true,
1064
+ supportsPromptCache: false,
1065
+ inputPrice: 0,
1066
+ outputPrice: 0
1067
+ },
1068
+ "gemini-2.0-flash-exp": {
1069
+ maxTokens: 8192,
1070
+ contextWindow: 1048576,
1071
+ supportsImages: true,
1072
+ supportsPromptCache: false,
1073
+ inputPrice: 0,
1074
+ outputPrice: 0
1075
+ },
1076
+ "gemini-1.5-flash-002": {
1077
+ maxTokens: 8192,
1078
+ contextWindow: 1048576,
1079
+ supportsImages: true,
1080
+ supportsPromptCache: true,
1081
+ inputPrice: 0.15,
1082
+ // This is the pricing for prompts above 128k tokens.
1083
+ outputPrice: 0.6,
1084
+ cacheReadsPrice: 0.0375,
1085
+ cacheWritesPrice: 1,
1086
+ tiers: [
1087
+ {
1088
+ contextWindow: 128e3,
1089
+ inputPrice: 0.075,
1090
+ outputPrice: 0.3,
1091
+ cacheReadsPrice: 0.01875
1092
+ },
1093
+ {
1094
+ contextWindow: Infinity,
1095
+ inputPrice: 0.15,
1096
+ outputPrice: 0.6,
1097
+ cacheReadsPrice: 0.0375
1098
+ }
1099
+ ]
1100
+ },
1101
+ "gemini-1.5-flash-exp-0827": {
1102
+ maxTokens: 8192,
1103
+ contextWindow: 1048576,
1104
+ supportsImages: true,
1105
+ supportsPromptCache: false,
1106
+ inputPrice: 0,
1107
+ outputPrice: 0
1108
+ },
1109
+ "gemini-1.5-flash-8b-exp-0827": {
1110
+ maxTokens: 8192,
1111
+ contextWindow: 1048576,
1112
+ supportsImages: true,
1113
+ supportsPromptCache: false,
1114
+ inputPrice: 0,
1115
+ outputPrice: 0
1116
+ },
1117
+ "gemini-1.5-pro-002": {
1118
+ maxTokens: 8192,
1119
+ contextWindow: 2097152,
1120
+ supportsImages: true,
1121
+ supportsPromptCache: false,
1122
+ inputPrice: 0,
1123
+ outputPrice: 0
1124
+ },
1125
+ "gemini-1.5-pro-exp-0827": {
1126
+ maxTokens: 8192,
1127
+ contextWindow: 2097152,
1128
+ supportsImages: true,
1129
+ supportsPromptCache: false,
1130
+ inputPrice: 0,
1131
+ outputPrice: 0
1132
+ },
1133
+ "gemini-exp-1206": {
1134
+ maxTokens: 8192,
1135
+ contextWindow: 2097152,
1136
+ supportsImages: true,
1137
+ supportsPromptCache: false,
1138
+ inputPrice: 0,
1139
+ outputPrice: 0
1140
+ }
1141
+ };
1142
+
1143
+ // src/providers/glama.ts
1144
+ var glamaDefaultModelId = "anthropic/claude-3-7-sonnet";
1145
+ var glamaDefaultModelInfo = {
1146
+ maxTokens: 8192,
1147
+ contextWindow: 2e5,
1148
+ supportsImages: true,
1149
+ supportsComputerUse: true,
1150
+ supportsPromptCache: true,
1151
+ inputPrice: 3,
1152
+ outputPrice: 15,
1153
+ cacheWritesPrice: 3.75,
1154
+ cacheReadsPrice: 0.3,
1155
+ description: "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. Claude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks. Read more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)"
1156
+ };
1157
+ var GLAMA_DEFAULT_TEMPERATURE = 0;
1158
+
1159
+ // src/providers/groq.ts
1160
+ var groqDefaultModelId = "llama-3.3-70b-versatile";
1161
+ var groqModels = {
1162
+ // Models based on API response: https://api.groq.com/openai/v1/models
1163
+ "llama-3.1-8b-instant": {
1164
+ maxTokens: 131072,
1165
+ contextWindow: 131072,
1166
+ supportsImages: false,
1167
+ supportsPromptCache: false,
1168
+ inputPrice: 0,
1169
+ outputPrice: 0,
1170
+ description: "Meta Llama 3.1 8B Instant model, 128K context."
1171
+ },
1172
+ "llama-3.3-70b-versatile": {
1173
+ maxTokens: 32768,
1174
+ contextWindow: 131072,
1175
+ supportsImages: false,
1176
+ supportsPromptCache: false,
1177
+ inputPrice: 0,
1178
+ outputPrice: 0,
1179
+ description: "Meta Llama 3.3 70B Versatile model, 128K context."
1180
+ },
1181
+ "meta-llama/llama-4-scout-17b-16e-instruct": {
1182
+ maxTokens: 8192,
1183
+ contextWindow: 131072,
1184
+ supportsImages: false,
1185
+ supportsPromptCache: false,
1186
+ inputPrice: 0,
1187
+ outputPrice: 0,
1188
+ description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
1189
+ },
1190
+ "meta-llama/llama-4-maverick-17b-128e-instruct": {
1191
+ maxTokens: 8192,
1192
+ contextWindow: 131072,
1193
+ supportsImages: false,
1194
+ supportsPromptCache: false,
1195
+ inputPrice: 0,
1196
+ outputPrice: 0,
1197
+ description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
1198
+ },
1199
+ "mistral-saba-24b": {
1200
+ maxTokens: 32768,
1201
+ contextWindow: 32768,
1202
+ supportsImages: false,
1203
+ supportsPromptCache: false,
1204
+ inputPrice: 0,
1205
+ outputPrice: 0,
1206
+ description: "Mistral Saba 24B model, 32K context."
1207
+ },
1208
+ "qwen-qwq-32b": {
1209
+ maxTokens: 131072,
1210
+ contextWindow: 131072,
1211
+ supportsImages: false,
1212
+ supportsPromptCache: false,
1213
+ inputPrice: 0,
1214
+ outputPrice: 0,
1215
+ description: "Alibaba Qwen QwQ 32B model, 128K context."
1216
+ },
1217
+ "deepseek-r1-distill-llama-70b": {
1218
+ maxTokens: 131072,
1219
+ contextWindow: 131072,
1220
+ supportsImages: false,
1221
+ supportsPromptCache: false,
1222
+ inputPrice: 0,
1223
+ outputPrice: 0,
1224
+ description: "DeepSeek R1 Distill Llama 70B model, 128K context."
1225
+ }
1226
+ };
1227
+
1228
+ // src/providers/lite-llm.ts
1229
+ var litellmDefaultModelId = "claude-3-7-sonnet-20250219";
1230
+ var litellmDefaultModelInfo = {
1231
+ maxTokens: 8192,
1232
+ contextWindow: 2e5,
1233
+ supportsImages: true,
1234
+ supportsComputerUse: true,
1235
+ supportsPromptCache: true,
1236
+ inputPrice: 3,
1237
+ outputPrice: 15,
1238
+ cacheWritesPrice: 3.75,
1239
+ cacheReadsPrice: 0.3
1240
+ };
1241
+ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1242
+ "claude-3-5-sonnet-latest",
1243
+ "claude-opus-4-20250514",
1244
+ "claude-sonnet-4-20250514",
1245
+ "claude-3-7-sonnet-latest",
1246
+ "claude-3-7-sonnet-20250219",
1247
+ "claude-3-5-sonnet-20241022",
1248
+ "vertex_ai/claude-3-5-sonnet",
1249
+ "vertex_ai/claude-3-5-sonnet-v2",
1250
+ "vertex_ai/claude-3-5-sonnet-v2@20241022",
1251
+ "vertex_ai/claude-3-7-sonnet@20250219",
1252
+ "vertex_ai/claude-opus-4@20250514",
1253
+ "vertex_ai/claude-sonnet-4@20250514",
1254
+ "openrouter/anthropic/claude-3.5-sonnet",
1255
+ "openrouter/anthropic/claude-3.5-sonnet:beta",
1256
+ "openrouter/anthropic/claude-3.7-sonnet",
1257
+ "openrouter/anthropic/claude-3.7-sonnet:beta",
1258
+ "anthropic.claude-opus-4-20250514-v1:0",
1259
+ "anthropic.claude-sonnet-4-20250514-v1:0",
1260
+ "anthropic.claude-3-7-sonnet-20250219-v1:0",
1261
+ "anthropic.claude-3-5-sonnet-20241022-v2:0",
1262
+ "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
1263
+ "us.anthropic.claude-3-7-sonnet-20250219-v1:0",
1264
+ "us.anthropic.claude-opus-4-20250514-v1:0",
1265
+ "us.anthropic.claude-sonnet-4-20250514-v1:0",
1266
+ "eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
1267
+ "eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
1268
+ "eu.anthropic.claude-opus-4-20250514-v1:0",
1269
+ "eu.anthropic.claude-sonnet-4-20250514-v1:0",
1270
+ "snowflake/claude-3-5-sonnet"
1271
+ ]);
1272
+
1273
+ // src/providers/lm-studio.ts
1274
+ var LMSTUDIO_DEFAULT_TEMPERATURE = 0;
1275
+
1276
+ // src/providers/mistral.ts
1277
+ var mistralDefaultModelId = "codestral-latest";
1278
+ var mistralModels = {
1279
+ "codestral-latest": {
1280
+ maxTokens: 256e3,
1281
+ contextWindow: 256e3,
1282
+ supportsImages: false,
1283
+ supportsPromptCache: false,
1284
+ inputPrice: 0.3,
1285
+ outputPrice: 0.9
1286
+ },
1287
+ "mistral-large-latest": {
1288
+ maxTokens: 131e3,
1289
+ contextWindow: 131e3,
1290
+ supportsImages: false,
1291
+ supportsPromptCache: false,
1292
+ inputPrice: 2,
1293
+ outputPrice: 6
1294
+ },
1295
+ "ministral-8b-latest": {
1296
+ maxTokens: 131e3,
1297
+ contextWindow: 131e3,
1298
+ supportsImages: false,
1299
+ supportsPromptCache: false,
1300
+ inputPrice: 0.1,
1301
+ outputPrice: 0.1
1302
+ },
1303
+ "ministral-3b-latest": {
1304
+ maxTokens: 131e3,
1305
+ contextWindow: 131e3,
1306
+ supportsImages: false,
1307
+ supportsPromptCache: false,
1308
+ inputPrice: 0.04,
1309
+ outputPrice: 0.04
1310
+ },
1311
+ "mistral-small-latest": {
1312
+ maxTokens: 32e3,
1313
+ contextWindow: 32e3,
1314
+ supportsImages: false,
1315
+ supportsPromptCache: false,
1316
+ inputPrice: 0.2,
1317
+ outputPrice: 0.6
1318
+ },
1319
+ "pixtral-large-latest": {
1320
+ maxTokens: 131e3,
1321
+ contextWindow: 131e3,
1322
+ supportsImages: true,
1323
+ supportsPromptCache: false,
1324
+ inputPrice: 2,
1325
+ outputPrice: 6
1326
+ }
1327
+ };
1328
+ var MISTRAL_DEFAULT_TEMPERATURE = 0;
1329
+
1330
+ // src/providers/openai.ts
1331
+ var openAiNativeDefaultModelId = "gpt-4.1";
1332
+ var openAiNativeModels = {
1333
+ "gpt-4.1": {
1334
+ maxTokens: 32768,
1335
+ contextWindow: 1047576,
1336
+ supportsImages: true,
1337
+ supportsPromptCache: true,
1338
+ inputPrice: 2,
1339
+ outputPrice: 8,
1340
+ cacheReadsPrice: 0.5
1341
+ },
1342
+ "gpt-4.1-mini": {
1343
+ maxTokens: 32768,
1344
+ contextWindow: 1047576,
1345
+ supportsImages: true,
1346
+ supportsPromptCache: true,
1347
+ inputPrice: 0.4,
1348
+ outputPrice: 1.6,
1349
+ cacheReadsPrice: 0.1
1350
+ },
1351
+ "gpt-4.1-nano": {
1352
+ maxTokens: 32768,
1353
+ contextWindow: 1047576,
1354
+ supportsImages: true,
1355
+ supportsPromptCache: true,
1356
+ inputPrice: 0.1,
1357
+ outputPrice: 0.4,
1358
+ cacheReadsPrice: 0.025
1359
+ },
1360
+ o3: {
1361
+ maxTokens: 1e5,
1362
+ contextWindow: 2e5,
1363
+ supportsImages: true,
1364
+ supportsPromptCache: true,
1365
+ inputPrice: 10,
1366
+ outputPrice: 40,
1367
+ cacheReadsPrice: 2.5,
1368
+ supportsReasoningEffort: true,
1369
+ reasoningEffort: "medium"
1370
+ },
1371
+ "o3-high": {
1372
+ maxTokens: 1e5,
1373
+ contextWindow: 2e5,
1374
+ supportsImages: true,
1375
+ supportsPromptCache: true,
1376
+ inputPrice: 10,
1377
+ outputPrice: 40,
1378
+ cacheReadsPrice: 2.5,
1379
+ reasoningEffort: "high"
1380
+ },
1381
+ "o3-low": {
1382
+ maxTokens: 1e5,
1383
+ contextWindow: 2e5,
1384
+ supportsImages: true,
1385
+ supportsPromptCache: true,
1386
+ inputPrice: 10,
1387
+ outputPrice: 40,
1388
+ cacheReadsPrice: 2.5,
1389
+ reasoningEffort: "low"
1390
+ },
1391
+ "o4-mini": {
1392
+ maxTokens: 1e5,
1393
+ contextWindow: 2e5,
1394
+ supportsImages: true,
1395
+ supportsPromptCache: true,
1396
+ inputPrice: 1.1,
1397
+ outputPrice: 4.4,
1398
+ cacheReadsPrice: 0.275,
1399
+ supportsReasoningEffort: true,
1400
+ reasoningEffort: "medium"
1401
+ },
1402
+ "o4-mini-high": {
1403
+ maxTokens: 1e5,
1404
+ contextWindow: 2e5,
1405
+ supportsImages: true,
1406
+ supportsPromptCache: true,
1407
+ inputPrice: 1.1,
1408
+ outputPrice: 4.4,
1409
+ cacheReadsPrice: 0.275,
1410
+ reasoningEffort: "high"
1411
+ },
1412
+ "o4-mini-low": {
1413
+ maxTokens: 1e5,
1414
+ contextWindow: 2e5,
1415
+ supportsImages: true,
1416
+ supportsPromptCache: true,
1417
+ inputPrice: 1.1,
1418
+ outputPrice: 4.4,
1419
+ cacheReadsPrice: 0.275,
1420
+ reasoningEffort: "low"
1421
+ },
1422
+ "o3-mini": {
1423
+ maxTokens: 1e5,
1424
+ contextWindow: 2e5,
1425
+ supportsImages: false,
1426
+ supportsPromptCache: true,
1427
+ inputPrice: 1.1,
1428
+ outputPrice: 4.4,
1429
+ cacheReadsPrice: 0.55,
1430
+ supportsReasoningEffort: true,
1431
+ reasoningEffort: "medium"
1432
+ },
1433
+ "o3-mini-high": {
1434
+ maxTokens: 1e5,
1435
+ contextWindow: 2e5,
1436
+ supportsImages: false,
1437
+ supportsPromptCache: true,
1438
+ inputPrice: 1.1,
1439
+ outputPrice: 4.4,
1440
+ cacheReadsPrice: 0.55,
1441
+ reasoningEffort: "high"
1442
+ },
1443
+ "o3-mini-low": {
1444
+ maxTokens: 1e5,
1445
+ contextWindow: 2e5,
1446
+ supportsImages: false,
1447
+ supportsPromptCache: true,
1448
+ inputPrice: 1.1,
1449
+ outputPrice: 4.4,
1450
+ cacheReadsPrice: 0.55,
1451
+ reasoningEffort: "low"
1452
+ },
1453
+ o1: {
1454
+ maxTokens: 1e5,
1455
+ contextWindow: 2e5,
1456
+ supportsImages: true,
1457
+ supportsPromptCache: true,
1458
+ inputPrice: 15,
1459
+ outputPrice: 60,
1460
+ cacheReadsPrice: 7.5
1461
+ },
1462
+ "o1-preview": {
1463
+ maxTokens: 32768,
1464
+ contextWindow: 128e3,
1465
+ supportsImages: true,
1466
+ supportsPromptCache: true,
1467
+ inputPrice: 15,
1468
+ outputPrice: 60,
1469
+ cacheReadsPrice: 7.5
1470
+ },
1471
+ "o1-mini": {
1472
+ maxTokens: 65536,
1473
+ contextWindow: 128e3,
1474
+ supportsImages: true,
1475
+ supportsPromptCache: true,
1476
+ inputPrice: 1.1,
1477
+ outputPrice: 4.4,
1478
+ cacheReadsPrice: 0.55
1479
+ },
1480
+ "gpt-4.5-preview": {
1481
+ maxTokens: 16384,
1482
+ contextWindow: 128e3,
1483
+ supportsImages: true,
1484
+ supportsPromptCache: true,
1485
+ inputPrice: 75,
1486
+ outputPrice: 150,
1487
+ cacheReadsPrice: 37.5
1488
+ },
1489
+ "gpt-4o": {
1490
+ maxTokens: 16384,
1491
+ contextWindow: 128e3,
1492
+ supportsImages: true,
1493
+ supportsPromptCache: true,
1494
+ inputPrice: 2.5,
1495
+ outputPrice: 10,
1496
+ cacheReadsPrice: 1.25
1497
+ },
1498
+ "gpt-4o-mini": {
1499
+ maxTokens: 16384,
1500
+ contextWindow: 128e3,
1501
+ supportsImages: true,
1502
+ supportsPromptCache: true,
1503
+ inputPrice: 0.15,
1504
+ outputPrice: 0.6,
1505
+ cacheReadsPrice: 0.075
1506
+ }
1507
+ };
1508
+ var openAiModelInfoSaneDefaults = {
1509
+ maxTokens: -1,
1510
+ contextWindow: 128e3,
1511
+ supportsImages: true,
1512
+ supportsPromptCache: false,
1513
+ inputPrice: 0,
1514
+ outputPrice: 0
1515
+ };
1516
+ var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
1517
+ var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
1518
+ var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
1519
+
1520
+ // src/providers/openrouter.ts
1521
+ var openRouterDefaultModelId = "anthropic/claude-sonnet-4";
1522
+ var openRouterDefaultModelInfo = {
1523
+ maxTokens: 8192,
1524
+ contextWindow: 2e5,
1525
+ supportsImages: true,
1526
+ supportsComputerUse: true,
1527
+ supportsPromptCache: true,
1528
+ inputPrice: 3,
1529
+ outputPrice: 15,
1530
+ cacheWritesPrice: 3.75,
1531
+ cacheReadsPrice: 0.3,
1532
+ description: "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. Claude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks. Read more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)"
1533
+ };
1534
+ var OPENROUTER_DEFAULT_PROVIDER_NAME = "[default]";
1535
+ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
1536
+ "anthropic/claude-3-haiku",
1537
+ "anthropic/claude-3-haiku:beta",
1538
+ "anthropic/claude-3-opus",
1539
+ "anthropic/claude-3-opus:beta",
1540
+ "anthropic/claude-3-sonnet",
1541
+ "anthropic/claude-3-sonnet:beta",
1542
+ "anthropic/claude-3.5-haiku",
1543
+ "anthropic/claude-3.5-haiku-20241022",
1544
+ "anthropic/claude-3.5-haiku-20241022:beta",
1545
+ "anthropic/claude-3.5-haiku:beta",
1546
+ "anthropic/claude-3.5-sonnet",
1547
+ "anthropic/claude-3.5-sonnet-20240620",
1548
+ "anthropic/claude-3.5-sonnet-20240620:beta",
1549
+ "anthropic/claude-3.5-sonnet:beta",
1550
+ "anthropic/claude-3.7-sonnet",
1551
+ "anthropic/claude-3.7-sonnet:beta",
1552
+ "anthropic/claude-3.7-sonnet:thinking",
1553
+ "anthropic/claude-sonnet-4",
1554
+ "anthropic/claude-opus-4",
1555
+ "google/gemini-2.5-pro-preview",
1556
+ "google/gemini-2.5-flash-preview",
1557
+ "google/gemini-2.5-flash-preview:thinking",
1558
+ "google/gemini-2.5-flash-preview-05-20",
1559
+ "google/gemini-2.5-flash-preview-05-20:thinking",
1560
+ "google/gemini-2.0-flash-001",
1561
+ "google/gemini-flash-1.5",
1562
+ "google/gemini-flash-1.5-8b"
1563
+ ]);
1564
+ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
1565
+ "anthropic/claude-3.5-sonnet",
1566
+ "anthropic/claude-3.5-sonnet:beta",
1567
+ "anthropic/claude-3.7-sonnet",
1568
+ "anthropic/claude-3.7-sonnet:beta",
1569
+ "anthropic/claude-3.7-sonnet:thinking",
1570
+ "anthropic/claude-sonnet-4",
1571
+ "anthropic/claude-opus-4"
1572
+ ]);
1573
+ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1574
+ "anthropic/claude-3.7-sonnet:beta",
1575
+ "anthropic/claude-3.7-sonnet:thinking",
1576
+ "anthropic/claude-opus-4",
1577
+ "anthropic/claude-sonnet-4",
1578
+ "google/gemini-2.5-flash-preview-05-20",
1579
+ "google/gemini-2.5-flash-preview-05-20:thinking"
1580
+ ]);
1581
+ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
1582
+ "anthropic/claude-3.7-sonnet:thinking",
1583
+ "google/gemini-2.5-flash-preview-05-20:thinking"
1584
+ ]);
1585
+
1586
+ // src/providers/requesty.ts
1587
+ var requestyDefaultModelId = "coding/claude-4-sonnet";
1588
+ var requestyDefaultModelInfo = {
1589
+ maxTokens: 8192,
1590
+ contextWindow: 2e5,
1591
+ supportsImages: true,
1592
+ supportsComputerUse: true,
1593
+ supportsPromptCache: true,
1594
+ inputPrice: 3,
1595
+ outputPrice: 15,
1596
+ cacheWritesPrice: 3.75,
1597
+ cacheReadsPrice: 0.3,
1598
+ description: "The best coding model, optimized by Requesty, and automatically routed to the fastest provider. Claude 4 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities."
1599
+ };
1600
+
1601
+ // src/providers/unbound.ts
1602
+ var unboundDefaultModelId = "anthropic/claude-3-7-sonnet-20250219";
1603
+ var unboundDefaultModelInfo = {
1604
+ maxTokens: 8192,
1605
+ contextWindow: 2e5,
1606
+ supportsImages: true,
1607
+ supportsPromptCache: true,
1608
+ inputPrice: 3,
1609
+ outputPrice: 15,
1610
+ cacheWritesPrice: 3.75,
1611
+ cacheReadsPrice: 0.3
1612
+ };
1613
+
1614
+ // src/providers/vertex.ts
1615
+ var vertexDefaultModelId = "claude-sonnet-4@20250514";
1616
+ var vertexModels = {
1617
+ "gemini-2.5-flash-preview-05-20:thinking": {
1618
+ maxTokens: 65535,
1619
+ contextWindow: 1048576,
1620
+ supportsImages: true,
1621
+ supportsPromptCache: true,
1622
+ inputPrice: 0.15,
1623
+ outputPrice: 3.5,
1624
+ maxThinkingTokens: 24576,
1625
+ supportsReasoningBudget: true,
1626
+ requiredReasoningBudget: true
1627
+ },
1628
+ "gemini-2.5-flash-preview-05-20": {
1629
+ maxTokens: 65535,
1630
+ contextWindow: 1048576,
1631
+ supportsImages: true,
1632
+ supportsPromptCache: true,
1633
+ inputPrice: 0.15,
1634
+ outputPrice: 0.6
1635
+ },
1636
+ "gemini-2.5-flash-preview-04-17:thinking": {
1637
+ maxTokens: 65535,
1638
+ contextWindow: 1048576,
1639
+ supportsImages: true,
1640
+ supportsPromptCache: false,
1641
+ inputPrice: 0.15,
1642
+ outputPrice: 3.5,
1643
+ maxThinkingTokens: 24576,
1644
+ supportsReasoningBudget: true,
1645
+ requiredReasoningBudget: true
1646
+ },
1647
+ "gemini-2.5-flash-preview-04-17": {
1648
+ maxTokens: 65535,
1649
+ contextWindow: 1048576,
1650
+ supportsImages: true,
1651
+ supportsPromptCache: false,
1652
+ inputPrice: 0.15,
1653
+ outputPrice: 0.6
1654
+ },
1655
+ "gemini-2.5-pro-preview-03-25": {
1656
+ maxTokens: 65535,
1657
+ contextWindow: 1048576,
1658
+ supportsImages: true,
1659
+ supportsPromptCache: true,
1660
+ inputPrice: 2.5,
1661
+ outputPrice: 15
1662
+ },
1663
+ "gemini-2.5-pro-preview-05-06": {
1664
+ maxTokens: 65535,
1665
+ contextWindow: 1048576,
1666
+ supportsImages: true,
1667
+ supportsPromptCache: true,
1668
+ inputPrice: 2.5,
1669
+ outputPrice: 15
1670
+ },
1671
+ "gemini-2.5-pro-exp-03-25": {
1672
+ maxTokens: 65535,
1673
+ contextWindow: 1048576,
1674
+ supportsImages: true,
1675
+ supportsPromptCache: false,
1676
+ inputPrice: 0,
1677
+ outputPrice: 0
1678
+ },
1679
+ "gemini-2.0-pro-exp-02-05": {
1680
+ maxTokens: 8192,
1681
+ contextWindow: 2097152,
1682
+ supportsImages: true,
1683
+ supportsPromptCache: false,
1684
+ inputPrice: 0,
1685
+ outputPrice: 0
1686
+ },
1687
+ "gemini-2.0-flash-001": {
1688
+ maxTokens: 8192,
1689
+ contextWindow: 1048576,
1690
+ supportsImages: true,
1691
+ supportsPromptCache: true,
1692
+ inputPrice: 0.15,
1693
+ outputPrice: 0.6
1694
+ },
1695
+ "gemini-2.0-flash-lite-001": {
1696
+ maxTokens: 8192,
1697
+ contextWindow: 1048576,
1698
+ supportsImages: true,
1699
+ supportsPromptCache: false,
1700
+ inputPrice: 0.075,
1701
+ outputPrice: 0.3
1702
+ },
1703
+ "gemini-2.0-flash-thinking-exp-01-21": {
1704
+ maxTokens: 8192,
1705
+ contextWindow: 32768,
1706
+ supportsImages: true,
1707
+ supportsPromptCache: false,
1708
+ inputPrice: 0,
1709
+ outputPrice: 0
1710
+ },
1711
+ "gemini-1.5-flash-002": {
1712
+ maxTokens: 8192,
1713
+ contextWindow: 1048576,
1714
+ supportsImages: true,
1715
+ supportsPromptCache: true,
1716
+ inputPrice: 0.075,
1717
+ outputPrice: 0.3
1718
+ },
1719
+ "gemini-1.5-pro-002": {
1720
+ maxTokens: 8192,
1721
+ contextWindow: 2097152,
1722
+ supportsImages: true,
1723
+ supportsPromptCache: false,
1724
+ inputPrice: 1.25,
1725
+ outputPrice: 5
1726
+ },
1727
+ "claude-sonnet-4@20250514": {
1728
+ maxTokens: 8192,
1729
+ contextWindow: 2e5,
1730
+ supportsImages: true,
1731
+ supportsComputerUse: true,
1732
+ supportsPromptCache: true,
1733
+ inputPrice: 3,
1734
+ outputPrice: 15,
1735
+ cacheWritesPrice: 3.75,
1736
+ cacheReadsPrice: 0.3,
1737
+ supportsReasoningBudget: true
1738
+ },
1739
+ "claude-opus-4@20250514": {
1740
+ maxTokens: 8192,
1741
+ contextWindow: 2e5,
1742
+ supportsImages: true,
1743
+ supportsComputerUse: true,
1744
+ supportsPromptCache: true,
1745
+ inputPrice: 15,
1746
+ outputPrice: 75,
1747
+ cacheWritesPrice: 18.75,
1748
+ cacheReadsPrice: 1.5
1749
+ },
1750
+ "claude-3-7-sonnet@20250219:thinking": {
1751
+ maxTokens: 64e3,
1752
+ contextWindow: 2e5,
1753
+ supportsImages: true,
1754
+ supportsComputerUse: true,
1755
+ supportsPromptCache: true,
1756
+ inputPrice: 3,
1757
+ outputPrice: 15,
1758
+ cacheWritesPrice: 3.75,
1759
+ cacheReadsPrice: 0.3,
1760
+ supportsReasoningBudget: true,
1761
+ requiredReasoningBudget: true
1762
+ },
1763
+ "claude-3-7-sonnet@20250219": {
1764
+ maxTokens: 8192,
1765
+ contextWindow: 2e5,
1766
+ supportsImages: true,
1767
+ supportsComputerUse: true,
1768
+ supportsPromptCache: true,
1769
+ inputPrice: 3,
1770
+ outputPrice: 15,
1771
+ cacheWritesPrice: 3.75,
1772
+ cacheReadsPrice: 0.3
1773
+ },
1774
+ "claude-3-5-sonnet-v2@20241022": {
1775
+ maxTokens: 8192,
1776
+ contextWindow: 2e5,
1777
+ supportsImages: true,
1778
+ supportsComputerUse: true,
1779
+ supportsPromptCache: true,
1780
+ inputPrice: 3,
1781
+ outputPrice: 15,
1782
+ cacheWritesPrice: 3.75,
1783
+ cacheReadsPrice: 0.3
1784
+ },
1785
+ "claude-3-5-sonnet@20240620": {
1786
+ maxTokens: 8192,
1787
+ contextWindow: 2e5,
1788
+ supportsImages: true,
1789
+ supportsPromptCache: true,
1790
+ inputPrice: 3,
1791
+ outputPrice: 15,
1792
+ cacheWritesPrice: 3.75,
1793
+ cacheReadsPrice: 0.3
1794
+ },
1795
+ "claude-3-5-haiku@20241022": {
1796
+ maxTokens: 8192,
1797
+ contextWindow: 2e5,
1798
+ supportsImages: false,
1799
+ supportsPromptCache: true,
1800
+ inputPrice: 1,
1801
+ outputPrice: 5,
1802
+ cacheWritesPrice: 1.25,
1803
+ cacheReadsPrice: 0.1
1804
+ },
1805
+ "claude-3-opus@20240229": {
1806
+ maxTokens: 4096,
1807
+ contextWindow: 2e5,
1808
+ supportsImages: true,
1809
+ supportsPromptCache: true,
1810
+ inputPrice: 15,
1811
+ outputPrice: 75,
1812
+ cacheWritesPrice: 18.75,
1813
+ cacheReadsPrice: 1.5
1814
+ },
1815
+ "claude-3-haiku@20240307": {
1816
+ maxTokens: 4096,
1817
+ contextWindow: 2e5,
1818
+ supportsImages: true,
1819
+ supportsPromptCache: true,
1820
+ inputPrice: 0.25,
1821
+ outputPrice: 1.25,
1822
+ cacheWritesPrice: 0.3,
1823
+ cacheReadsPrice: 0.03
1824
+ }
1825
+ };
1826
+ var VERTEX_REGIONS = [
1827
+ { value: "us-east5", label: "us-east5" },
1828
+ { value: "us-central1", label: "us-central1" },
1829
+ { value: "europe-west1", label: "europe-west1" },
1830
+ { value: "europe-west4", label: "europe-west4" },
1831
+ { value: "asia-southeast1", label: "asia-southeast1" }
1832
+ ];
1833
+
1834
+ // src/providers/vscode-llm.ts
1835
+ var vscodeLlmDefaultModelId = "claude-3.5-sonnet";
1836
+ var vscodeLlmModels = {
1837
+ "gpt-3.5-turbo": {
1838
+ contextWindow: 12114,
1839
+ supportsImages: false,
1840
+ supportsPromptCache: false,
1841
+ inputPrice: 0,
1842
+ outputPrice: 0,
1843
+ family: "gpt-3.5-turbo",
1844
+ version: "gpt-3.5-turbo-0613",
1845
+ name: "GPT 3.5 Turbo",
1846
+ supportsToolCalling: true,
1847
+ maxInputTokens: 12114
1848
+ },
1849
+ "gpt-4o-mini": {
1850
+ contextWindow: 12115,
1851
+ supportsImages: false,
1852
+ supportsPromptCache: false,
1853
+ inputPrice: 0,
1854
+ outputPrice: 0,
1855
+ family: "gpt-4o-mini",
1856
+ version: "gpt-4o-mini-2024-07-18",
1857
+ name: "GPT-4o mini",
1858
+ supportsToolCalling: true,
1859
+ maxInputTokens: 12115
1860
+ },
1861
+ "gpt-4": {
1862
+ contextWindow: 28501,
1863
+ supportsImages: false,
1864
+ supportsPromptCache: false,
1865
+ inputPrice: 0,
1866
+ outputPrice: 0,
1867
+ family: "gpt-4",
1868
+ version: "gpt-4-0613",
1869
+ name: "GPT 4",
1870
+ supportsToolCalling: true,
1871
+ maxInputTokens: 28501
1872
+ },
1873
+ "gpt-4-0125-preview": {
1874
+ contextWindow: 63826,
1875
+ supportsImages: false,
1876
+ supportsPromptCache: false,
1877
+ inputPrice: 0,
1878
+ outputPrice: 0,
1879
+ family: "gpt-4-turbo",
1880
+ version: "gpt-4-0125-preview",
1881
+ name: "GPT 4 Turbo",
1882
+ supportsToolCalling: true,
1883
+ maxInputTokens: 63826
1884
+ },
1885
+ "gpt-4o": {
1886
+ contextWindow: 63827,
1887
+ supportsImages: true,
1888
+ supportsPromptCache: false,
1889
+ inputPrice: 0,
1890
+ outputPrice: 0,
1891
+ family: "gpt-4o",
1892
+ version: "gpt-4o-2024-11-20",
1893
+ name: "GPT-4o",
1894
+ supportsToolCalling: true,
1895
+ maxInputTokens: 63827
1896
+ },
1897
+ o1: {
1898
+ contextWindow: 19827,
1899
+ supportsImages: false,
1900
+ supportsPromptCache: false,
1901
+ inputPrice: 0,
1902
+ outputPrice: 0,
1903
+ family: "o1-ga",
1904
+ version: "o1-2024-12-17",
1905
+ name: "o1 (Preview)",
1906
+ supportsToolCalling: true,
1907
+ maxInputTokens: 19827
1908
+ },
1909
+ "o3-mini": {
1910
+ contextWindow: 63827,
1911
+ supportsImages: false,
1912
+ supportsPromptCache: false,
1913
+ inputPrice: 0,
1914
+ outputPrice: 0,
1915
+ family: "o3-mini",
1916
+ version: "o3-mini-2025-01-31",
1917
+ name: "o3-mini",
1918
+ supportsToolCalling: true,
1919
+ maxInputTokens: 63827
1920
+ },
1921
+ "claude-3.5-sonnet": {
1922
+ contextWindow: 81638,
1923
+ supportsImages: true,
1924
+ supportsPromptCache: false,
1925
+ inputPrice: 0,
1926
+ outputPrice: 0,
1927
+ family: "claude-3.5-sonnet",
1928
+ version: "claude-3.5-sonnet",
1929
+ name: "Claude 3.5 Sonnet",
1930
+ supportsToolCalling: true,
1931
+ maxInputTokens: 81638
1932
+ },
1933
+ "gemini-2.0-flash-001": {
1934
+ contextWindow: 127827,
1935
+ supportsImages: true,
1936
+ supportsPromptCache: false,
1937
+ inputPrice: 0,
1938
+ outputPrice: 0,
1939
+ family: "gemini-2.0-flash",
1940
+ version: "gemini-2.0-flash-001",
1941
+ name: "Gemini 2.0 Flash",
1942
+ supportsToolCalling: false,
1943
+ maxInputTokens: 127827
1944
+ },
1945
+ "gemini-2.5-pro": {
1946
+ contextWindow: 63830,
1947
+ supportsImages: true,
1948
+ supportsPromptCache: false,
1949
+ inputPrice: 0,
1950
+ outputPrice: 0,
1951
+ family: "gemini-2.5-pro",
1952
+ version: "gemini-2.5-pro-preview-03-25",
1953
+ name: "Gemini 2.5 Pro (Preview)",
1954
+ supportsToolCalling: true,
1955
+ maxInputTokens: 63830
1956
+ },
1957
+ "o4-mini": {
1958
+ contextWindow: 111446,
1959
+ supportsImages: false,
1960
+ supportsPromptCache: false,
1961
+ inputPrice: 0,
1962
+ outputPrice: 0,
1963
+ family: "o4-mini",
1964
+ version: "o4-mini-2025-04-16",
1965
+ name: "o4-mini (Preview)",
1966
+ supportsToolCalling: true,
1967
+ maxInputTokens: 111446
1968
+ },
1969
+ "gpt-4.1": {
1970
+ contextWindow: 111446,
1971
+ supportsImages: true,
1972
+ supportsPromptCache: false,
1973
+ inputPrice: 0,
1974
+ outputPrice: 0,
1975
+ family: "gpt-4.1",
1976
+ version: "gpt-4.1-2025-04-14",
1977
+ name: "GPT-4.1 (Preview)",
1978
+ supportsToolCalling: true,
1979
+ maxInputTokens: 111446
1980
+ }
1981
+ };
1982
+
1983
+ // src/providers/xai.ts
1984
+ var xaiDefaultModelId = "grok-3";
1985
+ var xaiModels = {
1986
+ "grok-3-beta": {
1987
+ maxTokens: 8192,
1988
+ contextWindow: 131072,
1989
+ supportsImages: false,
1990
+ supportsPromptCache: false,
1991
+ inputPrice: 3,
1992
+ outputPrice: 15,
1993
+ description: "xAI's Grok-3 beta model with 131K context window"
1994
+ },
1995
+ "grok-3-fast-beta": {
1996
+ maxTokens: 8192,
1997
+ contextWindow: 131072,
1998
+ supportsImages: false,
1999
+ supportsPromptCache: false,
2000
+ inputPrice: 5,
2001
+ outputPrice: 25,
2002
+ description: "xAI's Grok-3 fast beta model with 131K context window"
2003
+ },
2004
+ "grok-3-mini-beta": {
2005
+ maxTokens: 8192,
2006
+ contextWindow: 131072,
2007
+ supportsImages: false,
2008
+ supportsPromptCache: false,
2009
+ inputPrice: 0.3,
2010
+ outputPrice: 0.5,
2011
+ description: "xAI's Grok-3 mini beta model with 131K context window",
2012
+ supportsReasoningEffort: true
2013
+ },
2014
+ "grok-3-mini-fast-beta": {
2015
+ maxTokens: 8192,
2016
+ contextWindow: 131072,
2017
+ supportsImages: false,
2018
+ supportsPromptCache: false,
2019
+ inputPrice: 0.6,
2020
+ outputPrice: 4,
2021
+ description: "xAI's Grok-3 mini fast beta model with 131K context window",
2022
+ supportsReasoningEffort: true
2023
+ },
2024
+ "grok-3": {
2025
+ maxTokens: 8192,
2026
+ contextWindow: 131072,
2027
+ supportsImages: false,
2028
+ supportsPromptCache: false,
2029
+ inputPrice: 3,
2030
+ outputPrice: 15,
2031
+ description: "xAI's Grok-3 model with 131K context window"
2032
+ },
2033
+ "grok-3-fast": {
2034
+ maxTokens: 8192,
2035
+ contextWindow: 131072,
2036
+ supportsImages: false,
2037
+ supportsPromptCache: false,
2038
+ inputPrice: 5,
2039
+ outputPrice: 25,
2040
+ description: "xAI's Grok-3 fast model with 131K context window"
2041
+ },
2042
+ "grok-3-mini": {
2043
+ maxTokens: 8192,
2044
+ contextWindow: 131072,
2045
+ supportsImages: false,
2046
+ supportsPromptCache: false,
2047
+ inputPrice: 0.3,
2048
+ outputPrice: 0.5,
2049
+ description: "xAI's Grok-3 mini model with 131K context window",
2050
+ supportsReasoningEffort: true
2051
+ },
2052
+ "grok-3-mini-fast": {
2053
+ maxTokens: 8192,
2054
+ contextWindow: 131072,
2055
+ supportsImages: false,
2056
+ supportsPromptCache: false,
2057
+ inputPrice: 0.6,
2058
+ outputPrice: 4,
2059
+ description: "xAI's Grok-3 mini fast model with 131K context window",
2060
+ supportsReasoningEffort: true
2061
+ },
2062
+ "grok-2-latest": {
2063
+ maxTokens: 8192,
2064
+ contextWindow: 131072,
2065
+ supportsImages: false,
2066
+ supportsPromptCache: false,
2067
+ inputPrice: 2,
2068
+ outputPrice: 10,
2069
+ description: "xAI's Grok-2 model - latest version with 131K context window"
2070
+ },
2071
+ "grok-2": {
2072
+ maxTokens: 8192,
2073
+ contextWindow: 131072,
2074
+ supportsImages: false,
2075
+ supportsPromptCache: false,
2076
+ inputPrice: 2,
2077
+ outputPrice: 10,
2078
+ description: "xAI's Grok-2 model with 131K context window"
2079
+ },
2080
+ "grok-2-1212": {
2081
+ maxTokens: 8192,
2082
+ contextWindow: 131072,
2083
+ supportsImages: false,
2084
+ supportsPromptCache: false,
2085
+ inputPrice: 2,
2086
+ outputPrice: 10,
2087
+ description: "xAI's Grok-2 model (version 1212) with 131K context window"
2088
+ },
2089
+ "grok-2-vision-latest": {
2090
+ maxTokens: 8192,
2091
+ contextWindow: 32768,
2092
+ supportsImages: true,
2093
+ supportsPromptCache: false,
2094
+ inputPrice: 2,
2095
+ outputPrice: 10,
2096
+ description: "xAI's Grok-2 Vision model - latest version with image support and 32K context window"
2097
+ },
2098
+ "grok-2-vision": {
2099
+ maxTokens: 8192,
2100
+ contextWindow: 32768,
2101
+ supportsImages: true,
2102
+ supportsPromptCache: false,
2103
+ inputPrice: 2,
2104
+ outputPrice: 10,
2105
+ description: "xAI's Grok-2 Vision model with image support and 32K context window"
2106
+ },
2107
+ "grok-2-vision-1212": {
2108
+ maxTokens: 8192,
2109
+ contextWindow: 32768,
2110
+ supportsImages: true,
2111
+ supportsPromptCache: false,
2112
+ inputPrice: 2,
2113
+ outputPrice: 10,
2114
+ description: "xAI's Grok-2 Vision model (version 1212) with image support and 32K context window"
2115
+ },
2116
+ "grok-vision-beta": {
2117
+ maxTokens: 8192,
2118
+ contextWindow: 8192,
2119
+ supportsImages: true,
2120
+ supportsPromptCache: false,
2121
+ inputPrice: 5,
2122
+ outputPrice: 15,
2123
+ description: "xAI's Grok Vision Beta model with image support and 8K context window"
2124
+ },
2125
+ "grok-beta": {
2126
+ maxTokens: 8192,
2127
+ contextWindow: 131072,
2128
+ supportsImages: false,
2129
+ supportsPromptCache: false,
2130
+ inputPrice: 5,
2131
+ outputPrice: 15,
2132
+ description: "xAI's Grok Beta model (legacy) with 131K context window"
2133
+ }
2134
+ };
2135
+
99
2136
  // src/codebase-index.ts
100
2137
  var import_zod = require("zod");
101
2138
  var codebaseIndexConfigSchema = import_zod.z.object({
@@ -139,16 +2176,19 @@ var organizationSettingsSchema = import_zod2.z.object({
139
2176
  maxReadFileLine: import_zod2.z.number().optional(),
140
2177
  fuzzyMatchThreshold: import_zod2.z.number().optional()
141
2178
  }).optional(),
2179
+ cloudSettings: import_zod2.z.object({
2180
+ recordTaskMessages: import_zod2.z.boolean().optional()
2181
+ }).optional(),
142
2182
  allowList: organizationAllowListSchema
143
2183
  });
144
2184
 
145
2185
  // src/experiment.ts
146
2186
  var import_zod3 = require("zod");
147
- var experimentIds = ["autoCondenseContext", "powerSteering"];
2187
+ var experimentIds = ["powerSteering", "concurrentFileReads"];
148
2188
  var experimentIdsSchema = import_zod3.z.enum(experimentIds);
149
2189
  var experimentsSchema = import_zod3.z.object({
150
- autoCondenseContext: import_zod3.z.boolean(),
151
- powerSteering: import_zod3.z.boolean()
2190
+ powerSteering: import_zod3.z.boolean(),
2191
+ concurrentFileReads: import_zod3.z.boolean()
152
2192
  });
153
2193
 
154
2194
  // src/global-settings.ts
@@ -562,6 +2602,7 @@ var clineSays = [
562
2602
  "rooignore_error",
563
2603
  "diff_error",
564
2604
  "condense_context",
2605
+ "condense_context_error",
565
2606
  "codebase_search_result"
566
2607
  ];
567
2608
  var clineSaySchema = import_zod7.z.enum(clineSays);
@@ -626,6 +2667,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
626
2667
  return TelemetryEventName2;
627
2668
  })(TelemetryEventName || {});
628
2669
  var appPropertiesSchema = import_zod8.z.object({
2670
+ appName: import_zod8.z.string(),
629
2671
  appVersion: import_zod8.z.string(),
630
2672
  vscodeVersion: import_zod8.z.string(),
631
2673
  platform: import_zod8.z.string(),
@@ -663,13 +2705,16 @@ var rooCodeTelemetryEventSchema = import_zod8.z.discriminatedUnion("type", [
663
2705
  "Schema Validation Error" /* SCHEMA_VALIDATION_ERROR */,
664
2706
  "Diff Application Error" /* DIFF_APPLICATION_ERROR */,
665
2707
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
666
- "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */
2708
+ "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */,
2709
+ "Context Condensed" /* CONTEXT_CONDENSED */,
2710
+ "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */
667
2711
  ]),
668
2712
  properties: telemetryPropertiesSchema
669
2713
  }),
670
2714
  import_zod8.z.object({
671
2715
  type: import_zod8.z.literal("Task Message" /* TASK_MESSAGE */),
672
2716
  properties: import_zod8.z.object({
2717
+ taskId: import_zod8.z.string(),
673
2718
  message: clineMessageSchema
674
2719
  })
675
2720
  }),
@@ -859,7 +2904,9 @@ var globalSettingsSchema = import_zod12.z.object({
859
2904
  alwaysAllowExecute: import_zod12.z.boolean().optional(),
860
2905
  allowedCommands: import_zod12.z.array(import_zod12.z.string()).optional(),
861
2906
  allowedMaxRequests: import_zod12.z.number().nullish(),
2907
+ autoCondenseContext: import_zod12.z.boolean().optional(),
862
2908
  autoCondenseContextPercent: import_zod12.z.number().optional(),
2909
+ maxConcurrentFileReads: import_zod12.z.number().optional(),
863
2910
  browserToolEnabled: import_zod12.z.boolean().optional(),
864
2911
  browserViewportSize: import_zod12.z.string().optional(),
865
2912
  screenshotQuality: import_zod12.z.number().optional(),
@@ -927,7 +2974,9 @@ var GLOBAL_SETTINGS_KEYS = keysOf()([
927
2974
  "alwaysAllowExecute",
928
2975
  "allowedCommands",
929
2976
  "allowedMaxRequests",
2977
+ "autoCondenseContext",
930
2978
  "autoCondenseContextPercent",
2979
+ "maxConcurrentFileReads",
931
2980
  "browserToolEnabled",
932
2981
  "browserViewportSize",
933
2982
  "screenshotQuality",
@@ -1171,18 +3220,44 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1171
3220
  ]);
1172
3221
  // Annotate the CommonJS export names for ESM import in node:
1173
3222
  0 && (module.exports = {
3223
+ ANTHROPIC_DEFAULT_MAX_TOKENS,
3224
+ BEDROCK_DEFAULT_TEMPERATURE,
3225
+ BEDROCK_MAX_TOKENS,
3226
+ BEDROCK_REGIONS,
3227
+ BEDROCK_REGION_INFO,
3228
+ DEEP_SEEK_DEFAULT_TEMPERATURE,
3229
+ GLAMA_DEFAULT_TEMPERATURE,
1174
3230
  GLOBAL_SETTINGS_KEYS,
1175
3231
  GLOBAL_STATE_KEYS,
1176
3232
  IpcMessageType,
1177
3233
  IpcOrigin,
3234
+ LITELLM_COMPUTER_USE_MODELS,
3235
+ LMSTUDIO_DEFAULT_TEMPERATURE,
3236
+ MISTRAL_DEFAULT_TEMPERATURE,
3237
+ OPENAI_AZURE_AI_INFERENCE_PATH,
3238
+ OPENAI_NATIVE_DEFAULT_TEMPERATURE,
3239
+ OPENROUTER_DEFAULT_PROVIDER_NAME,
3240
+ OPEN_ROUTER_COMPUTER_USE_MODELS,
3241
+ OPEN_ROUTER_PROMPT_CACHING_MODELS,
3242
+ OPEN_ROUTER_REASONING_BUDGET_MODELS,
3243
+ OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS,
1178
3244
  ORGANIZATION_ALLOW_ALL,
1179
3245
  PROVIDER_SETTINGS_KEYS,
1180
3246
  RooCodeEventName,
1181
3247
  SECRET_STATE_KEYS,
1182
3248
  TaskCommandName,
1183
3249
  TelemetryEventName,
3250
+ VERTEX_REGIONS,
1184
3251
  ackSchema,
3252
+ anthropicDefaultModelId,
3253
+ anthropicModels,
1185
3254
  appPropertiesSchema,
3255
+ azureOpenAiDefaultApiVersion,
3256
+ bedrockDefaultModelId,
3257
+ bedrockDefaultPromptRouterModelId,
3258
+ bedrockModels,
3259
+ chutesDefaultModelId,
3260
+ chutesModels,
1186
3261
  clineAskSchema,
1187
3262
  clineAsks,
1188
3263
  clineMessageSchema,
@@ -1198,10 +3273,18 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1198
3273
  customModePromptsSchema,
1199
3274
  customModesSettingsSchema,
1200
3275
  customSupportPromptsSchema,
3276
+ deepSeekDefaultModelId,
3277
+ deepSeekModels,
1201
3278
  experimentIds,
1202
3279
  experimentIdsSchema,
1203
3280
  experimentsSchema,
3281
+ geminiDefaultModelId,
3282
+ geminiModels,
3283
+ glamaDefaultModelId,
3284
+ glamaDefaultModelInfo,
1204
3285
  globalSettingsSchema,
3286
+ groqDefaultModelId,
3287
+ groqModels,
1205
3288
  groupEntrySchema,
1206
3289
  groupOptionsSchema,
1207
3290
  historyItemSchema,
@@ -1213,10 +3296,19 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1213
3296
  keysOf,
1214
3297
  languages,
1215
3298
  languagesSchema,
3299
+ litellmDefaultModelId,
3300
+ litellmDefaultModelInfo,
3301
+ mistralDefaultModelId,
3302
+ mistralModels,
1216
3303
  modeConfigSchema,
1217
3304
  modelInfoSchema,
1218
3305
  modelParameters,
1219
3306
  modelParametersSchema,
3307
+ openAiModelInfoSaneDefaults,
3308
+ openAiNativeDefaultModelId,
3309
+ openAiNativeModels,
3310
+ openRouterDefaultModelId,
3311
+ openRouterDefaultModelInfo,
1220
3312
  organizationAllowListSchema,
1221
3313
  organizationSettingsSchema,
1222
3314
  promptComponentSchema,
@@ -1227,6 +3319,8 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1227
3319
  providerSettingsSchemaDiscriminated,
1228
3320
  reasoningEfforts,
1229
3321
  reasoningEffortsSchema,
3322
+ requestyDefaultModelId,
3323
+ requestyDefaultModelInfo,
1230
3324
  rooCodeEventsSchema,
1231
3325
  rooCodeSettingsSchema,
1232
3326
  rooCodeTelemetryEventSchema,
@@ -1243,6 +3337,14 @@ var commandExecutionStatusSchema = import_zod14.z.discriminatedUnion("status", [
1243
3337
  toolNames,
1244
3338
  toolNamesSchema,
1245
3339
  toolProgressStatusSchema,
1246
- toolUsageSchema
3340
+ toolUsageSchema,
3341
+ unboundDefaultModelId,
3342
+ unboundDefaultModelInfo,
3343
+ vertexDefaultModelId,
3344
+ vertexModels,
3345
+ vscodeLlmDefaultModelId,
3346
+ vscodeLlmModels,
3347
+ xaiDefaultModelId,
3348
+ xaiModels
1247
3349
  });
1248
3350
  //# sourceMappingURL=index.cjs.map