hyperclaw 5.1.9 → 5.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (41) hide show
  1. package/dist/chat-BQMPVq39.js +310 -0
  2. package/dist/chat-DGUj_NDO.js +310 -0
  3. package/dist/daemon-DbWcGYXq.js +5 -0
  4. package/dist/daemon-DkdMP-kQ.js +316 -0
  5. package/dist/daemon-RoGBr7z8.js +316 -0
  6. package/dist/daemon-XtFu-VzD.js +5 -0
  7. package/dist/engine-BY0XLiQ1.js +7 -0
  8. package/dist/engine-CaG7uwkg.js +305 -0
  9. package/dist/engine-Ccmpd_MH.js +305 -0
  10. package/dist/engine-DeQdDkFO.js +7 -0
  11. package/dist/hyperclawbot-BD5gNWO8.js +505 -0
  12. package/dist/hyperclawbot-D6cmZ6Bg.js +505 -0
  13. package/dist/mcp-loader-CbcyUzpM.js +94 -0
  14. package/dist/onboard-0RpraIyZ.js +11 -0
  15. package/dist/onboard-B8zjN52U.js +4149 -0
  16. package/dist/onboard-CGGZRH3J.js +4149 -0
  17. package/dist/onboard-xgJmy17T.js +11 -0
  18. package/dist/orchestrator-BeK8XyJY.js +6 -0
  19. package/dist/orchestrator-Cj5XWVVe.js +189 -0
  20. package/dist/orchestrator-TsNkrbVG.js +6 -0
  21. package/dist/orchestrator-zS6Manjk.js +189 -0
  22. package/dist/providers-Bk6Y_v9m.js +681 -0
  23. package/dist/providers-BvScOYvt.js +5 -0
  24. package/dist/providers-DEJ8Jtml.js +1089 -0
  25. package/dist/providers-EyJsZ-sr.js +5 -0
  26. package/dist/run-main.js +27 -27
  27. package/dist/server-CeZNKjAL.js +4 -0
  28. package/dist/server-DBUnNWZY.js +1255 -0
  29. package/dist/server-DUVnyXaR.js +1255 -0
  30. package/dist/server-Dy7Q-YLx.js +4 -0
  31. package/dist/skill-runtime-Bsf3CXxL.js +102 -0
  32. package/dist/skill-runtime-C_54rn66.js +5 -0
  33. package/dist/skill-runtime-Do7NLtoa.js +5 -0
  34. package/dist/skill-runtime-fDLfk3m3.js +102 -0
  35. package/dist/src-CQOQRIzS.js +458 -0
  36. package/dist/src-DQ9Q89FM.js +458 -0
  37. package/dist/src-DXaT-nxD.js +63 -0
  38. package/dist/src-DrDuFPKq.js +63 -0
  39. package/dist/sub-agent-tools-Bfk5uqJo.js +39 -0
  40. package/dist/sub-agent-tools-XeiuvDg5.js +39 -0
  41. package/package.json +1 -1
@@ -0,0 +1,681 @@
1
+ const require_chunk = require('./chunk-jS-bbMI5.js');
2
+ const chalk = require_chunk.__toESM(require("chalk"));
3
+
4
+ //#region src/cli/providers.ts
5
+ function getProvider(id) {
6
+ return PROVIDERS.find((p) => p.id === id);
7
+ }
8
+ /** Providers that support voice note transcription. Shown in wizard. */
9
+ function getTranscriptionProviders() {
10
+ return PROVIDERS.filter((p) => p.supportsTranscription);
11
+ }
12
+ function formatModel(m) {
13
+ const badges = [];
14
+ if (m.flagship) badges.push(chalk.default.yellow("★"));
15
+ if (m.reasoning) badges.push(chalk.default.magenta("reasoning"));
16
+ if (m.fast) badges.push(chalk.default.green("fast"));
17
+ if (m.vision) badges.push(chalk.default.cyan("vision"));
18
+ const ctx = m.contextK >= 1e3 ? `${m.contextK}K` : `${m.contextK}K`;
19
+ return `${badges.join(" ")} ${m.name} ${chalk.default.gray(`ctx ${ctx}`)}`.trim();
20
+ }
21
+ var PROVIDERS;
22
+ var init_providers = require_chunk.__esm({ "src/cli/providers.ts"() {
23
+ PROVIDERS = [
24
+ {
25
+ id: "anthropic",
26
+ displayName: "🎭 Anthropic (API Key)",
27
+ authType: "api_key",
28
+ authLabel: "Anthropic API Key",
29
+ authHint: "console.anthropic.com → API Keys",
30
+ models: [
31
+ {
32
+ id: "claude-opus-4-6",
33
+ name: "Claude Opus 4.6",
34
+ contextK: 200,
35
+ reasoning: true,
36
+ flagship: true
37
+ },
38
+ {
39
+ id: "claude-sonnet-4-6",
40
+ name: "Claude Sonnet 4.6",
41
+ contextK: 200,
42
+ reasoning: true
43
+ },
44
+ {
45
+ id: "claude-haiku-4-5-20251001",
46
+ name: "Claude Haiku 4.5",
47
+ contextK: 200,
48
+ fast: true
49
+ }
50
+ ]
51
+ },
52
+ {
53
+ id: "anthropic-oauth",
54
+ displayName: "🎭 Anthropic (OAuth — Claude Code/Max)",
55
+ authType: "oauth",
56
+ authLabel: "Claude OAuth credentials",
57
+ authHint: "Reuses ~/.claude/.credentials.json (Claude Code CLI) or macOS Keychain \"Claude Code-credentials\"",
58
+ models: [{
59
+ id: "claude-opus-4-6",
60
+ name: "Claude Opus 4.6",
61
+ contextK: 200,
62
+ reasoning: true,
63
+ flagship: true
64
+ }, {
65
+ id: "claude-sonnet-4-6",
66
+ name: "Claude Sonnet 4.6",
67
+ contextK: 200,
68
+ reasoning: true
69
+ }]
70
+ },
71
+ {
72
+ id: "anthropic-setup-token",
73
+ displayName: "🎭 Anthropic (setup-token)",
74
+ authType: "api_key",
75
+ authLabel: "Anthropic setup-token",
76
+ authHint: "Run `claude setup-token` on any machine → paste the token here",
77
+ models: [{
78
+ id: "claude-opus-4-6",
79
+ name: "Claude Opus 4.6",
80
+ contextK: 200,
81
+ reasoning: true,
82
+ flagship: true
83
+ }, {
84
+ id: "claude-sonnet-4-6",
85
+ name: "Claude Sonnet 4.6",
86
+ contextK: 200,
87
+ reasoning: true
88
+ }]
89
+ },
90
+ {
91
+ id: "vercel-ai",
92
+ displayName: "▲ Vercel AI Gateway",
93
+ authType: "api_key",
94
+ authLabel: "Vercel AI Gateway API Key",
95
+ authHint: "vercel.com/docs/ai — multi-model proxy (AI_GATEWAY_API_KEY)",
96
+ baseUrl: "https://ai-gateway.vercel.sh/v1",
97
+ models: [
98
+ {
99
+ id: "openai/gpt-4o",
100
+ name: "GPT-4o (via Vercel)",
101
+ contextK: 128,
102
+ flagship: true
103
+ },
104
+ {
105
+ id: "anthropic/claude-sonnet-4-5",
106
+ name: "Claude Sonnet (via Vercel)",
107
+ contextK: 200
108
+ },
109
+ {
110
+ id: "google/gemini-2.0-flash",
111
+ name: "Gemini 2.0 Flash (via Vercel)",
112
+ contextK: 1e3,
113
+ fast: true
114
+ }
115
+ ]
116
+ },
117
+ {
118
+ id: "opencode-zen",
119
+ displayName: "🧘 OpenCode Zen (multi-model proxy)",
120
+ authType: "api_key",
121
+ authLabel: "OpenCode Zen API Key",
122
+ authHint: "opencode.ai/auth — OPENCODE_API_KEY",
123
+ baseUrl: "https://api.opencode.ai/v1",
124
+ models: [{
125
+ id: "auto",
126
+ name: "Auto (best available)",
127
+ contextK: 200,
128
+ flagship: true
129
+ }]
130
+ },
131
+ {
132
+ id: "openrouter",
133
+ displayName: "🌐 OpenRouter",
134
+ authType: "api_key",
135
+ authLabel: "OpenRouter API Key",
136
+ authHint: "openrouter.ai/keys",
137
+ baseUrl: "https://openrouter.ai/api/v1",
138
+ supportsTranscription: true,
139
+ models: [
140
+ {
141
+ id: "openrouter/auto",
142
+ name: "Auto (best available)",
143
+ contextK: 200,
144
+ flagship: true
145
+ },
146
+ {
147
+ id: "anthropic/claude-opus-4-6",
148
+ name: "Claude Opus 4.6",
149
+ contextK: 200,
150
+ reasoning: true
151
+ },
152
+ {
153
+ id: "anthropic/claude-sonnet-4-6",
154
+ name: "Claude Sonnet 4.6",
155
+ contextK: 200
156
+ },
157
+ {
158
+ id: "openai/gpt-4o",
159
+ name: "GPT-4o",
160
+ contextK: 128,
161
+ vision: true
162
+ },
163
+ {
164
+ id: "openai/o3",
165
+ name: "o3 (reasoning)",
166
+ contextK: 200,
167
+ reasoning: true
168
+ },
169
+ {
170
+ id: "google/gemini-2.5-pro-preview-03-25",
171
+ name: "Gemini 2.5 Pro",
172
+ contextK: 1e3,
173
+ reasoning: true
174
+ },
175
+ {
176
+ id: "google/gemini-2.0-flash",
177
+ name: "Gemini 2.0 Flash",
178
+ contextK: 1e3,
179
+ fast: true
180
+ },
181
+ {
182
+ id: "x-ai/grok-3-beta",
183
+ name: "Grok 3",
184
+ contextK: 131,
185
+ reasoning: true
186
+ },
187
+ {
188
+ id: "deepseek/deepseek-r1",
189
+ name: "DeepSeek R1",
190
+ contextK: 64,
191
+ reasoning: true
192
+ },
193
+ {
194
+ id: "meta-llama/llama-3.3-70b-instruct",
195
+ name: "Llama 3.3 70B",
196
+ contextK: 128
197
+ },
198
+ {
199
+ id: "qwen/qwen-2.5-72b-instruct",
200
+ name: "Qwen 2.5 72B",
201
+ contextK: 128
202
+ },
203
+ {
204
+ id: "mistralai/mistral-large",
205
+ name: "Mistral Large",
206
+ contextK: 128
207
+ }
208
+ ]
209
+ },
210
+ {
211
+ id: "openai",
212
+ displayName: "🧠 OpenAI",
213
+ authType: "api_key",
214
+ authLabel: "OpenAI API Key",
215
+ authHint: "platform.openai.com/api-keys",
216
+ baseUrl: "https://api.openai.com/v1",
217
+ supportsTranscription: true,
218
+ models: [
219
+ {
220
+ id: "gpt-4o",
221
+ name: "GPT-4o",
222
+ contextK: 128,
223
+ vision: true,
224
+ flagship: true
225
+ },
226
+ {
227
+ id: "gpt-4o-mini",
228
+ name: "GPT-4o Mini",
229
+ contextK: 128,
230
+ fast: true
231
+ },
232
+ {
233
+ id: "o3",
234
+ name: "o3 (reasoning)",
235
+ contextK: 200,
236
+ reasoning: true
237
+ },
238
+ {
239
+ id: "o4-mini",
240
+ name: "o4-mini (reasoning)",
241
+ contextK: 200,
242
+ reasoning: true,
243
+ fast: true
244
+ }
245
+ ]
246
+ },
247
+ {
248
+ id: "google",
249
+ displayName: "🔍 Google",
250
+ authType: "api_key",
251
+ authLabel: "Google AI API Key",
252
+ authHint: "aistudio.google.com/app/apikey",
253
+ baseUrl: "https://generativelanguage.googleapis.com/v1beta/openai",
254
+ supportsTranscription: true,
255
+ models: [
256
+ {
257
+ id: "gemini-2.5-pro-preview-03-25",
258
+ name: "Gemini 2.5 Pro (paid)",
259
+ contextK: 1e3,
260
+ reasoning: true,
261
+ flagship: true
262
+ },
263
+ {
264
+ id: "gemini-2.0-flash",
265
+ name: "Gemini 2.0 Flash (free)",
266
+ contextK: 1e3,
267
+ fast: true
268
+ },
269
+ {
270
+ id: "gemini-2.0-flash-lite",
271
+ name: "Gemini 2.0 Flash Lite (free)",
272
+ contextK: 1e3,
273
+ fast: true
274
+ },
275
+ {
276
+ id: "gemini-1.5-flash",
277
+ name: "Gemini 1.5 Flash (free)",
278
+ contextK: 1e3,
279
+ fast: true
280
+ }
281
+ ]
282
+ },
283
+ {
284
+ id: "xai",
285
+ displayName: "⚡ xAI (Grok)",
286
+ authType: "api_key",
287
+ authLabel: "xAI API Key",
288
+ authHint: "console.x.ai",
289
+ baseUrl: "https://api.x.ai/v1",
290
+ models: [{
291
+ id: "grok-3-beta",
292
+ name: "Grok 3",
293
+ contextK: 131,
294
+ reasoning: true,
295
+ flagship: true
296
+ }, {
297
+ id: "grok-3-mini-beta",
298
+ name: "Grok 3 Mini",
299
+ contextK: 131,
300
+ fast: true
301
+ }]
302
+ },
303
+ {
304
+ id: "minimax",
305
+ displayName: "🎯 MiniMax",
306
+ authType: "api_key",
307
+ authLabel: "MiniMax API Key",
308
+ authHint: "platform.minimaxi.com",
309
+ baseUrl: "https://api.minimaxi.chat/v1",
310
+ models: [{
311
+ id: "MiniMax-Text-01",
312
+ name: "MiniMax Text-01",
313
+ contextK: 1e3,
314
+ flagship: true
315
+ }, {
316
+ id: "abab6.5s-chat",
317
+ name: "ABAB 6.5S",
318
+ contextK: 245
319
+ }]
320
+ },
321
+ {
322
+ id: "moonshot",
323
+ displayName: "🌙 Moonshot (Kimi)",
324
+ authType: "api_key",
325
+ authLabel: "Moonshot API Key",
326
+ authHint: "platform.moonshot.cn",
327
+ baseUrl: "https://api.moonshot.cn/v1",
328
+ models: [{
329
+ id: "moonshot-v1-128k",
330
+ name: "Moonshot v1 128K",
331
+ contextK: 128,
332
+ flagship: true
333
+ }, {
334
+ id: "moonshot-v1-8k",
335
+ name: "Moonshot v1 8K",
336
+ contextK: 8,
337
+ fast: true
338
+ }]
339
+ },
340
+ {
341
+ id: "qwen",
342
+ displayName: "🐉 Qwen (Alibaba)",
343
+ authType: "api_key",
344
+ authLabel: "DashScope API Key",
345
+ authHint: "dashscope.aliyuncs.com",
346
+ baseUrl: "https://dashscope-intl.aliyuncs.com/compatible-mode/v1",
347
+ models: [
348
+ {
349
+ id: "qwen-max",
350
+ name: "Qwen Max",
351
+ contextK: 32,
352
+ flagship: true
353
+ },
354
+ {
355
+ id: "qwen-plus",
356
+ name: "Qwen Plus",
357
+ contextK: 128
358
+ },
359
+ {
360
+ id: "qwen-turbo",
361
+ name: "Qwen Turbo",
362
+ contextK: 128,
363
+ fast: true
364
+ },
365
+ {
366
+ id: "qwen3-235b-a22b",
367
+ name: "Qwen3 235B",
368
+ contextK: 32,
369
+ reasoning: true
370
+ }
371
+ ]
372
+ },
373
+ {
374
+ id: "zai",
375
+ displayName: "🔧 Z.AI (Zhipu GLM)",
376
+ authType: "api_key",
377
+ authLabel: "Z.AI API Key",
378
+ authHint: "open.bigmodel.cn/usercenter/apikeys",
379
+ baseUrl: "https://open.bigmodel.cn/api/paas/v4",
380
+ models: [
381
+ {
382
+ id: "glm-4-plus",
383
+ name: "GLM-4 Plus",
384
+ contextK: 128,
385
+ flagship: true
386
+ },
387
+ {
388
+ id: "glm-4-flash",
389
+ name: "GLM-4 Flash (free)",
390
+ contextK: 128,
391
+ fast: true
392
+ },
393
+ {
394
+ id: "glm-z1-flash",
395
+ name: "GLM-Z1 Flash (free, reasoning)",
396
+ contextK: 128,
397
+ reasoning: true,
398
+ fast: true
399
+ }
400
+ ]
401
+ },
402
+ {
403
+ id: "litellm",
404
+ displayName: "🔀 LiteLLM (proxy)",
405
+ authType: "api_key",
406
+ authLabel: "LiteLLM Master Key",
407
+ authHint: "Your self-hosted LiteLLM proxy key",
408
+ models: [{
409
+ id: "gpt-4o",
410
+ name: "GPT-4o (via proxy)",
411
+ contextK: 128,
412
+ flagship: true
413
+ }]
414
+ },
415
+ {
416
+ id: "cloudflare",
417
+ displayName: "☁️ Cloudflare AI Gateway",
418
+ authType: "api_key",
419
+ authLabel: "Cloudflare API Token",
420
+ authHint: "dash.cloudflare.com → AI → Gateway",
421
+ models: [{
422
+ id: "@cf/meta/llama-3.3-70b-instruct-fp8-fast",
423
+ name: "Llama 3.3 70B (CF)",
424
+ contextK: 128,
425
+ flagship: true
426
+ }]
427
+ },
428
+ {
429
+ id: "copilot",
430
+ displayName: "🤖 GitHub Copilot",
431
+ authType: "oauth",
432
+ authLabel: "GitHub OAuth Token",
433
+ authHint: "github.com/settings/tokens",
434
+ models: [{
435
+ id: "gpt-4o",
436
+ name: "GPT-4o (Copilot)",
437
+ contextK: 128,
438
+ flagship: true
439
+ }, {
440
+ id: "claude-sonnet-4-5",
441
+ name: "Claude Sonnet (Copilot)",
442
+ contextK: 200
443
+ }]
444
+ },
445
+ {
446
+ id: "groq",
447
+ displayName: "⚡ Groq (Fast Inference)",
448
+ authType: "api_key",
449
+ authLabel: "Groq API Key",
450
+ authHint: "console.groq.com/keys",
451
+ baseUrl: "https://api.groq.com/openai/v1",
452
+ models: [
453
+ {
454
+ id: "llama-3.3-70b-versatile",
455
+ name: "Llama 3.3 70B Versatile",
456
+ contextK: 128,
457
+ flagship: true
458
+ },
459
+ {
460
+ id: "llama-3.1-8b-instant",
461
+ name: "Llama 3.1 8B Instant",
462
+ contextK: 128,
463
+ fast: true
464
+ },
465
+ {
466
+ id: "mixtral-8x7b-32768",
467
+ name: "Mixtral 8x7B",
468
+ contextK: 32
469
+ },
470
+ {
471
+ id: "gemma2-9b-it",
472
+ name: "Gemma 2 9B",
473
+ contextK: 8,
474
+ fast: true
475
+ }
476
+ ]
477
+ },
478
+ {
479
+ id: "mistral",
480
+ displayName: "🌀 Mistral AI",
481
+ authType: "api_key",
482
+ authLabel: "Mistral API Key",
483
+ authHint: "console.mistral.ai/api-keys",
484
+ baseUrl: "https://api.mistral.ai/v1",
485
+ models: [
486
+ {
487
+ id: "mistral-large-latest",
488
+ name: "Mistral Large",
489
+ contextK: 128,
490
+ flagship: true
491
+ },
492
+ {
493
+ id: "mistral-medium-latest",
494
+ name: "Mistral Medium",
495
+ contextK: 128
496
+ },
497
+ {
498
+ id: "mistral-small-latest",
499
+ name: "Mistral Small",
500
+ contextK: 128,
501
+ fast: true
502
+ },
503
+ {
504
+ id: "codestral-latest",
505
+ name: "Codestral (code)",
506
+ contextK: 256
507
+ }
508
+ ]
509
+ },
510
+ {
511
+ id: "deepseek",
512
+ displayName: "🔬 DeepSeek",
513
+ authType: "api_key",
514
+ authLabel: "DeepSeek API Key",
515
+ authHint: "platform.deepseek.com/api_keys",
516
+ baseUrl: "https://api.deepseek.com/v1",
517
+ models: [{
518
+ id: "deepseek-chat",
519
+ name: "DeepSeek V3",
520
+ contextK: 64,
521
+ flagship: true
522
+ }, {
523
+ id: "deepseek-reasoner",
524
+ name: "DeepSeek R1 (reasoning)",
525
+ contextK: 64,
526
+ reasoning: true
527
+ }]
528
+ },
529
+ {
530
+ id: "perplexity",
531
+ displayName: "🔍 Perplexity (Search-augmented)",
532
+ authType: "api_key",
533
+ authLabel: "Perplexity API Key",
534
+ authHint: "perplexity.ai/settings/api",
535
+ baseUrl: "https://api.perplexity.ai",
536
+ models: [
537
+ {
538
+ id: "sonar-pro",
539
+ name: "Sonar Pro (search)",
540
+ contextK: 200,
541
+ flagship: true
542
+ },
543
+ {
544
+ id: "sonar",
545
+ name: "Sonar (search, fast)",
546
+ contextK: 128,
547
+ fast: true
548
+ },
549
+ {
550
+ id: "sonar-reasoning",
551
+ name: "Sonar Reasoning",
552
+ contextK: 128,
553
+ reasoning: true
554
+ }
555
+ ]
556
+ },
557
+ {
558
+ id: "huggingface",
559
+ displayName: "🤗 Hugging Face",
560
+ authType: "api_key",
561
+ authLabel: "HuggingFace API Token",
562
+ authHint: "huggingface.co/settings/tokens",
563
+ baseUrl: "https://api-inference.huggingface.co/v1",
564
+ models: [
565
+ {
566
+ id: "Qwen/Qwen2.5-72B-Instruct",
567
+ name: "Qwen 2.5 72B",
568
+ contextK: 128,
569
+ flagship: true
570
+ },
571
+ {
572
+ id: "meta-llama/Llama-3.3-70B-Instruct",
573
+ name: "Llama 3.3 70B",
574
+ contextK: 128
575
+ },
576
+ {
577
+ id: "mistralai/Mistral-7B-Instruct-v0.3",
578
+ name: "Mistral 7B",
579
+ contextK: 32,
580
+ fast: true
581
+ }
582
+ ]
583
+ },
584
+ {
585
+ id: "ollama",
586
+ displayName: "🦙 Ollama (Local)",
587
+ authType: "none",
588
+ authLabel: "No API key needed",
589
+ authHint: "ollama.ai — run `ollama serve` first",
590
+ baseUrl: "http://localhost:11434/v1",
591
+ models: [
592
+ {
593
+ id: "llama3.3",
594
+ name: "Llama 3.3 (local)",
595
+ contextK: 128,
596
+ flagship: true
597
+ },
598
+ {
599
+ id: "mistral",
600
+ name: "Mistral (local)",
601
+ contextK: 32
602
+ },
603
+ {
604
+ id: "codellama",
605
+ name: "CodeLlama (local)",
606
+ contextK: 16
607
+ },
608
+ {
609
+ id: "phi4",
610
+ name: "Phi-4 (local)",
611
+ contextK: 16,
612
+ fast: true
613
+ },
614
+ {
615
+ id: "__manual__",
616
+ name: "Enter model name manually",
617
+ contextK: 128
618
+ }
619
+ ]
620
+ },
621
+ {
622
+ id: "lmstudio",
623
+ displayName: "🖥️ LM Studio (Local)",
624
+ authType: "none",
625
+ authLabel: "No API key needed",
626
+ authHint: "lmstudio.ai — enable local server in app",
627
+ baseUrl: "http://localhost:1234/v1",
628
+ models: [{
629
+ id: "__manual__",
630
+ name: "Enter loaded model ID",
631
+ contextK: 128,
632
+ flagship: true
633
+ }]
634
+ },
635
+ {
636
+ id: "custom",
637
+ displayName: "🔌 Custom (OpenAI-compatible API)",
638
+ authType: "api_key",
639
+ authLabel: "API Key",
640
+ authHint: "Any OpenAI-compatible /chat/completions API (e.g. Ads Power, Proxies, new LLM APIs)",
641
+ models: [{
642
+ id: "__manual__",
643
+ name: "Enter model ID manually",
644
+ contextK: 128,
645
+ flagship: true
646
+ }]
647
+ }
648
+ ];
649
+ } });
650
+
651
+ //#endregion
652
+ Object.defineProperty(exports, 'PROVIDERS', {
653
+ enumerable: true,
654
+ get: function () {
655
+ return PROVIDERS;
656
+ }
657
+ });
658
+ Object.defineProperty(exports, 'formatModel', {
659
+ enumerable: true,
660
+ get: function () {
661
+ return formatModel;
662
+ }
663
+ });
664
+ Object.defineProperty(exports, 'getProvider', {
665
+ enumerable: true,
666
+ get: function () {
667
+ return getProvider;
668
+ }
669
+ });
670
+ Object.defineProperty(exports, 'getTranscriptionProviders', {
671
+ enumerable: true,
672
+ get: function () {
673
+ return getTranscriptionProviders;
674
+ }
675
+ });
676
+ Object.defineProperty(exports, 'init_providers', {
677
+ enumerable: true,
678
+ get: function () {
679
+ return init_providers;
680
+ }
681
+ });
@@ -0,0 +1,5 @@
1
+ const require_chunk = require('./chunk-jS-bbMI5.js');
2
+ const require_providers = require('./providers-Bk6Y_v9m.js');
3
+
4
+ require_providers.init_providers();
5
+ exports.getProvider = require_providers.getProvider;