libre-webui 0.6.2 → 0.7.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (57) hide show
  1. package/backend/dist/db.d.ts.map +1 -1
  2. package/backend/dist/db.js +17 -0
  3. package/backend/dist/db.js.map +1 -1
  4. package/backend/dist/routes/auth.js +2 -2
  5. package/backend/dist/routes/auth.js.map +1 -1
  6. package/backend/dist/routes/plugins.d.ts.map +1 -1
  7. package/backend/dist/routes/plugins.js +162 -0
  8. package/backend/dist/routes/plugins.js.map +1 -1
  9. package/backend/dist/services/authService.d.ts +1 -1
  10. package/backend/dist/services/authService.d.ts.map +1 -1
  11. package/backend/dist/services/authService.js +1 -1
  12. package/backend/dist/services/authService.js.map +1 -1
  13. package/backend/dist/services/pluginService.d.ts +9 -0
  14. package/backend/dist/services/pluginService.d.ts.map +1 -1
  15. package/backend/dist/services/pluginService.js +99 -24
  16. package/backend/dist/services/pluginService.js.map +1 -1
  17. package/backend/dist/services/pluginVariablesService.d.ts +36 -0
  18. package/backend/dist/services/pluginVariablesService.d.ts.map +1 -0
  19. package/backend/dist/services/pluginVariablesService.js +218 -0
  20. package/backend/dist/services/pluginVariablesService.js.map +1 -0
  21. package/backend/dist/types/index.d.ts +14 -0
  22. package/backend/dist/types/index.d.ts.map +1 -1
  23. package/backend/dist/types/index.js.map +1 -1
  24. package/frontend/dist/assets/index-DMlahgv7.js +51 -0
  25. package/frontend/dist/css/{index-Dh5G0uFZ.css → index-D9uKnbHR.css} +1 -1
  26. package/frontend/dist/index.html +6 -6
  27. package/frontend/dist/js/{ArtifactContainer-BLHcg4YG.js → ArtifactContainer-D_igWPrR.js} +2 -2
  28. package/frontend/dist/js/{ArtifactDemoPage-Br-19dKk.js → ArtifactDemoPage-BW64ikxs.js} +1 -1
  29. package/frontend/dist/js/{ChatPage-CQGVCtT_.js → ChatPage-CIUGXjJ_.js} +10 -10
  30. package/frontend/dist/js/{GalleryPage-CvA1Sox3.js → GalleryPage-CMqUQtns.js} +1 -1
  31. package/frontend/dist/js/ModelsPage-DQ2ud-YG.js +1 -0
  32. package/frontend/dist/js/PersonasPage-HqcbbtTL.js +13 -0
  33. package/frontend/dist/js/{UserManagementPage-H89n1w08.js → UserManagementPage-C4vtOCZ6.js} +1 -1
  34. package/frontend/dist/js/{markdown-vendor-DeGQOJEA.js → markdown-vendor-QwEm2Xso.js} +1 -1
  35. package/frontend/dist/js/{router-vendor-CYqbjR9t.js → router-vendor-CmbUVdlE.js} +5 -5
  36. package/frontend/dist/js/ui-vendor-DxZsuKzb.js +177 -0
  37. package/frontend/dist/js/utils-vendor-0Ymd3GFQ.js +6 -0
  38. package/package.json +1 -1
  39. package/plugins/anthropic.json +74 -13
  40. package/plugins/comfyui.json +45 -4
  41. package/plugins/elevenlabs.json +42 -2
  42. package/plugins/gemini.json +106 -45
  43. package/plugins/github.json +72 -11
  44. package/plugins/groq.json +74 -13
  45. package/plugins/huggingface.json +93 -6
  46. package/plugins/kyutai-tts-1.6b.json +76 -14
  47. package/plugins/kyutai-tts.json +32 -3
  48. package/plugins/mistral.json +122 -61
  49. package/plugins/openai-tts.json +38 -3
  50. package/plugins/openai.json +81 -3
  51. package/plugins/openrouter.json +402 -341
  52. package/plugins/qwen-tts.json +33 -3
  53. package/frontend/dist/assets/index-DE-xmTDC.js +0 -51
  54. package/frontend/dist/js/ModelsPage-9djD6Uj9.js +0 -1
  55. package/frontend/dist/js/PersonasPage-ajydrtrS.js +0 -13
  56. package/frontend/dist/js/ui-vendor-DnQP3_cJ.js +0 -177
  57. package/frontend/dist/js/utils-vendor-COrKO_vV.js +0 -6
@@ -48,11 +48,51 @@
48
48
  "Serena"
49
49
  ],
50
50
  "default_voice": "Rachel",
51
- "formats": ["mp3", "pcm", "ulaw"],
51
+ "formats": [
52
+ "mp3",
53
+ "pcm",
54
+ "ulaw"
55
+ ],
52
56
  "default_format": "mp3",
53
57
  "max_characters": 5000,
54
58
  "supports_streaming": true
55
59
  }
56
60
  }
57
- }
61
+ },
62
+ "variables": [
63
+ {
64
+ "name": "endpoint",
65
+ "type": "string",
66
+ "label": "API Endpoint",
67
+ "description": "Override the default API endpoint URL.",
68
+ "default": "https://api.elevenlabs.io/v1/text-to-speech/{voice_id}"
69
+ },
70
+ {
71
+ "name": "stability",
72
+ "type": "number",
73
+ "label": "Stability",
74
+ "description": "Voice stability (0-1). Higher values are more consistent.",
75
+ "default": 0.5,
76
+ "min": 0,
77
+ "max": 1
78
+ },
79
+ {
80
+ "name": "similarity_boost",
81
+ "type": "number",
82
+ "label": "Similarity Boost",
83
+ "description": "Voice similarity boost (0-1).",
84
+ "default": 0.75,
85
+ "min": 0,
86
+ "max": 1
87
+ },
88
+ {
89
+ "name": "speed",
90
+ "type": "number",
91
+ "label": "Speed",
92
+ "description": "Playback speed multiplier.",
93
+ "default": 1,
94
+ "min": 0.25,
95
+ "max": 4
96
+ }
97
+ ]
58
98
  }
@@ -9,49 +9,110 @@
9
9
  "key_env": "GEMINI_API_KEY"
10
10
  },
11
11
  "model_map": [
12
- "aqa",
13
- "deep-research-pro-preview-12-2025",
14
- "gemini-2.0-flash",
15
- "gemini-2.0-flash-001",
16
- "gemini-2.0-flash-exp",
17
- "gemini-2.0-flash-exp-image-generation",
18
- "gemini-2.0-flash-lite",
19
- "gemini-2.0-flash-lite-001",
20
- "gemini-2.0-flash-lite-preview",
21
- "gemini-2.0-flash-lite-preview-02-05",
22
- "gemini-2.5-computer-use-preview-10-2025",
23
- "gemini-2.5-flash",
24
- "gemini-2.5-flash-image",
25
- "gemini-2.5-flash-image-preview",
26
- "gemini-2.5-flash-lite",
27
- "gemini-2.5-flash-lite-preview-09-2025",
28
- "gemini-2.5-flash-preview-09-2025",
29
- "gemini-2.5-flash-preview-tts",
30
- "gemini-2.5-pro",
31
- "gemini-2.5-pro-preview-tts",
32
- "gemini-3-pro-image-preview",
33
- "gemini-3-pro-preview",
34
- "gemini-exp-1206",
35
- "gemini-flash-latest",
36
- "gemini-flash-lite-latest",
37
- "gemini-pro-latest",
38
- "gemini-robotics-er-1.5-preview",
39
- "gemma-3-12b-it",
40
- "gemma-3-1b-it",
41
- "gemma-3-27b-it",
42
- "gemma-3-4b-it",
43
- "gemma-3n-e2b-it",
44
- "gemma-3n-e4b-it",
45
- "imagen-4.0-fast-generate-001",
46
- "imagen-4.0-generate-001",
47
- "imagen-4.0-generate-preview-06-06",
48
- "imagen-4.0-ultra-generate-001",
49
- "imagen-4.0-ultra-generate-preview-06-06",
50
- "nano-banana-pro-preview",
51
- "veo-2.0-generate-001",
52
- "veo-3.0-fast-generate-001",
53
- "veo-3.0-generate-001",
54
- "veo-3.1-fast-generate-preview",
55
- "veo-3.1-generate-preview"
56
- ]
12
+ "aqa",
13
+ "deep-research-pro-preview-12-2025",
14
+ "gemini-2.0-flash",
15
+ "gemini-2.0-flash-001",
16
+ "gemini-2.0-flash-exp",
17
+ "gemini-2.0-flash-exp-image-generation",
18
+ "gemini-2.0-flash-lite",
19
+ "gemini-2.0-flash-lite-001",
20
+ "gemini-2.0-flash-lite-preview",
21
+ "gemini-2.0-flash-lite-preview-02-05",
22
+ "gemini-2.5-computer-use-preview-10-2025",
23
+ "gemini-2.5-flash",
24
+ "gemini-2.5-flash-image",
25
+ "gemini-2.5-flash-image-preview",
26
+ "gemini-2.5-flash-lite",
27
+ "gemini-2.5-flash-lite-preview-09-2025",
28
+ "gemini-2.5-flash-preview-09-2025",
29
+ "gemini-2.5-flash-preview-tts",
30
+ "gemini-2.5-pro",
31
+ "gemini-2.5-pro-preview-tts",
32
+ "gemini-3-pro-image-preview",
33
+ "gemini-3-pro-preview",
34
+ "gemini-exp-1206",
35
+ "gemini-flash-latest",
36
+ "gemini-flash-lite-latest",
37
+ "gemini-pro-latest",
38
+ "gemini-robotics-er-1.5-preview",
39
+ "gemma-3-12b-it",
40
+ "gemma-3-1b-it",
41
+ "gemma-3-27b-it",
42
+ "gemma-3-4b-it",
43
+ "gemma-3n-e2b-it",
44
+ "gemma-3n-e4b-it",
45
+ "imagen-4.0-fast-generate-001",
46
+ "imagen-4.0-generate-001",
47
+ "imagen-4.0-generate-preview-06-06",
48
+ "imagen-4.0-ultra-generate-001",
49
+ "imagen-4.0-ultra-generate-preview-06-06",
50
+ "nano-banana-pro-preview",
51
+ "veo-2.0-generate-001",
52
+ "veo-3.0-fast-generate-001",
53
+ "veo-3.0-generate-001",
54
+ "veo-3.1-fast-generate-preview",
55
+ "veo-3.1-generate-preview"
56
+ ],
57
+ "variables": [
58
+ {
59
+ "name": "endpoint",
60
+ "type": "string",
61
+ "label": "API Endpoint",
62
+ "description": "Override the default API endpoint URL.",
63
+ "default": "https://generativelanguage.googleapis.com/v1beta/models/{model}:generateContent"
64
+ },
65
+ {
66
+ "name": "temperature",
67
+ "type": "number",
68
+ "label": "Temperature",
69
+ "description": "Controls randomness. Lower values are more focused, higher values more creative.",
70
+ "default": 0.7,
71
+ "min": 0,
72
+ "max": 2
73
+ },
74
+ {
75
+ "name": "max_tokens",
76
+ "type": "number",
77
+ "label": "Max Tokens",
78
+ "description": "Maximum number of tokens to generate in the response.",
79
+ "default": 4096,
80
+ "min": 1,
81
+ "max": 128000
82
+ },
83
+ {
84
+ "name": "top_p",
85
+ "type": "number",
86
+ "label": "Top P",
87
+ "description": "Nucleus sampling. Lower values make output more focused.",
88
+ "default": 1,
89
+ "min": 0,
90
+ "max": 1
91
+ },
92
+ {
93
+ "name": "frequency_penalty",
94
+ "type": "number",
95
+ "label": "Frequency Penalty",
96
+ "description": "Reduces repetition of frequent tokens.",
97
+ "default": 0,
98
+ "min": -2,
99
+ "max": 2
100
+ },
101
+ {
102
+ "name": "presence_penalty",
103
+ "type": "number",
104
+ "label": "Presence Penalty",
105
+ "description": "Encourages the model to talk about new topics.",
106
+ "default": 0,
107
+ "min": -2,
108
+ "max": 2
109
+ },
110
+ {
111
+ "name": "stream",
112
+ "type": "boolean",
113
+ "label": "Stream Response",
114
+ "description": "Stream the response token by token.",
115
+ "default": true
116
+ }
117
+ ]
57
118
  }
@@ -9,15 +9,76 @@
9
9
  "key_env": "GITHUB_API_KEY"
10
10
  },
11
11
  "model_map": [
12
- "AI21-Jamba-Instruct",
13
- "Meta-Llama-3-70B-Instruct",
14
- "Meta-Llama-3-8B-Instruct",
15
- "Meta-Llama-3.1-405B-Instruct",
16
- "Meta-Llama-3.1-70B-Instruct",
17
- "Meta-Llama-3.1-8B-Instruct",
18
- "Mistral-Nemo",
19
- "Mistral-large-2407",
20
- "gpt-4o",
21
- "gpt-4o-mini"
22
- ]
12
+ "AI21-Jamba-Instruct",
13
+ "Meta-Llama-3-70B-Instruct",
14
+ "Meta-Llama-3-8B-Instruct",
15
+ "Meta-Llama-3.1-405B-Instruct",
16
+ "Meta-Llama-3.1-70B-Instruct",
17
+ "Meta-Llama-3.1-8B-Instruct",
18
+ "Mistral-Nemo",
19
+ "Mistral-large-2407",
20
+ "gpt-4o",
21
+ "gpt-4o-mini"
22
+ ],
23
+ "variables": [
24
+ {
25
+ "name": "endpoint",
26
+ "type": "string",
27
+ "label": "API Endpoint",
28
+ "description": "Override the default API endpoint URL.",
29
+ "default": "https://models.inference.ai.azure.com/chat/completions"
30
+ },
31
+ {
32
+ "name": "temperature",
33
+ "type": "number",
34
+ "label": "Temperature",
35
+ "description": "Controls randomness. Lower values are more focused, higher values more creative.",
36
+ "default": 0.7,
37
+ "min": 0,
38
+ "max": 2
39
+ },
40
+ {
41
+ "name": "max_tokens",
42
+ "type": "number",
43
+ "label": "Max Tokens",
44
+ "description": "Maximum number of tokens to generate in the response.",
45
+ "default": 4096,
46
+ "min": 1,
47
+ "max": 128000
48
+ },
49
+ {
50
+ "name": "top_p",
51
+ "type": "number",
52
+ "label": "Top P",
53
+ "description": "Nucleus sampling. Lower values make output more focused.",
54
+ "default": 1,
55
+ "min": 0,
56
+ "max": 1
57
+ },
58
+ {
59
+ "name": "frequency_penalty",
60
+ "type": "number",
61
+ "label": "Frequency Penalty",
62
+ "description": "Reduces repetition of frequent tokens.",
63
+ "default": 0,
64
+ "min": -2,
65
+ "max": 2
66
+ },
67
+ {
68
+ "name": "presence_penalty",
69
+ "type": "number",
70
+ "label": "Presence Penalty",
71
+ "description": "Encourages the model to talk about new topics.",
72
+ "default": 0,
73
+ "min": -2,
74
+ "max": 2
75
+ },
76
+ {
77
+ "name": "stream",
78
+ "type": "boolean",
79
+ "label": "Stream Response",
80
+ "description": "Stream the response token by token.",
81
+ "default": true
82
+ }
83
+ ]
23
84
  }
package/plugins/groq.json CHANGED
@@ -9,17 +9,78 @@
9
9
  "key_env": "GROQ_API_KEY"
10
10
  },
11
11
  "model_map": [
12
- "allam-2-7b",
13
- "groq/compound",
14
- "groq/compound-mini",
15
- "llama-3.1-8b-instant",
16
- "llama-3.3-70b-versatile",
17
- "meta-llama/llama-4-maverick-17b-128e-instruct",
18
- "meta-llama/llama-4-scout-17b-16e-instruct",
19
- "moonshotai/kimi-k2-instruct",
20
- "moonshotai/kimi-k2-instruct-0905",
21
- "openai/gpt-oss-120b",
22
- "openai/gpt-oss-20b",
23
- "qwen/qwen3-32b"
24
- ]
12
+ "allam-2-7b",
13
+ "groq/compound",
14
+ "groq/compound-mini",
15
+ "llama-3.1-8b-instant",
16
+ "llama-3.3-70b-versatile",
17
+ "meta-llama/llama-4-maverick-17b-128e-instruct",
18
+ "meta-llama/llama-4-scout-17b-16e-instruct",
19
+ "moonshotai/kimi-k2-instruct",
20
+ "moonshotai/kimi-k2-instruct-0905",
21
+ "openai/gpt-oss-120b",
22
+ "openai/gpt-oss-20b",
23
+ "qwen/qwen3-32b"
24
+ ],
25
+ "variables": [
26
+ {
27
+ "name": "endpoint",
28
+ "type": "string",
29
+ "label": "API Endpoint",
30
+ "description": "Override the default API endpoint URL.",
31
+ "default": "https://api.groq.com/openai/v1/chat/completions"
32
+ },
33
+ {
34
+ "name": "temperature",
35
+ "type": "number",
36
+ "label": "Temperature",
37
+ "description": "Controls randomness. Lower values are more focused, higher values more creative.",
38
+ "default": 0.7,
39
+ "min": 0,
40
+ "max": 2
41
+ },
42
+ {
43
+ "name": "max_tokens",
44
+ "type": "number",
45
+ "label": "Max Tokens",
46
+ "description": "Maximum number of tokens to generate in the response.",
47
+ "default": 4096,
48
+ "min": 1,
49
+ "max": 128000
50
+ },
51
+ {
52
+ "name": "top_p",
53
+ "type": "number",
54
+ "label": "Top P",
55
+ "description": "Nucleus sampling. Lower values make output more focused.",
56
+ "default": 1,
57
+ "min": 0,
58
+ "max": 1
59
+ },
60
+ {
61
+ "name": "frequency_penalty",
62
+ "type": "number",
63
+ "label": "Frequency Penalty",
64
+ "description": "Reduces repetition of frequent tokens.",
65
+ "default": 0,
66
+ "min": -2,
67
+ "max": 2
68
+ },
69
+ {
70
+ "name": "presence_penalty",
71
+ "type": "number",
72
+ "label": "Presence Penalty",
73
+ "description": "Encourages the model to talk about new topics.",
74
+ "default": 0,
75
+ "min": -2,
76
+ "max": 2
77
+ },
78
+ {
79
+ "name": "stream",
80
+ "type": "boolean",
81
+ "label": "Stream Response",
82
+ "description": "Stream the response token by token.",
83
+ "default": true
84
+ }
85
+ ]
25
86
  }
@@ -58,9 +58,15 @@
58
58
  "OuteAI/OuteTTS-0.3-1B"
59
59
  ],
60
60
  "config": {
61
- "voices": ["default"],
61
+ "voices": [
62
+ "default"
63
+ ],
62
64
  "default_voice": "default",
63
- "formats": ["flac", "wav", "mp3"],
65
+ "formats": [
66
+ "flac",
67
+ "wav",
68
+ "mp3"
69
+ ],
64
70
  "default_format": "flac",
65
71
  "max_characters": 5000,
66
72
  "supports_streaming": false
@@ -89,7 +95,10 @@
89
95
  "768x1024"
90
96
  ],
91
97
  "default_size": "1024x1024",
92
- "qualities": ["standard", "high"],
98
+ "qualities": [
99
+ "standard",
100
+ "high"
101
+ ],
93
102
  "default_quality": "standard",
94
103
  "max_prompt_length": 2000
95
104
  }
@@ -138,11 +147,89 @@
138
147
  "Systran/faster-whisper-large-v3"
139
148
  ],
140
149
  "config": {
141
- "formats": ["flac", "mp3", "wav", "webm", "ogg"],
150
+ "formats": [
151
+ "flac",
152
+ "mp3",
153
+ "wav",
154
+ "webm",
155
+ "ogg"
156
+ ],
142
157
  "max_duration_seconds": 300,
143
- "languages": ["en", "es", "fr", "de", "it", "pt", "nl", "ja", "ko", "zh"],
158
+ "languages": [
159
+ "en",
160
+ "es",
161
+ "fr",
162
+ "de",
163
+ "it",
164
+ "pt",
165
+ "nl",
166
+ "ja",
167
+ "ko",
168
+ "zh"
169
+ ],
144
170
  "supports_timestamps": true
145
171
  }
146
172
  }
147
- }
173
+ },
174
+ "variables": [
175
+ {
176
+ "name": "endpoint",
177
+ "type": "string",
178
+ "label": "API Endpoint",
179
+ "description": "Override the default API endpoint URL.",
180
+ "default": "https://router.huggingface.co/v1/chat/completions"
181
+ },
182
+ {
183
+ "name": "temperature",
184
+ "type": "number",
185
+ "label": "Temperature",
186
+ "description": "Controls randomness. Lower values are more focused, higher values more creative.",
187
+ "default": 0.7,
188
+ "min": 0,
189
+ "max": 2
190
+ },
191
+ {
192
+ "name": "max_tokens",
193
+ "type": "number",
194
+ "label": "Max Tokens",
195
+ "description": "Maximum number of tokens to generate in the response.",
196
+ "default": 4096,
197
+ "min": 1,
198
+ "max": 128000
199
+ },
200
+ {
201
+ "name": "top_p",
202
+ "type": "number",
203
+ "label": "Top P",
204
+ "description": "Nucleus sampling. Lower values make output more focused.",
205
+ "default": 1,
206
+ "min": 0,
207
+ "max": 1
208
+ },
209
+ {
210
+ "name": "frequency_penalty",
211
+ "type": "number",
212
+ "label": "Frequency Penalty",
213
+ "description": "Reduces repetition of frequent tokens.",
214
+ "default": 0,
215
+ "min": -2,
216
+ "max": 2
217
+ },
218
+ {
219
+ "name": "presence_penalty",
220
+ "type": "number",
221
+ "label": "Presence Penalty",
222
+ "description": "Encourages the model to talk about new topics.",
223
+ "default": 0,
224
+ "min": -2,
225
+ "max": 2
226
+ },
227
+ {
228
+ "name": "stream",
229
+ "type": "boolean",
230
+ "label": "Stream Response",
231
+ "description": "Stream the response token by token.",
232
+ "default": true
233
+ }
234
+ ]
148
235
  }
@@ -2,7 +2,7 @@
2
2
  "id": "kyutai-tts-1.6b",
3
3
  "name": "Kyutai TTS 1.6B",
4
4
  "type": "tts",
5
- "endpoint": "http://192.168.2.112:8201/v1/audio/speech",
5
+ "endpoint": "http://localhost:8201/v1/audio/speech",
6
6
  "auth": {
7
7
  "header": "",
8
8
  "key_env": ""
@@ -13,26 +13,70 @@
13
13
  ],
14
14
  "capabilities": {
15
15
  "tts": {
16
- "endpoint": "http://192.168.2.112:8201/v1/audio/speech",
16
+ "endpoint": "http://localhost:8201/v1/audio/speech",
17
17
  "model_map": [
18
18
  "kyutai-tts-1.6b",
19
19
  "tts-1-hd"
20
20
  ],
21
21
  "config": {
22
22
  "voices": [
23
- "alba", "alba-casual", "alba-merchant", "alba-announcer", "alba-moment",
24
- "angry-f", "calm-f", "confused-f", "desire-f", "fear-f", "happy-f", "sad-f", "neutral-f",
25
- "angry-m", "calm-m", "confused-m", "desire-m", "fear-m", "happy-m", "sad-m", "neutral-m",
26
- "expresso-happy", "expresso-sad", "expresso-angry", "expresso-calm", "expresso-confused",
27
- "expresso-sarcastic", "expresso-sleepy", "expresso-laughing", "expresso-awe",
28
- "expresso-desire", "expresso-whisper", "expresso-fast", "expresso-narration",
29
- "cml-12977-fr-f", "cml-1406-fr-m", "cml-2154-fr-f", "cml-4724-fr-m",
30
- "vctk-p225", "vctk-p226", "vctk-p228", "vctk-p231", "vctk-p255", "vctk-p277", "vctk-p292",
31
- "unmute-degaulle", "unmute-developer", "unmute-developpeuse", "unmute-narration",
32
- "unmute-default", "unmute-fabien", "voice-donation-dwp"
23
+ "alba",
24
+ "alba-casual",
25
+ "alba-merchant",
26
+ "alba-announcer",
27
+ "alba-moment",
28
+ "angry-f",
29
+ "calm-f",
30
+ "confused-f",
31
+ "desire-f",
32
+ "fear-f",
33
+ "happy-f",
34
+ "sad-f",
35
+ "neutral-f",
36
+ "angry-m",
37
+ "calm-m",
38
+ "confused-m",
39
+ "desire-m",
40
+ "fear-m",
41
+ "happy-m",
42
+ "sad-m",
43
+ "neutral-m",
44
+ "expresso-happy",
45
+ "expresso-sad",
46
+ "expresso-angry",
47
+ "expresso-calm",
48
+ "expresso-confused",
49
+ "expresso-sarcastic",
50
+ "expresso-sleepy",
51
+ "expresso-laughing",
52
+ "expresso-awe",
53
+ "expresso-desire",
54
+ "expresso-whisper",
55
+ "expresso-fast",
56
+ "expresso-narration",
57
+ "cml-12977-fr-f",
58
+ "cml-1406-fr-m",
59
+ "cml-2154-fr-f",
60
+ "cml-4724-fr-m",
61
+ "vctk-p225",
62
+ "vctk-p226",
63
+ "vctk-p228",
64
+ "vctk-p231",
65
+ "vctk-p255",
66
+ "vctk-p277",
67
+ "vctk-p292",
68
+ "unmute-degaulle",
69
+ "unmute-developer",
70
+ "unmute-developpeuse",
71
+ "unmute-narration",
72
+ "unmute-default",
73
+ "unmute-fabien",
74
+ "voice-donation-dwp"
33
75
  ],
34
76
  "default_voice": "alba",
35
- "formats": ["wav"],
77
+ "formats": [
78
+ "wav"
79
+ ],
36
80
  "default_format": "wav",
37
81
  "max_characters": 10000,
38
82
  "supports_streaming": true,
@@ -41,5 +85,23 @@
41
85
  }
42
86
  },
43
87
  "description": "Kyutai TTS 1.6B GPU-accelerated server (CUDA, MPS, CPU)",
44
- "documentation_url": "https://github.com/kyutai-labs/delayed-streams-modeling"
88
+ "documentation_url": "https://github.com/kyutai-labs/delayed-streams-modeling",
89
+ "variables": [
90
+ {
91
+ "name": "endpoint",
92
+ "type": "string",
93
+ "label": "API Endpoint",
94
+ "description": "Override the default API endpoint URL.",
95
+ "default": "http://localhost:8201/v1/audio/speech"
96
+ },
97
+ {
98
+ "name": "speed",
99
+ "type": "number",
100
+ "label": "Speed",
101
+ "description": "Playback speed multiplier.",
102
+ "default": 1,
103
+ "min": 0.25,
104
+ "max": 4
105
+ }
106
+ ]
45
107
  }
@@ -21,9 +21,20 @@
21
21
  "tts-1"
22
22
  ],
23
23
  "config": {
24
- "voices": ["Alba", "Marius", "Javert", "Jean", "Fantine", "Cosette", "Eponine", "Azelma"],
24
+ "voices": [
25
+ "Alba",
26
+ "Marius",
27
+ "Javert",
28
+ "Jean",
29
+ "Fantine",
30
+ "Cosette",
31
+ "Eponine",
32
+ "Azelma"
33
+ ],
25
34
  "default_voice": "Alba",
26
- "formats": ["wav"],
35
+ "formats": [
36
+ "wav"
37
+ ],
27
38
  "default_format": "wav",
28
39
  "max_characters": 10000,
29
40
  "supports_streaming": true,
@@ -32,5 +43,23 @@
32
43
  }
33
44
  },
34
45
  "description": "Kyutai Pocket TTS local server (CPU-based, voice cloning support)",
35
- "documentation_url": "https://github.com/kyutai-labs/pocket-tts"
46
+ "documentation_url": "https://github.com/kyutai-labs/pocket-tts",
47
+ "variables": [
48
+ {
49
+ "name": "endpoint",
50
+ "type": "string",
51
+ "label": "API Endpoint",
52
+ "description": "Override the default API endpoint URL.",
53
+ "default": "http://localhost:8200/v1/audio/speech"
54
+ },
55
+ {
56
+ "name": "speed",
57
+ "type": "number",
58
+ "label": "Speed",
59
+ "description": "Playback speed multiplier.",
60
+ "default": 1,
61
+ "min": 0.25,
62
+ "max": 4
63
+ }
64
+ ]
36
65
  }