libre-webui 0.6.2 → 0.7.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/backend/dist/db.d.ts.map +1 -1
- package/backend/dist/db.js +17 -0
- package/backend/dist/db.js.map +1 -1
- package/backend/dist/routes/auth.js +2 -2
- package/backend/dist/routes/auth.js.map +1 -1
- package/backend/dist/routes/plugins.d.ts.map +1 -1
- package/backend/dist/routes/plugins.js +162 -0
- package/backend/dist/routes/plugins.js.map +1 -1
- package/backend/dist/services/authService.d.ts +1 -1
- package/backend/dist/services/authService.d.ts.map +1 -1
- package/backend/dist/services/authService.js +1 -1
- package/backend/dist/services/authService.js.map +1 -1
- package/backend/dist/services/pluginService.d.ts +9 -0
- package/backend/dist/services/pluginService.d.ts.map +1 -1
- package/backend/dist/services/pluginService.js +99 -24
- package/backend/dist/services/pluginService.js.map +1 -1
- package/backend/dist/services/pluginVariablesService.d.ts +36 -0
- package/backend/dist/services/pluginVariablesService.d.ts.map +1 -0
- package/backend/dist/services/pluginVariablesService.js +218 -0
- package/backend/dist/services/pluginVariablesService.js.map +1 -0
- package/backend/dist/types/index.d.ts +14 -0
- package/backend/dist/types/index.d.ts.map +1 -1
- package/backend/dist/types/index.js.map +1 -1
- package/frontend/dist/assets/index-DMlahgv7.js +51 -0
- package/frontend/dist/css/{index-Dh5G0uFZ.css → index-D9uKnbHR.css} +1 -1
- package/frontend/dist/index.html +6 -6
- package/frontend/dist/js/{ArtifactContainer-BLHcg4YG.js → ArtifactContainer-D_igWPrR.js} +2 -2
- package/frontend/dist/js/{ArtifactDemoPage-Br-19dKk.js → ArtifactDemoPage-BW64ikxs.js} +1 -1
- package/frontend/dist/js/{ChatPage-CQGVCtT_.js → ChatPage-CIUGXjJ_.js} +10 -10
- package/frontend/dist/js/{GalleryPage-CvA1Sox3.js → GalleryPage-CMqUQtns.js} +1 -1
- package/frontend/dist/js/ModelsPage-DQ2ud-YG.js +1 -0
- package/frontend/dist/js/PersonasPage-HqcbbtTL.js +13 -0
- package/frontend/dist/js/{UserManagementPage-H89n1w08.js → UserManagementPage-C4vtOCZ6.js} +1 -1
- package/frontend/dist/js/{markdown-vendor-DeGQOJEA.js → markdown-vendor-QwEm2Xso.js} +1 -1
- package/frontend/dist/js/{router-vendor-CYqbjR9t.js → router-vendor-CmbUVdlE.js} +5 -5
- package/frontend/dist/js/ui-vendor-DxZsuKzb.js +177 -0
- package/frontend/dist/js/utils-vendor-0Ymd3GFQ.js +6 -0
- package/package.json +1 -1
- package/plugins/anthropic.json +74 -13
- package/plugins/comfyui.json +45 -4
- package/plugins/elevenlabs.json +42 -2
- package/plugins/gemini.json +106 -45
- package/plugins/github.json +72 -11
- package/plugins/groq.json +74 -13
- package/plugins/huggingface.json +93 -6
- package/plugins/kyutai-tts-1.6b.json +76 -14
- package/plugins/kyutai-tts.json +32 -3
- package/plugins/mistral.json +122 -61
- package/plugins/openai-tts.json +38 -3
- package/plugins/openai.json +81 -3
- package/plugins/openrouter.json +402 -341
- package/plugins/qwen-tts.json +33 -3
- package/frontend/dist/assets/index-DE-xmTDC.js +0 -51
- package/frontend/dist/js/ModelsPage-9djD6Uj9.js +0 -1
- package/frontend/dist/js/PersonasPage-ajydrtrS.js +0 -13
- package/frontend/dist/js/ui-vendor-DnQP3_cJ.js +0 -177
- package/frontend/dist/js/utils-vendor-COrKO_vV.js +0 -6
package/plugins/elevenlabs.json
CHANGED
|
@@ -48,11 +48,51 @@
|
|
|
48
48
|
"Serena"
|
|
49
49
|
],
|
|
50
50
|
"default_voice": "Rachel",
|
|
51
|
-
"formats": [
|
|
51
|
+
"formats": [
|
|
52
|
+
"mp3",
|
|
53
|
+
"pcm",
|
|
54
|
+
"ulaw"
|
|
55
|
+
],
|
|
52
56
|
"default_format": "mp3",
|
|
53
57
|
"max_characters": 5000,
|
|
54
58
|
"supports_streaming": true
|
|
55
59
|
}
|
|
56
60
|
}
|
|
57
|
-
}
|
|
61
|
+
},
|
|
62
|
+
"variables": [
|
|
63
|
+
{
|
|
64
|
+
"name": "endpoint",
|
|
65
|
+
"type": "string",
|
|
66
|
+
"label": "API Endpoint",
|
|
67
|
+
"description": "Override the default API endpoint URL.",
|
|
68
|
+
"default": "https://api.elevenlabs.io/v1/text-to-speech/{voice_id}"
|
|
69
|
+
},
|
|
70
|
+
{
|
|
71
|
+
"name": "stability",
|
|
72
|
+
"type": "number",
|
|
73
|
+
"label": "Stability",
|
|
74
|
+
"description": "Voice stability (0-1). Higher values are more consistent.",
|
|
75
|
+
"default": 0.5,
|
|
76
|
+
"min": 0,
|
|
77
|
+
"max": 1
|
|
78
|
+
},
|
|
79
|
+
{
|
|
80
|
+
"name": "similarity_boost",
|
|
81
|
+
"type": "number",
|
|
82
|
+
"label": "Similarity Boost",
|
|
83
|
+
"description": "Voice similarity boost (0-1).",
|
|
84
|
+
"default": 0.75,
|
|
85
|
+
"min": 0,
|
|
86
|
+
"max": 1
|
|
87
|
+
},
|
|
88
|
+
{
|
|
89
|
+
"name": "speed",
|
|
90
|
+
"type": "number",
|
|
91
|
+
"label": "Speed",
|
|
92
|
+
"description": "Playback speed multiplier.",
|
|
93
|
+
"default": 1,
|
|
94
|
+
"min": 0.25,
|
|
95
|
+
"max": 4
|
|
96
|
+
}
|
|
97
|
+
]
|
|
58
98
|
}
|
package/plugins/gemini.json
CHANGED
|
@@ -9,49 +9,110 @@
|
|
|
9
9
|
"key_env": "GEMINI_API_KEY"
|
|
10
10
|
},
|
|
11
11
|
"model_map": [
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
]
|
|
12
|
+
"aqa",
|
|
13
|
+
"deep-research-pro-preview-12-2025",
|
|
14
|
+
"gemini-2.0-flash",
|
|
15
|
+
"gemini-2.0-flash-001",
|
|
16
|
+
"gemini-2.0-flash-exp",
|
|
17
|
+
"gemini-2.0-flash-exp-image-generation",
|
|
18
|
+
"gemini-2.0-flash-lite",
|
|
19
|
+
"gemini-2.0-flash-lite-001",
|
|
20
|
+
"gemini-2.0-flash-lite-preview",
|
|
21
|
+
"gemini-2.0-flash-lite-preview-02-05",
|
|
22
|
+
"gemini-2.5-computer-use-preview-10-2025",
|
|
23
|
+
"gemini-2.5-flash",
|
|
24
|
+
"gemini-2.5-flash-image",
|
|
25
|
+
"gemini-2.5-flash-image-preview",
|
|
26
|
+
"gemini-2.5-flash-lite",
|
|
27
|
+
"gemini-2.5-flash-lite-preview-09-2025",
|
|
28
|
+
"gemini-2.5-flash-preview-09-2025",
|
|
29
|
+
"gemini-2.5-flash-preview-tts",
|
|
30
|
+
"gemini-2.5-pro",
|
|
31
|
+
"gemini-2.5-pro-preview-tts",
|
|
32
|
+
"gemini-3-pro-image-preview",
|
|
33
|
+
"gemini-3-pro-preview",
|
|
34
|
+
"gemini-exp-1206",
|
|
35
|
+
"gemini-flash-latest",
|
|
36
|
+
"gemini-flash-lite-latest",
|
|
37
|
+
"gemini-pro-latest",
|
|
38
|
+
"gemini-robotics-er-1.5-preview",
|
|
39
|
+
"gemma-3-12b-it",
|
|
40
|
+
"gemma-3-1b-it",
|
|
41
|
+
"gemma-3-27b-it",
|
|
42
|
+
"gemma-3-4b-it",
|
|
43
|
+
"gemma-3n-e2b-it",
|
|
44
|
+
"gemma-3n-e4b-it",
|
|
45
|
+
"imagen-4.0-fast-generate-001",
|
|
46
|
+
"imagen-4.0-generate-001",
|
|
47
|
+
"imagen-4.0-generate-preview-06-06",
|
|
48
|
+
"imagen-4.0-ultra-generate-001",
|
|
49
|
+
"imagen-4.0-ultra-generate-preview-06-06",
|
|
50
|
+
"nano-banana-pro-preview",
|
|
51
|
+
"veo-2.0-generate-001",
|
|
52
|
+
"veo-3.0-fast-generate-001",
|
|
53
|
+
"veo-3.0-generate-001",
|
|
54
|
+
"veo-3.1-fast-generate-preview",
|
|
55
|
+
"veo-3.1-generate-preview"
|
|
56
|
+
],
|
|
57
|
+
"variables": [
|
|
58
|
+
{
|
|
59
|
+
"name": "endpoint",
|
|
60
|
+
"type": "string",
|
|
61
|
+
"label": "API Endpoint",
|
|
62
|
+
"description": "Override the default API endpoint URL.",
|
|
63
|
+
"default": "https://generativelanguage.googleapis.com/v1beta/models/{model}:generateContent"
|
|
64
|
+
},
|
|
65
|
+
{
|
|
66
|
+
"name": "temperature",
|
|
67
|
+
"type": "number",
|
|
68
|
+
"label": "Temperature",
|
|
69
|
+
"description": "Controls randomness. Lower values are more focused, higher values more creative.",
|
|
70
|
+
"default": 0.7,
|
|
71
|
+
"min": 0,
|
|
72
|
+
"max": 2
|
|
73
|
+
},
|
|
74
|
+
{
|
|
75
|
+
"name": "max_tokens",
|
|
76
|
+
"type": "number",
|
|
77
|
+
"label": "Max Tokens",
|
|
78
|
+
"description": "Maximum number of tokens to generate in the response.",
|
|
79
|
+
"default": 4096,
|
|
80
|
+
"min": 1,
|
|
81
|
+
"max": 128000
|
|
82
|
+
},
|
|
83
|
+
{
|
|
84
|
+
"name": "top_p",
|
|
85
|
+
"type": "number",
|
|
86
|
+
"label": "Top P",
|
|
87
|
+
"description": "Nucleus sampling. Lower values make output more focused.",
|
|
88
|
+
"default": 1,
|
|
89
|
+
"min": 0,
|
|
90
|
+
"max": 1
|
|
91
|
+
},
|
|
92
|
+
{
|
|
93
|
+
"name": "frequency_penalty",
|
|
94
|
+
"type": "number",
|
|
95
|
+
"label": "Frequency Penalty",
|
|
96
|
+
"description": "Reduces repetition of frequent tokens.",
|
|
97
|
+
"default": 0,
|
|
98
|
+
"min": -2,
|
|
99
|
+
"max": 2
|
|
100
|
+
},
|
|
101
|
+
{
|
|
102
|
+
"name": "presence_penalty",
|
|
103
|
+
"type": "number",
|
|
104
|
+
"label": "Presence Penalty",
|
|
105
|
+
"description": "Encourages the model to talk about new topics.",
|
|
106
|
+
"default": 0,
|
|
107
|
+
"min": -2,
|
|
108
|
+
"max": 2
|
|
109
|
+
},
|
|
110
|
+
{
|
|
111
|
+
"name": "stream",
|
|
112
|
+
"type": "boolean",
|
|
113
|
+
"label": "Stream Response",
|
|
114
|
+
"description": "Stream the response token by token.",
|
|
115
|
+
"default": true
|
|
116
|
+
}
|
|
117
|
+
]
|
|
57
118
|
}
|
package/plugins/github.json
CHANGED
|
@@ -9,15 +9,76 @@
|
|
|
9
9
|
"key_env": "GITHUB_API_KEY"
|
|
10
10
|
},
|
|
11
11
|
"model_map": [
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
]
|
|
12
|
+
"AI21-Jamba-Instruct",
|
|
13
|
+
"Meta-Llama-3-70B-Instruct",
|
|
14
|
+
"Meta-Llama-3-8B-Instruct",
|
|
15
|
+
"Meta-Llama-3.1-405B-Instruct",
|
|
16
|
+
"Meta-Llama-3.1-70B-Instruct",
|
|
17
|
+
"Meta-Llama-3.1-8B-Instruct",
|
|
18
|
+
"Mistral-Nemo",
|
|
19
|
+
"Mistral-large-2407",
|
|
20
|
+
"gpt-4o",
|
|
21
|
+
"gpt-4o-mini"
|
|
22
|
+
],
|
|
23
|
+
"variables": [
|
|
24
|
+
{
|
|
25
|
+
"name": "endpoint",
|
|
26
|
+
"type": "string",
|
|
27
|
+
"label": "API Endpoint",
|
|
28
|
+
"description": "Override the default API endpoint URL.",
|
|
29
|
+
"default": "https://models.inference.ai.azure.com/chat/completions"
|
|
30
|
+
},
|
|
31
|
+
{
|
|
32
|
+
"name": "temperature",
|
|
33
|
+
"type": "number",
|
|
34
|
+
"label": "Temperature",
|
|
35
|
+
"description": "Controls randomness. Lower values are more focused, higher values more creative.",
|
|
36
|
+
"default": 0.7,
|
|
37
|
+
"min": 0,
|
|
38
|
+
"max": 2
|
|
39
|
+
},
|
|
40
|
+
{
|
|
41
|
+
"name": "max_tokens",
|
|
42
|
+
"type": "number",
|
|
43
|
+
"label": "Max Tokens",
|
|
44
|
+
"description": "Maximum number of tokens to generate in the response.",
|
|
45
|
+
"default": 4096,
|
|
46
|
+
"min": 1,
|
|
47
|
+
"max": 128000
|
|
48
|
+
},
|
|
49
|
+
{
|
|
50
|
+
"name": "top_p",
|
|
51
|
+
"type": "number",
|
|
52
|
+
"label": "Top P",
|
|
53
|
+
"description": "Nucleus sampling. Lower values make output more focused.",
|
|
54
|
+
"default": 1,
|
|
55
|
+
"min": 0,
|
|
56
|
+
"max": 1
|
|
57
|
+
},
|
|
58
|
+
{
|
|
59
|
+
"name": "frequency_penalty",
|
|
60
|
+
"type": "number",
|
|
61
|
+
"label": "Frequency Penalty",
|
|
62
|
+
"description": "Reduces repetition of frequent tokens.",
|
|
63
|
+
"default": 0,
|
|
64
|
+
"min": -2,
|
|
65
|
+
"max": 2
|
|
66
|
+
},
|
|
67
|
+
{
|
|
68
|
+
"name": "presence_penalty",
|
|
69
|
+
"type": "number",
|
|
70
|
+
"label": "Presence Penalty",
|
|
71
|
+
"description": "Encourages the model to talk about new topics.",
|
|
72
|
+
"default": 0,
|
|
73
|
+
"min": -2,
|
|
74
|
+
"max": 2
|
|
75
|
+
},
|
|
76
|
+
{
|
|
77
|
+
"name": "stream",
|
|
78
|
+
"type": "boolean",
|
|
79
|
+
"label": "Stream Response",
|
|
80
|
+
"description": "Stream the response token by token.",
|
|
81
|
+
"default": true
|
|
82
|
+
}
|
|
83
|
+
]
|
|
23
84
|
}
|
package/plugins/groq.json
CHANGED
|
@@ -9,17 +9,78 @@
|
|
|
9
9
|
"key_env": "GROQ_API_KEY"
|
|
10
10
|
},
|
|
11
11
|
"model_map": [
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
]
|
|
12
|
+
"allam-2-7b",
|
|
13
|
+
"groq/compound",
|
|
14
|
+
"groq/compound-mini",
|
|
15
|
+
"llama-3.1-8b-instant",
|
|
16
|
+
"llama-3.3-70b-versatile",
|
|
17
|
+
"meta-llama/llama-4-maverick-17b-128e-instruct",
|
|
18
|
+
"meta-llama/llama-4-scout-17b-16e-instruct",
|
|
19
|
+
"moonshotai/kimi-k2-instruct",
|
|
20
|
+
"moonshotai/kimi-k2-instruct-0905",
|
|
21
|
+
"openai/gpt-oss-120b",
|
|
22
|
+
"openai/gpt-oss-20b",
|
|
23
|
+
"qwen/qwen3-32b"
|
|
24
|
+
],
|
|
25
|
+
"variables": [
|
|
26
|
+
{
|
|
27
|
+
"name": "endpoint",
|
|
28
|
+
"type": "string",
|
|
29
|
+
"label": "API Endpoint",
|
|
30
|
+
"description": "Override the default API endpoint URL.",
|
|
31
|
+
"default": "https://api.groq.com/openai/v1/chat/completions"
|
|
32
|
+
},
|
|
33
|
+
{
|
|
34
|
+
"name": "temperature",
|
|
35
|
+
"type": "number",
|
|
36
|
+
"label": "Temperature",
|
|
37
|
+
"description": "Controls randomness. Lower values are more focused, higher values more creative.",
|
|
38
|
+
"default": 0.7,
|
|
39
|
+
"min": 0,
|
|
40
|
+
"max": 2
|
|
41
|
+
},
|
|
42
|
+
{
|
|
43
|
+
"name": "max_tokens",
|
|
44
|
+
"type": "number",
|
|
45
|
+
"label": "Max Tokens",
|
|
46
|
+
"description": "Maximum number of tokens to generate in the response.",
|
|
47
|
+
"default": 4096,
|
|
48
|
+
"min": 1,
|
|
49
|
+
"max": 128000
|
|
50
|
+
},
|
|
51
|
+
{
|
|
52
|
+
"name": "top_p",
|
|
53
|
+
"type": "number",
|
|
54
|
+
"label": "Top P",
|
|
55
|
+
"description": "Nucleus sampling. Lower values make output more focused.",
|
|
56
|
+
"default": 1,
|
|
57
|
+
"min": 0,
|
|
58
|
+
"max": 1
|
|
59
|
+
},
|
|
60
|
+
{
|
|
61
|
+
"name": "frequency_penalty",
|
|
62
|
+
"type": "number",
|
|
63
|
+
"label": "Frequency Penalty",
|
|
64
|
+
"description": "Reduces repetition of frequent tokens.",
|
|
65
|
+
"default": 0,
|
|
66
|
+
"min": -2,
|
|
67
|
+
"max": 2
|
|
68
|
+
},
|
|
69
|
+
{
|
|
70
|
+
"name": "presence_penalty",
|
|
71
|
+
"type": "number",
|
|
72
|
+
"label": "Presence Penalty",
|
|
73
|
+
"description": "Encourages the model to talk about new topics.",
|
|
74
|
+
"default": 0,
|
|
75
|
+
"min": -2,
|
|
76
|
+
"max": 2
|
|
77
|
+
},
|
|
78
|
+
{
|
|
79
|
+
"name": "stream",
|
|
80
|
+
"type": "boolean",
|
|
81
|
+
"label": "Stream Response",
|
|
82
|
+
"description": "Stream the response token by token.",
|
|
83
|
+
"default": true
|
|
84
|
+
}
|
|
85
|
+
]
|
|
25
86
|
}
|
package/plugins/huggingface.json
CHANGED
|
@@ -58,9 +58,15 @@
|
|
|
58
58
|
"OuteAI/OuteTTS-0.3-1B"
|
|
59
59
|
],
|
|
60
60
|
"config": {
|
|
61
|
-
"voices": [
|
|
61
|
+
"voices": [
|
|
62
|
+
"default"
|
|
63
|
+
],
|
|
62
64
|
"default_voice": "default",
|
|
63
|
-
"formats": [
|
|
65
|
+
"formats": [
|
|
66
|
+
"flac",
|
|
67
|
+
"wav",
|
|
68
|
+
"mp3"
|
|
69
|
+
],
|
|
64
70
|
"default_format": "flac",
|
|
65
71
|
"max_characters": 5000,
|
|
66
72
|
"supports_streaming": false
|
|
@@ -89,7 +95,10 @@
|
|
|
89
95
|
"768x1024"
|
|
90
96
|
],
|
|
91
97
|
"default_size": "1024x1024",
|
|
92
|
-
"qualities": [
|
|
98
|
+
"qualities": [
|
|
99
|
+
"standard",
|
|
100
|
+
"high"
|
|
101
|
+
],
|
|
93
102
|
"default_quality": "standard",
|
|
94
103
|
"max_prompt_length": 2000
|
|
95
104
|
}
|
|
@@ -138,11 +147,89 @@
|
|
|
138
147
|
"Systran/faster-whisper-large-v3"
|
|
139
148
|
],
|
|
140
149
|
"config": {
|
|
141
|
-
"formats": [
|
|
150
|
+
"formats": [
|
|
151
|
+
"flac",
|
|
152
|
+
"mp3",
|
|
153
|
+
"wav",
|
|
154
|
+
"webm",
|
|
155
|
+
"ogg"
|
|
156
|
+
],
|
|
142
157
|
"max_duration_seconds": 300,
|
|
143
|
-
"languages": [
|
|
158
|
+
"languages": [
|
|
159
|
+
"en",
|
|
160
|
+
"es",
|
|
161
|
+
"fr",
|
|
162
|
+
"de",
|
|
163
|
+
"it",
|
|
164
|
+
"pt",
|
|
165
|
+
"nl",
|
|
166
|
+
"ja",
|
|
167
|
+
"ko",
|
|
168
|
+
"zh"
|
|
169
|
+
],
|
|
144
170
|
"supports_timestamps": true
|
|
145
171
|
}
|
|
146
172
|
}
|
|
147
|
-
}
|
|
173
|
+
},
|
|
174
|
+
"variables": [
|
|
175
|
+
{
|
|
176
|
+
"name": "endpoint",
|
|
177
|
+
"type": "string",
|
|
178
|
+
"label": "API Endpoint",
|
|
179
|
+
"description": "Override the default API endpoint URL.",
|
|
180
|
+
"default": "https://router.huggingface.co/v1/chat/completions"
|
|
181
|
+
},
|
|
182
|
+
{
|
|
183
|
+
"name": "temperature",
|
|
184
|
+
"type": "number",
|
|
185
|
+
"label": "Temperature",
|
|
186
|
+
"description": "Controls randomness. Lower values are more focused, higher values more creative.",
|
|
187
|
+
"default": 0.7,
|
|
188
|
+
"min": 0,
|
|
189
|
+
"max": 2
|
|
190
|
+
},
|
|
191
|
+
{
|
|
192
|
+
"name": "max_tokens",
|
|
193
|
+
"type": "number",
|
|
194
|
+
"label": "Max Tokens",
|
|
195
|
+
"description": "Maximum number of tokens to generate in the response.",
|
|
196
|
+
"default": 4096,
|
|
197
|
+
"min": 1,
|
|
198
|
+
"max": 128000
|
|
199
|
+
},
|
|
200
|
+
{
|
|
201
|
+
"name": "top_p",
|
|
202
|
+
"type": "number",
|
|
203
|
+
"label": "Top P",
|
|
204
|
+
"description": "Nucleus sampling. Lower values make output more focused.",
|
|
205
|
+
"default": 1,
|
|
206
|
+
"min": 0,
|
|
207
|
+
"max": 1
|
|
208
|
+
},
|
|
209
|
+
{
|
|
210
|
+
"name": "frequency_penalty",
|
|
211
|
+
"type": "number",
|
|
212
|
+
"label": "Frequency Penalty",
|
|
213
|
+
"description": "Reduces repetition of frequent tokens.",
|
|
214
|
+
"default": 0,
|
|
215
|
+
"min": -2,
|
|
216
|
+
"max": 2
|
|
217
|
+
},
|
|
218
|
+
{
|
|
219
|
+
"name": "presence_penalty",
|
|
220
|
+
"type": "number",
|
|
221
|
+
"label": "Presence Penalty",
|
|
222
|
+
"description": "Encourages the model to talk about new topics.",
|
|
223
|
+
"default": 0,
|
|
224
|
+
"min": -2,
|
|
225
|
+
"max": 2
|
|
226
|
+
},
|
|
227
|
+
{
|
|
228
|
+
"name": "stream",
|
|
229
|
+
"type": "boolean",
|
|
230
|
+
"label": "Stream Response",
|
|
231
|
+
"description": "Stream the response token by token.",
|
|
232
|
+
"default": true
|
|
233
|
+
}
|
|
234
|
+
]
|
|
148
235
|
}
|
|
@@ -2,7 +2,7 @@
|
|
|
2
2
|
"id": "kyutai-tts-1.6b",
|
|
3
3
|
"name": "Kyutai TTS 1.6B",
|
|
4
4
|
"type": "tts",
|
|
5
|
-
"endpoint": "http://
|
|
5
|
+
"endpoint": "http://localhost:8201/v1/audio/speech",
|
|
6
6
|
"auth": {
|
|
7
7
|
"header": "",
|
|
8
8
|
"key_env": ""
|
|
@@ -13,26 +13,70 @@
|
|
|
13
13
|
],
|
|
14
14
|
"capabilities": {
|
|
15
15
|
"tts": {
|
|
16
|
-
"endpoint": "http://
|
|
16
|
+
"endpoint": "http://localhost:8201/v1/audio/speech",
|
|
17
17
|
"model_map": [
|
|
18
18
|
"kyutai-tts-1.6b",
|
|
19
19
|
"tts-1-hd"
|
|
20
20
|
],
|
|
21
21
|
"config": {
|
|
22
22
|
"voices": [
|
|
23
|
-
"alba",
|
|
24
|
-
"
|
|
25
|
-
"
|
|
26
|
-
"
|
|
27
|
-
"
|
|
28
|
-
"
|
|
29
|
-
"
|
|
30
|
-
"
|
|
31
|
-
"
|
|
32
|
-
"
|
|
23
|
+
"alba",
|
|
24
|
+
"alba-casual",
|
|
25
|
+
"alba-merchant",
|
|
26
|
+
"alba-announcer",
|
|
27
|
+
"alba-moment",
|
|
28
|
+
"angry-f",
|
|
29
|
+
"calm-f",
|
|
30
|
+
"confused-f",
|
|
31
|
+
"desire-f",
|
|
32
|
+
"fear-f",
|
|
33
|
+
"happy-f",
|
|
34
|
+
"sad-f",
|
|
35
|
+
"neutral-f",
|
|
36
|
+
"angry-m",
|
|
37
|
+
"calm-m",
|
|
38
|
+
"confused-m",
|
|
39
|
+
"desire-m",
|
|
40
|
+
"fear-m",
|
|
41
|
+
"happy-m",
|
|
42
|
+
"sad-m",
|
|
43
|
+
"neutral-m",
|
|
44
|
+
"expresso-happy",
|
|
45
|
+
"expresso-sad",
|
|
46
|
+
"expresso-angry",
|
|
47
|
+
"expresso-calm",
|
|
48
|
+
"expresso-confused",
|
|
49
|
+
"expresso-sarcastic",
|
|
50
|
+
"expresso-sleepy",
|
|
51
|
+
"expresso-laughing",
|
|
52
|
+
"expresso-awe",
|
|
53
|
+
"expresso-desire",
|
|
54
|
+
"expresso-whisper",
|
|
55
|
+
"expresso-fast",
|
|
56
|
+
"expresso-narration",
|
|
57
|
+
"cml-12977-fr-f",
|
|
58
|
+
"cml-1406-fr-m",
|
|
59
|
+
"cml-2154-fr-f",
|
|
60
|
+
"cml-4724-fr-m",
|
|
61
|
+
"vctk-p225",
|
|
62
|
+
"vctk-p226",
|
|
63
|
+
"vctk-p228",
|
|
64
|
+
"vctk-p231",
|
|
65
|
+
"vctk-p255",
|
|
66
|
+
"vctk-p277",
|
|
67
|
+
"vctk-p292",
|
|
68
|
+
"unmute-degaulle",
|
|
69
|
+
"unmute-developer",
|
|
70
|
+
"unmute-developpeuse",
|
|
71
|
+
"unmute-narration",
|
|
72
|
+
"unmute-default",
|
|
73
|
+
"unmute-fabien",
|
|
74
|
+
"voice-donation-dwp"
|
|
33
75
|
],
|
|
34
76
|
"default_voice": "alba",
|
|
35
|
-
"formats": [
|
|
77
|
+
"formats": [
|
|
78
|
+
"wav"
|
|
79
|
+
],
|
|
36
80
|
"default_format": "wav",
|
|
37
81
|
"max_characters": 10000,
|
|
38
82
|
"supports_streaming": true,
|
|
@@ -41,5 +85,23 @@
|
|
|
41
85
|
}
|
|
42
86
|
},
|
|
43
87
|
"description": "Kyutai TTS 1.6B GPU-accelerated server (CUDA, MPS, CPU)",
|
|
44
|
-
"documentation_url": "https://github.com/kyutai-labs/delayed-streams-modeling"
|
|
88
|
+
"documentation_url": "https://github.com/kyutai-labs/delayed-streams-modeling",
|
|
89
|
+
"variables": [
|
|
90
|
+
{
|
|
91
|
+
"name": "endpoint",
|
|
92
|
+
"type": "string",
|
|
93
|
+
"label": "API Endpoint",
|
|
94
|
+
"description": "Override the default API endpoint URL.",
|
|
95
|
+
"default": "http://localhost:8201/v1/audio/speech"
|
|
96
|
+
},
|
|
97
|
+
{
|
|
98
|
+
"name": "speed",
|
|
99
|
+
"type": "number",
|
|
100
|
+
"label": "Speed",
|
|
101
|
+
"description": "Playback speed multiplier.",
|
|
102
|
+
"default": 1,
|
|
103
|
+
"min": 0.25,
|
|
104
|
+
"max": 4
|
|
105
|
+
}
|
|
106
|
+
]
|
|
45
107
|
}
|
package/plugins/kyutai-tts.json
CHANGED
|
@@ -21,9 +21,20 @@
|
|
|
21
21
|
"tts-1"
|
|
22
22
|
],
|
|
23
23
|
"config": {
|
|
24
|
-
"voices": [
|
|
24
|
+
"voices": [
|
|
25
|
+
"Alba",
|
|
26
|
+
"Marius",
|
|
27
|
+
"Javert",
|
|
28
|
+
"Jean",
|
|
29
|
+
"Fantine",
|
|
30
|
+
"Cosette",
|
|
31
|
+
"Eponine",
|
|
32
|
+
"Azelma"
|
|
33
|
+
],
|
|
25
34
|
"default_voice": "Alba",
|
|
26
|
-
"formats": [
|
|
35
|
+
"formats": [
|
|
36
|
+
"wav"
|
|
37
|
+
],
|
|
27
38
|
"default_format": "wav",
|
|
28
39
|
"max_characters": 10000,
|
|
29
40
|
"supports_streaming": true,
|
|
@@ -32,5 +43,23 @@
|
|
|
32
43
|
}
|
|
33
44
|
},
|
|
34
45
|
"description": "Kyutai Pocket TTS local server (CPU-based, voice cloning support)",
|
|
35
|
-
"documentation_url": "https://github.com/kyutai-labs/pocket-tts"
|
|
46
|
+
"documentation_url": "https://github.com/kyutai-labs/pocket-tts",
|
|
47
|
+
"variables": [
|
|
48
|
+
{
|
|
49
|
+
"name": "endpoint",
|
|
50
|
+
"type": "string",
|
|
51
|
+
"label": "API Endpoint",
|
|
52
|
+
"description": "Override the default API endpoint URL.",
|
|
53
|
+
"default": "http://localhost:8200/v1/audio/speech"
|
|
54
|
+
},
|
|
55
|
+
{
|
|
56
|
+
"name": "speed",
|
|
57
|
+
"type": "number",
|
|
58
|
+
"label": "Speed",
|
|
59
|
+
"description": "Playback speed multiplier.",
|
|
60
|
+
"default": 1,
|
|
61
|
+
"min": 0.25,
|
|
62
|
+
"max": 4
|
|
63
|
+
}
|
|
64
|
+
]
|
|
36
65
|
}
|