@soederpop/luca 0.0.16 → 0.0.17
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/src/bootstrap/generated.ts +1 -1
- package/src/commands/chat.ts +41 -1
- package/src/introspection/generated.agi.ts +469 -469
- package/src/introspection/generated.node.ts +1073 -1073
- package/src/introspection/generated.web.ts +1 -1
- package/src/node/features/repl.ts +19 -6
- package/src/scaffolds/generated.ts +1 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
|
|
2
2
|
|
|
3
3
|
// Auto-generated introspection registry data
|
|
4
|
-
// Generated at: 2026-03-20T16:
|
|
4
|
+
// Generated at: 2026-03-20T16:25:11.024Z
|
|
5
5
|
|
|
6
6
|
setBuildTimeData('features.googleDocs', {
|
|
7
7
|
"id": "features.googleDocs",
|
|
@@ -9983,208 +9983,6 @@ setBuildTimeData('clients.supabase', {
|
|
|
9983
9983
|
]
|
|
9984
9984
|
});
|
|
9985
9985
|
|
|
9986
|
-
setBuildTimeData('clients.elevenlabs', {
|
|
9987
|
-
"id": "clients.elevenlabs",
|
|
9988
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
9989
|
-
"shortcut": "clients.elevenlabs",
|
|
9990
|
-
"className": "ElevenLabsClient",
|
|
9991
|
-
"methods": {
|
|
9992
|
-
"beforeRequest": {
|
|
9993
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
9994
|
-
"parameters": {},
|
|
9995
|
-
"required": [],
|
|
9996
|
-
"returns": "void"
|
|
9997
|
-
},
|
|
9998
|
-
"connect": {
|
|
9999
|
-
"description": "Validate the API key by listing available models.",
|
|
10000
|
-
"parameters": {},
|
|
10001
|
-
"required": [],
|
|
10002
|
-
"returns": "Promise<this>",
|
|
10003
|
-
"examples": [
|
|
10004
|
-
{
|
|
10005
|
-
"language": "ts",
|
|
10006
|
-
"code": "await el.connect()"
|
|
10007
|
-
}
|
|
10008
|
-
]
|
|
10009
|
-
},
|
|
10010
|
-
"listVoices": {
|
|
10011
|
-
"description": "List available voices with optional search and filtering.",
|
|
10012
|
-
"parameters": {
|
|
10013
|
-
"options": {
|
|
10014
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10015
|
-
"description": "Query parameters for filtering voices"
|
|
10016
|
-
}
|
|
10017
|
-
},
|
|
10018
|
-
"required": [],
|
|
10019
|
-
"returns": "Promise<any>",
|
|
10020
|
-
"examples": [
|
|
10021
|
-
{
|
|
10022
|
-
"language": "ts",
|
|
10023
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10024
|
-
}
|
|
10025
|
-
]
|
|
10026
|
-
},
|
|
10027
|
-
"getVoice": {
|
|
10028
|
-
"description": "Get details for a single voice.",
|
|
10029
|
-
"parameters": {
|
|
10030
|
-
"voiceId": {
|
|
10031
|
-
"type": "string",
|
|
10032
|
-
"description": "The voice ID to look up"
|
|
10033
|
-
}
|
|
10034
|
-
},
|
|
10035
|
-
"required": [
|
|
10036
|
-
"voiceId"
|
|
10037
|
-
],
|
|
10038
|
-
"returns": "Promise<any>",
|
|
10039
|
-
"examples": [
|
|
10040
|
-
{
|
|
10041
|
-
"language": "ts",
|
|
10042
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10043
|
-
}
|
|
10044
|
-
]
|
|
10045
|
-
},
|
|
10046
|
-
"listModels": {
|
|
10047
|
-
"description": "List available TTS models.",
|
|
10048
|
-
"parameters": {},
|
|
10049
|
-
"required": [],
|
|
10050
|
-
"returns": "Promise<any[]>",
|
|
10051
|
-
"examples": [
|
|
10052
|
-
{
|
|
10053
|
-
"language": "ts",
|
|
10054
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10055
|
-
}
|
|
10056
|
-
]
|
|
10057
|
-
},
|
|
10058
|
-
"synthesize": {
|
|
10059
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10060
|
-
"parameters": {
|
|
10061
|
-
"text": {
|
|
10062
|
-
"type": "string",
|
|
10063
|
-
"description": "The text to convert to speech"
|
|
10064
|
-
},
|
|
10065
|
-
"options": {
|
|
10066
|
-
"type": "SynthesizeOptions",
|
|
10067
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10068
|
-
"properties": {
|
|
10069
|
-
"voiceId": {
|
|
10070
|
-
"type": "string",
|
|
10071
|
-
"description": ""
|
|
10072
|
-
},
|
|
10073
|
-
"modelId": {
|
|
10074
|
-
"type": "string",
|
|
10075
|
-
"description": ""
|
|
10076
|
-
},
|
|
10077
|
-
"outputFormat": {
|
|
10078
|
-
"type": "string",
|
|
10079
|
-
"description": ""
|
|
10080
|
-
},
|
|
10081
|
-
"voiceSettings": {
|
|
10082
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10083
|
-
"description": ""
|
|
10084
|
-
},
|
|
10085
|
-
"disableCache": {
|
|
10086
|
-
"type": "boolean",
|
|
10087
|
-
"description": ""
|
|
10088
|
-
}
|
|
10089
|
-
}
|
|
10090
|
-
}
|
|
10091
|
-
},
|
|
10092
|
-
"required": [
|
|
10093
|
-
"text"
|
|
10094
|
-
],
|
|
10095
|
-
"returns": "Promise<Buffer>",
|
|
10096
|
-
"examples": [
|
|
10097
|
-
{
|
|
10098
|
-
"language": "ts",
|
|
10099
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10100
|
-
}
|
|
10101
|
-
]
|
|
10102
|
-
},
|
|
10103
|
-
"say": {
|
|
10104
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
10105
|
-
"parameters": {
|
|
10106
|
-
"text": {
|
|
10107
|
-
"type": "string",
|
|
10108
|
-
"description": "The text to convert to speech"
|
|
10109
|
-
},
|
|
10110
|
-
"outputPath": {
|
|
10111
|
-
"type": "string",
|
|
10112
|
-
"description": "File path to write the audio to"
|
|
10113
|
-
},
|
|
10114
|
-
"options": {
|
|
10115
|
-
"type": "SynthesizeOptions",
|
|
10116
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10117
|
-
"properties": {
|
|
10118
|
-
"voiceId": {
|
|
10119
|
-
"type": "string",
|
|
10120
|
-
"description": ""
|
|
10121
|
-
},
|
|
10122
|
-
"modelId": {
|
|
10123
|
-
"type": "string",
|
|
10124
|
-
"description": ""
|
|
10125
|
-
},
|
|
10126
|
-
"outputFormat": {
|
|
10127
|
-
"type": "string",
|
|
10128
|
-
"description": ""
|
|
10129
|
-
},
|
|
10130
|
-
"voiceSettings": {
|
|
10131
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10132
|
-
"description": ""
|
|
10133
|
-
},
|
|
10134
|
-
"disableCache": {
|
|
10135
|
-
"type": "boolean",
|
|
10136
|
-
"description": ""
|
|
10137
|
-
}
|
|
10138
|
-
}
|
|
10139
|
-
}
|
|
10140
|
-
},
|
|
10141
|
-
"required": [
|
|
10142
|
-
"text",
|
|
10143
|
-
"outputPath"
|
|
10144
|
-
],
|
|
10145
|
-
"returns": "Promise<string>",
|
|
10146
|
-
"examples": [
|
|
10147
|
-
{
|
|
10148
|
-
"language": "ts",
|
|
10149
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10150
|
-
}
|
|
10151
|
-
]
|
|
10152
|
-
}
|
|
10153
|
-
},
|
|
10154
|
-
"getters": {
|
|
10155
|
-
"apiKey": {
|
|
10156
|
-
"description": "The resolved API key from options or environment.",
|
|
10157
|
-
"returns": "string"
|
|
10158
|
-
}
|
|
10159
|
-
},
|
|
10160
|
-
"events": {
|
|
10161
|
-
"failure": {
|
|
10162
|
-
"name": "failure",
|
|
10163
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10164
|
-
"arguments": {}
|
|
10165
|
-
},
|
|
10166
|
-
"voices": {
|
|
10167
|
-
"name": "voices",
|
|
10168
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10169
|
-
"arguments": {}
|
|
10170
|
-
},
|
|
10171
|
-
"speech": {
|
|
10172
|
-
"name": "speech",
|
|
10173
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10174
|
-
"arguments": {}
|
|
10175
|
-
}
|
|
10176
|
-
},
|
|
10177
|
-
"state": {},
|
|
10178
|
-
"options": {},
|
|
10179
|
-
"envVars": [],
|
|
10180
|
-
"examples": [
|
|
10181
|
-
{
|
|
10182
|
-
"language": "ts",
|
|
10183
|
-
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10184
|
-
}
|
|
10185
|
-
]
|
|
10186
|
-
});
|
|
10187
|
-
|
|
10188
9986
|
setBuildTimeData('clients.comfyui', {
|
|
10189
9987
|
"id": "clients.comfyui",
|
|
10190
9988
|
"description": "ComfyUI client — execute Stable Diffusion workflows via the ComfyUI API. Connects to a ComfyUI instance to queue prompts, track execution via WebSocket or polling, and download generated images. Supports both UI-format and API-format workflows with automatic conversion.",
|
|
@@ -10390,39 +10188,241 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10390
10188
|
}
|
|
10391
10189
|
},
|
|
10392
10190
|
"events": {
|
|
10393
|
-
"execution_start": {
|
|
10394
|
-
"name": "execution_start",
|
|
10395
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10396
|
-
"arguments": {}
|
|
10397
|
-
},
|
|
10398
|
-
"execution_complete": {
|
|
10399
|
-
"name": "execution_complete",
|
|
10400
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10401
|
-
"arguments": {}
|
|
10402
|
-
},
|
|
10403
|
-
"executing": {
|
|
10404
|
-
"name": "executing",
|
|
10405
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10406
|
-
"arguments": {}
|
|
10407
|
-
},
|
|
10408
|
-
"progress": {
|
|
10409
|
-
"name": "progress",
|
|
10410
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10411
|
-
"arguments": {}
|
|
10412
|
-
},
|
|
10413
|
-
"executed": {
|
|
10414
|
-
"name": "executed",
|
|
10415
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10191
|
+
"execution_start": {
|
|
10192
|
+
"name": "execution_start",
|
|
10193
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10194
|
+
"arguments": {}
|
|
10195
|
+
},
|
|
10196
|
+
"execution_complete": {
|
|
10197
|
+
"name": "execution_complete",
|
|
10198
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10199
|
+
"arguments": {}
|
|
10200
|
+
},
|
|
10201
|
+
"executing": {
|
|
10202
|
+
"name": "executing",
|
|
10203
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10204
|
+
"arguments": {}
|
|
10205
|
+
},
|
|
10206
|
+
"progress": {
|
|
10207
|
+
"name": "progress",
|
|
10208
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10209
|
+
"arguments": {}
|
|
10210
|
+
},
|
|
10211
|
+
"executed": {
|
|
10212
|
+
"name": "executed",
|
|
10213
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10214
|
+
"arguments": {}
|
|
10215
|
+
},
|
|
10216
|
+
"execution_cached": {
|
|
10217
|
+
"name": "execution_cached",
|
|
10218
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10219
|
+
"arguments": {}
|
|
10220
|
+
},
|
|
10221
|
+
"execution_error": {
|
|
10222
|
+
"name": "execution_error",
|
|
10223
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10224
|
+
"arguments": {}
|
|
10225
|
+
}
|
|
10226
|
+
},
|
|
10227
|
+
"state": {},
|
|
10228
|
+
"options": {},
|
|
10229
|
+
"envVars": [],
|
|
10230
|
+
"examples": [
|
|
10231
|
+
{
|
|
10232
|
+
"language": "ts",
|
|
10233
|
+
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
10234
|
+
}
|
|
10235
|
+
]
|
|
10236
|
+
});
|
|
10237
|
+
|
|
10238
|
+
setBuildTimeData('clients.elevenlabs', {
|
|
10239
|
+
"id": "clients.elevenlabs",
|
|
10240
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
10241
|
+
"shortcut": "clients.elevenlabs",
|
|
10242
|
+
"className": "ElevenLabsClient",
|
|
10243
|
+
"methods": {
|
|
10244
|
+
"beforeRequest": {
|
|
10245
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
10246
|
+
"parameters": {},
|
|
10247
|
+
"required": [],
|
|
10248
|
+
"returns": "void"
|
|
10249
|
+
},
|
|
10250
|
+
"connect": {
|
|
10251
|
+
"description": "Validate the API key by listing available models.",
|
|
10252
|
+
"parameters": {},
|
|
10253
|
+
"required": [],
|
|
10254
|
+
"returns": "Promise<this>",
|
|
10255
|
+
"examples": [
|
|
10256
|
+
{
|
|
10257
|
+
"language": "ts",
|
|
10258
|
+
"code": "await el.connect()"
|
|
10259
|
+
}
|
|
10260
|
+
]
|
|
10261
|
+
},
|
|
10262
|
+
"listVoices": {
|
|
10263
|
+
"description": "List available voices with optional search and filtering.",
|
|
10264
|
+
"parameters": {
|
|
10265
|
+
"options": {
|
|
10266
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10267
|
+
"description": "Query parameters for filtering voices"
|
|
10268
|
+
}
|
|
10269
|
+
},
|
|
10270
|
+
"required": [],
|
|
10271
|
+
"returns": "Promise<any>",
|
|
10272
|
+
"examples": [
|
|
10273
|
+
{
|
|
10274
|
+
"language": "ts",
|
|
10275
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10276
|
+
}
|
|
10277
|
+
]
|
|
10278
|
+
},
|
|
10279
|
+
"getVoice": {
|
|
10280
|
+
"description": "Get details for a single voice.",
|
|
10281
|
+
"parameters": {
|
|
10282
|
+
"voiceId": {
|
|
10283
|
+
"type": "string",
|
|
10284
|
+
"description": "The voice ID to look up"
|
|
10285
|
+
}
|
|
10286
|
+
},
|
|
10287
|
+
"required": [
|
|
10288
|
+
"voiceId"
|
|
10289
|
+
],
|
|
10290
|
+
"returns": "Promise<any>",
|
|
10291
|
+
"examples": [
|
|
10292
|
+
{
|
|
10293
|
+
"language": "ts",
|
|
10294
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10295
|
+
}
|
|
10296
|
+
]
|
|
10297
|
+
},
|
|
10298
|
+
"listModels": {
|
|
10299
|
+
"description": "List available TTS models.",
|
|
10300
|
+
"parameters": {},
|
|
10301
|
+
"required": [],
|
|
10302
|
+
"returns": "Promise<any[]>",
|
|
10303
|
+
"examples": [
|
|
10304
|
+
{
|
|
10305
|
+
"language": "ts",
|
|
10306
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10307
|
+
}
|
|
10308
|
+
]
|
|
10309
|
+
},
|
|
10310
|
+
"synthesize": {
|
|
10311
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10312
|
+
"parameters": {
|
|
10313
|
+
"text": {
|
|
10314
|
+
"type": "string",
|
|
10315
|
+
"description": "The text to convert to speech"
|
|
10316
|
+
},
|
|
10317
|
+
"options": {
|
|
10318
|
+
"type": "SynthesizeOptions",
|
|
10319
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10320
|
+
"properties": {
|
|
10321
|
+
"voiceId": {
|
|
10322
|
+
"type": "string",
|
|
10323
|
+
"description": ""
|
|
10324
|
+
},
|
|
10325
|
+
"modelId": {
|
|
10326
|
+
"type": "string",
|
|
10327
|
+
"description": ""
|
|
10328
|
+
},
|
|
10329
|
+
"outputFormat": {
|
|
10330
|
+
"type": "string",
|
|
10331
|
+
"description": ""
|
|
10332
|
+
},
|
|
10333
|
+
"voiceSettings": {
|
|
10334
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10335
|
+
"description": ""
|
|
10336
|
+
},
|
|
10337
|
+
"disableCache": {
|
|
10338
|
+
"type": "boolean",
|
|
10339
|
+
"description": ""
|
|
10340
|
+
}
|
|
10341
|
+
}
|
|
10342
|
+
}
|
|
10343
|
+
},
|
|
10344
|
+
"required": [
|
|
10345
|
+
"text"
|
|
10346
|
+
],
|
|
10347
|
+
"returns": "Promise<Buffer>",
|
|
10348
|
+
"examples": [
|
|
10349
|
+
{
|
|
10350
|
+
"language": "ts",
|
|
10351
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10352
|
+
}
|
|
10353
|
+
]
|
|
10354
|
+
},
|
|
10355
|
+
"say": {
|
|
10356
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
10357
|
+
"parameters": {
|
|
10358
|
+
"text": {
|
|
10359
|
+
"type": "string",
|
|
10360
|
+
"description": "The text to convert to speech"
|
|
10361
|
+
},
|
|
10362
|
+
"outputPath": {
|
|
10363
|
+
"type": "string",
|
|
10364
|
+
"description": "File path to write the audio to"
|
|
10365
|
+
},
|
|
10366
|
+
"options": {
|
|
10367
|
+
"type": "SynthesizeOptions",
|
|
10368
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10369
|
+
"properties": {
|
|
10370
|
+
"voiceId": {
|
|
10371
|
+
"type": "string",
|
|
10372
|
+
"description": ""
|
|
10373
|
+
},
|
|
10374
|
+
"modelId": {
|
|
10375
|
+
"type": "string",
|
|
10376
|
+
"description": ""
|
|
10377
|
+
},
|
|
10378
|
+
"outputFormat": {
|
|
10379
|
+
"type": "string",
|
|
10380
|
+
"description": ""
|
|
10381
|
+
},
|
|
10382
|
+
"voiceSettings": {
|
|
10383
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10384
|
+
"description": ""
|
|
10385
|
+
},
|
|
10386
|
+
"disableCache": {
|
|
10387
|
+
"type": "boolean",
|
|
10388
|
+
"description": ""
|
|
10389
|
+
}
|
|
10390
|
+
}
|
|
10391
|
+
}
|
|
10392
|
+
},
|
|
10393
|
+
"required": [
|
|
10394
|
+
"text",
|
|
10395
|
+
"outputPath"
|
|
10396
|
+
],
|
|
10397
|
+
"returns": "Promise<string>",
|
|
10398
|
+
"examples": [
|
|
10399
|
+
{
|
|
10400
|
+
"language": "ts",
|
|
10401
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10402
|
+
}
|
|
10403
|
+
]
|
|
10404
|
+
}
|
|
10405
|
+
},
|
|
10406
|
+
"getters": {
|
|
10407
|
+
"apiKey": {
|
|
10408
|
+
"description": "The resolved API key from options or environment.",
|
|
10409
|
+
"returns": "string"
|
|
10410
|
+
}
|
|
10411
|
+
},
|
|
10412
|
+
"events": {
|
|
10413
|
+
"failure": {
|
|
10414
|
+
"name": "failure",
|
|
10415
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10416
10416
|
"arguments": {}
|
|
10417
10417
|
},
|
|
10418
|
-
"
|
|
10419
|
-
"name": "
|
|
10420
|
-
"description": "Event emitted by
|
|
10418
|
+
"voices": {
|
|
10419
|
+
"name": "voices",
|
|
10420
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10421
10421
|
"arguments": {}
|
|
10422
10422
|
},
|
|
10423
|
-
"
|
|
10424
|
-
"name": "
|
|
10425
|
-
"description": "Event emitted by
|
|
10423
|
+
"speech": {
|
|
10424
|
+
"name": "speech",
|
|
10425
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10426
10426
|
"arguments": {}
|
|
10427
10427
|
}
|
|
10428
10428
|
},
|
|
@@ -10432,7 +10432,7 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10432
10432
|
"examples": [
|
|
10433
10433
|
{
|
|
10434
10434
|
"language": "ts",
|
|
10435
|
-
"code": "const
|
|
10435
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10436
10436
|
}
|
|
10437
10437
|
]
|
|
10438
10438
|
});
|
|
@@ -23408,256 +23408,55 @@ export const introspectionData = [
|
|
|
23408
23408
|
"parameters": {
|
|
23409
23409
|
"channelName": {
|
|
23410
23410
|
"type": "string",
|
|
23411
|
-
"description": "The channel name to remove"
|
|
23412
|
-
}
|
|
23413
|
-
},
|
|
23414
|
-
"required": [
|
|
23415
|
-
"channelName"
|
|
23416
|
-
],
|
|
23417
|
-
"returns": "void"
|
|
23418
|
-
},
|
|
23419
|
-
"unsubscribeAll": {
|
|
23420
|
-
"description": "Unsubscribe and remove all realtime channels.",
|
|
23421
|
-
"parameters": {},
|
|
23422
|
-
"required": [],
|
|
23423
|
-
"returns": "void"
|
|
23424
|
-
},
|
|
23425
|
-
"connect": {
|
|
23426
|
-
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
23427
|
-
"parameters": {},
|
|
23428
|
-
"required": [],
|
|
23429
|
-
"returns": "void"
|
|
23430
|
-
},
|
|
23431
|
-
"disconnect": {
|
|
23432
|
-
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
23433
|
-
"parameters": {},
|
|
23434
|
-
"required": [],
|
|
23435
|
-
"returns": "void"
|
|
23436
|
-
}
|
|
23437
|
-
},
|
|
23438
|
-
"getters": {
|
|
23439
|
-
"sdk": {
|
|
23440
|
-
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
23441
|
-
"returns": "SupabaseSDKClient<any, any>"
|
|
23442
|
-
},
|
|
23443
|
-
"storage": {
|
|
23444
|
-
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
23445
|
-
"returns": "any"
|
|
23446
|
-
},
|
|
23447
|
-
"functions": {
|
|
23448
|
-
"description": "Returns the Supabase Functions client.",
|
|
23449
|
-
"returns": "any"
|
|
23450
|
-
}
|
|
23451
|
-
},
|
|
23452
|
-
"events": {},
|
|
23453
|
-
"state": {},
|
|
23454
|
-
"options": {},
|
|
23455
|
-
"envVars": [],
|
|
23456
|
-
"examples": [
|
|
23457
|
-
{
|
|
23458
|
-
"language": "ts",
|
|
23459
|
-
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
23460
|
-
}
|
|
23461
|
-
]
|
|
23462
|
-
},
|
|
23463
|
-
{
|
|
23464
|
-
"id": "clients.elevenlabs",
|
|
23465
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
23466
|
-
"shortcut": "clients.elevenlabs",
|
|
23467
|
-
"className": "ElevenLabsClient",
|
|
23468
|
-
"methods": {
|
|
23469
|
-
"beforeRequest": {
|
|
23470
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
23471
|
-
"parameters": {},
|
|
23472
|
-
"required": [],
|
|
23473
|
-
"returns": "void"
|
|
23474
|
-
},
|
|
23475
|
-
"connect": {
|
|
23476
|
-
"description": "Validate the API key by listing available models.",
|
|
23477
|
-
"parameters": {},
|
|
23478
|
-
"required": [],
|
|
23479
|
-
"returns": "Promise<this>",
|
|
23480
|
-
"examples": [
|
|
23481
|
-
{
|
|
23482
|
-
"language": "ts",
|
|
23483
|
-
"code": "await el.connect()"
|
|
23484
|
-
}
|
|
23485
|
-
]
|
|
23486
|
-
},
|
|
23487
|
-
"listVoices": {
|
|
23488
|
-
"description": "List available voices with optional search and filtering.",
|
|
23489
|
-
"parameters": {
|
|
23490
|
-
"options": {
|
|
23491
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
23492
|
-
"description": "Query parameters for filtering voices"
|
|
23493
|
-
}
|
|
23494
|
-
},
|
|
23495
|
-
"required": [],
|
|
23496
|
-
"returns": "Promise<any>",
|
|
23497
|
-
"examples": [
|
|
23498
|
-
{
|
|
23499
|
-
"language": "ts",
|
|
23500
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
23501
|
-
}
|
|
23502
|
-
]
|
|
23503
|
-
},
|
|
23504
|
-
"getVoice": {
|
|
23505
|
-
"description": "Get details for a single voice.",
|
|
23506
|
-
"parameters": {
|
|
23507
|
-
"voiceId": {
|
|
23508
|
-
"type": "string",
|
|
23509
|
-
"description": "The voice ID to look up"
|
|
23510
|
-
}
|
|
23511
|
-
},
|
|
23512
|
-
"required": [
|
|
23513
|
-
"voiceId"
|
|
23514
|
-
],
|
|
23515
|
-
"returns": "Promise<any>",
|
|
23516
|
-
"examples": [
|
|
23517
|
-
{
|
|
23518
|
-
"language": "ts",
|
|
23519
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
23520
|
-
}
|
|
23521
|
-
]
|
|
23522
|
-
},
|
|
23523
|
-
"listModels": {
|
|
23524
|
-
"description": "List available TTS models.",
|
|
23525
|
-
"parameters": {},
|
|
23526
|
-
"required": [],
|
|
23527
|
-
"returns": "Promise<any[]>",
|
|
23528
|
-
"examples": [
|
|
23529
|
-
{
|
|
23530
|
-
"language": "ts",
|
|
23531
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
23532
|
-
}
|
|
23533
|
-
]
|
|
23534
|
-
},
|
|
23535
|
-
"synthesize": {
|
|
23536
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
23537
|
-
"parameters": {
|
|
23538
|
-
"text": {
|
|
23539
|
-
"type": "string",
|
|
23540
|
-
"description": "The text to convert to speech"
|
|
23541
|
-
},
|
|
23542
|
-
"options": {
|
|
23543
|
-
"type": "SynthesizeOptions",
|
|
23544
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
23545
|
-
"properties": {
|
|
23546
|
-
"voiceId": {
|
|
23547
|
-
"type": "string",
|
|
23548
|
-
"description": ""
|
|
23549
|
-
},
|
|
23550
|
-
"modelId": {
|
|
23551
|
-
"type": "string",
|
|
23552
|
-
"description": ""
|
|
23553
|
-
},
|
|
23554
|
-
"outputFormat": {
|
|
23555
|
-
"type": "string",
|
|
23556
|
-
"description": ""
|
|
23557
|
-
},
|
|
23558
|
-
"voiceSettings": {
|
|
23559
|
-
"type": "ElevenLabsVoiceSettings",
|
|
23560
|
-
"description": ""
|
|
23561
|
-
},
|
|
23562
|
-
"disableCache": {
|
|
23563
|
-
"type": "boolean",
|
|
23564
|
-
"description": ""
|
|
23565
|
-
}
|
|
23566
|
-
}
|
|
23567
|
-
}
|
|
23568
|
-
},
|
|
23569
|
-
"required": [
|
|
23570
|
-
"text"
|
|
23571
|
-
],
|
|
23572
|
-
"returns": "Promise<Buffer>",
|
|
23573
|
-
"examples": [
|
|
23574
|
-
{
|
|
23575
|
-
"language": "ts",
|
|
23576
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
23577
|
-
}
|
|
23578
|
-
]
|
|
23579
|
-
},
|
|
23580
|
-
"say": {
|
|
23581
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
23582
|
-
"parameters": {
|
|
23583
|
-
"text": {
|
|
23584
|
-
"type": "string",
|
|
23585
|
-
"description": "The text to convert to speech"
|
|
23586
|
-
},
|
|
23587
|
-
"outputPath": {
|
|
23588
|
-
"type": "string",
|
|
23589
|
-
"description": "File path to write the audio to"
|
|
23590
|
-
},
|
|
23591
|
-
"options": {
|
|
23592
|
-
"type": "SynthesizeOptions",
|
|
23593
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
23594
|
-
"properties": {
|
|
23595
|
-
"voiceId": {
|
|
23596
|
-
"type": "string",
|
|
23597
|
-
"description": ""
|
|
23598
|
-
},
|
|
23599
|
-
"modelId": {
|
|
23600
|
-
"type": "string",
|
|
23601
|
-
"description": ""
|
|
23602
|
-
},
|
|
23603
|
-
"outputFormat": {
|
|
23604
|
-
"type": "string",
|
|
23605
|
-
"description": ""
|
|
23606
|
-
},
|
|
23607
|
-
"voiceSettings": {
|
|
23608
|
-
"type": "ElevenLabsVoiceSettings",
|
|
23609
|
-
"description": ""
|
|
23610
|
-
},
|
|
23611
|
-
"disableCache": {
|
|
23612
|
-
"type": "boolean",
|
|
23613
|
-
"description": ""
|
|
23614
|
-
}
|
|
23615
|
-
}
|
|
23616
|
-
}
|
|
23617
|
-
},
|
|
23618
|
-
"required": [
|
|
23619
|
-
"text",
|
|
23620
|
-
"outputPath"
|
|
23621
|
-
],
|
|
23622
|
-
"returns": "Promise<string>",
|
|
23623
|
-
"examples": [
|
|
23624
|
-
{
|
|
23625
|
-
"language": "ts",
|
|
23626
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
23411
|
+
"description": "The channel name to remove"
|
|
23627
23412
|
}
|
|
23628
|
-
|
|
23413
|
+
},
|
|
23414
|
+
"required": [
|
|
23415
|
+
"channelName"
|
|
23416
|
+
],
|
|
23417
|
+
"returns": "void"
|
|
23418
|
+
},
|
|
23419
|
+
"unsubscribeAll": {
|
|
23420
|
+
"description": "Unsubscribe and remove all realtime channels.",
|
|
23421
|
+
"parameters": {},
|
|
23422
|
+
"required": [],
|
|
23423
|
+
"returns": "void"
|
|
23424
|
+
},
|
|
23425
|
+
"connect": {
|
|
23426
|
+
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
23427
|
+
"parameters": {},
|
|
23428
|
+
"required": [],
|
|
23429
|
+
"returns": "void"
|
|
23430
|
+
},
|
|
23431
|
+
"disconnect": {
|
|
23432
|
+
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
23433
|
+
"parameters": {},
|
|
23434
|
+
"required": [],
|
|
23435
|
+
"returns": "void"
|
|
23629
23436
|
}
|
|
23630
23437
|
},
|
|
23631
23438
|
"getters": {
|
|
23632
|
-
"
|
|
23633
|
-
"description": "
|
|
23634
|
-
"returns": "
|
|
23635
|
-
}
|
|
23636
|
-
},
|
|
23637
|
-
"events": {
|
|
23638
|
-
"failure": {
|
|
23639
|
-
"name": "failure",
|
|
23640
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
23641
|
-
"arguments": {}
|
|
23439
|
+
"sdk": {
|
|
23440
|
+
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
23441
|
+
"returns": "SupabaseSDKClient<any, any>"
|
|
23642
23442
|
},
|
|
23643
|
-
"
|
|
23644
|
-
"
|
|
23645
|
-
"
|
|
23646
|
-
"arguments": {}
|
|
23443
|
+
"storage": {
|
|
23444
|
+
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
23445
|
+
"returns": "any"
|
|
23647
23446
|
},
|
|
23648
|
-
"
|
|
23649
|
-
"
|
|
23650
|
-
"
|
|
23651
|
-
"arguments": {}
|
|
23447
|
+
"functions": {
|
|
23448
|
+
"description": "Returns the Supabase Functions client.",
|
|
23449
|
+
"returns": "any"
|
|
23652
23450
|
}
|
|
23653
23451
|
},
|
|
23452
|
+
"events": {},
|
|
23654
23453
|
"state": {},
|
|
23655
23454
|
"options": {},
|
|
23656
23455
|
"envVars": [],
|
|
23657
23456
|
"examples": [
|
|
23658
23457
|
{
|
|
23659
23458
|
"language": "ts",
|
|
23660
|
-
"code": "const
|
|
23459
|
+
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
23661
23460
|
}
|
|
23662
23461
|
]
|
|
23663
23462
|
},
|
|
@@ -23912,6 +23711,207 @@ export const introspectionData = [
|
|
|
23912
23711
|
}
|
|
23913
23712
|
]
|
|
23914
23713
|
},
|
|
23714
|
+
{
|
|
23715
|
+
"id": "clients.elevenlabs",
|
|
23716
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
23717
|
+
"shortcut": "clients.elevenlabs",
|
|
23718
|
+
"className": "ElevenLabsClient",
|
|
23719
|
+
"methods": {
|
|
23720
|
+
"beforeRequest": {
|
|
23721
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
23722
|
+
"parameters": {},
|
|
23723
|
+
"required": [],
|
|
23724
|
+
"returns": "void"
|
|
23725
|
+
},
|
|
23726
|
+
"connect": {
|
|
23727
|
+
"description": "Validate the API key by listing available models.",
|
|
23728
|
+
"parameters": {},
|
|
23729
|
+
"required": [],
|
|
23730
|
+
"returns": "Promise<this>",
|
|
23731
|
+
"examples": [
|
|
23732
|
+
{
|
|
23733
|
+
"language": "ts",
|
|
23734
|
+
"code": "await el.connect()"
|
|
23735
|
+
}
|
|
23736
|
+
]
|
|
23737
|
+
},
|
|
23738
|
+
"listVoices": {
|
|
23739
|
+
"description": "List available voices with optional search and filtering.",
|
|
23740
|
+
"parameters": {
|
|
23741
|
+
"options": {
|
|
23742
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
23743
|
+
"description": "Query parameters for filtering voices"
|
|
23744
|
+
}
|
|
23745
|
+
},
|
|
23746
|
+
"required": [],
|
|
23747
|
+
"returns": "Promise<any>",
|
|
23748
|
+
"examples": [
|
|
23749
|
+
{
|
|
23750
|
+
"language": "ts",
|
|
23751
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
23752
|
+
}
|
|
23753
|
+
]
|
|
23754
|
+
},
|
|
23755
|
+
"getVoice": {
|
|
23756
|
+
"description": "Get details for a single voice.",
|
|
23757
|
+
"parameters": {
|
|
23758
|
+
"voiceId": {
|
|
23759
|
+
"type": "string",
|
|
23760
|
+
"description": "The voice ID to look up"
|
|
23761
|
+
}
|
|
23762
|
+
},
|
|
23763
|
+
"required": [
|
|
23764
|
+
"voiceId"
|
|
23765
|
+
],
|
|
23766
|
+
"returns": "Promise<any>",
|
|
23767
|
+
"examples": [
|
|
23768
|
+
{
|
|
23769
|
+
"language": "ts",
|
|
23770
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
23771
|
+
}
|
|
23772
|
+
]
|
|
23773
|
+
},
|
|
23774
|
+
"listModels": {
|
|
23775
|
+
"description": "List available TTS models.",
|
|
23776
|
+
"parameters": {},
|
|
23777
|
+
"required": [],
|
|
23778
|
+
"returns": "Promise<any[]>",
|
|
23779
|
+
"examples": [
|
|
23780
|
+
{
|
|
23781
|
+
"language": "ts",
|
|
23782
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
23783
|
+
}
|
|
23784
|
+
]
|
|
23785
|
+
},
|
|
23786
|
+
"synthesize": {
|
|
23787
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
23788
|
+
"parameters": {
|
|
23789
|
+
"text": {
|
|
23790
|
+
"type": "string",
|
|
23791
|
+
"description": "The text to convert to speech"
|
|
23792
|
+
},
|
|
23793
|
+
"options": {
|
|
23794
|
+
"type": "SynthesizeOptions",
|
|
23795
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
23796
|
+
"properties": {
|
|
23797
|
+
"voiceId": {
|
|
23798
|
+
"type": "string",
|
|
23799
|
+
"description": ""
|
|
23800
|
+
},
|
|
23801
|
+
"modelId": {
|
|
23802
|
+
"type": "string",
|
|
23803
|
+
"description": ""
|
|
23804
|
+
},
|
|
23805
|
+
"outputFormat": {
|
|
23806
|
+
"type": "string",
|
|
23807
|
+
"description": ""
|
|
23808
|
+
},
|
|
23809
|
+
"voiceSettings": {
|
|
23810
|
+
"type": "ElevenLabsVoiceSettings",
|
|
23811
|
+
"description": ""
|
|
23812
|
+
},
|
|
23813
|
+
"disableCache": {
|
|
23814
|
+
"type": "boolean",
|
|
23815
|
+
"description": ""
|
|
23816
|
+
}
|
|
23817
|
+
}
|
|
23818
|
+
}
|
|
23819
|
+
},
|
|
23820
|
+
"required": [
|
|
23821
|
+
"text"
|
|
23822
|
+
],
|
|
23823
|
+
"returns": "Promise<Buffer>",
|
|
23824
|
+
"examples": [
|
|
23825
|
+
{
|
|
23826
|
+
"language": "ts",
|
|
23827
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
23828
|
+
}
|
|
23829
|
+
]
|
|
23830
|
+
},
|
|
23831
|
+
"say": {
|
|
23832
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
23833
|
+
"parameters": {
|
|
23834
|
+
"text": {
|
|
23835
|
+
"type": "string",
|
|
23836
|
+
"description": "The text to convert to speech"
|
|
23837
|
+
},
|
|
23838
|
+
"outputPath": {
|
|
23839
|
+
"type": "string",
|
|
23840
|
+
"description": "File path to write the audio to"
|
|
23841
|
+
},
|
|
23842
|
+
"options": {
|
|
23843
|
+
"type": "SynthesizeOptions",
|
|
23844
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
23845
|
+
"properties": {
|
|
23846
|
+
"voiceId": {
|
|
23847
|
+
"type": "string",
|
|
23848
|
+
"description": ""
|
|
23849
|
+
},
|
|
23850
|
+
"modelId": {
|
|
23851
|
+
"type": "string",
|
|
23852
|
+
"description": ""
|
|
23853
|
+
},
|
|
23854
|
+
"outputFormat": {
|
|
23855
|
+
"type": "string",
|
|
23856
|
+
"description": ""
|
|
23857
|
+
},
|
|
23858
|
+
"voiceSettings": {
|
|
23859
|
+
"type": "ElevenLabsVoiceSettings",
|
|
23860
|
+
"description": ""
|
|
23861
|
+
},
|
|
23862
|
+
"disableCache": {
|
|
23863
|
+
"type": "boolean",
|
|
23864
|
+
"description": ""
|
|
23865
|
+
}
|
|
23866
|
+
}
|
|
23867
|
+
}
|
|
23868
|
+
},
|
|
23869
|
+
"required": [
|
|
23870
|
+
"text",
|
|
23871
|
+
"outputPath"
|
|
23872
|
+
],
|
|
23873
|
+
"returns": "Promise<string>",
|
|
23874
|
+
"examples": [
|
|
23875
|
+
{
|
|
23876
|
+
"language": "ts",
|
|
23877
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
23878
|
+
}
|
|
23879
|
+
]
|
|
23880
|
+
}
|
|
23881
|
+
},
|
|
23882
|
+
"getters": {
|
|
23883
|
+
"apiKey": {
|
|
23884
|
+
"description": "The resolved API key from options or environment.",
|
|
23885
|
+
"returns": "string"
|
|
23886
|
+
}
|
|
23887
|
+
},
|
|
23888
|
+
"events": {
|
|
23889
|
+
"failure": {
|
|
23890
|
+
"name": "failure",
|
|
23891
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23892
|
+
"arguments": {}
|
|
23893
|
+
},
|
|
23894
|
+
"voices": {
|
|
23895
|
+
"name": "voices",
|
|
23896
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23897
|
+
"arguments": {}
|
|
23898
|
+
},
|
|
23899
|
+
"speech": {
|
|
23900
|
+
"name": "speech",
|
|
23901
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23902
|
+
"arguments": {}
|
|
23903
|
+
}
|
|
23904
|
+
},
|
|
23905
|
+
"state": {},
|
|
23906
|
+
"options": {},
|
|
23907
|
+
"envVars": [],
|
|
23908
|
+
"examples": [
|
|
23909
|
+
{
|
|
23910
|
+
"language": "ts",
|
|
23911
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
23912
|
+
}
|
|
23913
|
+
]
|
|
23914
|
+
},
|
|
23915
23915
|
{
|
|
23916
23916
|
"id": "servers.mcp",
|
|
23917
23917
|
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|