@soederpop/luca 0.0.9 → 0.0.11
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/commands/release.ts +186 -0
- package/package.json +2 -2
- package/src/bootstrap/generated.ts +17 -3
- package/src/cli/cli.ts +20 -1
- package/src/introspection/generated.agi.ts +1138 -1138
- package/src/introspection/generated.node.ts +470 -470
- package/src/introspection/generated.web.ts +1 -1
- package/src/scaffolds/generated.ts +1 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
|
|
2
2
|
|
|
3
3
|
// Auto-generated introspection registry data
|
|
4
|
-
// Generated at: 2026-03-
|
|
4
|
+
// Generated at: 2026-03-20T00:15:39.136Z
|
|
5
5
|
|
|
6
6
|
setBuildTimeData('features.googleDocs', {
|
|
7
7
|
"id": "features.googleDocs",
|
|
@@ -9983,6 +9983,208 @@ setBuildTimeData('clients.supabase', {
|
|
|
9983
9983
|
]
|
|
9984
9984
|
});
|
|
9985
9985
|
|
|
9986
|
+
setBuildTimeData('clients.elevenlabs', {
|
|
9987
|
+
"id": "clients.elevenlabs",
|
|
9988
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
9989
|
+
"shortcut": "clients.elevenlabs",
|
|
9990
|
+
"className": "ElevenLabsClient",
|
|
9991
|
+
"methods": {
|
|
9992
|
+
"beforeRequest": {
|
|
9993
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
9994
|
+
"parameters": {},
|
|
9995
|
+
"required": [],
|
|
9996
|
+
"returns": "void"
|
|
9997
|
+
},
|
|
9998
|
+
"connect": {
|
|
9999
|
+
"description": "Validate the API key by listing available models.",
|
|
10000
|
+
"parameters": {},
|
|
10001
|
+
"required": [],
|
|
10002
|
+
"returns": "Promise<this>",
|
|
10003
|
+
"examples": [
|
|
10004
|
+
{
|
|
10005
|
+
"language": "ts",
|
|
10006
|
+
"code": "await el.connect()"
|
|
10007
|
+
}
|
|
10008
|
+
]
|
|
10009
|
+
},
|
|
10010
|
+
"listVoices": {
|
|
10011
|
+
"description": "List available voices with optional search and filtering.",
|
|
10012
|
+
"parameters": {
|
|
10013
|
+
"options": {
|
|
10014
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10015
|
+
"description": "Query parameters for filtering voices"
|
|
10016
|
+
}
|
|
10017
|
+
},
|
|
10018
|
+
"required": [],
|
|
10019
|
+
"returns": "Promise<any>",
|
|
10020
|
+
"examples": [
|
|
10021
|
+
{
|
|
10022
|
+
"language": "ts",
|
|
10023
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10024
|
+
}
|
|
10025
|
+
]
|
|
10026
|
+
},
|
|
10027
|
+
"getVoice": {
|
|
10028
|
+
"description": "Get details for a single voice.",
|
|
10029
|
+
"parameters": {
|
|
10030
|
+
"voiceId": {
|
|
10031
|
+
"type": "string",
|
|
10032
|
+
"description": "The voice ID to look up"
|
|
10033
|
+
}
|
|
10034
|
+
},
|
|
10035
|
+
"required": [
|
|
10036
|
+
"voiceId"
|
|
10037
|
+
],
|
|
10038
|
+
"returns": "Promise<any>",
|
|
10039
|
+
"examples": [
|
|
10040
|
+
{
|
|
10041
|
+
"language": "ts",
|
|
10042
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10043
|
+
}
|
|
10044
|
+
]
|
|
10045
|
+
},
|
|
10046
|
+
"listModels": {
|
|
10047
|
+
"description": "List available TTS models.",
|
|
10048
|
+
"parameters": {},
|
|
10049
|
+
"required": [],
|
|
10050
|
+
"returns": "Promise<any[]>",
|
|
10051
|
+
"examples": [
|
|
10052
|
+
{
|
|
10053
|
+
"language": "ts",
|
|
10054
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10055
|
+
}
|
|
10056
|
+
]
|
|
10057
|
+
},
|
|
10058
|
+
"synthesize": {
|
|
10059
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10060
|
+
"parameters": {
|
|
10061
|
+
"text": {
|
|
10062
|
+
"type": "string",
|
|
10063
|
+
"description": "The text to convert to speech"
|
|
10064
|
+
},
|
|
10065
|
+
"options": {
|
|
10066
|
+
"type": "SynthesizeOptions",
|
|
10067
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10068
|
+
"properties": {
|
|
10069
|
+
"voiceId": {
|
|
10070
|
+
"type": "string",
|
|
10071
|
+
"description": ""
|
|
10072
|
+
},
|
|
10073
|
+
"modelId": {
|
|
10074
|
+
"type": "string",
|
|
10075
|
+
"description": ""
|
|
10076
|
+
},
|
|
10077
|
+
"outputFormat": {
|
|
10078
|
+
"type": "string",
|
|
10079
|
+
"description": ""
|
|
10080
|
+
},
|
|
10081
|
+
"voiceSettings": {
|
|
10082
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10083
|
+
"description": ""
|
|
10084
|
+
},
|
|
10085
|
+
"disableCache": {
|
|
10086
|
+
"type": "boolean",
|
|
10087
|
+
"description": ""
|
|
10088
|
+
}
|
|
10089
|
+
}
|
|
10090
|
+
}
|
|
10091
|
+
},
|
|
10092
|
+
"required": [
|
|
10093
|
+
"text"
|
|
10094
|
+
],
|
|
10095
|
+
"returns": "Promise<Buffer>",
|
|
10096
|
+
"examples": [
|
|
10097
|
+
{
|
|
10098
|
+
"language": "ts",
|
|
10099
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10100
|
+
}
|
|
10101
|
+
]
|
|
10102
|
+
},
|
|
10103
|
+
"say": {
|
|
10104
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
10105
|
+
"parameters": {
|
|
10106
|
+
"text": {
|
|
10107
|
+
"type": "string",
|
|
10108
|
+
"description": "The text to convert to speech"
|
|
10109
|
+
},
|
|
10110
|
+
"outputPath": {
|
|
10111
|
+
"type": "string",
|
|
10112
|
+
"description": "File path to write the audio to"
|
|
10113
|
+
},
|
|
10114
|
+
"options": {
|
|
10115
|
+
"type": "SynthesizeOptions",
|
|
10116
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10117
|
+
"properties": {
|
|
10118
|
+
"voiceId": {
|
|
10119
|
+
"type": "string",
|
|
10120
|
+
"description": ""
|
|
10121
|
+
},
|
|
10122
|
+
"modelId": {
|
|
10123
|
+
"type": "string",
|
|
10124
|
+
"description": ""
|
|
10125
|
+
},
|
|
10126
|
+
"outputFormat": {
|
|
10127
|
+
"type": "string",
|
|
10128
|
+
"description": ""
|
|
10129
|
+
},
|
|
10130
|
+
"voiceSettings": {
|
|
10131
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10132
|
+
"description": ""
|
|
10133
|
+
},
|
|
10134
|
+
"disableCache": {
|
|
10135
|
+
"type": "boolean",
|
|
10136
|
+
"description": ""
|
|
10137
|
+
}
|
|
10138
|
+
}
|
|
10139
|
+
}
|
|
10140
|
+
},
|
|
10141
|
+
"required": [
|
|
10142
|
+
"text",
|
|
10143
|
+
"outputPath"
|
|
10144
|
+
],
|
|
10145
|
+
"returns": "Promise<string>",
|
|
10146
|
+
"examples": [
|
|
10147
|
+
{
|
|
10148
|
+
"language": "ts",
|
|
10149
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10150
|
+
}
|
|
10151
|
+
]
|
|
10152
|
+
}
|
|
10153
|
+
},
|
|
10154
|
+
"getters": {
|
|
10155
|
+
"apiKey": {
|
|
10156
|
+
"description": "The resolved API key from options or environment.",
|
|
10157
|
+
"returns": "string"
|
|
10158
|
+
}
|
|
10159
|
+
},
|
|
10160
|
+
"events": {
|
|
10161
|
+
"failure": {
|
|
10162
|
+
"name": "failure",
|
|
10163
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10164
|
+
"arguments": {}
|
|
10165
|
+
},
|
|
10166
|
+
"voices": {
|
|
10167
|
+
"name": "voices",
|
|
10168
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10169
|
+
"arguments": {}
|
|
10170
|
+
},
|
|
10171
|
+
"speech": {
|
|
10172
|
+
"name": "speech",
|
|
10173
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10174
|
+
"arguments": {}
|
|
10175
|
+
}
|
|
10176
|
+
},
|
|
10177
|
+
"state": {},
|
|
10178
|
+
"options": {},
|
|
10179
|
+
"envVars": [],
|
|
10180
|
+
"examples": [
|
|
10181
|
+
{
|
|
10182
|
+
"language": "ts",
|
|
10183
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10184
|
+
}
|
|
10185
|
+
]
|
|
10186
|
+
});
|
|
10187
|
+
|
|
9986
10188
|
setBuildTimeData('clients.comfyui', {
|
|
9987
10189
|
"id": "clients.comfyui",
|
|
9988
10190
|
"description": "ComfyUI client — execute Stable Diffusion workflows via the ComfyUI API. Connects to a ComfyUI instance to queue prompts, track execution via WebSocket or polling, and download generated images. Supports both UI-format and API-format workflows with automatic conversion.",
|
|
@@ -10182,247 +10384,45 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10182
10384
|
"description": "The unique client ID used for WebSocket session tracking.",
|
|
10183
10385
|
"returns": "string"
|
|
10184
10386
|
},
|
|
10185
|
-
"wsURL": {
|
|
10186
|
-
"description": "The WebSocket URL derived from baseURL or overridden via options.",
|
|
10187
|
-
"returns": "string"
|
|
10188
|
-
}
|
|
10189
|
-
},
|
|
10190
|
-
"events": {
|
|
10191
|
-
"execution_start": {
|
|
10192
|
-
"name": "execution_start",
|
|
10193
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10194
|
-
"arguments": {}
|
|
10195
|
-
},
|
|
10196
|
-
"execution_complete": {
|
|
10197
|
-
"name": "execution_complete",
|
|
10198
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10199
|
-
"arguments": {}
|
|
10200
|
-
},
|
|
10201
|
-
"executing": {
|
|
10202
|
-
"name": "executing",
|
|
10203
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10204
|
-
"arguments": {}
|
|
10205
|
-
},
|
|
10206
|
-
"progress": {
|
|
10207
|
-
"name": "progress",
|
|
10208
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10209
|
-
"arguments": {}
|
|
10210
|
-
},
|
|
10211
|
-
"executed": {
|
|
10212
|
-
"name": "executed",
|
|
10213
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10214
|
-
"arguments": {}
|
|
10215
|
-
},
|
|
10216
|
-
"execution_cached": {
|
|
10217
|
-
"name": "execution_cached",
|
|
10218
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10219
|
-
"arguments": {}
|
|
10220
|
-
},
|
|
10221
|
-
"execution_error": {
|
|
10222
|
-
"name": "execution_error",
|
|
10223
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10224
|
-
"arguments": {}
|
|
10225
|
-
}
|
|
10226
|
-
},
|
|
10227
|
-
"state": {},
|
|
10228
|
-
"options": {},
|
|
10229
|
-
"envVars": [],
|
|
10230
|
-
"examples": [
|
|
10231
|
-
{
|
|
10232
|
-
"language": "ts",
|
|
10233
|
-
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
10234
|
-
}
|
|
10235
|
-
]
|
|
10236
|
-
});
|
|
10237
|
-
|
|
10238
|
-
setBuildTimeData('clients.elevenlabs', {
|
|
10239
|
-
"id": "clients.elevenlabs",
|
|
10240
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
10241
|
-
"shortcut": "clients.elevenlabs",
|
|
10242
|
-
"className": "ElevenLabsClient",
|
|
10243
|
-
"methods": {
|
|
10244
|
-
"beforeRequest": {
|
|
10245
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
10246
|
-
"parameters": {},
|
|
10247
|
-
"required": [],
|
|
10248
|
-
"returns": "void"
|
|
10249
|
-
},
|
|
10250
|
-
"connect": {
|
|
10251
|
-
"description": "Validate the API key by listing available models.",
|
|
10252
|
-
"parameters": {},
|
|
10253
|
-
"required": [],
|
|
10254
|
-
"returns": "Promise<this>",
|
|
10255
|
-
"examples": [
|
|
10256
|
-
{
|
|
10257
|
-
"language": "ts",
|
|
10258
|
-
"code": "await el.connect()"
|
|
10259
|
-
}
|
|
10260
|
-
]
|
|
10261
|
-
},
|
|
10262
|
-
"listVoices": {
|
|
10263
|
-
"description": "List available voices with optional search and filtering.",
|
|
10264
|
-
"parameters": {
|
|
10265
|
-
"options": {
|
|
10266
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10267
|
-
"description": "Query parameters for filtering voices"
|
|
10268
|
-
}
|
|
10269
|
-
},
|
|
10270
|
-
"required": [],
|
|
10271
|
-
"returns": "Promise<any>",
|
|
10272
|
-
"examples": [
|
|
10273
|
-
{
|
|
10274
|
-
"language": "ts",
|
|
10275
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10276
|
-
}
|
|
10277
|
-
]
|
|
10278
|
-
},
|
|
10279
|
-
"getVoice": {
|
|
10280
|
-
"description": "Get details for a single voice.",
|
|
10281
|
-
"parameters": {
|
|
10282
|
-
"voiceId": {
|
|
10283
|
-
"type": "string",
|
|
10284
|
-
"description": "The voice ID to look up"
|
|
10285
|
-
}
|
|
10286
|
-
},
|
|
10287
|
-
"required": [
|
|
10288
|
-
"voiceId"
|
|
10289
|
-
],
|
|
10290
|
-
"returns": "Promise<any>",
|
|
10291
|
-
"examples": [
|
|
10292
|
-
{
|
|
10293
|
-
"language": "ts",
|
|
10294
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10295
|
-
}
|
|
10296
|
-
]
|
|
10297
|
-
},
|
|
10298
|
-
"listModels": {
|
|
10299
|
-
"description": "List available TTS models.",
|
|
10300
|
-
"parameters": {},
|
|
10301
|
-
"required": [],
|
|
10302
|
-
"returns": "Promise<any[]>",
|
|
10303
|
-
"examples": [
|
|
10304
|
-
{
|
|
10305
|
-
"language": "ts",
|
|
10306
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10307
|
-
}
|
|
10308
|
-
]
|
|
10309
|
-
},
|
|
10310
|
-
"synthesize": {
|
|
10311
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10312
|
-
"parameters": {
|
|
10313
|
-
"text": {
|
|
10314
|
-
"type": "string",
|
|
10315
|
-
"description": "The text to convert to speech"
|
|
10316
|
-
},
|
|
10317
|
-
"options": {
|
|
10318
|
-
"type": "SynthesizeOptions",
|
|
10319
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10320
|
-
"properties": {
|
|
10321
|
-
"voiceId": {
|
|
10322
|
-
"type": "string",
|
|
10323
|
-
"description": ""
|
|
10324
|
-
},
|
|
10325
|
-
"modelId": {
|
|
10326
|
-
"type": "string",
|
|
10327
|
-
"description": ""
|
|
10328
|
-
},
|
|
10329
|
-
"outputFormat": {
|
|
10330
|
-
"type": "string",
|
|
10331
|
-
"description": ""
|
|
10332
|
-
},
|
|
10333
|
-
"voiceSettings": {
|
|
10334
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10335
|
-
"description": ""
|
|
10336
|
-
},
|
|
10337
|
-
"disableCache": {
|
|
10338
|
-
"type": "boolean",
|
|
10339
|
-
"description": ""
|
|
10340
|
-
}
|
|
10341
|
-
}
|
|
10342
|
-
}
|
|
10343
|
-
},
|
|
10344
|
-
"required": [
|
|
10345
|
-
"text"
|
|
10346
|
-
],
|
|
10347
|
-
"returns": "Promise<Buffer>",
|
|
10348
|
-
"examples": [
|
|
10349
|
-
{
|
|
10350
|
-
"language": "ts",
|
|
10351
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10352
|
-
}
|
|
10353
|
-
]
|
|
10354
|
-
},
|
|
10355
|
-
"say": {
|
|
10356
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
10357
|
-
"parameters": {
|
|
10358
|
-
"text": {
|
|
10359
|
-
"type": "string",
|
|
10360
|
-
"description": "The text to convert to speech"
|
|
10361
|
-
},
|
|
10362
|
-
"outputPath": {
|
|
10363
|
-
"type": "string",
|
|
10364
|
-
"description": "File path to write the audio to"
|
|
10365
|
-
},
|
|
10366
|
-
"options": {
|
|
10367
|
-
"type": "SynthesizeOptions",
|
|
10368
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10369
|
-
"properties": {
|
|
10370
|
-
"voiceId": {
|
|
10371
|
-
"type": "string",
|
|
10372
|
-
"description": ""
|
|
10373
|
-
},
|
|
10374
|
-
"modelId": {
|
|
10375
|
-
"type": "string",
|
|
10376
|
-
"description": ""
|
|
10377
|
-
},
|
|
10378
|
-
"outputFormat": {
|
|
10379
|
-
"type": "string",
|
|
10380
|
-
"description": ""
|
|
10381
|
-
},
|
|
10382
|
-
"voiceSettings": {
|
|
10383
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10384
|
-
"description": ""
|
|
10385
|
-
},
|
|
10386
|
-
"disableCache": {
|
|
10387
|
-
"type": "boolean",
|
|
10388
|
-
"description": ""
|
|
10389
|
-
}
|
|
10390
|
-
}
|
|
10391
|
-
}
|
|
10392
|
-
},
|
|
10393
|
-
"required": [
|
|
10394
|
-
"text",
|
|
10395
|
-
"outputPath"
|
|
10396
|
-
],
|
|
10397
|
-
"returns": "Promise<string>",
|
|
10398
|
-
"examples": [
|
|
10399
|
-
{
|
|
10400
|
-
"language": "ts",
|
|
10401
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10402
|
-
}
|
|
10403
|
-
]
|
|
10404
|
-
}
|
|
10405
|
-
},
|
|
10406
|
-
"getters": {
|
|
10407
|
-
"apiKey": {
|
|
10408
|
-
"description": "The resolved API key from options or environment.",
|
|
10387
|
+
"wsURL": {
|
|
10388
|
+
"description": "The WebSocket URL derived from baseURL or overridden via options.",
|
|
10409
10389
|
"returns": "string"
|
|
10410
10390
|
}
|
|
10411
10391
|
},
|
|
10412
10392
|
"events": {
|
|
10413
|
-
"
|
|
10414
|
-
"name": "
|
|
10415
|
-
"description": "Event emitted by
|
|
10393
|
+
"execution_start": {
|
|
10394
|
+
"name": "execution_start",
|
|
10395
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10416
10396
|
"arguments": {}
|
|
10417
10397
|
},
|
|
10418
|
-
"
|
|
10419
|
-
"name": "
|
|
10420
|
-
"description": "Event emitted by
|
|
10398
|
+
"execution_complete": {
|
|
10399
|
+
"name": "execution_complete",
|
|
10400
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10421
10401
|
"arguments": {}
|
|
10422
10402
|
},
|
|
10423
|
-
"
|
|
10424
|
-
"name": "
|
|
10425
|
-
"description": "Event emitted by
|
|
10403
|
+
"executing": {
|
|
10404
|
+
"name": "executing",
|
|
10405
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10406
|
+
"arguments": {}
|
|
10407
|
+
},
|
|
10408
|
+
"progress": {
|
|
10409
|
+
"name": "progress",
|
|
10410
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10411
|
+
"arguments": {}
|
|
10412
|
+
},
|
|
10413
|
+
"executed": {
|
|
10414
|
+
"name": "executed",
|
|
10415
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10416
|
+
"arguments": {}
|
|
10417
|
+
},
|
|
10418
|
+
"execution_cached": {
|
|
10419
|
+
"name": "execution_cached",
|
|
10420
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10421
|
+
"arguments": {}
|
|
10422
|
+
},
|
|
10423
|
+
"execution_error": {
|
|
10424
|
+
"name": "execution_error",
|
|
10425
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10426
10426
|
"arguments": {}
|
|
10427
10427
|
}
|
|
10428
10428
|
},
|
|
@@ -10432,7 +10432,7 @@ setBuildTimeData('clients.elevenlabs', {
|
|
|
10432
10432
|
"examples": [
|
|
10433
10433
|
{
|
|
10434
10434
|
"language": "ts",
|
|
10435
|
-
"code": "const
|
|
10435
|
+
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
10436
10436
|
}
|
|
10437
10437
|
]
|
|
10438
10438
|
});
|
|
@@ -21109,53 +21109,254 @@ export const introspectionData = [
|
|
|
21109
21109
|
"type": "string",
|
|
21110
21110
|
"description": "The channel name to remove"
|
|
21111
21111
|
}
|
|
21112
|
-
},
|
|
21113
|
-
"required": [
|
|
21114
|
-
"channelName"
|
|
21115
|
-
],
|
|
21116
|
-
"returns": "void"
|
|
21117
|
-
},
|
|
21118
|
-
"unsubscribeAll": {
|
|
21119
|
-
"description": "Unsubscribe and remove all realtime channels.",
|
|
21120
|
-
"parameters": {},
|
|
21121
|
-
"required": [],
|
|
21122
|
-
"returns": "void"
|
|
21123
|
-
},
|
|
21124
|
-
"connect": {
|
|
21125
|
-
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
21126
|
-
"parameters": {},
|
|
21127
|
-
"required": [],
|
|
21128
|
-
"returns": "void"
|
|
21129
|
-
},
|
|
21130
|
-
"disconnect": {
|
|
21131
|
-
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
21132
|
-
"parameters": {},
|
|
21133
|
-
"required": [],
|
|
21134
|
-
"returns": "void"
|
|
21112
|
+
},
|
|
21113
|
+
"required": [
|
|
21114
|
+
"channelName"
|
|
21115
|
+
],
|
|
21116
|
+
"returns": "void"
|
|
21117
|
+
},
|
|
21118
|
+
"unsubscribeAll": {
|
|
21119
|
+
"description": "Unsubscribe and remove all realtime channels.",
|
|
21120
|
+
"parameters": {},
|
|
21121
|
+
"required": [],
|
|
21122
|
+
"returns": "void"
|
|
21123
|
+
},
|
|
21124
|
+
"connect": {
|
|
21125
|
+
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
21126
|
+
"parameters": {},
|
|
21127
|
+
"required": [],
|
|
21128
|
+
"returns": "void"
|
|
21129
|
+
},
|
|
21130
|
+
"disconnect": {
|
|
21131
|
+
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
21132
|
+
"parameters": {},
|
|
21133
|
+
"required": [],
|
|
21134
|
+
"returns": "void"
|
|
21135
|
+
}
|
|
21136
|
+
},
|
|
21137
|
+
"getters": {
|
|
21138
|
+
"sdk": {
|
|
21139
|
+
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
21140
|
+
"returns": "SupabaseSDKClient<any, any>"
|
|
21141
|
+
},
|
|
21142
|
+
"storage": {
|
|
21143
|
+
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
21144
|
+
"returns": "any"
|
|
21145
|
+
},
|
|
21146
|
+
"functions": {
|
|
21147
|
+
"description": "Returns the Supabase Functions client.",
|
|
21148
|
+
"returns": "any"
|
|
21149
|
+
}
|
|
21150
|
+
},
|
|
21151
|
+
"events": {},
|
|
21152
|
+
"state": {},
|
|
21153
|
+
"options": {},
|
|
21154
|
+
"envVars": [],
|
|
21155
|
+
"examples": [
|
|
21156
|
+
{
|
|
21157
|
+
"language": "ts",
|
|
21158
|
+
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
21159
|
+
}
|
|
21160
|
+
]
|
|
21161
|
+
},
|
|
21162
|
+
{
|
|
21163
|
+
"id": "clients.elevenlabs",
|
|
21164
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
21165
|
+
"shortcut": "clients.elevenlabs",
|
|
21166
|
+
"className": "ElevenLabsClient",
|
|
21167
|
+
"methods": {
|
|
21168
|
+
"beforeRequest": {
|
|
21169
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
21170
|
+
"parameters": {},
|
|
21171
|
+
"required": [],
|
|
21172
|
+
"returns": "void"
|
|
21173
|
+
},
|
|
21174
|
+
"connect": {
|
|
21175
|
+
"description": "Validate the API key by listing available models.",
|
|
21176
|
+
"parameters": {},
|
|
21177
|
+
"required": [],
|
|
21178
|
+
"returns": "Promise<this>",
|
|
21179
|
+
"examples": [
|
|
21180
|
+
{
|
|
21181
|
+
"language": "ts",
|
|
21182
|
+
"code": "await el.connect()"
|
|
21183
|
+
}
|
|
21184
|
+
]
|
|
21185
|
+
},
|
|
21186
|
+
"listVoices": {
|
|
21187
|
+
"description": "List available voices with optional search and filtering.",
|
|
21188
|
+
"parameters": {
|
|
21189
|
+
"options": {
|
|
21190
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
21191
|
+
"description": "Query parameters for filtering voices"
|
|
21192
|
+
}
|
|
21193
|
+
},
|
|
21194
|
+
"required": [],
|
|
21195
|
+
"returns": "Promise<any>",
|
|
21196
|
+
"examples": [
|
|
21197
|
+
{
|
|
21198
|
+
"language": "ts",
|
|
21199
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
21200
|
+
}
|
|
21201
|
+
]
|
|
21202
|
+
},
|
|
21203
|
+
"getVoice": {
|
|
21204
|
+
"description": "Get details for a single voice.",
|
|
21205
|
+
"parameters": {
|
|
21206
|
+
"voiceId": {
|
|
21207
|
+
"type": "string",
|
|
21208
|
+
"description": "The voice ID to look up"
|
|
21209
|
+
}
|
|
21210
|
+
},
|
|
21211
|
+
"required": [
|
|
21212
|
+
"voiceId"
|
|
21213
|
+
],
|
|
21214
|
+
"returns": "Promise<any>",
|
|
21215
|
+
"examples": [
|
|
21216
|
+
{
|
|
21217
|
+
"language": "ts",
|
|
21218
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
21219
|
+
}
|
|
21220
|
+
]
|
|
21221
|
+
},
|
|
21222
|
+
"listModels": {
|
|
21223
|
+
"description": "List available TTS models.",
|
|
21224
|
+
"parameters": {},
|
|
21225
|
+
"required": [],
|
|
21226
|
+
"returns": "Promise<any[]>",
|
|
21227
|
+
"examples": [
|
|
21228
|
+
{
|
|
21229
|
+
"language": "ts",
|
|
21230
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
21231
|
+
}
|
|
21232
|
+
]
|
|
21233
|
+
},
|
|
21234
|
+
"synthesize": {
|
|
21235
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
21236
|
+
"parameters": {
|
|
21237
|
+
"text": {
|
|
21238
|
+
"type": "string",
|
|
21239
|
+
"description": "The text to convert to speech"
|
|
21240
|
+
},
|
|
21241
|
+
"options": {
|
|
21242
|
+
"type": "SynthesizeOptions",
|
|
21243
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
21244
|
+
"properties": {
|
|
21245
|
+
"voiceId": {
|
|
21246
|
+
"type": "string",
|
|
21247
|
+
"description": ""
|
|
21248
|
+
},
|
|
21249
|
+
"modelId": {
|
|
21250
|
+
"type": "string",
|
|
21251
|
+
"description": ""
|
|
21252
|
+
},
|
|
21253
|
+
"outputFormat": {
|
|
21254
|
+
"type": "string",
|
|
21255
|
+
"description": ""
|
|
21256
|
+
},
|
|
21257
|
+
"voiceSettings": {
|
|
21258
|
+
"type": "ElevenLabsVoiceSettings",
|
|
21259
|
+
"description": ""
|
|
21260
|
+
},
|
|
21261
|
+
"disableCache": {
|
|
21262
|
+
"type": "boolean",
|
|
21263
|
+
"description": ""
|
|
21264
|
+
}
|
|
21265
|
+
}
|
|
21266
|
+
}
|
|
21267
|
+
},
|
|
21268
|
+
"required": [
|
|
21269
|
+
"text"
|
|
21270
|
+
],
|
|
21271
|
+
"returns": "Promise<Buffer>",
|
|
21272
|
+
"examples": [
|
|
21273
|
+
{
|
|
21274
|
+
"language": "ts",
|
|
21275
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
21276
|
+
}
|
|
21277
|
+
]
|
|
21278
|
+
},
|
|
21279
|
+
"say": {
|
|
21280
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
21281
|
+
"parameters": {
|
|
21282
|
+
"text": {
|
|
21283
|
+
"type": "string",
|
|
21284
|
+
"description": "The text to convert to speech"
|
|
21285
|
+
},
|
|
21286
|
+
"outputPath": {
|
|
21287
|
+
"type": "string",
|
|
21288
|
+
"description": "File path to write the audio to"
|
|
21289
|
+
},
|
|
21290
|
+
"options": {
|
|
21291
|
+
"type": "SynthesizeOptions",
|
|
21292
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
21293
|
+
"properties": {
|
|
21294
|
+
"voiceId": {
|
|
21295
|
+
"type": "string",
|
|
21296
|
+
"description": ""
|
|
21297
|
+
},
|
|
21298
|
+
"modelId": {
|
|
21299
|
+
"type": "string",
|
|
21300
|
+
"description": ""
|
|
21301
|
+
},
|
|
21302
|
+
"outputFormat": {
|
|
21303
|
+
"type": "string",
|
|
21304
|
+
"description": ""
|
|
21305
|
+
},
|
|
21306
|
+
"voiceSettings": {
|
|
21307
|
+
"type": "ElevenLabsVoiceSettings",
|
|
21308
|
+
"description": ""
|
|
21309
|
+
},
|
|
21310
|
+
"disableCache": {
|
|
21311
|
+
"type": "boolean",
|
|
21312
|
+
"description": ""
|
|
21313
|
+
}
|
|
21314
|
+
}
|
|
21315
|
+
}
|
|
21316
|
+
},
|
|
21317
|
+
"required": [
|
|
21318
|
+
"text",
|
|
21319
|
+
"outputPath"
|
|
21320
|
+
],
|
|
21321
|
+
"returns": "Promise<string>",
|
|
21322
|
+
"examples": [
|
|
21323
|
+
{
|
|
21324
|
+
"language": "ts",
|
|
21325
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
21326
|
+
}
|
|
21327
|
+
]
|
|
21135
21328
|
}
|
|
21136
21329
|
},
|
|
21137
21330
|
"getters": {
|
|
21138
|
-
"
|
|
21139
|
-
"description": "
|
|
21140
|
-
"returns": "
|
|
21331
|
+
"apiKey": {
|
|
21332
|
+
"description": "The resolved API key from options or environment.",
|
|
21333
|
+
"returns": "string"
|
|
21334
|
+
}
|
|
21335
|
+
},
|
|
21336
|
+
"events": {
|
|
21337
|
+
"failure": {
|
|
21338
|
+
"name": "failure",
|
|
21339
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21340
|
+
"arguments": {}
|
|
21141
21341
|
},
|
|
21142
|
-
"
|
|
21143
|
-
"
|
|
21144
|
-
"
|
|
21342
|
+
"voices": {
|
|
21343
|
+
"name": "voices",
|
|
21344
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21345
|
+
"arguments": {}
|
|
21145
21346
|
},
|
|
21146
|
-
"
|
|
21147
|
-
"
|
|
21148
|
-
"
|
|
21347
|
+
"speech": {
|
|
21348
|
+
"name": "speech",
|
|
21349
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21350
|
+
"arguments": {}
|
|
21149
21351
|
}
|
|
21150
21352
|
},
|
|
21151
|
-
"events": {},
|
|
21152
21353
|
"state": {},
|
|
21153
21354
|
"options": {},
|
|
21154
21355
|
"envVars": [],
|
|
21155
21356
|
"examples": [
|
|
21156
21357
|
{
|
|
21157
21358
|
"language": "ts",
|
|
21158
|
-
"code": "const
|
|
21359
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
21159
21360
|
}
|
|
21160
21361
|
]
|
|
21161
21362
|
},
|
|
@@ -21410,207 +21611,6 @@ export const introspectionData = [
|
|
|
21410
21611
|
}
|
|
21411
21612
|
]
|
|
21412
21613
|
},
|
|
21413
|
-
{
|
|
21414
|
-
"id": "clients.elevenlabs",
|
|
21415
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
21416
|
-
"shortcut": "clients.elevenlabs",
|
|
21417
|
-
"className": "ElevenLabsClient",
|
|
21418
|
-
"methods": {
|
|
21419
|
-
"beforeRequest": {
|
|
21420
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
21421
|
-
"parameters": {},
|
|
21422
|
-
"required": [],
|
|
21423
|
-
"returns": "void"
|
|
21424
|
-
},
|
|
21425
|
-
"connect": {
|
|
21426
|
-
"description": "Validate the API key by listing available models.",
|
|
21427
|
-
"parameters": {},
|
|
21428
|
-
"required": [],
|
|
21429
|
-
"returns": "Promise<this>",
|
|
21430
|
-
"examples": [
|
|
21431
|
-
{
|
|
21432
|
-
"language": "ts",
|
|
21433
|
-
"code": "await el.connect()"
|
|
21434
|
-
}
|
|
21435
|
-
]
|
|
21436
|
-
},
|
|
21437
|
-
"listVoices": {
|
|
21438
|
-
"description": "List available voices with optional search and filtering.",
|
|
21439
|
-
"parameters": {
|
|
21440
|
-
"options": {
|
|
21441
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
21442
|
-
"description": "Query parameters for filtering voices"
|
|
21443
|
-
}
|
|
21444
|
-
},
|
|
21445
|
-
"required": [],
|
|
21446
|
-
"returns": "Promise<any>",
|
|
21447
|
-
"examples": [
|
|
21448
|
-
{
|
|
21449
|
-
"language": "ts",
|
|
21450
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
21451
|
-
}
|
|
21452
|
-
]
|
|
21453
|
-
},
|
|
21454
|
-
"getVoice": {
|
|
21455
|
-
"description": "Get details for a single voice.",
|
|
21456
|
-
"parameters": {
|
|
21457
|
-
"voiceId": {
|
|
21458
|
-
"type": "string",
|
|
21459
|
-
"description": "The voice ID to look up"
|
|
21460
|
-
}
|
|
21461
|
-
},
|
|
21462
|
-
"required": [
|
|
21463
|
-
"voiceId"
|
|
21464
|
-
],
|
|
21465
|
-
"returns": "Promise<any>",
|
|
21466
|
-
"examples": [
|
|
21467
|
-
{
|
|
21468
|
-
"language": "ts",
|
|
21469
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
21470
|
-
}
|
|
21471
|
-
]
|
|
21472
|
-
},
|
|
21473
|
-
"listModels": {
|
|
21474
|
-
"description": "List available TTS models.",
|
|
21475
|
-
"parameters": {},
|
|
21476
|
-
"required": [],
|
|
21477
|
-
"returns": "Promise<any[]>",
|
|
21478
|
-
"examples": [
|
|
21479
|
-
{
|
|
21480
|
-
"language": "ts",
|
|
21481
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
21482
|
-
}
|
|
21483
|
-
]
|
|
21484
|
-
},
|
|
21485
|
-
"synthesize": {
|
|
21486
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
21487
|
-
"parameters": {
|
|
21488
|
-
"text": {
|
|
21489
|
-
"type": "string",
|
|
21490
|
-
"description": "The text to convert to speech"
|
|
21491
|
-
},
|
|
21492
|
-
"options": {
|
|
21493
|
-
"type": "SynthesizeOptions",
|
|
21494
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
21495
|
-
"properties": {
|
|
21496
|
-
"voiceId": {
|
|
21497
|
-
"type": "string",
|
|
21498
|
-
"description": ""
|
|
21499
|
-
},
|
|
21500
|
-
"modelId": {
|
|
21501
|
-
"type": "string",
|
|
21502
|
-
"description": ""
|
|
21503
|
-
},
|
|
21504
|
-
"outputFormat": {
|
|
21505
|
-
"type": "string",
|
|
21506
|
-
"description": ""
|
|
21507
|
-
},
|
|
21508
|
-
"voiceSettings": {
|
|
21509
|
-
"type": "ElevenLabsVoiceSettings",
|
|
21510
|
-
"description": ""
|
|
21511
|
-
},
|
|
21512
|
-
"disableCache": {
|
|
21513
|
-
"type": "boolean",
|
|
21514
|
-
"description": ""
|
|
21515
|
-
}
|
|
21516
|
-
}
|
|
21517
|
-
}
|
|
21518
|
-
},
|
|
21519
|
-
"required": [
|
|
21520
|
-
"text"
|
|
21521
|
-
],
|
|
21522
|
-
"returns": "Promise<Buffer>",
|
|
21523
|
-
"examples": [
|
|
21524
|
-
{
|
|
21525
|
-
"language": "ts",
|
|
21526
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
21527
|
-
}
|
|
21528
|
-
]
|
|
21529
|
-
},
|
|
21530
|
-
"say": {
|
|
21531
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
21532
|
-
"parameters": {
|
|
21533
|
-
"text": {
|
|
21534
|
-
"type": "string",
|
|
21535
|
-
"description": "The text to convert to speech"
|
|
21536
|
-
},
|
|
21537
|
-
"outputPath": {
|
|
21538
|
-
"type": "string",
|
|
21539
|
-
"description": "File path to write the audio to"
|
|
21540
|
-
},
|
|
21541
|
-
"options": {
|
|
21542
|
-
"type": "SynthesizeOptions",
|
|
21543
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
21544
|
-
"properties": {
|
|
21545
|
-
"voiceId": {
|
|
21546
|
-
"type": "string",
|
|
21547
|
-
"description": ""
|
|
21548
|
-
},
|
|
21549
|
-
"modelId": {
|
|
21550
|
-
"type": "string",
|
|
21551
|
-
"description": ""
|
|
21552
|
-
},
|
|
21553
|
-
"outputFormat": {
|
|
21554
|
-
"type": "string",
|
|
21555
|
-
"description": ""
|
|
21556
|
-
},
|
|
21557
|
-
"voiceSettings": {
|
|
21558
|
-
"type": "ElevenLabsVoiceSettings",
|
|
21559
|
-
"description": ""
|
|
21560
|
-
},
|
|
21561
|
-
"disableCache": {
|
|
21562
|
-
"type": "boolean",
|
|
21563
|
-
"description": ""
|
|
21564
|
-
}
|
|
21565
|
-
}
|
|
21566
|
-
}
|
|
21567
|
-
},
|
|
21568
|
-
"required": [
|
|
21569
|
-
"text",
|
|
21570
|
-
"outputPath"
|
|
21571
|
-
],
|
|
21572
|
-
"returns": "Promise<string>",
|
|
21573
|
-
"examples": [
|
|
21574
|
-
{
|
|
21575
|
-
"language": "ts",
|
|
21576
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
21577
|
-
}
|
|
21578
|
-
]
|
|
21579
|
-
}
|
|
21580
|
-
},
|
|
21581
|
-
"getters": {
|
|
21582
|
-
"apiKey": {
|
|
21583
|
-
"description": "The resolved API key from options or environment.",
|
|
21584
|
-
"returns": "string"
|
|
21585
|
-
}
|
|
21586
|
-
},
|
|
21587
|
-
"events": {
|
|
21588
|
-
"failure": {
|
|
21589
|
-
"name": "failure",
|
|
21590
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
21591
|
-
"arguments": {}
|
|
21592
|
-
},
|
|
21593
|
-
"voices": {
|
|
21594
|
-
"name": "voices",
|
|
21595
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
21596
|
-
"arguments": {}
|
|
21597
|
-
},
|
|
21598
|
-
"speech": {
|
|
21599
|
-
"name": "speech",
|
|
21600
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
21601
|
-
"arguments": {}
|
|
21602
|
-
}
|
|
21603
|
-
},
|
|
21604
|
-
"state": {},
|
|
21605
|
-
"options": {},
|
|
21606
|
-
"envVars": [],
|
|
21607
|
-
"examples": [
|
|
21608
|
-
{
|
|
21609
|
-
"language": "ts",
|
|
21610
|
-
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
21611
|
-
}
|
|
21612
|
-
]
|
|
21613
|
-
},
|
|
21614
21614
|
{
|
|
21615
21615
|
"id": "servers.mcp",
|
|
21616
21616
|
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|