@soederpop/luca 0.0.11 → 0.0.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +8 -3
- package/src/agi/features/claude-code.ts +7 -0
- package/src/bootstrap/generated.ts +1 -1
- package/src/commands/prompt.ts +3 -0
- package/src/introspection/generated.agi.ts +1002 -986
- package/src/introspection/generated.node.ts +469 -469
- package/src/introspection/generated.web.ts +1 -1
- package/src/scaffolds/generated.ts +1 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
|
|
2
2
|
|
|
3
3
|
// Auto-generated introspection registry data
|
|
4
|
-
// Generated at: 2026-03-
|
|
4
|
+
// Generated at: 2026-03-20T03:57:55.740Z
|
|
5
5
|
|
|
6
6
|
setBuildTimeData('features.googleDocs', {
|
|
7
7
|
"id": "features.googleDocs",
|
|
@@ -9983,208 +9983,6 @@ setBuildTimeData('clients.supabase', {
|
|
|
9983
9983
|
]
|
|
9984
9984
|
});
|
|
9985
9985
|
|
|
9986
|
-
setBuildTimeData('clients.elevenlabs', {
|
|
9987
|
-
"id": "clients.elevenlabs",
|
|
9988
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
9989
|
-
"shortcut": "clients.elevenlabs",
|
|
9990
|
-
"className": "ElevenLabsClient",
|
|
9991
|
-
"methods": {
|
|
9992
|
-
"beforeRequest": {
|
|
9993
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
9994
|
-
"parameters": {},
|
|
9995
|
-
"required": [],
|
|
9996
|
-
"returns": "void"
|
|
9997
|
-
},
|
|
9998
|
-
"connect": {
|
|
9999
|
-
"description": "Validate the API key by listing available models.",
|
|
10000
|
-
"parameters": {},
|
|
10001
|
-
"required": [],
|
|
10002
|
-
"returns": "Promise<this>",
|
|
10003
|
-
"examples": [
|
|
10004
|
-
{
|
|
10005
|
-
"language": "ts",
|
|
10006
|
-
"code": "await el.connect()"
|
|
10007
|
-
}
|
|
10008
|
-
]
|
|
10009
|
-
},
|
|
10010
|
-
"listVoices": {
|
|
10011
|
-
"description": "List available voices with optional search and filtering.",
|
|
10012
|
-
"parameters": {
|
|
10013
|
-
"options": {
|
|
10014
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10015
|
-
"description": "Query parameters for filtering voices"
|
|
10016
|
-
}
|
|
10017
|
-
},
|
|
10018
|
-
"required": [],
|
|
10019
|
-
"returns": "Promise<any>",
|
|
10020
|
-
"examples": [
|
|
10021
|
-
{
|
|
10022
|
-
"language": "ts",
|
|
10023
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10024
|
-
}
|
|
10025
|
-
]
|
|
10026
|
-
},
|
|
10027
|
-
"getVoice": {
|
|
10028
|
-
"description": "Get details for a single voice.",
|
|
10029
|
-
"parameters": {
|
|
10030
|
-
"voiceId": {
|
|
10031
|
-
"type": "string",
|
|
10032
|
-
"description": "The voice ID to look up"
|
|
10033
|
-
}
|
|
10034
|
-
},
|
|
10035
|
-
"required": [
|
|
10036
|
-
"voiceId"
|
|
10037
|
-
],
|
|
10038
|
-
"returns": "Promise<any>",
|
|
10039
|
-
"examples": [
|
|
10040
|
-
{
|
|
10041
|
-
"language": "ts",
|
|
10042
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10043
|
-
}
|
|
10044
|
-
]
|
|
10045
|
-
},
|
|
10046
|
-
"listModels": {
|
|
10047
|
-
"description": "List available TTS models.",
|
|
10048
|
-
"parameters": {},
|
|
10049
|
-
"required": [],
|
|
10050
|
-
"returns": "Promise<any[]>",
|
|
10051
|
-
"examples": [
|
|
10052
|
-
{
|
|
10053
|
-
"language": "ts",
|
|
10054
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10055
|
-
}
|
|
10056
|
-
]
|
|
10057
|
-
},
|
|
10058
|
-
"synthesize": {
|
|
10059
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10060
|
-
"parameters": {
|
|
10061
|
-
"text": {
|
|
10062
|
-
"type": "string",
|
|
10063
|
-
"description": "The text to convert to speech"
|
|
10064
|
-
},
|
|
10065
|
-
"options": {
|
|
10066
|
-
"type": "SynthesizeOptions",
|
|
10067
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10068
|
-
"properties": {
|
|
10069
|
-
"voiceId": {
|
|
10070
|
-
"type": "string",
|
|
10071
|
-
"description": ""
|
|
10072
|
-
},
|
|
10073
|
-
"modelId": {
|
|
10074
|
-
"type": "string",
|
|
10075
|
-
"description": ""
|
|
10076
|
-
},
|
|
10077
|
-
"outputFormat": {
|
|
10078
|
-
"type": "string",
|
|
10079
|
-
"description": ""
|
|
10080
|
-
},
|
|
10081
|
-
"voiceSettings": {
|
|
10082
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10083
|
-
"description": ""
|
|
10084
|
-
},
|
|
10085
|
-
"disableCache": {
|
|
10086
|
-
"type": "boolean",
|
|
10087
|
-
"description": ""
|
|
10088
|
-
}
|
|
10089
|
-
}
|
|
10090
|
-
}
|
|
10091
|
-
},
|
|
10092
|
-
"required": [
|
|
10093
|
-
"text"
|
|
10094
|
-
],
|
|
10095
|
-
"returns": "Promise<Buffer>",
|
|
10096
|
-
"examples": [
|
|
10097
|
-
{
|
|
10098
|
-
"language": "ts",
|
|
10099
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10100
|
-
}
|
|
10101
|
-
]
|
|
10102
|
-
},
|
|
10103
|
-
"say": {
|
|
10104
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
10105
|
-
"parameters": {
|
|
10106
|
-
"text": {
|
|
10107
|
-
"type": "string",
|
|
10108
|
-
"description": "The text to convert to speech"
|
|
10109
|
-
},
|
|
10110
|
-
"outputPath": {
|
|
10111
|
-
"type": "string",
|
|
10112
|
-
"description": "File path to write the audio to"
|
|
10113
|
-
},
|
|
10114
|
-
"options": {
|
|
10115
|
-
"type": "SynthesizeOptions",
|
|
10116
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10117
|
-
"properties": {
|
|
10118
|
-
"voiceId": {
|
|
10119
|
-
"type": "string",
|
|
10120
|
-
"description": ""
|
|
10121
|
-
},
|
|
10122
|
-
"modelId": {
|
|
10123
|
-
"type": "string",
|
|
10124
|
-
"description": ""
|
|
10125
|
-
},
|
|
10126
|
-
"outputFormat": {
|
|
10127
|
-
"type": "string",
|
|
10128
|
-
"description": ""
|
|
10129
|
-
},
|
|
10130
|
-
"voiceSettings": {
|
|
10131
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10132
|
-
"description": ""
|
|
10133
|
-
},
|
|
10134
|
-
"disableCache": {
|
|
10135
|
-
"type": "boolean",
|
|
10136
|
-
"description": ""
|
|
10137
|
-
}
|
|
10138
|
-
}
|
|
10139
|
-
}
|
|
10140
|
-
},
|
|
10141
|
-
"required": [
|
|
10142
|
-
"text",
|
|
10143
|
-
"outputPath"
|
|
10144
|
-
],
|
|
10145
|
-
"returns": "Promise<string>",
|
|
10146
|
-
"examples": [
|
|
10147
|
-
{
|
|
10148
|
-
"language": "ts",
|
|
10149
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10150
|
-
}
|
|
10151
|
-
]
|
|
10152
|
-
}
|
|
10153
|
-
},
|
|
10154
|
-
"getters": {
|
|
10155
|
-
"apiKey": {
|
|
10156
|
-
"description": "The resolved API key from options or environment.",
|
|
10157
|
-
"returns": "string"
|
|
10158
|
-
}
|
|
10159
|
-
},
|
|
10160
|
-
"events": {
|
|
10161
|
-
"failure": {
|
|
10162
|
-
"name": "failure",
|
|
10163
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10164
|
-
"arguments": {}
|
|
10165
|
-
},
|
|
10166
|
-
"voices": {
|
|
10167
|
-
"name": "voices",
|
|
10168
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10169
|
-
"arguments": {}
|
|
10170
|
-
},
|
|
10171
|
-
"speech": {
|
|
10172
|
-
"name": "speech",
|
|
10173
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10174
|
-
"arguments": {}
|
|
10175
|
-
}
|
|
10176
|
-
},
|
|
10177
|
-
"state": {},
|
|
10178
|
-
"options": {},
|
|
10179
|
-
"envVars": [],
|
|
10180
|
-
"examples": [
|
|
10181
|
-
{
|
|
10182
|
-
"language": "ts",
|
|
10183
|
-
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10184
|
-
}
|
|
10185
|
-
]
|
|
10186
|
-
});
|
|
10187
|
-
|
|
10188
9986
|
setBuildTimeData('clients.comfyui', {
|
|
10189
9987
|
"id": "clients.comfyui",
|
|
10190
9988
|
"description": "ComfyUI client — execute Stable Diffusion workflows via the ComfyUI API. Connects to a ComfyUI instance to queue prompts, track execution via WebSocket or polling, and download generated images. Supports both UI-format and API-format workflows with automatic conversion.",
|
|
@@ -10390,39 +10188,241 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10390
10188
|
}
|
|
10391
10189
|
},
|
|
10392
10190
|
"events": {
|
|
10393
|
-
"execution_start": {
|
|
10394
|
-
"name": "execution_start",
|
|
10395
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10396
|
-
"arguments": {}
|
|
10397
|
-
},
|
|
10398
|
-
"execution_complete": {
|
|
10399
|
-
"name": "execution_complete",
|
|
10400
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10401
|
-
"arguments": {}
|
|
10402
|
-
},
|
|
10403
|
-
"executing": {
|
|
10404
|
-
"name": "executing",
|
|
10405
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10406
|
-
"arguments": {}
|
|
10407
|
-
},
|
|
10408
|
-
"progress": {
|
|
10409
|
-
"name": "progress",
|
|
10410
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10411
|
-
"arguments": {}
|
|
10412
|
-
},
|
|
10413
|
-
"executed": {
|
|
10414
|
-
"name": "executed",
|
|
10415
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10191
|
+
"execution_start": {
|
|
10192
|
+
"name": "execution_start",
|
|
10193
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10194
|
+
"arguments": {}
|
|
10195
|
+
},
|
|
10196
|
+
"execution_complete": {
|
|
10197
|
+
"name": "execution_complete",
|
|
10198
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10199
|
+
"arguments": {}
|
|
10200
|
+
},
|
|
10201
|
+
"executing": {
|
|
10202
|
+
"name": "executing",
|
|
10203
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10204
|
+
"arguments": {}
|
|
10205
|
+
},
|
|
10206
|
+
"progress": {
|
|
10207
|
+
"name": "progress",
|
|
10208
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10209
|
+
"arguments": {}
|
|
10210
|
+
},
|
|
10211
|
+
"executed": {
|
|
10212
|
+
"name": "executed",
|
|
10213
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10214
|
+
"arguments": {}
|
|
10215
|
+
},
|
|
10216
|
+
"execution_cached": {
|
|
10217
|
+
"name": "execution_cached",
|
|
10218
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10219
|
+
"arguments": {}
|
|
10220
|
+
},
|
|
10221
|
+
"execution_error": {
|
|
10222
|
+
"name": "execution_error",
|
|
10223
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10224
|
+
"arguments": {}
|
|
10225
|
+
}
|
|
10226
|
+
},
|
|
10227
|
+
"state": {},
|
|
10228
|
+
"options": {},
|
|
10229
|
+
"envVars": [],
|
|
10230
|
+
"examples": [
|
|
10231
|
+
{
|
|
10232
|
+
"language": "ts",
|
|
10233
|
+
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
10234
|
+
}
|
|
10235
|
+
]
|
|
10236
|
+
});
|
|
10237
|
+
|
|
10238
|
+
setBuildTimeData('clients.elevenlabs', {
|
|
10239
|
+
"id": "clients.elevenlabs",
|
|
10240
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
10241
|
+
"shortcut": "clients.elevenlabs",
|
|
10242
|
+
"className": "ElevenLabsClient",
|
|
10243
|
+
"methods": {
|
|
10244
|
+
"beforeRequest": {
|
|
10245
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
10246
|
+
"parameters": {},
|
|
10247
|
+
"required": [],
|
|
10248
|
+
"returns": "void"
|
|
10249
|
+
},
|
|
10250
|
+
"connect": {
|
|
10251
|
+
"description": "Validate the API key by listing available models.",
|
|
10252
|
+
"parameters": {},
|
|
10253
|
+
"required": [],
|
|
10254
|
+
"returns": "Promise<this>",
|
|
10255
|
+
"examples": [
|
|
10256
|
+
{
|
|
10257
|
+
"language": "ts",
|
|
10258
|
+
"code": "await el.connect()"
|
|
10259
|
+
}
|
|
10260
|
+
]
|
|
10261
|
+
},
|
|
10262
|
+
"listVoices": {
|
|
10263
|
+
"description": "List available voices with optional search and filtering.",
|
|
10264
|
+
"parameters": {
|
|
10265
|
+
"options": {
|
|
10266
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10267
|
+
"description": "Query parameters for filtering voices"
|
|
10268
|
+
}
|
|
10269
|
+
},
|
|
10270
|
+
"required": [],
|
|
10271
|
+
"returns": "Promise<any>",
|
|
10272
|
+
"examples": [
|
|
10273
|
+
{
|
|
10274
|
+
"language": "ts",
|
|
10275
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10276
|
+
}
|
|
10277
|
+
]
|
|
10278
|
+
},
|
|
10279
|
+
"getVoice": {
|
|
10280
|
+
"description": "Get details for a single voice.",
|
|
10281
|
+
"parameters": {
|
|
10282
|
+
"voiceId": {
|
|
10283
|
+
"type": "string",
|
|
10284
|
+
"description": "The voice ID to look up"
|
|
10285
|
+
}
|
|
10286
|
+
},
|
|
10287
|
+
"required": [
|
|
10288
|
+
"voiceId"
|
|
10289
|
+
],
|
|
10290
|
+
"returns": "Promise<any>",
|
|
10291
|
+
"examples": [
|
|
10292
|
+
{
|
|
10293
|
+
"language": "ts",
|
|
10294
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10295
|
+
}
|
|
10296
|
+
]
|
|
10297
|
+
},
|
|
10298
|
+
"listModels": {
|
|
10299
|
+
"description": "List available TTS models.",
|
|
10300
|
+
"parameters": {},
|
|
10301
|
+
"required": [],
|
|
10302
|
+
"returns": "Promise<any[]>",
|
|
10303
|
+
"examples": [
|
|
10304
|
+
{
|
|
10305
|
+
"language": "ts",
|
|
10306
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10307
|
+
}
|
|
10308
|
+
]
|
|
10309
|
+
},
|
|
10310
|
+
"synthesize": {
|
|
10311
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10312
|
+
"parameters": {
|
|
10313
|
+
"text": {
|
|
10314
|
+
"type": "string",
|
|
10315
|
+
"description": "The text to convert to speech"
|
|
10316
|
+
},
|
|
10317
|
+
"options": {
|
|
10318
|
+
"type": "SynthesizeOptions",
|
|
10319
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10320
|
+
"properties": {
|
|
10321
|
+
"voiceId": {
|
|
10322
|
+
"type": "string",
|
|
10323
|
+
"description": ""
|
|
10324
|
+
},
|
|
10325
|
+
"modelId": {
|
|
10326
|
+
"type": "string",
|
|
10327
|
+
"description": ""
|
|
10328
|
+
},
|
|
10329
|
+
"outputFormat": {
|
|
10330
|
+
"type": "string",
|
|
10331
|
+
"description": ""
|
|
10332
|
+
},
|
|
10333
|
+
"voiceSettings": {
|
|
10334
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10335
|
+
"description": ""
|
|
10336
|
+
},
|
|
10337
|
+
"disableCache": {
|
|
10338
|
+
"type": "boolean",
|
|
10339
|
+
"description": ""
|
|
10340
|
+
}
|
|
10341
|
+
}
|
|
10342
|
+
}
|
|
10343
|
+
},
|
|
10344
|
+
"required": [
|
|
10345
|
+
"text"
|
|
10346
|
+
],
|
|
10347
|
+
"returns": "Promise<Buffer>",
|
|
10348
|
+
"examples": [
|
|
10349
|
+
{
|
|
10350
|
+
"language": "ts",
|
|
10351
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10352
|
+
}
|
|
10353
|
+
]
|
|
10354
|
+
},
|
|
10355
|
+
"say": {
|
|
10356
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
10357
|
+
"parameters": {
|
|
10358
|
+
"text": {
|
|
10359
|
+
"type": "string",
|
|
10360
|
+
"description": "The text to convert to speech"
|
|
10361
|
+
},
|
|
10362
|
+
"outputPath": {
|
|
10363
|
+
"type": "string",
|
|
10364
|
+
"description": "File path to write the audio to"
|
|
10365
|
+
},
|
|
10366
|
+
"options": {
|
|
10367
|
+
"type": "SynthesizeOptions",
|
|
10368
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10369
|
+
"properties": {
|
|
10370
|
+
"voiceId": {
|
|
10371
|
+
"type": "string",
|
|
10372
|
+
"description": ""
|
|
10373
|
+
},
|
|
10374
|
+
"modelId": {
|
|
10375
|
+
"type": "string",
|
|
10376
|
+
"description": ""
|
|
10377
|
+
},
|
|
10378
|
+
"outputFormat": {
|
|
10379
|
+
"type": "string",
|
|
10380
|
+
"description": ""
|
|
10381
|
+
},
|
|
10382
|
+
"voiceSettings": {
|
|
10383
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10384
|
+
"description": ""
|
|
10385
|
+
},
|
|
10386
|
+
"disableCache": {
|
|
10387
|
+
"type": "boolean",
|
|
10388
|
+
"description": ""
|
|
10389
|
+
}
|
|
10390
|
+
}
|
|
10391
|
+
}
|
|
10392
|
+
},
|
|
10393
|
+
"required": [
|
|
10394
|
+
"text",
|
|
10395
|
+
"outputPath"
|
|
10396
|
+
],
|
|
10397
|
+
"returns": "Promise<string>",
|
|
10398
|
+
"examples": [
|
|
10399
|
+
{
|
|
10400
|
+
"language": "ts",
|
|
10401
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10402
|
+
}
|
|
10403
|
+
]
|
|
10404
|
+
}
|
|
10405
|
+
},
|
|
10406
|
+
"getters": {
|
|
10407
|
+
"apiKey": {
|
|
10408
|
+
"description": "The resolved API key from options or environment.",
|
|
10409
|
+
"returns": "string"
|
|
10410
|
+
}
|
|
10411
|
+
},
|
|
10412
|
+
"events": {
|
|
10413
|
+
"failure": {
|
|
10414
|
+
"name": "failure",
|
|
10415
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10416
10416
|
"arguments": {}
|
|
10417
10417
|
},
|
|
10418
|
-
"
|
|
10419
|
-
"name": "
|
|
10420
|
-
"description": "Event emitted by
|
|
10418
|
+
"voices": {
|
|
10419
|
+
"name": "voices",
|
|
10420
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10421
10421
|
"arguments": {}
|
|
10422
10422
|
},
|
|
10423
|
-
"
|
|
10424
|
-
"name": "
|
|
10425
|
-
"description": "Event emitted by
|
|
10423
|
+
"speech": {
|
|
10424
|
+
"name": "speech",
|
|
10425
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10426
10426
|
"arguments": {}
|
|
10427
10427
|
}
|
|
10428
10428
|
},
|
|
@@ -10432,7 +10432,7 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10432
10432
|
"examples": [
|
|
10433
10433
|
{
|
|
10434
10434
|
"language": "ts",
|
|
10435
|
-
"code": "const
|
|
10435
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10436
10436
|
}
|
|
10437
10437
|
]
|
|
10438
10438
|
});
|
|
@@ -21107,256 +21107,55 @@ export const introspectionData = [
|
|
|
21107
21107
|
"parameters": {
|
|
21108
21108
|
"channelName": {
|
|
21109
21109
|
"type": "string",
|
|
21110
|
-
"description": "The channel name to remove"
|
|
21111
|
-
}
|
|
21112
|
-
},
|
|
21113
|
-
"required": [
|
|
21114
|
-
"channelName"
|
|
21115
|
-
],
|
|
21116
|
-
"returns": "void"
|
|
21117
|
-
},
|
|
21118
|
-
"unsubscribeAll": {
|
|
21119
|
-
"description": "Unsubscribe and remove all realtime channels.",
|
|
21120
|
-
"parameters": {},
|
|
21121
|
-
"required": [],
|
|
21122
|
-
"returns": "void"
|
|
21123
|
-
},
|
|
21124
|
-
"connect": {
|
|
21125
|
-
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
21126
|
-
"parameters": {},
|
|
21127
|
-
"required": [],
|
|
21128
|
-
"returns": "void"
|
|
21129
|
-
},
|
|
21130
|
-
"disconnect": {
|
|
21131
|
-
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
21132
|
-
"parameters": {},
|
|
21133
|
-
"required": [],
|
|
21134
|
-
"returns": "void"
|
|
21135
|
-
}
|
|
21136
|
-
},
|
|
21137
|
-
"getters": {
|
|
21138
|
-
"sdk": {
|
|
21139
|
-
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
21140
|
-
"returns": "SupabaseSDKClient<any, any>"
|
|
21141
|
-
},
|
|
21142
|
-
"storage": {
|
|
21143
|
-
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
21144
|
-
"returns": "any"
|
|
21145
|
-
},
|
|
21146
|
-
"functions": {
|
|
21147
|
-
"description": "Returns the Supabase Functions client.",
|
|
21148
|
-
"returns": "any"
|
|
21149
|
-
}
|
|
21150
|
-
},
|
|
21151
|
-
"events": {},
|
|
21152
|
-
"state": {},
|
|
21153
|
-
"options": {},
|
|
21154
|
-
"envVars": [],
|
|
21155
|
-
"examples": [
|
|
21156
|
-
{
|
|
21157
|
-
"language": "ts",
|
|
21158
|
-
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
21159
|
-
}
|
|
21160
|
-
]
|
|
21161
|
-
},
|
|
21162
|
-
{
|
|
21163
|
-
"id": "clients.elevenlabs",
|
|
21164
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
21165
|
-
"shortcut": "clients.elevenlabs",
|
|
21166
|
-
"className": "ElevenLabsClient",
|
|
21167
|
-
"methods": {
|
|
21168
|
-
"beforeRequest": {
|
|
21169
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
21170
|
-
"parameters": {},
|
|
21171
|
-
"required": [],
|
|
21172
|
-
"returns": "void"
|
|
21173
|
-
},
|
|
21174
|
-
"connect": {
|
|
21175
|
-
"description": "Validate the API key by listing available models.",
|
|
21176
|
-
"parameters": {},
|
|
21177
|
-
"required": [],
|
|
21178
|
-
"returns": "Promise<this>",
|
|
21179
|
-
"examples": [
|
|
21180
|
-
{
|
|
21181
|
-
"language": "ts",
|
|
21182
|
-
"code": "await el.connect()"
|
|
21183
|
-
}
|
|
21184
|
-
]
|
|
21185
|
-
},
|
|
21186
|
-
"listVoices": {
|
|
21187
|
-
"description": "List available voices with optional search and filtering.",
|
|
21188
|
-
"parameters": {
|
|
21189
|
-
"options": {
|
|
21190
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
21191
|
-
"description": "Query parameters for filtering voices"
|
|
21192
|
-
}
|
|
21193
|
-
},
|
|
21194
|
-
"required": [],
|
|
21195
|
-
"returns": "Promise<any>",
|
|
21196
|
-
"examples": [
|
|
21197
|
-
{
|
|
21198
|
-
"language": "ts",
|
|
21199
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
21200
|
-
}
|
|
21201
|
-
]
|
|
21202
|
-
},
|
|
21203
|
-
"getVoice": {
|
|
21204
|
-
"description": "Get details for a single voice.",
|
|
21205
|
-
"parameters": {
|
|
21206
|
-
"voiceId": {
|
|
21207
|
-
"type": "string",
|
|
21208
|
-
"description": "The voice ID to look up"
|
|
21209
|
-
}
|
|
21210
|
-
},
|
|
21211
|
-
"required": [
|
|
21212
|
-
"voiceId"
|
|
21213
|
-
],
|
|
21214
|
-
"returns": "Promise<any>",
|
|
21215
|
-
"examples": [
|
|
21216
|
-
{
|
|
21217
|
-
"language": "ts",
|
|
21218
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
21219
|
-
}
|
|
21220
|
-
]
|
|
21221
|
-
},
|
|
21222
|
-
"listModels": {
|
|
21223
|
-
"description": "List available TTS models.",
|
|
21224
|
-
"parameters": {},
|
|
21225
|
-
"required": [],
|
|
21226
|
-
"returns": "Promise<any[]>",
|
|
21227
|
-
"examples": [
|
|
21228
|
-
{
|
|
21229
|
-
"language": "ts",
|
|
21230
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
21231
|
-
}
|
|
21232
|
-
]
|
|
21233
|
-
},
|
|
21234
|
-
"synthesize": {
|
|
21235
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
21236
|
-
"parameters": {
|
|
21237
|
-
"text": {
|
|
21238
|
-
"type": "string",
|
|
21239
|
-
"description": "The text to convert to speech"
|
|
21240
|
-
},
|
|
21241
|
-
"options": {
|
|
21242
|
-
"type": "SynthesizeOptions",
|
|
21243
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
21244
|
-
"properties": {
|
|
21245
|
-
"voiceId": {
|
|
21246
|
-
"type": "string",
|
|
21247
|
-
"description": ""
|
|
21248
|
-
},
|
|
21249
|
-
"modelId": {
|
|
21250
|
-
"type": "string",
|
|
21251
|
-
"description": ""
|
|
21252
|
-
},
|
|
21253
|
-
"outputFormat": {
|
|
21254
|
-
"type": "string",
|
|
21255
|
-
"description": ""
|
|
21256
|
-
},
|
|
21257
|
-
"voiceSettings": {
|
|
21258
|
-
"type": "ElevenLabsVoiceSettings",
|
|
21259
|
-
"description": ""
|
|
21260
|
-
},
|
|
21261
|
-
"disableCache": {
|
|
21262
|
-
"type": "boolean",
|
|
21263
|
-
"description": ""
|
|
21264
|
-
}
|
|
21265
|
-
}
|
|
21266
|
-
}
|
|
21267
|
-
},
|
|
21268
|
-
"required": [
|
|
21269
|
-
"text"
|
|
21270
|
-
],
|
|
21271
|
-
"returns": "Promise<Buffer>",
|
|
21272
|
-
"examples": [
|
|
21273
|
-
{
|
|
21274
|
-
"language": "ts",
|
|
21275
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
21276
|
-
}
|
|
21277
|
-
]
|
|
21278
|
-
},
|
|
21279
|
-
"say": {
|
|
21280
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
21281
|
-
"parameters": {
|
|
21282
|
-
"text": {
|
|
21283
|
-
"type": "string",
|
|
21284
|
-
"description": "The text to convert to speech"
|
|
21285
|
-
},
|
|
21286
|
-
"outputPath": {
|
|
21287
|
-
"type": "string",
|
|
21288
|
-
"description": "File path to write the audio to"
|
|
21289
|
-
},
|
|
21290
|
-
"options": {
|
|
21291
|
-
"type": "SynthesizeOptions",
|
|
21292
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
21293
|
-
"properties": {
|
|
21294
|
-
"voiceId": {
|
|
21295
|
-
"type": "string",
|
|
21296
|
-
"description": ""
|
|
21297
|
-
},
|
|
21298
|
-
"modelId": {
|
|
21299
|
-
"type": "string",
|
|
21300
|
-
"description": ""
|
|
21301
|
-
},
|
|
21302
|
-
"outputFormat": {
|
|
21303
|
-
"type": "string",
|
|
21304
|
-
"description": ""
|
|
21305
|
-
},
|
|
21306
|
-
"voiceSettings": {
|
|
21307
|
-
"type": "ElevenLabsVoiceSettings",
|
|
21308
|
-
"description": ""
|
|
21309
|
-
},
|
|
21310
|
-
"disableCache": {
|
|
21311
|
-
"type": "boolean",
|
|
21312
|
-
"description": ""
|
|
21313
|
-
}
|
|
21314
|
-
}
|
|
21315
|
-
}
|
|
21316
|
-
},
|
|
21317
|
-
"required": [
|
|
21318
|
-
"text",
|
|
21319
|
-
"outputPath"
|
|
21320
|
-
],
|
|
21321
|
-
"returns": "Promise<string>",
|
|
21322
|
-
"examples": [
|
|
21323
|
-
{
|
|
21324
|
-
"language": "ts",
|
|
21325
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
21110
|
+
"description": "The channel name to remove"
|
|
21326
21111
|
}
|
|
21327
|
-
|
|
21112
|
+
},
|
|
21113
|
+
"required": [
|
|
21114
|
+
"channelName"
|
|
21115
|
+
],
|
|
21116
|
+
"returns": "void"
|
|
21117
|
+
},
|
|
21118
|
+
"unsubscribeAll": {
|
|
21119
|
+
"description": "Unsubscribe and remove all realtime channels.",
|
|
21120
|
+
"parameters": {},
|
|
21121
|
+
"required": [],
|
|
21122
|
+
"returns": "void"
|
|
21123
|
+
},
|
|
21124
|
+
"connect": {
|
|
21125
|
+
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
21126
|
+
"parameters": {},
|
|
21127
|
+
"required": [],
|
|
21128
|
+
"returns": "void"
|
|
21129
|
+
},
|
|
21130
|
+
"disconnect": {
|
|
21131
|
+
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
21132
|
+
"parameters": {},
|
|
21133
|
+
"required": [],
|
|
21134
|
+
"returns": "void"
|
|
21328
21135
|
}
|
|
21329
21136
|
},
|
|
21330
21137
|
"getters": {
|
|
21331
|
-
"
|
|
21332
|
-
"description": "
|
|
21333
|
-
"returns": "
|
|
21334
|
-
}
|
|
21335
|
-
},
|
|
21336
|
-
"events": {
|
|
21337
|
-
"failure": {
|
|
21338
|
-
"name": "failure",
|
|
21339
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
21340
|
-
"arguments": {}
|
|
21138
|
+
"sdk": {
|
|
21139
|
+
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
21140
|
+
"returns": "SupabaseSDKClient<any, any>"
|
|
21341
21141
|
},
|
|
21342
|
-
"
|
|
21343
|
-
"
|
|
21344
|
-
"
|
|
21345
|
-
"arguments": {}
|
|
21142
|
+
"storage": {
|
|
21143
|
+
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
21144
|
+
"returns": "any"
|
|
21346
21145
|
},
|
|
21347
|
-
"
|
|
21348
|
-
"
|
|
21349
|
-
"
|
|
21350
|
-
"arguments": {}
|
|
21146
|
+
"functions": {
|
|
21147
|
+
"description": "Returns the Supabase Functions client.",
|
|
21148
|
+
"returns": "any"
|
|
21351
21149
|
}
|
|
21352
21150
|
},
|
|
21151
|
+
"events": {},
|
|
21353
21152
|
"state": {},
|
|
21354
21153
|
"options": {},
|
|
21355
21154
|
"envVars": [],
|
|
21356
21155
|
"examples": [
|
|
21357
21156
|
{
|
|
21358
21157
|
"language": "ts",
|
|
21359
|
-
"code": "const
|
|
21158
|
+
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
21360
21159
|
}
|
|
21361
21160
|
]
|
|
21362
21161
|
},
|
|
@@ -21611,6 +21410,207 @@ export const introspectionData = [
|
|
|
21611
21410
|
}
|
|
21612
21411
|
]
|
|
21613
21412
|
},
|
|
21413
|
+
{
|
|
21414
|
+
"id": "clients.elevenlabs",
|
|
21415
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
21416
|
+
"shortcut": "clients.elevenlabs",
|
|
21417
|
+
"className": "ElevenLabsClient",
|
|
21418
|
+
"methods": {
|
|
21419
|
+
"beforeRequest": {
|
|
21420
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
21421
|
+
"parameters": {},
|
|
21422
|
+
"required": [],
|
|
21423
|
+
"returns": "void"
|
|
21424
|
+
},
|
|
21425
|
+
"connect": {
|
|
21426
|
+
"description": "Validate the API key by listing available models.",
|
|
21427
|
+
"parameters": {},
|
|
21428
|
+
"required": [],
|
|
21429
|
+
"returns": "Promise<this>",
|
|
21430
|
+
"examples": [
|
|
21431
|
+
{
|
|
21432
|
+
"language": "ts",
|
|
21433
|
+
"code": "await el.connect()"
|
|
21434
|
+
}
|
|
21435
|
+
]
|
|
21436
|
+
},
|
|
21437
|
+
"listVoices": {
|
|
21438
|
+
"description": "List available voices with optional search and filtering.",
|
|
21439
|
+
"parameters": {
|
|
21440
|
+
"options": {
|
|
21441
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
21442
|
+
"description": "Query parameters for filtering voices"
|
|
21443
|
+
}
|
|
21444
|
+
},
|
|
21445
|
+
"required": [],
|
|
21446
|
+
"returns": "Promise<any>",
|
|
21447
|
+
"examples": [
|
|
21448
|
+
{
|
|
21449
|
+
"language": "ts",
|
|
21450
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
21451
|
+
}
|
|
21452
|
+
]
|
|
21453
|
+
},
|
|
21454
|
+
"getVoice": {
|
|
21455
|
+
"description": "Get details for a single voice.",
|
|
21456
|
+
"parameters": {
|
|
21457
|
+
"voiceId": {
|
|
21458
|
+
"type": "string",
|
|
21459
|
+
"description": "The voice ID to look up"
|
|
21460
|
+
}
|
|
21461
|
+
},
|
|
21462
|
+
"required": [
|
|
21463
|
+
"voiceId"
|
|
21464
|
+
],
|
|
21465
|
+
"returns": "Promise<any>",
|
|
21466
|
+
"examples": [
|
|
21467
|
+
{
|
|
21468
|
+
"language": "ts",
|
|
21469
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
21470
|
+
}
|
|
21471
|
+
]
|
|
21472
|
+
},
|
|
21473
|
+
"listModels": {
|
|
21474
|
+
"description": "List available TTS models.",
|
|
21475
|
+
"parameters": {},
|
|
21476
|
+
"required": [],
|
|
21477
|
+
"returns": "Promise<any[]>",
|
|
21478
|
+
"examples": [
|
|
21479
|
+
{
|
|
21480
|
+
"language": "ts",
|
|
21481
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
21482
|
+
}
|
|
21483
|
+
]
|
|
21484
|
+
},
|
|
21485
|
+
"synthesize": {
|
|
21486
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
21487
|
+
"parameters": {
|
|
21488
|
+
"text": {
|
|
21489
|
+
"type": "string",
|
|
21490
|
+
"description": "The text to convert to speech"
|
|
21491
|
+
},
|
|
21492
|
+
"options": {
|
|
21493
|
+
"type": "SynthesizeOptions",
|
|
21494
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
21495
|
+
"properties": {
|
|
21496
|
+
"voiceId": {
|
|
21497
|
+
"type": "string",
|
|
21498
|
+
"description": ""
|
|
21499
|
+
},
|
|
21500
|
+
"modelId": {
|
|
21501
|
+
"type": "string",
|
|
21502
|
+
"description": ""
|
|
21503
|
+
},
|
|
21504
|
+
"outputFormat": {
|
|
21505
|
+
"type": "string",
|
|
21506
|
+
"description": ""
|
|
21507
|
+
},
|
|
21508
|
+
"voiceSettings": {
|
|
21509
|
+
"type": "ElevenLabsVoiceSettings",
|
|
21510
|
+
"description": ""
|
|
21511
|
+
},
|
|
21512
|
+
"disableCache": {
|
|
21513
|
+
"type": "boolean",
|
|
21514
|
+
"description": ""
|
|
21515
|
+
}
|
|
21516
|
+
}
|
|
21517
|
+
}
|
|
21518
|
+
},
|
|
21519
|
+
"required": [
|
|
21520
|
+
"text"
|
|
21521
|
+
],
|
|
21522
|
+
"returns": "Promise<Buffer>",
|
|
21523
|
+
"examples": [
|
|
21524
|
+
{
|
|
21525
|
+
"language": "ts",
|
|
21526
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
21527
|
+
}
|
|
21528
|
+
]
|
|
21529
|
+
},
|
|
21530
|
+
"say": {
|
|
21531
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
21532
|
+
"parameters": {
|
|
21533
|
+
"text": {
|
|
21534
|
+
"type": "string",
|
|
21535
|
+
"description": "The text to convert to speech"
|
|
21536
|
+
},
|
|
21537
|
+
"outputPath": {
|
|
21538
|
+
"type": "string",
|
|
21539
|
+
"description": "File path to write the audio to"
|
|
21540
|
+
},
|
|
21541
|
+
"options": {
|
|
21542
|
+
"type": "SynthesizeOptions",
|
|
21543
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
21544
|
+
"properties": {
|
|
21545
|
+
"voiceId": {
|
|
21546
|
+
"type": "string",
|
|
21547
|
+
"description": ""
|
|
21548
|
+
},
|
|
21549
|
+
"modelId": {
|
|
21550
|
+
"type": "string",
|
|
21551
|
+
"description": ""
|
|
21552
|
+
},
|
|
21553
|
+
"outputFormat": {
|
|
21554
|
+
"type": "string",
|
|
21555
|
+
"description": ""
|
|
21556
|
+
},
|
|
21557
|
+
"voiceSettings": {
|
|
21558
|
+
"type": "ElevenLabsVoiceSettings",
|
|
21559
|
+
"description": ""
|
|
21560
|
+
},
|
|
21561
|
+
"disableCache": {
|
|
21562
|
+
"type": "boolean",
|
|
21563
|
+
"description": ""
|
|
21564
|
+
}
|
|
21565
|
+
}
|
|
21566
|
+
}
|
|
21567
|
+
},
|
|
21568
|
+
"required": [
|
|
21569
|
+
"text",
|
|
21570
|
+
"outputPath"
|
|
21571
|
+
],
|
|
21572
|
+
"returns": "Promise<string>",
|
|
21573
|
+
"examples": [
|
|
21574
|
+
{
|
|
21575
|
+
"language": "ts",
|
|
21576
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
21577
|
+
}
|
|
21578
|
+
]
|
|
21579
|
+
}
|
|
21580
|
+
},
|
|
21581
|
+
"getters": {
|
|
21582
|
+
"apiKey": {
|
|
21583
|
+
"description": "The resolved API key from options or environment.",
|
|
21584
|
+
"returns": "string"
|
|
21585
|
+
}
|
|
21586
|
+
},
|
|
21587
|
+
"events": {
|
|
21588
|
+
"failure": {
|
|
21589
|
+
"name": "failure",
|
|
21590
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21591
|
+
"arguments": {}
|
|
21592
|
+
},
|
|
21593
|
+
"voices": {
|
|
21594
|
+
"name": "voices",
|
|
21595
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21596
|
+
"arguments": {}
|
|
21597
|
+
},
|
|
21598
|
+
"speech": {
|
|
21599
|
+
"name": "speech",
|
|
21600
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21601
|
+
"arguments": {}
|
|
21602
|
+
}
|
|
21603
|
+
},
|
|
21604
|
+
"state": {},
|
|
21605
|
+
"options": {},
|
|
21606
|
+
"envVars": [],
|
|
21607
|
+
"examples": [
|
|
21608
|
+
{
|
|
21609
|
+
"language": "ts",
|
|
21610
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
21611
|
+
}
|
|
21612
|
+
]
|
|
21613
|
+
},
|
|
21614
21614
|
{
|
|
21615
21615
|
"id": "servers.mcp",
|
|
21616
21616
|
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|