@soederpop/luca 0.0.17 → 0.0.19
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/assistants/codingAssistant/CORE.md +25 -0
- package/assistants/codingAssistant/hooks.ts +3 -0
- package/assistants/codingAssistant/tools.ts +108 -0
- package/package.json +2 -2
- package/src/agi/features/assistant.ts +48 -0
- package/src/bootstrap/generated.ts +1 -1
- package/src/commands/chat.ts +42 -6
- package/src/introspection/generated.agi.ts +471 -471
- package/src/introspection/generated.node.ts +1 -1
- package/src/introspection/generated.web.ts +1 -1
- package/src/scaffolds/generated.ts +1 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
|
|
2
2
|
|
|
3
3
|
// Auto-generated introspection registry data
|
|
4
|
-
// Generated at: 2026-03-
|
|
4
|
+
// Generated at: 2026-03-20T21:12:58.940Z
|
|
5
5
|
|
|
6
6
|
setBuildTimeData('features.googleDocs', {
|
|
7
7
|
"id": "features.googleDocs",
|
|
@@ -9787,6 +9787,208 @@ setBuildTimeData('clients.openai', {
|
|
|
9787
9787
|
]
|
|
9788
9788
|
});
|
|
9789
9789
|
|
|
9790
|
+
setBuildTimeData('clients.elevenlabs', {
|
|
9791
|
+
"id": "clients.elevenlabs",
|
|
9792
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
9793
|
+
"shortcut": "clients.elevenlabs",
|
|
9794
|
+
"className": "ElevenLabsClient",
|
|
9795
|
+
"methods": {
|
|
9796
|
+
"beforeRequest": {
|
|
9797
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
9798
|
+
"parameters": {},
|
|
9799
|
+
"required": [],
|
|
9800
|
+
"returns": "void"
|
|
9801
|
+
},
|
|
9802
|
+
"connect": {
|
|
9803
|
+
"description": "Validate the API key by listing available models.",
|
|
9804
|
+
"parameters": {},
|
|
9805
|
+
"required": [],
|
|
9806
|
+
"returns": "Promise<this>",
|
|
9807
|
+
"examples": [
|
|
9808
|
+
{
|
|
9809
|
+
"language": "ts",
|
|
9810
|
+
"code": "await el.connect()"
|
|
9811
|
+
}
|
|
9812
|
+
]
|
|
9813
|
+
},
|
|
9814
|
+
"listVoices": {
|
|
9815
|
+
"description": "List available voices with optional search and filtering.",
|
|
9816
|
+
"parameters": {
|
|
9817
|
+
"options": {
|
|
9818
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
9819
|
+
"description": "Query parameters for filtering voices"
|
|
9820
|
+
}
|
|
9821
|
+
},
|
|
9822
|
+
"required": [],
|
|
9823
|
+
"returns": "Promise<any>",
|
|
9824
|
+
"examples": [
|
|
9825
|
+
{
|
|
9826
|
+
"language": "ts",
|
|
9827
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
9828
|
+
}
|
|
9829
|
+
]
|
|
9830
|
+
},
|
|
9831
|
+
"getVoice": {
|
|
9832
|
+
"description": "Get details for a single voice.",
|
|
9833
|
+
"parameters": {
|
|
9834
|
+
"voiceId": {
|
|
9835
|
+
"type": "string",
|
|
9836
|
+
"description": "The voice ID to look up"
|
|
9837
|
+
}
|
|
9838
|
+
},
|
|
9839
|
+
"required": [
|
|
9840
|
+
"voiceId"
|
|
9841
|
+
],
|
|
9842
|
+
"returns": "Promise<any>",
|
|
9843
|
+
"examples": [
|
|
9844
|
+
{
|
|
9845
|
+
"language": "ts",
|
|
9846
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
9847
|
+
}
|
|
9848
|
+
]
|
|
9849
|
+
},
|
|
9850
|
+
"listModels": {
|
|
9851
|
+
"description": "List available TTS models.",
|
|
9852
|
+
"parameters": {},
|
|
9853
|
+
"required": [],
|
|
9854
|
+
"returns": "Promise<any[]>",
|
|
9855
|
+
"examples": [
|
|
9856
|
+
{
|
|
9857
|
+
"language": "ts",
|
|
9858
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
9859
|
+
}
|
|
9860
|
+
]
|
|
9861
|
+
},
|
|
9862
|
+
"synthesize": {
|
|
9863
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
9864
|
+
"parameters": {
|
|
9865
|
+
"text": {
|
|
9866
|
+
"type": "string",
|
|
9867
|
+
"description": "The text to convert to speech"
|
|
9868
|
+
},
|
|
9869
|
+
"options": {
|
|
9870
|
+
"type": "SynthesizeOptions",
|
|
9871
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
9872
|
+
"properties": {
|
|
9873
|
+
"voiceId": {
|
|
9874
|
+
"type": "string",
|
|
9875
|
+
"description": ""
|
|
9876
|
+
},
|
|
9877
|
+
"modelId": {
|
|
9878
|
+
"type": "string",
|
|
9879
|
+
"description": ""
|
|
9880
|
+
},
|
|
9881
|
+
"outputFormat": {
|
|
9882
|
+
"type": "string",
|
|
9883
|
+
"description": ""
|
|
9884
|
+
},
|
|
9885
|
+
"voiceSettings": {
|
|
9886
|
+
"type": "ElevenLabsVoiceSettings",
|
|
9887
|
+
"description": ""
|
|
9888
|
+
},
|
|
9889
|
+
"disableCache": {
|
|
9890
|
+
"type": "boolean",
|
|
9891
|
+
"description": ""
|
|
9892
|
+
}
|
|
9893
|
+
}
|
|
9894
|
+
}
|
|
9895
|
+
},
|
|
9896
|
+
"required": [
|
|
9897
|
+
"text"
|
|
9898
|
+
],
|
|
9899
|
+
"returns": "Promise<Buffer>",
|
|
9900
|
+
"examples": [
|
|
9901
|
+
{
|
|
9902
|
+
"language": "ts",
|
|
9903
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
9904
|
+
}
|
|
9905
|
+
]
|
|
9906
|
+
},
|
|
9907
|
+
"say": {
|
|
9908
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
9909
|
+
"parameters": {
|
|
9910
|
+
"text": {
|
|
9911
|
+
"type": "string",
|
|
9912
|
+
"description": "The text to convert to speech"
|
|
9913
|
+
},
|
|
9914
|
+
"outputPath": {
|
|
9915
|
+
"type": "string",
|
|
9916
|
+
"description": "File path to write the audio to"
|
|
9917
|
+
},
|
|
9918
|
+
"options": {
|
|
9919
|
+
"type": "SynthesizeOptions",
|
|
9920
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
9921
|
+
"properties": {
|
|
9922
|
+
"voiceId": {
|
|
9923
|
+
"type": "string",
|
|
9924
|
+
"description": ""
|
|
9925
|
+
},
|
|
9926
|
+
"modelId": {
|
|
9927
|
+
"type": "string",
|
|
9928
|
+
"description": ""
|
|
9929
|
+
},
|
|
9930
|
+
"outputFormat": {
|
|
9931
|
+
"type": "string",
|
|
9932
|
+
"description": ""
|
|
9933
|
+
},
|
|
9934
|
+
"voiceSettings": {
|
|
9935
|
+
"type": "ElevenLabsVoiceSettings",
|
|
9936
|
+
"description": ""
|
|
9937
|
+
},
|
|
9938
|
+
"disableCache": {
|
|
9939
|
+
"type": "boolean",
|
|
9940
|
+
"description": ""
|
|
9941
|
+
}
|
|
9942
|
+
}
|
|
9943
|
+
}
|
|
9944
|
+
},
|
|
9945
|
+
"required": [
|
|
9946
|
+
"text",
|
|
9947
|
+
"outputPath"
|
|
9948
|
+
],
|
|
9949
|
+
"returns": "Promise<string>",
|
|
9950
|
+
"examples": [
|
|
9951
|
+
{
|
|
9952
|
+
"language": "ts",
|
|
9953
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
9954
|
+
}
|
|
9955
|
+
]
|
|
9956
|
+
}
|
|
9957
|
+
},
|
|
9958
|
+
"getters": {
|
|
9959
|
+
"apiKey": {
|
|
9960
|
+
"description": "The resolved API key from options or environment.",
|
|
9961
|
+
"returns": "string"
|
|
9962
|
+
}
|
|
9963
|
+
},
|
|
9964
|
+
"events": {
|
|
9965
|
+
"failure": {
|
|
9966
|
+
"name": "failure",
|
|
9967
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
9968
|
+
"arguments": {}
|
|
9969
|
+
},
|
|
9970
|
+
"voices": {
|
|
9971
|
+
"name": "voices",
|
|
9972
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
9973
|
+
"arguments": {}
|
|
9974
|
+
},
|
|
9975
|
+
"speech": {
|
|
9976
|
+
"name": "speech",
|
|
9977
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
9978
|
+
"arguments": {}
|
|
9979
|
+
}
|
|
9980
|
+
},
|
|
9981
|
+
"state": {},
|
|
9982
|
+
"options": {},
|
|
9983
|
+
"envVars": [],
|
|
9984
|
+
"examples": [
|
|
9985
|
+
{
|
|
9986
|
+
"language": "ts",
|
|
9987
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
9988
|
+
}
|
|
9989
|
+
]
|
|
9990
|
+
});
|
|
9991
|
+
|
|
9790
9992
|
setBuildTimeData('clients.supabase', {
|
|
9791
9993
|
"id": "clients.supabase",
|
|
9792
9994
|
"description": "Supabase client for the Luca container system. Wraps the official `@supabase/supabase-js` SDK and exposes it through Luca's typed state, events, and introspection system. The SDK is isomorphic so this single implementation works in both Node and browser containers. Use `client.sdk` for full SDK access, or use the convenience wrappers for common operations (auth, database queries, storage, edge functions, realtime).",
|
|
@@ -10178,251 +10380,49 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10178
10380
|
}
|
|
10179
10381
|
},
|
|
10180
10382
|
"getters": {
|
|
10181
|
-
"clientId": {
|
|
10182
|
-
"description": "The unique client ID used for WebSocket session tracking.",
|
|
10183
|
-
"returns": "string"
|
|
10184
|
-
},
|
|
10185
|
-
"wsURL": {
|
|
10186
|
-
"description": "The WebSocket URL derived from baseURL or overridden via options.",
|
|
10187
|
-
"returns": "string"
|
|
10188
|
-
}
|
|
10189
|
-
},
|
|
10190
|
-
"events": {
|
|
10191
|
-
"execution_start": {
|
|
10192
|
-
"name": "execution_start",
|
|
10193
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10194
|
-
"arguments": {}
|
|
10195
|
-
},
|
|
10196
|
-
"execution_complete": {
|
|
10197
|
-
"name": "execution_complete",
|
|
10198
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10199
|
-
"arguments": {}
|
|
10200
|
-
},
|
|
10201
|
-
"executing": {
|
|
10202
|
-
"name": "executing",
|
|
10203
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10204
|
-
"arguments": {}
|
|
10205
|
-
},
|
|
10206
|
-
"progress": {
|
|
10207
|
-
"name": "progress",
|
|
10208
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10209
|
-
"arguments": {}
|
|
10210
|
-
},
|
|
10211
|
-
"executed": {
|
|
10212
|
-
"name": "executed",
|
|
10213
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10214
|
-
"arguments": {}
|
|
10215
|
-
},
|
|
10216
|
-
"execution_cached": {
|
|
10217
|
-
"name": "execution_cached",
|
|
10218
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10219
|
-
"arguments": {}
|
|
10220
|
-
},
|
|
10221
|
-
"execution_error": {
|
|
10222
|
-
"name": "execution_error",
|
|
10223
|
-
"description": "Event emitted by ComfyUIClient",
|
|
10224
|
-
"arguments": {}
|
|
10225
|
-
}
|
|
10226
|
-
},
|
|
10227
|
-
"state": {},
|
|
10228
|
-
"options": {},
|
|
10229
|
-
"envVars": [],
|
|
10230
|
-
"examples": [
|
|
10231
|
-
{
|
|
10232
|
-
"language": "ts",
|
|
10233
|
-
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
10234
|
-
}
|
|
10235
|
-
]
|
|
10236
|
-
});
|
|
10237
|
-
|
|
10238
|
-
setBuildTimeData('clients.elevenlabs', {
|
|
10239
|
-
"id": "clients.elevenlabs",
|
|
10240
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
10241
|
-
"shortcut": "clients.elevenlabs",
|
|
10242
|
-
"className": "ElevenLabsClient",
|
|
10243
|
-
"methods": {
|
|
10244
|
-
"beforeRequest": {
|
|
10245
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
10246
|
-
"parameters": {},
|
|
10247
|
-
"required": [],
|
|
10248
|
-
"returns": "void"
|
|
10249
|
-
},
|
|
10250
|
-
"connect": {
|
|
10251
|
-
"description": "Validate the API key by listing available models.",
|
|
10252
|
-
"parameters": {},
|
|
10253
|
-
"required": [],
|
|
10254
|
-
"returns": "Promise<this>",
|
|
10255
|
-
"examples": [
|
|
10256
|
-
{
|
|
10257
|
-
"language": "ts",
|
|
10258
|
-
"code": "await el.connect()"
|
|
10259
|
-
}
|
|
10260
|
-
]
|
|
10261
|
-
},
|
|
10262
|
-
"listVoices": {
|
|
10263
|
-
"description": "List available voices with optional search and filtering.",
|
|
10264
|
-
"parameters": {
|
|
10265
|
-
"options": {
|
|
10266
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10267
|
-
"description": "Query parameters for filtering voices"
|
|
10268
|
-
}
|
|
10269
|
-
},
|
|
10270
|
-
"required": [],
|
|
10271
|
-
"returns": "Promise<any>",
|
|
10272
|
-
"examples": [
|
|
10273
|
-
{
|
|
10274
|
-
"language": "ts",
|
|
10275
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10276
|
-
}
|
|
10277
|
-
]
|
|
10278
|
-
},
|
|
10279
|
-
"getVoice": {
|
|
10280
|
-
"description": "Get details for a single voice.",
|
|
10281
|
-
"parameters": {
|
|
10282
|
-
"voiceId": {
|
|
10283
|
-
"type": "string",
|
|
10284
|
-
"description": "The voice ID to look up"
|
|
10285
|
-
}
|
|
10286
|
-
},
|
|
10287
|
-
"required": [
|
|
10288
|
-
"voiceId"
|
|
10289
|
-
],
|
|
10290
|
-
"returns": "Promise<any>",
|
|
10291
|
-
"examples": [
|
|
10292
|
-
{
|
|
10293
|
-
"language": "ts",
|
|
10294
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10295
|
-
}
|
|
10296
|
-
]
|
|
10297
|
-
},
|
|
10298
|
-
"listModels": {
|
|
10299
|
-
"description": "List available TTS models.",
|
|
10300
|
-
"parameters": {},
|
|
10301
|
-
"required": [],
|
|
10302
|
-
"returns": "Promise<any[]>",
|
|
10303
|
-
"examples": [
|
|
10304
|
-
{
|
|
10305
|
-
"language": "ts",
|
|
10306
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10307
|
-
}
|
|
10308
|
-
]
|
|
10309
|
-
},
|
|
10310
|
-
"synthesize": {
|
|
10311
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10312
|
-
"parameters": {
|
|
10313
|
-
"text": {
|
|
10314
|
-
"type": "string",
|
|
10315
|
-
"description": "The text to convert to speech"
|
|
10316
|
-
},
|
|
10317
|
-
"options": {
|
|
10318
|
-
"type": "SynthesizeOptions",
|
|
10319
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10320
|
-
"properties": {
|
|
10321
|
-
"voiceId": {
|
|
10322
|
-
"type": "string",
|
|
10323
|
-
"description": ""
|
|
10324
|
-
},
|
|
10325
|
-
"modelId": {
|
|
10326
|
-
"type": "string",
|
|
10327
|
-
"description": ""
|
|
10328
|
-
},
|
|
10329
|
-
"outputFormat": {
|
|
10330
|
-
"type": "string",
|
|
10331
|
-
"description": ""
|
|
10332
|
-
},
|
|
10333
|
-
"voiceSettings": {
|
|
10334
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10335
|
-
"description": ""
|
|
10336
|
-
},
|
|
10337
|
-
"disableCache": {
|
|
10338
|
-
"type": "boolean",
|
|
10339
|
-
"description": ""
|
|
10340
|
-
}
|
|
10341
|
-
}
|
|
10342
|
-
}
|
|
10343
|
-
},
|
|
10344
|
-
"required": [
|
|
10345
|
-
"text"
|
|
10346
|
-
],
|
|
10347
|
-
"returns": "Promise<Buffer>",
|
|
10348
|
-
"examples": [
|
|
10349
|
-
{
|
|
10350
|
-
"language": "ts",
|
|
10351
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10352
|
-
}
|
|
10353
|
-
]
|
|
10354
|
-
},
|
|
10355
|
-
"say": {
|
|
10356
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
10357
|
-
"parameters": {
|
|
10358
|
-
"text": {
|
|
10359
|
-
"type": "string",
|
|
10360
|
-
"description": "The text to convert to speech"
|
|
10361
|
-
},
|
|
10362
|
-
"outputPath": {
|
|
10363
|
-
"type": "string",
|
|
10364
|
-
"description": "File path to write the audio to"
|
|
10365
|
-
},
|
|
10366
|
-
"options": {
|
|
10367
|
-
"type": "SynthesizeOptions",
|
|
10368
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10369
|
-
"properties": {
|
|
10370
|
-
"voiceId": {
|
|
10371
|
-
"type": "string",
|
|
10372
|
-
"description": ""
|
|
10373
|
-
},
|
|
10374
|
-
"modelId": {
|
|
10375
|
-
"type": "string",
|
|
10376
|
-
"description": ""
|
|
10377
|
-
},
|
|
10378
|
-
"outputFormat": {
|
|
10379
|
-
"type": "string",
|
|
10380
|
-
"description": ""
|
|
10381
|
-
},
|
|
10382
|
-
"voiceSettings": {
|
|
10383
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10384
|
-
"description": ""
|
|
10385
|
-
},
|
|
10386
|
-
"disableCache": {
|
|
10387
|
-
"type": "boolean",
|
|
10388
|
-
"description": ""
|
|
10389
|
-
}
|
|
10390
|
-
}
|
|
10391
|
-
}
|
|
10392
|
-
},
|
|
10393
|
-
"required": [
|
|
10394
|
-
"text",
|
|
10395
|
-
"outputPath"
|
|
10396
|
-
],
|
|
10397
|
-
"returns": "Promise<string>",
|
|
10398
|
-
"examples": [
|
|
10399
|
-
{
|
|
10400
|
-
"language": "ts",
|
|
10401
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10402
|
-
}
|
|
10403
|
-
]
|
|
10404
|
-
}
|
|
10405
|
-
},
|
|
10406
|
-
"getters": {
|
|
10407
|
-
"apiKey": {
|
|
10408
|
-
"description": "The resolved API key from options or environment.",
|
|
10383
|
+
"clientId": {
|
|
10384
|
+
"description": "The unique client ID used for WebSocket session tracking.",
|
|
10385
|
+
"returns": "string"
|
|
10386
|
+
},
|
|
10387
|
+
"wsURL": {
|
|
10388
|
+
"description": "The WebSocket URL derived from baseURL or overridden via options.",
|
|
10409
10389
|
"returns": "string"
|
|
10410
10390
|
}
|
|
10411
10391
|
},
|
|
10412
10392
|
"events": {
|
|
10413
|
-
"
|
|
10414
|
-
"name": "
|
|
10415
|
-
"description": "Event emitted by
|
|
10393
|
+
"execution_start": {
|
|
10394
|
+
"name": "execution_start",
|
|
10395
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10416
10396
|
"arguments": {}
|
|
10417
10397
|
},
|
|
10418
|
-
"
|
|
10419
|
-
"name": "
|
|
10420
|
-
"description": "Event emitted by
|
|
10398
|
+
"execution_complete": {
|
|
10399
|
+
"name": "execution_complete",
|
|
10400
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10421
10401
|
"arguments": {}
|
|
10422
10402
|
},
|
|
10423
|
-
"
|
|
10424
|
-
"name": "
|
|
10425
|
-
"description": "Event emitted by
|
|
10403
|
+
"executing": {
|
|
10404
|
+
"name": "executing",
|
|
10405
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10406
|
+
"arguments": {}
|
|
10407
|
+
},
|
|
10408
|
+
"progress": {
|
|
10409
|
+
"name": "progress",
|
|
10410
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10411
|
+
"arguments": {}
|
|
10412
|
+
},
|
|
10413
|
+
"executed": {
|
|
10414
|
+
"name": "executed",
|
|
10415
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10416
|
+
"arguments": {}
|
|
10417
|
+
},
|
|
10418
|
+
"execution_cached": {
|
|
10419
|
+
"name": "execution_cached",
|
|
10420
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10421
|
+
"arguments": {}
|
|
10422
|
+
},
|
|
10423
|
+
"execution_error": {
|
|
10424
|
+
"name": "execution_error",
|
|
10425
|
+
"description": "Event emitted by ComfyUIClient",
|
|
10426
10426
|
"arguments": {}
|
|
10427
10427
|
}
|
|
10428
10428
|
},
|
|
@@ -10432,7 +10432,7 @@ setBuildTimeData('clients.elevenlabs', {
|
|
|
10432
10432
|
"examples": [
|
|
10433
10433
|
{
|
|
10434
10434
|
"language": "ts",
|
|
10435
|
-
"code": "const
|
|
10435
|
+
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
10436
10436
|
}
|
|
10437
10437
|
]
|
|
10438
10438
|
});
|
|
@@ -23202,56 +23202,257 @@ export const introspectionData = [
|
|
|
23202
23202
|
}
|
|
23203
23203
|
},
|
|
23204
23204
|
"required": [
|
|
23205
|
-
"messages"
|
|
23205
|
+
"messages"
|
|
23206
|
+
],
|
|
23207
|
+
"returns": "Promise<string>",
|
|
23208
|
+
"examples": [
|
|
23209
|
+
{
|
|
23210
|
+
"language": "ts",
|
|
23211
|
+
"code": "const reply = await openai.chat([\n { role: 'system', content: 'You are a pirate.' },\n { role: 'user', content: 'Hello!' }\n])"
|
|
23212
|
+
}
|
|
23213
|
+
]
|
|
23214
|
+
}
|
|
23215
|
+
},
|
|
23216
|
+
"getters": {
|
|
23217
|
+
"defaultModel": {
|
|
23218
|
+
"description": "The default model used for completions, from options or 'gpt-4o'.",
|
|
23219
|
+
"returns": "string"
|
|
23220
|
+
},
|
|
23221
|
+
"raw": {
|
|
23222
|
+
"description": "The underlying OpenAI SDK instance for advanced use cases.",
|
|
23223
|
+
"returns": "OpenAI"
|
|
23224
|
+
}
|
|
23225
|
+
},
|
|
23226
|
+
"events": {
|
|
23227
|
+
"connected": {
|
|
23228
|
+
"name": "connected",
|
|
23229
|
+
"description": "Event emitted by OpenAIClient",
|
|
23230
|
+
"arguments": {}
|
|
23231
|
+
},
|
|
23232
|
+
"failure": {
|
|
23233
|
+
"name": "failure",
|
|
23234
|
+
"description": "Event emitted by OpenAIClient",
|
|
23235
|
+
"arguments": {}
|
|
23236
|
+
},
|
|
23237
|
+
"completion": {
|
|
23238
|
+
"name": "completion",
|
|
23239
|
+
"description": "Event emitted by OpenAIClient",
|
|
23240
|
+
"arguments": {}
|
|
23241
|
+
},
|
|
23242
|
+
"embedding": {
|
|
23243
|
+
"name": "embedding",
|
|
23244
|
+
"description": "Event emitted by OpenAIClient",
|
|
23245
|
+
"arguments": {}
|
|
23246
|
+
},
|
|
23247
|
+
"image": {
|
|
23248
|
+
"name": "image",
|
|
23249
|
+
"description": "Event emitted by OpenAIClient",
|
|
23250
|
+
"arguments": {}
|
|
23251
|
+
},
|
|
23252
|
+
"models": {
|
|
23253
|
+
"name": "models",
|
|
23254
|
+
"description": "Event emitted by OpenAIClient",
|
|
23255
|
+
"arguments": {}
|
|
23256
|
+
}
|
|
23257
|
+
},
|
|
23258
|
+
"state": {},
|
|
23259
|
+
"options": {},
|
|
23260
|
+
"envVars": [],
|
|
23261
|
+
"examples": [
|
|
23262
|
+
{
|
|
23263
|
+
"language": "ts",
|
|
23264
|
+
"code": "const openai = container.client('openai', { defaultModel: 'gpt-4o' })\nconst answer = await openai.ask('What is the meaning of life?')\nconsole.log(answer)"
|
|
23265
|
+
}
|
|
23266
|
+
]
|
|
23267
|
+
},
|
|
23268
|
+
{
|
|
23269
|
+
"id": "clients.elevenlabs",
|
|
23270
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
23271
|
+
"shortcut": "clients.elevenlabs",
|
|
23272
|
+
"className": "ElevenLabsClient",
|
|
23273
|
+
"methods": {
|
|
23274
|
+
"beforeRequest": {
|
|
23275
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
23276
|
+
"parameters": {},
|
|
23277
|
+
"required": [],
|
|
23278
|
+
"returns": "void"
|
|
23279
|
+
},
|
|
23280
|
+
"connect": {
|
|
23281
|
+
"description": "Validate the API key by listing available models.",
|
|
23282
|
+
"parameters": {},
|
|
23283
|
+
"required": [],
|
|
23284
|
+
"returns": "Promise<this>",
|
|
23285
|
+
"examples": [
|
|
23286
|
+
{
|
|
23287
|
+
"language": "ts",
|
|
23288
|
+
"code": "await el.connect()"
|
|
23289
|
+
}
|
|
23290
|
+
]
|
|
23291
|
+
},
|
|
23292
|
+
"listVoices": {
|
|
23293
|
+
"description": "List available voices with optional search and filtering.",
|
|
23294
|
+
"parameters": {
|
|
23295
|
+
"options": {
|
|
23296
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
23297
|
+
"description": "Query parameters for filtering voices"
|
|
23298
|
+
}
|
|
23299
|
+
},
|
|
23300
|
+
"required": [],
|
|
23301
|
+
"returns": "Promise<any>",
|
|
23302
|
+
"examples": [
|
|
23303
|
+
{
|
|
23304
|
+
"language": "ts",
|
|
23305
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
23306
|
+
}
|
|
23307
|
+
]
|
|
23308
|
+
},
|
|
23309
|
+
"getVoice": {
|
|
23310
|
+
"description": "Get details for a single voice.",
|
|
23311
|
+
"parameters": {
|
|
23312
|
+
"voiceId": {
|
|
23313
|
+
"type": "string",
|
|
23314
|
+
"description": "The voice ID to look up"
|
|
23315
|
+
}
|
|
23316
|
+
},
|
|
23317
|
+
"required": [
|
|
23318
|
+
"voiceId"
|
|
23319
|
+
],
|
|
23320
|
+
"returns": "Promise<any>",
|
|
23321
|
+
"examples": [
|
|
23322
|
+
{
|
|
23323
|
+
"language": "ts",
|
|
23324
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
23325
|
+
}
|
|
23326
|
+
]
|
|
23327
|
+
},
|
|
23328
|
+
"listModels": {
|
|
23329
|
+
"description": "List available TTS models.",
|
|
23330
|
+
"parameters": {},
|
|
23331
|
+
"required": [],
|
|
23332
|
+
"returns": "Promise<any[]>",
|
|
23333
|
+
"examples": [
|
|
23334
|
+
{
|
|
23335
|
+
"language": "ts",
|
|
23336
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
23337
|
+
}
|
|
23338
|
+
]
|
|
23339
|
+
},
|
|
23340
|
+
"synthesize": {
|
|
23341
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
23342
|
+
"parameters": {
|
|
23343
|
+
"text": {
|
|
23344
|
+
"type": "string",
|
|
23345
|
+
"description": "The text to convert to speech"
|
|
23346
|
+
},
|
|
23347
|
+
"options": {
|
|
23348
|
+
"type": "SynthesizeOptions",
|
|
23349
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
23350
|
+
"properties": {
|
|
23351
|
+
"voiceId": {
|
|
23352
|
+
"type": "string",
|
|
23353
|
+
"description": ""
|
|
23354
|
+
},
|
|
23355
|
+
"modelId": {
|
|
23356
|
+
"type": "string",
|
|
23357
|
+
"description": ""
|
|
23358
|
+
},
|
|
23359
|
+
"outputFormat": {
|
|
23360
|
+
"type": "string",
|
|
23361
|
+
"description": ""
|
|
23362
|
+
},
|
|
23363
|
+
"voiceSettings": {
|
|
23364
|
+
"type": "ElevenLabsVoiceSettings",
|
|
23365
|
+
"description": ""
|
|
23366
|
+
},
|
|
23367
|
+
"disableCache": {
|
|
23368
|
+
"type": "boolean",
|
|
23369
|
+
"description": ""
|
|
23370
|
+
}
|
|
23371
|
+
}
|
|
23372
|
+
}
|
|
23373
|
+
},
|
|
23374
|
+
"required": [
|
|
23375
|
+
"text"
|
|
23376
|
+
],
|
|
23377
|
+
"returns": "Promise<Buffer>",
|
|
23378
|
+
"examples": [
|
|
23379
|
+
{
|
|
23380
|
+
"language": "ts",
|
|
23381
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
23382
|
+
}
|
|
23383
|
+
]
|
|
23384
|
+
},
|
|
23385
|
+
"say": {
|
|
23386
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
23387
|
+
"parameters": {
|
|
23388
|
+
"text": {
|
|
23389
|
+
"type": "string",
|
|
23390
|
+
"description": "The text to convert to speech"
|
|
23391
|
+
},
|
|
23392
|
+
"outputPath": {
|
|
23393
|
+
"type": "string",
|
|
23394
|
+
"description": "File path to write the audio to"
|
|
23395
|
+
},
|
|
23396
|
+
"options": {
|
|
23397
|
+
"type": "SynthesizeOptions",
|
|
23398
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
23399
|
+
"properties": {
|
|
23400
|
+
"voiceId": {
|
|
23401
|
+
"type": "string",
|
|
23402
|
+
"description": ""
|
|
23403
|
+
},
|
|
23404
|
+
"modelId": {
|
|
23405
|
+
"type": "string",
|
|
23406
|
+
"description": ""
|
|
23407
|
+
},
|
|
23408
|
+
"outputFormat": {
|
|
23409
|
+
"type": "string",
|
|
23410
|
+
"description": ""
|
|
23411
|
+
},
|
|
23412
|
+
"voiceSettings": {
|
|
23413
|
+
"type": "ElevenLabsVoiceSettings",
|
|
23414
|
+
"description": ""
|
|
23415
|
+
},
|
|
23416
|
+
"disableCache": {
|
|
23417
|
+
"type": "boolean",
|
|
23418
|
+
"description": ""
|
|
23419
|
+
}
|
|
23420
|
+
}
|
|
23421
|
+
}
|
|
23422
|
+
},
|
|
23423
|
+
"required": [
|
|
23424
|
+
"text",
|
|
23425
|
+
"outputPath"
|
|
23206
23426
|
],
|
|
23207
23427
|
"returns": "Promise<string>",
|
|
23208
23428
|
"examples": [
|
|
23209
23429
|
{
|
|
23210
23430
|
"language": "ts",
|
|
23211
|
-
"code": "const
|
|
23431
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
23212
23432
|
}
|
|
23213
23433
|
]
|
|
23214
23434
|
}
|
|
23215
23435
|
},
|
|
23216
23436
|
"getters": {
|
|
23217
|
-
"
|
|
23218
|
-
"description": "The
|
|
23437
|
+
"apiKey": {
|
|
23438
|
+
"description": "The resolved API key from options or environment.",
|
|
23219
23439
|
"returns": "string"
|
|
23220
|
-
},
|
|
23221
|
-
"raw": {
|
|
23222
|
-
"description": "The underlying OpenAI SDK instance for advanced use cases.",
|
|
23223
|
-
"returns": "OpenAI"
|
|
23224
23440
|
}
|
|
23225
23441
|
},
|
|
23226
23442
|
"events": {
|
|
23227
|
-
"connected": {
|
|
23228
|
-
"name": "connected",
|
|
23229
|
-
"description": "Event emitted by OpenAIClient",
|
|
23230
|
-
"arguments": {}
|
|
23231
|
-
},
|
|
23232
23443
|
"failure": {
|
|
23233
23444
|
"name": "failure",
|
|
23234
|
-
"description": "Event emitted by
|
|
23235
|
-
"arguments": {}
|
|
23236
|
-
},
|
|
23237
|
-
"completion": {
|
|
23238
|
-
"name": "completion",
|
|
23239
|
-
"description": "Event emitted by OpenAIClient",
|
|
23240
|
-
"arguments": {}
|
|
23241
|
-
},
|
|
23242
|
-
"embedding": {
|
|
23243
|
-
"name": "embedding",
|
|
23244
|
-
"description": "Event emitted by OpenAIClient",
|
|
23445
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23245
23446
|
"arguments": {}
|
|
23246
23447
|
},
|
|
23247
|
-
"
|
|
23248
|
-
"name": "
|
|
23249
|
-
"description": "Event emitted by
|
|
23448
|
+
"voices": {
|
|
23449
|
+
"name": "voices",
|
|
23450
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23250
23451
|
"arguments": {}
|
|
23251
23452
|
},
|
|
23252
|
-
"
|
|
23253
|
-
"name": "
|
|
23254
|
-
"description": "Event emitted by
|
|
23453
|
+
"speech": {
|
|
23454
|
+
"name": "speech",
|
|
23455
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23255
23456
|
"arguments": {}
|
|
23256
23457
|
}
|
|
23257
23458
|
},
|
|
@@ -23261,7 +23462,7 @@ export const introspectionData = [
|
|
|
23261
23462
|
"examples": [
|
|
23262
23463
|
{
|
|
23263
23464
|
"language": "ts",
|
|
23264
|
-
"code": "const
|
|
23465
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
23265
23466
|
}
|
|
23266
23467
|
]
|
|
23267
23468
|
},
|
|
@@ -23711,207 +23912,6 @@ export const introspectionData = [
|
|
|
23711
23912
|
}
|
|
23712
23913
|
]
|
|
23713
23914
|
},
|
|
23714
|
-
{
|
|
23715
|
-
"id": "clients.elevenlabs",
|
|
23716
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
23717
|
-
"shortcut": "clients.elevenlabs",
|
|
23718
|
-
"className": "ElevenLabsClient",
|
|
23719
|
-
"methods": {
|
|
23720
|
-
"beforeRequest": {
|
|
23721
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
23722
|
-
"parameters": {},
|
|
23723
|
-
"required": [],
|
|
23724
|
-
"returns": "void"
|
|
23725
|
-
},
|
|
23726
|
-
"connect": {
|
|
23727
|
-
"description": "Validate the API key by listing available models.",
|
|
23728
|
-
"parameters": {},
|
|
23729
|
-
"required": [],
|
|
23730
|
-
"returns": "Promise<this>",
|
|
23731
|
-
"examples": [
|
|
23732
|
-
{
|
|
23733
|
-
"language": "ts",
|
|
23734
|
-
"code": "await el.connect()"
|
|
23735
|
-
}
|
|
23736
|
-
]
|
|
23737
|
-
},
|
|
23738
|
-
"listVoices": {
|
|
23739
|
-
"description": "List available voices with optional search and filtering.",
|
|
23740
|
-
"parameters": {
|
|
23741
|
-
"options": {
|
|
23742
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
23743
|
-
"description": "Query parameters for filtering voices"
|
|
23744
|
-
}
|
|
23745
|
-
},
|
|
23746
|
-
"required": [],
|
|
23747
|
-
"returns": "Promise<any>",
|
|
23748
|
-
"examples": [
|
|
23749
|
-
{
|
|
23750
|
-
"language": "ts",
|
|
23751
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
23752
|
-
}
|
|
23753
|
-
]
|
|
23754
|
-
},
|
|
23755
|
-
"getVoice": {
|
|
23756
|
-
"description": "Get details for a single voice.",
|
|
23757
|
-
"parameters": {
|
|
23758
|
-
"voiceId": {
|
|
23759
|
-
"type": "string",
|
|
23760
|
-
"description": "The voice ID to look up"
|
|
23761
|
-
}
|
|
23762
|
-
},
|
|
23763
|
-
"required": [
|
|
23764
|
-
"voiceId"
|
|
23765
|
-
],
|
|
23766
|
-
"returns": "Promise<any>",
|
|
23767
|
-
"examples": [
|
|
23768
|
-
{
|
|
23769
|
-
"language": "ts",
|
|
23770
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
23771
|
-
}
|
|
23772
|
-
]
|
|
23773
|
-
},
|
|
23774
|
-
"listModels": {
|
|
23775
|
-
"description": "List available TTS models.",
|
|
23776
|
-
"parameters": {},
|
|
23777
|
-
"required": [],
|
|
23778
|
-
"returns": "Promise<any[]>",
|
|
23779
|
-
"examples": [
|
|
23780
|
-
{
|
|
23781
|
-
"language": "ts",
|
|
23782
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
23783
|
-
}
|
|
23784
|
-
]
|
|
23785
|
-
},
|
|
23786
|
-
"synthesize": {
|
|
23787
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
23788
|
-
"parameters": {
|
|
23789
|
-
"text": {
|
|
23790
|
-
"type": "string",
|
|
23791
|
-
"description": "The text to convert to speech"
|
|
23792
|
-
},
|
|
23793
|
-
"options": {
|
|
23794
|
-
"type": "SynthesizeOptions",
|
|
23795
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
23796
|
-
"properties": {
|
|
23797
|
-
"voiceId": {
|
|
23798
|
-
"type": "string",
|
|
23799
|
-
"description": ""
|
|
23800
|
-
},
|
|
23801
|
-
"modelId": {
|
|
23802
|
-
"type": "string",
|
|
23803
|
-
"description": ""
|
|
23804
|
-
},
|
|
23805
|
-
"outputFormat": {
|
|
23806
|
-
"type": "string",
|
|
23807
|
-
"description": ""
|
|
23808
|
-
},
|
|
23809
|
-
"voiceSettings": {
|
|
23810
|
-
"type": "ElevenLabsVoiceSettings",
|
|
23811
|
-
"description": ""
|
|
23812
|
-
},
|
|
23813
|
-
"disableCache": {
|
|
23814
|
-
"type": "boolean",
|
|
23815
|
-
"description": ""
|
|
23816
|
-
}
|
|
23817
|
-
}
|
|
23818
|
-
}
|
|
23819
|
-
},
|
|
23820
|
-
"required": [
|
|
23821
|
-
"text"
|
|
23822
|
-
],
|
|
23823
|
-
"returns": "Promise<Buffer>",
|
|
23824
|
-
"examples": [
|
|
23825
|
-
{
|
|
23826
|
-
"language": "ts",
|
|
23827
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
23828
|
-
}
|
|
23829
|
-
]
|
|
23830
|
-
},
|
|
23831
|
-
"say": {
|
|
23832
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
23833
|
-
"parameters": {
|
|
23834
|
-
"text": {
|
|
23835
|
-
"type": "string",
|
|
23836
|
-
"description": "The text to convert to speech"
|
|
23837
|
-
},
|
|
23838
|
-
"outputPath": {
|
|
23839
|
-
"type": "string",
|
|
23840
|
-
"description": "File path to write the audio to"
|
|
23841
|
-
},
|
|
23842
|
-
"options": {
|
|
23843
|
-
"type": "SynthesizeOptions",
|
|
23844
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
23845
|
-
"properties": {
|
|
23846
|
-
"voiceId": {
|
|
23847
|
-
"type": "string",
|
|
23848
|
-
"description": ""
|
|
23849
|
-
},
|
|
23850
|
-
"modelId": {
|
|
23851
|
-
"type": "string",
|
|
23852
|
-
"description": ""
|
|
23853
|
-
},
|
|
23854
|
-
"outputFormat": {
|
|
23855
|
-
"type": "string",
|
|
23856
|
-
"description": ""
|
|
23857
|
-
},
|
|
23858
|
-
"voiceSettings": {
|
|
23859
|
-
"type": "ElevenLabsVoiceSettings",
|
|
23860
|
-
"description": ""
|
|
23861
|
-
},
|
|
23862
|
-
"disableCache": {
|
|
23863
|
-
"type": "boolean",
|
|
23864
|
-
"description": ""
|
|
23865
|
-
}
|
|
23866
|
-
}
|
|
23867
|
-
}
|
|
23868
|
-
},
|
|
23869
|
-
"required": [
|
|
23870
|
-
"text",
|
|
23871
|
-
"outputPath"
|
|
23872
|
-
],
|
|
23873
|
-
"returns": "Promise<string>",
|
|
23874
|
-
"examples": [
|
|
23875
|
-
{
|
|
23876
|
-
"language": "ts",
|
|
23877
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
23878
|
-
}
|
|
23879
|
-
]
|
|
23880
|
-
}
|
|
23881
|
-
},
|
|
23882
|
-
"getters": {
|
|
23883
|
-
"apiKey": {
|
|
23884
|
-
"description": "The resolved API key from options or environment.",
|
|
23885
|
-
"returns": "string"
|
|
23886
|
-
}
|
|
23887
|
-
},
|
|
23888
|
-
"events": {
|
|
23889
|
-
"failure": {
|
|
23890
|
-
"name": "failure",
|
|
23891
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
23892
|
-
"arguments": {}
|
|
23893
|
-
},
|
|
23894
|
-
"voices": {
|
|
23895
|
-
"name": "voices",
|
|
23896
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
23897
|
-
"arguments": {}
|
|
23898
|
-
},
|
|
23899
|
-
"speech": {
|
|
23900
|
-
"name": "speech",
|
|
23901
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
23902
|
-
"arguments": {}
|
|
23903
|
-
}
|
|
23904
|
-
},
|
|
23905
|
-
"state": {},
|
|
23906
|
-
"options": {},
|
|
23907
|
-
"envVars": [],
|
|
23908
|
-
"examples": [
|
|
23909
|
-
{
|
|
23910
|
-
"language": "ts",
|
|
23911
|
-
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
23912
|
-
}
|
|
23913
|
-
]
|
|
23914
|
-
},
|
|
23915
23915
|
{
|
|
23916
23916
|
"id": "servers.mcp",
|
|
23917
23917
|
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|