@soederpop/luca 0.0.15 → 0.0.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,7 +1,7 @@
1
1
  import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
2
2
 
3
3
  // Auto-generated introspection registry data
4
- // Generated at: 2026-03-20T05:45:32.252Z
4
+ // Generated at: 2026-03-20T16:25:11.024Z
5
5
 
6
6
  setBuildTimeData('features.googleDocs', {
7
7
  "id": "features.googleDocs",
@@ -9983,208 +9983,6 @@ setBuildTimeData('clients.supabase', {
9983
9983
  ]
9984
9984
  });
9985
9985
 
9986
- setBuildTimeData('clients.elevenlabs', {
9987
- "id": "clients.elevenlabs",
9988
- "description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
9989
- "shortcut": "clients.elevenlabs",
9990
- "className": "ElevenLabsClient",
9991
- "methods": {
9992
- "beforeRequest": {
9993
- "description": "Inject the xi-api-key header before each request.",
9994
- "parameters": {},
9995
- "required": [],
9996
- "returns": "void"
9997
- },
9998
- "connect": {
9999
- "description": "Validate the API key by listing available models.",
10000
- "parameters": {},
10001
- "required": [],
10002
- "returns": "Promise<this>",
10003
- "examples": [
10004
- {
10005
- "language": "ts",
10006
- "code": "await el.connect()"
10007
- }
10008
- ]
10009
- },
10010
- "listVoices": {
10011
- "description": "List available voices with optional search and filtering.",
10012
- "parameters": {
10013
- "options": {
10014
- "type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
10015
- "description": "Query parameters for filtering voices"
10016
- }
10017
- },
10018
- "required": [],
10019
- "returns": "Promise<any>",
10020
- "examples": [
10021
- {
10022
- "language": "ts",
10023
- "code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
10024
- }
10025
- ]
10026
- },
10027
- "getVoice": {
10028
- "description": "Get details for a single voice.",
10029
- "parameters": {
10030
- "voiceId": {
10031
- "type": "string",
10032
- "description": "The voice ID to look up"
10033
- }
10034
- },
10035
- "required": [
10036
- "voiceId"
10037
- ],
10038
- "returns": "Promise<any>",
10039
- "examples": [
10040
- {
10041
- "language": "ts",
10042
- "code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
10043
- }
10044
- ]
10045
- },
10046
- "listModels": {
10047
- "description": "List available TTS models.",
10048
- "parameters": {},
10049
- "required": [],
10050
- "returns": "Promise<any[]>",
10051
- "examples": [
10052
- {
10053
- "language": "ts",
10054
- "code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
10055
- }
10056
- ]
10057
- },
10058
- "synthesize": {
10059
- "description": "Synthesize speech from text, returning audio as a Buffer.",
10060
- "parameters": {
10061
- "text": {
10062
- "type": "string",
10063
- "description": "The text to convert to speech"
10064
- },
10065
- "options": {
10066
- "type": "SynthesizeOptions",
10067
- "description": "Voice, model, format, and voice settings overrides",
10068
- "properties": {
10069
- "voiceId": {
10070
- "type": "string",
10071
- "description": ""
10072
- },
10073
- "modelId": {
10074
- "type": "string",
10075
- "description": ""
10076
- },
10077
- "outputFormat": {
10078
- "type": "string",
10079
- "description": ""
10080
- },
10081
- "voiceSettings": {
10082
- "type": "ElevenLabsVoiceSettings",
10083
- "description": ""
10084
- },
10085
- "disableCache": {
10086
- "type": "boolean",
10087
- "description": ""
10088
- }
10089
- }
10090
- }
10091
- },
10092
- "required": [
10093
- "text"
10094
- ],
10095
- "returns": "Promise<Buffer>",
10096
- "examples": [
10097
- {
10098
- "language": "ts",
10099
- "code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
10100
- }
10101
- ]
10102
- },
10103
- "say": {
10104
- "description": "Synthesize speech and write the audio to a file.",
10105
- "parameters": {
10106
- "text": {
10107
- "type": "string",
10108
- "description": "The text to convert to speech"
10109
- },
10110
- "outputPath": {
10111
- "type": "string",
10112
- "description": "File path to write the audio to"
10113
- },
10114
- "options": {
10115
- "type": "SynthesizeOptions",
10116
- "description": "Voice, model, format, and voice settings overrides",
10117
- "properties": {
10118
- "voiceId": {
10119
- "type": "string",
10120
- "description": ""
10121
- },
10122
- "modelId": {
10123
- "type": "string",
10124
- "description": ""
10125
- },
10126
- "outputFormat": {
10127
- "type": "string",
10128
- "description": ""
10129
- },
10130
- "voiceSettings": {
10131
- "type": "ElevenLabsVoiceSettings",
10132
- "description": ""
10133
- },
10134
- "disableCache": {
10135
- "type": "boolean",
10136
- "description": ""
10137
- }
10138
- }
10139
- }
10140
- },
10141
- "required": [
10142
- "text",
10143
- "outputPath"
10144
- ],
10145
- "returns": "Promise<string>",
10146
- "examples": [
10147
- {
10148
- "language": "ts",
10149
- "code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
10150
- }
10151
- ]
10152
- }
10153
- },
10154
- "getters": {
10155
- "apiKey": {
10156
- "description": "The resolved API key from options or environment.",
10157
- "returns": "string"
10158
- }
10159
- },
10160
- "events": {
10161
- "failure": {
10162
- "name": "failure",
10163
- "description": "Event emitted by ElevenLabsClient",
10164
- "arguments": {}
10165
- },
10166
- "voices": {
10167
- "name": "voices",
10168
- "description": "Event emitted by ElevenLabsClient",
10169
- "arguments": {}
10170
- },
10171
- "speech": {
10172
- "name": "speech",
10173
- "description": "Event emitted by ElevenLabsClient",
10174
- "arguments": {}
10175
- }
10176
- },
10177
- "state": {},
10178
- "options": {},
10179
- "envVars": [],
10180
- "examples": [
10181
- {
10182
- "language": "ts",
10183
- "code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
10184
- }
10185
- ]
10186
- });
10187
-
10188
9986
  setBuildTimeData('clients.comfyui', {
10189
9987
  "id": "clients.comfyui",
10190
9988
  "description": "ComfyUI client — execute Stable Diffusion workflows via the ComfyUI API. Connects to a ComfyUI instance to queue prompts, track execution via WebSocket or polling, and download generated images. Supports both UI-format and API-format workflows with automatic conversion.",
@@ -10415,14 +10213,216 @@ setBuildTimeData('clients.comfyui', {
10415
10213
  "description": "Event emitted by ComfyUIClient",
10416
10214
  "arguments": {}
10417
10215
  },
10418
- "execution_cached": {
10419
- "name": "execution_cached",
10420
- "description": "Event emitted by ComfyUIClient",
10216
+ "execution_cached": {
10217
+ "name": "execution_cached",
10218
+ "description": "Event emitted by ComfyUIClient",
10219
+ "arguments": {}
10220
+ },
10221
+ "execution_error": {
10222
+ "name": "execution_error",
10223
+ "description": "Event emitted by ComfyUIClient",
10224
+ "arguments": {}
10225
+ }
10226
+ },
10227
+ "state": {},
10228
+ "options": {},
10229
+ "envVars": [],
10230
+ "examples": [
10231
+ {
10232
+ "language": "ts",
10233
+ "code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
10234
+ }
10235
+ ]
10236
+ });
10237
+
10238
+ setBuildTimeData('clients.elevenlabs', {
10239
+ "id": "clients.elevenlabs",
10240
+ "description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
10241
+ "shortcut": "clients.elevenlabs",
10242
+ "className": "ElevenLabsClient",
10243
+ "methods": {
10244
+ "beforeRequest": {
10245
+ "description": "Inject the xi-api-key header before each request.",
10246
+ "parameters": {},
10247
+ "required": [],
10248
+ "returns": "void"
10249
+ },
10250
+ "connect": {
10251
+ "description": "Validate the API key by listing available models.",
10252
+ "parameters": {},
10253
+ "required": [],
10254
+ "returns": "Promise<this>",
10255
+ "examples": [
10256
+ {
10257
+ "language": "ts",
10258
+ "code": "await el.connect()"
10259
+ }
10260
+ ]
10261
+ },
10262
+ "listVoices": {
10263
+ "description": "List available voices with optional search and filtering.",
10264
+ "parameters": {
10265
+ "options": {
10266
+ "type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
10267
+ "description": "Query parameters for filtering voices"
10268
+ }
10269
+ },
10270
+ "required": [],
10271
+ "returns": "Promise<any>",
10272
+ "examples": [
10273
+ {
10274
+ "language": "ts",
10275
+ "code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
10276
+ }
10277
+ ]
10278
+ },
10279
+ "getVoice": {
10280
+ "description": "Get details for a single voice.",
10281
+ "parameters": {
10282
+ "voiceId": {
10283
+ "type": "string",
10284
+ "description": "The voice ID to look up"
10285
+ }
10286
+ },
10287
+ "required": [
10288
+ "voiceId"
10289
+ ],
10290
+ "returns": "Promise<any>",
10291
+ "examples": [
10292
+ {
10293
+ "language": "ts",
10294
+ "code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
10295
+ }
10296
+ ]
10297
+ },
10298
+ "listModels": {
10299
+ "description": "List available TTS models.",
10300
+ "parameters": {},
10301
+ "required": [],
10302
+ "returns": "Promise<any[]>",
10303
+ "examples": [
10304
+ {
10305
+ "language": "ts",
10306
+ "code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
10307
+ }
10308
+ ]
10309
+ },
10310
+ "synthesize": {
10311
+ "description": "Synthesize speech from text, returning audio as a Buffer.",
10312
+ "parameters": {
10313
+ "text": {
10314
+ "type": "string",
10315
+ "description": "The text to convert to speech"
10316
+ },
10317
+ "options": {
10318
+ "type": "SynthesizeOptions",
10319
+ "description": "Voice, model, format, and voice settings overrides",
10320
+ "properties": {
10321
+ "voiceId": {
10322
+ "type": "string",
10323
+ "description": ""
10324
+ },
10325
+ "modelId": {
10326
+ "type": "string",
10327
+ "description": ""
10328
+ },
10329
+ "outputFormat": {
10330
+ "type": "string",
10331
+ "description": ""
10332
+ },
10333
+ "voiceSettings": {
10334
+ "type": "ElevenLabsVoiceSettings",
10335
+ "description": ""
10336
+ },
10337
+ "disableCache": {
10338
+ "type": "boolean",
10339
+ "description": ""
10340
+ }
10341
+ }
10342
+ }
10343
+ },
10344
+ "required": [
10345
+ "text"
10346
+ ],
10347
+ "returns": "Promise<Buffer>",
10348
+ "examples": [
10349
+ {
10350
+ "language": "ts",
10351
+ "code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
10352
+ }
10353
+ ]
10354
+ },
10355
+ "say": {
10356
+ "description": "Synthesize speech and write the audio to a file.",
10357
+ "parameters": {
10358
+ "text": {
10359
+ "type": "string",
10360
+ "description": "The text to convert to speech"
10361
+ },
10362
+ "outputPath": {
10363
+ "type": "string",
10364
+ "description": "File path to write the audio to"
10365
+ },
10366
+ "options": {
10367
+ "type": "SynthesizeOptions",
10368
+ "description": "Voice, model, format, and voice settings overrides",
10369
+ "properties": {
10370
+ "voiceId": {
10371
+ "type": "string",
10372
+ "description": ""
10373
+ },
10374
+ "modelId": {
10375
+ "type": "string",
10376
+ "description": ""
10377
+ },
10378
+ "outputFormat": {
10379
+ "type": "string",
10380
+ "description": ""
10381
+ },
10382
+ "voiceSettings": {
10383
+ "type": "ElevenLabsVoiceSettings",
10384
+ "description": ""
10385
+ },
10386
+ "disableCache": {
10387
+ "type": "boolean",
10388
+ "description": ""
10389
+ }
10390
+ }
10391
+ }
10392
+ },
10393
+ "required": [
10394
+ "text",
10395
+ "outputPath"
10396
+ ],
10397
+ "returns": "Promise<string>",
10398
+ "examples": [
10399
+ {
10400
+ "language": "ts",
10401
+ "code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
10402
+ }
10403
+ ]
10404
+ }
10405
+ },
10406
+ "getters": {
10407
+ "apiKey": {
10408
+ "description": "The resolved API key from options or environment.",
10409
+ "returns": "string"
10410
+ }
10411
+ },
10412
+ "events": {
10413
+ "failure": {
10414
+ "name": "failure",
10415
+ "description": "Event emitted by ElevenLabsClient",
10416
+ "arguments": {}
10417
+ },
10418
+ "voices": {
10419
+ "name": "voices",
10420
+ "description": "Event emitted by ElevenLabsClient",
10421
10421
  "arguments": {}
10422
10422
  },
10423
- "execution_error": {
10424
- "name": "execution_error",
10425
- "description": "Event emitted by ComfyUIClient",
10423
+ "speech": {
10424
+ "name": "speech",
10425
+ "description": "Event emitted by ElevenLabsClient",
10426
10426
  "arguments": {}
10427
10427
  }
10428
10428
  },
@@ -10432,7 +10432,7 @@ setBuildTimeData('clients.comfyui', {
10432
10432
  "examples": [
10433
10433
  {
10434
10434
  "language": "ts",
10435
- "code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
10435
+ "code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
10436
10436
  }
10437
10437
  ]
10438
10438
  });
@@ -10657,6 +10657,19 @@ setBuildTimeData('servers.express', {
10657
10657
  ],
10658
10658
  "returns": "Promise<this>"
10659
10659
  },
10660
+ "useEndpointModules": {
10661
+ "description": "",
10662
+ "parameters": {
10663
+ "modules": {
10664
+ "type": "EndpointModule[]",
10665
+ "description": "Parameter modules"
10666
+ }
10667
+ },
10668
+ "required": [
10669
+ "modules"
10670
+ ],
10671
+ "returns": "Promise<this>"
10672
+ },
10660
10673
  "serveOpenAPISpec": {
10661
10674
  "description": "",
10662
10675
  "parameters": {
@@ -23403,248 +23416,47 @@ export const introspectionData = [
23403
23416
  ],
23404
23417
  "returns": "void"
23405
23418
  },
23406
- "unsubscribeAll": {
23407
- "description": "Unsubscribe and remove all realtime channels.",
23408
- "parameters": {},
23409
- "required": [],
23410
- "returns": "void"
23411
- },
23412
- "connect": {
23413
- "description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
23414
- "parameters": {},
23415
- "required": [],
23416
- "returns": "void"
23417
- },
23418
- "disconnect": {
23419
- "description": "Disconnect by signing out and removing all realtime channels.",
23420
- "parameters": {},
23421
- "required": [],
23422
- "returns": "void"
23423
- }
23424
- },
23425
- "getters": {
23426
- "sdk": {
23427
- "description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
23428
- "returns": "SupabaseSDKClient<any, any>"
23429
- },
23430
- "storage": {
23431
- "description": "Returns the Supabase Storage client for managing buckets and files.",
23432
- "returns": "any"
23433
- },
23434
- "functions": {
23435
- "description": "Returns the Supabase Functions client.",
23436
- "returns": "any"
23437
- }
23438
- },
23439
- "events": {},
23440
- "state": {},
23441
- "options": {},
23442
- "envVars": [],
23443
- "examples": [
23444
- {
23445
- "language": "ts",
23446
- "code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
23447
- }
23448
- ]
23449
- },
23450
- {
23451
- "id": "clients.elevenlabs",
23452
- "description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
23453
- "shortcut": "clients.elevenlabs",
23454
- "className": "ElevenLabsClient",
23455
- "methods": {
23456
- "beforeRequest": {
23457
- "description": "Inject the xi-api-key header before each request.",
23458
- "parameters": {},
23459
- "required": [],
23460
- "returns": "void"
23461
- },
23462
- "connect": {
23463
- "description": "Validate the API key by listing available models.",
23464
- "parameters": {},
23465
- "required": [],
23466
- "returns": "Promise<this>",
23467
- "examples": [
23468
- {
23469
- "language": "ts",
23470
- "code": "await el.connect()"
23471
- }
23472
- ]
23473
- },
23474
- "listVoices": {
23475
- "description": "List available voices with optional search and filtering.",
23476
- "parameters": {
23477
- "options": {
23478
- "type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
23479
- "description": "Query parameters for filtering voices"
23480
- }
23481
- },
23482
- "required": [],
23483
- "returns": "Promise<any>",
23484
- "examples": [
23485
- {
23486
- "language": "ts",
23487
- "code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
23488
- }
23489
- ]
23490
- },
23491
- "getVoice": {
23492
- "description": "Get details for a single voice.",
23493
- "parameters": {
23494
- "voiceId": {
23495
- "type": "string",
23496
- "description": "The voice ID to look up"
23497
- }
23498
- },
23499
- "required": [
23500
- "voiceId"
23501
- ],
23502
- "returns": "Promise<any>",
23503
- "examples": [
23504
- {
23505
- "language": "ts",
23506
- "code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
23507
- }
23508
- ]
23509
- },
23510
- "listModels": {
23511
- "description": "List available TTS models.",
23512
- "parameters": {},
23513
- "required": [],
23514
- "returns": "Promise<any[]>",
23515
- "examples": [
23516
- {
23517
- "language": "ts",
23518
- "code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
23519
- }
23520
- ]
23521
- },
23522
- "synthesize": {
23523
- "description": "Synthesize speech from text, returning audio as a Buffer.",
23524
- "parameters": {
23525
- "text": {
23526
- "type": "string",
23527
- "description": "The text to convert to speech"
23528
- },
23529
- "options": {
23530
- "type": "SynthesizeOptions",
23531
- "description": "Voice, model, format, and voice settings overrides",
23532
- "properties": {
23533
- "voiceId": {
23534
- "type": "string",
23535
- "description": ""
23536
- },
23537
- "modelId": {
23538
- "type": "string",
23539
- "description": ""
23540
- },
23541
- "outputFormat": {
23542
- "type": "string",
23543
- "description": ""
23544
- },
23545
- "voiceSettings": {
23546
- "type": "ElevenLabsVoiceSettings",
23547
- "description": ""
23548
- },
23549
- "disableCache": {
23550
- "type": "boolean",
23551
- "description": ""
23552
- }
23553
- }
23554
- }
23555
- },
23556
- "required": [
23557
- "text"
23558
- ],
23559
- "returns": "Promise<Buffer>",
23560
- "examples": [
23561
- {
23562
- "language": "ts",
23563
- "code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
23564
- }
23565
- ]
23566
- },
23567
- "say": {
23568
- "description": "Synthesize speech and write the audio to a file.",
23569
- "parameters": {
23570
- "text": {
23571
- "type": "string",
23572
- "description": "The text to convert to speech"
23573
- },
23574
- "outputPath": {
23575
- "type": "string",
23576
- "description": "File path to write the audio to"
23577
- },
23578
- "options": {
23579
- "type": "SynthesizeOptions",
23580
- "description": "Voice, model, format, and voice settings overrides",
23581
- "properties": {
23582
- "voiceId": {
23583
- "type": "string",
23584
- "description": ""
23585
- },
23586
- "modelId": {
23587
- "type": "string",
23588
- "description": ""
23589
- },
23590
- "outputFormat": {
23591
- "type": "string",
23592
- "description": ""
23593
- },
23594
- "voiceSettings": {
23595
- "type": "ElevenLabsVoiceSettings",
23596
- "description": ""
23597
- },
23598
- "disableCache": {
23599
- "type": "boolean",
23600
- "description": ""
23601
- }
23602
- }
23603
- }
23604
- },
23605
- "required": [
23606
- "text",
23607
- "outputPath"
23608
- ],
23609
- "returns": "Promise<string>",
23610
- "examples": [
23611
- {
23612
- "language": "ts",
23613
- "code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
23614
- }
23615
- ]
23419
+ "unsubscribeAll": {
23420
+ "description": "Unsubscribe and remove all realtime channels.",
23421
+ "parameters": {},
23422
+ "required": [],
23423
+ "returns": "void"
23424
+ },
23425
+ "connect": {
23426
+ "description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
23427
+ "parameters": {},
23428
+ "required": [],
23429
+ "returns": "void"
23430
+ },
23431
+ "disconnect": {
23432
+ "description": "Disconnect by signing out and removing all realtime channels.",
23433
+ "parameters": {},
23434
+ "required": [],
23435
+ "returns": "void"
23616
23436
  }
23617
23437
  },
23618
23438
  "getters": {
23619
- "apiKey": {
23620
- "description": "The resolved API key from options or environment.",
23621
- "returns": "string"
23622
- }
23623
- },
23624
- "events": {
23625
- "failure": {
23626
- "name": "failure",
23627
- "description": "Event emitted by ElevenLabsClient",
23628
- "arguments": {}
23439
+ "sdk": {
23440
+ "description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
23441
+ "returns": "SupabaseSDKClient<any, any>"
23629
23442
  },
23630
- "voices": {
23631
- "name": "voices",
23632
- "description": "Event emitted by ElevenLabsClient",
23633
- "arguments": {}
23443
+ "storage": {
23444
+ "description": "Returns the Supabase Storage client for managing buckets and files.",
23445
+ "returns": "any"
23634
23446
  },
23635
- "speech": {
23636
- "name": "speech",
23637
- "description": "Event emitted by ElevenLabsClient",
23638
- "arguments": {}
23447
+ "functions": {
23448
+ "description": "Returns the Supabase Functions client.",
23449
+ "returns": "any"
23639
23450
  }
23640
23451
  },
23452
+ "events": {},
23641
23453
  "state": {},
23642
23454
  "options": {},
23643
23455
  "envVars": [],
23644
23456
  "examples": [
23645
23457
  {
23646
23458
  "language": "ts",
23647
- "code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
23459
+ "code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
23648
23460
  }
23649
23461
  ]
23650
23462
  },
@@ -23899,6 +23711,207 @@ export const introspectionData = [
23899
23711
  }
23900
23712
  ]
23901
23713
  },
23714
+ {
23715
+ "id": "clients.elevenlabs",
23716
+ "description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
23717
+ "shortcut": "clients.elevenlabs",
23718
+ "className": "ElevenLabsClient",
23719
+ "methods": {
23720
+ "beforeRequest": {
23721
+ "description": "Inject the xi-api-key header before each request.",
23722
+ "parameters": {},
23723
+ "required": [],
23724
+ "returns": "void"
23725
+ },
23726
+ "connect": {
23727
+ "description": "Validate the API key by listing available models.",
23728
+ "parameters": {},
23729
+ "required": [],
23730
+ "returns": "Promise<this>",
23731
+ "examples": [
23732
+ {
23733
+ "language": "ts",
23734
+ "code": "await el.connect()"
23735
+ }
23736
+ ]
23737
+ },
23738
+ "listVoices": {
23739
+ "description": "List available voices with optional search and filtering.",
23740
+ "parameters": {
23741
+ "options": {
23742
+ "type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
23743
+ "description": "Query parameters for filtering voices"
23744
+ }
23745
+ },
23746
+ "required": [],
23747
+ "returns": "Promise<any>",
23748
+ "examples": [
23749
+ {
23750
+ "language": "ts",
23751
+ "code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
23752
+ }
23753
+ ]
23754
+ },
23755
+ "getVoice": {
23756
+ "description": "Get details for a single voice.",
23757
+ "parameters": {
23758
+ "voiceId": {
23759
+ "type": "string",
23760
+ "description": "The voice ID to look up"
23761
+ }
23762
+ },
23763
+ "required": [
23764
+ "voiceId"
23765
+ ],
23766
+ "returns": "Promise<any>",
23767
+ "examples": [
23768
+ {
23769
+ "language": "ts",
23770
+ "code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
23771
+ }
23772
+ ]
23773
+ },
23774
+ "listModels": {
23775
+ "description": "List available TTS models.",
23776
+ "parameters": {},
23777
+ "required": [],
23778
+ "returns": "Promise<any[]>",
23779
+ "examples": [
23780
+ {
23781
+ "language": "ts",
23782
+ "code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
23783
+ }
23784
+ ]
23785
+ },
23786
+ "synthesize": {
23787
+ "description": "Synthesize speech from text, returning audio as a Buffer.",
23788
+ "parameters": {
23789
+ "text": {
23790
+ "type": "string",
23791
+ "description": "The text to convert to speech"
23792
+ },
23793
+ "options": {
23794
+ "type": "SynthesizeOptions",
23795
+ "description": "Voice, model, format, and voice settings overrides",
23796
+ "properties": {
23797
+ "voiceId": {
23798
+ "type": "string",
23799
+ "description": ""
23800
+ },
23801
+ "modelId": {
23802
+ "type": "string",
23803
+ "description": ""
23804
+ },
23805
+ "outputFormat": {
23806
+ "type": "string",
23807
+ "description": ""
23808
+ },
23809
+ "voiceSettings": {
23810
+ "type": "ElevenLabsVoiceSettings",
23811
+ "description": ""
23812
+ },
23813
+ "disableCache": {
23814
+ "type": "boolean",
23815
+ "description": ""
23816
+ }
23817
+ }
23818
+ }
23819
+ },
23820
+ "required": [
23821
+ "text"
23822
+ ],
23823
+ "returns": "Promise<Buffer>",
23824
+ "examples": [
23825
+ {
23826
+ "language": "ts",
23827
+ "code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
23828
+ }
23829
+ ]
23830
+ },
23831
+ "say": {
23832
+ "description": "Synthesize speech and write the audio to a file.",
23833
+ "parameters": {
23834
+ "text": {
23835
+ "type": "string",
23836
+ "description": "The text to convert to speech"
23837
+ },
23838
+ "outputPath": {
23839
+ "type": "string",
23840
+ "description": "File path to write the audio to"
23841
+ },
23842
+ "options": {
23843
+ "type": "SynthesizeOptions",
23844
+ "description": "Voice, model, format, and voice settings overrides",
23845
+ "properties": {
23846
+ "voiceId": {
23847
+ "type": "string",
23848
+ "description": ""
23849
+ },
23850
+ "modelId": {
23851
+ "type": "string",
23852
+ "description": ""
23853
+ },
23854
+ "outputFormat": {
23855
+ "type": "string",
23856
+ "description": ""
23857
+ },
23858
+ "voiceSettings": {
23859
+ "type": "ElevenLabsVoiceSettings",
23860
+ "description": ""
23861
+ },
23862
+ "disableCache": {
23863
+ "type": "boolean",
23864
+ "description": ""
23865
+ }
23866
+ }
23867
+ }
23868
+ },
23869
+ "required": [
23870
+ "text",
23871
+ "outputPath"
23872
+ ],
23873
+ "returns": "Promise<string>",
23874
+ "examples": [
23875
+ {
23876
+ "language": "ts",
23877
+ "code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
23878
+ }
23879
+ ]
23880
+ }
23881
+ },
23882
+ "getters": {
23883
+ "apiKey": {
23884
+ "description": "The resolved API key from options or environment.",
23885
+ "returns": "string"
23886
+ }
23887
+ },
23888
+ "events": {
23889
+ "failure": {
23890
+ "name": "failure",
23891
+ "description": "Event emitted by ElevenLabsClient",
23892
+ "arguments": {}
23893
+ },
23894
+ "voices": {
23895
+ "name": "voices",
23896
+ "description": "Event emitted by ElevenLabsClient",
23897
+ "arguments": {}
23898
+ },
23899
+ "speech": {
23900
+ "name": "speech",
23901
+ "description": "Event emitted by ElevenLabsClient",
23902
+ "arguments": {}
23903
+ }
23904
+ },
23905
+ "state": {},
23906
+ "options": {},
23907
+ "envVars": [],
23908
+ "examples": [
23909
+ {
23910
+ "language": "ts",
23911
+ "code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
23912
+ }
23913
+ ]
23914
+ },
23902
23915
  {
23903
23916
  "id": "servers.mcp",
23904
23917
  "description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
@@ -24118,6 +24131,19 @@ export const introspectionData = [
24118
24131
  ],
24119
24132
  "returns": "Promise<this>"
24120
24133
  },
24134
+ "useEndpointModules": {
24135
+ "description": "",
24136
+ "parameters": {
24137
+ "modules": {
24138
+ "type": "EndpointModule[]",
24139
+ "description": "Parameter modules"
24140
+ }
24141
+ },
24142
+ "required": [
24143
+ "modules"
24144
+ ],
24145
+ "returns": "Promise<this>"
24146
+ },
24121
24147
  "serveOpenAPISpec": {
24122
24148
  "description": "",
24123
24149
  "parameters": {