@soederpop/luca 0.0.25 → 0.0.26
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/src/agi/features/assistant.ts +14 -12
- package/src/agi/features/docs-reader.ts +25 -1
- package/src/bootstrap/generated.ts +1 -1
- package/src/cli/build-info.ts +2 -2
- package/src/command.ts +75 -0
- package/src/commands/describe.ts +29 -1089
- package/src/container-describer.ts +1098 -0
- package/src/container.ts +11 -0
- package/src/introspection/generated.agi.ts +513 -429
- package/src/introspection/generated.node.ts +503 -419
- package/src/introspection/generated.web.ts +9 -1
- package/src/node/features/content-db.ts +17 -0
- package/src/node/features/fs.ts +18 -0
- package/src/scaffolds/generated.ts +1 -1
- package/src/server.ts +40 -0
- package/src/servers/express.ts +2 -0
- package/src/servers/mcp.ts +1 -0
- package/src/servers/socket.ts +2 -0
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
|
|
2
2
|
|
|
3
3
|
// Auto-generated introspection registry data
|
|
4
|
-
// Generated at: 2026-03-
|
|
4
|
+
// Generated at: 2026-03-22T20:57:23.101Z
|
|
5
5
|
|
|
6
6
|
setBuildTimeData('features.googleDocs', {
|
|
7
7
|
"id": "features.googleDocs",
|
|
@@ -5298,6 +5298,23 @@ setBuildTimeData('features.fs', {
|
|
|
5298
5298
|
}
|
|
5299
5299
|
]
|
|
5300
5300
|
},
|
|
5301
|
+
"readFileSync": {
|
|
5302
|
+
"description": "Synchronously reads a file and returns its contents as a string. added this method because AI Assistants are understandly confused by this deviation from 2000's era node style",
|
|
5303
|
+
"parameters": {
|
|
5304
|
+
"path": {
|
|
5305
|
+
"type": "string",
|
|
5306
|
+
"description": "Parameter path"
|
|
5307
|
+
},
|
|
5308
|
+
"encoding": {
|
|
5309
|
+
"type": "BufferEncoding | null",
|
|
5310
|
+
"description": "Parameter encoding"
|
|
5311
|
+
}
|
|
5312
|
+
},
|
|
5313
|
+
"required": [
|
|
5314
|
+
"path"
|
|
5315
|
+
],
|
|
5316
|
+
"returns": "string | Buffer"
|
|
5317
|
+
},
|
|
5301
5318
|
"readFileAsync": {
|
|
5302
5319
|
"description": "Asynchronously reads a file and returns its contents as a string.",
|
|
5303
5320
|
"parameters": {
|
|
@@ -5340,6 +5357,19 @@ setBuildTimeData('features.fs', {
|
|
|
5340
5357
|
}
|
|
5341
5358
|
]
|
|
5342
5359
|
},
|
|
5360
|
+
"readJsonSync": {
|
|
5361
|
+
"description": "Read and parse a JSON file synchronously",
|
|
5362
|
+
"parameters": {
|
|
5363
|
+
"path": {
|
|
5364
|
+
"type": "string",
|
|
5365
|
+
"description": "Parameter path"
|
|
5366
|
+
}
|
|
5367
|
+
},
|
|
5368
|
+
"required": [
|
|
5369
|
+
"path"
|
|
5370
|
+
],
|
|
5371
|
+
"returns": "void"
|
|
5372
|
+
},
|
|
5343
5373
|
"readJsonAsync": {
|
|
5344
5374
|
"description": "Asynchronously reads and parses a JSON file.",
|
|
5345
5375
|
"parameters": {
|
|
@@ -9495,6 +9525,14 @@ setBuildTimeData('features.contentDb', {
|
|
|
9495
9525
|
"description": "Returns the available document ids in the collection",
|
|
9496
9526
|
"returns": "string[]"
|
|
9497
9527
|
},
|
|
9528
|
+
"modelDefinitionTable": {
|
|
9529
|
+
"description": "",
|
|
9530
|
+
"returns": "any"
|
|
9531
|
+
},
|
|
9532
|
+
"fileTree": {
|
|
9533
|
+
"description": "",
|
|
9534
|
+
"returns": "any"
|
|
9535
|
+
},
|
|
9498
9536
|
"searchIndexStatus": {
|
|
9499
9537
|
"description": "Get the current search index status.",
|
|
9500
9538
|
"returns": "any"
|
|
@@ -10109,6 +10147,208 @@ setBuildTimeData('clients.openai', {
|
|
|
10109
10147
|
]
|
|
10110
10148
|
});
|
|
10111
10149
|
|
|
10150
|
+
setBuildTimeData('clients.elevenlabs', {
|
|
10151
|
+
"id": "clients.elevenlabs",
|
|
10152
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
10153
|
+
"shortcut": "clients.elevenlabs",
|
|
10154
|
+
"className": "ElevenLabsClient",
|
|
10155
|
+
"methods": {
|
|
10156
|
+
"beforeRequest": {
|
|
10157
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
10158
|
+
"parameters": {},
|
|
10159
|
+
"required": [],
|
|
10160
|
+
"returns": "void"
|
|
10161
|
+
},
|
|
10162
|
+
"connect": {
|
|
10163
|
+
"description": "Validate the API key by listing available models.",
|
|
10164
|
+
"parameters": {},
|
|
10165
|
+
"required": [],
|
|
10166
|
+
"returns": "Promise<this>",
|
|
10167
|
+
"examples": [
|
|
10168
|
+
{
|
|
10169
|
+
"language": "ts",
|
|
10170
|
+
"code": "await el.connect()"
|
|
10171
|
+
}
|
|
10172
|
+
]
|
|
10173
|
+
},
|
|
10174
|
+
"listVoices": {
|
|
10175
|
+
"description": "List available voices with optional search and filtering.",
|
|
10176
|
+
"parameters": {
|
|
10177
|
+
"options": {
|
|
10178
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10179
|
+
"description": "Query parameters for filtering voices"
|
|
10180
|
+
}
|
|
10181
|
+
},
|
|
10182
|
+
"required": [],
|
|
10183
|
+
"returns": "Promise<any>",
|
|
10184
|
+
"examples": [
|
|
10185
|
+
{
|
|
10186
|
+
"language": "ts",
|
|
10187
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10188
|
+
}
|
|
10189
|
+
]
|
|
10190
|
+
},
|
|
10191
|
+
"getVoice": {
|
|
10192
|
+
"description": "Get details for a single voice.",
|
|
10193
|
+
"parameters": {
|
|
10194
|
+
"voiceId": {
|
|
10195
|
+
"type": "string",
|
|
10196
|
+
"description": "The voice ID to look up"
|
|
10197
|
+
}
|
|
10198
|
+
},
|
|
10199
|
+
"required": [
|
|
10200
|
+
"voiceId"
|
|
10201
|
+
],
|
|
10202
|
+
"returns": "Promise<any>",
|
|
10203
|
+
"examples": [
|
|
10204
|
+
{
|
|
10205
|
+
"language": "ts",
|
|
10206
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10207
|
+
}
|
|
10208
|
+
]
|
|
10209
|
+
},
|
|
10210
|
+
"listModels": {
|
|
10211
|
+
"description": "List available TTS models.",
|
|
10212
|
+
"parameters": {},
|
|
10213
|
+
"required": [],
|
|
10214
|
+
"returns": "Promise<any[]>",
|
|
10215
|
+
"examples": [
|
|
10216
|
+
{
|
|
10217
|
+
"language": "ts",
|
|
10218
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10219
|
+
}
|
|
10220
|
+
]
|
|
10221
|
+
},
|
|
10222
|
+
"synthesize": {
|
|
10223
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10224
|
+
"parameters": {
|
|
10225
|
+
"text": {
|
|
10226
|
+
"type": "string",
|
|
10227
|
+
"description": "The text to convert to speech"
|
|
10228
|
+
},
|
|
10229
|
+
"options": {
|
|
10230
|
+
"type": "SynthesizeOptions",
|
|
10231
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10232
|
+
"properties": {
|
|
10233
|
+
"voiceId": {
|
|
10234
|
+
"type": "string",
|
|
10235
|
+
"description": ""
|
|
10236
|
+
},
|
|
10237
|
+
"modelId": {
|
|
10238
|
+
"type": "string",
|
|
10239
|
+
"description": ""
|
|
10240
|
+
},
|
|
10241
|
+
"outputFormat": {
|
|
10242
|
+
"type": "string",
|
|
10243
|
+
"description": ""
|
|
10244
|
+
},
|
|
10245
|
+
"voiceSettings": {
|
|
10246
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10247
|
+
"description": ""
|
|
10248
|
+
},
|
|
10249
|
+
"disableCache": {
|
|
10250
|
+
"type": "boolean",
|
|
10251
|
+
"description": ""
|
|
10252
|
+
}
|
|
10253
|
+
}
|
|
10254
|
+
}
|
|
10255
|
+
},
|
|
10256
|
+
"required": [
|
|
10257
|
+
"text"
|
|
10258
|
+
],
|
|
10259
|
+
"returns": "Promise<Buffer>",
|
|
10260
|
+
"examples": [
|
|
10261
|
+
{
|
|
10262
|
+
"language": "ts",
|
|
10263
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10264
|
+
}
|
|
10265
|
+
]
|
|
10266
|
+
},
|
|
10267
|
+
"say": {
|
|
10268
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
10269
|
+
"parameters": {
|
|
10270
|
+
"text": {
|
|
10271
|
+
"type": "string",
|
|
10272
|
+
"description": "The text to convert to speech"
|
|
10273
|
+
},
|
|
10274
|
+
"outputPath": {
|
|
10275
|
+
"type": "string",
|
|
10276
|
+
"description": "File path to write the audio to"
|
|
10277
|
+
},
|
|
10278
|
+
"options": {
|
|
10279
|
+
"type": "SynthesizeOptions",
|
|
10280
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
10281
|
+
"properties": {
|
|
10282
|
+
"voiceId": {
|
|
10283
|
+
"type": "string",
|
|
10284
|
+
"description": ""
|
|
10285
|
+
},
|
|
10286
|
+
"modelId": {
|
|
10287
|
+
"type": "string",
|
|
10288
|
+
"description": ""
|
|
10289
|
+
},
|
|
10290
|
+
"outputFormat": {
|
|
10291
|
+
"type": "string",
|
|
10292
|
+
"description": ""
|
|
10293
|
+
},
|
|
10294
|
+
"voiceSettings": {
|
|
10295
|
+
"type": "ElevenLabsVoiceSettings",
|
|
10296
|
+
"description": ""
|
|
10297
|
+
},
|
|
10298
|
+
"disableCache": {
|
|
10299
|
+
"type": "boolean",
|
|
10300
|
+
"description": ""
|
|
10301
|
+
}
|
|
10302
|
+
}
|
|
10303
|
+
}
|
|
10304
|
+
},
|
|
10305
|
+
"required": [
|
|
10306
|
+
"text",
|
|
10307
|
+
"outputPath"
|
|
10308
|
+
],
|
|
10309
|
+
"returns": "Promise<string>",
|
|
10310
|
+
"examples": [
|
|
10311
|
+
{
|
|
10312
|
+
"language": "ts",
|
|
10313
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10314
|
+
}
|
|
10315
|
+
]
|
|
10316
|
+
}
|
|
10317
|
+
},
|
|
10318
|
+
"getters": {
|
|
10319
|
+
"apiKey": {
|
|
10320
|
+
"description": "The resolved API key from options or environment.",
|
|
10321
|
+
"returns": "string"
|
|
10322
|
+
}
|
|
10323
|
+
},
|
|
10324
|
+
"events": {
|
|
10325
|
+
"failure": {
|
|
10326
|
+
"name": "failure",
|
|
10327
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10328
|
+
"arguments": {}
|
|
10329
|
+
},
|
|
10330
|
+
"voices": {
|
|
10331
|
+
"name": "voices",
|
|
10332
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10333
|
+
"arguments": {}
|
|
10334
|
+
},
|
|
10335
|
+
"speech": {
|
|
10336
|
+
"name": "speech",
|
|
10337
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
10338
|
+
"arguments": {}
|
|
10339
|
+
}
|
|
10340
|
+
},
|
|
10341
|
+
"state": {},
|
|
10342
|
+
"options": {},
|
|
10343
|
+
"envVars": [],
|
|
10344
|
+
"examples": [
|
|
10345
|
+
{
|
|
10346
|
+
"language": "ts",
|
|
10347
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10348
|
+
}
|
|
10349
|
+
]
|
|
10350
|
+
});
|
|
10351
|
+
|
|
10112
10352
|
setBuildTimeData('clients.supabase', {
|
|
10113
10353
|
"id": "clients.supabase",
|
|
10114
10354
|
"description": "Supabase client for the Luca container system. Wraps the official `@supabase/supabase-js` SDK and exposes it through Luca's typed state, events, and introspection system. The SDK is isomorphic so this single implementation works in both Node and browser containers. Use `client.sdk` for full SDK access, or use the convenience wrappers for common operations (auth, database queries, storage, edge functions, realtime).",
|
|
@@ -10557,213 +10797,11 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10557
10797
|
]
|
|
10558
10798
|
});
|
|
10559
10799
|
|
|
10560
|
-
setBuildTimeData('
|
|
10561
|
-
"id": "
|
|
10562
|
-
"description": "
|
|
10563
|
-
"shortcut": "
|
|
10564
|
-
"className": "
|
|
10565
|
-
"methods": {
|
|
10566
|
-
"beforeRequest": {
|
|
10567
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
10568
|
-
"parameters": {},
|
|
10569
|
-
"required": [],
|
|
10570
|
-
"returns": "void"
|
|
10571
|
-
},
|
|
10572
|
-
"connect": {
|
|
10573
|
-
"description": "Validate the API key by listing available models.",
|
|
10574
|
-
"parameters": {},
|
|
10575
|
-
"required": [],
|
|
10576
|
-
"returns": "Promise<this>",
|
|
10577
|
-
"examples": [
|
|
10578
|
-
{
|
|
10579
|
-
"language": "ts",
|
|
10580
|
-
"code": "await el.connect()"
|
|
10581
|
-
}
|
|
10582
|
-
]
|
|
10583
|
-
},
|
|
10584
|
-
"listVoices": {
|
|
10585
|
-
"description": "List available voices with optional search and filtering.",
|
|
10586
|
-
"parameters": {
|
|
10587
|
-
"options": {
|
|
10588
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
10589
|
-
"description": "Query parameters for filtering voices"
|
|
10590
|
-
}
|
|
10591
|
-
},
|
|
10592
|
-
"required": [],
|
|
10593
|
-
"returns": "Promise<any>",
|
|
10594
|
-
"examples": [
|
|
10595
|
-
{
|
|
10596
|
-
"language": "ts",
|
|
10597
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
10598
|
-
}
|
|
10599
|
-
]
|
|
10600
|
-
},
|
|
10601
|
-
"getVoice": {
|
|
10602
|
-
"description": "Get details for a single voice.",
|
|
10603
|
-
"parameters": {
|
|
10604
|
-
"voiceId": {
|
|
10605
|
-
"type": "string",
|
|
10606
|
-
"description": "The voice ID to look up"
|
|
10607
|
-
}
|
|
10608
|
-
},
|
|
10609
|
-
"required": [
|
|
10610
|
-
"voiceId"
|
|
10611
|
-
],
|
|
10612
|
-
"returns": "Promise<any>",
|
|
10613
|
-
"examples": [
|
|
10614
|
-
{
|
|
10615
|
-
"language": "ts",
|
|
10616
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
10617
|
-
}
|
|
10618
|
-
]
|
|
10619
|
-
},
|
|
10620
|
-
"listModels": {
|
|
10621
|
-
"description": "List available TTS models.",
|
|
10622
|
-
"parameters": {},
|
|
10623
|
-
"required": [],
|
|
10624
|
-
"returns": "Promise<any[]>",
|
|
10625
|
-
"examples": [
|
|
10626
|
-
{
|
|
10627
|
-
"language": "ts",
|
|
10628
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
10629
|
-
}
|
|
10630
|
-
]
|
|
10631
|
-
},
|
|
10632
|
-
"synthesize": {
|
|
10633
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
10634
|
-
"parameters": {
|
|
10635
|
-
"text": {
|
|
10636
|
-
"type": "string",
|
|
10637
|
-
"description": "The text to convert to speech"
|
|
10638
|
-
},
|
|
10639
|
-
"options": {
|
|
10640
|
-
"type": "SynthesizeOptions",
|
|
10641
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10642
|
-
"properties": {
|
|
10643
|
-
"voiceId": {
|
|
10644
|
-
"type": "string",
|
|
10645
|
-
"description": ""
|
|
10646
|
-
},
|
|
10647
|
-
"modelId": {
|
|
10648
|
-
"type": "string",
|
|
10649
|
-
"description": ""
|
|
10650
|
-
},
|
|
10651
|
-
"outputFormat": {
|
|
10652
|
-
"type": "string",
|
|
10653
|
-
"description": ""
|
|
10654
|
-
},
|
|
10655
|
-
"voiceSettings": {
|
|
10656
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10657
|
-
"description": ""
|
|
10658
|
-
},
|
|
10659
|
-
"disableCache": {
|
|
10660
|
-
"type": "boolean",
|
|
10661
|
-
"description": ""
|
|
10662
|
-
}
|
|
10663
|
-
}
|
|
10664
|
-
}
|
|
10665
|
-
},
|
|
10666
|
-
"required": [
|
|
10667
|
-
"text"
|
|
10668
|
-
],
|
|
10669
|
-
"returns": "Promise<Buffer>",
|
|
10670
|
-
"examples": [
|
|
10671
|
-
{
|
|
10672
|
-
"language": "ts",
|
|
10673
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
10674
|
-
}
|
|
10675
|
-
]
|
|
10676
|
-
},
|
|
10677
|
-
"say": {
|
|
10678
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
10679
|
-
"parameters": {
|
|
10680
|
-
"text": {
|
|
10681
|
-
"type": "string",
|
|
10682
|
-
"description": "The text to convert to speech"
|
|
10683
|
-
},
|
|
10684
|
-
"outputPath": {
|
|
10685
|
-
"type": "string",
|
|
10686
|
-
"description": "File path to write the audio to"
|
|
10687
|
-
},
|
|
10688
|
-
"options": {
|
|
10689
|
-
"type": "SynthesizeOptions",
|
|
10690
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
10691
|
-
"properties": {
|
|
10692
|
-
"voiceId": {
|
|
10693
|
-
"type": "string",
|
|
10694
|
-
"description": ""
|
|
10695
|
-
},
|
|
10696
|
-
"modelId": {
|
|
10697
|
-
"type": "string",
|
|
10698
|
-
"description": ""
|
|
10699
|
-
},
|
|
10700
|
-
"outputFormat": {
|
|
10701
|
-
"type": "string",
|
|
10702
|
-
"description": ""
|
|
10703
|
-
},
|
|
10704
|
-
"voiceSettings": {
|
|
10705
|
-
"type": "ElevenLabsVoiceSettings",
|
|
10706
|
-
"description": ""
|
|
10707
|
-
},
|
|
10708
|
-
"disableCache": {
|
|
10709
|
-
"type": "boolean",
|
|
10710
|
-
"description": ""
|
|
10711
|
-
}
|
|
10712
|
-
}
|
|
10713
|
-
}
|
|
10714
|
-
},
|
|
10715
|
-
"required": [
|
|
10716
|
-
"text",
|
|
10717
|
-
"outputPath"
|
|
10718
|
-
],
|
|
10719
|
-
"returns": "Promise<string>",
|
|
10720
|
-
"examples": [
|
|
10721
|
-
{
|
|
10722
|
-
"language": "ts",
|
|
10723
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
10724
|
-
}
|
|
10725
|
-
]
|
|
10726
|
-
}
|
|
10727
|
-
},
|
|
10728
|
-
"getters": {
|
|
10729
|
-
"apiKey": {
|
|
10730
|
-
"description": "The resolved API key from options or environment.",
|
|
10731
|
-
"returns": "string"
|
|
10732
|
-
}
|
|
10733
|
-
},
|
|
10734
|
-
"events": {
|
|
10735
|
-
"failure": {
|
|
10736
|
-
"name": "failure",
|
|
10737
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10738
|
-
"arguments": {}
|
|
10739
|
-
},
|
|
10740
|
-
"voices": {
|
|
10741
|
-
"name": "voices",
|
|
10742
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10743
|
-
"arguments": {}
|
|
10744
|
-
},
|
|
10745
|
-
"speech": {
|
|
10746
|
-
"name": "speech",
|
|
10747
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
10748
|
-
"arguments": {}
|
|
10749
|
-
}
|
|
10750
|
-
},
|
|
10751
|
-
"state": {},
|
|
10752
|
-
"options": {},
|
|
10753
|
-
"envVars": [],
|
|
10754
|
-
"examples": [
|
|
10755
|
-
{
|
|
10756
|
-
"language": "ts",
|
|
10757
|
-
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
10758
|
-
}
|
|
10759
|
-
]
|
|
10760
|
-
});
|
|
10761
|
-
|
|
10762
|
-
setBuildTimeData('servers.mcp', {
|
|
10763
|
-
"id": "servers.mcp",
|
|
10764
|
-
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|
|
10765
|
-
"shortcut": "servers.mcp",
|
|
10766
|
-
"className": "MCPServer",
|
|
10800
|
+
setBuildTimeData('servers.mcp', {
|
|
10801
|
+
"id": "servers.mcp",
|
|
10802
|
+
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|
|
10803
|
+
"shortcut": "servers.mcp",
|
|
10804
|
+
"className": "MCPServer",
|
|
10767
10805
|
"methods": {
|
|
10768
10806
|
"tool": {
|
|
10769
10807
|
"description": "Register an MCP tool. The tool's Zod schema is converted to JSON Schema for the protocol listing, and used for runtime argument validation. Tool handlers can return a string (auto-wrapped as text content) or a full CallToolResult object for advanced responses (images, errors, etc).",
|
|
@@ -11479,6 +11517,10 @@ setContainerBuildTimeData('Container', {
|
|
|
11479
11517
|
"description": "Returns a map of enabled feature shortcut IDs to their instances.",
|
|
11480
11518
|
"returns": "Partial<AvailableInstanceTypes<Features>>"
|
|
11481
11519
|
},
|
|
11520
|
+
"describer": {
|
|
11521
|
+
"description": "Lazy-initialized ContainerDescriber for introspecting registries, helpers, and members.",
|
|
11522
|
+
"returns": "ContainerDescriber"
|
|
11523
|
+
},
|
|
11482
11524
|
"context": {
|
|
11483
11525
|
"description": "The Container's context is an object that contains the enabled features, the container itself, and any additional context that has been added to the container. All helper instances that are created by the container will have access to the shared context.",
|
|
11484
11526
|
"returns": "ContainerContext<Features> & Partial<AvailableInstanceTypes<AvailableFeatures>>"
|
|
@@ -16829,6 +16871,23 @@ export const introspectionData = [
|
|
|
16829
16871
|
}
|
|
16830
16872
|
]
|
|
16831
16873
|
},
|
|
16874
|
+
"readFileSync": {
|
|
16875
|
+
"description": "Synchronously reads a file and returns its contents as a string. added this method because AI Assistants are understandly confused by this deviation from 2000's era node style",
|
|
16876
|
+
"parameters": {
|
|
16877
|
+
"path": {
|
|
16878
|
+
"type": "string",
|
|
16879
|
+
"description": "Parameter path"
|
|
16880
|
+
},
|
|
16881
|
+
"encoding": {
|
|
16882
|
+
"type": "BufferEncoding | null",
|
|
16883
|
+
"description": "Parameter encoding"
|
|
16884
|
+
}
|
|
16885
|
+
},
|
|
16886
|
+
"required": [
|
|
16887
|
+
"path"
|
|
16888
|
+
],
|
|
16889
|
+
"returns": "string | Buffer"
|
|
16890
|
+
},
|
|
16832
16891
|
"readFileAsync": {
|
|
16833
16892
|
"description": "Asynchronously reads a file and returns its contents as a string.",
|
|
16834
16893
|
"parameters": {
|
|
@@ -16871,6 +16930,19 @@ export const introspectionData = [
|
|
|
16871
16930
|
}
|
|
16872
16931
|
]
|
|
16873
16932
|
},
|
|
16933
|
+
"readJsonSync": {
|
|
16934
|
+
"description": "Read and parse a JSON file synchronously",
|
|
16935
|
+
"parameters": {
|
|
16936
|
+
"path": {
|
|
16937
|
+
"type": "string",
|
|
16938
|
+
"description": "Parameter path"
|
|
16939
|
+
}
|
|
16940
|
+
},
|
|
16941
|
+
"required": [
|
|
16942
|
+
"path"
|
|
16943
|
+
],
|
|
16944
|
+
"returns": "void"
|
|
16945
|
+
},
|
|
16874
16946
|
"readJsonAsync": {
|
|
16875
16947
|
"description": "Asynchronously reads and parses a JSON file.",
|
|
16876
16948
|
"parameters": {
|
|
@@ -21011,6 +21083,14 @@ export const introspectionData = [
|
|
|
21011
21083
|
"description": "Returns the available document ids in the collection",
|
|
21012
21084
|
"returns": "string[]"
|
|
21013
21085
|
},
|
|
21086
|
+
"modelDefinitionTable": {
|
|
21087
|
+
"description": "",
|
|
21088
|
+
"returns": "any"
|
|
21089
|
+
},
|
|
21090
|
+
"fileTree": {
|
|
21091
|
+
"description": "",
|
|
21092
|
+
"returns": "any"
|
|
21093
|
+
},
|
|
21014
21094
|
"searchIndexStatus": {
|
|
21015
21095
|
"description": "Get the current search index status.",
|
|
21016
21096
|
"returns": "any"
|
|
@@ -21621,20 +21701,221 @@ export const introspectionData = [
|
|
|
21621
21701
|
]
|
|
21622
21702
|
},
|
|
21623
21703
|
{
|
|
21624
|
-
"id": "clients.
|
|
21625
|
-
"description": "
|
|
21626
|
-
"shortcut": "clients.
|
|
21627
|
-
"className": "
|
|
21704
|
+
"id": "clients.elevenlabs",
|
|
21705
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
21706
|
+
"shortcut": "clients.elevenlabs",
|
|
21707
|
+
"className": "ElevenLabsClient",
|
|
21628
21708
|
"methods": {
|
|
21629
|
-
"
|
|
21630
|
-
"description": "
|
|
21709
|
+
"beforeRequest": {
|
|
21710
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
21711
|
+
"parameters": {},
|
|
21712
|
+
"required": [],
|
|
21713
|
+
"returns": "void"
|
|
21714
|
+
},
|
|
21715
|
+
"connect": {
|
|
21716
|
+
"description": "Validate the API key by listing available models.",
|
|
21717
|
+
"parameters": {},
|
|
21718
|
+
"required": [],
|
|
21719
|
+
"returns": "Promise<this>",
|
|
21720
|
+
"examples": [
|
|
21721
|
+
{
|
|
21722
|
+
"language": "ts",
|
|
21723
|
+
"code": "await el.connect()"
|
|
21724
|
+
}
|
|
21725
|
+
]
|
|
21726
|
+
},
|
|
21727
|
+
"listVoices": {
|
|
21728
|
+
"description": "List available voices with optional search and filtering.",
|
|
21631
21729
|
"parameters": {
|
|
21632
|
-
"
|
|
21633
|
-
"type": "string",
|
|
21634
|
-
"description": "
|
|
21730
|
+
"options": {
|
|
21731
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
21732
|
+
"description": "Query parameters for filtering voices"
|
|
21635
21733
|
}
|
|
21636
21734
|
},
|
|
21637
|
-
"required": [
|
|
21735
|
+
"required": [],
|
|
21736
|
+
"returns": "Promise<any>",
|
|
21737
|
+
"examples": [
|
|
21738
|
+
{
|
|
21739
|
+
"language": "ts",
|
|
21740
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
21741
|
+
}
|
|
21742
|
+
]
|
|
21743
|
+
},
|
|
21744
|
+
"getVoice": {
|
|
21745
|
+
"description": "Get details for a single voice.",
|
|
21746
|
+
"parameters": {
|
|
21747
|
+
"voiceId": {
|
|
21748
|
+
"type": "string",
|
|
21749
|
+
"description": "The voice ID to look up"
|
|
21750
|
+
}
|
|
21751
|
+
},
|
|
21752
|
+
"required": [
|
|
21753
|
+
"voiceId"
|
|
21754
|
+
],
|
|
21755
|
+
"returns": "Promise<any>",
|
|
21756
|
+
"examples": [
|
|
21757
|
+
{
|
|
21758
|
+
"language": "ts",
|
|
21759
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
21760
|
+
}
|
|
21761
|
+
]
|
|
21762
|
+
},
|
|
21763
|
+
"listModels": {
|
|
21764
|
+
"description": "List available TTS models.",
|
|
21765
|
+
"parameters": {},
|
|
21766
|
+
"required": [],
|
|
21767
|
+
"returns": "Promise<any[]>",
|
|
21768
|
+
"examples": [
|
|
21769
|
+
{
|
|
21770
|
+
"language": "ts",
|
|
21771
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
21772
|
+
}
|
|
21773
|
+
]
|
|
21774
|
+
},
|
|
21775
|
+
"synthesize": {
|
|
21776
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
21777
|
+
"parameters": {
|
|
21778
|
+
"text": {
|
|
21779
|
+
"type": "string",
|
|
21780
|
+
"description": "The text to convert to speech"
|
|
21781
|
+
},
|
|
21782
|
+
"options": {
|
|
21783
|
+
"type": "SynthesizeOptions",
|
|
21784
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
21785
|
+
"properties": {
|
|
21786
|
+
"voiceId": {
|
|
21787
|
+
"type": "string",
|
|
21788
|
+
"description": ""
|
|
21789
|
+
},
|
|
21790
|
+
"modelId": {
|
|
21791
|
+
"type": "string",
|
|
21792
|
+
"description": ""
|
|
21793
|
+
},
|
|
21794
|
+
"outputFormat": {
|
|
21795
|
+
"type": "string",
|
|
21796
|
+
"description": ""
|
|
21797
|
+
},
|
|
21798
|
+
"voiceSettings": {
|
|
21799
|
+
"type": "ElevenLabsVoiceSettings",
|
|
21800
|
+
"description": ""
|
|
21801
|
+
},
|
|
21802
|
+
"disableCache": {
|
|
21803
|
+
"type": "boolean",
|
|
21804
|
+
"description": ""
|
|
21805
|
+
}
|
|
21806
|
+
}
|
|
21807
|
+
}
|
|
21808
|
+
},
|
|
21809
|
+
"required": [
|
|
21810
|
+
"text"
|
|
21811
|
+
],
|
|
21812
|
+
"returns": "Promise<Buffer>",
|
|
21813
|
+
"examples": [
|
|
21814
|
+
{
|
|
21815
|
+
"language": "ts",
|
|
21816
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
21817
|
+
}
|
|
21818
|
+
]
|
|
21819
|
+
},
|
|
21820
|
+
"say": {
|
|
21821
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
21822
|
+
"parameters": {
|
|
21823
|
+
"text": {
|
|
21824
|
+
"type": "string",
|
|
21825
|
+
"description": "The text to convert to speech"
|
|
21826
|
+
},
|
|
21827
|
+
"outputPath": {
|
|
21828
|
+
"type": "string",
|
|
21829
|
+
"description": "File path to write the audio to"
|
|
21830
|
+
},
|
|
21831
|
+
"options": {
|
|
21832
|
+
"type": "SynthesizeOptions",
|
|
21833
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
21834
|
+
"properties": {
|
|
21835
|
+
"voiceId": {
|
|
21836
|
+
"type": "string",
|
|
21837
|
+
"description": ""
|
|
21838
|
+
},
|
|
21839
|
+
"modelId": {
|
|
21840
|
+
"type": "string",
|
|
21841
|
+
"description": ""
|
|
21842
|
+
},
|
|
21843
|
+
"outputFormat": {
|
|
21844
|
+
"type": "string",
|
|
21845
|
+
"description": ""
|
|
21846
|
+
},
|
|
21847
|
+
"voiceSettings": {
|
|
21848
|
+
"type": "ElevenLabsVoiceSettings",
|
|
21849
|
+
"description": ""
|
|
21850
|
+
},
|
|
21851
|
+
"disableCache": {
|
|
21852
|
+
"type": "boolean",
|
|
21853
|
+
"description": ""
|
|
21854
|
+
}
|
|
21855
|
+
}
|
|
21856
|
+
}
|
|
21857
|
+
},
|
|
21858
|
+
"required": [
|
|
21859
|
+
"text",
|
|
21860
|
+
"outputPath"
|
|
21861
|
+
],
|
|
21862
|
+
"returns": "Promise<string>",
|
|
21863
|
+
"examples": [
|
|
21864
|
+
{
|
|
21865
|
+
"language": "ts",
|
|
21866
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
21867
|
+
}
|
|
21868
|
+
]
|
|
21869
|
+
}
|
|
21870
|
+
},
|
|
21871
|
+
"getters": {
|
|
21872
|
+
"apiKey": {
|
|
21873
|
+
"description": "The resolved API key from options or environment.",
|
|
21874
|
+
"returns": "string"
|
|
21875
|
+
}
|
|
21876
|
+
},
|
|
21877
|
+
"events": {
|
|
21878
|
+
"failure": {
|
|
21879
|
+
"name": "failure",
|
|
21880
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21881
|
+
"arguments": {}
|
|
21882
|
+
},
|
|
21883
|
+
"voices": {
|
|
21884
|
+
"name": "voices",
|
|
21885
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21886
|
+
"arguments": {}
|
|
21887
|
+
},
|
|
21888
|
+
"speech": {
|
|
21889
|
+
"name": "speech",
|
|
21890
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
21891
|
+
"arguments": {}
|
|
21892
|
+
}
|
|
21893
|
+
},
|
|
21894
|
+
"state": {},
|
|
21895
|
+
"options": {},
|
|
21896
|
+
"envVars": [],
|
|
21897
|
+
"examples": [
|
|
21898
|
+
{
|
|
21899
|
+
"language": "ts",
|
|
21900
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
21901
|
+
}
|
|
21902
|
+
]
|
|
21903
|
+
},
|
|
21904
|
+
{
|
|
21905
|
+
"id": "clients.supabase",
|
|
21906
|
+
"description": "Supabase client for the Luca container system. Wraps the official `@supabase/supabase-js` SDK and exposes it through Luca's typed state, events, and introspection system. The SDK is isomorphic so this single implementation works in both Node and browser containers. Use `client.sdk` for full SDK access, or use the convenience wrappers for common operations (auth, database queries, storage, edge functions, realtime).",
|
|
21907
|
+
"shortcut": "clients.supabase",
|
|
21908
|
+
"className": "SupabaseClient",
|
|
21909
|
+
"methods": {
|
|
21910
|
+
"from": {
|
|
21911
|
+
"description": "Start a query on a Postgres table or view.",
|
|
21912
|
+
"parameters": {
|
|
21913
|
+
"table": {
|
|
21914
|
+
"type": "string",
|
|
21915
|
+
"description": "The table or view name to query"
|
|
21916
|
+
}
|
|
21917
|
+
},
|
|
21918
|
+
"required": [
|
|
21638
21919
|
"table"
|
|
21639
21920
|
],
|
|
21640
21921
|
"returns": "void"
|
|
@@ -22066,207 +22347,6 @@ export const introspectionData = [
|
|
|
22066
22347
|
}
|
|
22067
22348
|
]
|
|
22068
22349
|
},
|
|
22069
|
-
{
|
|
22070
|
-
"id": "clients.elevenlabs",
|
|
22071
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
22072
|
-
"shortcut": "clients.elevenlabs",
|
|
22073
|
-
"className": "ElevenLabsClient",
|
|
22074
|
-
"methods": {
|
|
22075
|
-
"beforeRequest": {
|
|
22076
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
22077
|
-
"parameters": {},
|
|
22078
|
-
"required": [],
|
|
22079
|
-
"returns": "void"
|
|
22080
|
-
},
|
|
22081
|
-
"connect": {
|
|
22082
|
-
"description": "Validate the API key by listing available models.",
|
|
22083
|
-
"parameters": {},
|
|
22084
|
-
"required": [],
|
|
22085
|
-
"returns": "Promise<this>",
|
|
22086
|
-
"examples": [
|
|
22087
|
-
{
|
|
22088
|
-
"language": "ts",
|
|
22089
|
-
"code": "await el.connect()"
|
|
22090
|
-
}
|
|
22091
|
-
]
|
|
22092
|
-
},
|
|
22093
|
-
"listVoices": {
|
|
22094
|
-
"description": "List available voices with optional search and filtering.",
|
|
22095
|
-
"parameters": {
|
|
22096
|
-
"options": {
|
|
22097
|
-
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
22098
|
-
"description": "Query parameters for filtering voices"
|
|
22099
|
-
}
|
|
22100
|
-
},
|
|
22101
|
-
"required": [],
|
|
22102
|
-
"returns": "Promise<any>",
|
|
22103
|
-
"examples": [
|
|
22104
|
-
{
|
|
22105
|
-
"language": "ts",
|
|
22106
|
-
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
22107
|
-
}
|
|
22108
|
-
]
|
|
22109
|
-
},
|
|
22110
|
-
"getVoice": {
|
|
22111
|
-
"description": "Get details for a single voice.",
|
|
22112
|
-
"parameters": {
|
|
22113
|
-
"voiceId": {
|
|
22114
|
-
"type": "string",
|
|
22115
|
-
"description": "The voice ID to look up"
|
|
22116
|
-
}
|
|
22117
|
-
},
|
|
22118
|
-
"required": [
|
|
22119
|
-
"voiceId"
|
|
22120
|
-
],
|
|
22121
|
-
"returns": "Promise<any>",
|
|
22122
|
-
"examples": [
|
|
22123
|
-
{
|
|
22124
|
-
"language": "ts",
|
|
22125
|
-
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
22126
|
-
}
|
|
22127
|
-
]
|
|
22128
|
-
},
|
|
22129
|
-
"listModels": {
|
|
22130
|
-
"description": "List available TTS models.",
|
|
22131
|
-
"parameters": {},
|
|
22132
|
-
"required": [],
|
|
22133
|
-
"returns": "Promise<any[]>",
|
|
22134
|
-
"examples": [
|
|
22135
|
-
{
|
|
22136
|
-
"language": "ts",
|
|
22137
|
-
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
22138
|
-
}
|
|
22139
|
-
]
|
|
22140
|
-
},
|
|
22141
|
-
"synthesize": {
|
|
22142
|
-
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
22143
|
-
"parameters": {
|
|
22144
|
-
"text": {
|
|
22145
|
-
"type": "string",
|
|
22146
|
-
"description": "The text to convert to speech"
|
|
22147
|
-
},
|
|
22148
|
-
"options": {
|
|
22149
|
-
"type": "SynthesizeOptions",
|
|
22150
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
22151
|
-
"properties": {
|
|
22152
|
-
"voiceId": {
|
|
22153
|
-
"type": "string",
|
|
22154
|
-
"description": ""
|
|
22155
|
-
},
|
|
22156
|
-
"modelId": {
|
|
22157
|
-
"type": "string",
|
|
22158
|
-
"description": ""
|
|
22159
|
-
},
|
|
22160
|
-
"outputFormat": {
|
|
22161
|
-
"type": "string",
|
|
22162
|
-
"description": ""
|
|
22163
|
-
},
|
|
22164
|
-
"voiceSettings": {
|
|
22165
|
-
"type": "ElevenLabsVoiceSettings",
|
|
22166
|
-
"description": ""
|
|
22167
|
-
},
|
|
22168
|
-
"disableCache": {
|
|
22169
|
-
"type": "boolean",
|
|
22170
|
-
"description": ""
|
|
22171
|
-
}
|
|
22172
|
-
}
|
|
22173
|
-
}
|
|
22174
|
-
},
|
|
22175
|
-
"required": [
|
|
22176
|
-
"text"
|
|
22177
|
-
],
|
|
22178
|
-
"returns": "Promise<Buffer>",
|
|
22179
|
-
"examples": [
|
|
22180
|
-
{
|
|
22181
|
-
"language": "ts",
|
|
22182
|
-
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
22183
|
-
}
|
|
22184
|
-
]
|
|
22185
|
-
},
|
|
22186
|
-
"say": {
|
|
22187
|
-
"description": "Synthesize speech and write the audio to a file.",
|
|
22188
|
-
"parameters": {
|
|
22189
|
-
"text": {
|
|
22190
|
-
"type": "string",
|
|
22191
|
-
"description": "The text to convert to speech"
|
|
22192
|
-
},
|
|
22193
|
-
"outputPath": {
|
|
22194
|
-
"type": "string",
|
|
22195
|
-
"description": "File path to write the audio to"
|
|
22196
|
-
},
|
|
22197
|
-
"options": {
|
|
22198
|
-
"type": "SynthesizeOptions",
|
|
22199
|
-
"description": "Voice, model, format, and voice settings overrides",
|
|
22200
|
-
"properties": {
|
|
22201
|
-
"voiceId": {
|
|
22202
|
-
"type": "string",
|
|
22203
|
-
"description": ""
|
|
22204
|
-
},
|
|
22205
|
-
"modelId": {
|
|
22206
|
-
"type": "string",
|
|
22207
|
-
"description": ""
|
|
22208
|
-
},
|
|
22209
|
-
"outputFormat": {
|
|
22210
|
-
"type": "string",
|
|
22211
|
-
"description": ""
|
|
22212
|
-
},
|
|
22213
|
-
"voiceSettings": {
|
|
22214
|
-
"type": "ElevenLabsVoiceSettings",
|
|
22215
|
-
"description": ""
|
|
22216
|
-
},
|
|
22217
|
-
"disableCache": {
|
|
22218
|
-
"type": "boolean",
|
|
22219
|
-
"description": ""
|
|
22220
|
-
}
|
|
22221
|
-
}
|
|
22222
|
-
}
|
|
22223
|
-
},
|
|
22224
|
-
"required": [
|
|
22225
|
-
"text",
|
|
22226
|
-
"outputPath"
|
|
22227
|
-
],
|
|
22228
|
-
"returns": "Promise<string>",
|
|
22229
|
-
"examples": [
|
|
22230
|
-
{
|
|
22231
|
-
"language": "ts",
|
|
22232
|
-
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
22233
|
-
}
|
|
22234
|
-
]
|
|
22235
|
-
}
|
|
22236
|
-
},
|
|
22237
|
-
"getters": {
|
|
22238
|
-
"apiKey": {
|
|
22239
|
-
"description": "The resolved API key from options or environment.",
|
|
22240
|
-
"returns": "string"
|
|
22241
|
-
}
|
|
22242
|
-
},
|
|
22243
|
-
"events": {
|
|
22244
|
-
"failure": {
|
|
22245
|
-
"name": "failure",
|
|
22246
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
22247
|
-
"arguments": {}
|
|
22248
|
-
},
|
|
22249
|
-
"voices": {
|
|
22250
|
-
"name": "voices",
|
|
22251
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
22252
|
-
"arguments": {}
|
|
22253
|
-
},
|
|
22254
|
-
"speech": {
|
|
22255
|
-
"name": "speech",
|
|
22256
|
-
"description": "Event emitted by ElevenLabsClient",
|
|
22257
|
-
"arguments": {}
|
|
22258
|
-
}
|
|
22259
|
-
},
|
|
22260
|
-
"state": {},
|
|
22261
|
-
"options": {},
|
|
22262
|
-
"envVars": [],
|
|
22263
|
-
"examples": [
|
|
22264
|
-
{
|
|
22265
|
-
"language": "ts",
|
|
22266
|
-
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
22267
|
-
}
|
|
22268
|
-
]
|
|
22269
|
-
},
|
|
22270
22350
|
{
|
|
22271
22351
|
"id": "servers.mcp",
|
|
22272
22352
|
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|
|
@@ -22986,6 +23066,10 @@ export const containerIntrospectionData = [
|
|
|
22986
23066
|
"description": "Returns a map of enabled feature shortcut IDs to their instances.",
|
|
22987
23067
|
"returns": "Partial<AvailableInstanceTypes<Features>>"
|
|
22988
23068
|
},
|
|
23069
|
+
"describer": {
|
|
23070
|
+
"description": "Lazy-initialized ContainerDescriber for introspecting registries, helpers, and members.",
|
|
23071
|
+
"returns": "ContainerDescriber"
|
|
23072
|
+
},
|
|
22989
23073
|
"context": {
|
|
22990
23074
|
"description": "The Container's context is an object that contains the enabled features, the container itself, and any additional context that has been added to the container. All helper instances that are created by the container will have access to the shared context.",
|
|
22991
23075
|
"returns": "ContainerContext<Features> & Partial<AvailableInstanceTypes<AvailableFeatures>>"
|