@soederpop/luca 0.0.12 → 0.0.15
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/assistants/architect/CORE.md +3 -0
- package/assistants/architect/hooks.ts +3 -0
- package/assistants/architect/tools.ts +10 -0
- package/package.json +8 -3
- package/src/agi/features/assistant.ts +5 -0
- package/src/agi/features/conversation.ts +67 -4
- package/src/bootstrap/generated.ts +1 -1
- package/src/commands/chat.ts +12 -0
- package/src/introspection/generated.agi.ts +1031 -1015
- package/src/introspection/generated.node.ts +1 -1
- package/src/introspection/generated.web.ts +1 -1
- package/src/scaffolds/generated.ts +1 -1
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import { setBuildTimeData, setContainerBuildTimeData } from './index.js';
|
|
2
2
|
|
|
3
3
|
// Auto-generated introspection registry data
|
|
4
|
-
// Generated at: 2026-03-
|
|
4
|
+
// Generated at: 2026-03-20T05:45:32.252Z
|
|
5
5
|
|
|
6
6
|
setBuildTimeData('features.googleDocs', {
|
|
7
7
|
"id": "features.googleDocs",
|
|
@@ -9519,175 +9519,443 @@ setBuildTimeData('clients.websocket', {
|
|
|
9519
9519
|
]
|
|
9520
9520
|
});
|
|
9521
9521
|
|
|
9522
|
-
setBuildTimeData('clients.
|
|
9523
|
-
"id": "clients.
|
|
9524
|
-
"description": "
|
|
9525
|
-
"shortcut": "clients.
|
|
9526
|
-
"className": "
|
|
9522
|
+
setBuildTimeData('clients.openai', {
|
|
9523
|
+
"id": "clients.openai",
|
|
9524
|
+
"description": "OpenAI client — wraps the OpenAI SDK for chat completions, responses API, embeddings, and image generation. Provides convenience methods for common operations while tracking token usage and request counts. Supports both the Chat Completions API and the newer Responses API.",
|
|
9525
|
+
"shortcut": "clients.openai",
|
|
9526
|
+
"className": "OpenAIClient",
|
|
9527
9527
|
"methods": {
|
|
9528
|
-
"
|
|
9529
|
-
"description": "
|
|
9528
|
+
"connect": {
|
|
9529
|
+
"description": "Test the API connection by listing models.",
|
|
9530
|
+
"parameters": {},
|
|
9531
|
+
"required": [],
|
|
9532
|
+
"returns": "Promise<this>",
|
|
9533
|
+
"examples": [
|
|
9534
|
+
{
|
|
9535
|
+
"language": "ts",
|
|
9536
|
+
"code": "await openai.connect()"
|
|
9537
|
+
}
|
|
9538
|
+
]
|
|
9539
|
+
},
|
|
9540
|
+
"createChatCompletion": {
|
|
9541
|
+
"description": "Create a chat completion using the Chat Completions API.",
|
|
9530
9542
|
"parameters": {
|
|
9531
|
-
"
|
|
9532
|
-
"type": "
|
|
9533
|
-
"description": "
|
|
9543
|
+
"messages": {
|
|
9544
|
+
"type": "OpenAI.Chat.Completions.ChatCompletionMessageParam[]",
|
|
9545
|
+
"description": "Array of chat messages"
|
|
9546
|
+
},
|
|
9547
|
+
"options": {
|
|
9548
|
+
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
9549
|
+
"description": "Additional parameters for the completion"
|
|
9534
9550
|
}
|
|
9535
9551
|
},
|
|
9536
9552
|
"required": [
|
|
9537
|
-
"
|
|
9553
|
+
"messages"
|
|
9538
9554
|
],
|
|
9539
|
-
"returns": "
|
|
9555
|
+
"returns": "Promise<OpenAI.Chat.Completions.ChatCompletion>",
|
|
9556
|
+
"examples": [
|
|
9557
|
+
{
|
|
9558
|
+
"language": "ts",
|
|
9559
|
+
"code": "const response = await openai.createChatCompletion([\n { role: 'system', content: 'You are a helpful assistant.' },\n { role: 'user', content: 'Hello!' }\n])\nconsole.log(response.choices[0]?.message?.content)"
|
|
9560
|
+
}
|
|
9561
|
+
]
|
|
9540
9562
|
},
|
|
9541
|
-
"
|
|
9542
|
-
"description": "
|
|
9563
|
+
"createResponse": {
|
|
9564
|
+
"description": "Create a response using the Responses API.",
|
|
9543
9565
|
"parameters": {
|
|
9544
|
-
"
|
|
9545
|
-
"type": "string",
|
|
9546
|
-
"description": "The
|
|
9547
|
-
},
|
|
9548
|
-
"params": {
|
|
9549
|
-
"type": "Record<string, unknown>",
|
|
9550
|
-
"description": "Arguments to pass to the function"
|
|
9566
|
+
"input": {
|
|
9567
|
+
"type": "OpenAI.Responses.ResponseInput | string",
|
|
9568
|
+
"description": "The input prompt or message array"
|
|
9551
9569
|
},
|
|
9552
9570
|
"options": {
|
|
9553
|
-
"type": "
|
|
9554
|
-
"description": "
|
|
9571
|
+
"type": "Partial<OpenAI.Responses.ResponseCreateParamsNonStreaming>",
|
|
9572
|
+
"description": "Additional parameters for the response"
|
|
9555
9573
|
}
|
|
9556
9574
|
},
|
|
9557
9575
|
"required": [
|
|
9558
|
-
"
|
|
9576
|
+
"input"
|
|
9559
9577
|
],
|
|
9560
|
-
"returns": "
|
|
9578
|
+
"returns": "Promise<OpenAI.Responses.Response>",
|
|
9579
|
+
"examples": [
|
|
9580
|
+
{
|
|
9581
|
+
"language": "ts",
|
|
9582
|
+
"code": "const response = await openai.createResponse('Explain quantum computing')"
|
|
9583
|
+
}
|
|
9584
|
+
]
|
|
9561
9585
|
},
|
|
9562
|
-
"
|
|
9563
|
-
"description": "
|
|
9586
|
+
"streamResponse": {
|
|
9587
|
+
"description": "Stream a response using the Responses API.",
|
|
9564
9588
|
"parameters": {
|
|
9565
|
-
"
|
|
9566
|
-
"type": "string",
|
|
9567
|
-
"description": "
|
|
9589
|
+
"input": {
|
|
9590
|
+
"type": "OpenAI.Responses.ResponseInput | string",
|
|
9591
|
+
"description": "The input prompt or message array"
|
|
9568
9592
|
},
|
|
9569
|
-
"
|
|
9570
|
-
"type": "
|
|
9571
|
-
"description": "
|
|
9593
|
+
"options": {
|
|
9594
|
+
"type": "Partial<OpenAI.Responses.ResponseCreateParamsStreaming>",
|
|
9595
|
+
"description": "Additional parameters for the streaming response"
|
|
9572
9596
|
}
|
|
9573
9597
|
},
|
|
9574
9598
|
"required": [
|
|
9575
|
-
"
|
|
9576
|
-
"password"
|
|
9599
|
+
"input"
|
|
9577
9600
|
],
|
|
9578
|
-
"returns": "
|
|
9601
|
+
"returns": "Promise<AsyncIterable<OpenAI.Responses.ResponseStreamEvent>>",
|
|
9602
|
+
"examples": [
|
|
9603
|
+
{
|
|
9604
|
+
"language": "ts",
|
|
9605
|
+
"code": "const stream = await openai.streamResponse('Write a poem')\nfor await (const event of stream) {\n if (event.type === 'response.output_text.delta') {\n process.stdout.write(event.delta)\n }\n}"
|
|
9606
|
+
}
|
|
9607
|
+
]
|
|
9579
9608
|
},
|
|
9580
|
-
"
|
|
9581
|
-
"description": "Create a
|
|
9609
|
+
"createCompletion": {
|
|
9610
|
+
"description": "Create a legacy text completion.",
|
|
9582
9611
|
"parameters": {
|
|
9583
|
-
"
|
|
9612
|
+
"prompt": {
|
|
9584
9613
|
"type": "string",
|
|
9585
|
-
"description": "
|
|
9614
|
+
"description": "The text prompt to complete"
|
|
9586
9615
|
},
|
|
9587
|
-
"
|
|
9588
|
-
"type": "
|
|
9589
|
-
"description": "
|
|
9616
|
+
"options": {
|
|
9617
|
+
"type": "Partial<OpenAI.Completions.CompletionCreateParams>",
|
|
9618
|
+
"description": "Additional parameters for the completion"
|
|
9590
9619
|
}
|
|
9591
9620
|
},
|
|
9592
9621
|
"required": [
|
|
9593
|
-
"
|
|
9594
|
-
"password"
|
|
9622
|
+
"prompt"
|
|
9595
9623
|
],
|
|
9596
|
-
"returns": "
|
|
9597
|
-
|
|
9598
|
-
|
|
9599
|
-
|
|
9600
|
-
|
|
9601
|
-
|
|
9602
|
-
|
|
9603
|
-
},
|
|
9604
|
-
"getSession": {
|
|
9605
|
-
"description": "Get the current session, if any.",
|
|
9606
|
-
"parameters": {},
|
|
9607
|
-
"required": [],
|
|
9608
|
-
"returns": "void"
|
|
9609
|
-
},
|
|
9610
|
-
"getUser": {
|
|
9611
|
-
"description": "Get the current user, if any.",
|
|
9612
|
-
"parameters": {},
|
|
9613
|
-
"required": [],
|
|
9614
|
-
"returns": "void"
|
|
9624
|
+
"returns": "Promise<OpenAI.Completions.Completion>",
|
|
9625
|
+
"examples": [
|
|
9626
|
+
{
|
|
9627
|
+
"language": "ts",
|
|
9628
|
+
"code": "const response = await openai.createCompletion('Once upon a time')"
|
|
9629
|
+
}
|
|
9630
|
+
]
|
|
9615
9631
|
},
|
|
9616
|
-
"
|
|
9617
|
-
"description": "
|
|
9632
|
+
"createEmbedding": {
|
|
9633
|
+
"description": "Create text embeddings for semantic search or similarity comparisons.",
|
|
9618
9634
|
"parameters": {
|
|
9619
|
-
"
|
|
9620
|
-
"type": "string",
|
|
9621
|
-
"description": "
|
|
9635
|
+
"input": {
|
|
9636
|
+
"type": "string | string[]",
|
|
9637
|
+
"description": "A string or array of strings to embed"
|
|
9622
9638
|
},
|
|
9623
|
-
"
|
|
9624
|
-
"type": "
|
|
9625
|
-
"description": "
|
|
9639
|
+
"options": {
|
|
9640
|
+
"type": "Partial<OpenAI.Embeddings.EmbeddingCreateParams>",
|
|
9641
|
+
"description": "Additional parameters (model, etc.)"
|
|
9626
9642
|
}
|
|
9627
9643
|
},
|
|
9628
9644
|
"required": [
|
|
9629
|
-
"
|
|
9645
|
+
"input"
|
|
9630
9646
|
],
|
|
9631
|
-
"returns": "
|
|
9647
|
+
"returns": "Promise<OpenAI.Embeddings.CreateEmbeddingResponse>",
|
|
9648
|
+
"examples": [
|
|
9649
|
+
{
|
|
9650
|
+
"language": "ts",
|
|
9651
|
+
"code": "const response = await openai.createEmbedding('Hello world')\nconsole.log(response.data[0].embedding.length)"
|
|
9652
|
+
}
|
|
9653
|
+
]
|
|
9632
9654
|
},
|
|
9633
|
-
"
|
|
9634
|
-
"description": "
|
|
9655
|
+
"createImage": {
|
|
9656
|
+
"description": "Generate an image from a text prompt using DALL-E.",
|
|
9635
9657
|
"parameters": {
|
|
9636
|
-
"
|
|
9637
|
-
"type": "string",
|
|
9638
|
-
"description": "A name for this subscription channel"
|
|
9639
|
-
},
|
|
9640
|
-
"table": {
|
|
9658
|
+
"prompt": {
|
|
9641
9659
|
"type": "string",
|
|
9642
|
-
"description": "
|
|
9643
|
-
},
|
|
9644
|
-
"callback": {
|
|
9645
|
-
"type": "(payload: any) => void",
|
|
9646
|
-
"description": "Called with the payload on each change"
|
|
9660
|
+
"description": "Description of the image to generate"
|
|
9647
9661
|
},
|
|
9648
|
-
"
|
|
9649
|
-
"type": "
|
|
9650
|
-
"description": "
|
|
9662
|
+
"options": {
|
|
9663
|
+
"type": "Partial<OpenAI.Images.ImageGenerateParams>",
|
|
9664
|
+
"description": "Additional parameters (size, n, etc.)"
|
|
9651
9665
|
}
|
|
9652
9666
|
},
|
|
9653
9667
|
"required": [
|
|
9654
|
-
"
|
|
9655
|
-
"table",
|
|
9656
|
-
"callback"
|
|
9668
|
+
"prompt"
|
|
9657
9669
|
],
|
|
9658
|
-
"returns": "
|
|
9670
|
+
"returns": "Promise<OpenAI.Images.ImagesResponse>",
|
|
9671
|
+
"examples": [
|
|
9672
|
+
{
|
|
9673
|
+
"language": "ts",
|
|
9674
|
+
"code": "const response = await openai.createImage('A sunset over mountains')\nconsole.log(response.data[0].url)"
|
|
9675
|
+
}
|
|
9676
|
+
]
|
|
9659
9677
|
},
|
|
9660
|
-
"
|
|
9661
|
-
"description": "
|
|
9678
|
+
"listModels": {
|
|
9679
|
+
"description": "List all available models.",
|
|
9680
|
+
"parameters": {},
|
|
9681
|
+
"required": [],
|
|
9682
|
+
"returns": "Promise<OpenAI.Models.ModelsPage>",
|
|
9683
|
+
"examples": [
|
|
9684
|
+
{
|
|
9685
|
+
"language": "ts",
|
|
9686
|
+
"code": "const models = await openai.listModels()"
|
|
9687
|
+
}
|
|
9688
|
+
]
|
|
9689
|
+
},
|
|
9690
|
+
"ask": {
|
|
9691
|
+
"description": "Ask a single question and get a text response. Convenience wrapper around `createChatCompletion` for simple Q&A.",
|
|
9662
9692
|
"parameters": {
|
|
9663
|
-
"
|
|
9693
|
+
"question": {
|
|
9664
9694
|
"type": "string",
|
|
9665
|
-
"description": "The
|
|
9695
|
+
"description": "The question to ask"
|
|
9696
|
+
},
|
|
9697
|
+
"options": {
|
|
9698
|
+
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
9699
|
+
"description": "Additional completion parameters"
|
|
9666
9700
|
}
|
|
9667
9701
|
},
|
|
9668
9702
|
"required": [
|
|
9669
|
-
"
|
|
9703
|
+
"question"
|
|
9670
9704
|
],
|
|
9671
|
-
"returns": "
|
|
9672
|
-
|
|
9673
|
-
|
|
9674
|
-
|
|
9675
|
-
|
|
9676
|
-
|
|
9677
|
-
|
|
9678
|
-
},
|
|
9679
|
-
"connect": {
|
|
9680
|
-
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
9681
|
-
"parameters": {},
|
|
9682
|
-
"required": [],
|
|
9683
|
-
"returns": "void"
|
|
9705
|
+
"returns": "Promise<string>",
|
|
9706
|
+
"examples": [
|
|
9707
|
+
{
|
|
9708
|
+
"language": "ts",
|
|
9709
|
+
"code": "const answer = await openai.ask('What is 2 + 2?')\nconsole.log(answer) // '4'"
|
|
9710
|
+
}
|
|
9711
|
+
]
|
|
9684
9712
|
},
|
|
9685
|
-
"
|
|
9686
|
-
"description": "
|
|
9687
|
-
"parameters": {
|
|
9688
|
-
|
|
9689
|
-
|
|
9690
|
-
|
|
9713
|
+
"chat": {
|
|
9714
|
+
"description": "Send a multi-turn conversation and get a text response. Convenience wrapper around `createChatCompletion` that returns just the text.",
|
|
9715
|
+
"parameters": {
|
|
9716
|
+
"messages": {
|
|
9717
|
+
"type": "OpenAI.Chat.Completions.ChatCompletionMessageParam[]",
|
|
9718
|
+
"description": "Array of chat messages"
|
|
9719
|
+
},
|
|
9720
|
+
"options": {
|
|
9721
|
+
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
9722
|
+
"description": "Additional completion parameters"
|
|
9723
|
+
}
|
|
9724
|
+
},
|
|
9725
|
+
"required": [
|
|
9726
|
+
"messages"
|
|
9727
|
+
],
|
|
9728
|
+
"returns": "Promise<string>",
|
|
9729
|
+
"examples": [
|
|
9730
|
+
{
|
|
9731
|
+
"language": "ts",
|
|
9732
|
+
"code": "const reply = await openai.chat([\n { role: 'system', content: 'You are a pirate.' },\n { role: 'user', content: 'Hello!' }\n])"
|
|
9733
|
+
}
|
|
9734
|
+
]
|
|
9735
|
+
}
|
|
9736
|
+
},
|
|
9737
|
+
"getters": {
|
|
9738
|
+
"defaultModel": {
|
|
9739
|
+
"description": "The default model used for completions, from options or 'gpt-4o'.",
|
|
9740
|
+
"returns": "string"
|
|
9741
|
+
},
|
|
9742
|
+
"raw": {
|
|
9743
|
+
"description": "The underlying OpenAI SDK instance for advanced use cases.",
|
|
9744
|
+
"returns": "OpenAI"
|
|
9745
|
+
}
|
|
9746
|
+
},
|
|
9747
|
+
"events": {
|
|
9748
|
+
"connected": {
|
|
9749
|
+
"name": "connected",
|
|
9750
|
+
"description": "Event emitted by OpenAIClient",
|
|
9751
|
+
"arguments": {}
|
|
9752
|
+
},
|
|
9753
|
+
"failure": {
|
|
9754
|
+
"name": "failure",
|
|
9755
|
+
"description": "Event emitted by OpenAIClient",
|
|
9756
|
+
"arguments": {}
|
|
9757
|
+
},
|
|
9758
|
+
"completion": {
|
|
9759
|
+
"name": "completion",
|
|
9760
|
+
"description": "Event emitted by OpenAIClient",
|
|
9761
|
+
"arguments": {}
|
|
9762
|
+
},
|
|
9763
|
+
"embedding": {
|
|
9764
|
+
"name": "embedding",
|
|
9765
|
+
"description": "Event emitted by OpenAIClient",
|
|
9766
|
+
"arguments": {}
|
|
9767
|
+
},
|
|
9768
|
+
"image": {
|
|
9769
|
+
"name": "image",
|
|
9770
|
+
"description": "Event emitted by OpenAIClient",
|
|
9771
|
+
"arguments": {}
|
|
9772
|
+
},
|
|
9773
|
+
"models": {
|
|
9774
|
+
"name": "models",
|
|
9775
|
+
"description": "Event emitted by OpenAIClient",
|
|
9776
|
+
"arguments": {}
|
|
9777
|
+
}
|
|
9778
|
+
},
|
|
9779
|
+
"state": {},
|
|
9780
|
+
"options": {},
|
|
9781
|
+
"envVars": [],
|
|
9782
|
+
"examples": [
|
|
9783
|
+
{
|
|
9784
|
+
"language": "ts",
|
|
9785
|
+
"code": "const openai = container.client('openai', { defaultModel: 'gpt-4o' })\nconst answer = await openai.ask('What is the meaning of life?')\nconsole.log(answer)"
|
|
9786
|
+
}
|
|
9787
|
+
]
|
|
9788
|
+
});
|
|
9789
|
+
|
|
9790
|
+
setBuildTimeData('clients.supabase', {
|
|
9791
|
+
"id": "clients.supabase",
|
|
9792
|
+
"description": "Supabase client for the Luca container system. Wraps the official `@supabase/supabase-js` SDK and exposes it through Luca's typed state, events, and introspection system. The SDK is isomorphic so this single implementation works in both Node and browser containers. Use `client.sdk` for full SDK access, or use the convenience wrappers for common operations (auth, database queries, storage, edge functions, realtime).",
|
|
9793
|
+
"shortcut": "clients.supabase",
|
|
9794
|
+
"className": "SupabaseClient",
|
|
9795
|
+
"methods": {
|
|
9796
|
+
"from": {
|
|
9797
|
+
"description": "Start a query on a Postgres table or view.",
|
|
9798
|
+
"parameters": {
|
|
9799
|
+
"table": {
|
|
9800
|
+
"type": "string",
|
|
9801
|
+
"description": "The table or view name to query"
|
|
9802
|
+
}
|
|
9803
|
+
},
|
|
9804
|
+
"required": [
|
|
9805
|
+
"table"
|
|
9806
|
+
],
|
|
9807
|
+
"returns": "void"
|
|
9808
|
+
},
|
|
9809
|
+
"rpc": {
|
|
9810
|
+
"description": "Call a Postgres function (RPC).",
|
|
9811
|
+
"parameters": {
|
|
9812
|
+
"fn": {
|
|
9813
|
+
"type": "string",
|
|
9814
|
+
"description": "The function name"
|
|
9815
|
+
},
|
|
9816
|
+
"params": {
|
|
9817
|
+
"type": "Record<string, unknown>",
|
|
9818
|
+
"description": "Arguments to pass to the function"
|
|
9819
|
+
},
|
|
9820
|
+
"options": {
|
|
9821
|
+
"type": "{ head?: boolean; get?: boolean; count?: \"exact\" | \"planned\" | \"estimated\" }",
|
|
9822
|
+
"description": "Optional settings (head, get, count)"
|
|
9823
|
+
}
|
|
9824
|
+
},
|
|
9825
|
+
"required": [
|
|
9826
|
+
"fn"
|
|
9827
|
+
],
|
|
9828
|
+
"returns": "void"
|
|
9829
|
+
},
|
|
9830
|
+
"signInWithPassword": {
|
|
9831
|
+
"description": "Sign in with email and password.",
|
|
9832
|
+
"parameters": {
|
|
9833
|
+
"email": {
|
|
9834
|
+
"type": "string",
|
|
9835
|
+
"description": "Parameter email"
|
|
9836
|
+
},
|
|
9837
|
+
"password": {
|
|
9838
|
+
"type": "string",
|
|
9839
|
+
"description": "Parameter password"
|
|
9840
|
+
}
|
|
9841
|
+
},
|
|
9842
|
+
"required": [
|
|
9843
|
+
"email",
|
|
9844
|
+
"password"
|
|
9845
|
+
],
|
|
9846
|
+
"returns": "void"
|
|
9847
|
+
},
|
|
9848
|
+
"signUp": {
|
|
9849
|
+
"description": "Create a new user account with email and password.",
|
|
9850
|
+
"parameters": {
|
|
9851
|
+
"email": {
|
|
9852
|
+
"type": "string",
|
|
9853
|
+
"description": "Parameter email"
|
|
9854
|
+
},
|
|
9855
|
+
"password": {
|
|
9856
|
+
"type": "string",
|
|
9857
|
+
"description": "Parameter password"
|
|
9858
|
+
}
|
|
9859
|
+
},
|
|
9860
|
+
"required": [
|
|
9861
|
+
"email",
|
|
9862
|
+
"password"
|
|
9863
|
+
],
|
|
9864
|
+
"returns": "void"
|
|
9865
|
+
},
|
|
9866
|
+
"signOut": {
|
|
9867
|
+
"description": "Sign the current user out.",
|
|
9868
|
+
"parameters": {},
|
|
9869
|
+
"required": [],
|
|
9870
|
+
"returns": "void"
|
|
9871
|
+
},
|
|
9872
|
+
"getSession": {
|
|
9873
|
+
"description": "Get the current session, if any.",
|
|
9874
|
+
"parameters": {},
|
|
9875
|
+
"required": [],
|
|
9876
|
+
"returns": "void"
|
|
9877
|
+
},
|
|
9878
|
+
"getUser": {
|
|
9879
|
+
"description": "Get the current user, if any.",
|
|
9880
|
+
"parameters": {},
|
|
9881
|
+
"required": [],
|
|
9882
|
+
"returns": "void"
|
|
9883
|
+
},
|
|
9884
|
+
"invoke": {
|
|
9885
|
+
"description": "Invoke a Supabase Edge Function by name.",
|
|
9886
|
+
"parameters": {
|
|
9887
|
+
"name": {
|
|
9888
|
+
"type": "string",
|
|
9889
|
+
"description": "Parameter name"
|
|
9890
|
+
},
|
|
9891
|
+
"body": {
|
|
9892
|
+
"type": "any",
|
|
9893
|
+
"description": "Parameter body"
|
|
9894
|
+
}
|
|
9895
|
+
},
|
|
9896
|
+
"required": [
|
|
9897
|
+
"name"
|
|
9898
|
+
],
|
|
9899
|
+
"returns": "void"
|
|
9900
|
+
},
|
|
9901
|
+
"subscribe": {
|
|
9902
|
+
"description": "Subscribe to realtime changes on a Postgres table.",
|
|
9903
|
+
"parameters": {
|
|
9904
|
+
"channelName": {
|
|
9905
|
+
"type": "string",
|
|
9906
|
+
"description": "A name for this subscription channel"
|
|
9907
|
+
},
|
|
9908
|
+
"table": {
|
|
9909
|
+
"type": "string",
|
|
9910
|
+
"description": "The table to listen to"
|
|
9911
|
+
},
|
|
9912
|
+
"callback": {
|
|
9913
|
+
"type": "(payload: any) => void",
|
|
9914
|
+
"description": "Called with the payload on each change"
|
|
9915
|
+
},
|
|
9916
|
+
"event": {
|
|
9917
|
+
"type": "\"INSERT\" | \"UPDATE\" | \"DELETE\" | \"*\"",
|
|
9918
|
+
"description": "The event type to listen for (default: all changes)"
|
|
9919
|
+
}
|
|
9920
|
+
},
|
|
9921
|
+
"required": [
|
|
9922
|
+
"channelName",
|
|
9923
|
+
"table",
|
|
9924
|
+
"callback"
|
|
9925
|
+
],
|
|
9926
|
+
"returns": "RealtimeChannel"
|
|
9927
|
+
},
|
|
9928
|
+
"unsubscribe": {
|
|
9929
|
+
"description": "Unsubscribe and remove a realtime channel by name.",
|
|
9930
|
+
"parameters": {
|
|
9931
|
+
"channelName": {
|
|
9932
|
+
"type": "string",
|
|
9933
|
+
"description": "The channel name to remove"
|
|
9934
|
+
}
|
|
9935
|
+
},
|
|
9936
|
+
"required": [
|
|
9937
|
+
"channelName"
|
|
9938
|
+
],
|
|
9939
|
+
"returns": "void"
|
|
9940
|
+
},
|
|
9941
|
+
"unsubscribeAll": {
|
|
9942
|
+
"description": "Unsubscribe and remove all realtime channels.",
|
|
9943
|
+
"parameters": {},
|
|
9944
|
+
"required": [],
|
|
9945
|
+
"returns": "void"
|
|
9946
|
+
},
|
|
9947
|
+
"connect": {
|
|
9948
|
+
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
9949
|
+
"parameters": {},
|
|
9950
|
+
"required": [],
|
|
9951
|
+
"returns": "void"
|
|
9952
|
+
},
|
|
9953
|
+
"disconnect": {
|
|
9954
|
+
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
9955
|
+
"parameters": {},
|
|
9956
|
+
"required": [],
|
|
9957
|
+
"returns": "void"
|
|
9958
|
+
}
|
|
9691
9959
|
},
|
|
9692
9960
|
"getters": {
|
|
9693
9961
|
"sdk": {
|
|
@@ -10169,279 +10437,11 @@ setBuildTimeData('clients.comfyui', {
|
|
|
10169
10437
|
]
|
|
10170
10438
|
});
|
|
10171
10439
|
|
|
10172
|
-
setBuildTimeData('
|
|
10173
|
-
"id": "
|
|
10174
|
-
"description": "
|
|
10175
|
-
"shortcut": "
|
|
10176
|
-
"className": "
|
|
10177
|
-
"methods": {
|
|
10178
|
-
"connect": {
|
|
10179
|
-
"description": "Test the API connection by listing models.",
|
|
10180
|
-
"parameters": {},
|
|
10181
|
-
"required": [],
|
|
10182
|
-
"returns": "Promise<this>",
|
|
10183
|
-
"examples": [
|
|
10184
|
-
{
|
|
10185
|
-
"language": "ts",
|
|
10186
|
-
"code": "await openai.connect()"
|
|
10187
|
-
}
|
|
10188
|
-
]
|
|
10189
|
-
},
|
|
10190
|
-
"createChatCompletion": {
|
|
10191
|
-
"description": "Create a chat completion using the Chat Completions API.",
|
|
10192
|
-
"parameters": {
|
|
10193
|
-
"messages": {
|
|
10194
|
-
"type": "OpenAI.Chat.Completions.ChatCompletionMessageParam[]",
|
|
10195
|
-
"description": "Array of chat messages"
|
|
10196
|
-
},
|
|
10197
|
-
"options": {
|
|
10198
|
-
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
10199
|
-
"description": "Additional parameters for the completion"
|
|
10200
|
-
}
|
|
10201
|
-
},
|
|
10202
|
-
"required": [
|
|
10203
|
-
"messages"
|
|
10204
|
-
],
|
|
10205
|
-
"returns": "Promise<OpenAI.Chat.Completions.ChatCompletion>",
|
|
10206
|
-
"examples": [
|
|
10207
|
-
{
|
|
10208
|
-
"language": "ts",
|
|
10209
|
-
"code": "const response = await openai.createChatCompletion([\n { role: 'system', content: 'You are a helpful assistant.' },\n { role: 'user', content: 'Hello!' }\n])\nconsole.log(response.choices[0]?.message?.content)"
|
|
10210
|
-
}
|
|
10211
|
-
]
|
|
10212
|
-
},
|
|
10213
|
-
"createResponse": {
|
|
10214
|
-
"description": "Create a response using the Responses API.",
|
|
10215
|
-
"parameters": {
|
|
10216
|
-
"input": {
|
|
10217
|
-
"type": "OpenAI.Responses.ResponseInput | string",
|
|
10218
|
-
"description": "The input prompt or message array"
|
|
10219
|
-
},
|
|
10220
|
-
"options": {
|
|
10221
|
-
"type": "Partial<OpenAI.Responses.ResponseCreateParamsNonStreaming>",
|
|
10222
|
-
"description": "Additional parameters for the response"
|
|
10223
|
-
}
|
|
10224
|
-
},
|
|
10225
|
-
"required": [
|
|
10226
|
-
"input"
|
|
10227
|
-
],
|
|
10228
|
-
"returns": "Promise<OpenAI.Responses.Response>",
|
|
10229
|
-
"examples": [
|
|
10230
|
-
{
|
|
10231
|
-
"language": "ts",
|
|
10232
|
-
"code": "const response = await openai.createResponse('Explain quantum computing')"
|
|
10233
|
-
}
|
|
10234
|
-
]
|
|
10235
|
-
},
|
|
10236
|
-
"streamResponse": {
|
|
10237
|
-
"description": "Stream a response using the Responses API.",
|
|
10238
|
-
"parameters": {
|
|
10239
|
-
"input": {
|
|
10240
|
-
"type": "OpenAI.Responses.ResponseInput | string",
|
|
10241
|
-
"description": "The input prompt or message array"
|
|
10242
|
-
},
|
|
10243
|
-
"options": {
|
|
10244
|
-
"type": "Partial<OpenAI.Responses.ResponseCreateParamsStreaming>",
|
|
10245
|
-
"description": "Additional parameters for the streaming response"
|
|
10246
|
-
}
|
|
10247
|
-
},
|
|
10248
|
-
"required": [
|
|
10249
|
-
"input"
|
|
10250
|
-
],
|
|
10251
|
-
"returns": "Promise<AsyncIterable<OpenAI.Responses.ResponseStreamEvent>>",
|
|
10252
|
-
"examples": [
|
|
10253
|
-
{
|
|
10254
|
-
"language": "ts",
|
|
10255
|
-
"code": "const stream = await openai.streamResponse('Write a poem')\nfor await (const event of stream) {\n if (event.type === 'response.output_text.delta') {\n process.stdout.write(event.delta)\n }\n}"
|
|
10256
|
-
}
|
|
10257
|
-
]
|
|
10258
|
-
},
|
|
10259
|
-
"createCompletion": {
|
|
10260
|
-
"description": "Create a legacy text completion.",
|
|
10261
|
-
"parameters": {
|
|
10262
|
-
"prompt": {
|
|
10263
|
-
"type": "string",
|
|
10264
|
-
"description": "The text prompt to complete"
|
|
10265
|
-
},
|
|
10266
|
-
"options": {
|
|
10267
|
-
"type": "Partial<OpenAI.Completions.CompletionCreateParams>",
|
|
10268
|
-
"description": "Additional parameters for the completion"
|
|
10269
|
-
}
|
|
10270
|
-
},
|
|
10271
|
-
"required": [
|
|
10272
|
-
"prompt"
|
|
10273
|
-
],
|
|
10274
|
-
"returns": "Promise<OpenAI.Completions.Completion>",
|
|
10275
|
-
"examples": [
|
|
10276
|
-
{
|
|
10277
|
-
"language": "ts",
|
|
10278
|
-
"code": "const response = await openai.createCompletion('Once upon a time')"
|
|
10279
|
-
}
|
|
10280
|
-
]
|
|
10281
|
-
},
|
|
10282
|
-
"createEmbedding": {
|
|
10283
|
-
"description": "Create text embeddings for semantic search or similarity comparisons.",
|
|
10284
|
-
"parameters": {
|
|
10285
|
-
"input": {
|
|
10286
|
-
"type": "string | string[]",
|
|
10287
|
-
"description": "A string or array of strings to embed"
|
|
10288
|
-
},
|
|
10289
|
-
"options": {
|
|
10290
|
-
"type": "Partial<OpenAI.Embeddings.EmbeddingCreateParams>",
|
|
10291
|
-
"description": "Additional parameters (model, etc.)"
|
|
10292
|
-
}
|
|
10293
|
-
},
|
|
10294
|
-
"required": [
|
|
10295
|
-
"input"
|
|
10296
|
-
],
|
|
10297
|
-
"returns": "Promise<OpenAI.Embeddings.CreateEmbeddingResponse>",
|
|
10298
|
-
"examples": [
|
|
10299
|
-
{
|
|
10300
|
-
"language": "ts",
|
|
10301
|
-
"code": "const response = await openai.createEmbedding('Hello world')\nconsole.log(response.data[0].embedding.length)"
|
|
10302
|
-
}
|
|
10303
|
-
]
|
|
10304
|
-
},
|
|
10305
|
-
"createImage": {
|
|
10306
|
-
"description": "Generate an image from a text prompt using DALL-E.",
|
|
10307
|
-
"parameters": {
|
|
10308
|
-
"prompt": {
|
|
10309
|
-
"type": "string",
|
|
10310
|
-
"description": "Description of the image to generate"
|
|
10311
|
-
},
|
|
10312
|
-
"options": {
|
|
10313
|
-
"type": "Partial<OpenAI.Images.ImageGenerateParams>",
|
|
10314
|
-
"description": "Additional parameters (size, n, etc.)"
|
|
10315
|
-
}
|
|
10316
|
-
},
|
|
10317
|
-
"required": [
|
|
10318
|
-
"prompt"
|
|
10319
|
-
],
|
|
10320
|
-
"returns": "Promise<OpenAI.Images.ImagesResponse>",
|
|
10321
|
-
"examples": [
|
|
10322
|
-
{
|
|
10323
|
-
"language": "ts",
|
|
10324
|
-
"code": "const response = await openai.createImage('A sunset over mountains')\nconsole.log(response.data[0].url)"
|
|
10325
|
-
}
|
|
10326
|
-
]
|
|
10327
|
-
},
|
|
10328
|
-
"listModels": {
|
|
10329
|
-
"description": "List all available models.",
|
|
10330
|
-
"parameters": {},
|
|
10331
|
-
"required": [],
|
|
10332
|
-
"returns": "Promise<OpenAI.Models.ModelsPage>",
|
|
10333
|
-
"examples": [
|
|
10334
|
-
{
|
|
10335
|
-
"language": "ts",
|
|
10336
|
-
"code": "const models = await openai.listModels()"
|
|
10337
|
-
}
|
|
10338
|
-
]
|
|
10339
|
-
},
|
|
10340
|
-
"ask": {
|
|
10341
|
-
"description": "Ask a single question and get a text response. Convenience wrapper around `createChatCompletion` for simple Q&A.",
|
|
10342
|
-
"parameters": {
|
|
10343
|
-
"question": {
|
|
10344
|
-
"type": "string",
|
|
10345
|
-
"description": "The question to ask"
|
|
10346
|
-
},
|
|
10347
|
-
"options": {
|
|
10348
|
-
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
10349
|
-
"description": "Additional completion parameters"
|
|
10350
|
-
}
|
|
10351
|
-
},
|
|
10352
|
-
"required": [
|
|
10353
|
-
"question"
|
|
10354
|
-
],
|
|
10355
|
-
"returns": "Promise<string>",
|
|
10356
|
-
"examples": [
|
|
10357
|
-
{
|
|
10358
|
-
"language": "ts",
|
|
10359
|
-
"code": "const answer = await openai.ask('What is 2 + 2?')\nconsole.log(answer) // '4'"
|
|
10360
|
-
}
|
|
10361
|
-
]
|
|
10362
|
-
},
|
|
10363
|
-
"chat": {
|
|
10364
|
-
"description": "Send a multi-turn conversation and get a text response. Convenience wrapper around `createChatCompletion` that returns just the text.",
|
|
10365
|
-
"parameters": {
|
|
10366
|
-
"messages": {
|
|
10367
|
-
"type": "OpenAI.Chat.Completions.ChatCompletionMessageParam[]",
|
|
10368
|
-
"description": "Array of chat messages"
|
|
10369
|
-
},
|
|
10370
|
-
"options": {
|
|
10371
|
-
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
10372
|
-
"description": "Additional completion parameters"
|
|
10373
|
-
}
|
|
10374
|
-
},
|
|
10375
|
-
"required": [
|
|
10376
|
-
"messages"
|
|
10377
|
-
],
|
|
10378
|
-
"returns": "Promise<string>",
|
|
10379
|
-
"examples": [
|
|
10380
|
-
{
|
|
10381
|
-
"language": "ts",
|
|
10382
|
-
"code": "const reply = await openai.chat([\n { role: 'system', content: 'You are a pirate.' },\n { role: 'user', content: 'Hello!' }\n])"
|
|
10383
|
-
}
|
|
10384
|
-
]
|
|
10385
|
-
}
|
|
10386
|
-
},
|
|
10387
|
-
"getters": {
|
|
10388
|
-
"defaultModel": {
|
|
10389
|
-
"description": "The default model used for completions, from options or 'gpt-4o'.",
|
|
10390
|
-
"returns": "string"
|
|
10391
|
-
},
|
|
10392
|
-
"raw": {
|
|
10393
|
-
"description": "The underlying OpenAI SDK instance for advanced use cases.",
|
|
10394
|
-
"returns": "OpenAI"
|
|
10395
|
-
}
|
|
10396
|
-
},
|
|
10397
|
-
"events": {
|
|
10398
|
-
"connected": {
|
|
10399
|
-
"name": "connected",
|
|
10400
|
-
"description": "Event emitted by OpenAIClient",
|
|
10401
|
-
"arguments": {}
|
|
10402
|
-
},
|
|
10403
|
-
"failure": {
|
|
10404
|
-
"name": "failure",
|
|
10405
|
-
"description": "Event emitted by OpenAIClient",
|
|
10406
|
-
"arguments": {}
|
|
10407
|
-
},
|
|
10408
|
-
"completion": {
|
|
10409
|
-
"name": "completion",
|
|
10410
|
-
"description": "Event emitted by OpenAIClient",
|
|
10411
|
-
"arguments": {}
|
|
10412
|
-
},
|
|
10413
|
-
"embedding": {
|
|
10414
|
-
"name": "embedding",
|
|
10415
|
-
"description": "Event emitted by OpenAIClient",
|
|
10416
|
-
"arguments": {}
|
|
10417
|
-
},
|
|
10418
|
-
"image": {
|
|
10419
|
-
"name": "image",
|
|
10420
|
-
"description": "Event emitted by OpenAIClient",
|
|
10421
|
-
"arguments": {}
|
|
10422
|
-
},
|
|
10423
|
-
"models": {
|
|
10424
|
-
"name": "models",
|
|
10425
|
-
"description": "Event emitted by OpenAIClient",
|
|
10426
|
-
"arguments": {}
|
|
10427
|
-
}
|
|
10428
|
-
},
|
|
10429
|
-
"state": {},
|
|
10430
|
-
"options": {},
|
|
10431
|
-
"envVars": [],
|
|
10432
|
-
"examples": [
|
|
10433
|
-
{
|
|
10434
|
-
"language": "ts",
|
|
10435
|
-
"code": "const openai = container.client('openai', { defaultModel: 'gpt-4o' })\nconst answer = await openai.ask('What is the meaning of life?')\nconsole.log(answer)"
|
|
10436
|
-
}
|
|
10437
|
-
]
|
|
10438
|
-
});
|
|
10439
|
-
|
|
10440
|
-
setBuildTimeData('servers.mcp', {
|
|
10441
|
-
"id": "servers.mcp",
|
|
10442
|
-
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|
|
10443
|
-
"shortcut": "servers.mcp",
|
|
10444
|
-
"className": "MCPServer",
|
|
10440
|
+
setBuildTimeData('servers.mcp', {
|
|
10441
|
+
"id": "servers.mcp",
|
|
10442
|
+
"description": "MCP (Model Context Protocol) server for exposing tools, resources, and prompts to AI clients like Claude Code. Uses the low-level MCP SDK Server class directly with Zod 4 native JSON Schema conversion. Register tools, resources, and prompts programmatically, then start the server over stdio (for CLI integration) or HTTP (for remote access).",
|
|
10443
|
+
"shortcut": "servers.mcp",
|
|
10444
|
+
"className": "MCPServer",
|
|
10445
10445
|
"methods": {
|
|
10446
10446
|
"tool": {
|
|
10447
10447
|
"description": "Register an MCP tool. The tool's Zod schema is converted to JSON Schema for the protocol listing, and used for runtime argument validation. Tool handlers can return a string (auto-wrapped as text content) or a full CallToolResult object for advanced responses (images, errors, etc).",
|
|
@@ -11980,6 +11980,10 @@ setBuildTimeData('features.claudeCode', {
|
|
|
11980
11980
|
"settingsFile": {
|
|
11981
11981
|
"type": "string",
|
|
11982
11982
|
"description": "Path to a custom settings file."
|
|
11983
|
+
},
|
|
11984
|
+
"chrome": {
|
|
11985
|
+
"type": "boolean",
|
|
11986
|
+
"description": "Launch Claude Code with a Chrome browser tool."
|
|
11983
11987
|
}
|
|
11984
11988
|
}
|
|
11985
11989
|
}
|
|
@@ -12129,6 +12133,10 @@ setBuildTimeData('features.claudeCode', {
|
|
|
12129
12133
|
"settingsFile": {
|
|
12130
12134
|
"type": "string",
|
|
12131
12135
|
"description": "Path to a custom settings file."
|
|
12136
|
+
},
|
|
12137
|
+
"chrome": {
|
|
12138
|
+
"type": "boolean",
|
|
12139
|
+
"description": "Launch Claude Code with a Chrome browser tool."
|
|
12132
12140
|
}
|
|
12133
12141
|
}
|
|
12134
12142
|
}
|
|
@@ -22978,388 +22986,259 @@ export const introspectionData = [
|
|
|
22978
22986
|
]
|
|
22979
22987
|
},
|
|
22980
22988
|
{
|
|
22981
|
-
"id": "clients.
|
|
22982
|
-
"description": "
|
|
22983
|
-
"shortcut": "clients.
|
|
22984
|
-
"className": "
|
|
22989
|
+
"id": "clients.openai",
|
|
22990
|
+
"description": "OpenAI client — wraps the OpenAI SDK for chat completions, responses API, embeddings, and image generation. Provides convenience methods for common operations while tracking token usage and request counts. Supports both the Chat Completions API and the newer Responses API.",
|
|
22991
|
+
"shortcut": "clients.openai",
|
|
22992
|
+
"className": "OpenAIClient",
|
|
22985
22993
|
"methods": {
|
|
22986
|
-
"
|
|
22987
|
-
"description": "
|
|
22988
|
-
"parameters": {
|
|
22989
|
-
|
|
22990
|
-
|
|
22991
|
-
|
|
22994
|
+
"connect": {
|
|
22995
|
+
"description": "Test the API connection by listing models.",
|
|
22996
|
+
"parameters": {},
|
|
22997
|
+
"required": [],
|
|
22998
|
+
"returns": "Promise<this>",
|
|
22999
|
+
"examples": [
|
|
23000
|
+
{
|
|
23001
|
+
"language": "ts",
|
|
23002
|
+
"code": "await openai.connect()"
|
|
22992
23003
|
}
|
|
22993
|
-
|
|
22994
|
-
"required": [
|
|
22995
|
-
"table"
|
|
22996
|
-
],
|
|
22997
|
-
"returns": "void"
|
|
23004
|
+
]
|
|
22998
23005
|
},
|
|
22999
|
-
"
|
|
23000
|
-
"description": "
|
|
23006
|
+
"createChatCompletion": {
|
|
23007
|
+
"description": "Create a chat completion using the Chat Completions API.",
|
|
23001
23008
|
"parameters": {
|
|
23002
|
-
"
|
|
23003
|
-
"type": "
|
|
23004
|
-
"description": "
|
|
23005
|
-
},
|
|
23006
|
-
"params": {
|
|
23007
|
-
"type": "Record<string, unknown>",
|
|
23008
|
-
"description": "Arguments to pass to the function"
|
|
23009
|
+
"messages": {
|
|
23010
|
+
"type": "OpenAI.Chat.Completions.ChatCompletionMessageParam[]",
|
|
23011
|
+
"description": "Array of chat messages"
|
|
23009
23012
|
},
|
|
23010
23013
|
"options": {
|
|
23011
|
-
"type": "
|
|
23012
|
-
"description": "
|
|
23013
|
-
}
|
|
23014
|
-
},
|
|
23015
|
-
"required": [
|
|
23016
|
-
"fn"
|
|
23017
|
-
],
|
|
23018
|
-
"returns": "void"
|
|
23019
|
-
},
|
|
23020
|
-
"signInWithPassword": {
|
|
23021
|
-
"description": "Sign in with email and password.",
|
|
23022
|
-
"parameters": {
|
|
23023
|
-
"email": {
|
|
23024
|
-
"type": "string",
|
|
23025
|
-
"description": "Parameter email"
|
|
23026
|
-
},
|
|
23027
|
-
"password": {
|
|
23028
|
-
"type": "string",
|
|
23029
|
-
"description": "Parameter password"
|
|
23030
|
-
}
|
|
23031
|
-
},
|
|
23032
|
-
"required": [
|
|
23033
|
-
"email",
|
|
23034
|
-
"password"
|
|
23035
|
-
],
|
|
23036
|
-
"returns": "void"
|
|
23037
|
-
},
|
|
23038
|
-
"signUp": {
|
|
23039
|
-
"description": "Create a new user account with email and password.",
|
|
23040
|
-
"parameters": {
|
|
23041
|
-
"email": {
|
|
23042
|
-
"type": "string",
|
|
23043
|
-
"description": "Parameter email"
|
|
23044
|
-
},
|
|
23045
|
-
"password": {
|
|
23046
|
-
"type": "string",
|
|
23047
|
-
"description": "Parameter password"
|
|
23014
|
+
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
23015
|
+
"description": "Additional parameters for the completion"
|
|
23048
23016
|
}
|
|
23049
23017
|
},
|
|
23050
23018
|
"required": [
|
|
23051
|
-
"
|
|
23052
|
-
"password"
|
|
23019
|
+
"messages"
|
|
23053
23020
|
],
|
|
23054
|
-
"returns": "
|
|
23055
|
-
|
|
23056
|
-
|
|
23057
|
-
|
|
23058
|
-
|
|
23059
|
-
"required": [],
|
|
23060
|
-
"returns": "void"
|
|
23061
|
-
},
|
|
23062
|
-
"getSession": {
|
|
23063
|
-
"description": "Get the current session, if any.",
|
|
23064
|
-
"parameters": {},
|
|
23065
|
-
"required": [],
|
|
23066
|
-
"returns": "void"
|
|
23067
|
-
},
|
|
23068
|
-
"getUser": {
|
|
23069
|
-
"description": "Get the current user, if any.",
|
|
23070
|
-
"parameters": {},
|
|
23071
|
-
"required": [],
|
|
23072
|
-
"returns": "void"
|
|
23073
|
-
},
|
|
23074
|
-
"invoke": {
|
|
23075
|
-
"description": "Invoke a Supabase Edge Function by name.",
|
|
23076
|
-
"parameters": {
|
|
23077
|
-
"name": {
|
|
23078
|
-
"type": "string",
|
|
23079
|
-
"description": "Parameter name"
|
|
23080
|
-
},
|
|
23081
|
-
"body": {
|
|
23082
|
-
"type": "any",
|
|
23083
|
-
"description": "Parameter body"
|
|
23021
|
+
"returns": "Promise<OpenAI.Chat.Completions.ChatCompletion>",
|
|
23022
|
+
"examples": [
|
|
23023
|
+
{
|
|
23024
|
+
"language": "ts",
|
|
23025
|
+
"code": "const response = await openai.createChatCompletion([\n { role: 'system', content: 'You are a helpful assistant.' },\n { role: 'user', content: 'Hello!' }\n])\nconsole.log(response.choices[0]?.message?.content)"
|
|
23084
23026
|
}
|
|
23085
|
-
|
|
23086
|
-
"required": [
|
|
23087
|
-
"name"
|
|
23088
|
-
],
|
|
23089
|
-
"returns": "void"
|
|
23027
|
+
]
|
|
23090
23028
|
},
|
|
23091
|
-
"
|
|
23092
|
-
"description": "
|
|
23029
|
+
"createResponse": {
|
|
23030
|
+
"description": "Create a response using the Responses API.",
|
|
23093
23031
|
"parameters": {
|
|
23094
|
-
"
|
|
23095
|
-
"type": "string",
|
|
23096
|
-
"description": "
|
|
23097
|
-
},
|
|
23098
|
-
"table": {
|
|
23099
|
-
"type": "string",
|
|
23100
|
-
"description": "The table to listen to"
|
|
23101
|
-
},
|
|
23102
|
-
"callback": {
|
|
23103
|
-
"type": "(payload: any) => void",
|
|
23104
|
-
"description": "Called with the payload on each change"
|
|
23032
|
+
"input": {
|
|
23033
|
+
"type": "OpenAI.Responses.ResponseInput | string",
|
|
23034
|
+
"description": "The input prompt or message array"
|
|
23105
23035
|
},
|
|
23106
|
-
"
|
|
23107
|
-
"type": "
|
|
23108
|
-
"description": "
|
|
23036
|
+
"options": {
|
|
23037
|
+
"type": "Partial<OpenAI.Responses.ResponseCreateParamsNonStreaming>",
|
|
23038
|
+
"description": "Additional parameters for the response"
|
|
23109
23039
|
}
|
|
23110
23040
|
},
|
|
23111
23041
|
"required": [
|
|
23112
|
-
"
|
|
23113
|
-
"table",
|
|
23114
|
-
"callback"
|
|
23042
|
+
"input"
|
|
23115
23043
|
],
|
|
23116
|
-
"returns": "
|
|
23117
|
-
|
|
23118
|
-
|
|
23119
|
-
|
|
23120
|
-
|
|
23121
|
-
"channelName": {
|
|
23122
|
-
"type": "string",
|
|
23123
|
-
"description": "The channel name to remove"
|
|
23044
|
+
"returns": "Promise<OpenAI.Responses.Response>",
|
|
23045
|
+
"examples": [
|
|
23046
|
+
{
|
|
23047
|
+
"language": "ts",
|
|
23048
|
+
"code": "const response = await openai.createResponse('Explain quantum computing')"
|
|
23124
23049
|
}
|
|
23125
|
-
|
|
23126
|
-
"required": [
|
|
23127
|
-
"channelName"
|
|
23128
|
-
],
|
|
23129
|
-
"returns": "void"
|
|
23130
|
-
},
|
|
23131
|
-
"unsubscribeAll": {
|
|
23132
|
-
"description": "Unsubscribe and remove all realtime channels.",
|
|
23133
|
-
"parameters": {},
|
|
23134
|
-
"required": [],
|
|
23135
|
-
"returns": "void"
|
|
23136
|
-
},
|
|
23137
|
-
"connect": {
|
|
23138
|
-
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
23139
|
-
"parameters": {},
|
|
23140
|
-
"required": [],
|
|
23141
|
-
"returns": "void"
|
|
23142
|
-
},
|
|
23143
|
-
"disconnect": {
|
|
23144
|
-
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
23145
|
-
"parameters": {},
|
|
23146
|
-
"required": [],
|
|
23147
|
-
"returns": "void"
|
|
23148
|
-
}
|
|
23149
|
-
},
|
|
23150
|
-
"getters": {
|
|
23151
|
-
"sdk": {
|
|
23152
|
-
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
23153
|
-
"returns": "SupabaseSDKClient<any, any>"
|
|
23154
|
-
},
|
|
23155
|
-
"storage": {
|
|
23156
|
-
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
23157
|
-
"returns": "any"
|
|
23158
|
-
},
|
|
23159
|
-
"functions": {
|
|
23160
|
-
"description": "Returns the Supabase Functions client.",
|
|
23161
|
-
"returns": "any"
|
|
23162
|
-
}
|
|
23163
|
-
},
|
|
23164
|
-
"events": {},
|
|
23165
|
-
"state": {},
|
|
23166
|
-
"options": {},
|
|
23167
|
-
"envVars": [],
|
|
23168
|
-
"examples": [
|
|
23169
|
-
{
|
|
23170
|
-
"language": "ts",
|
|
23171
|
-
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
23172
|
-
}
|
|
23173
|
-
]
|
|
23174
|
-
},
|
|
23175
|
-
{
|
|
23176
|
-
"id": "clients.elevenlabs",
|
|
23177
|
-
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
23178
|
-
"shortcut": "clients.elevenlabs",
|
|
23179
|
-
"className": "ElevenLabsClient",
|
|
23180
|
-
"methods": {
|
|
23181
|
-
"beforeRequest": {
|
|
23182
|
-
"description": "Inject the xi-api-key header before each request.",
|
|
23183
|
-
"parameters": {},
|
|
23184
|
-
"required": [],
|
|
23185
|
-
"returns": "void"
|
|
23050
|
+
]
|
|
23186
23051
|
},
|
|
23187
|
-
"
|
|
23188
|
-
"description": "
|
|
23189
|
-
"parameters": {
|
|
23190
|
-
|
|
23191
|
-
|
|
23052
|
+
"streamResponse": {
|
|
23053
|
+
"description": "Stream a response using the Responses API.",
|
|
23054
|
+
"parameters": {
|
|
23055
|
+
"input": {
|
|
23056
|
+
"type": "OpenAI.Responses.ResponseInput | string",
|
|
23057
|
+
"description": "The input prompt or message array"
|
|
23058
|
+
},
|
|
23059
|
+
"options": {
|
|
23060
|
+
"type": "Partial<OpenAI.Responses.ResponseCreateParamsStreaming>",
|
|
23061
|
+
"description": "Additional parameters for the streaming response"
|
|
23062
|
+
}
|
|
23063
|
+
},
|
|
23064
|
+
"required": [
|
|
23065
|
+
"input"
|
|
23066
|
+
],
|
|
23067
|
+
"returns": "Promise<AsyncIterable<OpenAI.Responses.ResponseStreamEvent>>",
|
|
23192
23068
|
"examples": [
|
|
23193
23069
|
{
|
|
23194
23070
|
"language": "ts",
|
|
23195
|
-
"code": "await
|
|
23071
|
+
"code": "const stream = await openai.streamResponse('Write a poem')\nfor await (const event of stream) {\n if (event.type === 'response.output_text.delta') {\n process.stdout.write(event.delta)\n }\n}"
|
|
23196
23072
|
}
|
|
23197
23073
|
]
|
|
23198
23074
|
},
|
|
23199
|
-
"
|
|
23200
|
-
"description": "
|
|
23075
|
+
"createCompletion": {
|
|
23076
|
+
"description": "Create a legacy text completion.",
|
|
23201
23077
|
"parameters": {
|
|
23078
|
+
"prompt": {
|
|
23079
|
+
"type": "string",
|
|
23080
|
+
"description": "The text prompt to complete"
|
|
23081
|
+
},
|
|
23202
23082
|
"options": {
|
|
23203
|
-
"type": "
|
|
23204
|
-
"description": "
|
|
23083
|
+
"type": "Partial<OpenAI.Completions.CompletionCreateParams>",
|
|
23084
|
+
"description": "Additional parameters for the completion"
|
|
23205
23085
|
}
|
|
23206
23086
|
},
|
|
23207
|
-
"required": [
|
|
23208
|
-
|
|
23087
|
+
"required": [
|
|
23088
|
+
"prompt"
|
|
23089
|
+
],
|
|
23090
|
+
"returns": "Promise<OpenAI.Completions.Completion>",
|
|
23209
23091
|
"examples": [
|
|
23210
23092
|
{
|
|
23211
23093
|
"language": "ts",
|
|
23212
|
-
"code": "const
|
|
23094
|
+
"code": "const response = await openai.createCompletion('Once upon a time')"
|
|
23213
23095
|
}
|
|
23214
23096
|
]
|
|
23215
23097
|
},
|
|
23216
|
-
"
|
|
23217
|
-
"description": "
|
|
23098
|
+
"createEmbedding": {
|
|
23099
|
+
"description": "Create text embeddings for semantic search or similarity comparisons.",
|
|
23218
23100
|
"parameters": {
|
|
23219
|
-
"
|
|
23101
|
+
"input": {
|
|
23102
|
+
"type": "string | string[]",
|
|
23103
|
+
"description": "A string or array of strings to embed"
|
|
23104
|
+
},
|
|
23105
|
+
"options": {
|
|
23106
|
+
"type": "Partial<OpenAI.Embeddings.EmbeddingCreateParams>",
|
|
23107
|
+
"description": "Additional parameters (model, etc.)"
|
|
23108
|
+
}
|
|
23109
|
+
},
|
|
23110
|
+
"required": [
|
|
23111
|
+
"input"
|
|
23112
|
+
],
|
|
23113
|
+
"returns": "Promise<OpenAI.Embeddings.CreateEmbeddingResponse>",
|
|
23114
|
+
"examples": [
|
|
23115
|
+
{
|
|
23116
|
+
"language": "ts",
|
|
23117
|
+
"code": "const response = await openai.createEmbedding('Hello world')\nconsole.log(response.data[0].embedding.length)"
|
|
23118
|
+
}
|
|
23119
|
+
]
|
|
23120
|
+
},
|
|
23121
|
+
"createImage": {
|
|
23122
|
+
"description": "Generate an image from a text prompt using DALL-E.",
|
|
23123
|
+
"parameters": {
|
|
23124
|
+
"prompt": {
|
|
23220
23125
|
"type": "string",
|
|
23221
|
-
"description": "
|
|
23126
|
+
"description": "Description of the image to generate"
|
|
23127
|
+
},
|
|
23128
|
+
"options": {
|
|
23129
|
+
"type": "Partial<OpenAI.Images.ImageGenerateParams>",
|
|
23130
|
+
"description": "Additional parameters (size, n, etc.)"
|
|
23222
23131
|
}
|
|
23223
23132
|
},
|
|
23224
23133
|
"required": [
|
|
23225
|
-
"
|
|
23134
|
+
"prompt"
|
|
23226
23135
|
],
|
|
23227
|
-
"returns": "Promise<
|
|
23136
|
+
"returns": "Promise<OpenAI.Images.ImagesResponse>",
|
|
23228
23137
|
"examples": [
|
|
23229
23138
|
{
|
|
23230
23139
|
"language": "ts",
|
|
23231
|
-
"code": "const
|
|
23140
|
+
"code": "const response = await openai.createImage('A sunset over mountains')\nconsole.log(response.data[0].url)"
|
|
23232
23141
|
}
|
|
23233
23142
|
]
|
|
23234
23143
|
},
|
|
23235
23144
|
"listModels": {
|
|
23236
|
-
"description": "List available
|
|
23145
|
+
"description": "List all available models.",
|
|
23237
23146
|
"parameters": {},
|
|
23238
23147
|
"required": [],
|
|
23239
|
-
"returns": "Promise<
|
|
23148
|
+
"returns": "Promise<OpenAI.Models.ModelsPage>",
|
|
23240
23149
|
"examples": [
|
|
23241
23150
|
{
|
|
23242
23151
|
"language": "ts",
|
|
23243
|
-
"code": "const models = await
|
|
23152
|
+
"code": "const models = await openai.listModels()"
|
|
23244
23153
|
}
|
|
23245
23154
|
]
|
|
23246
23155
|
},
|
|
23247
|
-
"
|
|
23248
|
-
"description": "
|
|
23156
|
+
"ask": {
|
|
23157
|
+
"description": "Ask a single question and get a text response. Convenience wrapper around `createChatCompletion` for simple Q&A.",
|
|
23249
23158
|
"parameters": {
|
|
23250
|
-
"
|
|
23159
|
+
"question": {
|
|
23251
23160
|
"type": "string",
|
|
23252
|
-
"description": "The
|
|
23161
|
+
"description": "The question to ask"
|
|
23253
23162
|
},
|
|
23254
23163
|
"options": {
|
|
23255
|
-
"type": "
|
|
23256
|
-
"description": "
|
|
23257
|
-
"properties": {
|
|
23258
|
-
"voiceId": {
|
|
23259
|
-
"type": "string",
|
|
23260
|
-
"description": ""
|
|
23261
|
-
},
|
|
23262
|
-
"modelId": {
|
|
23263
|
-
"type": "string",
|
|
23264
|
-
"description": ""
|
|
23265
|
-
},
|
|
23266
|
-
"outputFormat": {
|
|
23267
|
-
"type": "string",
|
|
23268
|
-
"description": ""
|
|
23269
|
-
},
|
|
23270
|
-
"voiceSettings": {
|
|
23271
|
-
"type": "ElevenLabsVoiceSettings",
|
|
23272
|
-
"description": ""
|
|
23273
|
-
},
|
|
23274
|
-
"disableCache": {
|
|
23275
|
-
"type": "boolean",
|
|
23276
|
-
"description": ""
|
|
23277
|
-
}
|
|
23278
|
-
}
|
|
23164
|
+
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
23165
|
+
"description": "Additional completion parameters"
|
|
23279
23166
|
}
|
|
23280
23167
|
},
|
|
23281
23168
|
"required": [
|
|
23282
|
-
"
|
|
23169
|
+
"question"
|
|
23283
23170
|
],
|
|
23284
|
-
"returns": "Promise<
|
|
23171
|
+
"returns": "Promise<string>",
|
|
23285
23172
|
"examples": [
|
|
23286
23173
|
{
|
|
23287
23174
|
"language": "ts",
|
|
23288
|
-
"code": "const
|
|
23175
|
+
"code": "const answer = await openai.ask('What is 2 + 2?')\nconsole.log(answer) // '4'"
|
|
23289
23176
|
}
|
|
23290
23177
|
]
|
|
23291
23178
|
},
|
|
23292
|
-
"
|
|
23293
|
-
"description": "
|
|
23179
|
+
"chat": {
|
|
23180
|
+
"description": "Send a multi-turn conversation and get a text response. Convenience wrapper around `createChatCompletion` that returns just the text.",
|
|
23294
23181
|
"parameters": {
|
|
23295
|
-
"
|
|
23296
|
-
"type": "
|
|
23297
|
-
"description": "
|
|
23298
|
-
},
|
|
23299
|
-
"outputPath": {
|
|
23300
|
-
"type": "string",
|
|
23301
|
-
"description": "File path to write the audio to"
|
|
23182
|
+
"messages": {
|
|
23183
|
+
"type": "OpenAI.Chat.Completions.ChatCompletionMessageParam[]",
|
|
23184
|
+
"description": "Array of chat messages"
|
|
23302
23185
|
},
|
|
23303
23186
|
"options": {
|
|
23304
|
-
"type": "
|
|
23305
|
-
"description": "
|
|
23306
|
-
"properties": {
|
|
23307
|
-
"voiceId": {
|
|
23308
|
-
"type": "string",
|
|
23309
|
-
"description": ""
|
|
23310
|
-
},
|
|
23311
|
-
"modelId": {
|
|
23312
|
-
"type": "string",
|
|
23313
|
-
"description": ""
|
|
23314
|
-
},
|
|
23315
|
-
"outputFormat": {
|
|
23316
|
-
"type": "string",
|
|
23317
|
-
"description": ""
|
|
23318
|
-
},
|
|
23319
|
-
"voiceSettings": {
|
|
23320
|
-
"type": "ElevenLabsVoiceSettings",
|
|
23321
|
-
"description": ""
|
|
23322
|
-
},
|
|
23323
|
-
"disableCache": {
|
|
23324
|
-
"type": "boolean",
|
|
23325
|
-
"description": ""
|
|
23326
|
-
}
|
|
23327
|
-
}
|
|
23187
|
+
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
23188
|
+
"description": "Additional completion parameters"
|
|
23328
23189
|
}
|
|
23329
23190
|
},
|
|
23330
23191
|
"required": [
|
|
23331
|
-
"
|
|
23332
|
-
"outputPath"
|
|
23192
|
+
"messages"
|
|
23333
23193
|
],
|
|
23334
23194
|
"returns": "Promise<string>",
|
|
23335
23195
|
"examples": [
|
|
23336
23196
|
{
|
|
23337
23197
|
"language": "ts",
|
|
23338
|
-
"code": "const
|
|
23198
|
+
"code": "const reply = await openai.chat([\n { role: 'system', content: 'You are a pirate.' },\n { role: 'user', content: 'Hello!' }\n])"
|
|
23339
23199
|
}
|
|
23340
23200
|
]
|
|
23341
23201
|
}
|
|
23342
23202
|
},
|
|
23343
23203
|
"getters": {
|
|
23344
|
-
"
|
|
23345
|
-
"description": "The
|
|
23204
|
+
"defaultModel": {
|
|
23205
|
+
"description": "The default model used for completions, from options or 'gpt-4o'.",
|
|
23346
23206
|
"returns": "string"
|
|
23207
|
+
},
|
|
23208
|
+
"raw": {
|
|
23209
|
+
"description": "The underlying OpenAI SDK instance for advanced use cases.",
|
|
23210
|
+
"returns": "OpenAI"
|
|
23347
23211
|
}
|
|
23348
23212
|
},
|
|
23349
23213
|
"events": {
|
|
23214
|
+
"connected": {
|
|
23215
|
+
"name": "connected",
|
|
23216
|
+
"description": "Event emitted by OpenAIClient",
|
|
23217
|
+
"arguments": {}
|
|
23218
|
+
},
|
|
23350
23219
|
"failure": {
|
|
23351
23220
|
"name": "failure",
|
|
23352
|
-
"description": "Event emitted by
|
|
23221
|
+
"description": "Event emitted by OpenAIClient",
|
|
23353
23222
|
"arguments": {}
|
|
23354
23223
|
},
|
|
23355
|
-
"
|
|
23356
|
-
"name": "
|
|
23357
|
-
"description": "Event emitted by
|
|
23224
|
+
"completion": {
|
|
23225
|
+
"name": "completion",
|
|
23226
|
+
"description": "Event emitted by OpenAIClient",
|
|
23358
23227
|
"arguments": {}
|
|
23359
23228
|
},
|
|
23360
|
-
"
|
|
23361
|
-
"name": "
|
|
23362
|
-
"description": "Event emitted by
|
|
23229
|
+
"embedding": {
|
|
23230
|
+
"name": "embedding",
|
|
23231
|
+
"description": "Event emitted by OpenAIClient",
|
|
23232
|
+
"arguments": {}
|
|
23233
|
+
},
|
|
23234
|
+
"image": {
|
|
23235
|
+
"name": "image",
|
|
23236
|
+
"description": "Event emitted by OpenAIClient",
|
|
23237
|
+
"arguments": {}
|
|
23238
|
+
},
|
|
23239
|
+
"models": {
|
|
23240
|
+
"name": "models",
|
|
23241
|
+
"description": "Event emitted by OpenAIClient",
|
|
23363
23242
|
"arguments": {}
|
|
23364
23243
|
}
|
|
23365
23244
|
},
|
|
@@ -23369,248 +23248,393 @@ export const introspectionData = [
|
|
|
23369
23248
|
"examples": [
|
|
23370
23249
|
{
|
|
23371
23250
|
"language": "ts",
|
|
23372
|
-
"code": "const
|
|
23251
|
+
"code": "const openai = container.client('openai', { defaultModel: 'gpt-4o' })\nconst answer = await openai.ask('What is the meaning of life?')\nconsole.log(answer)"
|
|
23373
23252
|
}
|
|
23374
23253
|
]
|
|
23375
23254
|
},
|
|
23376
23255
|
{
|
|
23377
|
-
"id": "clients.
|
|
23378
|
-
"description": "
|
|
23379
|
-
"shortcut": "clients.
|
|
23380
|
-
"className": "
|
|
23256
|
+
"id": "clients.supabase",
|
|
23257
|
+
"description": "Supabase client for the Luca container system. Wraps the official `@supabase/supabase-js` SDK and exposes it through Luca's typed state, events, and introspection system. The SDK is isomorphic so this single implementation works in both Node and browser containers. Use `client.sdk` for full SDK access, or use the convenience wrappers for common operations (auth, database queries, storage, edge functions, realtime).",
|
|
23258
|
+
"shortcut": "clients.supabase",
|
|
23259
|
+
"className": "SupabaseClient",
|
|
23381
23260
|
"methods": {
|
|
23382
|
-
"
|
|
23383
|
-
"description": "
|
|
23261
|
+
"from": {
|
|
23262
|
+
"description": "Start a query on a Postgres table or view.",
|
|
23263
|
+
"parameters": {
|
|
23264
|
+
"table": {
|
|
23265
|
+
"type": "string",
|
|
23266
|
+
"description": "The table or view name to query"
|
|
23267
|
+
}
|
|
23268
|
+
},
|
|
23269
|
+
"required": [
|
|
23270
|
+
"table"
|
|
23271
|
+
],
|
|
23272
|
+
"returns": "void"
|
|
23273
|
+
},
|
|
23274
|
+
"rpc": {
|
|
23275
|
+
"description": "Call a Postgres function (RPC).",
|
|
23276
|
+
"parameters": {
|
|
23277
|
+
"fn": {
|
|
23278
|
+
"type": "string",
|
|
23279
|
+
"description": "The function name"
|
|
23280
|
+
},
|
|
23281
|
+
"params": {
|
|
23282
|
+
"type": "Record<string, unknown>",
|
|
23283
|
+
"description": "Arguments to pass to the function"
|
|
23284
|
+
},
|
|
23285
|
+
"options": {
|
|
23286
|
+
"type": "{ head?: boolean; get?: boolean; count?: \"exact\" | \"planned\" | \"estimated\" }",
|
|
23287
|
+
"description": "Optional settings (head, get, count)"
|
|
23288
|
+
}
|
|
23289
|
+
},
|
|
23290
|
+
"required": [
|
|
23291
|
+
"fn"
|
|
23292
|
+
],
|
|
23293
|
+
"returns": "void"
|
|
23294
|
+
},
|
|
23295
|
+
"signInWithPassword": {
|
|
23296
|
+
"description": "Sign in with email and password.",
|
|
23384
23297
|
"parameters": {
|
|
23385
|
-
"
|
|
23386
|
-
"type": "
|
|
23387
|
-
"description": "
|
|
23298
|
+
"email": {
|
|
23299
|
+
"type": "string",
|
|
23300
|
+
"description": "Parameter email"
|
|
23388
23301
|
},
|
|
23389
|
-
"
|
|
23302
|
+
"password": {
|
|
23390
23303
|
"type": "string",
|
|
23391
|
-
"description": "
|
|
23304
|
+
"description": "Parameter password"
|
|
23392
23305
|
}
|
|
23393
23306
|
},
|
|
23394
23307
|
"required": [
|
|
23395
|
-
"
|
|
23308
|
+
"email",
|
|
23309
|
+
"password"
|
|
23396
23310
|
],
|
|
23397
|
-
"returns": "
|
|
23398
|
-
"examples": [
|
|
23399
|
-
{
|
|
23400
|
-
"language": "ts",
|
|
23401
|
-
"code": "const { prompt_id } = await comfy.queuePrompt(apiWorkflow)"
|
|
23402
|
-
}
|
|
23403
|
-
]
|
|
23404
|
-
},
|
|
23405
|
-
"getQueue": {
|
|
23406
|
-
"description": "Get the current prompt queue status.",
|
|
23407
|
-
"parameters": {},
|
|
23408
|
-
"required": [],
|
|
23409
|
-
"returns": "Promise<{ queue_running: any[]; queue_pending: any[] }>"
|
|
23311
|
+
"returns": "void"
|
|
23410
23312
|
},
|
|
23411
|
-
"
|
|
23412
|
-
"description": "
|
|
23313
|
+
"signUp": {
|
|
23314
|
+
"description": "Create a new user account with email and password.",
|
|
23413
23315
|
"parameters": {
|
|
23414
|
-
"
|
|
23316
|
+
"email": {
|
|
23415
23317
|
"type": "string",
|
|
23416
|
-
"description": "
|
|
23318
|
+
"description": "Parameter email"
|
|
23319
|
+
},
|
|
23320
|
+
"password": {
|
|
23321
|
+
"type": "string",
|
|
23322
|
+
"description": "Parameter password"
|
|
23417
23323
|
}
|
|
23418
23324
|
},
|
|
23419
|
-
"required": [
|
|
23420
|
-
|
|
23325
|
+
"required": [
|
|
23326
|
+
"email",
|
|
23327
|
+
"password"
|
|
23328
|
+
],
|
|
23329
|
+
"returns": "void"
|
|
23421
23330
|
},
|
|
23422
|
-
"
|
|
23423
|
-
"description": "
|
|
23331
|
+
"signOut": {
|
|
23332
|
+
"description": "Sign the current user out.",
|
|
23424
23333
|
"parameters": {},
|
|
23425
23334
|
"required": [],
|
|
23426
|
-
"returns": "
|
|
23335
|
+
"returns": "void"
|
|
23427
23336
|
},
|
|
23428
|
-
"
|
|
23429
|
-
"description": "Get
|
|
23430
|
-
"parameters": {
|
|
23431
|
-
"nodeClass": {
|
|
23432
|
-
"type": "string",
|
|
23433
|
-
"description": "If provided, returns info for this node type only"
|
|
23434
|
-
}
|
|
23435
|
-
},
|
|
23337
|
+
"getSession": {
|
|
23338
|
+
"description": "Get the current session, if any.",
|
|
23339
|
+
"parameters": {},
|
|
23436
23340
|
"required": [],
|
|
23437
|
-
"returns": "
|
|
23341
|
+
"returns": "void"
|
|
23438
23342
|
},
|
|
23439
|
-
"
|
|
23440
|
-
"description": "
|
|
23343
|
+
"getUser": {
|
|
23344
|
+
"description": "Get the current user, if any.",
|
|
23441
23345
|
"parameters": {},
|
|
23442
23346
|
"required": [],
|
|
23443
|
-
"returns": "
|
|
23347
|
+
"returns": "void"
|
|
23444
23348
|
},
|
|
23445
|
-
"
|
|
23446
|
-
"description": "
|
|
23349
|
+
"invoke": {
|
|
23350
|
+
"description": "Invoke a Supabase Edge Function by name.",
|
|
23447
23351
|
"parameters": {
|
|
23448
|
-
"
|
|
23352
|
+
"name": {
|
|
23449
23353
|
"type": "string",
|
|
23450
|
-
"description": "
|
|
23354
|
+
"description": "Parameter name"
|
|
23355
|
+
},
|
|
23356
|
+
"body": {
|
|
23357
|
+
"type": "any",
|
|
23358
|
+
"description": "Parameter body"
|
|
23451
23359
|
}
|
|
23452
23360
|
},
|
|
23453
|
-
"required": [
|
|
23454
|
-
|
|
23455
|
-
|
|
23456
|
-
|
|
23457
|
-
"description": "List available embedding models.",
|
|
23458
|
-
"parameters": {},
|
|
23459
|
-
"required": [],
|
|
23460
|
-
"returns": "Promise<string[]>"
|
|
23361
|
+
"required": [
|
|
23362
|
+
"name"
|
|
23363
|
+
],
|
|
23364
|
+
"returns": "void"
|
|
23461
23365
|
},
|
|
23462
|
-
"
|
|
23463
|
-
"description": "
|
|
23366
|
+
"subscribe": {
|
|
23367
|
+
"description": "Subscribe to realtime changes on a Postgres table.",
|
|
23464
23368
|
"parameters": {
|
|
23465
|
-
"
|
|
23466
|
-
"type": "
|
|
23467
|
-
"description": "
|
|
23369
|
+
"channelName": {
|
|
23370
|
+
"type": "string",
|
|
23371
|
+
"description": "A name for this subscription channel"
|
|
23468
23372
|
},
|
|
23469
|
-
"
|
|
23373
|
+
"table": {
|
|
23470
23374
|
"type": "string",
|
|
23471
|
-
"description": "
|
|
23375
|
+
"description": "The table to listen to"
|
|
23472
23376
|
},
|
|
23473
|
-
"
|
|
23474
|
-
"type": "
|
|
23475
|
-
"description": "
|
|
23377
|
+
"callback": {
|
|
23378
|
+
"type": "(payload: any) => void",
|
|
23379
|
+
"description": "Called with the payload on each change"
|
|
23380
|
+
},
|
|
23381
|
+
"event": {
|
|
23382
|
+
"type": "\"INSERT\" | \"UPDATE\" | \"DELETE\" | \"*\"",
|
|
23383
|
+
"description": "The event type to listen for (default: all changes)"
|
|
23476
23384
|
}
|
|
23477
23385
|
},
|
|
23478
23386
|
"required": [
|
|
23479
|
-
"
|
|
23480
|
-
"
|
|
23387
|
+
"channelName",
|
|
23388
|
+
"table",
|
|
23389
|
+
"callback"
|
|
23481
23390
|
],
|
|
23482
|
-
"returns": "
|
|
23391
|
+
"returns": "RealtimeChannel"
|
|
23483
23392
|
},
|
|
23484
|
-
"
|
|
23485
|
-
"description": "
|
|
23393
|
+
"unsubscribe": {
|
|
23394
|
+
"description": "Unsubscribe and remove a realtime channel by name.",
|
|
23486
23395
|
"parameters": {
|
|
23487
|
-
"
|
|
23396
|
+
"channelName": {
|
|
23488
23397
|
"type": "string",
|
|
23489
|
-
"description": "The
|
|
23490
|
-
},
|
|
23491
|
-
"subfolder": {
|
|
23492
|
-
"type": "any",
|
|
23493
|
-
"description": "Subfolder within the output directory"
|
|
23494
|
-
},
|
|
23495
|
-
"type": {
|
|
23496
|
-
"type": "any",
|
|
23497
|
-
"description": "Image type ('output', 'input', 'temp')"
|
|
23398
|
+
"description": "The channel name to remove"
|
|
23498
23399
|
}
|
|
23499
23400
|
},
|
|
23500
23401
|
"required": [
|
|
23501
|
-
"
|
|
23402
|
+
"channelName"
|
|
23502
23403
|
],
|
|
23503
|
-
"returns": "
|
|
23404
|
+
"returns": "void"
|
|
23504
23405
|
},
|
|
23505
|
-
"
|
|
23506
|
-
"description": "
|
|
23406
|
+
"unsubscribeAll": {
|
|
23407
|
+
"description": "Unsubscribe and remove all realtime channels.",
|
|
23507
23408
|
"parameters": {},
|
|
23508
23409
|
"required": [],
|
|
23509
|
-
"returns": "
|
|
23410
|
+
"returns": "void"
|
|
23510
23411
|
},
|
|
23511
|
-
"
|
|
23512
|
-
"description": "
|
|
23412
|
+
"connect": {
|
|
23413
|
+
"description": "Connect is a no-op since the Supabase SDK initializes on construction. The client is ready to use immediately after creation.",
|
|
23414
|
+
"parameters": {},
|
|
23415
|
+
"required": [],
|
|
23416
|
+
"returns": "void"
|
|
23417
|
+
},
|
|
23418
|
+
"disconnect": {
|
|
23419
|
+
"description": "Disconnect by signing out and removing all realtime channels.",
|
|
23420
|
+
"parameters": {},
|
|
23421
|
+
"required": [],
|
|
23422
|
+
"returns": "void"
|
|
23423
|
+
}
|
|
23424
|
+
},
|
|
23425
|
+
"getters": {
|
|
23426
|
+
"sdk": {
|
|
23427
|
+
"description": "Returns the raw Supabase SDK client for full access to all SDK methods.",
|
|
23428
|
+
"returns": "SupabaseSDKClient<any, any>"
|
|
23429
|
+
},
|
|
23430
|
+
"storage": {
|
|
23431
|
+
"description": "Returns the Supabase Storage client for managing buckets and files.",
|
|
23432
|
+
"returns": "any"
|
|
23433
|
+
},
|
|
23434
|
+
"functions": {
|
|
23435
|
+
"description": "Returns the Supabase Functions client.",
|
|
23436
|
+
"returns": "any"
|
|
23437
|
+
}
|
|
23438
|
+
},
|
|
23439
|
+
"events": {},
|
|
23440
|
+
"state": {},
|
|
23441
|
+
"options": {},
|
|
23442
|
+
"envVars": [],
|
|
23443
|
+
"examples": [
|
|
23444
|
+
{
|
|
23445
|
+
"language": "ts",
|
|
23446
|
+
"code": "const supabase = container.client('supabase', {\n supabaseUrl: 'https://xyz.supabase.co',\n supabaseKey: 'your-anon-key',\n})\n\n// Query data\nconst { data } = await supabase.from('users').select('*')\n\n// Auth\nawait supabase.signInWithPassword('user@example.com', 'password')\n\n// Realtime\nsupabase.subscribe('changes', 'users', (payload) => {\n console.log('Change:', payload)\n})"
|
|
23447
|
+
}
|
|
23448
|
+
]
|
|
23449
|
+
},
|
|
23450
|
+
{
|
|
23451
|
+
"id": "clients.elevenlabs",
|
|
23452
|
+
"description": "ElevenLabs client — text-to-speech synthesis via the ElevenLabs REST API. Provides methods for listing voices, listing models, and generating speech audio. Audio is returned as a Buffer; use `say()` for a convenience method that writes to disk.",
|
|
23453
|
+
"shortcut": "clients.elevenlabs",
|
|
23454
|
+
"className": "ElevenLabsClient",
|
|
23455
|
+
"methods": {
|
|
23456
|
+
"beforeRequest": {
|
|
23457
|
+
"description": "Inject the xi-api-key header before each request.",
|
|
23458
|
+
"parameters": {},
|
|
23459
|
+
"required": [],
|
|
23460
|
+
"returns": "void"
|
|
23461
|
+
},
|
|
23462
|
+
"connect": {
|
|
23463
|
+
"description": "Validate the API key by listing available models.",
|
|
23464
|
+
"parameters": {},
|
|
23465
|
+
"required": [],
|
|
23466
|
+
"returns": "Promise<this>",
|
|
23467
|
+
"examples": [
|
|
23468
|
+
{
|
|
23469
|
+
"language": "ts",
|
|
23470
|
+
"code": "await el.connect()"
|
|
23471
|
+
}
|
|
23472
|
+
]
|
|
23473
|
+
},
|
|
23474
|
+
"listVoices": {
|
|
23475
|
+
"description": "List available voices with optional search and filtering.",
|
|
23476
|
+
"parameters": {
|
|
23477
|
+
"options": {
|
|
23478
|
+
"type": "{\n search?: string\n category?: string\n voice_type?: string\n page_size?: number\n next_page_token?: string\n }",
|
|
23479
|
+
"description": "Query parameters for filtering voices"
|
|
23480
|
+
}
|
|
23481
|
+
},
|
|
23482
|
+
"required": [],
|
|
23483
|
+
"returns": "Promise<any>",
|
|
23484
|
+
"examples": [
|
|
23485
|
+
{
|
|
23486
|
+
"language": "ts",
|
|
23487
|
+
"code": "const voices = await el.listVoices()\nconst premade = await el.listVoices({ category: 'premade' })"
|
|
23488
|
+
}
|
|
23489
|
+
]
|
|
23490
|
+
},
|
|
23491
|
+
"getVoice": {
|
|
23492
|
+
"description": "Get details for a single voice.",
|
|
23493
|
+
"parameters": {
|
|
23494
|
+
"voiceId": {
|
|
23495
|
+
"type": "string",
|
|
23496
|
+
"description": "The voice ID to look up"
|
|
23497
|
+
}
|
|
23498
|
+
},
|
|
23499
|
+
"required": [
|
|
23500
|
+
"voiceId"
|
|
23501
|
+
],
|
|
23502
|
+
"returns": "Promise<any>",
|
|
23503
|
+
"examples": [
|
|
23504
|
+
{
|
|
23505
|
+
"language": "ts",
|
|
23506
|
+
"code": "const voice = await el.getVoice('21m00Tcm4TlvDq8ikWAM')\nconsole.log(voice.name, voice.settings)"
|
|
23507
|
+
}
|
|
23508
|
+
]
|
|
23509
|
+
},
|
|
23510
|
+
"listModels": {
|
|
23511
|
+
"description": "List available TTS models.",
|
|
23513
23512
|
"parameters": {},
|
|
23514
23513
|
"required": [],
|
|
23515
|
-
"returns": "
|
|
23514
|
+
"returns": "Promise<any[]>",
|
|
23515
|
+
"examples": [
|
|
23516
|
+
{
|
|
23517
|
+
"language": "ts",
|
|
23518
|
+
"code": "const models = await el.listModels()\nconsole.log(models.map(m => m.model_id))"
|
|
23519
|
+
}
|
|
23520
|
+
]
|
|
23516
23521
|
},
|
|
23517
|
-
"
|
|
23518
|
-
"description": "
|
|
23522
|
+
"synthesize": {
|
|
23523
|
+
"description": "Synthesize speech from text, returning audio as a Buffer.",
|
|
23519
23524
|
"parameters": {
|
|
23520
|
-
"
|
|
23521
|
-
"type": "
|
|
23522
|
-
"description": "
|
|
23525
|
+
"text": {
|
|
23526
|
+
"type": "string",
|
|
23527
|
+
"description": "The text to convert to speech"
|
|
23528
|
+
},
|
|
23529
|
+
"options": {
|
|
23530
|
+
"type": "SynthesizeOptions",
|
|
23531
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
23532
|
+
"properties": {
|
|
23533
|
+
"voiceId": {
|
|
23534
|
+
"type": "string",
|
|
23535
|
+
"description": ""
|
|
23536
|
+
},
|
|
23537
|
+
"modelId": {
|
|
23538
|
+
"type": "string",
|
|
23539
|
+
"description": ""
|
|
23540
|
+
},
|
|
23541
|
+
"outputFormat": {
|
|
23542
|
+
"type": "string",
|
|
23543
|
+
"description": ""
|
|
23544
|
+
},
|
|
23545
|
+
"voiceSettings": {
|
|
23546
|
+
"type": "ElevenLabsVoiceSettings",
|
|
23547
|
+
"description": ""
|
|
23548
|
+
},
|
|
23549
|
+
"disableCache": {
|
|
23550
|
+
"type": "boolean",
|
|
23551
|
+
"description": ""
|
|
23552
|
+
}
|
|
23553
|
+
}
|
|
23523
23554
|
}
|
|
23524
23555
|
},
|
|
23525
23556
|
"required": [
|
|
23526
|
-
"
|
|
23557
|
+
"text"
|
|
23527
23558
|
],
|
|
23528
|
-
"returns": "Promise<
|
|
23559
|
+
"returns": "Promise<Buffer>",
|
|
23560
|
+
"examples": [
|
|
23561
|
+
{
|
|
23562
|
+
"language": "ts",
|
|
23563
|
+
"code": "const audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data\n\nconst custom = await el.synthesize('Hello', {\n voiceId: '21m00Tcm4TlvDq8ikWAM',\n voiceSettings: { stability: 0.5, similarityBoost: 0.8 }\n})"
|
|
23564
|
+
}
|
|
23565
|
+
]
|
|
23529
23566
|
},
|
|
23530
|
-
"
|
|
23531
|
-
"description": "
|
|
23567
|
+
"say": {
|
|
23568
|
+
"description": "Synthesize speech and write the audio to a file.",
|
|
23532
23569
|
"parameters": {
|
|
23533
|
-
"
|
|
23534
|
-
"type": "
|
|
23535
|
-
"description": "
|
|
23570
|
+
"text": {
|
|
23571
|
+
"type": "string",
|
|
23572
|
+
"description": "The text to convert to speech"
|
|
23536
23573
|
},
|
|
23537
|
-
"
|
|
23538
|
-
"type": "
|
|
23539
|
-
"description": "
|
|
23574
|
+
"outputPath": {
|
|
23575
|
+
"type": "string",
|
|
23576
|
+
"description": "File path to write the audio to"
|
|
23540
23577
|
},
|
|
23541
23578
|
"options": {
|
|
23542
|
-
"type": "
|
|
23543
|
-
"description": "
|
|
23579
|
+
"type": "SynthesizeOptions",
|
|
23580
|
+
"description": "Voice, model, format, and voice settings overrides",
|
|
23544
23581
|
"properties": {
|
|
23545
|
-
"
|
|
23546
|
-
"type": "
|
|
23547
|
-
"description": "
|
|
23548
|
-
},
|
|
23549
|
-
"pollInterval": {
|
|
23550
|
-
"type": "number",
|
|
23551
|
-
"description": "Polling interval in ms (default 1000)"
|
|
23582
|
+
"voiceId": {
|
|
23583
|
+
"type": "string",
|
|
23584
|
+
"description": ""
|
|
23552
23585
|
},
|
|
23553
|
-
"
|
|
23554
|
-
"type": "
|
|
23555
|
-
"description": "
|
|
23586
|
+
"modelId": {
|
|
23587
|
+
"type": "string",
|
|
23588
|
+
"description": ""
|
|
23556
23589
|
},
|
|
23557
|
-
"
|
|
23590
|
+
"outputFormat": {
|
|
23558
23591
|
"type": "string",
|
|
23559
|
-
"description": "
|
|
23592
|
+
"description": ""
|
|
23593
|
+
},
|
|
23594
|
+
"voiceSettings": {
|
|
23595
|
+
"type": "ElevenLabsVoiceSettings",
|
|
23596
|
+
"description": ""
|
|
23597
|
+
},
|
|
23598
|
+
"disableCache": {
|
|
23599
|
+
"type": "boolean",
|
|
23600
|
+
"description": ""
|
|
23560
23601
|
}
|
|
23561
23602
|
}
|
|
23562
23603
|
}
|
|
23563
23604
|
},
|
|
23564
23605
|
"required": [
|
|
23565
|
-
"
|
|
23606
|
+
"text",
|
|
23607
|
+
"outputPath"
|
|
23566
23608
|
],
|
|
23567
|
-
"returns": "Promise<
|
|
23609
|
+
"returns": "Promise<string>",
|
|
23610
|
+
"examples": [
|
|
23611
|
+
{
|
|
23612
|
+
"language": "ts",
|
|
23613
|
+
"code": "const path = await el.say('Hello world', './hello.mp3')\nconsole.log(`Audio saved to ${path}`)"
|
|
23614
|
+
}
|
|
23615
|
+
]
|
|
23568
23616
|
}
|
|
23569
23617
|
},
|
|
23570
23618
|
"getters": {
|
|
23571
|
-
"
|
|
23572
|
-
"description": "The
|
|
23573
|
-
"returns": "string"
|
|
23574
|
-
},
|
|
23575
|
-
"wsURL": {
|
|
23576
|
-
"description": "The WebSocket URL derived from baseURL or overridden via options.",
|
|
23619
|
+
"apiKey": {
|
|
23620
|
+
"description": "The resolved API key from options or environment.",
|
|
23577
23621
|
"returns": "string"
|
|
23578
23622
|
}
|
|
23579
23623
|
},
|
|
23580
23624
|
"events": {
|
|
23581
|
-
"
|
|
23582
|
-
"name": "
|
|
23583
|
-
"description": "Event emitted by
|
|
23584
|
-
"arguments": {}
|
|
23585
|
-
},
|
|
23586
|
-
"execution_complete": {
|
|
23587
|
-
"name": "execution_complete",
|
|
23588
|
-
"description": "Event emitted by ComfyUIClient",
|
|
23589
|
-
"arguments": {}
|
|
23590
|
-
},
|
|
23591
|
-
"executing": {
|
|
23592
|
-
"name": "executing",
|
|
23593
|
-
"description": "Event emitted by ComfyUIClient",
|
|
23594
|
-
"arguments": {}
|
|
23595
|
-
},
|
|
23596
|
-
"progress": {
|
|
23597
|
-
"name": "progress",
|
|
23598
|
-
"description": "Event emitted by ComfyUIClient",
|
|
23599
|
-
"arguments": {}
|
|
23600
|
-
},
|
|
23601
|
-
"executed": {
|
|
23602
|
-
"name": "executed",
|
|
23603
|
-
"description": "Event emitted by ComfyUIClient",
|
|
23625
|
+
"failure": {
|
|
23626
|
+
"name": "failure",
|
|
23627
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23604
23628
|
"arguments": {}
|
|
23605
23629
|
},
|
|
23606
|
-
"
|
|
23607
|
-
"name": "
|
|
23608
|
-
"description": "Event emitted by
|
|
23630
|
+
"voices": {
|
|
23631
|
+
"name": "voices",
|
|
23632
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23609
23633
|
"arguments": {}
|
|
23610
23634
|
},
|
|
23611
|
-
"
|
|
23612
|
-
"name": "
|
|
23613
|
-
"description": "Event emitted by
|
|
23635
|
+
"speech": {
|
|
23636
|
+
"name": "speech",
|
|
23637
|
+
"description": "Event emitted by ElevenLabsClient",
|
|
23614
23638
|
"arguments": {}
|
|
23615
23639
|
}
|
|
23616
23640
|
},
|
|
@@ -23620,264 +23644,248 @@ export const introspectionData = [
|
|
|
23620
23644
|
"examples": [
|
|
23621
23645
|
{
|
|
23622
23646
|
"language": "ts",
|
|
23623
|
-
"code": "const
|
|
23647
|
+
"code": "const el = container.client('elevenlabs')\nawait el.connect()\nconst voices = await el.listVoices()\nconst audio = await el.synthesize('Hello world')\n// audio is a Buffer of mp3 data"
|
|
23624
23648
|
}
|
|
23625
23649
|
]
|
|
23626
23650
|
},
|
|
23627
23651
|
{
|
|
23628
|
-
"id": "clients.
|
|
23629
|
-
"description": "
|
|
23630
|
-
"shortcut": "clients.
|
|
23631
|
-
"className": "
|
|
23652
|
+
"id": "clients.comfyui",
|
|
23653
|
+
"description": "ComfyUI client — execute Stable Diffusion workflows via the ComfyUI API. Connects to a ComfyUI instance to queue prompts, track execution via WebSocket or polling, and download generated images. Supports both UI-format and API-format workflows with automatic conversion.",
|
|
23654
|
+
"shortcut": "clients.comfyui",
|
|
23655
|
+
"className": "ComfyUIClient",
|
|
23632
23656
|
"methods": {
|
|
23633
|
-
"
|
|
23634
|
-
"description": "
|
|
23635
|
-
"parameters": {},
|
|
23636
|
-
"required": [],
|
|
23637
|
-
"returns": "Promise<this>",
|
|
23638
|
-
"examples": [
|
|
23639
|
-
{
|
|
23640
|
-
"language": "ts",
|
|
23641
|
-
"code": "await openai.connect()"
|
|
23642
|
-
}
|
|
23643
|
-
]
|
|
23644
|
-
},
|
|
23645
|
-
"createChatCompletion": {
|
|
23646
|
-
"description": "Create a chat completion using the Chat Completions API.",
|
|
23657
|
+
"queuePrompt": {
|
|
23658
|
+
"description": "Queue a prompt (API-format workflow) for execution.",
|
|
23647
23659
|
"parameters": {
|
|
23648
|
-
"
|
|
23649
|
-
"type": "
|
|
23650
|
-
"description": "
|
|
23660
|
+
"prompt": {
|
|
23661
|
+
"type": "Record<string, any>",
|
|
23662
|
+
"description": "The API-format workflow object"
|
|
23651
23663
|
},
|
|
23652
|
-
"
|
|
23653
|
-
"type": "
|
|
23654
|
-
"description": "
|
|
23664
|
+
"clientId": {
|
|
23665
|
+
"type": "string",
|
|
23666
|
+
"description": "Override the client ID for this request"
|
|
23655
23667
|
}
|
|
23656
23668
|
},
|
|
23657
23669
|
"required": [
|
|
23658
|
-
"
|
|
23670
|
+
"prompt"
|
|
23659
23671
|
],
|
|
23660
|
-
"returns": "Promise<
|
|
23672
|
+
"returns": "Promise<{ prompt_id: string; number: number }>",
|
|
23661
23673
|
"examples": [
|
|
23662
23674
|
{
|
|
23663
23675
|
"language": "ts",
|
|
23664
|
-
"code": "const
|
|
23676
|
+
"code": "const { prompt_id } = await comfy.queuePrompt(apiWorkflow)"
|
|
23665
23677
|
}
|
|
23666
23678
|
]
|
|
23667
23679
|
},
|
|
23668
|
-
"
|
|
23669
|
-
"description": "
|
|
23680
|
+
"getQueue": {
|
|
23681
|
+
"description": "Get the current prompt queue status.",
|
|
23682
|
+
"parameters": {},
|
|
23683
|
+
"required": [],
|
|
23684
|
+
"returns": "Promise<{ queue_running: any[]; queue_pending: any[] }>"
|
|
23685
|
+
},
|
|
23686
|
+
"getHistory": {
|
|
23687
|
+
"description": "Get execution history, optionally for a specific prompt.",
|
|
23670
23688
|
"parameters": {
|
|
23671
|
-
"
|
|
23672
|
-
"type": "
|
|
23673
|
-
"description": "
|
|
23674
|
-
},
|
|
23675
|
-
"options": {
|
|
23676
|
-
"type": "Partial<OpenAI.Responses.ResponseCreateParamsNonStreaming>",
|
|
23677
|
-
"description": "Additional parameters for the response"
|
|
23689
|
+
"promptId": {
|
|
23690
|
+
"type": "string",
|
|
23691
|
+
"description": "If provided, returns history for this prompt only"
|
|
23678
23692
|
}
|
|
23679
23693
|
},
|
|
23680
|
-
"required": [
|
|
23681
|
-
|
|
23682
|
-
],
|
|
23683
|
-
"returns": "Promise<OpenAI.Responses.Response>",
|
|
23684
|
-
"examples": [
|
|
23685
|
-
{
|
|
23686
|
-
"language": "ts",
|
|
23687
|
-
"code": "const response = await openai.createResponse('Explain quantum computing')"
|
|
23688
|
-
}
|
|
23689
|
-
]
|
|
23694
|
+
"required": [],
|
|
23695
|
+
"returns": "Promise<Record<string, any>>"
|
|
23690
23696
|
},
|
|
23691
|
-
"
|
|
23692
|
-
"description": "
|
|
23693
|
-
"parameters": {
|
|
23694
|
-
|
|
23695
|
-
|
|
23696
|
-
|
|
23697
|
-
|
|
23698
|
-
|
|
23699
|
-
|
|
23700
|
-
|
|
23697
|
+
"getSystemStats": {
|
|
23698
|
+
"description": "Get system stats including GPU memory and queue info.",
|
|
23699
|
+
"parameters": {},
|
|
23700
|
+
"required": [],
|
|
23701
|
+
"returns": "Promise<any>"
|
|
23702
|
+
},
|
|
23703
|
+
"getObjectInfo": {
|
|
23704
|
+
"description": "Get node type info with input/output schemas.",
|
|
23705
|
+
"parameters": {
|
|
23706
|
+
"nodeClass": {
|
|
23707
|
+
"type": "string",
|
|
23708
|
+
"description": "If provided, returns info for this node type only"
|
|
23701
23709
|
}
|
|
23702
23710
|
},
|
|
23703
|
-
"required": [
|
|
23704
|
-
|
|
23705
|
-
],
|
|
23706
|
-
"returns": "Promise<AsyncIterable<OpenAI.Responses.ResponseStreamEvent>>",
|
|
23707
|
-
"examples": [
|
|
23708
|
-
{
|
|
23709
|
-
"language": "ts",
|
|
23710
|
-
"code": "const stream = await openai.streamResponse('Write a poem')\nfor await (const event of stream) {\n if (event.type === 'response.output_text.delta') {\n process.stdout.write(event.delta)\n }\n}"
|
|
23711
|
-
}
|
|
23712
|
-
]
|
|
23711
|
+
"required": [],
|
|
23712
|
+
"returns": "Promise<any>"
|
|
23713
23713
|
},
|
|
23714
|
-
"
|
|
23715
|
-
"description": "
|
|
23714
|
+
"interrupt": {
|
|
23715
|
+
"description": "Interrupt the currently executing prompt.",
|
|
23716
|
+
"parameters": {},
|
|
23717
|
+
"required": [],
|
|
23718
|
+
"returns": "Promise<void>"
|
|
23719
|
+
},
|
|
23720
|
+
"getModels": {
|
|
23721
|
+
"description": "List available models, optionally filtered by type.",
|
|
23716
23722
|
"parameters": {
|
|
23717
|
-
"
|
|
23723
|
+
"type": {
|
|
23718
23724
|
"type": "string",
|
|
23719
|
-
"description": "
|
|
23720
|
-
},
|
|
23721
|
-
"options": {
|
|
23722
|
-
"type": "Partial<OpenAI.Completions.CompletionCreateParams>",
|
|
23723
|
-
"description": "Additional parameters for the completion"
|
|
23725
|
+
"description": "Model type filter (e.g., 'checkpoints', 'loras')"
|
|
23724
23726
|
}
|
|
23725
23727
|
},
|
|
23726
|
-
"required": [
|
|
23727
|
-
|
|
23728
|
-
],
|
|
23729
|
-
"returns": "Promise<OpenAI.Completions.Completion>",
|
|
23730
|
-
"examples": [
|
|
23731
|
-
{
|
|
23732
|
-
"language": "ts",
|
|
23733
|
-
"code": "const response = await openai.createCompletion('Once upon a time')"
|
|
23734
|
-
}
|
|
23735
|
-
]
|
|
23728
|
+
"required": [],
|
|
23729
|
+
"returns": "Promise<string[]>"
|
|
23736
23730
|
},
|
|
23737
|
-
"
|
|
23738
|
-
"description": "
|
|
23731
|
+
"getEmbeddings": {
|
|
23732
|
+
"description": "List available embedding models.",
|
|
23733
|
+
"parameters": {},
|
|
23734
|
+
"required": [],
|
|
23735
|
+
"returns": "Promise<string[]>"
|
|
23736
|
+
},
|
|
23737
|
+
"uploadImage": {
|
|
23738
|
+
"description": "Upload an image to ComfyUI's input directory.",
|
|
23739
23739
|
"parameters": {
|
|
23740
|
-
"
|
|
23741
|
-
"type": "
|
|
23742
|
-
"description": "
|
|
23740
|
+
"file": {
|
|
23741
|
+
"type": "Buffer | Blob",
|
|
23742
|
+
"description": "The image data as Buffer or Blob"
|
|
23743
23743
|
},
|
|
23744
|
-
"
|
|
23745
|
-
"type": "
|
|
23746
|
-
"description": "
|
|
23744
|
+
"filename": {
|
|
23745
|
+
"type": "string",
|
|
23746
|
+
"description": "File name for the upload"
|
|
23747
|
+
},
|
|
23748
|
+
"opts": {
|
|
23749
|
+
"type": "{ subfolder?: string; type?: string; overwrite?: boolean }",
|
|
23750
|
+
"description": "Upload options (subfolder, type, overwrite)"
|
|
23747
23751
|
}
|
|
23748
23752
|
},
|
|
23749
23753
|
"required": [
|
|
23750
|
-
"
|
|
23754
|
+
"file",
|
|
23755
|
+
"filename"
|
|
23751
23756
|
],
|
|
23752
|
-
"returns": "Promise<
|
|
23753
|
-
"examples": [
|
|
23754
|
-
{
|
|
23755
|
-
"language": "ts",
|
|
23756
|
-
"code": "const response = await openai.createEmbedding('Hello world')\nconsole.log(response.data[0].embedding.length)"
|
|
23757
|
-
}
|
|
23758
|
-
]
|
|
23757
|
+
"returns": "Promise<any>"
|
|
23759
23758
|
},
|
|
23760
|
-
"
|
|
23761
|
-
"description": "
|
|
23759
|
+
"viewImage": {
|
|
23760
|
+
"description": "Download a generated image from ComfyUI as a Buffer.",
|
|
23762
23761
|
"parameters": {
|
|
23763
|
-
"
|
|
23762
|
+
"filename": {
|
|
23764
23763
|
"type": "string",
|
|
23765
|
-
"description": "
|
|
23764
|
+
"description": "The image filename"
|
|
23766
23765
|
},
|
|
23767
|
-
"
|
|
23768
|
-
"type": "
|
|
23769
|
-
"description": "
|
|
23766
|
+
"subfolder": {
|
|
23767
|
+
"type": "any",
|
|
23768
|
+
"description": "Subfolder within the output directory"
|
|
23769
|
+
},
|
|
23770
|
+
"type": {
|
|
23771
|
+
"type": "any",
|
|
23772
|
+
"description": "Image type ('output', 'input', 'temp')"
|
|
23770
23773
|
}
|
|
23771
23774
|
},
|
|
23772
23775
|
"required": [
|
|
23773
|
-
"
|
|
23776
|
+
"filename"
|
|
23774
23777
|
],
|
|
23775
|
-
"returns": "Promise<
|
|
23776
|
-
"examples": [
|
|
23777
|
-
{
|
|
23778
|
-
"language": "ts",
|
|
23779
|
-
"code": "const response = await openai.createImage('A sunset over mountains')\nconsole.log(response.data[0].url)"
|
|
23780
|
-
}
|
|
23781
|
-
]
|
|
23778
|
+
"returns": "Promise<Buffer>"
|
|
23782
23779
|
},
|
|
23783
|
-
"
|
|
23784
|
-
"description": "
|
|
23780
|
+
"connectWs": {
|
|
23781
|
+
"description": "Open a WebSocket connection to ComfyUI for real-time execution tracking. Events emitted: `execution_start`, `executing`, `progress`, `executed`, `execution_cached`, `execution_error`, `execution_complete`.",
|
|
23785
23782
|
"parameters": {},
|
|
23786
23783
|
"required": [],
|
|
23787
|
-
"returns": "Promise<
|
|
23788
|
-
"examples": [
|
|
23789
|
-
{
|
|
23790
|
-
"language": "ts",
|
|
23791
|
-
"code": "const models = await openai.listModels()"
|
|
23792
|
-
}
|
|
23793
|
-
]
|
|
23784
|
+
"returns": "Promise<void>"
|
|
23794
23785
|
},
|
|
23795
|
-
"
|
|
23796
|
-
"description": "
|
|
23786
|
+
"disconnectWs": {
|
|
23787
|
+
"description": "Close the WebSocket connection.",
|
|
23788
|
+
"parameters": {},
|
|
23789
|
+
"required": [],
|
|
23790
|
+
"returns": "void"
|
|
23791
|
+
},
|
|
23792
|
+
"toApiFormat": {
|
|
23793
|
+
"description": "Convert a UI-format workflow to the API format that /prompt expects. Requires a running ComfyUI instance to fetch `object_info` so we can map positional `widgets_values` to their named input fields. If the workflow is already in API format, it's returned as-is.",
|
|
23797
23794
|
"parameters": {
|
|
23798
|
-
"
|
|
23799
|
-
"type": "string",
|
|
23800
|
-
"description": "
|
|
23801
|
-
},
|
|
23802
|
-
"options": {
|
|
23803
|
-
"type": "Partial<OpenAI.Chat.Completions.ChatCompletionCreateParams>",
|
|
23804
|
-
"description": "Additional completion parameters"
|
|
23795
|
+
"workflow": {
|
|
23796
|
+
"type": "Record<string, any>",
|
|
23797
|
+
"description": "Parameter workflow"
|
|
23805
23798
|
}
|
|
23806
23799
|
},
|
|
23807
23800
|
"required": [
|
|
23808
|
-
"
|
|
23801
|
+
"workflow"
|
|
23809
23802
|
],
|
|
23810
|
-
"returns": "Promise<string
|
|
23811
|
-
"examples": [
|
|
23812
|
-
{
|
|
23813
|
-
"language": "ts",
|
|
23814
|
-
"code": "const answer = await openai.ask('What is 2 + 2?')\nconsole.log(answer) // '4'"
|
|
23815
|
-
}
|
|
23816
|
-
]
|
|
23803
|
+
"returns": "Promise<Record<string, any>>"
|
|
23817
23804
|
},
|
|
23818
|
-
"
|
|
23819
|
-
"description": "
|
|
23805
|
+
"runWorkflow": {
|
|
23806
|
+
"description": "Run a ComfyUI workflow with optional runtime input overrides. Inputs can be provided in two forms: **Direct node mapping** (when no `inputMap` in options): ``` { '3': { seed: 42 }, '6': { text: 'a cat' } } ``` **Named inputs** (when `inputMap` is provided in options): ``` inputs: { positive_prompt: 'a cat', seed: 42 } options.inputMap: { positive_prompt: { nodeId: '6', field: 'text' }, seed: { nodeId: '3', field: 'seed' } } ```",
|
|
23820
23807
|
"parameters": {
|
|
23821
|
-
"
|
|
23822
|
-
"type": "
|
|
23823
|
-
"description": "
|
|
23808
|
+
"workflow": {
|
|
23809
|
+
"type": "Record<string, any>",
|
|
23810
|
+
"description": "Parameter workflow"
|
|
23811
|
+
},
|
|
23812
|
+
"inputs": {
|
|
23813
|
+
"type": "Record<string, any>",
|
|
23814
|
+
"description": "Parameter inputs"
|
|
23824
23815
|
},
|
|
23825
23816
|
"options": {
|
|
23826
|
-
"type": "
|
|
23827
|
-
"description": "
|
|
23817
|
+
"type": "WorkflowRunOptions",
|
|
23818
|
+
"description": "Parameter options",
|
|
23819
|
+
"properties": {
|
|
23820
|
+
"poll": {
|
|
23821
|
+
"type": "boolean",
|
|
23822
|
+
"description": "Use polling instead of WebSocket for tracking execution"
|
|
23823
|
+
},
|
|
23824
|
+
"pollInterval": {
|
|
23825
|
+
"type": "number",
|
|
23826
|
+
"description": "Polling interval in ms (default 1000)"
|
|
23827
|
+
},
|
|
23828
|
+
"inputMap": {
|
|
23829
|
+
"type": "InputMapping",
|
|
23830
|
+
"description": "Named input mapping: semantic name -> { nodeId, field }"
|
|
23831
|
+
},
|
|
23832
|
+
"outputDir": {
|
|
23833
|
+
"type": "string",
|
|
23834
|
+
"description": "If provided, output images are downloaded to this directory"
|
|
23835
|
+
}
|
|
23836
|
+
}
|
|
23828
23837
|
}
|
|
23829
23838
|
},
|
|
23830
23839
|
"required": [
|
|
23831
|
-
"
|
|
23840
|
+
"workflow"
|
|
23832
23841
|
],
|
|
23833
|
-
"returns": "Promise<
|
|
23834
|
-
"examples": [
|
|
23835
|
-
{
|
|
23836
|
-
"language": "ts",
|
|
23837
|
-
"code": "const reply = await openai.chat([\n { role: 'system', content: 'You are a pirate.' },\n { role: 'user', content: 'Hello!' }\n])"
|
|
23838
|
-
}
|
|
23839
|
-
]
|
|
23842
|
+
"returns": "Promise<WorkflowResult>"
|
|
23840
23843
|
}
|
|
23841
23844
|
},
|
|
23842
23845
|
"getters": {
|
|
23843
|
-
"
|
|
23844
|
-
"description": "The
|
|
23846
|
+
"clientId": {
|
|
23847
|
+
"description": "The unique client ID used for WebSocket session tracking.",
|
|
23845
23848
|
"returns": "string"
|
|
23846
23849
|
},
|
|
23847
|
-
"
|
|
23848
|
-
"description": "The
|
|
23849
|
-
"returns": "
|
|
23850
|
+
"wsURL": {
|
|
23851
|
+
"description": "The WebSocket URL derived from baseURL or overridden via options.",
|
|
23852
|
+
"returns": "string"
|
|
23850
23853
|
}
|
|
23851
23854
|
},
|
|
23852
23855
|
"events": {
|
|
23853
|
-
"
|
|
23854
|
-
"name": "
|
|
23855
|
-
"description": "Event emitted by
|
|
23856
|
+
"execution_start": {
|
|
23857
|
+
"name": "execution_start",
|
|
23858
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23856
23859
|
"arguments": {}
|
|
23857
23860
|
},
|
|
23858
|
-
"
|
|
23859
|
-
"name": "
|
|
23860
|
-
"description": "Event emitted by
|
|
23861
|
+
"execution_complete": {
|
|
23862
|
+
"name": "execution_complete",
|
|
23863
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23861
23864
|
"arguments": {}
|
|
23862
23865
|
},
|
|
23863
|
-
"
|
|
23864
|
-
"name": "
|
|
23865
|
-
"description": "Event emitted by
|
|
23866
|
+
"executing": {
|
|
23867
|
+
"name": "executing",
|
|
23868
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23866
23869
|
"arguments": {}
|
|
23867
23870
|
},
|
|
23868
|
-
"
|
|
23869
|
-
"name": "
|
|
23870
|
-
"description": "Event emitted by
|
|
23871
|
+
"progress": {
|
|
23872
|
+
"name": "progress",
|
|
23873
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23871
23874
|
"arguments": {}
|
|
23872
23875
|
},
|
|
23873
|
-
"
|
|
23874
|
-
"name": "
|
|
23875
|
-
"description": "Event emitted by
|
|
23876
|
+
"executed": {
|
|
23877
|
+
"name": "executed",
|
|
23878
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23876
23879
|
"arguments": {}
|
|
23877
23880
|
},
|
|
23878
|
-
"
|
|
23879
|
-
"name": "
|
|
23880
|
-
"description": "Event emitted by
|
|
23881
|
+
"execution_cached": {
|
|
23882
|
+
"name": "execution_cached",
|
|
23883
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23884
|
+
"arguments": {}
|
|
23885
|
+
},
|
|
23886
|
+
"execution_error": {
|
|
23887
|
+
"name": "execution_error",
|
|
23888
|
+
"description": "Event emitted by ComfyUIClient",
|
|
23881
23889
|
"arguments": {}
|
|
23882
23890
|
}
|
|
23883
23891
|
},
|
|
@@ -23887,7 +23895,7 @@ export const introspectionData = [
|
|
|
23887
23895
|
"examples": [
|
|
23888
23896
|
{
|
|
23889
23897
|
"language": "ts",
|
|
23890
|
-
"code": "const
|
|
23898
|
+
"code": "const comfy = container.client('comfyui', { baseURL: 'http://localhost:8188' })\nconst result = await comfy.runWorkflow(workflow, {\n '6': { text: 'a beautiful sunset' }\n})\nconsole.log(result.images)"
|
|
23891
23899
|
}
|
|
23892
23900
|
]
|
|
23893
23901
|
},
|
|
@@ -25426,6 +25434,10 @@ export const introspectionData = [
|
|
|
25426
25434
|
"settingsFile": {
|
|
25427
25435
|
"type": "string",
|
|
25428
25436
|
"description": "Path to a custom settings file."
|
|
25437
|
+
},
|
|
25438
|
+
"chrome": {
|
|
25439
|
+
"type": "boolean",
|
|
25440
|
+
"description": "Launch Claude Code with a Chrome browser tool."
|
|
25429
25441
|
}
|
|
25430
25442
|
}
|
|
25431
25443
|
}
|
|
@@ -25575,6 +25587,10 @@ export const introspectionData = [
|
|
|
25575
25587
|
"settingsFile": {
|
|
25576
25588
|
"type": "string",
|
|
25577
25589
|
"description": "Path to a custom settings file."
|
|
25590
|
+
},
|
|
25591
|
+
"chrome": {
|
|
25592
|
+
"type": "boolean",
|
|
25593
|
+
"description": "Launch Claude Code with a Chrome browser tool."
|
|
25578
25594
|
}
|
|
25579
25595
|
}
|
|
25580
25596
|
}
|