@decocms/bindings 0.2.4 → 1.0.0-test.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (54) hide show
  1. package/README.md +3 -3
  2. package/package.json +9 -35
  3. package/src/core/binder.ts +241 -0
  4. package/src/core/client/README.md +3 -0
  5. package/{dist/core/client/http-client-transport.js → src/core/client/http-client-transport.ts} +24 -12
  6. package/src/core/client/index.ts +1 -0
  7. package/src/core/client/mcp-client.ts +149 -0
  8. package/src/core/client/mcp.ts +93 -0
  9. package/src/core/client/proxy.ts +151 -0
  10. package/src/core/connection.ts +38 -0
  11. package/src/core/subset.ts +514 -0
  12. package/src/index.ts +15 -0
  13. package/src/well-known/agent.ts +60 -0
  14. package/src/well-known/collections.ts +416 -0
  15. package/src/well-known/language-model.ts +383 -0
  16. package/test/index.test.ts +942 -0
  17. package/tsconfig.json +11 -0
  18. package/vitest.config.ts +8 -0
  19. package/dist/core/binder.d.ts +0 -3
  20. package/dist/core/binder.js +0 -77
  21. package/dist/core/binder.js.map +0 -1
  22. package/dist/core/client/http-client-transport.d.ts +0 -12
  23. package/dist/core/client/http-client-transport.js.map +0 -1
  24. package/dist/core/client/index.d.ts +0 -3
  25. package/dist/core/client/index.js +0 -5
  26. package/dist/core/client/index.js.map +0 -1
  27. package/dist/core/client/mcp-client.d.ts +0 -233
  28. package/dist/core/client/mcp-client.js +0 -99
  29. package/dist/core/client/mcp-client.js.map +0 -1
  30. package/dist/core/client/mcp.d.ts +0 -3
  31. package/dist/core/client/mcp.js +0 -29
  32. package/dist/core/client/mcp.js.map +0 -1
  33. package/dist/core/client/proxy.d.ts +0 -10
  34. package/dist/core/client/proxy.js +0 -104
  35. package/dist/core/client/proxy.js.map +0 -1
  36. package/dist/core/connection.d.ts +0 -30
  37. package/dist/core/connection.js +0 -1
  38. package/dist/core/connection.js.map +0 -1
  39. package/dist/core/subset.d.ts +0 -17
  40. package/dist/core/subset.js +0 -319
  41. package/dist/core/subset.js.map +0 -1
  42. package/dist/index-D0aUdNls.d.ts +0 -153
  43. package/dist/index.d.ts +0 -3
  44. package/dist/index.js +0 -7
  45. package/dist/index.js.map +0 -1
  46. package/dist/well-known/agent.d.ts +0 -903
  47. package/dist/well-known/agent.js +0 -27
  48. package/dist/well-known/agent.js.map +0 -1
  49. package/dist/well-known/collections.d.ts +0 -537
  50. package/dist/well-known/collections.js +0 -134
  51. package/dist/well-known/collections.js.map +0 -1
  52. package/dist/well-known/language-model.d.ts +0 -2836
  53. package/dist/well-known/language-model.js +0 -209
  54. package/dist/well-known/language-model.js.map +0 -1
@@ -1,209 +0,0 @@
1
- import { z } from "zod/v3";
2
- import { bindingClient } from "../core/binder";
3
- import {
4
- BaseCollectionEntitySchema,
5
- createCollectionBindings
6
- } from "./collections";
7
- const LanguageModelCallOptionsSchema = z.object({
8
- // Core parameters
9
- prompt: z.any().describe(
10
- "A language mode prompt is a standardized prompt type (messages, system, etc.)"
11
- ),
12
- // Generation parameters
13
- maxOutputTokens: z.number().optional().describe("Maximum number of tokens to generate"),
14
- temperature: z.number().optional().describe(
15
- "Temperature setting. The range depends on the provider and model"
16
- ),
17
- topP: z.number().optional().describe("Nucleus sampling parameter"),
18
- topK: z.number().optional().describe(
19
- "Only sample from the top K options for each subsequent token. Used to remove long tail low probability responses"
20
- ),
21
- presencePenalty: z.number().optional().describe(
22
- "Presence penalty setting. It affects the likelihood of the model to repeat information that is already in the prompt"
23
- ),
24
- frequencyPenalty: z.number().optional().describe(
25
- "Frequency penalty setting. It affects the likelihood of the model to repeatedly use the same words or phrases"
26
- ),
27
- seed: z.number().optional().describe(
28
- "The seed (integer) to use for random sampling. If set and supported by the model, calls will generate deterministic results"
29
- ),
30
- // Stop sequences
31
- stopSequences: z.array(z.string()).optional().describe(
32
- "Stop sequences. If set, the model will stop generating text when one of the stop sequences is generated"
33
- ),
34
- // Response format
35
- responseFormat: z.union([
36
- z.object({ type: z.literal("text") }),
37
- z.object({
38
- type: z.literal("json"),
39
- schema: z.any().optional().describe("JSON schema that the generated output should conform to"),
40
- name: z.string().optional().describe("Name of output that should be generated"),
41
- description: z.string().optional().describe("Description of the output that should be generated")
42
- })
43
- ]).optional().describe(
44
- "Response format. The output can either be text or JSON. Default is text"
45
- ),
46
- // Tools
47
- tools: z.array(z.any()).optional().describe("The tools that are available for the model"),
48
- toolChoice: z.any().optional().describe("Specifies how the tool should be selected. Defaults to 'auto'"),
49
- // Stream options
50
- includeRawChunks: z.boolean().optional().describe(
51
- "Include raw chunks in the stream. Only applicable for streaming calls"
52
- ),
53
- // Abort signal
54
- abortSignal: z.any().optional().describe("Abort signal for cancelling the operation"),
55
- // Additional options
56
- headers: z.record(z.string(), z.union([z.string(), z.undefined()])).optional().describe("Additional HTTP headers to be sent with the request"),
57
- providerOptions: z.any().optional().describe("Additional provider-specific options")
58
- });
59
- const LanguageModelGenerateOutputSchema = z.object({
60
- // Ordered content that the model has generated
61
- content: z.array(z.any()).describe(
62
- "Ordered content that the model has generated (text, tool-calls, reasoning, files, sources)"
63
- ),
64
- // Finish reason (required)
65
- finishReason: z.enum([
66
- "stop",
67
- "length",
68
- "content-filter",
69
- "tool-calls",
70
- "error",
71
- "other",
72
- "unknown"
73
- ]).describe("Reason why generation stopped"),
74
- // Usage information (required)
75
- usage: z.object({
76
- inputTokens: z.number().optional(),
77
- outputTokens: z.number().optional(),
78
- totalTokens: z.number().optional(),
79
- reasoningTokens: z.number().optional()
80
- }).passthrough().transform((val) => ({
81
- inputTokens: val.inputTokens,
82
- outputTokens: val.outputTokens,
83
- totalTokens: val.totalTokens,
84
- reasoningTokens: val.reasoningTokens,
85
- ...val
86
- })).describe("Usage information for the language model call"),
87
- // Provider metadata
88
- providerMetadata: z.any().optional().describe("Additional provider-specific metadata"),
89
- // Request information for telemetry and debugging
90
- request: z.object({
91
- body: z.any().optional().describe("Request HTTP body sent to the provider API")
92
- }).optional().describe("Optional request information for telemetry and debugging"),
93
- // Response information for telemetry and debugging
94
- response: z.object({
95
- id: z.string().optional().describe("ID for the generated response"),
96
- timestamp: z.date().optional().describe("Timestamp for the start of the generated response"),
97
- modelId: z.string().optional().describe("The ID of the response model that was used"),
98
- headers: z.record(z.string(), z.string()).optional().describe("Response headers"),
99
- body: z.any().optional().describe("Response HTTP body")
100
- }).optional().describe("Optional response information for telemetry and debugging"),
101
- // Warnings for the call (required)
102
- warnings: z.array(z.any()).describe("Warnings for the call, e.g. unsupported settings")
103
- });
104
- const LanguageModelStreamOutputSchema = z.object({
105
- // Stream of language model output parts
106
- stream: z.any().describe("ReadableStream of LanguageModelV2StreamPart"),
107
- // Request information for telemetry and debugging
108
- request: z.object({
109
- body: z.any().optional().describe("Request HTTP body sent to the provider API")
110
- }).optional().describe("Optional request information for telemetry and debugging"),
111
- // Response information
112
- response: z.object({
113
- headers: z.record(z.string(), z.string()).optional().describe("Response headers")
114
- }).optional().describe("Optional response data")
115
- });
116
- const LanguageModelMetadataSchema = z.object({
117
- supportedUrls: z.record(z.string(), z.array(z.string())).describe("Supported URL patterns by media type for the provider")
118
- });
119
- const ModelSchema = z.object({
120
- modelId: z.string().describe("The ID of the model"),
121
- // Model-specific fields
122
- logo: z.string().nullable(),
123
- description: z.string().nullable(),
124
- capabilities: z.array(z.string()),
125
- limits: z.object({
126
- contextWindow: z.number(),
127
- maxOutputTokens: z.number()
128
- }).nullable(),
129
- costs: z.object({
130
- input: z.number(),
131
- output: z.number()
132
- }).nullable(),
133
- // Provider information
134
- provider: z.enum([
135
- "openai",
136
- "anthropic",
137
- "google",
138
- "x-ai",
139
- "deepseek",
140
- "openai-compatible",
141
- "openrouter"
142
- ]).nullable()
143
- });
144
- const LanguageModelInputSchema = z.object({
145
- modelId: z.string().describe("The ID of the model"),
146
- callOptions: LanguageModelCallOptionsSchema
147
- });
148
- const ModelCollectionEntitySchema = BaseCollectionEntitySchema.extend({
149
- // Model-specific fields
150
- logo: z.string().nullable(),
151
- description: z.string().nullable(),
152
- capabilities: z.array(z.string()),
153
- limits: z.object({
154
- contextWindow: z.number(),
155
- maxOutputTokens: z.number()
156
- }).nullable(),
157
- costs: z.object({
158
- input: z.number(),
159
- output: z.number()
160
- }).nullable(),
161
- // Provider information
162
- provider: z.enum([
163
- "openai",
164
- "anthropic",
165
- "google",
166
- "xai",
167
- "deepseek",
168
- "openai-compatible",
169
- "openrouter"
170
- ]).nullable()
171
- });
172
- const LLM_COLLECTION_BINDING = createCollectionBindings(
173
- "llm",
174
- ModelCollectionEntitySchema,
175
- { readOnly: true }
176
- );
177
- const LANGUAGE_MODEL_BINDING = [
178
- {
179
- name: "LLM_METADATA",
180
- inputSchema: z.object({
181
- modelId: z.string().describe("The ID of the model")
182
- }),
183
- outputSchema: LanguageModelMetadataSchema
184
- },
185
- {
186
- name: "LLM_DO_STREAM",
187
- inputSchema: LanguageModelInputSchema,
188
- streamable: true
189
- },
190
- {
191
- name: "LLM_DO_GENERATE",
192
- inputSchema: LanguageModelInputSchema,
193
- outputSchema: LanguageModelGenerateOutputSchema
194
- },
195
- ...LLM_COLLECTION_BINDING
196
- ];
197
- const LanguageModelBinding = bindingClient(LANGUAGE_MODEL_BINDING);
198
- export {
199
- LANGUAGE_MODEL_BINDING,
200
- LanguageModelBinding,
201
- LanguageModelCallOptionsSchema,
202
- LanguageModelGenerateOutputSchema,
203
- LanguageModelInputSchema,
204
- LanguageModelMetadataSchema,
205
- LanguageModelStreamOutputSchema,
206
- ModelCollectionEntitySchema,
207
- ModelSchema
208
- };
209
- //# sourceMappingURL=language-model.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../src/well-known/language-model.ts"],"sourcesContent":["/**\n * Language Model Well-Known Binding\n *\n * Defines the interface for AI model providers.\n * Any MCP that implements this binding can provide AI models and streaming endpoints.\n *\n * This binding includes:\n * - LLM operations (metadata, stream, generate)\n * - Collection bindings for LIST and GET operations (read-only)\n * - Streaming endpoint information is included directly in the model entity schema.\n */\n\nimport { z } from \"zod/v3\";\nimport { bindingClient, type ToolBinder } from \"../core/binder\";\nimport {\n BaseCollectionEntitySchema,\n createCollectionBindings,\n} from \"./collections\";\n\n/**\n * Language Model Call Options Schema\n * Based on LanguageModelV2CallOptions from @ai-sdk/provider\n */\nexport const LanguageModelCallOptionsSchema = z.object({\n // Core parameters\n prompt: z\n .any()\n .describe(\n \"A language mode prompt is a standardized prompt type (messages, system, etc.)\",\n ),\n\n // Generation parameters\n maxOutputTokens: z\n .number()\n .optional()\n .describe(\"Maximum number of tokens to generate\"),\n temperature: z\n .number()\n .optional()\n .describe(\n \"Temperature setting. The range depends on the provider and model\",\n ),\n topP: z.number().optional().describe(\"Nucleus sampling parameter\"),\n topK: z\n .number()\n .optional()\n .describe(\n \"Only sample from the top K options for each subsequent token. Used to remove long tail low probability responses\",\n ),\n presencePenalty: z\n .number()\n .optional()\n .describe(\n \"Presence penalty setting. It affects the likelihood of the model to repeat information that is already in the prompt\",\n ),\n frequencyPenalty: z\n .number()\n .optional()\n .describe(\n \"Frequency penalty setting. It affects the likelihood of the model to repeatedly use the same words or phrases\",\n ),\n seed: z\n .number()\n .optional()\n .describe(\n \"The seed (integer) to use for random sampling. If set and supported by the model, calls will generate deterministic results\",\n ),\n\n // Stop sequences\n stopSequences: z\n .array(z.string())\n .optional()\n .describe(\n \"Stop sequences. If set, the model will stop generating text when one of the stop sequences is generated\",\n ),\n\n // Response format\n responseFormat: z\n .union([\n z.object({ type: z.literal(\"text\") }),\n z.object({\n type: z.literal(\"json\"),\n schema: z\n .any()\n .optional()\n .describe(\"JSON schema that the generated output should conform to\"),\n name: z\n .string()\n .optional()\n .describe(\"Name of output that should be generated\"),\n description: z\n .string()\n .optional()\n .describe(\"Description of the output that should be generated\"),\n }),\n ])\n .optional()\n .describe(\n \"Response format. The output can either be text or JSON. Default is text\",\n ),\n\n // Tools\n tools: z\n .array(z.any())\n .optional()\n .describe(\"The tools that are available for the model\"),\n toolChoice: z\n .any()\n .optional()\n .describe(\"Specifies how the tool should be selected. Defaults to 'auto'\"),\n\n // Stream options\n includeRawChunks: z\n .boolean()\n .optional()\n .describe(\n \"Include raw chunks in the stream. Only applicable for streaming calls\",\n ),\n\n // Abort signal\n abortSignal: z\n .any()\n .optional()\n .describe(\"Abort signal for cancelling the operation\"),\n\n // Additional options\n headers: z\n .record(z.string(), z.union([z.string(), z.undefined()]))\n .optional()\n .describe(\"Additional HTTP headers to be sent with the request\"),\n providerOptions: z\n .any()\n .optional()\n .describe(\"Additional provider-specific options\"),\n});\n\n/**\n * Language Model Generate Output Schema\n * Based on the return type of LanguageModelV2.doGenerate from @ai-sdk/provider\n */\nexport const LanguageModelGenerateOutputSchema = z.object({\n // Ordered content that the model has generated\n content: z\n .array(z.any())\n .describe(\n \"Ordered content that the model has generated (text, tool-calls, reasoning, files, sources)\",\n ),\n\n // Finish reason (required)\n finishReason: z\n .enum([\n \"stop\",\n \"length\",\n \"content-filter\",\n \"tool-calls\",\n \"error\",\n \"other\",\n \"unknown\",\n ])\n .describe(\"Reason why generation stopped\"),\n\n // Usage information (required)\n usage: z\n .object({\n inputTokens: z.number().optional(),\n outputTokens: z.number().optional(),\n totalTokens: z.number().optional(),\n reasoningTokens: z.number().optional(),\n })\n .passthrough()\n .transform((val) => ({\n inputTokens: val.inputTokens,\n outputTokens: val.outputTokens,\n totalTokens: val.totalTokens,\n reasoningTokens: val.reasoningTokens,\n ...val,\n }))\n .describe(\"Usage information for the language model call\"),\n\n // Provider metadata\n providerMetadata: z\n .any()\n .optional()\n .describe(\"Additional provider-specific metadata\"),\n\n // Request information for telemetry and debugging\n request: z\n .object({\n body: z\n .any()\n .optional()\n .describe(\"Request HTTP body sent to the provider API\"),\n })\n .optional()\n .describe(\"Optional request information for telemetry and debugging\"),\n\n // Response information for telemetry and debugging\n response: z\n .object({\n id: z.string().optional().describe(\"ID for the generated response\"),\n timestamp: z\n .date()\n .optional()\n .describe(\"Timestamp for the start of the generated response\"),\n modelId: z\n .string()\n .optional()\n .describe(\"The ID of the response model that was used\"),\n headers: z\n .record(z.string(), z.string())\n .optional()\n .describe(\"Response headers\"),\n body: z.any().optional().describe(\"Response HTTP body\"),\n })\n .optional()\n .describe(\"Optional response information for telemetry and debugging\"),\n\n // Warnings for the call (required)\n warnings: z\n .array(z.any())\n .describe(\"Warnings for the call, e.g. unsupported settings\"),\n});\n\n/**\n * Language Model Stream Output Schema\n * Based on the return type of LanguageModelV2.doStream from @ai-sdk/provider\n */\nexport const LanguageModelStreamOutputSchema = z.object({\n // Stream of language model output parts\n stream: z.any().describe(\"ReadableStream of LanguageModelV2StreamPart\"),\n\n // Request information for telemetry and debugging\n request: z\n .object({\n body: z\n .any()\n .optional()\n .describe(\"Request HTTP body sent to the provider API\"),\n })\n .optional()\n .describe(\"Optional request information for telemetry and debugging\"),\n\n // Response information\n response: z\n .object({\n headers: z\n .record(z.string(), z.string())\n .optional()\n .describe(\"Response headers\"),\n })\n .optional()\n .describe(\"Optional response data\"),\n});\n\nexport const LanguageModelMetadataSchema = z.object({\n supportedUrls: z\n .record(z.string(), z.array(z.string()))\n .describe(\"Supported URL patterns by media type for the provider\"),\n});\n\n/**\n * Simple Model schema for LLM operations\n */\nexport const ModelSchema = z.object({\n modelId: z.string().describe(\"The ID of the model\"),\n // Model-specific fields\n logo: z.string().nullable(),\n description: z.string().nullable(),\n capabilities: z.array(z.string()),\n limits: z\n .object({\n contextWindow: z.number(),\n maxOutputTokens: z.number(),\n })\n .nullable(),\n costs: z\n .object({\n input: z.number(),\n output: z.number(),\n })\n .nullable(),\n // Provider information\n provider: z\n .enum([\n \"openai\",\n \"anthropic\",\n \"google\",\n \"x-ai\",\n \"deepseek\",\n \"openai-compatible\",\n \"openrouter\",\n ])\n .nullable(),\n});\n\nexport const LanguageModelInputSchema = z.object({\n modelId: z.string().describe(\"The ID of the model\"),\n callOptions: LanguageModelCallOptionsSchema,\n});\n\n/**\n * Model entity schema for AI models (Collection Entity)\n * Extends BaseCollectionEntitySchema with model-specific fields\n * Base schema already includes: id, title, created_at, updated_at, created_by, updated_by\n */\nexport const ModelCollectionEntitySchema = BaseCollectionEntitySchema.extend({\n // Model-specific fields\n logo: z.string().nullable(),\n description: z.string().nullable(),\n capabilities: z.array(z.string()),\n limits: z\n .object({\n contextWindow: z.number(),\n maxOutputTokens: z.number(),\n })\n .nullable(),\n costs: z\n .object({\n input: z.number(),\n output: z.number(),\n })\n .nullable(),\n // Provider information\n provider: z\n .enum([\n \"openai\",\n \"anthropic\",\n \"google\",\n \"xai\",\n \"deepseek\",\n \"openai-compatible\",\n \"openrouter\",\n ])\n .nullable(),\n});\n\n/**\n * LLM Collection Binding (internal)\n *\n * Collection bindings for language models (read-only).\n * Provides LIST and GET operations for AI models.\n */\nconst LLM_COLLECTION_BINDING = createCollectionBindings(\n \"llm\",\n ModelCollectionEntitySchema,\n { readOnly: true },\n);\n\n/**\n * Language Model Binding\n *\n * Defines the interface for AI model providers.\n * Any MCP that implements this binding can provide AI models.\n *\n * Required tools:\n * - LLM_METADATA: Get metadata for a specific model\n * - LLM_DO_STREAM: Stream a language model response\n * - LLM_DO_GENERATE: Generate a language model response\n * - COLLECTION_LLM_LIST: List available AI models with their capabilities\n * - COLLECTION_LLM_GET: Get a single model by ID\n */\nexport const LANGUAGE_MODEL_BINDING = [\n {\n name: \"LLM_METADATA\" as const,\n inputSchema: z.object({\n modelId: z.string().describe(\"The ID of the model\"),\n }),\n outputSchema: LanguageModelMetadataSchema,\n },\n {\n name: \"LLM_DO_STREAM\" as const,\n inputSchema: LanguageModelInputSchema,\n streamable: true,\n },\n {\n name: \"LLM_DO_GENERATE\" as const,\n inputSchema: LanguageModelInputSchema,\n outputSchema: LanguageModelGenerateOutputSchema,\n },\n ...LLM_COLLECTION_BINDING,\n] satisfies ToolBinder[];\n\nexport const LanguageModelBinding = bindingClient(LANGUAGE_MODEL_BINDING);\n"],"mappings":"AAYA,SAAS,SAAS;AAClB,SAAS,qBAAsC;AAC/C;AAAA,EACE;AAAA,EACA;AAAA,OACK;AAMA,MAAM,iCAAiC,EAAE,OAAO;AAAA;AAAA,EAErD,QAAQ,EACL,IAAI,EACJ;AAAA,IACC;AAAA,EACF;AAAA;AAAA,EAGF,iBAAiB,EACd,OAAO,EACP,SAAS,EACT,SAAS,sCAAsC;AAAA,EAClD,aAAa,EACV,OAAO,EACP,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA,EACF,MAAM,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,4BAA4B;AAAA,EACjE,MAAM,EACH,OAAO,EACP,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA,EACF,iBAAiB,EACd,OAAO,EACP,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA,EACF,kBAAkB,EACf,OAAO,EACP,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA,EACF,MAAM,EACH,OAAO,EACP,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA;AAAA,EAGF,eAAe,EACZ,MAAM,EAAE,OAAO,CAAC,EAChB,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA;AAAA,EAGF,gBAAgB,EACb,MAAM;AAAA,IACL,EAAE,OAAO,EAAE,MAAM,EAAE,QAAQ,MAAM,EAAE,CAAC;AAAA,IACpC,EAAE,OAAO;AAAA,MACP,MAAM,EAAE,QAAQ,MAAM;AAAA,MACtB,QAAQ,EACL,IAAI,EACJ,SAAS,EACT,SAAS,yDAAyD;AAAA,MACrE,MAAM,EACH,OAAO,EACP,SAAS,EACT,SAAS,yCAAyC;AAAA,MACrD,aAAa,EACV,OAAO,EACP,SAAS,EACT,SAAS,oDAAoD;AAAA,IAClE,CAAC;AAAA,EACH,CAAC,EACA,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA;AAAA,EAGF,OAAO,EACJ,MAAM,EAAE,IAAI,CAAC,EACb,SAAS,EACT,SAAS,4CAA4C;AAAA,EACxD,YAAY,EACT,IAAI,EACJ,SAAS,EACT,SAAS,+DAA+D;AAAA;AAAA,EAG3E,kBAAkB,EACf,QAAQ,EACR,SAAS,EACT;AAAA,IACC;AAAA,EACF;AAAA;AAAA,EAGF,aAAa,EACV,IAAI,EACJ,SAAS,EACT,SAAS,2CAA2C;AAAA;AAAA,EAGvD,SAAS,EACN,OAAO,EAAE,OAAO,GAAG,EAAE,MAAM,CAAC,EAAE,OAAO,GAAG,EAAE,UAAU,CAAC,CAAC,CAAC,EACvD,SAAS,EACT,SAAS,qDAAqD;AAAA,EACjE,iBAAiB,EACd,IAAI,EACJ,SAAS,EACT,SAAS,sCAAsC;AACpD,CAAC;AAMM,MAAM,oCAAoC,EAAE,OAAO;AAAA;AAAA,EAExD,SAAS,EACN,MAAM,EAAE,IAAI,CAAC,EACb;AAAA,IACC;AAAA,EACF;AAAA;AAAA,EAGF,cAAc,EACX,KAAK;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,CAAC,EACA,SAAS,+BAA+B;AAAA;AAAA,EAG3C,OAAO,EACJ,OAAO;AAAA,IACN,aAAa,EAAE,OAAO,EAAE,SAAS;AAAA,IACjC,cAAc,EAAE,OAAO,EAAE,SAAS;AAAA,IAClC,aAAa,EAAE,OAAO,EAAE,SAAS;AAAA,IACjC,iBAAiB,EAAE,OAAO,EAAE,SAAS;AAAA,EACvC,CAAC,EACA,YAAY,EACZ,UAAU,CAAC,SAAS;AAAA,IACnB,aAAa,IAAI;AAAA,IACjB,cAAc,IAAI;AAAA,IAClB,aAAa,IAAI;AAAA,IACjB,iBAAiB,IAAI;AAAA,IACrB,GAAG;AAAA,EACL,EAAE,EACD,SAAS,+CAA+C;AAAA;AAAA,EAG3D,kBAAkB,EACf,IAAI,EACJ,SAAS,EACT,SAAS,uCAAuC;AAAA;AAAA,EAGnD,SAAS,EACN,OAAO;AAAA,IACN,MAAM,EACH,IAAI,EACJ,SAAS,EACT,SAAS,4CAA4C;AAAA,EAC1D,CAAC,EACA,SAAS,EACT,SAAS,0DAA0D;AAAA;AAAA,EAGtE,UAAU,EACP,OAAO;AAAA,IACN,IAAI,EAAE,OAAO,EAAE,SAAS,EAAE,SAAS,+BAA+B;AAAA,IAClE,WAAW,EACR,KAAK,EACL,SAAS,EACT,SAAS,mDAAmD;AAAA,IAC/D,SAAS,EACN,OAAO,EACP,SAAS,EACT,SAAS,4CAA4C;AAAA,IACxD,SAAS,EACN,OAAO,EAAE,OAAO,GAAG,EAAE,OAAO,CAAC,EAC7B,SAAS,EACT,SAAS,kBAAkB;AAAA,IAC9B,MAAM,EAAE,IAAI,EAAE,SAAS,EAAE,SAAS,oBAAoB;AAAA,EACxD,CAAC,EACA,SAAS,EACT,SAAS,2DAA2D;AAAA;AAAA,EAGvE,UAAU,EACP,MAAM,EAAE,IAAI,CAAC,EACb,SAAS,kDAAkD;AAChE,CAAC;AAMM,MAAM,kCAAkC,EAAE,OAAO;AAAA;AAAA,EAEtD,QAAQ,EAAE,IAAI,EAAE,SAAS,6CAA6C;AAAA;AAAA,EAGtE,SAAS,EACN,OAAO;AAAA,IACN,MAAM,EACH,IAAI,EACJ,SAAS,EACT,SAAS,4CAA4C;AAAA,EAC1D,CAAC,EACA,SAAS,EACT,SAAS,0DAA0D;AAAA;AAAA,EAGtE,UAAU,EACP,OAAO;AAAA,IACN,SAAS,EACN,OAAO,EAAE,OAAO,GAAG,EAAE,OAAO,CAAC,EAC7B,SAAS,EACT,SAAS,kBAAkB;AAAA,EAChC,CAAC,EACA,SAAS,EACT,SAAS,wBAAwB;AACtC,CAAC;AAEM,MAAM,8BAA8B,EAAE,OAAO;AAAA,EAClD,eAAe,EACZ,OAAO,EAAE,OAAO,GAAG,EAAE,MAAM,EAAE,OAAO,CAAC,CAAC,EACtC,SAAS,uDAAuD;AACrE,CAAC;AAKM,MAAM,cAAc,EAAE,OAAO;AAAA,EAClC,SAAS,EAAE,OAAO,EAAE,SAAS,qBAAqB;AAAA;AAAA,EAElD,MAAM,EAAE,OAAO,EAAE,SAAS;AAAA,EAC1B,aAAa,EAAE,OAAO,EAAE,SAAS;AAAA,EACjC,cAAc,EAAE,MAAM,EAAE,OAAO,CAAC;AAAA,EAChC,QAAQ,EACL,OAAO;AAAA,IACN,eAAe,EAAE,OAAO;AAAA,IACxB,iBAAiB,EAAE,OAAO;AAAA,EAC5B,CAAC,EACA,SAAS;AAAA,EACZ,OAAO,EACJ,OAAO;AAAA,IACN,OAAO,EAAE,OAAO;AAAA,IAChB,QAAQ,EAAE,OAAO;AAAA,EACnB,CAAC,EACA,SAAS;AAAA;AAAA,EAEZ,UAAU,EACP,KAAK;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,CAAC,EACA,SAAS;AACd,CAAC;AAEM,MAAM,2BAA2B,EAAE,OAAO;AAAA,EAC/C,SAAS,EAAE,OAAO,EAAE,SAAS,qBAAqB;AAAA,EAClD,aAAa;AACf,CAAC;AAOM,MAAM,8BAA8B,2BAA2B,OAAO;AAAA;AAAA,EAE3E,MAAM,EAAE,OAAO,EAAE,SAAS;AAAA,EAC1B,aAAa,EAAE,OAAO,EAAE,SAAS;AAAA,EACjC,cAAc,EAAE,MAAM,EAAE,OAAO,CAAC;AAAA,EAChC,QAAQ,EACL,OAAO;AAAA,IACN,eAAe,EAAE,OAAO;AAAA,IACxB,iBAAiB,EAAE,OAAO;AAAA,EAC5B,CAAC,EACA,SAAS;AAAA,EACZ,OAAO,EACJ,OAAO;AAAA,IACN,OAAO,EAAE,OAAO;AAAA,IAChB,QAAQ,EAAE,OAAO;AAAA,EACnB,CAAC,EACA,SAAS;AAAA;AAAA,EAEZ,UAAU,EACP,KAAK;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,CAAC,EACA,SAAS;AACd,CAAC;AAQD,MAAM,yBAAyB;AAAA,EAC7B;AAAA,EACA;AAAA,EACA,EAAE,UAAU,KAAK;AACnB;AAeO,MAAM,yBAAyB;AAAA,EACpC;AAAA,IACE,MAAM;AAAA,IACN,aAAa,EAAE,OAAO;AAAA,MACpB,SAAS,EAAE,OAAO,EAAE,SAAS,qBAAqB;AAAA,IACpD,CAAC;AAAA,IACD,cAAc;AAAA,EAChB;AAAA,EACA;AAAA,IACE,MAAM;AAAA,IACN,aAAa;AAAA,IACb,YAAY;AAAA,EACd;AAAA,EACA;AAAA,IACE,MAAM;AAAA,IACN,aAAa;AAAA,IACb,cAAc;AAAA,EAChB;AAAA,EACA,GAAG;AACL;AAEO,MAAM,uBAAuB,cAAc,sBAAsB;","names":[]}