llama-cpp-capacitor 0.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/plugin.js ADDED
@@ -0,0 +1,534 @@
1
+ var capacitorLlamaCpp = (function (exports, core) {
2
+ 'use strict';
3
+
4
+ var __rest = (undefined && undefined.__rest) || function (s, e) {
5
+ var t = {};
6
+ for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p) && e.indexOf(p) < 0)
7
+ t[p] = s[p];
8
+ if (s != null && typeof Object.getOwnPropertySymbols === "function")
9
+ for (var i = 0, p = Object.getOwnPropertySymbols(s); i < p.length; i++) {
10
+ if (e.indexOf(p[i]) < 0 && Object.prototype.propertyIsEnumerable.call(s, p[i]))
11
+ t[p[i]] = s[p[i]];
12
+ }
13
+ return t;
14
+ };
15
+ var _a, _b, _c;
16
+ // Constants
17
+ const LLAMACPP_MTMD_DEFAULT_MEDIA_MARKER = '<__media__>';
18
+ // Event names
19
+ const EVENT_ON_INIT_CONTEXT_PROGRESS = '@LlamaCpp_onInitContextProgress';
20
+ const EVENT_ON_TOKEN = '@LlamaCpp_onToken';
21
+ const EVENT_ON_NATIVE_LOG = '@LlamaCpp_onNativeLog';
22
+ // Register the plugin
23
+ const LlamaCpp = core.registerPlugin('LlamaCpp');
24
+ // Log listeners management
25
+ const logListeners = [];
26
+ // Set up native log listener
27
+ LlamaCpp.addListener(EVENT_ON_NATIVE_LOG, (evt) => {
28
+ logListeners.forEach((listener) => listener(evt.level, evt.text));
29
+ });
30
+ // Trigger unset to use default log callback
31
+ (_c = (_b = (_a = LlamaCpp === null || LlamaCpp === void 0 ? void 0 : LlamaCpp.toggleNativeLog) === null || _a === void 0 ? void 0 : _a.call(LlamaCpp, { enabled: false })) === null || _b === void 0 ? void 0 : _b.catch) === null || _c === void 0 ? void 0 : _c.call(_b, () => { });
32
+ const RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER = LLAMACPP_MTMD_DEFAULT_MEDIA_MARKER;
33
+ const validCacheTypes = [
34
+ 'f16',
35
+ 'f32',
36
+ 'bf16',
37
+ 'q8_0',
38
+ 'q4_0',
39
+ 'q4_1',
40
+ 'iq4_nl',
41
+ 'q5_0',
42
+ 'q5_1',
43
+ ];
44
+ const getJsonSchema = (responseFormat) => {
45
+ var _a;
46
+ if ((responseFormat === null || responseFormat === void 0 ? void 0 : responseFormat.type) === 'json_schema') {
47
+ return (_a = responseFormat.json_schema) === null || _a === void 0 ? void 0 : _a.schema;
48
+ }
49
+ if ((responseFormat === null || responseFormat === void 0 ? void 0 : responseFormat.type) === 'json_object') {
50
+ return responseFormat.schema || {};
51
+ }
52
+ return null;
53
+ };
54
+ class LlamaContext {
55
+ constructor({ contextId, gpu, reasonNoGPU, model }) {
56
+ this.gpu = false;
57
+ this.reasonNoGPU = '';
58
+ this.id = contextId;
59
+ this.gpu = gpu;
60
+ this.reasonNoGPU = reasonNoGPU;
61
+ this.model = model;
62
+ }
63
+ /**
64
+ * Load cached prompt & completion state from a file.
65
+ */
66
+ async loadSession(filepath) {
67
+ let path = filepath;
68
+ if (path.startsWith('file://'))
69
+ path = path.slice(7);
70
+ return LlamaCpp.loadSession({ contextId: this.id, filepath: path });
71
+ }
72
+ /**
73
+ * Save current cached prompt & completion state to a file.
74
+ */
75
+ async saveSession(filepath, options) {
76
+ return LlamaCpp.saveSession({
77
+ contextId: this.id,
78
+ filepath,
79
+ size: (options === null || options === void 0 ? void 0 : options.tokenSize) || -1
80
+ });
81
+ }
82
+ isLlamaChatSupported() {
83
+ return !!this.model.chatTemplates.llamaChat;
84
+ }
85
+ isJinjaSupported() {
86
+ const { minja } = this.model.chatTemplates;
87
+ return !!(minja === null || minja === void 0 ? void 0 : minja.toolUse) || !!(minja === null || minja === void 0 ? void 0 : minja.default);
88
+ }
89
+ async getFormattedChat(messages, template, params) {
90
+ var _a;
91
+ const mediaPaths = [];
92
+ const chat = messages.map((msg) => {
93
+ if (Array.isArray(msg.content)) {
94
+ const content = msg.content.map((part) => {
95
+ var _a;
96
+ // Handle multimodal content
97
+ if (part.type === 'image_url') {
98
+ let path = ((_a = part.image_url) === null || _a === void 0 ? void 0 : _a.url) || '';
99
+ if (path === null || path === void 0 ? void 0 : path.startsWith('file://'))
100
+ path = path.slice(7);
101
+ mediaPaths.push(path);
102
+ return {
103
+ type: 'text',
104
+ text: RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER,
105
+ };
106
+ }
107
+ else if (part.type === 'input_audio') {
108
+ const { input_audio: audio } = part;
109
+ if (!audio)
110
+ throw new Error('input_audio is required');
111
+ const { format } = audio;
112
+ if (format != 'wav' && format != 'mp3') {
113
+ throw new Error(`Unsupported audio format: ${format}`);
114
+ }
115
+ if (audio.url) {
116
+ const path = audio.url.replace(/file:\/\//, '');
117
+ mediaPaths.push(path);
118
+ }
119
+ else if (audio.data) {
120
+ mediaPaths.push(audio.data);
121
+ }
122
+ return {
123
+ type: 'text',
124
+ text: RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER,
125
+ };
126
+ }
127
+ return part;
128
+ });
129
+ return Object.assign(Object.assign({}, msg), { content });
130
+ }
131
+ return msg;
132
+ });
133
+ const useJinja = this.isJinjaSupported() && (params === null || params === void 0 ? void 0 : params.jinja);
134
+ let tmpl;
135
+ if (template)
136
+ tmpl = template; // Force replace if provided
137
+ const jsonSchema = getJsonSchema(params === null || params === void 0 ? void 0 : params.response_format);
138
+ const result = await LlamaCpp.getFormattedChat({
139
+ contextId: this.id,
140
+ messages: JSON.stringify(chat),
141
+ chatTemplate: tmpl,
142
+ params: {
143
+ jinja: useJinja,
144
+ json_schema: jsonSchema ? JSON.stringify(jsonSchema) : undefined,
145
+ tools: (params === null || params === void 0 ? void 0 : params.tools) ? JSON.stringify(params.tools) : undefined,
146
+ parallel_tool_calls: (params === null || params === void 0 ? void 0 : params.parallel_tool_calls) ? JSON.stringify(params.parallel_tool_calls)
147
+ : undefined,
148
+ tool_choice: params === null || params === void 0 ? void 0 : params.tool_choice,
149
+ enable_thinking: (_a = params === null || params === void 0 ? void 0 : params.enable_thinking) !== null && _a !== void 0 ? _a : true,
150
+ add_generation_prompt: params === null || params === void 0 ? void 0 : params.add_generation_prompt,
151
+ now: typeof (params === null || params === void 0 ? void 0 : params.now) === 'number' ? params.now.toString() : params === null || params === void 0 ? void 0 : params.now,
152
+ chat_template_kwargs: (params === null || params === void 0 ? void 0 : params.chat_template_kwargs) ? JSON.stringify(Object.entries(params.chat_template_kwargs).reduce((acc, [key, value]) => {
153
+ acc[key] = JSON.stringify(value); // Each value is a stringified JSON object
154
+ return acc;
155
+ }, {})) : undefined,
156
+ },
157
+ });
158
+ if (!useJinja) {
159
+ return {
160
+ type: 'llama-chat',
161
+ prompt: result,
162
+ has_media: mediaPaths.length > 0,
163
+ media_paths: mediaPaths,
164
+ };
165
+ }
166
+ const jinjaResult = result;
167
+ jinjaResult.type = 'jinja';
168
+ jinjaResult.has_media = mediaPaths.length > 0;
169
+ jinjaResult.media_paths = mediaPaths;
170
+ return jinjaResult;
171
+ }
172
+ /**
173
+ * Generate a completion based on the provided parameters
174
+ * @param params Completion parameters including prompt or messages
175
+ * @param callback Optional callback for token-by-token streaming
176
+ * @returns Promise resolving to the completion result
177
+ *
178
+ * Note: For multimodal support, you can include an media_paths parameter.
179
+ * This will process the images and add them to the context before generating text.
180
+ * Multimodal support must be enabled via initMultimodal() first.
181
+ */
182
+ async completion(params, callback) {
183
+ const nativeParams = Object.assign(Object.assign({}, params), { prompt: params.prompt || '', emit_partial_completion: !!callback });
184
+ if (params.messages) {
185
+ const formattedResult = await this.getFormattedChat(params.messages, params.chat_template || params.chatTemplate, {
186
+ jinja: params.jinja,
187
+ tools: params.tools,
188
+ parallel_tool_calls: params.parallel_tool_calls,
189
+ tool_choice: params.tool_choice,
190
+ enable_thinking: params.enable_thinking,
191
+ add_generation_prompt: params.add_generation_prompt,
192
+ now: params.now,
193
+ chat_template_kwargs: params.chat_template_kwargs,
194
+ });
195
+ if (formattedResult.type === 'jinja') {
196
+ const jinjaResult = formattedResult;
197
+ nativeParams.prompt = jinjaResult.prompt || '';
198
+ if (typeof jinjaResult.chat_format === 'number')
199
+ nativeParams.chat_format = jinjaResult.chat_format;
200
+ if (jinjaResult.grammar)
201
+ nativeParams.grammar = jinjaResult.grammar;
202
+ if (typeof jinjaResult.grammar_lazy === 'boolean')
203
+ nativeParams.grammar_lazy = jinjaResult.grammar_lazy;
204
+ if (jinjaResult.grammar_triggers)
205
+ nativeParams.grammar_triggers = jinjaResult.grammar_triggers;
206
+ if (jinjaResult.preserved_tokens)
207
+ nativeParams.preserved_tokens = jinjaResult.preserved_tokens;
208
+ if (jinjaResult.additional_stops) {
209
+ if (!nativeParams.stop)
210
+ nativeParams.stop = [];
211
+ nativeParams.stop.push(...jinjaResult.additional_stops);
212
+ }
213
+ if (jinjaResult.has_media) {
214
+ nativeParams.media_paths = jinjaResult.media_paths;
215
+ }
216
+ }
217
+ else if (formattedResult.type === 'llama-chat') {
218
+ const llamaChatResult = formattedResult;
219
+ nativeParams.prompt = llamaChatResult.prompt || '';
220
+ if (llamaChatResult.has_media) {
221
+ nativeParams.media_paths = llamaChatResult.media_paths;
222
+ }
223
+ }
224
+ }
225
+ else {
226
+ nativeParams.prompt = params.prompt || '';
227
+ }
228
+ // If media_paths were explicitly provided or extracted from messages, use them
229
+ if (!nativeParams.media_paths && params.media_paths) {
230
+ nativeParams.media_paths = params.media_paths;
231
+ }
232
+ if (nativeParams.response_format && !nativeParams.grammar) {
233
+ const jsonSchema = getJsonSchema(params.response_format);
234
+ if (jsonSchema)
235
+ nativeParams.json_schema = JSON.stringify(jsonSchema);
236
+ }
237
+ let tokenListener = callback &&
238
+ LlamaCpp.addListener(EVENT_ON_TOKEN, (evt) => {
239
+ const { contextId, tokenResult } = evt;
240
+ if (contextId !== this.id)
241
+ return;
242
+ callback(tokenResult);
243
+ });
244
+ if (!nativeParams.prompt)
245
+ throw new Error('Prompt is required');
246
+ const promise = LlamaCpp.completion({ contextId: this.id, params: nativeParams });
247
+ return promise
248
+ .then((completionResult) => {
249
+ tokenListener === null || tokenListener === void 0 ? void 0 : tokenListener.remove();
250
+ tokenListener = null;
251
+ return completionResult;
252
+ })
253
+ .catch((err) => {
254
+ tokenListener === null || tokenListener === void 0 ? void 0 : tokenListener.remove();
255
+ tokenListener = null;
256
+ throw err;
257
+ });
258
+ }
259
+ stopCompletion() {
260
+ return LlamaCpp.stopCompletion({ contextId: this.id });
261
+ }
262
+ /**
263
+ * Tokenize text or text with images
264
+ * @param text Text to tokenize
265
+ * @param params.media_paths Array of image paths to tokenize (if multimodal is enabled)
266
+ * @returns Promise resolving to the tokenize result
267
+ */
268
+ tokenize(text, { media_paths: mediaPaths, } = {}) {
269
+ return LlamaCpp.tokenize({ contextId: this.id, text, imagePaths: mediaPaths });
270
+ }
271
+ detokenize(tokens) {
272
+ return LlamaCpp.detokenize({ contextId: this.id, tokens });
273
+ }
274
+ embedding(text, params) {
275
+ return LlamaCpp.embedding({ contextId: this.id, text, params: params || {} });
276
+ }
277
+ /**
278
+ * Rerank documents based on relevance to a query
279
+ * @param query The query text to rank documents against
280
+ * @param documents Array of document texts to rank
281
+ * @param params Optional reranking parameters
282
+ * @returns Promise resolving to an array of ranking results with scores and indices
283
+ */
284
+ async rerank(query, documents, params) {
285
+ const results = await LlamaCpp.rerank({
286
+ contextId: this.id,
287
+ query,
288
+ documents,
289
+ params: params || {}
290
+ });
291
+ // Sort by score descending and add document text if requested
292
+ return results
293
+ .map((result) => (Object.assign(Object.assign({}, result), { document: documents[result.index] })))
294
+ .sort((a, b) => b.score - a.score);
295
+ }
296
+ async bench(pp, tg, pl, nr) {
297
+ const result = await LlamaCpp.bench({ contextId: this.id, pp, tg, pl, nr });
298
+ const [modelDesc, modelSize, modelNParams, ppAvg, ppStd, tgAvg, tgStd] = JSON.parse(result);
299
+ return {
300
+ modelDesc,
301
+ modelSize,
302
+ modelNParams,
303
+ ppAvg,
304
+ ppStd,
305
+ tgAvg,
306
+ tgStd,
307
+ };
308
+ }
309
+ async applyLoraAdapters(loraList) {
310
+ let loraAdapters = [];
311
+ if (loraList)
312
+ loraAdapters = loraList.map((l) => ({
313
+ path: l.path.replace(/file:\/\//, ''),
314
+ scaled: l.scaled,
315
+ }));
316
+ return LlamaCpp.applyLoraAdapters({ contextId: this.id, loraAdapters });
317
+ }
318
+ async removeLoraAdapters() {
319
+ return LlamaCpp.removeLoraAdapters({ contextId: this.id });
320
+ }
321
+ async getLoadedLoraAdapters() {
322
+ return LlamaCpp.getLoadedLoraAdapters({ contextId: this.id });
323
+ }
324
+ /**
325
+ * Initialize multimodal support with a mmproj file
326
+ * @param params Parameters for multimodal support
327
+ * @param params.path Path to the multimodal projector file
328
+ * @param params.use_gpu Whether to use GPU
329
+ * @returns Promise resolving to true if initialization was successful
330
+ */
331
+ async initMultimodal({ path, use_gpu: useGpu, }) {
332
+ if (path.startsWith('file://'))
333
+ path = path.slice(7);
334
+ return LlamaCpp.initMultimodal({
335
+ contextId: this.id,
336
+ params: {
337
+ path,
338
+ use_gpu: useGpu !== null && useGpu !== void 0 ? useGpu : true,
339
+ },
340
+ });
341
+ }
342
+ /**
343
+ * Check if multimodal support is enabled
344
+ * @returns Promise resolving to true if multimodal is enabled
345
+ */
346
+ async isMultimodalEnabled() {
347
+ return await LlamaCpp.isMultimodalEnabled({ contextId: this.id });
348
+ }
349
+ /**
350
+ * Check multimodal support
351
+ * @returns Promise resolving to an object with vision and audio support
352
+ */
353
+ async getMultimodalSupport() {
354
+ return await LlamaCpp.getMultimodalSupport({ contextId: this.id });
355
+ }
356
+ /**
357
+ * Release multimodal support
358
+ * @returns Promise resolving to void
359
+ */
360
+ async releaseMultimodal() {
361
+ return await LlamaCpp.releaseMultimodal({ contextId: this.id });
362
+ }
363
+ /**
364
+ * Initialize TTS support with a vocoder model
365
+ * @param params Parameters for TTS support
366
+ * @param params.path Path to the vocoder model
367
+ * @param params.n_batch Batch size for the vocoder model
368
+ * @returns Promise resolving to true if initialization was successful
369
+ */
370
+ async initVocoder({ path, n_batch: nBatch }) {
371
+ if (path.startsWith('file://'))
372
+ path = path.slice(7);
373
+ return await LlamaCpp.initVocoder({
374
+ contextId: this.id,
375
+ params: { path, n_batch: nBatch }
376
+ });
377
+ }
378
+ /**
379
+ * Check if TTS support is enabled
380
+ * @returns Promise resolving to true if TTS is enabled
381
+ */
382
+ async isVocoderEnabled() {
383
+ return await LlamaCpp.isVocoderEnabled({ contextId: this.id });
384
+ }
385
+ /**
386
+ * Get a formatted audio completion prompt
387
+ * @param speakerJsonStr JSON string representing the speaker
388
+ * @param textToSpeak Text to speak
389
+ * @returns Promise resolving to the formatted audio completion result with prompt and grammar
390
+ */
391
+ async getFormattedAudioCompletion(speaker, textToSpeak) {
392
+ return await LlamaCpp.getFormattedAudioCompletion({
393
+ contextId: this.id,
394
+ speakerJsonStr: speaker ? JSON.stringify(speaker) : '',
395
+ textToSpeak,
396
+ });
397
+ }
398
+ /**
399
+ * Get guide tokens for audio completion
400
+ * @param textToSpeak Text to speak
401
+ * @returns Promise resolving to the guide tokens
402
+ */
403
+ async getAudioCompletionGuideTokens(textToSpeak) {
404
+ return await LlamaCpp.getAudioCompletionGuideTokens({ contextId: this.id, textToSpeak });
405
+ }
406
+ /**
407
+ * Decode audio tokens
408
+ * @param tokens Array of audio tokens
409
+ * @returns Promise resolving to the decoded audio tokens
410
+ */
411
+ async decodeAudioTokens(tokens) {
412
+ return await LlamaCpp.decodeAudioTokens({ contextId: this.id, tokens });
413
+ }
414
+ /**
415
+ * Release TTS support
416
+ * @returns Promise resolving to void
417
+ */
418
+ async releaseVocoder() {
419
+ return await LlamaCpp.releaseVocoder({ contextId: this.id });
420
+ }
421
+ async release() {
422
+ return LlamaCpp.releaseContext({ contextId: this.id });
423
+ }
424
+ }
425
+ async function toggleNativeLog(enabled) {
426
+ return LlamaCpp.toggleNativeLog({ enabled });
427
+ }
428
+ function addNativeLogListener(listener) {
429
+ logListeners.push(listener);
430
+ return {
431
+ remove: () => {
432
+ logListeners.splice(logListeners.indexOf(listener), 1);
433
+ },
434
+ };
435
+ }
436
+ async function setContextLimit(limit) {
437
+ return LlamaCpp.setContextLimit({ limit });
438
+ }
439
+ let contextIdCounter = 0;
440
+ const contextIdRandom = () => process.env.NODE_ENV === 'test' ? 0 : Math.floor(Math.random() * 100000);
441
+ const modelInfoSkip = [
442
+ // Large fields
443
+ 'tokenizer.ggml.tokens',
444
+ 'tokenizer.ggml.token_type',
445
+ 'tokenizer.ggml.merges',
446
+ 'tokenizer.ggml.scores',
447
+ ];
448
+ async function loadLlamaModelInfo(model) {
449
+ let path = model;
450
+ if (path.startsWith('file://'))
451
+ path = path.slice(7);
452
+ return LlamaCpp.modelInfo({ path, skip: modelInfoSkip });
453
+ }
454
+ const poolTypeMap = {
455
+ // -1 is unspecified as undefined
456
+ none: 0,
457
+ mean: 1,
458
+ cls: 2,
459
+ last: 3,
460
+ rank: 4,
461
+ };
462
+ async function initLlama(_a, onProgress) {
463
+ var { model, is_model_asset: isModelAsset, pooling_type: poolingType, lora, lora_list: loraList } = _a, rest = __rest(_a, ["model", "is_model_asset", "pooling_type", "lora", "lora_list"]);
464
+ let path = model;
465
+ if (path.startsWith('file://'))
466
+ path = path.slice(7);
467
+ let loraPath = lora;
468
+ if (loraPath === null || loraPath === void 0 ? void 0 : loraPath.startsWith('file://'))
469
+ loraPath = loraPath.slice(7);
470
+ let loraAdapters = [];
471
+ if (loraList)
472
+ loraAdapters = loraList.map((l) => ({
473
+ path: l.path.replace(/file:\/\//, ''),
474
+ scaled: l.scaled,
475
+ }));
476
+ const contextId = contextIdCounter + contextIdRandom();
477
+ contextIdCounter += 1;
478
+ let removeProgressListener = null;
479
+ if (onProgress) {
480
+ removeProgressListener = LlamaCpp.addListener(EVENT_ON_INIT_CONTEXT_PROGRESS, (evt) => {
481
+ if (evt.contextId !== contextId)
482
+ return;
483
+ onProgress(evt.progress);
484
+ });
485
+ }
486
+ const poolType = poolTypeMap[poolingType];
487
+ if (rest.cache_type_k && !validCacheTypes.includes(rest.cache_type_k)) {
488
+ console.warn(`[LlamaCpp] initLlama: Invalid cache K type: ${rest.cache_type_k}, falling back to f16`);
489
+ delete rest.cache_type_k;
490
+ }
491
+ if (rest.cache_type_v && !validCacheTypes.includes(rest.cache_type_v)) {
492
+ console.warn(`[LlamaCpp] initLlama: Invalid cache V type: ${rest.cache_type_v}, falling back to f16`);
493
+ delete rest.cache_type_v;
494
+ }
495
+ const { gpu, reasonNoGPU, model: modelDetails, androidLib, } = await LlamaCpp.initContext({
496
+ contextId,
497
+ params: Object.assign({ model: path, is_model_asset: !!isModelAsset, use_progress_callback: !!onProgress, pooling_type: poolType, lora: loraPath, lora_list: loraAdapters }, rest),
498
+ }).catch((err) => {
499
+ removeProgressListener === null || removeProgressListener === void 0 ? void 0 : removeProgressListener.remove();
500
+ throw err;
501
+ });
502
+ removeProgressListener === null || removeProgressListener === void 0 ? void 0 : removeProgressListener.remove();
503
+ return new LlamaContext({
504
+ contextId,
505
+ gpu,
506
+ reasonNoGPU,
507
+ model: modelDetails,
508
+ androidLib,
509
+ });
510
+ }
511
+ async function releaseAllLlama() {
512
+ return LlamaCpp.releaseAllContexts();
513
+ }
514
+ const BuildInfo = {
515
+ number: '1.0.0',
516
+ commit: 'capacitor-llama-cpp',
517
+ };
518
+
519
+ exports.BuildInfo = BuildInfo;
520
+ exports.LLAMACPP_MTMD_DEFAULT_MEDIA_MARKER = LLAMACPP_MTMD_DEFAULT_MEDIA_MARKER;
521
+ exports.LlamaContext = LlamaContext;
522
+ exports.LlamaCpp = LlamaCpp;
523
+ exports.RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER = RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER;
524
+ exports.addNativeLogListener = addNativeLogListener;
525
+ exports.initLlama = initLlama;
526
+ exports.loadLlamaModelInfo = loadLlamaModelInfo;
527
+ exports.releaseAllLlama = releaseAllLlama;
528
+ exports.setContextLimit = setContextLimit;
529
+ exports.toggleNativeLog = toggleNativeLog;
530
+
531
+ return exports;
532
+
533
+ })({}, capacitorExports);
534
+ //# sourceMappingURL=plugin.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"plugin.js","sources":["esm/index.js"],"sourcesContent":["var __rest = (this && this.__rest) || function (s, e) {\r\n var t = {};\r\n for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p) && e.indexOf(p) < 0)\r\n t[p] = s[p];\r\n if (s != null && typeof Object.getOwnPropertySymbols === \"function\")\r\n for (var i = 0, p = Object.getOwnPropertySymbols(s); i < p.length; i++) {\r\n if (e.indexOf(p[i]) < 0 && Object.prototype.propertyIsEnumerable.call(s, p[i]))\r\n t[p[i]] = s[p[i]];\r\n }\r\n return t;\r\n};\r\nvar _a, _b, _c;\r\nimport { registerPlugin } from '@capacitor/core';\r\n// Constants\r\nexport const LLAMACPP_MTMD_DEFAULT_MEDIA_MARKER = '<__media__>';\r\n// Event names\r\nconst EVENT_ON_INIT_CONTEXT_PROGRESS = '@LlamaCpp_onInitContextProgress';\r\nconst EVENT_ON_TOKEN = '@LlamaCpp_onToken';\r\nconst EVENT_ON_NATIVE_LOG = '@LlamaCpp_onNativeLog';\r\n// Register the plugin\r\nconst LlamaCpp = registerPlugin('LlamaCpp');\r\n// Log listeners management\r\nconst logListeners = [];\r\n// Set up native log listener\r\nLlamaCpp.addListener(EVENT_ON_NATIVE_LOG, (evt) => {\r\n logListeners.forEach((listener) => listener(evt.level, evt.text));\r\n});\r\n// Trigger unset to use default log callback\r\n(_c = (_b = (_a = LlamaCpp === null || LlamaCpp === void 0 ? void 0 : LlamaCpp.toggleNativeLog) === null || _a === void 0 ? void 0 : _a.call(LlamaCpp, { enabled: false })) === null || _b === void 0 ? void 0 : _b.catch) === null || _c === void 0 ? void 0 : _c.call(_b, () => { });\r\nexport const RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER = LLAMACPP_MTMD_DEFAULT_MEDIA_MARKER;\r\nconst validCacheTypes = [\r\n 'f16',\r\n 'f32',\r\n 'bf16',\r\n 'q8_0',\r\n 'q4_0',\r\n 'q4_1',\r\n 'iq4_nl',\r\n 'q5_0',\r\n 'q5_1',\r\n];\r\nconst getJsonSchema = (responseFormat) => {\r\n var _a;\r\n if ((responseFormat === null || responseFormat === void 0 ? void 0 : responseFormat.type) === 'json_schema') {\r\n return (_a = responseFormat.json_schema) === null || _a === void 0 ? void 0 : _a.schema;\r\n }\r\n if ((responseFormat === null || responseFormat === void 0 ? void 0 : responseFormat.type) === 'json_object') {\r\n return responseFormat.schema || {};\r\n }\r\n return null;\r\n};\r\nexport class LlamaContext {\r\n constructor({ contextId, gpu, reasonNoGPU, model }) {\r\n this.gpu = false;\r\n this.reasonNoGPU = '';\r\n this.id = contextId;\r\n this.gpu = gpu;\r\n this.reasonNoGPU = reasonNoGPU;\r\n this.model = model;\r\n }\r\n /**\r\n * Load cached prompt & completion state from a file.\r\n */\r\n async loadSession(filepath) {\r\n let path = filepath;\r\n if (path.startsWith('file://'))\r\n path = path.slice(7);\r\n return LlamaCpp.loadSession({ contextId: this.id, filepath: path });\r\n }\r\n /**\r\n * Save current cached prompt & completion state to a file.\r\n */\r\n async saveSession(filepath, options) {\r\n return LlamaCpp.saveSession({\r\n contextId: this.id,\r\n filepath,\r\n size: (options === null || options === void 0 ? void 0 : options.tokenSize) || -1\r\n });\r\n }\r\n isLlamaChatSupported() {\r\n return !!this.model.chatTemplates.llamaChat;\r\n }\r\n isJinjaSupported() {\r\n const { minja } = this.model.chatTemplates;\r\n return !!(minja === null || minja === void 0 ? void 0 : minja.toolUse) || !!(minja === null || minja === void 0 ? void 0 : minja.default);\r\n }\r\n async getFormattedChat(messages, template, params) {\r\n var _a;\r\n const mediaPaths = [];\r\n const chat = messages.map((msg) => {\r\n if (Array.isArray(msg.content)) {\r\n const content = msg.content.map((part) => {\r\n var _a;\r\n // Handle multimodal content\r\n if (part.type === 'image_url') {\r\n let path = ((_a = part.image_url) === null || _a === void 0 ? void 0 : _a.url) || '';\r\n if (path === null || path === void 0 ? void 0 : path.startsWith('file://'))\r\n path = path.slice(7);\r\n mediaPaths.push(path);\r\n return {\r\n type: 'text',\r\n text: RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER,\r\n };\r\n }\r\n else if (part.type === 'input_audio') {\r\n const { input_audio: audio } = part;\r\n if (!audio)\r\n throw new Error('input_audio is required');\r\n const { format } = audio;\r\n if (format != 'wav' && format != 'mp3') {\r\n throw new Error(`Unsupported audio format: ${format}`);\r\n }\r\n if (audio.url) {\r\n const path = audio.url.replace(/file:\\/\\//, '');\r\n mediaPaths.push(path);\r\n }\r\n else if (audio.data) {\r\n mediaPaths.push(audio.data);\r\n }\r\n return {\r\n type: 'text',\r\n text: RNLLAMA_MTMD_DEFAULT_MEDIA_MARKER,\r\n };\r\n }\r\n return part;\r\n });\r\n return Object.assign(Object.assign({}, msg), { content });\r\n }\r\n return msg;\r\n });\r\n const useJinja = this.isJinjaSupported() && (params === null || params === void 0 ? void 0 : params.jinja);\r\n let tmpl;\r\n if (template)\r\n tmpl = template; // Force replace if provided\r\n const jsonSchema = getJsonSchema(params === null || params === void 0 ? void 0 : params.response_format);\r\n const result = await LlamaCpp.getFormattedChat({\r\n contextId: this.id,\r\n messages: JSON.stringify(chat),\r\n chatTemplate: tmpl,\r\n params: {\r\n jinja: useJinja,\r\n json_schema: jsonSchema ? JSON.stringify(jsonSchema) : undefined,\r\n tools: (params === null || params === void 0 ? void 0 : params.tools) ? JSON.stringify(params.tools) : undefined,\r\n parallel_tool_calls: (params === null || params === void 0 ? void 0 : params.parallel_tool_calls) ? JSON.stringify(params.parallel_tool_calls)\r\n : undefined,\r\n tool_choice: params === null || params === void 0 ? void 0 : params.tool_choice,\r\n enable_thinking: (_a = params === null || params === void 0 ? void 0 : params.enable_thinking) !== null && _a !== void 0 ? _a : true,\r\n add_generation_prompt: params === null || params === void 0 ? void 0 : params.add_generation_prompt,\r\n now: typeof (params === null || params === void 0 ? void 0 : params.now) === 'number' ? params.now.toString() : params === null || params === void 0 ? void 0 : params.now,\r\n chat_template_kwargs: (params === null || params === void 0 ? void 0 : params.chat_template_kwargs) ? JSON.stringify(Object.entries(params.chat_template_kwargs).reduce((acc, [key, value]) => {\r\n acc[key] = JSON.stringify(value); // Each value is a stringified JSON object\r\n return acc;\r\n }, {})) : undefined,\r\n },\r\n });\r\n if (!useJinja) {\r\n return {\r\n type: 'llama-chat',\r\n prompt: result,\r\n has_media: mediaPaths.length > 0,\r\n media_paths: mediaPaths,\r\n };\r\n }\r\n const jinjaResult = result;\r\n jinjaResult.type = 'jinja';\r\n jinjaResult.has_media = mediaPaths.length > 0;\r\n jinjaResult.media_paths = mediaPaths;\r\n return jinjaResult;\r\n }\r\n /**\r\n * Generate a completion based on the provided parameters\r\n * @param params Completion parameters including prompt or messages\r\n * @param callback Optional callback for token-by-token streaming\r\n * @returns Promise resolving to the completion result\r\n *\r\n * Note: For multimodal support, you can include an media_paths parameter.\r\n * This will process the images and add them to the context before generating text.\r\n * Multimodal support must be enabled via initMultimodal() first.\r\n */\r\n async completion(params, callback) {\r\n const nativeParams = Object.assign(Object.assign({}, params), { prompt: params.prompt || '', emit_partial_completion: !!callback });\r\n if (params.messages) {\r\n const formattedResult = await this.getFormattedChat(params.messages, params.chat_template || params.chatTemplate, {\r\n jinja: params.jinja,\r\n tools: params.tools,\r\n parallel_tool_calls: params.parallel_tool_calls,\r\n tool_choice: params.tool_choice,\r\n enable_thinking: params.enable_thinking,\r\n add_generation_prompt: params.add_generation_prompt,\r\n now: params.now,\r\n chat_template_kwargs: params.chat_template_kwargs,\r\n });\r\n if (formattedResult.type === 'jinja') {\r\n const jinjaResult = formattedResult;\r\n nativeParams.prompt = jinjaResult.prompt || '';\r\n if (typeof jinjaResult.chat_format === 'number')\r\n nativeParams.chat_format = jinjaResult.chat_format;\r\n if (jinjaResult.grammar)\r\n nativeParams.grammar = jinjaResult.grammar;\r\n if (typeof jinjaResult.grammar_lazy === 'boolean')\r\n nativeParams.grammar_lazy = jinjaResult.grammar_lazy;\r\n if (jinjaResult.grammar_triggers)\r\n nativeParams.grammar_triggers = jinjaResult.grammar_triggers;\r\n if (jinjaResult.preserved_tokens)\r\n nativeParams.preserved_tokens = jinjaResult.preserved_tokens;\r\n if (jinjaResult.additional_stops) {\r\n if (!nativeParams.stop)\r\n nativeParams.stop = [];\r\n nativeParams.stop.push(...jinjaResult.additional_stops);\r\n }\r\n if (jinjaResult.has_media) {\r\n nativeParams.media_paths = jinjaResult.media_paths;\r\n }\r\n }\r\n else if (formattedResult.type === 'llama-chat') {\r\n const llamaChatResult = formattedResult;\r\n nativeParams.prompt = llamaChatResult.prompt || '';\r\n if (llamaChatResult.has_media) {\r\n nativeParams.media_paths = llamaChatResult.media_paths;\r\n }\r\n }\r\n }\r\n else {\r\n nativeParams.prompt = params.prompt || '';\r\n }\r\n // If media_paths were explicitly provided or extracted from messages, use them\r\n if (!nativeParams.media_paths && params.media_paths) {\r\n nativeParams.media_paths = params.media_paths;\r\n }\r\n if (nativeParams.response_format && !nativeParams.grammar) {\r\n const jsonSchema = getJsonSchema(params.response_format);\r\n if (jsonSchema)\r\n nativeParams.json_schema = JSON.stringify(jsonSchema);\r\n }\r\n let tokenListener = callback &&\r\n LlamaCpp.addListener(EVENT_ON_TOKEN, (evt) => {\r\n const { contextId, tokenResult } = evt;\r\n if (contextId !== this.id)\r\n return;\r\n callback(tokenResult);\r\n });\r\n if (!nativeParams.prompt)\r\n throw new Error('Prompt is required');\r\n const promise = LlamaCpp.completion({ contextId: this.id, params: nativeParams });\r\n return promise\r\n .then((completionResult) => {\r\n tokenListener === null || tokenListener === void 0 ? void 0 : tokenListener.remove();\r\n tokenListener = null;\r\n return completionResult;\r\n })\r\n .catch((err) => {\r\n tokenListener === null || tokenListener === void 0 ? void 0 : tokenListener.remove();\r\n tokenListener = null;\r\n throw err;\r\n });\r\n }\r\n stopCompletion() {\r\n return LlamaCpp.stopCompletion({ contextId: this.id });\r\n }\r\n /**\r\n * Tokenize text or text with images\r\n * @param text Text to tokenize\r\n * @param params.media_paths Array of image paths to tokenize (if multimodal is enabled)\r\n * @returns Promise resolving to the tokenize result\r\n */\r\n tokenize(text, { media_paths: mediaPaths, } = {}) {\r\n return LlamaCpp.tokenize({ contextId: this.id, text, imagePaths: mediaPaths });\r\n }\r\n detokenize(tokens) {\r\n return LlamaCpp.detokenize({ contextId: this.id, tokens });\r\n }\r\n embedding(text, params) {\r\n return LlamaCpp.embedding({ contextId: this.id, text, params: params || {} });\r\n }\r\n /**\r\n * Rerank documents based on relevance to a query\r\n * @param query The query text to rank documents against\r\n * @param documents Array of document texts to rank\r\n * @param params Optional reranking parameters\r\n * @returns Promise resolving to an array of ranking results with scores and indices\r\n */\r\n async rerank(query, documents, params) {\r\n const results = await LlamaCpp.rerank({\r\n contextId: this.id,\r\n query,\r\n documents,\r\n params: params || {}\r\n });\r\n // Sort by score descending and add document text if requested\r\n return results\r\n .map((result) => (Object.assign(Object.assign({}, result), { document: documents[result.index] })))\r\n .sort((a, b) => b.score - a.score);\r\n }\r\n async bench(pp, tg, pl, nr) {\r\n const result = await LlamaCpp.bench({ contextId: this.id, pp, tg, pl, nr });\r\n const [modelDesc, modelSize, modelNParams, ppAvg, ppStd, tgAvg, tgStd] = JSON.parse(result);\r\n return {\r\n modelDesc,\r\n modelSize,\r\n modelNParams,\r\n ppAvg,\r\n ppStd,\r\n tgAvg,\r\n tgStd,\r\n };\r\n }\r\n async applyLoraAdapters(loraList) {\r\n let loraAdapters = [];\r\n if (loraList)\r\n loraAdapters = loraList.map((l) => ({\r\n path: l.path.replace(/file:\\/\\//, ''),\r\n scaled: l.scaled,\r\n }));\r\n return LlamaCpp.applyLoraAdapters({ contextId: this.id, loraAdapters });\r\n }\r\n async removeLoraAdapters() {\r\n return LlamaCpp.removeLoraAdapters({ contextId: this.id });\r\n }\r\n async getLoadedLoraAdapters() {\r\n return LlamaCpp.getLoadedLoraAdapters({ contextId: this.id });\r\n }\r\n /**\r\n * Initialize multimodal support with a mmproj file\r\n * @param params Parameters for multimodal support\r\n * @param params.path Path to the multimodal projector file\r\n * @param params.use_gpu Whether to use GPU\r\n * @returns Promise resolving to true if initialization was successful\r\n */\r\n async initMultimodal({ path, use_gpu: useGpu, }) {\r\n if (path.startsWith('file://'))\r\n path = path.slice(7);\r\n return LlamaCpp.initMultimodal({\r\n contextId: this.id,\r\n params: {\r\n path,\r\n use_gpu: useGpu !== null && useGpu !== void 0 ? useGpu : true,\r\n },\r\n });\r\n }\r\n /**\r\n * Check if multimodal support is enabled\r\n * @returns Promise resolving to true if multimodal is enabled\r\n */\r\n async isMultimodalEnabled() {\r\n return await LlamaCpp.isMultimodalEnabled({ contextId: this.id });\r\n }\r\n /**\r\n * Check multimodal support\r\n * @returns Promise resolving to an object with vision and audio support\r\n */\r\n async getMultimodalSupport() {\r\n return await LlamaCpp.getMultimodalSupport({ contextId: this.id });\r\n }\r\n /**\r\n * Release multimodal support\r\n * @returns Promise resolving to void\r\n */\r\n async releaseMultimodal() {\r\n return await LlamaCpp.releaseMultimodal({ contextId: this.id });\r\n }\r\n /**\r\n * Initialize TTS support with a vocoder model\r\n * @param params Parameters for TTS support\r\n * @param params.path Path to the vocoder model\r\n * @param params.n_batch Batch size for the vocoder model\r\n * @returns Promise resolving to true if initialization was successful\r\n */\r\n async initVocoder({ path, n_batch: nBatch }) {\r\n if (path.startsWith('file://'))\r\n path = path.slice(7);\r\n return await LlamaCpp.initVocoder({\r\n contextId: this.id,\r\n params: { path, n_batch: nBatch }\r\n });\r\n }\r\n /**\r\n * Check if TTS support is enabled\r\n * @returns Promise resolving to true if TTS is enabled\r\n */\r\n async isVocoderEnabled() {\r\n return await LlamaCpp.isVocoderEnabled({ contextId: this.id });\r\n }\r\n /**\r\n * Get a formatted audio completion prompt\r\n * @param speakerJsonStr JSON string representing the speaker\r\n * @param textToSpeak Text to speak\r\n * @returns Promise resolving to the formatted audio completion result with prompt and grammar\r\n */\r\n async getFormattedAudioCompletion(speaker, textToSpeak) {\r\n return await LlamaCpp.getFormattedAudioCompletion({\r\n contextId: this.id,\r\n speakerJsonStr: speaker ? JSON.stringify(speaker) : '',\r\n textToSpeak,\r\n });\r\n }\r\n /**\r\n * Get guide tokens for audio completion\r\n * @param textToSpeak Text to speak\r\n * @returns Promise resolving to the guide tokens\r\n */\r\n async getAudioCompletionGuideTokens(textToSpeak) {\r\n return await LlamaCpp.getAudioCompletionGuideTokens({ contextId: this.id, textToSpeak });\r\n }\r\n /**\r\n * Decode audio tokens\r\n * @param tokens Array of audio tokens\r\n * @returns Promise resolving to the decoded audio tokens\r\n */\r\n async decodeAudioTokens(tokens) {\r\n return await LlamaCpp.decodeAudioTokens({ contextId: this.id, tokens });\r\n }\r\n /**\r\n * Release TTS support\r\n * @returns Promise resolving to void\r\n */\r\n async releaseVocoder() {\r\n return await LlamaCpp.releaseVocoder({ contextId: this.id });\r\n }\r\n async release() {\r\n return LlamaCpp.releaseContext({ contextId: this.id });\r\n }\r\n}\r\nexport async function toggleNativeLog(enabled) {\r\n return LlamaCpp.toggleNativeLog({ enabled });\r\n}\r\nexport function addNativeLogListener(listener) {\r\n logListeners.push(listener);\r\n return {\r\n remove: () => {\r\n logListeners.splice(logListeners.indexOf(listener), 1);\r\n },\r\n };\r\n}\r\nexport async function setContextLimit(limit) {\r\n return LlamaCpp.setContextLimit({ limit });\r\n}\r\nlet contextIdCounter = 0;\r\nconst contextIdRandom = () => process.env.NODE_ENV === 'test' ? 0 : Math.floor(Math.random() * 100000);\r\nconst modelInfoSkip = [\r\n // Large fields\r\n 'tokenizer.ggml.tokens',\r\n 'tokenizer.ggml.token_type',\r\n 'tokenizer.ggml.merges',\r\n 'tokenizer.ggml.scores',\r\n];\r\nexport async function loadLlamaModelInfo(model) {\r\n let path = model;\r\n if (path.startsWith('file://'))\r\n path = path.slice(7);\r\n return LlamaCpp.modelInfo({ path, skip: modelInfoSkip });\r\n}\r\nconst poolTypeMap = {\r\n // -1 is unspecified as undefined\r\n none: 0,\r\n mean: 1,\r\n cls: 2,\r\n last: 3,\r\n rank: 4,\r\n};\r\nexport async function initLlama(_a, onProgress) {\r\n var { model, is_model_asset: isModelAsset, pooling_type: poolingType, lora, lora_list: loraList } = _a, rest = __rest(_a, [\"model\", \"is_model_asset\", \"pooling_type\", \"lora\", \"lora_list\"]);\r\n let path = model;\r\n if (path.startsWith('file://'))\r\n path = path.slice(7);\r\n let loraPath = lora;\r\n if (loraPath === null || loraPath === void 0 ? void 0 : loraPath.startsWith('file://'))\r\n loraPath = loraPath.slice(7);\r\n let loraAdapters = [];\r\n if (loraList)\r\n loraAdapters = loraList.map((l) => ({\r\n path: l.path.replace(/file:\\/\\//, ''),\r\n scaled: l.scaled,\r\n }));\r\n const contextId = contextIdCounter + contextIdRandom();\r\n contextIdCounter += 1;\r\n let removeProgressListener = null;\r\n if (onProgress) {\r\n removeProgressListener = LlamaCpp.addListener(EVENT_ON_INIT_CONTEXT_PROGRESS, (evt) => {\r\n if (evt.contextId !== contextId)\r\n return;\r\n onProgress(evt.progress);\r\n });\r\n }\r\n const poolType = poolTypeMap[poolingType];\r\n if (rest.cache_type_k && !validCacheTypes.includes(rest.cache_type_k)) {\r\n console.warn(`[LlamaCpp] initLlama: Invalid cache K type: ${rest.cache_type_k}, falling back to f16`);\r\n delete rest.cache_type_k;\r\n }\r\n if (rest.cache_type_v && !validCacheTypes.includes(rest.cache_type_v)) {\r\n console.warn(`[LlamaCpp] initLlama: Invalid cache V type: ${rest.cache_type_v}, falling back to f16`);\r\n delete rest.cache_type_v;\r\n }\r\n const { gpu, reasonNoGPU, model: modelDetails, androidLib, } = await LlamaCpp.initContext({\r\n contextId,\r\n params: Object.assign({ model: path, is_model_asset: !!isModelAsset, use_progress_callback: !!onProgress, pooling_type: poolType, lora: loraPath, lora_list: loraAdapters }, rest),\r\n }).catch((err) => {\r\n removeProgressListener === null || removeProgressListener === void 0 ? void 0 : removeProgressListener.remove();\r\n throw err;\r\n });\r\n removeProgressListener === null || removeProgressListener === void 0 ? void 0 : removeProgressListener.remove();\r\n return new LlamaContext({\r\n contextId,\r\n gpu,\r\n reasonNoGPU,\r\n model: modelDetails,\r\n androidLib,\r\n });\r\n}\r\nexport async function releaseAllLlama() {\r\n return LlamaCpp.releaseAllContexts();\r\n}\r\nexport const BuildInfo = {\r\n number: '1.0.0',\r\n commit: 'capacitor-llama-cpp',\r\n};\r\n// Re-export the plugin for direct access\r\nexport { LlamaCpp };\r\n//# sourceMappingURL=index.js.map"],"names":["this","registerPlugin"],"mappings":";;;IAAA,IAAI,MAAM,GAAG,CAACA,SAAI,IAAIA,SAAI,CAAC,MAAM,KAAK,UAAU,CAAC,EAAE,CAAC,EAAE;IACtD,IAAI,IAAI,CAAC,GAAG,EAAE,CAAC;IACf,IAAI,KAAK,IAAI,CAAC,IAAI,CAAC,EAAE,IAAI,MAAM,CAAC,SAAS,CAAC,cAAc,CAAC,IAAI,CAAC,CAAC,EAAE,CAAC,CAAC,IAAI,CAAC,CAAC,OAAO,CAAC,CAAC,CAAC,GAAG,CAAC;IACvF,QAAQ,CAAC,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC,CAAC;IACpB,IAAI,IAAI,CAAC,IAAI,IAAI,IAAI,OAAO,MAAM,CAAC,qBAAqB,KAAK,UAAU;IACvE,QAAQ,KAAK,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,GAAG,MAAM,CAAC,qBAAqB,CAAC,CAAC,CAAC,EAAE,CAAC,GAAG,CAAC,CAAC,MAAM,EAAE,CAAC,EAAE,EAAE;IAChF,YAAY,IAAI,CAAC,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,GAAG,CAAC,IAAI,MAAM,CAAC,SAAS,CAAC,oBAAoB,CAAC,IAAI,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,CAAC,CAAC;IAC1F,gBAAgB,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC;IAClC,QAAQ,CAAC;IACT,IAAI,OAAO,CAAC,CAAC;IACb,CAAC,CAAC;IACF,IAAI,EAAE,EAAE,EAAE,EAAE,EAAE,CAAC;IAEf;AACY,UAAC,kCAAkC,GAAG,cAAc;IAChE;IACA,MAAM,8BAA8B,GAAG,iCAAiC,CAAC;IACzE,MAAM,cAAc,GAAG,mBAAmB,CAAC;IAC3C,MAAM,mBAAmB,GAAG,uBAAuB,CAAC;IACpD;AACK,UAAC,QAAQ,GAAGC,mBAAc,CAAC,UAAU,EAAE;IAC5C;IACA,MAAM,YAAY,GAAG,EAAE,CAAC;IACxB;IACA,QAAQ,CAAC,WAAW,CAAC,mBAAmB,EAAE,CAAC,GAAG,KAAK;IACnD,IAAI,YAAY,CAAC,OAAO,CAAC,CAAC,QAAQ,KAAK,QAAQ,CAAC,GAAG,CAAC,KAAK,EAAE,GAAG,CAAC,IAAI,CAAC,CAAC,CAAC;IACtE,CAAC,CAAC,CAAC;IACH;IACA,CAAC,EAAE,GAAG,CAAC,EAAE,GAAG,CAAC,EAAE,GAAG,QAAQ,KAAK,IAAI,IAAI,QAAQ,KAAK,MAAM,GAAG,MAAM,GAAG,QAAQ,CAAC,eAAe,MAAM,IAAI,IAAI,EAAE,KAAK,MAAM,GAAG,MAAM,GAAG,EAAE,CAAC,IAAI,CAAC,QAAQ,EAAE,EAAE,OAAO,EAAE,KAAK,EAAE,CAAC,MAAM,IAAI,IAAI,EAAE,KAAK,MAAM,GAAG,MAAM,GAAG,EAAE,CAAC,KAAK,MAAM,IAAI,IAAI,EAAE,KAAK,MAAM,GAAG,MAAM,GAAG,EAAE,CAAC,IAAI,CAAC,EAAE,EAAE,MAAM,EAAE,CAAC,CAAC,CAAC;AAC3Q,UAAC,iCAAiC,GAAG,mCAAmC;IACpF,MAAM,eAAe,GAAG;IACxB,IAAI,KAAK;IACT,IAAI,KAAK;IACT,IAAI,MAAM;IACV,IAAI,MAAM;IACV,IAAI,MAAM;IACV,IAAI,MAAM;IACV,IAAI,QAAQ;IACZ,IAAI,MAAM;IACV,IAAI,MAAM;IACV,CAAC,CAAC;IACF,MAAM,aAAa,GAAG,CAAC,cAAc,KAAK;IAC1C,IAAI,IAAI,EAAE,CAAC;IACX,IAAI,IAAI,CAAC,cAAc,KAAK,IAAI,IAAI,cAAc,KAAK,MAAM,GAAG,MAAM,GAAG,cAAc,CAAC,IAAI,MAAM,aAAa,EAAE;IACjH,QAAQ,OAAO,CAAC,EAAE,GAAG,cAAc,CAAC,WAAW,MAAM,IAAI,IAAI,EAAE,KAAK,MAAM,GAAG,MAAM,GAAG,EAAE,CAAC,MAAM,CAAC;IAChG,IAAI,CAAC;IACL,IAAI,IAAI,CAAC,cAAc,KAAK,IAAI,IAAI,cAAc,KAAK,MAAM,GAAG,MAAM,GAAG,cAAc,CAAC,IAAI,MAAM,aAAa,EAAE;IACjH,QAAQ,OAAO,cAAc,CAAC,MAAM,IAAI,EAAE,CAAC;IAC3C,IAAI,CAAC;IACL,IAAI,OAAO,IAAI,CAAC;IAChB,CAAC,CAAC;IACK,MAAM,YAAY,CAAC;IAC1B,IAAI,WAAW,CAAC,EAAE,SAAS,EAAE,GAAG,EAAE,WAAW,EAAE,KAAK,EAAE,EAAE;IACxD,QAAQ,IAAI,CAAC,GAAG,GAAG,KAAK,CAAC;IACzB,QAAQ,IAAI,CAAC,WAAW,GAAG,EAAE,CAAC;IAC9B,QAAQ,IAAI,CAAC,EAAE,GAAG,SAAS,CAAC;IAC5B,QAAQ,IAAI,CAAC,GAAG,GAAG,GAAG,CAAC;IACvB,QAAQ,IAAI,CAAC,WAAW,GAAG,WAAW,CAAC;IACvC,QAAQ,IAAI,CAAC,KAAK,GAAG,KAAK,CAAC;IAC3B,IAAI,CAAC;IACL;IACA;IACA;IACA,IAAI,MAAM,WAAW,CAAC,QAAQ,EAAE;IAChC,QAAQ,IAAI,IAAI,GAAG,QAAQ,CAAC;IAC5B,QAAQ,IAAI,IAAI,CAAC,UAAU,CAAC,SAAS,CAAC;IACtC,YAAY,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IACjC,QAAQ,OAAO,QAAQ,CAAC,WAAW,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,QAAQ,EAAE,IAAI,EAAE,CAAC,CAAC;IAC5E,IAAI,CAAC;IACL;IACA;IACA;IACA,IAAI,MAAM,WAAW,CAAC,QAAQ,EAAE,OAAO,EAAE;IACzC,QAAQ,OAAO,QAAQ,CAAC,WAAW,CAAC;IACpC,YAAY,SAAS,EAAE,IAAI,CAAC,EAAE;IAC9B,YAAY,QAAQ;IACpB,YAAY,IAAI,EAAE,CAAC,OAAO,KAAK,IAAI,IAAI,OAAO,KAAK,MAAM,GAAG,MAAM,GAAG,OAAO,CAAC,SAAS,KAAK,EAAE;IAC7F,SAAS,CAAC,CAAC;IACX,IAAI,CAAC;IACL,IAAI,oBAAoB,GAAG;IAC3B,QAAQ,OAAO,CAAC,CAAC,IAAI,CAAC,KAAK,CAAC,aAAa,CAAC,SAAS,CAAC;IACpD,IAAI,CAAC;IACL,IAAI,gBAAgB,GAAG;IACvB,QAAQ,MAAM,EAAE,KAAK,EAAE,GAAG,IAAI,CAAC,KAAK,CAAC,aAAa,CAAC;IACnD,QAAQ,OAAO,CAAC,EAAE,KAAK,KAAK,IAAI,IAAI,KAAK,KAAK,MAAM,GAAG,MAAM,GAAG,KAAK,CAAC,OAAO,CAAC,IAAI,CAAC,EAAE,KAAK,KAAK,IAAI,IAAI,KAAK,KAAK,MAAM,GAAG,MAAM,GAAG,KAAK,CAAC,OAAO,CAAC,CAAC;IAClJ,IAAI,CAAC;IACL,IAAI,MAAM,gBAAgB,CAAC,QAAQ,EAAE,QAAQ,EAAE,MAAM,EAAE;IACvD,QAAQ,IAAI,EAAE,CAAC;IACf,QAAQ,MAAM,UAAU,GAAG,EAAE,CAAC;IAC9B,QAAQ,MAAM,IAAI,GAAG,QAAQ,CAAC,GAAG,CAAC,CAAC,GAAG,KAAK;IAC3C,YAAY,IAAI,KAAK,CAAC,OAAO,CAAC,GAAG,CAAC,OAAO,CAAC,EAAE;IAC5C,gBAAgB,MAAM,OAAO,GAAG,GAAG,CAAC,OAAO,CAAC,GAAG,CAAC,CAAC,IAAI,KAAK;IAC1D,oBAAoB,IAAI,EAAE,CAAC;IAC3B;IACA,oBAAoB,IAAI,IAAI,CAAC,IAAI,KAAK,WAAW,EAAE;IACnD,wBAAwB,IAAI,IAAI,GAAG,CAAC,CAAC,EAAE,GAAG,IAAI,CAAC,SAAS,MAAM,IAAI,IAAI,EAAE,KAAK,MAAM,GAAG,MAAM,GAAG,EAAE,CAAC,GAAG,KAAK,EAAE,CAAC;IAC7G,wBAAwB,IAAI,IAAI,KAAK,IAAI,IAAI,IAAI,KAAK,MAAM,GAAG,MAAM,GAAG,IAAI,CAAC,UAAU,CAAC,SAAS,CAAC;IAClG,4BAA4B,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IACjD,wBAAwB,UAAU,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC;IAC9C,wBAAwB,OAAO;IAC/B,4BAA4B,IAAI,EAAE,MAAM;IACxC,4BAA4B,IAAI,EAAE,iCAAiC;IACnE,yBAAyB,CAAC;IAC1B,oBAAoB,CAAC;IACrB,yBAAyB,IAAI,IAAI,CAAC,IAAI,KAAK,aAAa,EAAE;IAC1D,wBAAwB,MAAM,EAAE,WAAW,EAAE,KAAK,EAAE,GAAG,IAAI,CAAC;IAC5D,wBAAwB,IAAI,CAAC,KAAK;IAClC,4BAA4B,MAAM,IAAI,KAAK,CAAC,yBAAyB,CAAC,CAAC;IACvE,wBAAwB,MAAM,EAAE,MAAM,EAAE,GAAG,KAAK,CAAC;IACjD,wBAAwB,IAAI,MAAM,IAAI,KAAK,IAAI,MAAM,IAAI,KAAK,EAAE;IAChE,4BAA4B,MAAM,IAAI,KAAK,CAAC,CAAC,0BAA0B,EAAE,MAAM,CAAC,CAAC,CAAC,CAAC;IACnF,wBAAwB,CAAC;IACzB,wBAAwB,IAAI,KAAK,CAAC,GAAG,EAAE;IACvC,4BAA4B,MAAM,IAAI,GAAG,KAAK,CAAC,GAAG,CAAC,OAAO,CAAC,WAAW,EAAE,EAAE,CAAC,CAAC;IAC5E,4BAA4B,UAAU,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC;IAClD,wBAAwB,CAAC;IACzB,6BAA6B,IAAI,KAAK,CAAC,IAAI,EAAE;IAC7C,4BAA4B,UAAU,CAAC,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,CAAC;IACxD,wBAAwB,CAAC;IACzB,wBAAwB,OAAO;IAC/B,4BAA4B,IAAI,EAAE,MAAM;IACxC,4BAA4B,IAAI,EAAE,iCAAiC;IACnE,yBAAyB,CAAC;IAC1B,oBAAoB,CAAC;IACrB,oBAAoB,OAAO,IAAI,CAAC;IAChC,gBAAgB,CAAC,CAAC,CAAC;IACnB,gBAAgB,OAAO,MAAM,CAAC,MAAM,CAAC,MAAM,CAAC,MAAM,CAAC,EAAE,EAAE,GAAG,CAAC,EAAE,EAAE,OAAO,EAAE,CAAC,CAAC;IAC1E,YAAY,CAAC;IACb,YAAY,OAAO,GAAG,CAAC;IACvB,QAAQ,CAAC,CAAC,CAAC;IACX,QAAQ,MAAM,QAAQ,GAAG,IAAI,CAAC,gBAAgB,EAAE,KAAK,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,KAAK,CAAC,CAAC;IACnH,QAAQ,IAAI,IAAI,CAAC;IACjB,QAAQ,IAAI,QAAQ;IACpB,YAAY,IAAI,GAAG,QAAQ,CAAC;IAC5B,QAAQ,MAAM,UAAU,GAAG,aAAa,CAAC,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,eAAe,CAAC,CAAC;IACjH,QAAQ,MAAM,MAAM,GAAG,MAAM,QAAQ,CAAC,gBAAgB,CAAC;IACvD,YAAY,SAAS,EAAE,IAAI,CAAC,EAAE;IAC9B,YAAY,QAAQ,EAAE,IAAI,CAAC,SAAS,CAAC,IAAI,CAAC;IAC1C,YAAY,YAAY,EAAE,IAAI;IAC9B,YAAY,MAAM,EAAE;IACpB,gBAAgB,KAAK,EAAE,QAAQ;IAC/B,gBAAgB,WAAW,EAAE,UAAU,GAAG,IAAI,CAAC,SAAS,CAAC,UAAU,CAAC,GAAG,SAAS;IAChF,gBAAgB,KAAK,EAAE,CAAC,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,KAAK,IAAI,IAAI,CAAC,SAAS,CAAC,MAAM,CAAC,KAAK,CAAC,GAAG,SAAS;IAChI,gBAAgB,mBAAmB,EAAE,CAAC,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,mBAAmB,IAAI,IAAI,CAAC,SAAS,CAAC,MAAM,CAAC,mBAAmB,CAAC;IAC9J,sBAAsB,SAAS;IAC/B,gBAAgB,WAAW,EAAE,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,WAAW;IAC/F,gBAAgB,eAAe,EAAE,CAAC,EAAE,GAAG,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,eAAe,MAAM,IAAI,IAAI,EAAE,KAAK,MAAM,GAAG,EAAE,GAAG,IAAI;IACpJ,gBAAgB,qBAAqB,EAAE,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,qBAAqB;IACnH,gBAAgB,GAAG,EAAE,QAAQ,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,GAAG,CAAC,KAAK,QAAQ,GAAG,MAAM,CAAC,GAAG,CAAC,QAAQ,EAAE,GAAG,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,GAAG;IAC1L,gBAAgB,oBAAoB,EAAE,CAAC,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,MAAM,CAAC,oBAAoB,IAAI,IAAI,CAAC,SAAS,CAAC,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,oBAAoB,CAAC,CAAC,MAAM,CAAC,CAAC,GAAG,EAAE,CAAC,GAAG,EAAE,KAAK,CAAC,KAAK;IAC/M,oBAAoB,GAAG,CAAC,GAAG,CAAC,GAAG,IAAI,CAAC,SAAS,CAAC,KAAK,CAAC,CAAC;IACrD,oBAAoB,OAAO,GAAG,CAAC;IAC/B,gBAAgB,CAAC,EAAE,EAAE,CAAC,CAAC,GAAG,SAAS;IACnC,aAAa;IACb,SAAS,CAAC,CAAC;IACX,QAAQ,IAAI,CAAC,QAAQ,EAAE;IACvB,YAAY,OAAO;IACnB,gBAAgB,IAAI,EAAE,YAAY;IAClC,gBAAgB,MAAM,EAAE,MAAM;IAC9B,gBAAgB,SAAS,EAAE,UAAU,CAAC,MAAM,GAAG,CAAC;IAChD,gBAAgB,WAAW,EAAE,UAAU;IACvC,aAAa,CAAC;IACd,QAAQ,CAAC;IACT,QAAQ,MAAM,WAAW,GAAG,MAAM,CAAC;IACnC,QAAQ,WAAW,CAAC,IAAI,GAAG,OAAO,CAAC;IACnC,QAAQ,WAAW,CAAC,SAAS,GAAG,UAAU,CAAC,MAAM,GAAG,CAAC,CAAC;IACtD,QAAQ,WAAW,CAAC,WAAW,GAAG,UAAU,CAAC;IAC7C,QAAQ,OAAO,WAAW,CAAC;IAC3B,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,UAAU,CAAC,MAAM,EAAE,QAAQ,EAAE;IACvC,QAAQ,MAAM,YAAY,GAAG,MAAM,CAAC,MAAM,CAAC,MAAM,CAAC,MAAM,CAAC,EAAE,EAAE,MAAM,CAAC,EAAE,EAAE,MAAM,EAAE,MAAM,CAAC,MAAM,IAAI,EAAE,EAAE,uBAAuB,EAAE,CAAC,CAAC,QAAQ,EAAE,CAAC,CAAC;IAC5I,QAAQ,IAAI,MAAM,CAAC,QAAQ,EAAE;IAC7B,YAAY,MAAM,eAAe,GAAG,MAAM,IAAI,CAAC,gBAAgB,CAAC,MAAM,CAAC,QAAQ,EAAE,MAAM,CAAC,aAAa,IAAI,MAAM,CAAC,YAAY,EAAE;IAC9H,gBAAgB,KAAK,EAAE,MAAM,CAAC,KAAK;IACnC,gBAAgB,KAAK,EAAE,MAAM,CAAC,KAAK;IACnC,gBAAgB,mBAAmB,EAAE,MAAM,CAAC,mBAAmB;IAC/D,gBAAgB,WAAW,EAAE,MAAM,CAAC,WAAW;IAC/C,gBAAgB,eAAe,EAAE,MAAM,CAAC,eAAe;IACvD,gBAAgB,qBAAqB,EAAE,MAAM,CAAC,qBAAqB;IACnE,gBAAgB,GAAG,EAAE,MAAM,CAAC,GAAG;IAC/B,gBAAgB,oBAAoB,EAAE,MAAM,CAAC,oBAAoB;IACjE,aAAa,CAAC,CAAC;IACf,YAAY,IAAI,eAAe,CAAC,IAAI,KAAK,OAAO,EAAE;IAClD,gBAAgB,MAAM,WAAW,GAAG,eAAe,CAAC;IACpD,gBAAgB,YAAY,CAAC,MAAM,GAAG,WAAW,CAAC,MAAM,IAAI,EAAE,CAAC;IAC/D,gBAAgB,IAAI,OAAO,WAAW,CAAC,WAAW,KAAK,QAAQ;IAC/D,oBAAoB,YAAY,CAAC,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;IACvE,gBAAgB,IAAI,WAAW,CAAC,OAAO;IACvC,oBAAoB,YAAY,CAAC,OAAO,GAAG,WAAW,CAAC,OAAO,CAAC;IAC/D,gBAAgB,IAAI,OAAO,WAAW,CAAC,YAAY,KAAK,SAAS;IACjE,oBAAoB,YAAY,CAAC,YAAY,GAAG,WAAW,CAAC,YAAY,CAAC;IACzE,gBAAgB,IAAI,WAAW,CAAC,gBAAgB;IAChD,oBAAoB,YAAY,CAAC,gBAAgB,GAAG,WAAW,CAAC,gBAAgB,CAAC;IACjF,gBAAgB,IAAI,WAAW,CAAC,gBAAgB;IAChD,oBAAoB,YAAY,CAAC,gBAAgB,GAAG,WAAW,CAAC,gBAAgB,CAAC;IACjF,gBAAgB,IAAI,WAAW,CAAC,gBAAgB,EAAE;IAClD,oBAAoB,IAAI,CAAC,YAAY,CAAC,IAAI;IAC1C,wBAAwB,YAAY,CAAC,IAAI,GAAG,EAAE,CAAC;IAC/C,oBAAoB,YAAY,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,WAAW,CAAC,gBAAgB,CAAC,CAAC;IAC5E,gBAAgB,CAAC;IACjB,gBAAgB,IAAI,WAAW,CAAC,SAAS,EAAE;IAC3C,oBAAoB,YAAY,CAAC,WAAW,GAAG,WAAW,CAAC,WAAW,CAAC;IACvE,gBAAgB,CAAC;IACjB,YAAY,CAAC;IACb,iBAAiB,IAAI,eAAe,CAAC,IAAI,KAAK,YAAY,EAAE;IAC5D,gBAAgB,MAAM,eAAe,GAAG,eAAe,CAAC;IACxD,gBAAgB,YAAY,CAAC,MAAM,GAAG,eAAe,CAAC,MAAM,IAAI,EAAE,CAAC;IACnE,gBAAgB,IAAI,eAAe,CAAC,SAAS,EAAE;IAC/C,oBAAoB,YAAY,CAAC,WAAW,GAAG,eAAe,CAAC,WAAW,CAAC;IAC3E,gBAAgB,CAAC;IACjB,YAAY,CAAC;IACb,QAAQ,CAAC;IACT,aAAa;IACb,YAAY,YAAY,CAAC,MAAM,GAAG,MAAM,CAAC,MAAM,IAAI,EAAE,CAAC;IACtD,QAAQ,CAAC;IACT;IACA,QAAQ,IAAI,CAAC,YAAY,CAAC,WAAW,IAAI,MAAM,CAAC,WAAW,EAAE;IAC7D,YAAY,YAAY,CAAC,WAAW,GAAG,MAAM,CAAC,WAAW,CAAC;IAC1D,QAAQ,CAAC;IACT,QAAQ,IAAI,YAAY,CAAC,eAAe,IAAI,CAAC,YAAY,CAAC,OAAO,EAAE;IACnE,YAAY,MAAM,UAAU,GAAG,aAAa,CAAC,MAAM,CAAC,eAAe,CAAC,CAAC;IACrE,YAAY,IAAI,UAAU;IAC1B,gBAAgB,YAAY,CAAC,WAAW,GAAG,IAAI,CAAC,SAAS,CAAC,UAAU,CAAC,CAAC;IACtE,QAAQ,CAAC;IACT,QAAQ,IAAI,aAAa,GAAG,QAAQ;IACpC,YAAY,QAAQ,CAAC,WAAW,CAAC,cAAc,EAAE,CAAC,GAAG,KAAK;IAC1D,gBAAgB,MAAM,EAAE,SAAS,EAAE,WAAW,EAAE,GAAG,GAAG,CAAC;IACvD,gBAAgB,IAAI,SAAS,KAAK,IAAI,CAAC,EAAE;IACzC,oBAAoB,OAAO;IAC3B,gBAAgB,QAAQ,CAAC,WAAW,CAAC,CAAC;IACtC,YAAY,CAAC,CAAC,CAAC;IACf,QAAQ,IAAI,CAAC,YAAY,CAAC,MAAM;IAChC,YAAY,MAAM,IAAI,KAAK,CAAC,oBAAoB,CAAC,CAAC;IAClD,QAAQ,MAAM,OAAO,GAAG,QAAQ,CAAC,UAAU,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,MAAM,EAAE,YAAY,EAAE,CAAC,CAAC;IAC1F,QAAQ,OAAO,OAAO;IACtB,aAAa,IAAI,CAAC,CAAC,gBAAgB,KAAK;IACxC,YAAY,aAAa,KAAK,IAAI,IAAI,aAAa,KAAK,MAAM,GAAG,MAAM,GAAG,aAAa,CAAC,MAAM,EAAE,CAAC;IACjG,YAAY,aAAa,GAAG,IAAI,CAAC;IACjC,YAAY,OAAO,gBAAgB,CAAC;IACpC,QAAQ,CAAC,CAAC;IACV,aAAa,KAAK,CAAC,CAAC,GAAG,KAAK;IAC5B,YAAY,aAAa,KAAK,IAAI,IAAI,aAAa,KAAK,MAAM,GAAG,MAAM,GAAG,aAAa,CAAC,MAAM,EAAE,CAAC;IACjG,YAAY,aAAa,GAAG,IAAI,CAAC;IACjC,YAAY,MAAM,GAAG,CAAC;IACtB,QAAQ,CAAC,CAAC,CAAC;IACX,IAAI,CAAC;IACL,IAAI,cAAc,GAAG;IACrB,QAAQ,OAAO,QAAQ,CAAC,cAAc,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IAC/D,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA;IACA,IAAI,QAAQ,CAAC,IAAI,EAAE,EAAE,WAAW,EAAE,UAAU,GAAG,GAAG,EAAE,EAAE;IACtD,QAAQ,OAAO,QAAQ,CAAC,QAAQ,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,IAAI,EAAE,UAAU,EAAE,UAAU,EAAE,CAAC,CAAC;IACvF,IAAI,CAAC;IACL,IAAI,UAAU,CAAC,MAAM,EAAE;IACvB,QAAQ,OAAO,QAAQ,CAAC,UAAU,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,MAAM,EAAE,CAAC,CAAC;IACnE,IAAI,CAAC;IACL,IAAI,SAAS,CAAC,IAAI,EAAE,MAAM,EAAE;IAC5B,QAAQ,OAAO,QAAQ,CAAC,SAAS,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,IAAI,EAAE,MAAM,EAAE,MAAM,IAAI,EAAE,EAAE,CAAC,CAAC;IACtF,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,MAAM,CAAC,KAAK,EAAE,SAAS,EAAE,MAAM,EAAE;IAC3C,QAAQ,MAAM,OAAO,GAAG,MAAM,QAAQ,CAAC,MAAM,CAAC;IAC9C,YAAY,SAAS,EAAE,IAAI,CAAC,EAAE;IAC9B,YAAY,KAAK;IACjB,YAAY,SAAS;IACrB,YAAY,MAAM,EAAE,MAAM,IAAI,EAAE;IAChC,SAAS,CAAC,CAAC;IACX;IACA,QAAQ,OAAO,OAAO;IACtB,aAAa,GAAG,CAAC,CAAC,MAAM,MAAM,MAAM,CAAC,MAAM,CAAC,MAAM,CAAC,MAAM,CAAC,EAAE,EAAE,MAAM,CAAC,EAAE,EAAE,QAAQ,EAAE,SAAS,CAAC,MAAM,CAAC,KAAK,CAAC,EAAE,CAAC,CAAC,CAAC;IAC/G,aAAa,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,KAAK,CAAC,CAAC,KAAK,GAAG,CAAC,CAAC,KAAK,CAAC,CAAC;IAC/C,IAAI,CAAC;IACL,IAAI,MAAM,KAAK,CAAC,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE;IAChC,QAAQ,MAAM,MAAM,GAAG,MAAM,QAAQ,CAAC,KAAK,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,EAAE,CAAC,CAAC;IACpF,QAAQ,MAAM,CAAC,SAAS,EAAE,SAAS,EAAE,YAAY,EAAE,KAAK,EAAE,KAAK,EAAE,KAAK,EAAE,KAAK,CAAC,GAAG,IAAI,CAAC,KAAK,CAAC,MAAM,CAAC,CAAC;IACpG,QAAQ,OAAO;IACf,YAAY,SAAS;IACrB,YAAY,SAAS;IACrB,YAAY,YAAY;IACxB,YAAY,KAAK;IACjB,YAAY,KAAK;IACjB,YAAY,KAAK;IACjB,YAAY,KAAK;IACjB,SAAS,CAAC;IACV,IAAI,CAAC;IACL,IAAI,MAAM,iBAAiB,CAAC,QAAQ,EAAE;IACtC,QAAQ,IAAI,YAAY,GAAG,EAAE,CAAC;IAC9B,QAAQ,IAAI,QAAQ;IACpB,YAAY,YAAY,GAAG,QAAQ,CAAC,GAAG,CAAC,CAAC,CAAC,MAAM;IAChD,gBAAgB,IAAI,EAAE,CAAC,CAAC,IAAI,CAAC,OAAO,CAAC,WAAW,EAAE,EAAE,CAAC;IACrD,gBAAgB,MAAM,EAAE,CAAC,CAAC,MAAM;IAChC,aAAa,CAAC,CAAC,CAAC;IAChB,QAAQ,OAAO,QAAQ,CAAC,iBAAiB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,YAAY,EAAE,CAAC,CAAC;IAChF,IAAI,CAAC;IACL,IAAI,MAAM,kBAAkB,GAAG;IAC/B,QAAQ,OAAO,QAAQ,CAAC,kBAAkB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IACnE,IAAI,CAAC;IACL,IAAI,MAAM,qBAAqB,GAAG;IAClC,QAAQ,OAAO,QAAQ,CAAC,qBAAqB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IACtE,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,cAAc,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,MAAM,GAAG,EAAE;IACrD,QAAQ,IAAI,IAAI,CAAC,UAAU,CAAC,SAAS,CAAC;IACtC,YAAY,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IACjC,QAAQ,OAAO,QAAQ,CAAC,cAAc,CAAC;IACvC,YAAY,SAAS,EAAE,IAAI,CAAC,EAAE;IAC9B,YAAY,MAAM,EAAE;IACpB,gBAAgB,IAAI;IACpB,gBAAgB,OAAO,EAAE,MAAM,KAAK,IAAI,IAAI,MAAM,KAAK,MAAM,GAAG,MAAM,GAAG,IAAI;IAC7E,aAAa;IACb,SAAS,CAAC,CAAC;IACX,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA,IAAI,MAAM,mBAAmB,GAAG;IAChC,QAAQ,OAAO,MAAM,QAAQ,CAAC,mBAAmB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IAC1E,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA,IAAI,MAAM,oBAAoB,GAAG;IACjC,QAAQ,OAAO,MAAM,QAAQ,CAAC,oBAAoB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IAC3E,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA,IAAI,MAAM,iBAAiB,GAAG;IAC9B,QAAQ,OAAO,MAAM,QAAQ,CAAC,iBAAiB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IACxE,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,WAAW,CAAC,EAAE,IAAI,EAAE,OAAO,EAAE,MAAM,EAAE,EAAE;IACjD,QAAQ,IAAI,IAAI,CAAC,UAAU,CAAC,SAAS,CAAC;IACtC,YAAY,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IACjC,QAAQ,OAAO,MAAM,QAAQ,CAAC,WAAW,CAAC;IAC1C,YAAY,SAAS,EAAE,IAAI,CAAC,EAAE;IAC9B,YAAY,MAAM,EAAE,EAAE,IAAI,EAAE,OAAO,EAAE,MAAM,EAAE;IAC7C,SAAS,CAAC,CAAC;IACX,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA,IAAI,MAAM,gBAAgB,GAAG;IAC7B,QAAQ,OAAO,MAAM,QAAQ,CAAC,gBAAgB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IACvE,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,2BAA2B,CAAC,OAAO,EAAE,WAAW,EAAE;IAC5D,QAAQ,OAAO,MAAM,QAAQ,CAAC,2BAA2B,CAAC;IAC1D,YAAY,SAAS,EAAE,IAAI,CAAC,EAAE;IAC9B,YAAY,cAAc,EAAE,OAAO,GAAG,IAAI,CAAC,SAAS,CAAC,OAAO,CAAC,GAAG,EAAE;IAClE,YAAY,WAAW;IACvB,SAAS,CAAC,CAAC;IACX,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,6BAA6B,CAAC,WAAW,EAAE;IACrD,QAAQ,OAAO,MAAM,QAAQ,CAAC,6BAA6B,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,WAAW,EAAE,CAAC,CAAC;IACjG,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA;IACA,IAAI,MAAM,iBAAiB,CAAC,MAAM,EAAE;IACpC,QAAQ,OAAO,MAAM,QAAQ,CAAC,iBAAiB,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,MAAM,EAAE,CAAC,CAAC;IAChF,IAAI,CAAC;IACL;IACA;IACA;IACA;IACA,IAAI,MAAM,cAAc,GAAG;IAC3B,QAAQ,OAAO,MAAM,QAAQ,CAAC,cAAc,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IACrE,IAAI,CAAC;IACL,IAAI,MAAM,OAAO,GAAG;IACpB,QAAQ,OAAO,QAAQ,CAAC,cAAc,CAAC,EAAE,SAAS,EAAE,IAAI,CAAC,EAAE,EAAE,CAAC,CAAC;IAC/D,IAAI,CAAC;IACL,CAAC;IACM,eAAe,eAAe,CAAC,OAAO,EAAE;IAC/C,IAAI,OAAO,QAAQ,CAAC,eAAe,CAAC,EAAE,OAAO,EAAE,CAAC,CAAC;IACjD,CAAC;IACM,SAAS,oBAAoB,CAAC,QAAQ,EAAE;IAC/C,IAAI,YAAY,CAAC,IAAI,CAAC,QAAQ,CAAC,CAAC;IAChC,IAAI,OAAO;IACX,QAAQ,MAAM,EAAE,MAAM;IACtB,YAAY,YAAY,CAAC,MAAM,CAAC,YAAY,CAAC,OAAO,CAAC,QAAQ,CAAC,EAAE,CAAC,CAAC,CAAC;IACnE,QAAQ,CAAC;IACT,KAAK,CAAC;IACN,CAAC;IACM,eAAe,eAAe,CAAC,KAAK,EAAE;IAC7C,IAAI,OAAO,QAAQ,CAAC,eAAe,CAAC,EAAE,KAAK,EAAE,CAAC,CAAC;IAC/C,CAAC;IACD,IAAI,gBAAgB,GAAG,CAAC,CAAC;IACzB,MAAM,eAAe,GAAG,MAAM,OAAO,CAAC,GAAG,CAAC,QAAQ,KAAK,MAAM,GAAG,CAAC,GAAG,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,MAAM,EAAE,GAAG,MAAM,CAAC,CAAC;IACvG,MAAM,aAAa,GAAG;IACtB;IACA,IAAI,uBAAuB;IAC3B,IAAI,2BAA2B;IAC/B,IAAI,uBAAuB;IAC3B,IAAI,uBAAuB;IAC3B,CAAC,CAAC;IACK,eAAe,kBAAkB,CAAC,KAAK,EAAE;IAChD,IAAI,IAAI,IAAI,GAAG,KAAK,CAAC;IACrB,IAAI,IAAI,IAAI,CAAC,UAAU,CAAC,SAAS,CAAC;IAClC,QAAQ,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IAC7B,IAAI,OAAO,QAAQ,CAAC,SAAS,CAAC,EAAE,IAAI,EAAE,IAAI,EAAE,aAAa,EAAE,CAAC,CAAC;IAC7D,CAAC;IACD,MAAM,WAAW,GAAG;IACpB;IACA,IAAI,IAAI,EAAE,CAAC;IACX,IAAI,IAAI,EAAE,CAAC;IACX,IAAI,GAAG,EAAE,CAAC;IACV,IAAI,IAAI,EAAE,CAAC;IACX,IAAI,IAAI,EAAE,CAAC;IACX,CAAC,CAAC;IACK,eAAe,SAAS,CAAC,EAAE,EAAE,UAAU,EAAE;IAChD,IAAI,IAAI,EAAE,KAAK,EAAE,cAAc,EAAE,YAAY,EAAE,YAAY,EAAE,WAAW,EAAE,IAAI,EAAE,SAAS,EAAE,QAAQ,EAAE,GAAG,EAAE,EAAE,IAAI,GAAG,MAAM,CAAC,EAAE,EAAE,CAAC,OAAO,EAAE,gBAAgB,EAAE,cAAc,EAAE,MAAM,EAAE,WAAW,CAAC,CAAC,CAAC;IAChM,IAAI,IAAI,IAAI,GAAG,KAAK,CAAC;IACrB,IAAI,IAAI,IAAI,CAAC,UAAU,CAAC,SAAS,CAAC;IAClC,QAAQ,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IAC7B,IAAI,IAAI,QAAQ,GAAG,IAAI,CAAC;IACxB,IAAI,IAAI,QAAQ,KAAK,IAAI,IAAI,QAAQ,KAAK,MAAM,GAAG,MAAM,GAAG,QAAQ,CAAC,UAAU,CAAC,SAAS,CAAC;IAC1F,QAAQ,QAAQ,GAAG,QAAQ,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC;IACrC,IAAI,IAAI,YAAY,GAAG,EAAE,CAAC;IAC1B,IAAI,IAAI,QAAQ;IAChB,QAAQ,YAAY,GAAG,QAAQ,CAAC,GAAG,CAAC,CAAC,CAAC,MAAM;IAC5C,YAAY,IAAI,EAAE,CAAC,CAAC,IAAI,CAAC,OAAO,CAAC,WAAW,EAAE,EAAE,CAAC;IACjD,YAAY,MAAM,EAAE,CAAC,CAAC,MAAM;IAC5B,SAAS,CAAC,CAAC,CAAC;IACZ,IAAI,MAAM,SAAS,GAAG,gBAAgB,GAAG,eAAe,EAAE,CAAC;IAC3D,IAAI,gBAAgB,IAAI,CAAC,CAAC;IAC1B,IAAI,IAAI,sBAAsB,GAAG,IAAI,CAAC;IACtC,IAAI,IAAI,UAAU,EAAE;IACpB,QAAQ,sBAAsB,GAAG,QAAQ,CAAC,WAAW,CAAC,8BAA8B,EAAE,CAAC,GAAG,KAAK;IAC/F,YAAY,IAAI,GAAG,CAAC,SAAS,KAAK,SAAS;IAC3C,gBAAgB,OAAO;IACvB,YAAY,UAAU,CAAC,GAAG,CAAC,QAAQ,CAAC,CAAC;IACrC,QAAQ,CAAC,CAAC,CAAC;IACX,IAAI,CAAC;IACL,IAAI,MAAM,QAAQ,GAAG,WAAW,CAAC,WAAW,CAAC,CAAC;IAC9C,IAAI,IAAI,IAAI,CAAC,YAAY,IAAI,CAAC,eAAe,CAAC,QAAQ,CAAC,IAAI,CAAC,YAAY,CAAC,EAAE;IAC3E,QAAQ,OAAO,CAAC,IAAI,CAAC,CAAC,4CAA4C,EAAE,IAAI,CAAC,YAAY,CAAC,qBAAqB,CAAC,CAAC,CAAC;IAC9G,QAAQ,OAAO,IAAI,CAAC,YAAY,CAAC;IACjC,IAAI,CAAC;IACL,IAAI,IAAI,IAAI,CAAC,YAAY,IAAI,CAAC,eAAe,CAAC,QAAQ,CAAC,IAAI,CAAC,YAAY,CAAC,EAAE;IAC3E,QAAQ,OAAO,CAAC,IAAI,CAAC,CAAC,4CAA4C,EAAE,IAAI,CAAC,YAAY,CAAC,qBAAqB,CAAC,CAAC,CAAC;IAC9G,QAAQ,OAAO,IAAI,CAAC,YAAY,CAAC;IACjC,IAAI,CAAC;IACL,IAAI,MAAM,EAAE,GAAG,EAAE,WAAW,EAAE,KAAK,EAAE,YAAY,EAAE,UAAU,GAAG,GAAG,MAAM,QAAQ,CAAC,WAAW,CAAC;IAC9F,QAAQ,SAAS;IACjB,QAAQ,MAAM,EAAE,MAAM,CAAC,MAAM,CAAC,EAAE,KAAK,EAAE,IAAI,EAAE,cAAc,EAAE,CAAC,CAAC,YAAY,EAAE,qBAAqB,EAAE,CAAC,CAAC,UAAU,EAAE,YAAY,EAAE,QAAQ,EAAE,IAAI,EAAE,QAAQ,EAAE,SAAS,EAAE,YAAY,EAAE,EAAE,IAAI,CAAC;IAC1L,KAAK,CAAC,CAAC,KAAK,CAAC,CAAC,GAAG,KAAK;IACtB,QAAQ,sBAAsB,KAAK,IAAI,IAAI,sBAAsB,KAAK,MAAM,GAAG,MAAM,GAAG,sBAAsB,CAAC,MAAM,EAAE,CAAC;IACxH,QAAQ,MAAM,GAAG,CAAC;IAClB,IAAI,CAAC,CAAC,CAAC;IACP,IAAI,sBAAsB,KAAK,IAAI,IAAI,sBAAsB,KAAK,MAAM,GAAG,MAAM,GAAG,sBAAsB,CAAC,MAAM,EAAE,CAAC;IACpH,IAAI,OAAO,IAAI,YAAY,CAAC;IAC5B,QAAQ,SAAS;IACjB,QAAQ,GAAG;IACX,QAAQ,WAAW;IACnB,QAAQ,KAAK,EAAE,YAAY;IAC3B,QAAQ,UAAU;IAClB,KAAK,CAAC,CAAC;IACP,CAAC;IACM,eAAe,eAAe,GAAG;IACxC,IAAI,OAAO,QAAQ,CAAC,kBAAkB,EAAE,CAAC;IACzC,CAAC;AACW,UAAC,SAAS,GAAG;IACzB,IAAI,MAAM,EAAE,OAAO;IACnB,IAAI,MAAM,EAAE,qBAAqB;IACjC;;;;;;;;;;;;;;;;;;;;"}