@atcute/lex-cli 2.4.0 → 2.5.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +107 -10
- package/dist/cli.js +10 -168
- package/dist/cli.js.map +1 -1
- package/dist/codegen.d.ts.map +1 -1
- package/dist/codegen.js +76 -78
- package/dist/codegen.js.map +1 -1
- package/dist/commands/export.d.ts +17 -0
- package/dist/commands/export.d.ts.map +1 -0
- package/dist/commands/export.js +76 -0
- package/dist/commands/export.js.map +1 -0
- package/dist/commands/generate.d.ts +17 -0
- package/dist/commands/generate.d.ts.map +1 -0
- package/dist/commands/generate.js +136 -0
- package/dist/commands/generate.js.map +1 -0
- package/dist/commands/pull.d.ts +17 -0
- package/dist/commands/pull.d.ts.map +1 -0
- package/dist/{pull.js → commands/pull.js} +35 -81
- package/dist/commands/pull.js.map +1 -0
- package/dist/config.d.ts +68 -6
- package/dist/config.d.ts.map +1 -1
- package/dist/config.js +54 -3
- package/dist/config.js.map +1 -1
- package/dist/git.d.ts.map +1 -1
- package/dist/git.js.map +1 -1
- package/dist/index.d.ts +65 -1
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js.map +1 -1
- package/dist/lexicon-loader.d.ts +17 -0
- package/dist/lexicon-loader.d.ts.map +1 -0
- package/dist/lexicon-loader.js +167 -0
- package/dist/lexicon-loader.js.map +1 -0
- package/dist/lexicon-metadata.js.map +1 -1
- package/dist/pull-sources/atproto.d.ts +17 -0
- package/dist/pull-sources/atproto.d.ts.map +1 -0
- package/dist/pull-sources/atproto.js +192 -0
- package/dist/pull-sources/atproto.js.map +1 -0
- package/dist/pull-sources/git.d.ts +15 -0
- package/dist/pull-sources/git.d.ts.map +1 -0
- package/dist/pull-sources/git.js +80 -0
- package/dist/pull-sources/git.js.map +1 -0
- package/dist/pull-sources/types.d.ts +16 -0
- package/dist/pull-sources/types.d.ts.map +1 -0
- package/dist/pull-sources/types.js +2 -0
- package/dist/pull-sources/types.js.map +1 -0
- package/dist/shared-options.d.ts +6 -0
- package/dist/shared-options.d.ts.map +1 -0
- package/dist/shared-options.js +11 -0
- package/dist/shared-options.js.map +1 -0
- package/package.json +12 -9
- package/src/cli.ts +9 -210
- package/src/codegen.ts +90 -88
- package/src/commands/export.ts +106 -0
- package/src/commands/generate.ts +170 -0
- package/src/{pull.ts → commands/pull.ts} +49 -116
- package/src/config.ts +67 -4
- package/src/lexicon-loader.ts +201 -0
- package/src/pull-sources/atproto.ts +243 -0
- package/src/pull-sources/git.ts +103 -0
- package/src/pull-sources/types.ts +18 -0
- package/src/shared-options.ts +13 -0
- package/dist/pull.d.ts +0 -7
- package/dist/pull.d.ts.map +0 -1
- package/dist/pull.js.map +0 -1
package/README.md
CHANGED
|
@@ -1,6 +1,10 @@
|
|
|
1
1
|
# @atcute/lex-cli
|
|
2
2
|
|
|
3
|
-
|
|
3
|
+
generate TypeScript schemas from lexicon documents.
|
|
4
|
+
|
|
5
|
+
```sh
|
|
6
|
+
npm install @atcute/lex-cli
|
|
7
|
+
```
|
|
4
8
|
|
|
5
9
|
## quick start
|
|
6
10
|
|
|
@@ -17,22 +21,88 @@ export default defineLexiconConfig({
|
|
|
17
21
|
});
|
|
18
22
|
```
|
|
19
23
|
|
|
20
|
-
then run the tool:
|
|
24
|
+
then run the tool (it automatically finds `lex.config.js` or `lex.config.ts`):
|
|
21
25
|
|
|
22
26
|
```
|
|
23
|
-
npm exec lex-cli generate
|
|
27
|
+
npm exec lex-cli generate
|
|
28
|
+
```
|
|
29
|
+
|
|
30
|
+
## authoring lexicons with TypeScript
|
|
31
|
+
|
|
32
|
+
instead of writing lexicons as JSON documents, you can author them programmatically using the
|
|
33
|
+
builder functions from `@atcute/lexicon-doc/builder`.
|
|
34
|
+
|
|
35
|
+
```ts
|
|
36
|
+
// file: lexicons-src/com/example/bookmark.ts
|
|
37
|
+
import { array, document, object, record, required, string } from '@atcute/lexicon-doc/builder';
|
|
38
|
+
|
|
39
|
+
export default document({
|
|
40
|
+
id: 'com.example.bookmark',
|
|
41
|
+
defs: {
|
|
42
|
+
main: record({
|
|
43
|
+
key: 'tid',
|
|
44
|
+
description: 'a saved link to come back to later',
|
|
45
|
+
record: object({
|
|
46
|
+
properties: {
|
|
47
|
+
subject: required(string({ format: 'uri' })),
|
|
48
|
+
createdAt: required(string({ format: 'datetime' })),
|
|
49
|
+
tags: array({ items: string(), description: 'tags for organizing bookmarks' }),
|
|
50
|
+
},
|
|
51
|
+
}),
|
|
52
|
+
}),
|
|
53
|
+
},
|
|
54
|
+
});
|
|
55
|
+
```
|
|
56
|
+
|
|
57
|
+
update your config to point to TypeScript files:
|
|
58
|
+
|
|
59
|
+
```ts
|
|
60
|
+
// file: lex.config.js
|
|
61
|
+
import { defineLexiconConfig } from '@atcute/lex-cli';
|
|
62
|
+
|
|
63
|
+
export default defineLexiconConfig({
|
|
64
|
+
files: ['lexicons-src/**/*.ts'],
|
|
65
|
+
outdir: 'src/lexicons/',
|
|
66
|
+
});
|
|
67
|
+
```
|
|
68
|
+
|
|
69
|
+
### exporting lexicons to JSON
|
|
70
|
+
|
|
71
|
+
if you need the actual JSON lexicon documents (e.g., for publishing or sharing), configure the
|
|
72
|
+
export command:
|
|
73
|
+
|
|
74
|
+
```ts
|
|
75
|
+
export default defineLexiconConfig({
|
|
76
|
+
files: ['lexicons-src/**/*.ts'],
|
|
77
|
+
outdir: 'src/lexicons/',
|
|
78
|
+
export: {
|
|
79
|
+
outdir: 'lexicons/',
|
|
80
|
+
clean: true,
|
|
81
|
+
},
|
|
82
|
+
});
|
|
83
|
+
```
|
|
84
|
+
|
|
85
|
+
then run:
|
|
86
|
+
|
|
87
|
+
```
|
|
88
|
+
npm exec lex-cli export
|
|
24
89
|
```
|
|
25
90
|
|
|
26
91
|
## pulling lexicons
|
|
27
92
|
|
|
28
|
-
|
|
29
|
-
|
|
93
|
+
you can pull lexicon files from other sources.
|
|
94
|
+
|
|
95
|
+
### git sources
|
|
96
|
+
|
|
97
|
+
pull lexicons from git repositories using sparse checkout:
|
|
30
98
|
|
|
31
99
|
```ts
|
|
32
100
|
// file: lex.config.js
|
|
33
101
|
import { defineLexiconConfig } from '@atcute/lex-cli';
|
|
34
102
|
|
|
35
103
|
export default defineLexiconConfig({
|
|
104
|
+
files: ['lexicons/**/*.json'],
|
|
105
|
+
outdir: 'src/lexicons/',
|
|
36
106
|
pull: {
|
|
37
107
|
outdir: 'lexicons/',
|
|
38
108
|
clean: true,
|
|
@@ -45,23 +115,50 @@ export default defineLexiconConfig({
|
|
|
45
115
|
},
|
|
46
116
|
],
|
|
47
117
|
},
|
|
118
|
+
});
|
|
119
|
+
```
|
|
120
|
+
|
|
121
|
+
### atproto sources
|
|
122
|
+
|
|
123
|
+
pull lexicons directly from the AT Protocol network.
|
|
124
|
+
|
|
125
|
+
```ts
|
|
126
|
+
export default defineLexiconConfig({
|
|
48
127
|
files: ['lexicons/**/*.json'],
|
|
49
128
|
outdir: 'src/lexicons/',
|
|
129
|
+
pull: {
|
|
130
|
+
outdir: 'lexicons/',
|
|
131
|
+
sources: [
|
|
132
|
+
{
|
|
133
|
+
type: 'atproto',
|
|
134
|
+
mode: 'nsids',
|
|
135
|
+
nsids: ['app.bsky.feed.post', 'app.bsky.actor.profile'],
|
|
136
|
+
},
|
|
137
|
+
{
|
|
138
|
+
type: 'atproto',
|
|
139
|
+
mode: 'authority',
|
|
140
|
+
authority: 'atproto-lexicons.bsky.social',
|
|
141
|
+
pattern: ['com.atproto.*'], // optional
|
|
142
|
+
},
|
|
143
|
+
],
|
|
144
|
+
},
|
|
50
145
|
});
|
|
51
146
|
```
|
|
52
147
|
|
|
148
|
+
### running the pull command
|
|
149
|
+
|
|
53
150
|
pull the lexicons to disk, then generate types from them:
|
|
54
151
|
|
|
55
152
|
```
|
|
56
|
-
npm exec lex-cli pull
|
|
57
|
-
npm exec lex-cli generate
|
|
153
|
+
npm exec lex-cli pull
|
|
154
|
+
npm exec lex-cli generate
|
|
58
155
|
```
|
|
59
156
|
|
|
60
157
|
## publishing your schemas
|
|
61
158
|
|
|
62
|
-
if you're packaging your generated schemas as a publishable library, add the `atcute:lexicons`
|
|
63
|
-
|
|
64
|
-
|
|
159
|
+
if you're packaging your generated schemas as a publishable library, add the `atcute:lexicons` field
|
|
160
|
+
to your package.json. this allows other projects to automatically discover and import your schemas
|
|
161
|
+
without manual configuration.
|
|
65
162
|
|
|
66
163
|
```json
|
|
67
164
|
{
|
package/dist/cli.js
CHANGED
|
@@ -1,175 +1,17 @@
|
|
|
1
|
-
import * as fs from 'node:fs/promises';
|
|
2
|
-
import * as path from 'node:path';
|
|
3
|
-
import { lexiconDoc, refineLexiconDoc } from '@atcute/lexicon-doc';
|
|
4
|
-
import { object } from '@optique/core/constructs';
|
|
5
|
-
import { command, constant, option } from '@optique/core/primitives';
|
|
6
1
|
import { or } from '@optique/core/constructs';
|
|
7
2
|
import { run } from '@optique/run';
|
|
8
|
-
import {
|
|
9
|
-
import
|
|
10
|
-
import {
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
import { runPull } from './pull.js';
|
|
14
|
-
/**
|
|
15
|
-
* Resolves package imports to ImportMapping[]
|
|
16
|
-
*/
|
|
17
|
-
const resolveImportsToMappings = async (imports, configDirname) => {
|
|
18
|
-
const mappings = [];
|
|
19
|
-
for (const packageName of imports) {
|
|
20
|
-
// Walk up from config directory to find package in node_modules
|
|
21
|
-
let packageJson;
|
|
22
|
-
let currentDir = configDirname;
|
|
23
|
-
let found = false;
|
|
24
|
-
while (currentDir !== path.dirname(currentDir)) {
|
|
25
|
-
const candidatePath = path.join(currentDir, 'node_modules', packageName, 'package.json');
|
|
26
|
-
try {
|
|
27
|
-
const content = await fs.readFile(candidatePath, 'utf8');
|
|
28
|
-
packageJson = JSON.parse(content);
|
|
29
|
-
found = true;
|
|
30
|
-
break;
|
|
31
|
-
}
|
|
32
|
-
catch (err) {
|
|
33
|
-
// Only continue to parent if file not found
|
|
34
|
-
if (err.code !== 'ENOENT') {
|
|
35
|
-
console.error(pc.bold(pc.red(`failed to read package.json for "${packageName}":`)));
|
|
36
|
-
console.error(err);
|
|
37
|
-
process.exit(1);
|
|
38
|
-
}
|
|
39
|
-
// Not found, try parent directory
|
|
40
|
-
currentDir = path.dirname(currentDir);
|
|
41
|
-
}
|
|
42
|
-
}
|
|
43
|
-
if (!found) {
|
|
44
|
-
console.error(pc.bold(pc.red(`failed to resolve package "${packageName}"`)));
|
|
45
|
-
console.error(`Could not find package in node_modules starting from ${configDirname}`);
|
|
46
|
-
process.exit(1);
|
|
47
|
-
}
|
|
48
|
-
// Validate package.json
|
|
49
|
-
const result = packageJsonSchema.try(packageJson, { mode: 'passthrough' });
|
|
50
|
-
if (!result.ok) {
|
|
51
|
-
console.error(pc.bold(pc.red(`invalid atcute:lexicons in "${packageName}":`)));
|
|
52
|
-
console.error(result.message);
|
|
53
|
-
for (const issue of result.issues) {
|
|
54
|
-
console.log(`- ${issue.code} at .${issue.path.join('.')}`);
|
|
55
|
-
}
|
|
56
|
-
process.exit(1);
|
|
57
|
-
}
|
|
58
|
-
const lexicons = result.value['atcute:lexicons'];
|
|
59
|
-
if (!lexicons?.mappings) {
|
|
60
|
-
continue;
|
|
61
|
-
}
|
|
62
|
-
// Convert mapping to ImportMapping[]
|
|
63
|
-
for (const [pattern, entry] of Object.entries(lexicons.mappings)) {
|
|
64
|
-
const isWildcard = pattern.endsWith('.*');
|
|
65
|
-
mappings.push({
|
|
66
|
-
nsid: [pattern],
|
|
67
|
-
imports: (nsid) => {
|
|
68
|
-
// Check if pattern matches
|
|
69
|
-
if (isWildcard) {
|
|
70
|
-
if (!nsid.startsWith(pattern.slice(0, -1))) {
|
|
71
|
-
throw new Error(`NSID ${nsid} does not match pattern ${pattern}`);
|
|
72
|
-
}
|
|
73
|
-
}
|
|
74
|
-
else {
|
|
75
|
-
if (nsid !== pattern) {
|
|
76
|
-
throw new Error(`NSID ${nsid} does not match pattern ${pattern}`);
|
|
77
|
-
}
|
|
78
|
-
}
|
|
79
|
-
const nsidPrefix = isWildcard ? pattern.slice(0, -2) : pattern;
|
|
80
|
-
const nsidRemainder = isWildcard ? nsid.slice(nsidPrefix.length + 1) : '';
|
|
81
|
-
let expandedPath = entry.path
|
|
82
|
-
.replaceAll('{{nsid}}', nsid.replaceAll('.', '/'))
|
|
83
|
-
.replaceAll('{{nsid_remainder}}', nsidRemainder.replaceAll('.', '/'))
|
|
84
|
-
.replaceAll('{{nsid_prefix}}', nsidPrefix.replaceAll('.', '/'));
|
|
85
|
-
if (expandedPath === '.') {
|
|
86
|
-
expandedPath = packageName;
|
|
87
|
-
}
|
|
88
|
-
else if (expandedPath.startsWith('./')) {
|
|
89
|
-
expandedPath = `${packageName}/${expandedPath.slice(2)}`;
|
|
90
|
-
}
|
|
91
|
-
return {
|
|
92
|
-
type: entry.type,
|
|
93
|
-
from: expandedPath,
|
|
94
|
-
};
|
|
95
|
-
},
|
|
96
|
-
});
|
|
97
|
-
}
|
|
98
|
-
}
|
|
99
|
-
return mappings;
|
|
100
|
-
};
|
|
101
|
-
const parser = or(command('generate', object({
|
|
102
|
-
type: constant('generate'),
|
|
103
|
-
config: option('-c', '--config', pathParser({ metavar: 'CONFIG' })),
|
|
104
|
-
})), command('pull', object({
|
|
105
|
-
type: constant('pull'),
|
|
106
|
-
config: option('-c', '--config', pathParser({ metavar: 'CONFIG' })),
|
|
107
|
-
})));
|
|
108
|
-
const result = run(parser, { programName: 'lex-cli' });
|
|
3
|
+
import { exportCommandSchema, runExport } from './commands/export.js';
|
|
4
|
+
import { generateCommandSchema, runGenerate } from './commands/generate.js';
|
|
5
|
+
import { pullCommandSchema, runPull } from './commands/pull.js';
|
|
6
|
+
const parser = or(generateCommandSchema, pullCommandSchema, exportCommandSchema);
|
|
7
|
+
const result = run(parser, { programName: 'lex-cli', help: 'both' });
|
|
109
8
|
if (result.type === 'generate') {
|
|
110
|
-
|
|
111
|
-
// Resolve imports to mappings
|
|
112
|
-
const importMappings = config.imports ? await resolveImportsToMappings(config.imports, config.root) : [];
|
|
113
|
-
const allMappings = [...importMappings, ...(config.mappings ?? [])];
|
|
114
|
-
const documents = [];
|
|
115
|
-
for await (const filename of fs.glob(config.files, { cwd: config.root })) {
|
|
116
|
-
let source;
|
|
117
|
-
try {
|
|
118
|
-
source = await fs.readFile(path.join(config.root, filename), 'utf8');
|
|
119
|
-
}
|
|
120
|
-
catch (err) {
|
|
121
|
-
console.error(pc.bold(pc.red(`file read error with "${filename}"`)));
|
|
122
|
-
console.error(err);
|
|
123
|
-
process.exit(1);
|
|
124
|
-
}
|
|
125
|
-
let json;
|
|
126
|
-
try {
|
|
127
|
-
json = JSON.parse(source);
|
|
128
|
-
}
|
|
129
|
-
catch (err) {
|
|
130
|
-
console.error(pc.bold(pc.red(`json parse error in "${filename}"`)));
|
|
131
|
-
console.error(err);
|
|
132
|
-
process.exit(1);
|
|
133
|
-
}
|
|
134
|
-
const result = lexiconDoc.try(json, { mode: 'strip' });
|
|
135
|
-
if (!result.ok) {
|
|
136
|
-
console.error(pc.bold(pc.red(`schema validation failed for "${filename}"`)));
|
|
137
|
-
console.error(result.message);
|
|
138
|
-
for (const issue of result.issues) {
|
|
139
|
-
console.log(`- ${issue.code} at .${issue.path.join('.')}`);
|
|
140
|
-
}
|
|
141
|
-
process.exit(1);
|
|
142
|
-
}
|
|
143
|
-
const issues = refineLexiconDoc(result.value, true);
|
|
144
|
-
if (issues.length > 0) {
|
|
145
|
-
console.error(pc.bold(pc.red(`lint validation failed for "${filename}"`)));
|
|
146
|
-
for (const issue of issues) {
|
|
147
|
-
console.log(`- ${issue.message} at .${issue.path.join('.')}`);
|
|
148
|
-
}
|
|
149
|
-
process.exit(1);
|
|
150
|
-
}
|
|
151
|
-
documents.push(result.value);
|
|
152
|
-
}
|
|
153
|
-
const generationResult = await generateLexiconApi({
|
|
154
|
-
documents: documents,
|
|
155
|
-
mappings: allMappings,
|
|
156
|
-
modules: {
|
|
157
|
-
importSuffix: config.modules?.importSuffix ?? '.js',
|
|
158
|
-
},
|
|
159
|
-
prettier: {
|
|
160
|
-
cwd: process.cwd(),
|
|
161
|
-
},
|
|
162
|
-
});
|
|
163
|
-
const outdir = path.join(config.root, config.outdir);
|
|
164
|
-
for (const file of generationResult.files) {
|
|
165
|
-
const filename = path.join(outdir, file.filename);
|
|
166
|
-
const dirname = path.dirname(filename);
|
|
167
|
-
await fs.mkdir(dirname, { recursive: true });
|
|
168
|
-
await fs.writeFile(filename, file.code);
|
|
169
|
-
}
|
|
9
|
+
await runGenerate(result);
|
|
170
10
|
}
|
|
171
11
|
else if (result.type === 'pull') {
|
|
172
|
-
|
|
173
|
-
|
|
12
|
+
await runPull(result);
|
|
13
|
+
}
|
|
14
|
+
else if (result.type === 'export') {
|
|
15
|
+
await runExport(result);
|
|
174
16
|
}
|
|
175
17
|
//# sourceMappingURL=cli.js.map
|
package/dist/cli.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"cli.js","sourceRoot":"","sources":["../src/cli.ts"],"names":[],"mappings":"AAAA,OAAO,
|
|
1
|
+
{"version":3,"file":"cli.js","sourceRoot":"","sources":["../src/cli.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,EAAE,EAAE,MAAM,0BAA0B,CAAC;AAC9C,OAAO,EAAE,GAAG,EAAE,MAAM,cAAc,CAAC;AAEnC,OAAO,EAAE,mBAAmB,EAAE,SAAS,EAAE,MAAM,sBAAsB,CAAC;AACtE,OAAO,EAAE,qBAAqB,EAAE,WAAW,EAAE,MAAM,wBAAwB,CAAC;AAC5E,OAAO,EAAE,iBAAiB,EAAE,OAAO,EAAE,MAAM,oBAAoB,CAAC;AAEhE,MAAM,MAAM,GAAG,EAAE,CAAC,qBAAqB,EAAE,iBAAiB,EAAE,mBAAmB,CAAC,CAAC;AAEjF,MAAM,MAAM,GAAG,GAAG,CAAC,MAAM,EAAE,EAAE,WAAW,EAAE,SAAS,EAAE,IAAI,EAAE,MAAM,EAAE,CAAC,CAAC;AAErE,IAAI,MAAM,CAAC,IAAI,KAAK,UAAU,EAAE,CAAC;IAChC,MAAM,WAAW,CAAC,MAAM,CAAC,CAAC;AAC3B,CAAC;KAAM,IAAI,MAAM,CAAC,IAAI,KAAK,MAAM,EAAE,CAAC;IACnC,MAAM,OAAO,CAAC,MAAM,CAAC,CAAC;AACvB,CAAC;KAAM,IAAI,MAAM,CAAC,IAAI,KAAK,QAAQ,EAAE,CAAC;IACrC,MAAM,SAAS,CAAC,MAAM,CAAC,CAAC;AACzB,CAAC"}
|
package/dist/codegen.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"codegen.d.ts","sourceRoot":"","sources":["../src/codegen.ts"],"names":[],"mappings":"AAIA,OAAO,KAAK,EAEX,UAAU,EAWV,MAAM,qBAAqB,CAAC;AAE7B,MAAM,WAAW,UAAU;IAC1B,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;CACb;AAED,MAAM,WAAW,aAAa;IAC7B,IAAI,EAAE,MAAM,EAAE,CAAC;IACf,OAAO,EAAE,MAAM,GAAG,CAAC,CAAC,IAAI,EAAE,MAAM,KAAK;QAAE,IAAI,EAAE,OAAO,GAAG,WAAW,CAAC;QAAC,IAAI,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC;CACpF;AAED,MAAM,WAAW,iBAAiB;IACjC,SAAS,EAAE,UAAU,EAAE,CAAC;IACxB,QAAQ,EAAE,aAAa,EAAE,CAAC;IAC1B,OAAO,EAAE;QACR,YAAY,EAAE,MAAM,CAAC;KACrB,CAAC;IACF,QAAQ,EAAE;QACT,GAAG,EAAE,MAAM,CAAC;KACZ,CAAC;CACF;AAED,MAAM,WAAW,gBAAgB;IAChC,KAAK,EAAE,UAAU,EAAE,CAAC;CACpB;
|
|
1
|
+
{"version":3,"file":"codegen.d.ts","sourceRoot":"","sources":["../src/codegen.ts"],"names":[],"mappings":"AAIA,OAAO,KAAK,EAEX,UAAU,EAWV,MAAM,qBAAqB,CAAC;AAE7B,MAAM,WAAW,UAAU;IAC1B,QAAQ,EAAE,MAAM,CAAC;IACjB,IAAI,EAAE,MAAM,CAAC;CACb;AAED,MAAM,WAAW,aAAa;IAC7B,IAAI,EAAE,MAAM,EAAE,CAAC;IACf,OAAO,EAAE,MAAM,GAAG,CAAC,CAAC,IAAI,EAAE,MAAM,KAAK;QAAE,IAAI,EAAE,OAAO,GAAG,WAAW,CAAC;QAAC,IAAI,EAAE,MAAM,CAAA;KAAE,CAAC,CAAC;CACpF;AAED,MAAM,WAAW,iBAAiB;IACjC,SAAS,EAAE,UAAU,EAAE,CAAC;IACxB,QAAQ,EAAE,aAAa,EAAE,CAAC;IAC1B,OAAO,EAAE;QACR,YAAY,EAAE,MAAM,CAAC;KACrB,CAAC;IACF,QAAQ,EAAE;QACT,GAAG,EAAE,MAAM,CAAC;KACZ,CAAC;CACF;AAED,MAAM,WAAW,gBAAgB;IAChC,KAAK,EAAE,UAAU,EAAE,CAAC;CACpB;AAkDD,eAAO,MAAM,kBAAkB,wDAmT9B,CAAC"}
|