@atcute/lex-cli 2.2.2 → 2.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +47 -6
- package/dist/cli.js +152 -70
- package/dist/cli.js.map +1 -1
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js.map +1 -1
- package/dist/lexicon-metadata.d.ts +37 -0
- package/dist/lexicon-metadata.d.ts.map +1 -0
- package/dist/lexicon-metadata.js +42 -0
- package/dist/lexicon-metadata.js.map +1 -0
- package/package.json +10 -5
- package/schema/lexicon-package.schema.json +38 -0
- package/src/cli.ts +171 -77
- package/src/index.ts +1 -0
- package/src/lexicon-metadata.ts +59 -0
package/README.md
CHANGED
|
@@ -23,12 +23,38 @@ then run the tool:
|
|
|
23
23
|
npm exec lex-cli generate -c ./lex.config.js
|
|
24
24
|
```
|
|
25
25
|
|
|
26
|
-
|
|
26
|
+
## publishing your schemas
|
|
27
|
+
|
|
28
|
+
if you're packaging your generated schemas as a publishable library, add the `atcute:lexicons`
|
|
29
|
+
field to your package.json. this allows other projects to automatically discover and import your
|
|
30
|
+
schemas without manual configuration.
|
|
31
|
+
|
|
32
|
+
```json
|
|
33
|
+
{
|
|
34
|
+
"name": "@example/my-schemas",
|
|
35
|
+
"atcute:lexicons": {
|
|
36
|
+
"mappings": {
|
|
37
|
+
"com.example.*": {
|
|
38
|
+
"type": "namespace",
|
|
39
|
+
"path": "./types/{{nsid_remainder}}"
|
|
40
|
+
}
|
|
41
|
+
}
|
|
42
|
+
}
|
|
43
|
+
}
|
|
44
|
+
```
|
|
45
|
+
|
|
46
|
+
the `path` field supports several template expansions:
|
|
47
|
+
|
|
48
|
+
- `.` or `./` at the start is replaced with the package name (e.g., `./types/foo` becomes
|
|
49
|
+
`@example/my-schemas/types/foo`, or `.` becomes `@example/my-schemas`)
|
|
50
|
+
- `{{nsid}}` - the full NSID with dots replaced by slashes (e.g., `com/example/foo/bar`)
|
|
51
|
+
- `{{nsid_prefix}}` - the part before the wildcard (e.g., `com/example`)
|
|
52
|
+
- `{{nsid_remainder}}` - the part after the prefix (e.g., `foo/bar`)
|
|
27
53
|
|
|
28
54
|
## external references
|
|
29
55
|
|
|
30
56
|
when your lexicons reference types from namespaces outside your configured files, you'll need to
|
|
31
|
-
configure
|
|
57
|
+
configure how these references are resolved.
|
|
32
58
|
|
|
33
59
|
for example, if your lexicon references a type from another namespace:
|
|
34
60
|
|
|
@@ -54,7 +80,25 @@ for example, if your lexicon references a type from another namespace:
|
|
|
54
80
|
}
|
|
55
81
|
```
|
|
56
82
|
|
|
57
|
-
|
|
83
|
+
the simplest way to resolve external references is using the `imports` array with packages that
|
|
84
|
+
provide the `atcute:lexicons` metadata:
|
|
85
|
+
|
|
86
|
+
```ts
|
|
87
|
+
// file: lex.config.js
|
|
88
|
+
import { defineLexiconConfig } from '@atcute/lex-cli';
|
|
89
|
+
|
|
90
|
+
export default defineLexiconConfig({
|
|
91
|
+
files: ['lexicons/**/*.json'],
|
|
92
|
+
outdir: 'src/lexicons/',
|
|
93
|
+
imports: ['@atcute/atproto', '@atcute/bluesky'],
|
|
94
|
+
});
|
|
95
|
+
```
|
|
96
|
+
|
|
97
|
+
the CLI will automatically discover the namespace mappings from each package's `atcute:lexicons`
|
|
98
|
+
field in their package.json.
|
|
99
|
+
|
|
100
|
+
for packages without metadata, or when you need more fine-grained control over import resolution,
|
|
101
|
+
use the `mappings` configuration instead:
|
|
58
102
|
|
|
59
103
|
```ts
|
|
60
104
|
// file: lex.config.js
|
|
@@ -81,6 +125,3 @@ export default defineLexiconConfig({
|
|
|
81
125
|
],
|
|
82
126
|
});
|
|
83
127
|
```
|
|
84
|
-
|
|
85
|
-
with this configuration, any reference to a lexicon in the `com.atproto.*` or `app.bsky.*` namespace
|
|
86
|
-
will be imported from `@atcute/atproto` or `@atcute/bluesky`, respectively.
|
package/dist/cli.js
CHANGED
|
@@ -1,85 +1,167 @@
|
|
|
1
1
|
import * as fs from 'node:fs/promises';
|
|
2
2
|
import * as path from 'node:path';
|
|
3
3
|
import * as url from 'node:url';
|
|
4
|
-
import {
|
|
4
|
+
import { object } from '@optique/core/constructs';
|
|
5
|
+
import { command, constant, option } from '@optique/core/primitives';
|
|
6
|
+
import { run } from '@optique/run';
|
|
7
|
+
import { path as pathParser } from '@optique/run/valueparser';
|
|
5
8
|
import pc from 'picocolors';
|
|
6
9
|
import { lexiconDoc } from '@atcute/lexicon-doc';
|
|
7
10
|
import { generateLexiconApi } from './codegen.js';
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
const configDirname = path.dirname(configFilename);
|
|
22
|
-
let config;
|
|
23
|
-
try {
|
|
24
|
-
const configURL = url.pathToFileURL(configFilename);
|
|
25
|
-
const configMod = (await import(configURL.href));
|
|
26
|
-
config = configMod.default;
|
|
27
|
-
}
|
|
28
|
-
catch (err) {
|
|
29
|
-
console.error(pc.bold(pc.red(`failed to import config:`)));
|
|
30
|
-
console.error(err);
|
|
31
|
-
return 1;
|
|
32
|
-
}
|
|
33
|
-
const documents = [];
|
|
34
|
-
for await (const filename of fs.glob(config.files, { cwd: configDirname })) {
|
|
35
|
-
let source;
|
|
36
|
-
try {
|
|
37
|
-
source = await fs.readFile(path.join(configDirname, filename), 'utf8');
|
|
38
|
-
}
|
|
39
|
-
catch (err) {
|
|
40
|
-
console.error(pc.bold(pc.red(`file read error with "${filename}"`)));
|
|
41
|
-
console.error(err);
|
|
42
|
-
return 1;
|
|
43
|
-
}
|
|
44
|
-
let json;
|
|
11
|
+
import { validatePackageJson } from './lexicon-metadata.js';
|
|
12
|
+
/**
|
|
13
|
+
* Resolves package imports to ImportMapping[]
|
|
14
|
+
*/
|
|
15
|
+
const resolveImportsToMappings = async (imports, configDirname) => {
|
|
16
|
+
const mappings = [];
|
|
17
|
+
for (const packageName of imports) {
|
|
18
|
+
// Walk up from config directory to find package in node_modules
|
|
19
|
+
let packageJson;
|
|
20
|
+
let currentDir = configDirname;
|
|
21
|
+
let found = false;
|
|
22
|
+
while (currentDir !== path.dirname(currentDir)) {
|
|
23
|
+
const candidatePath = path.join(currentDir, 'node_modules', packageName, 'package.json');
|
|
45
24
|
try {
|
|
46
|
-
|
|
25
|
+
const content = await fs.readFile(candidatePath, 'utf8');
|
|
26
|
+
packageJson = JSON.parse(content);
|
|
27
|
+
found = true;
|
|
28
|
+
break;
|
|
47
29
|
}
|
|
48
30
|
catch (err) {
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
if (!result.ok) {
|
|
55
|
-
console.error(pc.bold(pc.red(`schema validation failed for "${filename}"`)));
|
|
56
|
-
console.error(result.message);
|
|
57
|
-
for (const issue of result.issues) {
|
|
58
|
-
console.log(`- ${issue.code} at .${issue.path.join('.')}`);
|
|
31
|
+
// Only continue to parent if file not found
|
|
32
|
+
if (err.code !== 'ENOENT') {
|
|
33
|
+
console.error(pc.bold(pc.red(`failed to read package.json for "${packageName}":`)));
|
|
34
|
+
console.error(err);
|
|
35
|
+
process.exit(1);
|
|
59
36
|
}
|
|
60
|
-
|
|
37
|
+
// Not found, try parent directory
|
|
38
|
+
currentDir = path.dirname(currentDir);
|
|
61
39
|
}
|
|
62
|
-
documents.push(result.value);
|
|
63
40
|
}
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
}
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
41
|
+
if (!found) {
|
|
42
|
+
console.error(pc.bold(pc.red(`failed to resolve package "${packageName}"`)));
|
|
43
|
+
console.error(`Could not find package in node_modules starting from ${configDirname}`);
|
|
44
|
+
process.exit(1);
|
|
45
|
+
}
|
|
46
|
+
// Validate package.json
|
|
47
|
+
const result = validatePackageJson(packageJson);
|
|
48
|
+
if (!result.success) {
|
|
49
|
+
console.error(pc.bold(pc.red(`invalid atcute:lexicons in "${packageName}":`)));
|
|
50
|
+
console.error(result.issues);
|
|
51
|
+
process.exit(1);
|
|
52
|
+
}
|
|
53
|
+
const lexicons = result.output['atcute:lexicons'];
|
|
54
|
+
if (!lexicons?.mappings) {
|
|
55
|
+
continue;
|
|
56
|
+
}
|
|
57
|
+
// Convert mapping to ImportMapping[]
|
|
58
|
+
for (const [pattern, entry] of Object.entries(lexicons.mappings)) {
|
|
59
|
+
const isWildcard = pattern.endsWith('.*');
|
|
60
|
+
mappings.push({
|
|
61
|
+
nsid: [pattern],
|
|
62
|
+
imports: (nsid) => {
|
|
63
|
+
// Check if pattern matches
|
|
64
|
+
if (isWildcard) {
|
|
65
|
+
if (!nsid.startsWith(pattern.slice(0, -1))) {
|
|
66
|
+
throw new Error(`NSID ${nsid} does not match pattern ${pattern}`);
|
|
67
|
+
}
|
|
68
|
+
}
|
|
69
|
+
else {
|
|
70
|
+
if (nsid !== pattern) {
|
|
71
|
+
throw new Error(`NSID ${nsid} does not match pattern ${pattern}`);
|
|
72
|
+
}
|
|
73
|
+
}
|
|
74
|
+
const nsidPrefix = isWildcard ? pattern.slice(0, -2) : pattern;
|
|
75
|
+
const nsidRemainder = isWildcard ? nsid.slice(nsidPrefix.length + 1) : '';
|
|
76
|
+
let expandedPath = entry.path
|
|
77
|
+
.replaceAll('{{nsid}}', nsid.replaceAll('.', '/'))
|
|
78
|
+
.replaceAll('{{nsid_remainder}}', nsidRemainder.replaceAll('.', '/'))
|
|
79
|
+
.replaceAll('{{nsid_prefix}}', nsidPrefix.replaceAll('.', '/'));
|
|
80
|
+
if (expandedPath === '.') {
|
|
81
|
+
expandedPath = packageName;
|
|
82
|
+
}
|
|
83
|
+
else if (expandedPath.startsWith('./')) {
|
|
84
|
+
expandedPath = `${packageName}/${expandedPath.slice(2)}`;
|
|
85
|
+
}
|
|
86
|
+
return {
|
|
87
|
+
type: entry.type,
|
|
88
|
+
from: expandedPath,
|
|
89
|
+
};
|
|
90
|
+
},
|
|
91
|
+
});
|
|
92
|
+
}
|
|
93
|
+
}
|
|
94
|
+
return mappings;
|
|
95
|
+
};
|
|
96
|
+
const parser = command('generate', object({
|
|
97
|
+
type: constant('generate'),
|
|
98
|
+
config: option('-c', '--config', pathParser({ metavar: 'CONFIG' })),
|
|
99
|
+
}));
|
|
100
|
+
const result = run(parser, { programName: 'lex-cli' });
|
|
101
|
+
if (result.type === 'generate') {
|
|
102
|
+
const configFilename = path.resolve(result.config);
|
|
103
|
+
const configDirname = path.dirname(configFilename);
|
|
104
|
+
let config;
|
|
105
|
+
try {
|
|
106
|
+
const configURL = url.pathToFileURL(configFilename);
|
|
107
|
+
const configMod = (await import(configURL.href));
|
|
108
|
+
config = configMod.default;
|
|
109
|
+
}
|
|
110
|
+
catch (err) {
|
|
111
|
+
console.error(pc.bold(pc.red(`failed to import config:`)));
|
|
112
|
+
console.error(err);
|
|
113
|
+
process.exit(1);
|
|
114
|
+
}
|
|
115
|
+
// Resolve imports to mappings
|
|
116
|
+
const importMappings = config.imports ? await resolveImportsToMappings(config.imports, configDirname) : [];
|
|
117
|
+
const allMappings = [...importMappings, ...(config.mappings ?? [])];
|
|
118
|
+
const documents = [];
|
|
119
|
+
for await (const filename of fs.glob(config.files, { cwd: configDirname })) {
|
|
120
|
+
let source;
|
|
121
|
+
try {
|
|
122
|
+
source = await fs.readFile(path.join(configDirname, filename), 'utf8');
|
|
123
|
+
}
|
|
124
|
+
catch (err) {
|
|
125
|
+
console.error(pc.bold(pc.red(`file read error with "${filename}"`)));
|
|
126
|
+
console.error(err);
|
|
127
|
+
process.exit(1);
|
|
128
|
+
}
|
|
129
|
+
let json;
|
|
130
|
+
try {
|
|
131
|
+
json = JSON.parse(source);
|
|
132
|
+
}
|
|
133
|
+
catch (err) {
|
|
134
|
+
console.error(pc.bold(pc.red(`json parse error in "${filename}"`)));
|
|
135
|
+
console.error(err);
|
|
136
|
+
process.exit(1);
|
|
80
137
|
}
|
|
138
|
+
const result = lexiconDoc.try(json, { mode: 'strip' });
|
|
139
|
+
if (!result.ok) {
|
|
140
|
+
console.error(pc.bold(pc.red(`schema validation failed for "${filename}"`)));
|
|
141
|
+
console.error(result.message);
|
|
142
|
+
for (const issue of result.issues) {
|
|
143
|
+
console.log(`- ${issue.code} at .${issue.path.join('.')}`);
|
|
144
|
+
}
|
|
145
|
+
process.exit(1);
|
|
146
|
+
}
|
|
147
|
+
documents.push(result.value);
|
|
148
|
+
}
|
|
149
|
+
const generationResult = await generateLexiconApi({
|
|
150
|
+
documents: documents,
|
|
151
|
+
mappings: allMappings,
|
|
152
|
+
modules: {
|
|
153
|
+
importSuffix: config.modules?.importSuffix ?? '.js',
|
|
154
|
+
},
|
|
155
|
+
prettier: {
|
|
156
|
+
cwd: process.cwd(),
|
|
157
|
+
},
|
|
158
|
+
});
|
|
159
|
+
const outdir = path.join(configDirname, config.outdir);
|
|
160
|
+
for (const file of generationResult.files) {
|
|
161
|
+
const filename = path.join(outdir, file.filename);
|
|
162
|
+
const dirname = path.dirname(filename);
|
|
163
|
+
await fs.mkdir(dirname, { recursive: true });
|
|
164
|
+
await fs.writeFile(filename, file.code);
|
|
81
165
|
}
|
|
82
|
-
}
|
|
83
|
-
const exitCode = await program.run(process.argv.slice(2));
|
|
84
|
-
process.exitCode = exitCode;
|
|
166
|
+
}
|
|
85
167
|
//# sourceMappingURL=cli.js.map
|
package/dist/cli.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"cli.js","sourceRoot":"","sources":["../src/cli.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,MAAM,kBAAkB,CAAC;AACvC,OAAO,KAAK,IAAI,MAAM,WAAW,CAAC;AAClC,OAAO,KAAK,GAAG,MAAM,UAAU,CAAC;AAEhC,OAAO,EAAE,
|
|
1
|
+
{"version":3,"file":"cli.js","sourceRoot":"","sources":["../src/cli.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,MAAM,kBAAkB,CAAC;AACvC,OAAO,KAAK,IAAI,MAAM,WAAW,CAAC;AAClC,OAAO,KAAK,GAAG,MAAM,UAAU,CAAC;AAEhC,OAAO,EAAE,MAAM,EAAE,MAAM,0BAA0B,CAAC;AAClD,OAAO,EAAE,OAAO,EAAE,QAAQ,EAAE,MAAM,EAAE,MAAM,0BAA0B,CAAC;AACrE,OAAO,EAAE,GAAG,EAAE,MAAM,cAAc,CAAC;AACnC,OAAO,EAAE,IAAI,IAAI,UAAU,EAAE,MAAM,0BAA0B,CAAC;AAC9D,OAAO,EAAE,MAAM,YAAY,CAAC;AAE5B,OAAO,EAAE,UAAU,EAAmB,MAAM,qBAAqB,CAAC;AAElE,OAAO,EAAE,kBAAkB,EAAsB,MAAM,cAAc,CAAC;AAEtE,OAAO,EAAE,mBAAmB,EAAE,MAAM,uBAAuB,CAAC;AAE5D;;GAEG;AACH,MAAM,wBAAwB,GAAG,KAAK,EACrC,OAAiB,EACjB,aAAqB,EACM,EAAE;IAC7B,MAAM,QAAQ,GAAoB,EAAE,CAAC;IAErC,KAAK,MAAM,WAAW,IAAI,OAAO,EAAE,CAAC;QACnC,gEAAgE;QAChE,IAAI,WAAoB,CAAC;QACzB,IAAI,UAAU,GAAG,aAAa,CAAC;QAC/B,IAAI,KAAK,GAAG,KAAK,CAAC;QAElB,OAAO,UAAU,KAAK,IAAI,CAAC,OAAO,CAAC,UAAU,CAAC,EAAE,CAAC;YAChD,MAAM,aAAa,GAAG,IAAI,CAAC,IAAI,CAAC,UAAU,EAAE,cAAc,EAAE,WAAW,EAAE,cAAc,CAAC,CAAC;YACzF,IAAI,CAAC;gBACJ,MAAM,OAAO,GAAG,MAAM,EAAE,CAAC,QAAQ,CAAC,aAAa,EAAE,MAAM,CAAC,CAAC;gBACzD,WAAW,GAAG,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC;gBAClC,KAAK,GAAG,IAAI,CAAC;gBACb,MAAM;YACP,CAAC;YAAC,OAAO,GAAQ,EAAE,CAAC;gBACnB,4CAA4C;gBAC5C,IAAI,GAAG,CAAC,IAAI,KAAK,QAAQ,EAAE,CAAC;oBAC3B,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,oCAAoC,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;oBACpF,OAAO,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC;oBACnB,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;gBACjB,CAAC;gBAED,kCAAkC;gBAClC,UAAU,GAAG,IAAI,CAAC,OAAO,CAAC,UAAU,CAAC,CAAC;YACvC,CAAC;QACF,CAAC;QAED,IAAI,CAAC,KAAK,EAAE,CAAC;YACZ,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,8BAA8B,WAAW,GAAG,CAAC,CAAC,CAAC,CAAC;YAC7E,OAAO,CAAC,KAAK,CAAC,wDAAwD,aAAa,EAAE,CAAC,CAAC;YACvF,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;QACjB,CAAC;QAED,wBAAwB;QACxB,MAAM,MAAM,GAAG,mBAAmB,CAAC,WAAW,CAAC,CAAC;QAChD,IAAI,CAAC,MAAM,CAAC,OAAO,EAAE,CAAC;YACrB,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,+BAA+B,WAAW,IAAI,CAAC,CAAC,CAAC,CAAC;YAC/E,OAAO,CAAC,KAAK,CAAC,MAAM,CAAC,MAAM,CAAC,CAAC;YAC7B,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;QACjB,CAAC;QAED,MAAM,QAAQ,GAAG,MAAM,CAAC,MAAM,CAAC,iBAAiB,CAAC,CAAC;QAClD,IAAI,CAAC,QAAQ,EAAE,QAAQ,EAAE,CAAC;YACzB,SAAS;QACV,CAAC;QAED,qCAAqC;QACrC,KAAK,MAAM,CAAC,OAAO,EAAE,KAAK,CAAC,IAAI,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,QAAQ,CAAC,EAAE,CAAC;YAClE,MAAM,UAAU,GAAG,OAAO,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC;YAE1C,QAAQ,CAAC,IAAI,CAAC;gBACb,IAAI,EAAE,CAAC,OAAO,CAAC;gBACf,OAAO,EAAE,CAAC,IAAY,EAAE,EAAE;oBACzB,2BAA2B;oBAC3B,IAAI,UAAU,EAAE,CAAC;wBAChB,IAAI,CAAC,IAAI,CAAC,UAAU,CAAC,OAAO,CAAC,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,CAAC,EAAE,CAAC;4BAC5C,MAAM,IAAI,KAAK,CAAC,QAAQ,IAAI,2BAA2B,OAAO,EAAE,CAAC,CAAC;wBACnE,CAAC;oBACF,CAAC;yBAAM,CAAC;wBACP,IAAI,IAAI,KAAK,OAAO,EAAE,CAAC;4BACtB,MAAM,IAAI,KAAK,CAAC,QAAQ,IAAI,2BAA2B,OAAO,EAAE,CAAC,CAAC;wBACnE,CAAC;oBACF,CAAC;oBAED,MAAM,UAAU,GAAG,UAAU,CAAC,CAAC,CAAC,OAAO,CAAC,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,OAAO,CAAC;oBAC/D,MAAM,aAAa,GAAG,UAAU,CAAC,CAAC,CAAC,IAAI,CAAC,KAAK,CAAC,UAAU,CAAC,MAAM,GAAG,CAAC,CAAC,CAAC,CAAC,CAAC,EAAE,CAAC;oBAE1E,IAAI,YAAY,GAAG,KAAK,CAAC,IAAI;yBAC3B,UAAU,CAAC,UAAU,EAAE,IAAI,CAAC,UAAU,CAAC,GAAG,EAAE,GAAG,CAAC,CAAC;yBACjD,UAAU,CAAC,oBAAoB,EAAE,aAAa,CAAC,UAAU,CAAC,GAAG,EAAE,GAAG,CAAC,CAAC;yBACpE,UAAU,CAAC,iBAAiB,EAAE,UAAU,CAAC,UAAU,CAAC,GAAG,EAAE,GAAG,CAAC,CAAC,CAAC;oBAEjE,IAAI,YAAY,KAAK,GAAG,EAAE,CAAC;wBAC1B,YAAY,GAAG,WAAW,CAAC;oBAC5B,CAAC;yBAAM,IAAI,YAAY,CAAC,UAAU,CAAC,IAAI,CAAC,EAAE,CAAC;wBAC1C,YAAY,GAAG,GAAG,WAAW,IAAI,YAAY,CAAC,KAAK,CAAC,CAAC,CAAC,EAAE,CAAC;oBAC1D,CAAC;oBAED,OAAO;wBACN,IAAI,EAAE,KAAK,CAAC,IAAI;wBAChB,IAAI,EAAE,YAAY;qBAClB,CAAC;gBACH,CAAC;aACD,CAAC,CAAC;QACJ,CAAC;IACF,CAAC;IAED,OAAO,QAAQ,CAAC;AACjB,CAAC,CAAC;AAEF,MAAM,MAAM,GAAG,OAAO,CACrB,UAAU,EACV,MAAM,CAAC;IACN,IAAI,EAAE,QAAQ,CAAC,UAAU,CAAC;IAC1B,MAAM,EAAE,MAAM,CAAC,IAAI,EAAE,UAAU,EAAE,UAAU,CAAC,EAAE,OAAO,EAAE,QAAQ,EAAE,CAAC,CAAC;CACnE,CAAC,CACF,CAAC;AAEF,MAAM,MAAM,GAAG,GAAG,CAAC,MAAM,EAAE,EAAE,WAAW,EAAE,SAAS,EAAE,CAAC,CAAC;AAEvD,IAAI,MAAM,CAAC,IAAI,KAAK,UAAU,EAAE,CAAC;IAChC,MAAM,cAAc,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,MAAM,CAAC,CAAC;IACnD,MAAM,aAAa,GAAG,IAAI,CAAC,OAAO,CAAC,cAAc,CAAC,CAAC;IAEnD,IAAI,MAAqB,CAAC;IAC1B,IAAI,CAAC;QACJ,MAAM,SAAS,GAAG,GAAG,CAAC,aAAa,CAAC,cAAc,CAAC,CAAC;QACpD,MAAM,SAAS,GAAG,CAAC,MAAM,MAAM,CAAC,SAAS,CAAC,IAAI,CAAC,CAA+B,CAAC;QAC/E,MAAM,GAAG,SAAS,CAAC,OAAO,CAAC;IAC5B,CAAC;IAAC,OAAO,GAAG,EAAE,CAAC;QACd,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,0BAA0B,CAAC,CAAC,CAAC,CAAC;QAC3D,OAAO,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC;QAEnB,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;IACjB,CAAC;IAED,8BAA8B;IAC9B,MAAM,cAAc,GAAG,MAAM,CAAC,OAAO,CAAC,CAAC,CAAC,MAAM,wBAAwB,CAAC,MAAM,CAAC,OAAO,EAAE,aAAa,CAAC,CAAC,CAAC,CAAC,EAAE,CAAC;IAC3G,MAAM,WAAW,GAAG,CAAC,GAAG,cAAc,EAAE,GAAG,CAAC,MAAM,CAAC,QAAQ,IAAI,EAAE,CAAC,CAAC,CAAC;IAEpE,MAAM,SAAS,GAAiB,EAAE,CAAC;IAEnC,IAAI,KAAK,EAAE,MAAM,QAAQ,IAAI,EAAE,CAAC,IAAI,CAAC,MAAM,CAAC,KAAK,EAAE,EAAE,GAAG,EAAE,aAAa,EAAE,CAAC,EAAE,CAAC;QAC5E,IAAI,MAAc,CAAC;QACnB,IAAI,CAAC;YACJ,MAAM,GAAG,MAAM,EAAE,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,aAAa,EAAE,QAAQ,CAAC,EAAE,MAAM,CAAC,CAAC;QACxE,CAAC;QAAC,OAAO,GAAG,EAAE,CAAC;YACd,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,yBAAyB,QAAQ,GAAG,CAAC,CAAC,CAAC,CAAC;YACrE,OAAO,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC;YAEnB,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;QACjB,CAAC;QAED,IAAI,IAAa,CAAC;QAClB,IAAI,CAAC;YACJ,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,MAAM,CAAC,CAAC;QAC3B,CAAC;QAAC,OAAO,GAAG,EAAE,CAAC;YACd,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,wBAAwB,QAAQ,GAAG,CAAC,CAAC,CAAC,CAAC;YACpE,OAAO,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC;YAEnB,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;QACjB,CAAC;QAED,MAAM,MAAM,GAAG,UAAU,CAAC,GAAG,CAAC,IAAI,EAAE,EAAE,IAAI,EAAE,OAAO,EAAE,CAAC,CAAC;QACvD,IAAI,CAAC,MAAM,CAAC,EAAE,EAAE,CAAC;YAChB,OAAO,CAAC,KAAK,CAAC,EAAE,CAAC,IAAI,CAAC,EAAE,CAAC,GAAG,CAAC,iCAAiC,QAAQ,GAAG,CAAC,CAAC,CAAC,CAAC;YAC7E,OAAO,CAAC,KAAK,CAAC,MAAM,CAAC,OAAO,CAAC,CAAC;YAE9B,KAAK,MAAM,KAAK,IAAI,MAAM,CAAC,MAAM,EAAE,CAAC;gBACnC,OAAO,CAAC,GAAG,CAAC,KAAK,KAAK,CAAC,IAAI,QAAQ,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,CAAC;YAC5D,CAAC;YAED,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC;QACjB,CAAC;QAED,SAAS,CAAC,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAC;IAC9B,CAAC;IAED,MAAM,gBAAgB,GAAG,MAAM,kBAAkB,CAAC;QACjD,SAAS,EAAE,SAAS;QACpB,QAAQ,EAAE,WAAW;QACrB,OAAO,EAAE;YACR,YAAY,EAAE,MAAM,CAAC,OAAO,EAAE,YAAY,IAAI,KAAK;SACnD;QACD,QAAQ,EAAE;YACT,GAAG,EAAE,OAAO,CAAC,GAAG,EAAE;SAClB;KACD,CAAC,CAAC;IAEH,MAAM,MAAM,GAAG,IAAI,CAAC,IAAI,CAAC,aAAa,EAAE,MAAM,CAAC,MAAM,CAAC,CAAC;IAEvD,KAAK,MAAM,IAAI,IAAI,gBAAgB,CAAC,KAAK,EAAE,CAAC;QAC3C,MAAM,QAAQ,GAAG,IAAI,CAAC,IAAI,CAAC,MAAM,EAAE,IAAI,CAAC,QAAQ,CAAC,CAAC;QAClD,MAAM,OAAO,GAAG,IAAI,CAAC,OAAO,CAAC,QAAQ,CAAC,CAAC;QAEvC,MAAM,EAAE,CAAC,KAAK,CAAC,OAAO,EAAE,EAAE,SAAS,EAAE,IAAI,EAAE,CAAC,CAAC;QAC7C,MAAM,EAAE,CAAC,SAAS,CAAC,QAAQ,EAAE,IAAI,CAAC,IAAI,CAAC,CAAC;IACzC,CAAC;AACF,CAAC"}
|
package/dist/index.d.ts
CHANGED
package/dist/index.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,cAAc,CAAC;AAElD,MAAM,WAAW,aAAa;IAC7B,MAAM,EAAE,MAAM,CAAC;IACf,KAAK,EAAE,MAAM,EAAE,CAAC;IAChB,QAAQ,CAAC,EAAE,aAAa,EAAE,CAAC;IAC3B,OAAO,CAAC,EAAE;QACT,YAAY,CAAC,EAAE,MAAM,CAAC;KACtB,CAAC;CACF;AAED,eAAO,MAAM,mBAAmB,GAAI,QAAQ,aAAa,KAAG,aAE3D,CAAC"}
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,cAAc,CAAC;AAElD,MAAM,WAAW,aAAa;IAC7B,MAAM,EAAE,MAAM,CAAC;IACf,KAAK,EAAE,MAAM,EAAE,CAAC;IAChB,OAAO,CAAC,EAAE,MAAM,EAAE,CAAC;IACnB,QAAQ,CAAC,EAAE,aAAa,EAAE,CAAC;IAC3B,OAAO,CAAC,EAAE;QACT,YAAY,CAAC,EAAE,MAAM,CAAC;KACtB,CAAC;CACF;AAED,eAAO,MAAM,mBAAmB,GAAI,QAAQ,aAAa,KAAG,aAE3D,CAAC"}
|
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAYA,MAAM,CAAC,MAAM,mBAAmB,GAAG,CAAC,MAAqB,EAAiB,EAAE;IAC3E,OAAO,MAAM,CAAC;AACf,CAAC,CAAC"}
|
|
@@ -0,0 +1,37 @@
|
|
|
1
|
+
import * as v from 'valibot';
|
|
2
|
+
/**
|
|
3
|
+
* Schema for a single lexicon mapping entry
|
|
4
|
+
*/
|
|
5
|
+
declare const lexiconMappingEntry: v.ObjectSchema<{
|
|
6
|
+
readonly type: v.PicklistSchema<["namespace", "named"], undefined>;
|
|
7
|
+
readonly path: v.SchemaWithPipe<readonly [v.StringSchema<undefined>, v.RegexAction<string, "path must be \".\" or start with \"./\"">]>;
|
|
8
|
+
}, undefined>;
|
|
9
|
+
/**
|
|
10
|
+
* Schema for the atcute:lexicons field in package.json
|
|
11
|
+
*/
|
|
12
|
+
declare const atcuteLexiconsField: v.ObjectSchema<{
|
|
13
|
+
readonly mappings: v.OptionalSchema<v.RecordSchema<v.SchemaWithPipe<readonly [v.StringSchema<undefined>, v.CheckAction<string, "invalid NSID pattern (must be valid NSID or end with .*)">]>, v.ObjectSchema<{
|
|
14
|
+
readonly type: v.PicklistSchema<["namespace", "named"], undefined>;
|
|
15
|
+
readonly path: v.SchemaWithPipe<readonly [v.StringSchema<undefined>, v.RegexAction<string, "path must be \".\" or start with \"./\"">]>;
|
|
16
|
+
}, undefined>, undefined>, undefined>;
|
|
17
|
+
}, undefined>;
|
|
18
|
+
/**
|
|
19
|
+
* Schema for package.json with atcute:lexicons field
|
|
20
|
+
*/
|
|
21
|
+
export declare const packageJsonSchema: v.LooseObjectSchema<{
|
|
22
|
+
readonly 'atcute:lexicons': v.OptionalSchema<v.ObjectSchema<{
|
|
23
|
+
readonly mappings: v.OptionalSchema<v.RecordSchema<v.SchemaWithPipe<readonly [v.StringSchema<undefined>, v.CheckAction<string, "invalid NSID pattern (must be valid NSID or end with .*)">]>, v.ObjectSchema<{
|
|
24
|
+
readonly type: v.PicklistSchema<["namespace", "named"], undefined>;
|
|
25
|
+
readonly path: v.SchemaWithPipe<readonly [v.StringSchema<undefined>, v.RegexAction<string, "path must be \".\" or start with \"./\"">]>;
|
|
26
|
+
}, undefined>, undefined>, undefined>;
|
|
27
|
+
}, undefined>, undefined>;
|
|
28
|
+
}, undefined>;
|
|
29
|
+
export type LexiconMappingEntry = v.InferOutput<typeof lexiconMappingEntry>;
|
|
30
|
+
export type AtcuteLexiconsField = v.InferOutput<typeof atcuteLexiconsField>;
|
|
31
|
+
export type PackageJsonWithLexicons = v.InferOutput<typeof packageJsonSchema>;
|
|
32
|
+
/**
|
|
33
|
+
* Validates a package.json object against the schema
|
|
34
|
+
*/
|
|
35
|
+
export declare const validatePackageJson: (data: unknown) => v.SafeParseResult<typeof packageJsonSchema>;
|
|
36
|
+
export {};
|
|
37
|
+
//# sourceMappingURL=lexicon-metadata.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"lexicon-metadata.d.ts","sourceRoot":"","sources":["../src/lexicon-metadata.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,CAAC,MAAM,SAAS,CAAC;AAmB7B;;GAEG;AACH,QAAA,MAAM,mBAAmB;;;aAGvB,CAAC;AAEH;;GAEG;AACH,QAAA,MAAM,mBAAmB;;;;;aAUvB,CAAC;AAEH;;GAEG;AACH,eAAO,MAAM,iBAAiB;;;;;;;aAE5B,CAAC;AAEH,MAAM,MAAM,mBAAmB,GAAG,CAAC,CAAC,WAAW,CAAC,OAAO,mBAAmB,CAAC,CAAC;AAC5E,MAAM,MAAM,mBAAmB,GAAG,CAAC,CAAC,WAAW,CAAC,OAAO,mBAAmB,CAAC,CAAC;AAC5E,MAAM,MAAM,uBAAuB,GAAG,CAAC,CAAC,WAAW,CAAC,OAAO,iBAAiB,CAAC,CAAC;AAE9E;;GAEG;AACH,eAAO,MAAM,mBAAmB,GAAI,MAAM,OAAO,KAAG,CAAC,CAAC,eAAe,CAAC,OAAO,iBAAiB,CAE7F,CAAC"}
|
|
@@ -0,0 +1,42 @@
|
|
|
1
|
+
import * as v from 'valibot';
|
|
2
|
+
import { isNsid } from '@atcute/lexicons/syntax';
|
|
3
|
+
/**
|
|
4
|
+
* Validates if a string is a valid NSID pattern (exact or wildcard)
|
|
5
|
+
* - Exact: "com.atproto.repo.getRecord"
|
|
6
|
+
* - Wildcard: "com.atproto.*"
|
|
7
|
+
*/
|
|
8
|
+
const isValidLexiconPattern = (pattern) => {
|
|
9
|
+
if (pattern.endsWith('.*')) {
|
|
10
|
+
// For wildcards, remove the .* and validate the prefix as an NSID segment
|
|
11
|
+
const prefix = pattern.slice(0, -2);
|
|
12
|
+
// Add a dummy segment to make it a valid NSID for validation
|
|
13
|
+
return isNsid(prefix + '.x');
|
|
14
|
+
}
|
|
15
|
+
return isNsid(pattern);
|
|
16
|
+
};
|
|
17
|
+
/**
|
|
18
|
+
* Schema for a single lexicon mapping entry
|
|
19
|
+
*/
|
|
20
|
+
const lexiconMappingEntry = v.object({
|
|
21
|
+
type: v.picklist(['namespace', 'named']),
|
|
22
|
+
path: v.pipe(v.string(), v.regex(/^\.$|^\.\//, `path must be "." or start with "./"`)),
|
|
23
|
+
});
|
|
24
|
+
/**
|
|
25
|
+
* Schema for the atcute:lexicons field in package.json
|
|
26
|
+
*/
|
|
27
|
+
const atcuteLexiconsField = v.object({
|
|
28
|
+
mappings: v.optional(v.record(v.pipe(v.string(), v.check(isValidLexiconPattern, `invalid NSID pattern (must be valid NSID or end with .*)`)), lexiconMappingEntry)),
|
|
29
|
+
});
|
|
30
|
+
/**
|
|
31
|
+
* Schema for package.json with atcute:lexicons field
|
|
32
|
+
*/
|
|
33
|
+
export const packageJsonSchema = v.looseObject({
|
|
34
|
+
'atcute:lexicons': v.optional(atcuteLexiconsField),
|
|
35
|
+
});
|
|
36
|
+
/**
|
|
37
|
+
* Validates a package.json object against the schema
|
|
38
|
+
*/
|
|
39
|
+
export const validatePackageJson = (data) => {
|
|
40
|
+
return v.safeParse(packageJsonSchema, data);
|
|
41
|
+
};
|
|
42
|
+
//# sourceMappingURL=lexicon-metadata.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"lexicon-metadata.js","sourceRoot":"","sources":["../src/lexicon-metadata.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,CAAC,MAAM,SAAS,CAAC;AAE7B,OAAO,EAAE,MAAM,EAAE,MAAM,yBAAyB,CAAC;AAEjD;;;;GAIG;AACH,MAAM,qBAAqB,GAAG,CAAC,OAAe,EAAW,EAAE;IAC1D,IAAI,OAAO,CAAC,QAAQ,CAAC,IAAI,CAAC,EAAE,CAAC;QAC5B,0EAA0E;QAC1E,MAAM,MAAM,GAAG,OAAO,CAAC,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,CAAC;QACpC,6DAA6D;QAC7D,OAAO,MAAM,CAAC,MAAM,GAAG,IAAI,CAAC,CAAC;IAC9B,CAAC;IACD,OAAO,MAAM,CAAC,OAAO,CAAC,CAAC;AACxB,CAAC,CAAC;AAEF;;GAEG;AACH,MAAM,mBAAmB,GAAG,CAAC,CAAC,MAAM,CAAC;IACpC,IAAI,EAAE,CAAC,CAAC,QAAQ,CAAC,CAAC,WAAW,EAAE,OAAO,CAAC,CAAC;IACxC,IAAI,EAAE,CAAC,CAAC,IAAI,CAAC,CAAC,CAAC,MAAM,EAAE,EAAE,CAAC,CAAC,KAAK,CAAC,YAAY,EAAE,qCAAqC,CAAC,CAAC;CACtF,CAAC,CAAC;AAEH;;GAEG;AACH,MAAM,mBAAmB,GAAG,CAAC,CAAC,MAAM,CAAC;IACpC,QAAQ,EAAE,CAAC,CAAC,QAAQ,CACnB,CAAC,CAAC,MAAM,CACP,CAAC,CAAC,IAAI,CACL,CAAC,CAAC,MAAM,EAAE,EACV,CAAC,CAAC,KAAK,CAAC,qBAAqB,EAAE,0DAA0D,CAAC,CAC1F,EACD,mBAAmB,CACnB,CACD;CACD,CAAC,CAAC;AAEH;;GAEG;AACH,MAAM,CAAC,MAAM,iBAAiB,GAAG,CAAC,CAAC,WAAW,CAAC;IAC9C,iBAAiB,EAAE,CAAC,CAAC,QAAQ,CAAC,mBAAmB,CAAC;CAClD,CAAC,CAAC;AAMH;;GAEG;AACH,MAAM,CAAC,MAAM,mBAAmB,GAAG,CAAC,IAAa,EAA+C,EAAE;IACjG,OAAO,CAAC,CAAC,SAAS,CAAC,iBAAiB,EAAE,IAAI,CAAC,CAAC;AAC7C,CAAC,CAAC"}
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"type": "module",
|
|
3
3
|
"name": "@atcute/lex-cli",
|
|
4
|
-
"version": "2.
|
|
4
|
+
"version": "2.3.0",
|
|
5
5
|
"description": "cli tool to generate type definitions for atcute",
|
|
6
6
|
"license": "0BSD",
|
|
7
7
|
"repository": {
|
|
@@ -13,7 +13,8 @@
|
|
|
13
13
|
"src/",
|
|
14
14
|
"!src/**/*.bench.ts",
|
|
15
15
|
"!src/**/*.test.ts",
|
|
16
|
-
"cli.mjs"
|
|
16
|
+
"cli.mjs",
|
|
17
|
+
"schema/"
|
|
17
18
|
],
|
|
18
19
|
"bin": "./cli.mjs",
|
|
19
20
|
"exports": {
|
|
@@ -21,17 +22,21 @@
|
|
|
21
22
|
},
|
|
22
23
|
"dependencies": {
|
|
23
24
|
"@badrap/valita": "^0.4.6",
|
|
24
|
-
"@
|
|
25
|
+
"@optique/core": "^0.6.1",
|
|
26
|
+
"@optique/run": "^0.6.1",
|
|
25
27
|
"picocolors": "^1.1.1",
|
|
26
28
|
"prettier": "^3.6.2",
|
|
27
|
-
"
|
|
29
|
+
"valibot": "^1.0.0",
|
|
30
|
+
"@atcute/lexicon-doc": "^1.1.3"
|
|
28
31
|
},
|
|
29
32
|
"devDependencies": {
|
|
30
33
|
"@types/node": "^22.18.0",
|
|
34
|
+
"@valibot/to-json-schema": "^1.3.0",
|
|
31
35
|
"@atcute/lexicons": "^1.2.2"
|
|
32
36
|
},
|
|
33
37
|
"scripts": {
|
|
34
|
-
"build": "tsc",
|
|
38
|
+
"build": "pnpm run generate:schema && tsc",
|
|
39
|
+
"generate:schema": "node scripts/generate-schema.ts",
|
|
35
40
|
"prepublish": "rm -rf dist; pnpm run build"
|
|
36
41
|
}
|
|
37
42
|
}
|
|
@@ -0,0 +1,38 @@
|
|
|
1
|
+
{
|
|
2
|
+
"$schema": "http://json-schema.org/draft-07/schema#",
|
|
3
|
+
"$id": "https://unpkg.com/@atcute/lex-cli/schema/lexicon-package.schema.json",
|
|
4
|
+
"title": "package.json with atcute:lexicons",
|
|
5
|
+
"description": "JSON Schema for package.json with atcute:lexicons field for lexicon import mappings",
|
|
6
|
+
"type": "object",
|
|
7
|
+
"properties": {
|
|
8
|
+
"atcute:lexicons": {
|
|
9
|
+
"type": "object",
|
|
10
|
+
"properties": {
|
|
11
|
+
"mappings": {
|
|
12
|
+
"type": "object",
|
|
13
|
+
"additionalProperties": {
|
|
14
|
+
"type": "object",
|
|
15
|
+
"properties": {
|
|
16
|
+
"type": {
|
|
17
|
+
"enum": [
|
|
18
|
+
"namespace",
|
|
19
|
+
"named"
|
|
20
|
+
]
|
|
21
|
+
},
|
|
22
|
+
"path": {
|
|
23
|
+
"type": "string",
|
|
24
|
+
"pattern": "^\\.$|^\\.\\/"
|
|
25
|
+
}
|
|
26
|
+
},
|
|
27
|
+
"required": [
|
|
28
|
+
"type",
|
|
29
|
+
"path"
|
|
30
|
+
]
|
|
31
|
+
}
|
|
32
|
+
}
|
|
33
|
+
},
|
|
34
|
+
"required": []
|
|
35
|
+
}
|
|
36
|
+
},
|
|
37
|
+
"required": []
|
|
38
|
+
}
|
package/src/cli.ts
CHANGED
|
@@ -2,108 +2,202 @@ import * as fs from 'node:fs/promises';
|
|
|
2
2
|
import * as path from 'node:path';
|
|
3
3
|
import * as url from 'node:url';
|
|
4
4
|
|
|
5
|
-
import {
|
|
5
|
+
import { object } from '@optique/core/constructs';
|
|
6
|
+
import { command, constant, option } from '@optique/core/primitives';
|
|
7
|
+
import { run } from '@optique/run';
|
|
8
|
+
import { path as pathParser } from '@optique/run/valueparser';
|
|
6
9
|
import pc from 'picocolors';
|
|
7
10
|
|
|
8
11
|
import { lexiconDoc, type LexiconDoc } from '@atcute/lexicon-doc';
|
|
9
12
|
|
|
10
|
-
import { generateLexiconApi } from './codegen.js';
|
|
13
|
+
import { generateLexiconApi, type ImportMapping } from './codegen.js';
|
|
11
14
|
import type { LexiconConfig } from './index.js';
|
|
15
|
+
import { validatePackageJson } from './lexicon-metadata.js';
|
|
16
|
+
|
|
17
|
+
/**
|
|
18
|
+
* Resolves package imports to ImportMapping[]
|
|
19
|
+
*/
|
|
20
|
+
const resolveImportsToMappings = async (
|
|
21
|
+
imports: string[],
|
|
22
|
+
configDirname: string,
|
|
23
|
+
): Promise<ImportMapping[]> => {
|
|
24
|
+
const mappings: ImportMapping[] = [];
|
|
25
|
+
|
|
26
|
+
for (const packageName of imports) {
|
|
27
|
+
// Walk up from config directory to find package in node_modules
|
|
28
|
+
let packageJson: unknown;
|
|
29
|
+
let currentDir = configDirname;
|
|
30
|
+
let found = false;
|
|
31
|
+
|
|
32
|
+
while (currentDir !== path.dirname(currentDir)) {
|
|
33
|
+
const candidatePath = path.join(currentDir, 'node_modules', packageName, 'package.json');
|
|
34
|
+
try {
|
|
35
|
+
const content = await fs.readFile(candidatePath, 'utf8');
|
|
36
|
+
packageJson = JSON.parse(content);
|
|
37
|
+
found = true;
|
|
38
|
+
break;
|
|
39
|
+
} catch (err: any) {
|
|
40
|
+
// Only continue to parent if file not found
|
|
41
|
+
if (err.code !== 'ENOENT') {
|
|
42
|
+
console.error(pc.bold(pc.red(`failed to read package.json for "${packageName}":`)));
|
|
43
|
+
console.error(err);
|
|
44
|
+
process.exit(1);
|
|
45
|
+
}
|
|
12
46
|
|
|
13
|
-
|
|
47
|
+
// Not found, try parent directory
|
|
48
|
+
currentDir = path.dirname(currentDir);
|
|
49
|
+
}
|
|
50
|
+
}
|
|
14
51
|
|
|
15
|
-
|
|
52
|
+
if (!found) {
|
|
53
|
+
console.error(pc.bold(pc.red(`failed to resolve package "${packageName}"`)));
|
|
54
|
+
console.error(`Could not find package in node_modules starting from ${configDirname}`);
|
|
55
|
+
process.exit(1);
|
|
56
|
+
}
|
|
16
57
|
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
58
|
+
// Validate package.json
|
|
59
|
+
const result = validatePackageJson(packageJson);
|
|
60
|
+
if (!result.success) {
|
|
61
|
+
console.error(pc.bold(pc.red(`invalid atcute:lexicons in "${packageName}":`)));
|
|
62
|
+
console.error(result.issues);
|
|
63
|
+
process.exit(1);
|
|
64
|
+
}
|
|
20
65
|
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
66
|
+
const lexicons = result.output['atcute:lexicons'];
|
|
67
|
+
if (!lexicons?.mappings) {
|
|
68
|
+
continue;
|
|
69
|
+
}
|
|
24
70
|
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
71
|
+
// Convert mapping to ImportMapping[]
|
|
72
|
+
for (const [pattern, entry] of Object.entries(lexicons.mappings)) {
|
|
73
|
+
const isWildcard = pattern.endsWith('.*');
|
|
74
|
+
|
|
75
|
+
mappings.push({
|
|
76
|
+
nsid: [pattern],
|
|
77
|
+
imports: (nsid: string) => {
|
|
78
|
+
// Check if pattern matches
|
|
79
|
+
if (isWildcard) {
|
|
80
|
+
if (!nsid.startsWith(pattern.slice(0, -1))) {
|
|
81
|
+
throw new Error(`NSID ${nsid} does not match pattern ${pattern}`);
|
|
82
|
+
}
|
|
83
|
+
} else {
|
|
84
|
+
if (nsid !== pattern) {
|
|
85
|
+
throw new Error(`NSID ${nsid} does not match pattern ${pattern}`);
|
|
86
|
+
}
|
|
87
|
+
}
|
|
29
88
|
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
const configDirname = path.dirname(configFilename);
|
|
89
|
+
const nsidPrefix = isWildcard ? pattern.slice(0, -2) : pattern;
|
|
90
|
+
const nsidRemainder = isWildcard ? nsid.slice(nsidPrefix.length + 1) : '';
|
|
33
91
|
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
config = configMod.default;
|
|
39
|
-
} catch (err) {
|
|
40
|
-
console.error(pc.bold(pc.red(`failed to import config:`)));
|
|
41
|
-
console.error(err);
|
|
42
|
-
|
|
43
|
-
return 1;
|
|
44
|
-
}
|
|
92
|
+
let expandedPath = entry.path
|
|
93
|
+
.replaceAll('{{nsid}}', nsid.replaceAll('.', '/'))
|
|
94
|
+
.replaceAll('{{nsid_remainder}}', nsidRemainder.replaceAll('.', '/'))
|
|
95
|
+
.replaceAll('{{nsid_prefix}}', nsidPrefix.replaceAll('.', '/'));
|
|
45
96
|
|
|
46
|
-
|
|
97
|
+
if (expandedPath === '.') {
|
|
98
|
+
expandedPath = packageName;
|
|
99
|
+
} else if (expandedPath.startsWith('./')) {
|
|
100
|
+
expandedPath = `${packageName}/${expandedPath.slice(2)}`;
|
|
101
|
+
}
|
|
47
102
|
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
}
|
|
53
|
-
|
|
54
|
-
|
|
103
|
+
return {
|
|
104
|
+
type: entry.type,
|
|
105
|
+
from: expandedPath,
|
|
106
|
+
};
|
|
107
|
+
},
|
|
108
|
+
});
|
|
109
|
+
}
|
|
110
|
+
}
|
|
55
111
|
|
|
56
|
-
|
|
57
|
-
|
|
112
|
+
return mappings;
|
|
113
|
+
};
|
|
58
114
|
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
115
|
+
const parser = command(
|
|
116
|
+
'generate',
|
|
117
|
+
object({
|
|
118
|
+
type: constant('generate'),
|
|
119
|
+
config: option('-c', '--config', pathParser({ metavar: 'CONFIG' })),
|
|
120
|
+
}),
|
|
121
|
+
);
|
|
65
122
|
|
|
66
|
-
|
|
67
|
-
}
|
|
123
|
+
const result = run(parser, { programName: 'lex-cli' });
|
|
68
124
|
|
|
69
|
-
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
console.error(result.message);
|
|
125
|
+
if (result.type === 'generate') {
|
|
126
|
+
const configFilename = path.resolve(result.config);
|
|
127
|
+
const configDirname = path.dirname(configFilename);
|
|
73
128
|
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
129
|
+
let config: LexiconConfig;
|
|
130
|
+
try {
|
|
131
|
+
const configURL = url.pathToFileURL(configFilename);
|
|
132
|
+
const configMod = (await import(configURL.href)) as { default: LexiconConfig };
|
|
133
|
+
config = configMod.default;
|
|
134
|
+
} catch (err) {
|
|
135
|
+
console.error(pc.bold(pc.red(`failed to import config:`)));
|
|
136
|
+
console.error(err);
|
|
77
137
|
|
|
78
|
-
|
|
79
|
-
|
|
138
|
+
process.exit(1);
|
|
139
|
+
}
|
|
80
140
|
|
|
81
|
-
|
|
82
|
-
|
|
141
|
+
// Resolve imports to mappings
|
|
142
|
+
const importMappings = config.imports ? await resolveImportsToMappings(config.imports, configDirname) : [];
|
|
143
|
+
const allMappings = [...importMappings, ...(config.mappings ?? [])];
|
|
83
144
|
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
145
|
+
const documents: LexiconDoc[] = [];
|
|
146
|
+
|
|
147
|
+
for await (const filename of fs.glob(config.files, { cwd: configDirname })) {
|
|
148
|
+
let source: string;
|
|
149
|
+
try {
|
|
150
|
+
source = await fs.readFile(path.join(configDirname, filename), 'utf8');
|
|
151
|
+
} catch (err) {
|
|
152
|
+
console.error(pc.bold(pc.red(`file read error with "${filename}"`)));
|
|
153
|
+
console.error(err);
|
|
154
|
+
|
|
155
|
+
process.exit(1);
|
|
156
|
+
}
|
|
94
157
|
|
|
95
|
-
|
|
158
|
+
let json: unknown;
|
|
159
|
+
try {
|
|
160
|
+
json = JSON.parse(source);
|
|
161
|
+
} catch (err) {
|
|
162
|
+
console.error(pc.bold(pc.red(`json parse error in "${filename}"`)));
|
|
163
|
+
console.error(err);
|
|
96
164
|
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
165
|
+
process.exit(1);
|
|
166
|
+
}
|
|
167
|
+
|
|
168
|
+
const result = lexiconDoc.try(json, { mode: 'strip' });
|
|
169
|
+
if (!result.ok) {
|
|
170
|
+
console.error(pc.bold(pc.red(`schema validation failed for "${filename}"`)));
|
|
171
|
+
console.error(result.message);
|
|
100
172
|
|
|
101
|
-
|
|
102
|
-
|
|
173
|
+
for (const issue of result.issues) {
|
|
174
|
+
console.log(`- ${issue.code} at .${issue.path.join('.')}`);
|
|
103
175
|
}
|
|
176
|
+
|
|
177
|
+
process.exit(1);
|
|
104
178
|
}
|
|
105
|
-
},
|
|
106
|
-
);
|
|
107
179
|
|
|
108
|
-
|
|
109
|
-
|
|
180
|
+
documents.push(result.value);
|
|
181
|
+
}
|
|
182
|
+
|
|
183
|
+
const generationResult = await generateLexiconApi({
|
|
184
|
+
documents: documents,
|
|
185
|
+
mappings: allMappings,
|
|
186
|
+
modules: {
|
|
187
|
+
importSuffix: config.modules?.importSuffix ?? '.js',
|
|
188
|
+
},
|
|
189
|
+
prettier: {
|
|
190
|
+
cwd: process.cwd(),
|
|
191
|
+
},
|
|
192
|
+
});
|
|
193
|
+
|
|
194
|
+
const outdir = path.join(configDirname, config.outdir);
|
|
195
|
+
|
|
196
|
+
for (const file of generationResult.files) {
|
|
197
|
+
const filename = path.join(outdir, file.filename);
|
|
198
|
+
const dirname = path.dirname(filename);
|
|
199
|
+
|
|
200
|
+
await fs.mkdir(dirname, { recursive: true });
|
|
201
|
+
await fs.writeFile(filename, file.code);
|
|
202
|
+
}
|
|
203
|
+
}
|
package/src/index.ts
CHANGED
|
@@ -0,0 +1,59 @@
|
|
|
1
|
+
import * as v from 'valibot';
|
|
2
|
+
|
|
3
|
+
import { isNsid } from '@atcute/lexicons/syntax';
|
|
4
|
+
|
|
5
|
+
/**
|
|
6
|
+
* Validates if a string is a valid NSID pattern (exact or wildcard)
|
|
7
|
+
* - Exact: "com.atproto.repo.getRecord"
|
|
8
|
+
* - Wildcard: "com.atproto.*"
|
|
9
|
+
*/
|
|
10
|
+
const isValidLexiconPattern = (pattern: string): boolean => {
|
|
11
|
+
if (pattern.endsWith('.*')) {
|
|
12
|
+
// For wildcards, remove the .* and validate the prefix as an NSID segment
|
|
13
|
+
const prefix = pattern.slice(0, -2);
|
|
14
|
+
// Add a dummy segment to make it a valid NSID for validation
|
|
15
|
+
return isNsid(prefix + '.x');
|
|
16
|
+
}
|
|
17
|
+
return isNsid(pattern);
|
|
18
|
+
};
|
|
19
|
+
|
|
20
|
+
/**
|
|
21
|
+
* Schema for a single lexicon mapping entry
|
|
22
|
+
*/
|
|
23
|
+
const lexiconMappingEntry = v.object({
|
|
24
|
+
type: v.picklist(['namespace', 'named']),
|
|
25
|
+
path: v.pipe(v.string(), v.regex(/^\.$|^\.\//, `path must be "." or start with "./"`)),
|
|
26
|
+
});
|
|
27
|
+
|
|
28
|
+
/**
|
|
29
|
+
* Schema for the atcute:lexicons field in package.json
|
|
30
|
+
*/
|
|
31
|
+
const atcuteLexiconsField = v.object({
|
|
32
|
+
mappings: v.optional(
|
|
33
|
+
v.record(
|
|
34
|
+
v.pipe(
|
|
35
|
+
v.string(),
|
|
36
|
+
v.check(isValidLexiconPattern, `invalid NSID pattern (must be valid NSID or end with .*)`),
|
|
37
|
+
),
|
|
38
|
+
lexiconMappingEntry,
|
|
39
|
+
),
|
|
40
|
+
),
|
|
41
|
+
});
|
|
42
|
+
|
|
43
|
+
/**
|
|
44
|
+
* Schema for package.json with atcute:lexicons field
|
|
45
|
+
*/
|
|
46
|
+
export const packageJsonSchema = v.looseObject({
|
|
47
|
+
'atcute:lexicons': v.optional(atcuteLexiconsField),
|
|
48
|
+
});
|
|
49
|
+
|
|
50
|
+
export type LexiconMappingEntry = v.InferOutput<typeof lexiconMappingEntry>;
|
|
51
|
+
export type AtcuteLexiconsField = v.InferOutput<typeof atcuteLexiconsField>;
|
|
52
|
+
export type PackageJsonWithLexicons = v.InferOutput<typeof packageJsonSchema>;
|
|
53
|
+
|
|
54
|
+
/**
|
|
55
|
+
* Validates a package.json object against the schema
|
|
56
|
+
*/
|
|
57
|
+
export const validatePackageJson = (data: unknown): v.SafeParseResult<typeof packageJsonSchema> => {
|
|
58
|
+
return v.safeParse(packageJsonSchema, data);
|
|
59
|
+
};
|