@storm-software/terraform-tools 0.61.1 → 0.62.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +9 -0
- package/README.md +1 -1
- package/dist/{chunk-VWGJH5XZ.mjs → chunk-4UUC5PL5.mjs} +1 -1
- package/dist/{chunk-LHLWLQYI.mjs → chunk-ASN7N3MD.mjs} +1 -1
- package/dist/{chunk-RIOD6BBP.js → chunk-C3TATPNG.js} +2 -2
- package/dist/{chunk-HJPLYKIK.js → chunk-EBBFG7FE.js} +2 -2
- package/dist/{chunk-OTXSJUFN.js → chunk-EDARDEZL.js} +2 -2
- package/dist/{chunk-PTGHK4O2.mjs → chunk-LDL6GDGI.mjs} +1 -1
- package/dist/{chunk-W3KFGH6F.mjs → chunk-LLEL65DB.mjs} +1 -1
- package/dist/{chunk-IRQ6ATAT.js → chunk-O6ULIZTX.js} +2 -2
- package/dist/{chunk-A3AZGCZW.mjs → chunk-OW3Z6C4U.mjs} +6 -4
- package/dist/{chunk-XLCDBNTQ.js → chunk-OZAMHBE3.js} +2 -2
- package/dist/{chunk-3HCCGFVA.js → chunk-P7TIR2I6.js} +552 -249
- package/dist/{chunk-QEMBK6MY.mjs → chunk-VD7U2ISL.mjs} +552 -249
- package/dist/{chunk-TTTXG2FP.mjs → chunk-VM73HSDX.mjs} +1 -1
- package/dist/{chunk-QDREZDRA.js → chunk-YAQKI2DU.js} +110 -108
- package/dist/executors.js +6 -6
- package/dist/executors.mjs +6 -6
- package/dist/generators.js +3 -3
- package/dist/generators.mjs +2 -2
- package/dist/index.js +8 -8
- package/dist/index.mjs +7 -7
- package/dist/src/base/index.js +3 -3
- package/dist/src/base/index.mjs +2 -2
- package/dist/src/base/terraform-executor.js +3 -3
- package/dist/src/base/terraform-executor.mjs +2 -2
- package/dist/src/executors/apply/executor.js +4 -4
- package/dist/src/executors/apply/executor.mjs +3 -3
- package/dist/src/executors/destroy/executor.js +4 -4
- package/dist/src/executors/destroy/executor.mjs +3 -3
- package/dist/src/executors/output/executor.js +4 -4
- package/dist/src/executors/output/executor.mjs +3 -3
- package/dist/src/executors/plan/executor.js +4 -4
- package/dist/src/executors/plan/executor.mjs +3 -3
- package/dist/src/generators/init/init.js +3 -3
- package/dist/src/generators/init/init.mjs +2 -2
- package/package.json +2 -2
|
@@ -16,7 +16,8 @@
|
|
|
16
16
|
|
|
17
17
|
|
|
18
18
|
|
|
19
|
-
|
|
19
|
+
|
|
20
|
+
var _chunkP7TIR2I6js = require('./chunk-P7TIR2I6.js');
|
|
20
21
|
|
|
21
22
|
|
|
22
23
|
var _chunk3RG5ZIWIjs = require('./chunk-3RG5ZIWI.js');
|
|
@@ -146,7 +147,7 @@ async function cargoBuildExecutor(options, context) {
|
|
|
146
147
|
const command = buildCargoCommand("build", options, context);
|
|
147
148
|
return await cargoCommand(...command);
|
|
148
149
|
}
|
|
149
|
-
var executor_default =
|
|
150
|
+
var executor_default = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
150
151
|
"Cargo - Build",
|
|
151
152
|
cargoBuildExecutor,
|
|
152
153
|
{
|
|
@@ -166,7 +167,7 @@ async function cargoCheckExecutor(options, context) {
|
|
|
166
167
|
const command = buildCargoCommand("check", options, context);
|
|
167
168
|
return await cargoCommand(...command);
|
|
168
169
|
}
|
|
169
|
-
var executor_default2 =
|
|
170
|
+
var executor_default2 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
170
171
|
"Cargo - Check",
|
|
171
172
|
cargoCheckExecutor,
|
|
172
173
|
{
|
|
@@ -185,7 +186,7 @@ async function cargoClippyExecutor(options, context) {
|
|
|
185
186
|
const command = buildCargoCommand("clippy", options, context);
|
|
186
187
|
return await cargoCommand(...command);
|
|
187
188
|
}
|
|
188
|
-
var executor_default3 =
|
|
189
|
+
var executor_default3 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
189
190
|
"Cargo - Clippy",
|
|
190
191
|
cargoClippyExecutor,
|
|
191
192
|
{
|
|
@@ -208,7 +209,7 @@ async function cargoDocExecutor(options, context) {
|
|
|
208
209
|
const command = buildCargoCommand("doc", options, context);
|
|
209
210
|
return await cargoCommand(...command);
|
|
210
211
|
}
|
|
211
|
-
var executor_default4 =
|
|
212
|
+
var executor_default4 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
212
213
|
"Cargo - Doc",
|
|
213
214
|
cargoDocExecutor,
|
|
214
215
|
{
|
|
@@ -234,7 +235,7 @@ async function cargoFormatExecutor(options, context) {
|
|
|
234
235
|
const command = buildCargoCommand("fmt", options, context);
|
|
235
236
|
return await cargoCommand(...command);
|
|
236
237
|
}
|
|
237
|
-
var executor_default5 =
|
|
238
|
+
var executor_default5 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
238
239
|
"Cargo - Format",
|
|
239
240
|
cargoFormatExecutor,
|
|
240
241
|
{
|
|
@@ -313,9 +314,9 @@ var copyAssets = async (config, assets, outputPath, projectRoot, sourceRoot, gen
|
|
|
313
314
|
output: "src/"
|
|
314
315
|
});
|
|
315
316
|
}
|
|
316
|
-
|
|
317
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0,
|
|
317
318
|
`\u{1F4DD} Copying the following assets to the output directory:
|
|
318
|
-
${pendingAssets.map((pendingAsset) => typeof pendingAsset === "string" ? ` - ${pendingAsset} -> ${outputPath}` : ` - ${pendingAsset.input}/${pendingAsset.glob} -> ${
|
|
319
|
+
${pendingAssets.map((pendingAsset) => typeof pendingAsset === "string" ? ` - ${pendingAsset} -> ${outputPath}` : ` - ${pendingAsset.input}/${pendingAsset.glob} -> ${_chunkP7TIR2I6js.joinPaths.call(void 0, outputPath, pendingAsset.output)}`).join("\n")}`,
|
|
319
320
|
config
|
|
320
321
|
);
|
|
321
322
|
const assetHandler = new (0, _copyassetshandler.CopyAssetsHandler)({
|
|
@@ -325,20 +326,20 @@ ${pendingAssets.map((pendingAsset) => typeof pendingAsset === "string" ? ` - ${p
|
|
|
325
326
|
assets: pendingAssets
|
|
326
327
|
});
|
|
327
328
|
await assetHandler.processAllAssetsOnce();
|
|
328
|
-
|
|
329
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0, "Completed copying assets to the output directory", config);
|
|
329
330
|
if (includeSrc === true) {
|
|
330
|
-
|
|
331
|
-
`\u{1F4DD} Adding banner and writing source files: ${
|
|
331
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0,
|
|
332
|
+
`\u{1F4DD} Adding banner and writing source files: ${_chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
332
333
|
outputPath,
|
|
333
334
|
"src"
|
|
334
335
|
)}`,
|
|
335
336
|
config
|
|
336
337
|
);
|
|
337
338
|
const files = await _glob.glob.call(void 0, [
|
|
338
|
-
|
|
339
|
-
|
|
340
|
-
|
|
341
|
-
|
|
339
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, config.workspaceRoot, outputPath, "src/**/*.ts"),
|
|
340
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, config.workspaceRoot, outputPath, "src/**/*.tsx"),
|
|
341
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, config.workspaceRoot, outputPath, "src/**/*.js"),
|
|
342
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, config.workspaceRoot, outputPath, "src/**/*.jsx")
|
|
342
343
|
]);
|
|
343
344
|
await Promise.allSettled(
|
|
344
345
|
files.map(
|
|
@@ -393,7 +394,7 @@ var addPackageDependencies = async (workspaceRoot3, projectRoot, projectName, pa
|
|
|
393
394
|
)) {
|
|
394
395
|
const projectNode = project.node;
|
|
395
396
|
if (projectNode.data.root) {
|
|
396
|
-
const projectPackageJsonPath =
|
|
397
|
+
const projectPackageJsonPath = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
397
398
|
workspaceRoot3,
|
|
398
399
|
projectNode.data.root,
|
|
399
400
|
"package.json"
|
|
@@ -411,11 +412,11 @@ var addPackageDependencies = async (workspaceRoot3, projectRoot, projectName, pa
|
|
|
411
412
|
}
|
|
412
413
|
}
|
|
413
414
|
if (localPackages.length > 0) {
|
|
414
|
-
|
|
415
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0,
|
|
415
416
|
`\u{1F4E6} Adding local packages to package.json: ${localPackages.map((p) => p.name).join(", ")}`
|
|
416
417
|
);
|
|
417
418
|
const projectJsonFile = await _promises.readFile.call(void 0,
|
|
418
|
-
|
|
419
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, projectRoot, "project.json"),
|
|
419
420
|
"utf8"
|
|
420
421
|
);
|
|
421
422
|
const projectJson = JSON.parse(projectJsonFile);
|
|
@@ -428,7 +429,7 @@ var addPackageDependencies = async (workspaceRoot3, projectRoot, projectName, pa
|
|
|
428
429
|
}
|
|
429
430
|
const implicitDependencies = _optionalChain([projectConfigurations, 'access', _13 => _13.projects, 'optionalAccess', _14 => _14[projectName2], 'access', _15 => _15.implicitDependencies, 'optionalAccess', _16 => _16.reduce, 'call', _17 => _17((ret, dep) => {
|
|
430
431
|
if (_optionalChain([projectConfigurations, 'access', _18 => _18.projects, 'optionalAccess', _19 => _19[dep]])) {
|
|
431
|
-
const depPackageJsonPath =
|
|
432
|
+
const depPackageJsonPath = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
432
433
|
workspaceRoot3,
|
|
433
434
|
projectConfigurations.projects[dep].root,
|
|
434
435
|
"package.json"
|
|
@@ -459,14 +460,14 @@ var addPackageDependencies = async (workspaceRoot3, projectRoot, projectName, pa
|
|
|
459
460
|
return ret;
|
|
460
461
|
}, _nullishCoalesce(packageJson.devDependencies, () => ( {})));
|
|
461
462
|
} else {
|
|
462
|
-
|
|
463
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0, "\u{1F4E6} No local packages dependencies to add to package.json");
|
|
463
464
|
}
|
|
464
465
|
return packageJson;
|
|
465
466
|
};
|
|
466
467
|
var addWorkspacePackageJsonFields = async (workspaceConfig, projectRoot, sourceRoot, projectName, includeSrc = false, packageJson) => {
|
|
467
|
-
const workspaceRoot3 = workspaceConfig.workspaceRoot ? workspaceConfig.workspaceRoot :
|
|
468
|
+
const workspaceRoot3 = workspaceConfig.workspaceRoot ? workspaceConfig.workspaceRoot : _chunkP7TIR2I6js.findWorkspaceRoot.call(void 0, );
|
|
468
469
|
const workspacePackageJsonContent = await _promises.readFile.call(void 0,
|
|
469
|
-
|
|
470
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, workspaceRoot3, "package.json"),
|
|
470
471
|
"utf8"
|
|
471
472
|
);
|
|
472
473
|
const workspacePackageJson = JSON.parse(workspacePackageJsonContent);
|
|
@@ -477,7 +478,7 @@ var addWorkspacePackageJsonFields = async (workspaceConfig, projectRoot, sourceR
|
|
|
477
478
|
if (distSrc.startsWith("/")) {
|
|
478
479
|
distSrc = distSrc.substring(1);
|
|
479
480
|
}
|
|
480
|
-
packageJson.source ??= `${
|
|
481
|
+
packageJson.source ??= `${_chunkP7TIR2I6js.joinPaths.call(void 0, distSrc, "index.ts").replaceAll("\\", "/")}`;
|
|
481
482
|
}
|
|
482
483
|
packageJson.files ??= ["dist/**/*"];
|
|
483
484
|
if (includeSrc === true && !packageJson.files.includes("src")) {
|
|
@@ -502,7 +503,7 @@ var addWorkspacePackageJsonFields = async (workspaceConfig, projectRoot, sourceR
|
|
|
502
503
|
packageJson.contributors = [packageJson.author];
|
|
503
504
|
}
|
|
504
505
|
packageJson.repository ??= workspacePackageJson.repository;
|
|
505
|
-
packageJson.repository.directory ??= projectRoot ? projectRoot :
|
|
506
|
+
packageJson.repository.directory ??= projectRoot ? projectRoot : _chunkP7TIR2I6js.joinPaths.call(void 0, "packages", projectName);
|
|
506
507
|
return packageJson;
|
|
507
508
|
};
|
|
508
509
|
|
|
@@ -536,11 +537,11 @@ var _createtaskgraph = require('nx/src/tasks-runner/create-task-graph');
|
|
|
536
537
|
// ../esbuild/src/assets.ts
|
|
537
538
|
async function copyBuildAssets(context) {
|
|
538
539
|
if (!_optionalChain([context, 'access', _28 => _28.result, 'optionalAccess', _29 => _29.errors, 'access', _30 => _30.length]) && _optionalChain([context, 'access', _31 => _31.options, 'access', _32 => _32.assets, 'optionalAccess', _33 => _33.length])) {
|
|
539
|
-
|
|
540
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0,
|
|
540
541
|
` \u{1F4CB} Copying ${context.options.assets.length} asset files to output directory: ${context.outputPath}`,
|
|
541
542
|
context.workspaceConfig
|
|
542
543
|
);
|
|
543
|
-
const stopwatch =
|
|
544
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, `${context.options.name} asset copy`);
|
|
544
545
|
await copyAssets(
|
|
545
546
|
context.workspaceConfig,
|
|
546
547
|
_nullishCoalesce(context.options.assets, () => ( [])),
|
|
@@ -621,15 +622,15 @@ async function resolveContext(userOptions) {
|
|
|
621
622
|
if (!workspaceRoot3) {
|
|
622
623
|
throw new Error("Cannot find Nx workspace root");
|
|
623
624
|
}
|
|
624
|
-
const workspaceConfig = await
|
|
625
|
+
const workspaceConfig = await _chunkP7TIR2I6js.getWorkspaceConfig.call(void 0, true, {
|
|
625
626
|
workspaceRoot: workspaceRoot3.dir
|
|
626
627
|
});
|
|
627
|
-
|
|
628
|
-
const stopwatch =
|
|
628
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, " \u2699\uFE0F Resolving build options", workspaceConfig);
|
|
629
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, "Build options resolution");
|
|
629
630
|
const projectGraph = await _devkit.createProjectGraphAsync.call(void 0, {
|
|
630
631
|
exitOnError: true
|
|
631
632
|
});
|
|
632
|
-
const projectJsonPath =
|
|
633
|
+
const projectJsonPath = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
633
634
|
workspaceRoot3.dir,
|
|
634
635
|
projectRoot,
|
|
635
636
|
"project.json"
|
|
@@ -648,7 +649,7 @@ async function resolveContext(userOptions) {
|
|
|
648
649
|
}
|
|
649
650
|
const options = _defu2.default.call(void 0, userOptions, DEFAULT_BUILD_OPTIONS);
|
|
650
651
|
options.name ??= projectName;
|
|
651
|
-
const packageJsonPath =
|
|
652
|
+
const packageJsonPath = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
652
653
|
workspaceRoot3.dir,
|
|
653
654
|
options.projectRoot,
|
|
654
655
|
"package.json"
|
|
@@ -660,7 +661,7 @@ async function resolveContext(userOptions) {
|
|
|
660
661
|
const define = _defu2.default.call(void 0, _nullishCoalesce(options.define, () => ( {})), _nullishCoalesce(env, () => ( {})));
|
|
661
662
|
const resolvedOptions = {
|
|
662
663
|
...options,
|
|
663
|
-
tsconfig: userOptions.tsconfig === null ? void 0 : userOptions.tsconfig ? userOptions.tsconfig :
|
|
664
|
+
tsconfig: userOptions.tsconfig === null ? void 0 : userOptions.tsconfig ? userOptions.tsconfig : _chunkP7TIR2I6js.joinPaths.call(void 0, workspaceRoot3.dir, projectRoot, "tsconfig.json"),
|
|
664
665
|
metafile: userOptions.mode === "development",
|
|
665
666
|
clean: false,
|
|
666
667
|
env,
|
|
@@ -685,8 +686,8 @@ async function resolveContext(userOptions) {
|
|
|
685
686
|
projectConfigurations,
|
|
686
687
|
projectName,
|
|
687
688
|
projectGraph,
|
|
688
|
-
sourceRoot: resolvedOptions.sourceRoot || projectJson.sourceRoot ||
|
|
689
|
-
outputPath: resolvedOptions.outputPath ||
|
|
689
|
+
sourceRoot: resolvedOptions.sourceRoot || projectJson.sourceRoot || _chunkP7TIR2I6js.joinPaths.call(void 0, resolvedOptions.projectRoot, "src"),
|
|
690
|
+
outputPath: resolvedOptions.outputPath || _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
690
691
|
workspaceConfig.workspaceRoot,
|
|
691
692
|
"dist",
|
|
692
693
|
resolvedOptions.projectRoot
|
|
@@ -695,10 +696,10 @@ async function resolveContext(userOptions) {
|
|
|
695
696
|
};
|
|
696
697
|
context.options.esbuildPlugins = [..._nullishCoalesce(context.options.esbuildPlugins, () => ( []))];
|
|
697
698
|
if (context.options.verbose) {
|
|
698
|
-
|
|
699
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0,
|
|
699
700
|
` \u2699\uFE0F Build options resolved:
|
|
700
701
|
|
|
701
|
-
${
|
|
702
|
+
${_chunkP7TIR2I6js.formatLogMessage.call(void 0, context.options)}`,
|
|
702
703
|
workspaceConfig
|
|
703
704
|
);
|
|
704
705
|
}
|
|
@@ -710,10 +711,10 @@ ${_chunk3HCCGFVAjs.formatLogMessage.call(void 0, context.options)}`,
|
|
|
710
711
|
|
|
711
712
|
|
|
712
713
|
async function generatePackageJson(context) {
|
|
713
|
-
if (context.options.generatePackageJson !== false && _fs.existsSync.call(void 0,
|
|
714
|
-
|
|
715
|
-
const stopwatch =
|
|
716
|
-
const packageJsonPath =
|
|
714
|
+
if (context.options.generatePackageJson !== false && _fs.existsSync.call(void 0, _chunkP7TIR2I6js.joinPaths.call(void 0, context.options.projectRoot, "package.json"))) {
|
|
715
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, " \u270D\uFE0F Writing package.json file", context.workspaceConfig);
|
|
716
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, "Write package.json file");
|
|
717
|
+
const packageJsonPath = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
717
718
|
context.options.projectRoot,
|
|
718
719
|
"project.json"
|
|
719
720
|
);
|
|
@@ -721,7 +722,7 @@ async function generatePackageJson(context) {
|
|
|
721
722
|
throw new Error("Cannot find package.json configuration");
|
|
722
723
|
}
|
|
723
724
|
const packageJsonFile = await _promises2.default.readFile(
|
|
724
|
-
|
|
725
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
725
726
|
context.workspaceConfig.workspaceRoot,
|
|
726
727
|
context.options.projectRoot,
|
|
727
728
|
"package.json"
|
|
@@ -827,7 +828,7 @@ async function generatePackageJson(context) {
|
|
|
827
828
|
}
|
|
828
829
|
}
|
|
829
830
|
await _devkit.writeJsonFile.call(void 0,
|
|
830
|
-
|
|
831
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, context.outputPath, "package.json"),
|
|
831
832
|
packageJson
|
|
832
833
|
);
|
|
833
834
|
stopwatch();
|
|
@@ -902,10 +903,10 @@ var depsCheckPlugin = (bundle) => ({
|
|
|
902
903
|
const filteredMissingDeps = missingDependencies.filter((dep) => {
|
|
903
904
|
return !missingIgnore.some((pattern) => dep.match(pattern)) && !peerDependencies.includes(dep);
|
|
904
905
|
});
|
|
905
|
-
|
|
906
|
+
_chunkP7TIR2I6js.writeWarning.call(void 0,
|
|
906
907
|
`Unused Dependencies: ${JSON.stringify(filteredUnusedDeps)}`
|
|
907
908
|
);
|
|
908
|
-
|
|
909
|
+
_chunkP7TIR2I6js.writeError.call(void 0,
|
|
909
910
|
`Missing Dependencies: ${JSON.stringify(filteredMissingDeps)}`
|
|
910
911
|
);
|
|
911
912
|
if (filteredMissingDeps.length > 0) {
|
|
@@ -920,14 +921,14 @@ ${JSON.stringify(filteredMissingDeps)}
|
|
|
920
921
|
// ../esbuild/src/tsup.ts
|
|
921
922
|
var _tsup = require('tsup');
|
|
922
923
|
async function executeTsup(context) {
|
|
923
|
-
|
|
924
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0,
|
|
924
925
|
` \u{1F680} Running ${context.options.name} build`,
|
|
925
926
|
context.workspaceConfig
|
|
926
927
|
);
|
|
927
|
-
const stopwatch =
|
|
928
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, `${context.options.name} build`);
|
|
928
929
|
await _tsup.build.call(void 0, {
|
|
929
930
|
...context.options,
|
|
930
|
-
outDir: context.options.distDir ?
|
|
931
|
+
outDir: context.options.distDir ? _chunkP7TIR2I6js.joinPaths.call(void 0, context.outputPath, context.options.distDir) : context.outputPath,
|
|
931
932
|
workspaceConfig: context.workspaceConfig
|
|
932
933
|
});
|
|
933
934
|
stopwatch();
|
|
@@ -938,17 +939,17 @@ async function executeTsup(context) {
|
|
|
938
939
|
async function reportResults(context) {
|
|
939
940
|
if (_optionalChain([context, 'access', _36 => _36.result, 'optionalAccess', _37 => _37.errors, 'access', _38 => _38.length]) === 0) {
|
|
940
941
|
if (context.result.warnings.length > 0) {
|
|
941
|
-
|
|
942
|
+
_chunkP7TIR2I6js.writeWarning.call(void 0,
|
|
942
943
|
` \u{1F6A7} The following warnings occurred during the build: ${context.result.warnings.map((warning) => warning.text).join("\n")}`,
|
|
943
944
|
context.workspaceConfig
|
|
944
945
|
);
|
|
945
946
|
}
|
|
946
|
-
|
|
947
|
+
_chunkP7TIR2I6js.writeSuccess.call(void 0,
|
|
947
948
|
` \u{1F4E6} The ${context.options.name} build completed successfully`,
|
|
948
949
|
context.workspaceConfig
|
|
949
950
|
);
|
|
950
951
|
} else if (_optionalChain([context, 'access', _39 => _39.result, 'optionalAccess', _40 => _40.errors]) && _optionalChain([context, 'access', _41 => _41.result, 'optionalAccess', _42 => _42.errors, 'access', _43 => _43.length]) > 0) {
|
|
951
|
-
|
|
952
|
+
_chunkP7TIR2I6js.writeError.call(void 0,
|
|
952
953
|
` \u274C The ${context.options.name} build failed with the following errors: ${context.result.errors.map((error) => error.text).join("\n")}`,
|
|
953
954
|
context.workspaceConfig
|
|
954
955
|
);
|
|
@@ -985,19 +986,19 @@ async function dependencyCheck(options) {
|
|
|
985
986
|
}
|
|
986
987
|
async function cleanOutputPath(context) {
|
|
987
988
|
if (context.clean !== false && context.outputPath) {
|
|
988
|
-
|
|
989
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0,
|
|
989
990
|
` \u{1F9F9} Cleaning ${context.options.name} output path: ${context.outputPath}`,
|
|
990
991
|
context.workspaceConfig
|
|
991
992
|
);
|
|
992
|
-
const stopwatch =
|
|
993
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, `${context.options.name} output clean`);
|
|
993
994
|
await cleanDirectories(context.outputPath);
|
|
994
995
|
stopwatch();
|
|
995
996
|
}
|
|
996
997
|
return context;
|
|
997
998
|
}
|
|
998
999
|
async function build2(options) {
|
|
999
|
-
|
|
1000
|
-
const stopwatch =
|
|
1000
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, ` \u26A1 Executing Storm ESBuild pipeline`);
|
|
1001
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, "ESBuild pipeline");
|
|
1001
1002
|
try {
|
|
1002
1003
|
const opts = Array.isArray(options) ? options : [options];
|
|
1003
1004
|
if (opts.length === 0) {
|
|
@@ -1012,9 +1013,9 @@ async function build2(options) {
|
|
|
1012
1013
|
executeTsup(context)
|
|
1013
1014
|
]);
|
|
1014
1015
|
await reportResults(context);
|
|
1015
|
-
|
|
1016
|
+
_chunkP7TIR2I6js.writeSuccess.call(void 0, " \u{1F3C1} ESBuild pipeline build completed successfully");
|
|
1016
1017
|
} catch (error) {
|
|
1017
|
-
|
|
1018
|
+
_chunkP7TIR2I6js.writeFatal.call(void 0,
|
|
1018
1019
|
"Fatal errors that the build process could not recover from have occured. The build process has been terminated."
|
|
1019
1020
|
);
|
|
1020
1021
|
throw error;
|
|
@@ -1025,7 +1026,7 @@ async function build2(options) {
|
|
|
1025
1026
|
|
|
1026
1027
|
// ../workspace-tools/src/executors/esbuild/executor.ts
|
|
1027
1028
|
async function esbuildExecutorFn(options, context, config) {
|
|
1028
|
-
|
|
1029
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0, "\u{1F4E6} Running Storm ESBuild executor on the workspace", config);
|
|
1029
1030
|
if (!_optionalChain([context, 'access', _44 => _44.projectsConfigurations, 'optionalAccess', _45 => _45.projects]) || !context.projectName || !context.projectsConfigurations.projects[context.projectName] || !_optionalChain([context, 'access', _46 => _46.projectsConfigurations, 'access', _47 => _47.projects, 'access', _48 => _48[context.projectName], 'optionalAccess', _49 => _49.root])) {
|
|
1030
1031
|
throw new Error(
|
|
1031
1032
|
"The Build process failed because the context is not valid. Please run this command from a workspace."
|
|
@@ -1043,7 +1044,7 @@ async function esbuildExecutorFn(options, context, config) {
|
|
|
1043
1044
|
success: true
|
|
1044
1045
|
};
|
|
1045
1046
|
}
|
|
1046
|
-
var executor_default6 =
|
|
1047
|
+
var executor_default6 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
1047
1048
|
"Storm ESBuild build",
|
|
1048
1049
|
esbuildExecutorFn,
|
|
1049
1050
|
{
|
|
@@ -1160,14 +1161,14 @@ async function sizeLimitExecutorFn(options, context, config) {
|
|
|
1160
1161
|
"The Size-Limit process failed because the context is not valid. Please run this command from a workspace."
|
|
1161
1162
|
);
|
|
1162
1163
|
}
|
|
1163
|
-
|
|
1164
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0, `\u{1F4CF} Running Size-Limit on ${context.projectName}`, config);
|
|
1164
1165
|
_sizelimit2.default.call(void 0, [_file2.default, _esbuild3.default, _esbuildwhy2.default], {
|
|
1165
1166
|
checks: _nullishCoalesce(_nullishCoalesce(options.entry, () => ( _optionalChain([context, 'access', _61 => _61.projectsConfigurations, 'access', _62 => _62.projects, 'access', _63 => _63[context.projectName], 'optionalAccess', _64 => _64.sourceRoot]))), () => ( _devkit.joinPathFragments.call(void 0,
|
|
1166
1167
|
_nullishCoalesce(_optionalChain([context, 'access', _65 => _65.projectsConfigurations, 'access', _66 => _66.projects, 'access', _67 => _67[context.projectName], 'optionalAccess', _68 => _68.root]), () => ( "./")),
|
|
1167
1168
|
"src"
|
|
1168
1169
|
)))
|
|
1169
1170
|
}).then((result) => {
|
|
1170
|
-
|
|
1171
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0,
|
|
1171
1172
|
`\u{1F4CF} ${context.projectName} Size-Limit result: ${JSON.stringify(result)}`,
|
|
1172
1173
|
config
|
|
1173
1174
|
);
|
|
@@ -1176,7 +1177,7 @@ async function sizeLimitExecutorFn(options, context, config) {
|
|
|
1176
1177
|
success: true
|
|
1177
1178
|
};
|
|
1178
1179
|
}
|
|
1179
|
-
var executor_default7 =
|
|
1180
|
+
var executor_default7 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
1180
1181
|
"Size-Limit Performance Test Executor",
|
|
1181
1182
|
sizeLimitExecutorFn,
|
|
1182
1183
|
{
|
|
@@ -1194,12 +1195,12 @@ var _fsextra = require('fs-extra');
|
|
|
1194
1195
|
var _TypiaProgrammerjs = require('typia/lib/programmers/TypiaProgrammer.js');
|
|
1195
1196
|
async function typiaExecutorFn(options, _, config) {
|
|
1196
1197
|
if (options.clean !== false) {
|
|
1197
|
-
|
|
1198
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0, `\u{1F9F9} Cleaning output path: ${options.outputPath}`, config);
|
|
1198
1199
|
_fsextra.removeSync.call(void 0, options.outputPath);
|
|
1199
1200
|
}
|
|
1200
1201
|
await Promise.all(
|
|
1201
1202
|
options.entry.map((entry) => {
|
|
1202
|
-
|
|
1203
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0, `\u{1F680} Running Typia on entry: ${entry}`, config);
|
|
1203
1204
|
return _TypiaProgrammerjs.TypiaProgrammer.build({
|
|
1204
1205
|
input: entry,
|
|
1205
1206
|
output: options.outputPath,
|
|
@@ -1211,7 +1212,7 @@ async function typiaExecutorFn(options, _, config) {
|
|
|
1211
1212
|
success: true
|
|
1212
1213
|
};
|
|
1213
1214
|
}
|
|
1214
|
-
var executor_default8 =
|
|
1215
|
+
var executor_default8 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
1215
1216
|
"Typia runtime validation generator",
|
|
1216
1217
|
typiaExecutorFn,
|
|
1217
1218
|
{
|
|
@@ -1232,7 +1233,7 @@ var executor_default8 = _chunk3HCCGFVAjs.withRunExecutor.call(void 0,
|
|
|
1232
1233
|
|
|
1233
1234
|
var _jiti = require('jiti');
|
|
1234
1235
|
async function unbuildExecutorFn(options, context, config) {
|
|
1235
|
-
|
|
1236
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0, "\u{1F4E6} Running Storm Unbuild executor on the workspace", config);
|
|
1236
1237
|
if (!_optionalChain([context, 'access', _69 => _69.projectsConfigurations, 'optionalAccess', _70 => _70.projects]) || !context.projectName || !context.projectsConfigurations.projects[context.projectName]) {
|
|
1237
1238
|
throw new Error(
|
|
1238
1239
|
"The Build process failed because the context is not valid. Please run this command from a workspace root directory."
|
|
@@ -1249,7 +1250,7 @@ async function unbuildExecutorFn(options, context, config) {
|
|
|
1249
1250
|
);
|
|
1250
1251
|
}
|
|
1251
1252
|
const jiti = _jiti.createJiti.call(void 0, config.workspaceRoot, {
|
|
1252
|
-
fsCache: config.skipCache ? false :
|
|
1253
|
+
fsCache: config.skipCache ? false : _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
1253
1254
|
config.workspaceRoot,
|
|
1254
1255
|
config.directories.cache || "node_modules/.cache/storm",
|
|
1255
1256
|
"jiti"
|
|
@@ -1271,7 +1272,7 @@ async function unbuildExecutorFn(options, context, config) {
|
|
|
1271
1272
|
{
|
|
1272
1273
|
stubOptions: {
|
|
1273
1274
|
jiti: {
|
|
1274
|
-
fsCache: config.skipCache ? false :
|
|
1275
|
+
fsCache: config.skipCache ? false : _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
1275
1276
|
config.workspaceRoot,
|
|
1276
1277
|
config.directories.cache || "node_modules/.cache/storm",
|
|
1277
1278
|
"jiti"
|
|
@@ -1300,7 +1301,7 @@ async function unbuildExecutorFn(options, context, config) {
|
|
|
1300
1301
|
success: true
|
|
1301
1302
|
};
|
|
1302
1303
|
}
|
|
1303
|
-
var executor_default9 =
|
|
1304
|
+
var executor_default9 = _chunkP7TIR2I6js.withRunExecutor.call(void 0,
|
|
1304
1305
|
"TypeScript Unbuild build",
|
|
1305
1306
|
unbuildExecutorFn,
|
|
1306
1307
|
{
|
|
@@ -1331,45 +1332,45 @@ var executor_default9 = _chunk3HCCGFVAjs.withRunExecutor.call(void 0,
|
|
|
1331
1332
|
var withRunGenerator = (name, generatorFn, generatorOptions = {
|
|
1332
1333
|
skipReadingConfig: false
|
|
1333
1334
|
}) => async (tree, _options) => {
|
|
1334
|
-
const stopwatch =
|
|
1335
|
+
const stopwatch = _chunkP7TIR2I6js.getStopwatch.call(void 0, name);
|
|
1335
1336
|
let options = _options;
|
|
1336
1337
|
let config;
|
|
1337
1338
|
try {
|
|
1338
|
-
|
|
1339
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0, `\u26A1 Running the ${name} generator...
|
|
1339
1340
|
|
|
1340
1341
|
`, config);
|
|
1341
|
-
const workspaceRoot3 =
|
|
1342
|
+
const workspaceRoot3 = _chunkP7TIR2I6js.findWorkspaceRoot.call(void 0, );
|
|
1342
1343
|
if (!generatorOptions.skipReadingConfig) {
|
|
1343
|
-
|
|
1344
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0,
|
|
1344
1345
|
`Loading the Storm Config from environment variables and storm.config.js file...
|
|
1345
1346
|
- workspaceRoot: ${workspaceRoot3}`,
|
|
1346
1347
|
config
|
|
1347
1348
|
);
|
|
1348
|
-
config = await
|
|
1349
|
+
config = await _chunkP7TIR2I6js.getConfig.call(void 0, workspaceRoot3);
|
|
1349
1350
|
}
|
|
1350
1351
|
if (_optionalChain([generatorOptions, 'optionalAccess', _71 => _71.hooks, 'optionalAccess', _72 => _72.applyDefaultOptions])) {
|
|
1351
|
-
|
|
1352
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, "Running the applyDefaultOptions hook...", config);
|
|
1352
1353
|
options = await Promise.resolve(
|
|
1353
1354
|
generatorOptions.hooks.applyDefaultOptions(options, config)
|
|
1354
1355
|
);
|
|
1355
|
-
|
|
1356
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, "Completed the applyDefaultOptions hook", config);
|
|
1356
1357
|
}
|
|
1357
|
-
|
|
1358
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0,
|
|
1358
1359
|
`Generator schema options \u2699\uFE0F
|
|
1359
1360
|
${Object.keys(_nullishCoalesce(options, () => ( {}))).map((key) => ` - ${key}=${JSON.stringify(options[key])}`).join("\n")}`,
|
|
1360
1361
|
config
|
|
1361
1362
|
);
|
|
1362
|
-
const tokenized = await
|
|
1363
|
+
const tokenized = await _chunkP7TIR2I6js.applyWorkspaceTokens.call(void 0,
|
|
1363
1364
|
options,
|
|
1364
1365
|
{ workspaceRoot: tree.root, config },
|
|
1365
|
-
|
|
1366
|
+
_chunkP7TIR2I6js.applyWorkspaceBaseTokens
|
|
1366
1367
|
);
|
|
1367
1368
|
if (_optionalChain([generatorOptions, 'optionalAccess', _73 => _73.hooks, 'optionalAccess', _74 => _74.preProcess])) {
|
|
1368
|
-
|
|
1369
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, "Running the preProcess hook...", config);
|
|
1369
1370
|
await Promise.resolve(
|
|
1370
1371
|
generatorOptions.hooks.preProcess(tokenized, config)
|
|
1371
1372
|
);
|
|
1372
|
-
|
|
1373
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, "Completed the preProcess hook", config);
|
|
1373
1374
|
}
|
|
1374
1375
|
const result = await Promise.resolve(
|
|
1375
1376
|
generatorFn(tree, tokenized, config)
|
|
@@ -1384,21 +1385,21 @@ ${Object.keys(_nullishCoalesce(options, () => ( {}))).map((key) => ` - ${key}=${
|
|
|
1384
1385
|
}
|
|
1385
1386
|
}
|
|
1386
1387
|
if (_optionalChain([generatorOptions, 'optionalAccess', _84 => _84.hooks, 'optionalAccess', _85 => _85.postProcess])) {
|
|
1387
|
-
|
|
1388
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, "Running the postProcess hook...", config);
|
|
1388
1389
|
await Promise.resolve(generatorOptions.hooks.postProcess(config));
|
|
1389
|
-
|
|
1390
|
+
_chunkP7TIR2I6js.writeDebug.call(void 0, "Completed the postProcess hook", config);
|
|
1390
1391
|
}
|
|
1391
1392
|
return () => {
|
|
1392
|
-
|
|
1393
|
+
_chunkP7TIR2I6js.writeSuccess.call(void 0, `Completed running the ${name} generator!
|
|
1393
1394
|
`, config);
|
|
1394
1395
|
};
|
|
1395
1396
|
} catch (error) {
|
|
1396
1397
|
return () => {
|
|
1397
|
-
|
|
1398
|
+
_chunkP7TIR2I6js.writeFatal.call(void 0,
|
|
1398
1399
|
"A fatal error occurred while running the generator - the process was forced to terminate",
|
|
1399
1400
|
config
|
|
1400
1401
|
);
|
|
1401
|
-
|
|
1402
|
+
_chunkP7TIR2I6js.writeError.call(void 0,
|
|
1402
1403
|
`An exception was thrown in the generator's process
|
|
1403
1404
|
- Details: ${error.message}
|
|
1404
1405
|
- Stacktrace: ${error.stack}`,
|
|
@@ -1464,16 +1465,16 @@ async function typeScriptLibraryGeneratorFn(tree, options, config) {
|
|
|
1464
1465
|
const projectConfig = {
|
|
1465
1466
|
root: normalized.directory,
|
|
1466
1467
|
projectType: "library",
|
|
1467
|
-
sourceRoot:
|
|
1468
|
+
sourceRoot: _chunkP7TIR2I6js.joinPaths.call(void 0, _nullishCoalesce(normalized.directory, () => ( "")), "src"),
|
|
1468
1469
|
targets: {
|
|
1469
1470
|
build: {
|
|
1470
1471
|
executor: options.buildExecutor,
|
|
1471
1472
|
outputs: ["{options.outputPath}"],
|
|
1472
1473
|
options: {
|
|
1473
|
-
entry: [
|
|
1474
|
+
entry: [_chunkP7TIR2I6js.joinPaths.call(void 0, normalized.projectRoot, "src", "index.ts")],
|
|
1474
1475
|
outputPath: getOutputPath(normalized),
|
|
1475
|
-
tsconfig:
|
|
1476
|
-
project:
|
|
1476
|
+
tsconfig: _chunkP7TIR2I6js.joinPaths.call(void 0, normalized.projectRoot, "tsconfig.json"),
|
|
1477
|
+
project: _chunkP7TIR2I6js.joinPaths.call(void 0, normalized.projectRoot, "package.json"),
|
|
1477
1478
|
defaultConfiguration: "production",
|
|
1478
1479
|
platform: "neutral",
|
|
1479
1480
|
assets: [
|
|
@@ -1530,7 +1531,7 @@ async function typeScriptLibraryGeneratorFn(tree, options, config) {
|
|
|
1530
1531
|
if (!normalized.importPath) {
|
|
1531
1532
|
normalized.importPath = normalized.name;
|
|
1532
1533
|
}
|
|
1533
|
-
const packageJsonPath =
|
|
1534
|
+
const packageJsonPath = _chunkP7TIR2I6js.joinPaths.call(void 0, normalized.projectRoot, "package.json");
|
|
1534
1535
|
if (tree.exists(packageJsonPath)) {
|
|
1535
1536
|
_devkit.updateJson.call(void 0, tree, packageJsonPath, (json) => {
|
|
1536
1537
|
if (!normalized.importPath) {
|
|
@@ -1587,14 +1588,14 @@ async function typeScriptLibraryGeneratorFn(tree, options, config) {
|
|
|
1587
1588
|
}));
|
|
1588
1589
|
}
|
|
1589
1590
|
_js.addTsConfigPath.call(void 0, tree, normalized.importPath, [
|
|
1590
|
-
|
|
1591
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
1591
1592
|
normalized.projectRoot,
|
|
1592
1593
|
"./src",
|
|
1593
1594
|
`index.${normalized.js ? "js" : "ts"}`
|
|
1594
1595
|
)
|
|
1595
1596
|
]);
|
|
1596
|
-
_js.addTsConfigPath.call(void 0, tree,
|
|
1597
|
-
|
|
1597
|
+
_js.addTsConfigPath.call(void 0, tree, _chunkP7TIR2I6js.joinPaths.call(void 0, normalized.importPath, "/*"), [
|
|
1598
|
+
_chunkP7TIR2I6js.joinPaths.call(void 0, normalized.projectRoot, "./src", "/*")
|
|
1598
1599
|
]);
|
|
1599
1600
|
if (tree.exists("package.json")) {
|
|
1600
1601
|
const packageJson = _devkit.readJson.call(void 0, tree, "package.json");
|
|
@@ -1605,7 +1606,7 @@ async function typeScriptLibraryGeneratorFn(tree, options, config) {
|
|
|
1605
1606
|
description = packageJson.description;
|
|
1606
1607
|
}
|
|
1607
1608
|
}
|
|
1608
|
-
const tsconfigPath =
|
|
1609
|
+
const tsconfigPath = _chunkP7TIR2I6js.joinPaths.call(void 0, normalized.projectRoot, "tsconfig.json");
|
|
1609
1610
|
if (tree.exists(tsconfigPath)) {
|
|
1610
1611
|
_devkit.updateJson.call(void 0, tree, tsconfigPath, (json) => {
|
|
1611
1612
|
json.composite ??= true;
|
|
@@ -1633,7 +1634,7 @@ function getOutputPath(options) {
|
|
|
1633
1634
|
} else {
|
|
1634
1635
|
parts.push(options.projectRoot);
|
|
1635
1636
|
}
|
|
1636
|
-
return
|
|
1637
|
+
return _chunkP7TIR2I6js.joinPaths.call(void 0, ...parts);
|
|
1637
1638
|
}
|
|
1638
1639
|
function createProjectTsConfigJson(tree, options) {
|
|
1639
1640
|
const tsconfig = {
|
|
@@ -1641,7 +1642,7 @@ function createProjectTsConfigJson(tree, options) {
|
|
|
1641
1642
|
..._nullishCoalesce(_optionalChain([options, 'optionalAccess', _100 => _100.tsconfigOptions]), () => ( {})),
|
|
1642
1643
|
compilerOptions: {
|
|
1643
1644
|
...options.rootProject ? _js.tsConfigBaseOptions : {},
|
|
1644
|
-
outDir:
|
|
1645
|
+
outDir: _chunkP7TIR2I6js.joinPaths.call(void 0, _devkit.offsetFromRoot.call(void 0, options.projectRoot), "dist/out-tsc"),
|
|
1645
1646
|
noEmit: true,
|
|
1646
1647
|
..._nullishCoalesce(_optionalChain([options, 'optionalAccess', _101 => _101.tsconfigOptions, 'optionalAccess', _102 => _102.compilerOptions]), () => ( {}))
|
|
1647
1648
|
},
|
|
@@ -1659,7 +1660,7 @@ function createProjectTsConfigJson(tree, options) {
|
|
|
1659
1660
|
"src/**/*.test.ts"
|
|
1660
1661
|
]
|
|
1661
1662
|
};
|
|
1662
|
-
_devkit.writeJson.call(void 0, tree,
|
|
1663
|
+
_devkit.writeJson.call(void 0, tree, _chunkP7TIR2I6js.joinPaths.call(void 0, options.projectRoot, "tsconfig.json"), tsconfig);
|
|
1663
1664
|
}
|
|
1664
1665
|
async function normalizeOptions(tree, options, config) {
|
|
1665
1666
|
let importPath = options.importPath;
|
|
@@ -1724,7 +1725,7 @@ async function normalizeOptions(tree, options, config) {
|
|
|
1724
1725
|
|
|
1725
1726
|
// ../workspace-tools/src/generators/browser-library/generator.ts
|
|
1726
1727
|
async function browserLibraryGeneratorFn(tree, schema, config) {
|
|
1727
|
-
const filesDir =
|
|
1728
|
+
const filesDir = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
1728
1729
|
__dirname,
|
|
1729
1730
|
"src",
|
|
1730
1731
|
"generators",
|
|
@@ -1807,37 +1808,38 @@ var generator_default = withRunGenerator(
|
|
|
1807
1808
|
|
|
1808
1809
|
var _zod = require('zod'); var z = _interopRequireWildcard(_zod);
|
|
1809
1810
|
async function configSchemaGeneratorFn(tree, options, config) {
|
|
1810
|
-
|
|
1811
|
+
_chunkP7TIR2I6js.writeInfo.call(void 0,
|
|
1811
1812
|
"\u{1F4E6} Running Storm Workspace Configuration JSON Schema generator",
|
|
1812
1813
|
config
|
|
1813
1814
|
);
|
|
1814
|
-
|
|
1815
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0,
|
|
1815
1816
|
`Determining the Storm Workspace Configuration JSON Schema...`,
|
|
1816
1817
|
config
|
|
1817
1818
|
);
|
|
1818
|
-
const jsonSchema = z.toJSONSchema(
|
|
1819
|
-
target: "draft-7"
|
|
1819
|
+
const jsonSchema = z.toJSONSchema(_chunkP7TIR2I6js.workspaceConfigSchema, {
|
|
1820
|
+
target: "draft-7",
|
|
1821
|
+
metadata: _chunkP7TIR2I6js.schemaRegistry
|
|
1820
1822
|
});
|
|
1821
1823
|
jsonSchema.$id ??= "https://public.storm-cdn.com/schemas/storm-workspace.schema.json";
|
|
1822
1824
|
jsonSchema.title ??= "Storm Workspace Configuration JSON Schema";
|
|
1823
1825
|
jsonSchema.description ??= "This JSON Schema defines the structure of the Storm Workspace configuration file (`storm-workspace.json`). It is used to validate the configuration file and ensure that it adheres to the expected format.";
|
|
1824
|
-
|
|
1826
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0, jsonSchema, config);
|
|
1825
1827
|
if (!options.outputFile) {
|
|
1826
1828
|
throw new Error(
|
|
1827
1829
|
"The `outputFile` option is required. Please specify the output file path."
|
|
1828
1830
|
);
|
|
1829
1831
|
}
|
|
1830
1832
|
const outputPath = options.outputFile.replaceAll("{workspaceRoot}", "").replaceAll(
|
|
1831
|
-
_nullishCoalesce(_optionalChain([config, 'optionalAccess', _111 => _111.workspaceRoot]), () => (
|
|
1833
|
+
_nullishCoalesce(_optionalChain([config, 'optionalAccess', _111 => _111.workspaceRoot]), () => ( _chunkP7TIR2I6js.findWorkspaceRoot.call(void 0, ))),
|
|
1832
1834
|
options.outputFile.startsWith("./") ? "" : "./"
|
|
1833
1835
|
);
|
|
1834
|
-
|
|
1836
|
+
_chunkP7TIR2I6js.writeTrace.call(void 0,
|
|
1835
1837
|
`\u{1F4DD} Writing Storm Configuration JSON Schema to "${outputPath}"`,
|
|
1836
1838
|
config
|
|
1837
1839
|
);
|
|
1838
1840
|
_devkit.writeJson.call(void 0, tree, outputPath, jsonSchema, { spaces: 2 });
|
|
1839
1841
|
await _devkit.formatFiles.call(void 0, tree);
|
|
1840
|
-
|
|
1842
|
+
_chunkP7TIR2I6js.writeSuccess.call(void 0,
|
|
1841
1843
|
"\u{1F680} Storm Configuration JSON Schema creation has completed successfully!",
|
|
1842
1844
|
config
|
|
1843
1845
|
);
|
|
@@ -1893,7 +1895,7 @@ async function initGenerator(tree, schema) {
|
|
|
1893
1895
|
|
|
1894
1896
|
|
|
1895
1897
|
async function neutralLibraryGeneratorFn(tree, schema, config) {
|
|
1896
|
-
const filesDir =
|
|
1898
|
+
const filesDir = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
1897
1899
|
__dirname,
|
|
1898
1900
|
"src",
|
|
1899
1901
|
"generators",
|
|
@@ -1952,7 +1954,7 @@ var generator_default3 = withRunGenerator(
|
|
|
1952
1954
|
|
|
1953
1955
|
|
|
1954
1956
|
async function nodeLibraryGeneratorFn(tree, schema, config) {
|
|
1955
|
-
const filesDir =
|
|
1957
|
+
const filesDir = _chunkP7TIR2I6js.joinPaths.call(void 0,
|
|
1956
1958
|
__dirname,
|
|
1957
1959
|
"src",
|
|
1958
1960
|
"generators",
|