contensis-cli 1.6.1-beta.2 → 1.6.1-beta.20
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.js +58301 -11
- package/dist/index.js.map +4 -4
- package/dist/shell.js +58040 -52
- package/dist/shell.js.map +4 -4
- package/esbuild.config.js +8 -13
- package/package.json +3 -2
- package/src/commands/create.ts +48 -0
- package/src/commands/dev.ts +12 -2
- package/src/commands/globalOptions.ts +5 -3
- package/src/commands/import.ts +46 -5
- package/src/commands/push.ts +8 -1
- package/src/commands/set.ts +69 -1
- package/src/factories/RequestHandlerFactory.ts +19 -8
- package/src/localisation/en-GB.ts +22 -4
- package/src/mappers/DevRequests-to-RequestHanderCliArgs.ts +25 -17
- package/src/models/CliService.d.ts +5 -1
- package/src/providers/GitHubCliModuleProvider.ts +8 -8
- package/src/providers/HttpProvider.ts +2 -2
- package/src/providers/file-provider.ts +3 -0
- package/src/services/ContensisCliService.ts +265 -23
- package/src/services/ContensisDevService.ts +99 -35
- package/src/shell.ts +7 -1
- package/src/util/console.printer.ts +52 -0
- package/src/util/logger.ts +3 -4
- package/src/util/os.ts +5 -3
- package/src/version.ts +1 -1
- package/dist/commands/connect.js +0 -44
- package/dist/commands/connect.js.map +0 -7
- package/dist/commands/copy.js +0 -78
- package/dist/commands/copy.js.map +0 -7
- package/dist/commands/create.js +0 -170
- package/dist/commands/create.js.map +0 -7
- package/dist/commands/dev.js +0 -78
- package/dist/commands/dev.js.map +0 -7
- package/dist/commands/diff.js +0 -57
- package/dist/commands/diff.js.map +0 -7
- package/dist/commands/execute.js +0 -103
- package/dist/commands/execute.js.map +0 -7
- package/dist/commands/get.js +0 -330
- package/dist/commands/get.js.map +0 -7
- package/dist/commands/globalOptions.js +0 -252
- package/dist/commands/globalOptions.js.map +0 -7
- package/dist/commands/import.js +0 -229
- package/dist/commands/import.js.map +0 -7
- package/dist/commands/index.js +0 -107
- package/dist/commands/index.js.map +0 -7
- package/dist/commands/list.js +0 -195
- package/dist/commands/list.js.map +0 -7
- package/dist/commands/login.js +0 -55
- package/dist/commands/login.js.map +0 -7
- package/dist/commands/push.js +0 -229
- package/dist/commands/push.js.map +0 -7
- package/dist/commands/remove.js +0 -193
- package/dist/commands/remove.js.map +0 -7
- package/dist/commands/set.js +0 -180
- package/dist/commands/set.js.map +0 -7
- package/dist/commands/update.js +0 -70
- package/dist/commands/update.js.map +0 -7
- package/dist/factories/RequestHandlerFactory.js +0 -221
- package/dist/factories/RequestHandlerFactory.js.map +0 -7
- package/dist/localisation/en-GB.js +0 -552
- package/dist/localisation/en-GB.js.map +0 -7
- package/dist/mappers/ContensisCliService-to-RequestHanderSiteConfigYaml.js +0 -56
- package/dist/mappers/ContensisCliService-to-RequestHanderSiteConfigYaml.js.map +0 -7
- package/dist/mappers/DevInit-to-CIWorkflow.js +0 -409
- package/dist/mappers/DevInit-to-CIWorkflow.js.map +0 -7
- package/dist/mappers/DevInit-to-RolePermissions.js +0 -56
- package/dist/mappers/DevInit-to-RolePermissions.js.map +0 -7
- package/dist/mappers/DevRequests-to-RequestHanderCliArgs.js +0 -161
- package/dist/mappers/DevRequests-to-RequestHanderCliArgs.js.map +0 -7
- package/dist/mappers/MixedFileData.js +0 -83
- package/dist/mappers/MixedFileData.js.map +0 -7
- package/dist/models/AppError.d.js +0 -2
- package/dist/models/AppError.d.js.map +0 -7
- package/dist/models/Cache.d.js +0 -2
- package/dist/models/Cache.d.js.map +0 -7
- package/dist/models/CliService.d.js +0 -17
- package/dist/models/CliService.d.js.map +0 -7
- package/dist/models/DevService.d.js +0 -17
- package/dist/models/DevService.d.js.map +0 -7
- package/dist/models/JsModules.d.js +0 -2
- package/dist/models/JsModules.d.js.map +0 -7
- package/dist/providers/CredentialProvider.js +0 -126
- package/dist/providers/CredentialProvider.js.map +0 -7
- package/dist/providers/GitHubCliModuleProvider.js +0 -115
- package/dist/providers/GitHubCliModuleProvider.js.map +0 -7
- package/dist/providers/HttpProvider.js +0 -73
- package/dist/providers/HttpProvider.js.map +0 -7
- package/dist/providers/ManifestProvider.js +0 -50
- package/dist/providers/ManifestProvider.js.map +0 -7
- package/dist/providers/SessionCacheProvider.js +0 -132
- package/dist/providers/SessionCacheProvider.js.map +0 -7
- package/dist/providers/file-provider.js +0 -174
- package/dist/providers/file-provider.js.map +0 -7
- package/dist/services/ContensisAuthService.js +0 -88
- package/dist/services/ContensisAuthService.js.map +0 -7
- package/dist/services/ContensisCliService.js +0 -2598
- package/dist/services/ContensisCliService.js.map +0 -7
- package/dist/services/ContensisDevService.js +0 -387
- package/dist/services/ContensisDevService.js.map +0 -7
- package/dist/services/ContensisRoleService.js +0 -112
- package/dist/services/ContensisRoleService.js.map +0 -7
- package/dist/util/api-ids.js +0 -110
- package/dist/util/api-ids.js.map +0 -7
- package/dist/util/assert.js +0 -62
- package/dist/util/assert.js.map +0 -7
- package/dist/util/console.printer.js +0 -425
- package/dist/util/console.printer.js.map +0 -7
- package/dist/util/csv.formatter.js +0 -118
- package/dist/util/csv.formatter.js.map +0 -7
- package/dist/util/debug.js +0 -29
- package/dist/util/debug.js.map +0 -7
- package/dist/util/diff.js +0 -118
- package/dist/util/diff.js.map +0 -7
- package/dist/util/dotenv.js +0 -56
- package/dist/util/dotenv.js.map +0 -7
- package/dist/util/error.js +0 -36
- package/dist/util/error.js.map +0 -7
- package/dist/util/fetch.js +0 -69
- package/dist/util/fetch.js.map +0 -7
- package/dist/util/find.js +0 -39
- package/dist/util/find.js.map +0 -7
- package/dist/util/git.js +0 -145
- package/dist/util/git.js.map +0 -7
- package/dist/util/gitignore.js +0 -61
- package/dist/util/gitignore.js.map +0 -7
- package/dist/util/html.formatter.js +0 -70
- package/dist/util/html.formatter.js.map +0 -7
- package/dist/util/index.js +0 -68
- package/dist/util/index.js.map +0 -7
- package/dist/util/json.formatter.js +0 -66
- package/dist/util/json.formatter.js.map +0 -7
- package/dist/util/logger.js +0 -304
- package/dist/util/logger.js.map +0 -7
- package/dist/util/os.js +0 -46
- package/dist/util/os.js.map +0 -7
- package/dist/util/timers.js +0 -49
- package/dist/util/timers.js.map +0 -7
- package/dist/util/xml.formatter.js +0 -62
- package/dist/util/xml.formatter.js.map +0 -7
- package/dist/util/yaml.js +0 -44
- package/dist/util/yaml.js.map +0 -7
- package/dist/version.js +0 -29
- package/dist/version.js.map +0 -7
|
@@ -1,115 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
var __create = Object.create;
|
|
3
|
-
var __defProp = Object.defineProperty;
|
|
4
|
-
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
5
|
-
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
6
|
-
var __getProtoOf = Object.getPrototypeOf;
|
|
7
|
-
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
8
|
-
var __export = (target, all) => {
|
|
9
|
-
for (var name in all)
|
|
10
|
-
__defProp(target, name, { get: all[name], enumerable: true });
|
|
11
|
-
};
|
|
12
|
-
var __copyProps = (to, from, except, desc) => {
|
|
13
|
-
if (from && typeof from === "object" || typeof from === "function") {
|
|
14
|
-
for (let key of __getOwnPropNames(from))
|
|
15
|
-
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
16
|
-
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
17
|
-
}
|
|
18
|
-
return to;
|
|
19
|
-
};
|
|
20
|
-
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
|
|
21
|
-
// If the importer is in node compatibility mode or this is not an ESM
|
|
22
|
-
// file that has been converted to a CommonJS file using a Babel-
|
|
23
|
-
// compatible transform (i.e. "__esModule" has not been set), then set
|
|
24
|
-
// "default" to the CommonJS "module.exports" for node compatibility.
|
|
25
|
-
isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
|
|
26
|
-
mod
|
|
27
|
-
));
|
|
28
|
-
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
29
|
-
var GitHubCliModuleProvider_exports = {};
|
|
30
|
-
__export(GitHubCliModuleProvider_exports, {
|
|
31
|
-
default: () => GitHubCliModuleProvider_default
|
|
32
|
-
});
|
|
33
|
-
module.exports = __toCommonJS(GitHubCliModuleProvider_exports);
|
|
34
|
-
var import_os = __toESM(require("os"));
|
|
35
|
-
var import_adm_zip = __toESM(require("adm-zip"));
|
|
36
|
-
var import_HttpProvider = __toESM(require("../providers/HttpProvider"));
|
|
37
|
-
var import_file_provider = require("./file-provider");
|
|
38
|
-
var import_fetch = require("../util/fetch");
|
|
39
|
-
class GitHubCliModuleProvider {
|
|
40
|
-
http;
|
|
41
|
-
repo;
|
|
42
|
-
baseUrl = "https://api.github.com/repos";
|
|
43
|
-
get releases_url() {
|
|
44
|
-
return `${this.baseUrl}/${this.repo}/releases`;
|
|
45
|
-
}
|
|
46
|
-
get latest_release_url() {
|
|
47
|
-
return `${this.baseUrl}/${this.repo}/releases/latest`;
|
|
48
|
-
}
|
|
49
|
-
download;
|
|
50
|
-
constructor(repo) {
|
|
51
|
-
this.http = new import_HttpProvider.default();
|
|
52
|
-
this.repo = repo;
|
|
53
|
-
}
|
|
54
|
-
async FindLatestRelease(version) {
|
|
55
|
-
const { http, latest_release_url, releases_url } = this;
|
|
56
|
-
const responses = await Promise.all([
|
|
57
|
-
http.get(latest_release_url, {
|
|
58
|
-
doRetry: (0, import_fetch.doRetry)({ silent: true })
|
|
59
|
-
}),
|
|
60
|
-
http.get(releases_url)
|
|
61
|
-
]);
|
|
62
|
-
const [latestErr, latest, latestResponse] = responses[0];
|
|
63
|
-
const [releasesErr, releases] = responses[1];
|
|
64
|
-
if (releasesErr) {
|
|
65
|
-
throw new Error(`Unable to get releases`, { cause: releasesErr });
|
|
66
|
-
} else if (!releases || releases.length === 0)
|
|
67
|
-
throw new Error(`No releases available`);
|
|
68
|
-
else if (version) {
|
|
69
|
-
const release = releases.find(
|
|
70
|
-
(r) => r.tag_name.toLowerCase() === version.toLowerCase()
|
|
71
|
-
);
|
|
72
|
-
if (release) return release;
|
|
73
|
-
else throw new Error(`No release for ${version} found`);
|
|
74
|
-
} else if (latestErr && !latest) {
|
|
75
|
-
if ((latestResponse == null ? void 0 : latestResponse.status) === 404 && (releases == null ? void 0 : releases.length)) {
|
|
76
|
-
const release = releases.find((r) => r.prerelease) || releases[0];
|
|
77
|
-
if (release) return release;
|
|
78
|
-
}
|
|
79
|
-
} else {
|
|
80
|
-
return latest;
|
|
81
|
-
}
|
|
82
|
-
}
|
|
83
|
-
async DownloadRelease(release, {
|
|
84
|
-
cmd,
|
|
85
|
-
path,
|
|
86
|
-
platforms,
|
|
87
|
-
unzip = true
|
|
88
|
-
}) {
|
|
89
|
-
const platform = platforms.find((p) => p[0] === import_os.default.platform()) || [
|
|
90
|
-
import_os.default.platform(),
|
|
91
|
-
import_os.default.platform()
|
|
92
|
-
];
|
|
93
|
-
const asset = release.assets.find(
|
|
94
|
-
(r) => r.name.toLowerCase().includes(platform[1])
|
|
95
|
-
);
|
|
96
|
-
if (asset) {
|
|
97
|
-
const filePath = (0, import_file_provider.joinPath)(path, asset.name);
|
|
98
|
-
(0, import_file_provider.removeDirectory)(path);
|
|
99
|
-
(0, import_file_provider.checkDir)(filePath);
|
|
100
|
-
await this.http.downloadFile(asset.browser_download_url, filePath);
|
|
101
|
-
if (unzip && asset.name.endsWith(".zip")) {
|
|
102
|
-
const zipFile = new import_adm_zip.default(filePath);
|
|
103
|
-
zipFile.extractAllTo(path);
|
|
104
|
-
(0, import_file_provider.removeFile)(filePath);
|
|
105
|
-
}
|
|
106
|
-
if (import_os.default.platform() !== "win32") (0, import_file_provider.addExecutePermission)((0, import_file_provider.joinPath)(path, cmd));
|
|
107
|
-
} else
|
|
108
|
-
throw new Error(
|
|
109
|
-
`no asset found in release ${release.tag_name} for platform ${import_os.default.platform()}
|
|
110
|
-
${release.html_url}`
|
|
111
|
-
);
|
|
112
|
-
}
|
|
113
|
-
}
|
|
114
|
-
var GitHubCliModuleProvider_default = GitHubCliModuleProvider;
|
|
115
|
-
//# sourceMappingURL=GitHubCliModuleProvider.js.map
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"version": 3,
|
|
3
|
-
"sources": ["../../src/providers/GitHubCliModuleProvider.ts"],
|
|
4
|
-
"sourcesContent": ["import os from 'os';\nimport Zip from 'adm-zip';\nimport type { Endpoints } from '@octokit/types';\nimport HttpProvider from '~/providers/HttpProvider';\nimport {\n addExecutePermission,\n checkDir,\n joinPath,\n removeDirectory,\n removeFile,\n} from './file-provider';\nimport { doRetry } from '~/util/fetch';\n\ntype GitHubApiRelease =\n Endpoints['GET /repos/{owner}/{repo}/releases/latest']['response']['data'];\n\nclass GitHubCliModuleProvider {\n http: HttpProvider;\n repo: string;\n baseUrl = 'https://api.github.com/repos';\n\n get releases_url() {\n return `${this.baseUrl}/${this.repo}/releases`;\n }\n get latest_release_url() {\n return `${this.baseUrl}/${this.repo}/releases/latest`;\n }\n\n download?: {\n tag: string;\n name: string;\n url: string;\n browser_url: string;\n };\n\n constructor(repo: string) {\n this.http = new HttpProvider();\n this.repo = repo;\n }\n\n async FindLatestRelease(version?: string) {\n const { http, latest_release_url, releases_url } = this;\n // return latest tag version is:\n\n const responses = await Promise.all([\n http.get<GitHubApiRelease>(latest_release_url, {\n doRetry: doRetry({ silent: true }),\n }),\n http.get<GitHubApiRelease[]>(releases_url),\n ]);\n\n const [latestErr, latest, latestResponse] = responses[0];\n const [releasesErr, releases] = responses[1];\n\n if (releasesErr) {\n throw new Error(`Unable to get releases`, { cause: releasesErr });\n } else if (!releases || releases.length === 0)\n throw new Error(`No releases available`);\n else if (version) {\n const release = releases.find(\n r => r.tag_name.toLowerCase() === version.toLowerCase()\n );\n if (release) return release;\n else throw new Error(`No release for ${version} found`);\n } else if (latestErr && !latest) {\n if (latestResponse?.status === 404 && releases?.length) {\n // No latest release, check releases for prerelease version, fallback to last release\n const release = releases.find(r => r.prerelease) || releases[0];\n\n if (release) return release;\n }\n } else {\n return latest;\n }\n }\n\n async DownloadRelease(\n release: GitHubApiRelease,\n {\n cmd,\n path,\n platforms,\n unzip = true,\n }: {\n cmd: string;\n path: string;\n unzip?: boolean;\n platforms: [NodeJS.Platform, string][];\n }\n ) {\n // find os-specific asset\n const platform = platforms.find(p => p[0] === os.platform()) || [\n os.platform(),\n os.platform(),\n ];\n\n const asset = release.assets.find(r =>\n r.name.toLowerCase().includes(platform[1])\n );\n\n // download asset\n if (asset) {\n const filePath = joinPath(path, asset.name);\n removeDirectory(path);\n checkDir(filePath);\n await this.http.downloadFile(asset.browser_download_url, filePath);\n\n if (unzip && asset.name.endsWith('.zip')) {\n // unzip the downloaded file\n const zipFile = new Zip(filePath);\n zipFile.extractAllTo(path);\n\n // delete the downloaded zip file\n removeFile(filePath);\n }\n\n if (os.platform() !== 'win32') addExecutePermission(joinPath(path, cmd));\n } else\n throw new Error(\n `no asset found in release ${\n release.tag_name\n } for platform ${os.platform()}\\n${release.html_url}`\n );\n }\n}\n\nexport default GitHubCliModuleProvider;\n"],
|
|
5
|
-
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,gBAAe;AACf,qBAAgB;AAEhB,0BAAyB;AACzB,2BAMO;AACP,mBAAwB;AAKxB,MAAM,wBAAwB;AAAA,EAC5B;AAAA,EACA;AAAA,EACA,UAAU;AAAA,EAEV,IAAI,eAAe;AACjB,WAAO,GAAG,KAAK,OAAO,IAAI,KAAK,IAAI;AAAA,EACrC;AAAA,EACA,IAAI,qBAAqB;AACvB,WAAO,GAAG,KAAK,OAAO,IAAI,KAAK,IAAI;AAAA,EACrC;AAAA,EAEA;AAAA,EAOA,YAAY,MAAc;AACxB,SAAK,OAAO,IAAI,oBAAAA,QAAa;AAC7B,SAAK,OAAO;AAAA,EACd;AAAA,EAEA,MAAM,kBAAkB,SAAkB;AACxC,UAAM,EAAE,MAAM,oBAAoB,aAAa,IAAI;AAGnD,UAAM,YAAY,MAAM,QAAQ,IAAI;AAAA,MAClC,KAAK,IAAsB,oBAAoB;AAAA,QAC7C,aAAS,sBAAQ,EAAE,QAAQ,KAAK,CAAC;AAAA,MACnC,CAAC;AAAA,MACD,KAAK,IAAwB,YAAY;AAAA,IAC3C,CAAC;AAED,UAAM,CAAC,WAAW,QAAQ,cAAc,IAAI,UAAU,CAAC;AACvD,UAAM,CAAC,aAAa,QAAQ,IAAI,UAAU,CAAC;AAE3C,QAAI,aAAa;AACf,YAAM,IAAI,MAAM,0BAA0B,EAAE,OAAO,YAAY,CAAC;AAAA,IAClE,WAAW,CAAC,YAAY,SAAS,WAAW;AAC1C,YAAM,IAAI,MAAM,uBAAuB;AAAA,aAChC,SAAS;AAChB,YAAM,UAAU,SAAS;AAAA,QACvB,OAAK,EAAE,SAAS,YAAY,MAAM,QAAQ,YAAY;AAAA,MACxD;AACA,UAAI,QAAS,QAAO;AAAA,UACf,OAAM,IAAI,MAAM,kBAAkB,OAAO,QAAQ;AAAA,IACxD,WAAW,aAAa,CAAC,QAAQ;AAC/B,WAAI,iDAAgB,YAAW,QAAO,qCAAU,SAAQ;AAEtD,cAAM,UAAU,SAAS,KAAK,OAAK,EAAE,UAAU,KAAK,SAAS,CAAC;AAE9D,YAAI,QAAS,QAAO;AAAA,MACtB;AAAA,IACF,OAAO;AACL,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,MAAM,gBACJ,SACA;AAAA,IACE;AAAA,IACA;AAAA,IACA;AAAA,IACA,QAAQ;AAAA,EACV,GAMA;AAEA,UAAM,WAAW,UAAU,KAAK,OAAK,EAAE,CAAC,MAAM,UAAAC,QAAG,SAAS,CAAC,KAAK;AAAA,MAC9D,UAAAA,QAAG,SAAS;AAAA,MACZ,UAAAA,QAAG,SAAS;AAAA,IACd;AAEA,UAAM,QAAQ,QAAQ,OAAO;AAAA,MAAK,OAChC,EAAE,KAAK,YAAY,EAAE,SAAS,SAAS,CAAC,CAAC;AAAA,IAC3C;AAGA,QAAI,OAAO;AACT,YAAM,eAAW,+BAAS,MAAM,MAAM,IAAI;AAC1C,gDAAgB,IAAI;AACpB,yCAAS,QAAQ;AACjB,YAAM,KAAK,KAAK,aAAa,MAAM,sBAAsB,QAAQ;AAEjE,UAAI,SAAS,MAAM,KAAK,SAAS,MAAM,GAAG;AAExC,cAAM,UAAU,IAAI,eAAAC,QAAI,QAAQ;AAChC,gBAAQ,aAAa,IAAI;AAGzB,6CAAW,QAAQ;AAAA,MACrB;AAEA,UAAI,UAAAD,QAAG,SAAS,MAAM,QAAS,oDAAqB,+BAAS,MAAM,GAAG,CAAC;AAAA,IACzE;AACE,YAAM,IAAI;AAAA,QACR,6BACE,QAAQ,QACV,iBAAiB,UAAAA,QAAG,SAAS,CAAC;AAAA,EAAK,QAAQ,QAAQ;AAAA,MACrD;AAAA,EACJ;AACF;AAEA,IAAO,kCAAQ;",
|
|
6
|
-
"names": ["HttpProvider", "os", "Zip"]
|
|
7
|
-
}
|
|
@@ -1,73 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
var __create = Object.create;
|
|
3
|
-
var __defProp = Object.defineProperty;
|
|
4
|
-
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
5
|
-
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
6
|
-
var __getProtoOf = Object.getPrototypeOf;
|
|
7
|
-
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
8
|
-
var __export = (target, all) => {
|
|
9
|
-
for (var name in all)
|
|
10
|
-
__defProp(target, name, { get: all[name], enumerable: true });
|
|
11
|
-
};
|
|
12
|
-
var __copyProps = (to2, from, except, desc) => {
|
|
13
|
-
if (from && typeof from === "object" || typeof from === "function") {
|
|
14
|
-
for (let key of __getOwnPropNames(from))
|
|
15
|
-
if (!__hasOwnProp.call(to2, key) && key !== except)
|
|
16
|
-
__defProp(to2, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
17
|
-
}
|
|
18
|
-
return to2;
|
|
19
|
-
};
|
|
20
|
-
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
|
|
21
|
-
// If the importer is in node compatibility mode or this is not an ESM
|
|
22
|
-
// file that has been converted to a CommonJS file using a Babel-
|
|
23
|
-
// compatible transform (i.e. "__esModule" has not been set), then set
|
|
24
|
-
// "default" to the CommonJS "module.exports" for node compatibility.
|
|
25
|
-
isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
|
|
26
|
-
mod
|
|
27
|
-
));
|
|
28
|
-
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
29
|
-
var HttpProvider_exports = {};
|
|
30
|
-
__export(HttpProvider_exports, {
|
|
31
|
-
default: () => HttpProvider_default
|
|
32
|
-
});
|
|
33
|
-
module.exports = __toCommonJS(HttpProvider_exports);
|
|
34
|
-
var import_await_to_js = __toESM(require("await-to-js"));
|
|
35
|
-
var import_fs = __toESM(require("fs"));
|
|
36
|
-
var import_stream = require("stream");
|
|
37
|
-
var import_promises = require("stream/promises");
|
|
38
|
-
var import_assert = require("../util/assert");
|
|
39
|
-
var import_fetch = require("../util/fetch");
|
|
40
|
-
class HttpProvider {
|
|
41
|
-
constructor() {
|
|
42
|
-
}
|
|
43
|
-
async get(url, init = {}) {
|
|
44
|
-
return this.fetch(url, { method: "GET", ...init });
|
|
45
|
-
}
|
|
46
|
-
async fetch(uri, init = {}) {
|
|
47
|
-
const [error, response] = await (0, import_await_to_js.default)((0, import_fetch.enhancedFetch)(uri, init));
|
|
48
|
-
if (response && !error) {
|
|
49
|
-
const [bodyError, text] = await (0, import_await_to_js.default)(response.text());
|
|
50
|
-
if (bodyError) return [bodyError, void 0, response];
|
|
51
|
-
if ((0, import_assert.isJson)(text)) {
|
|
52
|
-
const err = !response.status || !response.ok ? (0, import_assert.tryParse)(text) : void 0;
|
|
53
|
-
const payload = response.status && response.ok ? (0, import_assert.tryParse)(text) : void 0;
|
|
54
|
-
return [err, payload, response];
|
|
55
|
-
}
|
|
56
|
-
return [
|
|
57
|
-
response.ok ? null : new Error(text),
|
|
58
|
-
response.ok ? text : void 0,
|
|
59
|
-
response
|
|
60
|
-
];
|
|
61
|
-
}
|
|
62
|
-
return [error, void 0, response];
|
|
63
|
-
}
|
|
64
|
-
async downloadFile(url, destination) {
|
|
65
|
-
const res = await fetch(url);
|
|
66
|
-
if (res.ok && res.body !== null) {
|
|
67
|
-
const fileStream = import_fs.default.createWriteStream(destination, { flags: "wx" });
|
|
68
|
-
await (0, import_promises.finished)(import_stream.Readable.fromWeb(res.body).pipe(fileStream));
|
|
69
|
-
}
|
|
70
|
-
}
|
|
71
|
-
}
|
|
72
|
-
var HttpProvider_default = HttpProvider;
|
|
73
|
-
//# sourceMappingURL=HttpProvider.js.map
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"version": 3,
|
|
3
|
-
"sources": ["../../src/providers/HttpProvider.ts"],
|
|
4
|
-
"sourcesContent": ["import to from 'await-to-js';\nimport { FetchInit } from 'enterprise-fetch';\nimport fs from 'fs';\nimport { Readable } from 'stream';\nimport { finished } from 'stream/promises';\n\nimport { isJson, tryParse } from '~/util/assert';\nimport { enhancedFetch } from '~/util/fetch';\nclass HttpProvider {\n constructor() {}\n\n async get<T = any>(url: string, init: FetchInit = {}) {\n return this.fetch<T>(url, { method: 'GET', ...init });\n }\n\n async fetch<T = any>(\n uri: string,\n init: FetchInit = {}\n ): Promise<[Error | null, T | undefined, Response | undefined]> {\n const [error, response] = await to(enhancedFetch(uri, init));\n\n if (response && !error) {\n const [bodyError, text] = await to(response.text());\n if (bodyError) return [bodyError, undefined, response];\n if (isJson(text)) {\n const err =\n !response.status || !response.ok ? tryParse(text) : undefined;\n const payload =\n response.status && response.ok ? tryParse(text) : undefined;\n return [err, payload, response];\n }\n return [\n response.ok ? null : new Error(text),\n response.ok ? (text as unknown as T) : undefined,\n response,\n ];\n }\n return [error, undefined, response];\n }\n\n async downloadFile(url: string, destination: string) {\n const res = await fetch(url);\n if (res.ok && res.body !== null) {\n const fileStream = fs.createWriteStream(destination, { flags: 'wx' });\n await finished(Readable.fromWeb(res.body as any).pipe(fileStream));\n }\n }\n}\n\nexport default HttpProvider;\n"],
|
|
5
|
-
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,yBAAe;AAEf,gBAAe;AACf,oBAAyB;AACzB,sBAAyB;AAEzB,oBAAiC;AACjC,mBAA8B;AAC9B,MAAM,aAAa;AAAA,EACjB,cAAc;AAAA,EAAC;AAAA,EAEf,MAAM,IAAa,KAAa,OAAkB,CAAC,GAAG;AACpD,WAAO,KAAK,MAAS,KAAK,EAAE,QAAQ,OAAO,GAAG,KAAK,CAAC;AAAA,EACtD;AAAA,EAEA,MAAM,MACJ,KACA,OAAkB,CAAC,GAC2C;AAC9D,UAAM,CAAC,OAAO,QAAQ,IAAI,UAAM,mBAAAA,aAAG,4BAAc,KAAK,IAAI,CAAC;AAE3D,QAAI,YAAY,CAAC,OAAO;AACtB,YAAM,CAAC,WAAW,IAAI,IAAI,UAAM,mBAAAA,SAAG,SAAS,KAAK,CAAC;AAClD,UAAI,UAAW,QAAO,CAAC,WAAW,QAAW,QAAQ;AACrD,cAAI,sBAAO,IAAI,GAAG;AAChB,cAAM,MACJ,CAAC,SAAS,UAAU,CAAC,SAAS,SAAK,wBAAS,IAAI,IAAI;AACtD,cAAM,UACJ,SAAS,UAAU,SAAS,SAAK,wBAAS,IAAI,IAAI;AACpD,eAAO,CAAC,KAAK,SAAS,QAAQ;AAAA,MAChC;AACA,aAAO;AAAA,QACL,SAAS,KAAK,OAAO,IAAI,MAAM,IAAI;AAAA,QACnC,SAAS,KAAM,OAAwB;AAAA,QACvC;AAAA,MACF;AAAA,IACF;AACA,WAAO,CAAC,OAAO,QAAW,QAAQ;AAAA,EACpC;AAAA,EAEA,MAAM,aAAa,KAAa,aAAqB;AACnD,UAAM,MAAM,MAAM,MAAM,GAAG;AAC3B,QAAI,IAAI,MAAM,IAAI,SAAS,MAAM;AAC/B,YAAM,aAAa,UAAAC,QAAG,kBAAkB,aAAa,EAAE,OAAO,KAAK,CAAC;AACpE,gBAAM,0BAAS,uBAAS,QAAQ,IAAI,IAAW,EAAE,KAAK,UAAU,CAAC;AAAA,IACnE;AAAA,EACF;AACF;AAEA,IAAO,uBAAQ;",
|
|
6
|
-
"names": ["to", "fs"]
|
|
7
|
-
}
|
|
@@ -1,50 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
var __defProp = Object.defineProperty;
|
|
3
|
-
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
4
|
-
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
5
|
-
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
6
|
-
var __export = (target, all) => {
|
|
7
|
-
for (var name in all)
|
|
8
|
-
__defProp(target, name, { get: all[name], enumerable: true });
|
|
9
|
-
};
|
|
10
|
-
var __copyProps = (to, from, except, desc) => {
|
|
11
|
-
if (from && typeof from === "object" || typeof from === "function") {
|
|
12
|
-
for (let key of __getOwnPropNames(from))
|
|
13
|
-
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
14
|
-
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
15
|
-
}
|
|
16
|
-
return to;
|
|
17
|
-
};
|
|
18
|
-
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
19
|
-
var ManifestProvider_exports = {};
|
|
20
|
-
__export(ManifestProvider_exports, {
|
|
21
|
-
default: () => ManifestProvider_default
|
|
22
|
-
});
|
|
23
|
-
module.exports = __toCommonJS(ManifestProvider_exports);
|
|
24
|
-
var import_assert = require("../util/assert");
|
|
25
|
-
var import_file_provider = require("./file-provider");
|
|
26
|
-
const MANIFEST_PATH = (0, import_file_provider.appPath)("cli-manifest.json");
|
|
27
|
-
class ManifestProvider {
|
|
28
|
-
manifest;
|
|
29
|
-
constructor() {
|
|
30
|
-
const manifest = (0, import_assert.tryParse)((0, import_file_provider.readFile)(MANIFEST_PATH));
|
|
31
|
-
this.manifest = manifest || {};
|
|
32
|
-
}
|
|
33
|
-
get() {
|
|
34
|
-
return this.manifest;
|
|
35
|
-
}
|
|
36
|
-
getModule(name) {
|
|
37
|
-
var _a;
|
|
38
|
-
return (_a = this.manifest) == null ? void 0 : _a[name];
|
|
39
|
-
}
|
|
40
|
-
writeModule(name, moduleInfo) {
|
|
41
|
-
if (this.manifest) this.manifest[name] = moduleInfo;
|
|
42
|
-
else
|
|
43
|
-
this.manifest = {
|
|
44
|
-
[name]: moduleInfo
|
|
45
|
-
};
|
|
46
|
-
(0, import_file_provider.writeFile)(MANIFEST_PATH, JSON.stringify(this.manifest, null, 2));
|
|
47
|
-
}
|
|
48
|
-
}
|
|
49
|
-
var ManifestProvider_default = ManifestProvider;
|
|
50
|
-
//# sourceMappingURL=ManifestProvider.js.map
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"version": 3,
|
|
3
|
-
"sources": ["../../src/providers/ManifestProvider.ts"],
|
|
4
|
-
"sourcesContent": ["import { tryParse } from '~/util/assert';\nimport { appPath, readFile, writeFile } from './file-provider';\n\nexport type CliModule = {\n github: string;\n version: string;\n install?: string;\n cmd?: string;\n};\n\ntype CliManifest = {\n [moduleName: string]: CliModule;\n};\n\nconst MANIFEST_PATH = appPath('cli-manifest.json');\n\nclass ManifestProvider {\n private manifest: CliManifest;\n\n constructor() {\n const manifest = tryParse(readFile(MANIFEST_PATH));\n this.manifest = manifest || {};\n }\n\n get() {\n return this.manifest;\n }\n\n getModule(name: string) {\n return this.manifest?.[name];\n }\n writeModule(name: string, moduleInfo: CliModule) {\n if (this.manifest) this.manifest[name] = moduleInfo;\n else\n this.manifest = {\n [name]: moduleInfo,\n };\n\n writeFile(MANIFEST_PATH, JSON.stringify(this.manifest, null, 2));\n }\n}\n\nexport default ManifestProvider;\n"],
|
|
5
|
-
"mappings": ";;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAAyB;AACzB,2BAA6C;AAa7C,MAAM,oBAAgB,8BAAQ,mBAAmB;AAEjD,MAAM,iBAAiB;AAAA,EACb;AAAA,EAER,cAAc;AACZ,UAAM,eAAW,4BAAS,+BAAS,aAAa,CAAC;AACjD,SAAK,WAAW,YAAY,CAAC;AAAA,EAC/B;AAAA,EAEA,MAAM;AACJ,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,UAAU,MAAc;AA5B1B;AA6BI,YAAO,UAAK,aAAL,mBAAgB;AAAA,EACzB;AAAA,EACA,YAAY,MAAc,YAAuB;AAC/C,QAAI,KAAK,SAAU,MAAK,SAAS,IAAI,IAAI;AAAA;AAEvC,WAAK,WAAW;AAAA,QACd,CAAC,IAAI,GAAG;AAAA,MACV;AAEF,wCAAU,eAAe,KAAK,UAAU,KAAK,UAAU,MAAM,CAAC,CAAC;AAAA,EACjE;AACF;AAEA,IAAO,2BAAQ;",
|
|
6
|
-
"names": []
|
|
7
|
-
}
|
|
@@ -1,132 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
var __create = Object.create;
|
|
3
|
-
var __defProp = Object.defineProperty;
|
|
4
|
-
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
5
|
-
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
6
|
-
var __getProtoOf = Object.getPrototypeOf;
|
|
7
|
-
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
8
|
-
var __export = (target, all) => {
|
|
9
|
-
for (var name in all)
|
|
10
|
-
__defProp(target, name, { get: all[name], enumerable: true });
|
|
11
|
-
};
|
|
12
|
-
var __copyProps = (to, from, except, desc) => {
|
|
13
|
-
if (from && typeof from === "object" || typeof from === "function") {
|
|
14
|
-
for (let key of __getOwnPropNames(from))
|
|
15
|
-
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
16
|
-
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
17
|
-
}
|
|
18
|
-
return to;
|
|
19
|
-
};
|
|
20
|
-
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
|
|
21
|
-
// If the importer is in node compatibility mode or this is not an ESM
|
|
22
|
-
// file that has been converted to a CommonJS file using a Babel-
|
|
23
|
-
// compatible transform (i.e. "__esModule" has not been set), then set
|
|
24
|
-
// "default" to the CommonJS "module.exports" for node compatibility.
|
|
25
|
-
isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
|
|
26
|
-
mod
|
|
27
|
-
));
|
|
28
|
-
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
29
|
-
var SessionCacheProvider_exports = {};
|
|
30
|
-
__export(SessionCacheProvider_exports, {
|
|
31
|
-
default: () => SessionCacheProvider_default
|
|
32
|
-
});
|
|
33
|
-
module.exports = __toCommonJS(SessionCacheProvider_exports);
|
|
34
|
-
var import_fs = __toESM(require("fs"));
|
|
35
|
-
var import_path = __toESM(require("path"));
|
|
36
|
-
var import_cloneDeep = __toESM(require("lodash/cloneDeep"));
|
|
37
|
-
var import_mergeWith = __toESM(require("lodash/mergeWith"));
|
|
38
|
-
var import_unionBy = __toESM(require("lodash/unionBy"));
|
|
39
|
-
var import_file_provider = require("./file-provider");
|
|
40
|
-
var import_assert = require("../util/assert");
|
|
41
|
-
var import_logger = require("../util/logger");
|
|
42
|
-
var import_version = require("../version");
|
|
43
|
-
class SessionCacheProvider {
|
|
44
|
-
localFilePath;
|
|
45
|
-
cache = {};
|
|
46
|
-
constructor() {
|
|
47
|
-
this.localFilePath = import_path.default.join(import_file_provider.appRootDir, "environments.json");
|
|
48
|
-
(0, import_file_provider.checkDir)(this.localFilePath);
|
|
49
|
-
this.cache = {
|
|
50
|
-
currentTimestamp: (/* @__PURE__ */ new Date()).toISOString(),
|
|
51
|
-
environments: {},
|
|
52
|
-
history: [],
|
|
53
|
-
version: import_version.LIB_VERSION
|
|
54
|
-
};
|
|
55
|
-
this.ReadCacheFromDisk();
|
|
56
|
-
}
|
|
57
|
-
ReadCacheFromDisk = () => {
|
|
58
|
-
try {
|
|
59
|
-
if (import_fs.default.existsSync(this.localFilePath)) {
|
|
60
|
-
const raw = import_fs.default.readFileSync(this.localFilePath, "utf-8");
|
|
61
|
-
if ((0, import_assert.isJson)(raw)) this.cache = (0, import_assert.tryParse)(raw);
|
|
62
|
-
} else {
|
|
63
|
-
this.WriteCacheToDisk();
|
|
64
|
-
}
|
|
65
|
-
} catch (ex) {
|
|
66
|
-
import_logger.Logger.error(ex);
|
|
67
|
-
}
|
|
68
|
-
};
|
|
69
|
-
WriteCacheToDisk = () => {
|
|
70
|
-
try {
|
|
71
|
-
import_fs.default.writeFileSync(this.localFilePath, JSON.stringify(this.cache, null, 2));
|
|
72
|
-
} catch (ex) {
|
|
73
|
-
import_logger.Logger.error(ex);
|
|
74
|
-
}
|
|
75
|
-
};
|
|
76
|
-
Get = () => (0, import_cloneDeep.default)(this.cache);
|
|
77
|
-
Update = (updateContent) => {
|
|
78
|
-
try {
|
|
79
|
-
this.cache = (0, import_mergeWith.default)(this.cache, updateContent, (val, src, key) => {
|
|
80
|
-
if (key === "history" && ((src == null ? void 0 : src[0]) && typeof src[0] === "object" && "createdDate" in src[0] || (val == null ? void 0 : val[0]) && typeof val[0] === "object" && "createdDate" in val[0])) {
|
|
81
|
-
return (0, import_unionBy.default)(val, src, "createdDate");
|
|
82
|
-
}
|
|
83
|
-
if (key === "projects") return Array.isArray(val) ? val : src;
|
|
84
|
-
if (Array.isArray(val)) return val.concat(src);
|
|
85
|
-
});
|
|
86
|
-
this.cache.currentTimestamp = (/* @__PURE__ */ new Date()).toISOString();
|
|
87
|
-
this.cache.version = import_version.LIB_VERSION;
|
|
88
|
-
this.WriteCacheToDisk();
|
|
89
|
-
} catch (ex) {
|
|
90
|
-
import_logger.Logger.error(`Problem updating environments.json`);
|
|
91
|
-
import_logger.Logger.error(ex);
|
|
92
|
-
}
|
|
93
|
-
return this.Get();
|
|
94
|
-
};
|
|
95
|
-
UpdateEnv = (updateContent, env = this.cache.currentEnvironment, setCurrentEnv = true) => {
|
|
96
|
-
try {
|
|
97
|
-
const environment = this.cache.environments[env || ""];
|
|
98
|
-
this.cache.environments[env || ""] = {
|
|
99
|
-
...environment,
|
|
100
|
-
...updateContent
|
|
101
|
-
};
|
|
102
|
-
this.Update({
|
|
103
|
-
currentEnvironment: setCurrentEnv ? env : this.cache.currentEnvironment,
|
|
104
|
-
environments: this.cache.environments
|
|
105
|
-
});
|
|
106
|
-
} catch (ex) {
|
|
107
|
-
import_logger.Logger.error(
|
|
108
|
-
`Problem updating environment "${env}" in environments.json`
|
|
109
|
-
);
|
|
110
|
-
import_logger.Logger.error(ex);
|
|
111
|
-
}
|
|
112
|
-
return this.Get();
|
|
113
|
-
};
|
|
114
|
-
RemoveEnv = (env) => {
|
|
115
|
-
try {
|
|
116
|
-
const environment = this.cache.environments[env || ""];
|
|
117
|
-
if (environment) delete this.cache.environments[env || ""];
|
|
118
|
-
this.Update({
|
|
119
|
-
currentEnvironment: this.cache.currentEnvironment === env ? "" : this.cache.currentEnvironment,
|
|
120
|
-
environments: this.cache.environments
|
|
121
|
-
});
|
|
122
|
-
} catch (ex) {
|
|
123
|
-
import_logger.Logger.error(
|
|
124
|
-
`Problem removing environment "${env}" in environments.json`
|
|
125
|
-
);
|
|
126
|
-
import_logger.Logger.error(ex);
|
|
127
|
-
}
|
|
128
|
-
return this.Get();
|
|
129
|
-
};
|
|
130
|
-
}
|
|
131
|
-
var SessionCacheProvider_default = SessionCacheProvider;
|
|
132
|
-
//# sourceMappingURL=SessionCacheProvider.js.map
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"version": 3,
|
|
3
|
-
"sources": ["../../src/providers/SessionCacheProvider.ts"],
|
|
4
|
-
"sourcesContent": ["import fs from 'fs';\nimport path from 'path';\nimport clone from 'lodash/cloneDeep';\nimport mergeWith from 'lodash/mergeWith';\nimport unionBy from 'lodash/unionBy';\nimport { appRootDir, checkDir } from './file-provider';\nimport { isJson, tryParse } from '~/util/assert';\nimport { Logger } from '~/util/logger';\nimport { LIB_VERSION } from '~/version';\n\nclass SessionCacheProvider {\n private localFilePath: string;\n private cache = {} as SessionCache;\n\n constructor() {\n this.localFilePath = path.join(appRootDir, 'environments.json');\n checkDir(this.localFilePath); // ensure the .contensis home folder exists\n this.cache = {\n currentTimestamp: new Date().toISOString(),\n environments: {},\n history: [],\n version: LIB_VERSION,\n };\n this.ReadCacheFromDisk();\n }\n\n private ReadCacheFromDisk = () => {\n try {\n if (fs.existsSync(this.localFilePath)) {\n const raw = fs.readFileSync(this.localFilePath, 'utf-8');\n if (isJson(raw)) this.cache = tryParse(raw);\n } else {\n this.WriteCacheToDisk();\n }\n } catch (ex: any) {\n // Problem reading or parsing cache file\n Logger.error(ex);\n }\n };\n\n private WriteCacheToDisk = () => {\n try {\n fs.writeFileSync(this.localFilePath, JSON.stringify(this.cache, null, 2));\n } catch (ex: any) {\n // Problem writing session cache to file\n Logger.error(ex);\n }\n };\n\n Get = () => clone(this.cache);\n\n Update = (updateContent: Partial<SessionCache>) => {\n try {\n this.cache = mergeWith(this.cache, updateContent, (val, src, key) => {\n if (\n key === 'history' &&\n ((src?.[0] &&\n typeof src[0] === 'object' &&\n 'createdDate' in src[0]) ||\n (val?.[0] && typeof val[0] === 'object' && 'createdDate' in val[0]))\n ) {\n return unionBy(val, src, 'createdDate');\n }\n if (key === 'projects') return Array.isArray(val) ? val : src;\n\n if (Array.isArray(val)) return val.concat(src);\n });\n\n this.cache.currentTimestamp = new Date().toISOString();\n this.cache.version = LIB_VERSION;\n\n this.WriteCacheToDisk();\n } catch (ex: any) {\n // Problem merging cache data for update\n Logger.error(`Problem updating environments.json`);\n Logger.error(ex);\n }\n return this.Get();\n };\n\n UpdateEnv = (\n updateContent: Partial<EnvironmentCache>,\n env = this.cache.currentEnvironment,\n setCurrentEnv = true\n ) => {\n try {\n const environment = this.cache.environments[env || ''];\n\n this.cache.environments[env || ''] = {\n ...environment,\n ...updateContent,\n };\n this.Update({\n currentEnvironment: setCurrentEnv ? env : this.cache.currentEnvironment,\n environments: this.cache.environments,\n });\n } catch (ex: any) {\n // Problem merging cache data for update\n Logger.error(\n `Problem updating environment \"${env}\" in environments.json`\n );\n Logger.error(ex);\n }\n return this.Get();\n };\n\n RemoveEnv = (env: string) => {\n try {\n const environment = this.cache.environments[env || ''];\n if (environment) delete this.cache.environments[env || ''];\n\n this.Update({\n currentEnvironment:\n this.cache.currentEnvironment === env\n ? ''\n : this.cache.currentEnvironment,\n environments: this.cache.environments,\n });\n } catch (ex: any) {\n // Problem merging cache data for update\n Logger.error(\n `Problem removing environment \"${env}\" in environments.json`\n );\n Logger.error(ex);\n }\n return this.Get();\n };\n}\n\nexport default SessionCacheProvider;\n"],
|
|
5
|
-
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,gBAAe;AACf,kBAAiB;AACjB,uBAAkB;AAClB,uBAAsB;AACtB,qBAAoB;AACpB,2BAAqC;AACrC,oBAAiC;AACjC,oBAAuB;AACvB,qBAA4B;AAE5B,MAAM,qBAAqB;AAAA,EACjB;AAAA,EACA,QAAQ,CAAC;AAAA,EAEjB,cAAc;AACZ,SAAK,gBAAgB,YAAAA,QAAK,KAAK,iCAAY,mBAAmB;AAC9D,uCAAS,KAAK,aAAa;AAC3B,SAAK,QAAQ;AAAA,MACX,mBAAkB,oBAAI,KAAK,GAAE,YAAY;AAAA,MACzC,cAAc,CAAC;AAAA,MACf,SAAS,CAAC;AAAA,MACV,SAAS;AAAA,IACX;AACA,SAAK,kBAAkB;AAAA,EACzB;AAAA,EAEQ,oBAAoB,MAAM;AAChC,QAAI;AACF,UAAI,UAAAC,QAAG,WAAW,KAAK,aAAa,GAAG;AACrC,cAAM,MAAM,UAAAA,QAAG,aAAa,KAAK,eAAe,OAAO;AACvD,gBAAI,sBAAO,GAAG,EAAG,MAAK,YAAQ,wBAAS,GAAG;AAAA,MAC5C,OAAO;AACL,aAAK,iBAAiB;AAAA,MACxB;AAAA,IACF,SAAS,IAAS;AAEhB,2BAAO,MAAM,EAAE;AAAA,IACjB;AAAA,EACF;AAAA,EAEQ,mBAAmB,MAAM;AAC/B,QAAI;AACF,gBAAAA,QAAG,cAAc,KAAK,eAAe,KAAK,UAAU,KAAK,OAAO,MAAM,CAAC,CAAC;AAAA,IAC1E,SAAS,IAAS;AAEhB,2BAAO,MAAM,EAAE;AAAA,IACjB;AAAA,EACF;AAAA,EAEA,MAAM,UAAM,iBAAAC,SAAM,KAAK,KAAK;AAAA,EAE5B,SAAS,CAAC,kBAAyC;AACjD,QAAI;AACF,WAAK,YAAQ,iBAAAC,SAAU,KAAK,OAAO,eAAe,CAAC,KAAK,KAAK,QAAQ;AACnE,YACE,QAAQ,eACN,2BAAM,OACN,OAAO,IAAI,CAAC,MAAM,YAClB,iBAAiB,IAAI,CAAC,MACrB,2BAAM,OAAM,OAAO,IAAI,CAAC,MAAM,YAAY,iBAAiB,IAAI,CAAC,IACnE;AACA,qBAAO,eAAAC,SAAQ,KAAK,KAAK,aAAa;AAAA,QACxC;AACA,YAAI,QAAQ,WAAY,QAAO,MAAM,QAAQ,GAAG,IAAI,MAAM;AAE1D,YAAI,MAAM,QAAQ,GAAG,EAAG,QAAO,IAAI,OAAO,GAAG;AAAA,MAC/C,CAAC;AAED,WAAK,MAAM,oBAAmB,oBAAI,KAAK,GAAE,YAAY;AACrD,WAAK,MAAM,UAAU;AAErB,WAAK,iBAAiB;AAAA,IACxB,SAAS,IAAS;AAEhB,2BAAO,MAAM,oCAAoC;AACjD,2BAAO,MAAM,EAAE;AAAA,IACjB;AACA,WAAO,KAAK,IAAI;AAAA,EAClB;AAAA,EAEA,YAAY,CACV,eACA,MAAM,KAAK,MAAM,oBACjB,gBAAgB,SACb;AACH,QAAI;AACF,YAAM,cAAc,KAAK,MAAM,aAAa,OAAO,EAAE;AAErD,WAAK,MAAM,aAAa,OAAO,EAAE,IAAI;AAAA,QACnC,GAAG;AAAA,QACH,GAAG;AAAA,MACL;AACA,WAAK,OAAO;AAAA,QACV,oBAAoB,gBAAgB,MAAM,KAAK,MAAM;AAAA,QACrD,cAAc,KAAK,MAAM;AAAA,MAC3B,CAAC;AAAA,IACH,SAAS,IAAS;AAEhB,2BAAO;AAAA,QACL,iCAAiC,GAAG;AAAA,MACtC;AACA,2BAAO,MAAM,EAAE;AAAA,IACjB;AACA,WAAO,KAAK,IAAI;AAAA,EAClB;AAAA,EAEA,YAAY,CAAC,QAAgB;AAC3B,QAAI;AACF,YAAM,cAAc,KAAK,MAAM,aAAa,OAAO,EAAE;AACrD,UAAI,YAAa,QAAO,KAAK,MAAM,aAAa,OAAO,EAAE;AAEzD,WAAK,OAAO;AAAA,QACV,oBACE,KAAK,MAAM,uBAAuB,MAC9B,KACA,KAAK,MAAM;AAAA,QACjB,cAAc,KAAK,MAAM;AAAA,MAC3B,CAAC;AAAA,IACH,SAAS,IAAS;AAEhB,2BAAO;AAAA,QACL,iCAAiC,GAAG;AAAA,MACtC;AACA,2BAAO,MAAM,EAAE;AAAA,IACjB;AACA,WAAO,KAAK,IAAI;AAAA,EAClB;AACF;AAEA,IAAO,+BAAQ;",
|
|
6
|
-
"names": ["path", "fs", "clone", "mergeWith", "unionBy"]
|
|
7
|
-
}
|
|
@@ -1,174 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
var __create = Object.create;
|
|
3
|
-
var __defProp = Object.defineProperty;
|
|
4
|
-
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
|
5
|
-
var __getOwnPropNames = Object.getOwnPropertyNames;
|
|
6
|
-
var __getProtoOf = Object.getPrototypeOf;
|
|
7
|
-
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
8
|
-
var __export = (target, all) => {
|
|
9
|
-
for (var name in all)
|
|
10
|
-
__defProp(target, name, { get: all[name], enumerable: true });
|
|
11
|
-
};
|
|
12
|
-
var __copyProps = (to, from, except, desc) => {
|
|
13
|
-
if (from && typeof from === "object" || typeof from === "function") {
|
|
14
|
-
for (let key of __getOwnPropNames(from))
|
|
15
|
-
if (!__hasOwnProp.call(to, key) && key !== except)
|
|
16
|
-
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
|
|
17
|
-
}
|
|
18
|
-
return to;
|
|
19
|
-
};
|
|
20
|
-
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
|
|
21
|
-
// If the importer is in node compatibility mode or this is not an ESM
|
|
22
|
-
// file that has been converted to a CommonJS file using a Babel-
|
|
23
|
-
// compatible transform (i.e. "__esModule" has not been set), then set
|
|
24
|
-
// "default" to the CommonJS "module.exports" for node compatibility.
|
|
25
|
-
isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
|
|
26
|
-
mod
|
|
27
|
-
));
|
|
28
|
-
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
|
29
|
-
var file_provider_exports = {};
|
|
30
|
-
__export(file_provider_exports, {
|
|
31
|
-
addExecutePermission: () => addExecutePermission,
|
|
32
|
-
appPath: () => appPath,
|
|
33
|
-
appRootDir: () => appRootDir,
|
|
34
|
-
checkDir: () => checkDir,
|
|
35
|
-
cwdPath: () => cwdPath,
|
|
36
|
-
joinPath: () => joinPath,
|
|
37
|
-
moveFile: () => moveFile,
|
|
38
|
-
readFile: () => readFile,
|
|
39
|
-
readFileAsJSON: () => readFileAsJSON,
|
|
40
|
-
readFiles: () => readFiles,
|
|
41
|
-
removeDirectory: () => removeDirectory,
|
|
42
|
-
removeFile: () => removeFile,
|
|
43
|
-
writeFile: () => writeFile
|
|
44
|
-
});
|
|
45
|
-
module.exports = __toCommonJS(file_provider_exports);
|
|
46
|
-
var import_fs = __toESM(require("fs"));
|
|
47
|
-
var import_os = require("os");
|
|
48
|
-
var import_path = __toESM(require("path"));
|
|
49
|
-
var import_assert = require("../util/assert");
|
|
50
|
-
var import_csv = require("../util/csv.formatter");
|
|
51
|
-
var import_json = require("../util/json.formatter");
|
|
52
|
-
var import_logger = require("../util/logger");
|
|
53
|
-
var import_xml = require("../util/xml.formatter");
|
|
54
|
-
const userHomeDir = (0, import_os.homedir)();
|
|
55
|
-
const appRootDir = process.env.CONTAINER_CONTEXT === "true" ? process.cwd() : import_path.default.join(userHomeDir, ".contensis/");
|
|
56
|
-
const readFile = (filePath) => {
|
|
57
|
-
const directoryPath = appPath(filePath);
|
|
58
|
-
if (import_fs.default.existsSync(directoryPath)) {
|
|
59
|
-
const file = import_fs.default.readFileSync(directoryPath, "utf8");
|
|
60
|
-
return file;
|
|
61
|
-
} else {
|
|
62
|
-
return void 0;
|
|
63
|
-
}
|
|
64
|
-
};
|
|
65
|
-
const readFiles = (directory, createDirectory = true) => {
|
|
66
|
-
const directoryPath = appPath(directory);
|
|
67
|
-
if (import_fs.default.existsSync(directoryPath)) {
|
|
68
|
-
const files = import_fs.default.readdirSync(directoryPath);
|
|
69
|
-
return files;
|
|
70
|
-
} else if (createDirectory) {
|
|
71
|
-
import_fs.default.mkdirSync(directoryPath, { recursive: true });
|
|
72
|
-
return [];
|
|
73
|
-
} else {
|
|
74
|
-
throw new Error(`ENOENT: Directory does not exist ${directoryPath}`);
|
|
75
|
-
}
|
|
76
|
-
};
|
|
77
|
-
const writeFile = (filePath, content) => {
|
|
78
|
-
const directoryPath = appPath(filePath);
|
|
79
|
-
import_fs.default.writeFileSync(directoryPath, content, { encoding: "utf-8" });
|
|
80
|
-
};
|
|
81
|
-
const removeFile = (filePath) => {
|
|
82
|
-
const directoryPath = appPath(filePath);
|
|
83
|
-
if (import_fs.default.existsSync(directoryPath)) {
|
|
84
|
-
import_fs.default.rmSync(directoryPath);
|
|
85
|
-
}
|
|
86
|
-
};
|
|
87
|
-
const removeDirectory = (filePath) => {
|
|
88
|
-
const directoryPath = appPath(filePath);
|
|
89
|
-
if (import_fs.default.existsSync(directoryPath)) {
|
|
90
|
-
import_fs.default.rmSync(directoryPath, { force: true, recursive: true });
|
|
91
|
-
}
|
|
92
|
-
};
|
|
93
|
-
const moveFile = (file, fromPath, toPath) => {
|
|
94
|
-
const from = import_path.default.join(appRootDir, `${fromPath}${file}`);
|
|
95
|
-
const to = import_path.default.join(appRootDir, `${toPath}${file}`);
|
|
96
|
-
if (import_fs.default.existsSync(from)) {
|
|
97
|
-
checkDir(toPath);
|
|
98
|
-
import_fs.default.rename(from, to, (err) => {
|
|
99
|
-
if (err)
|
|
100
|
-
console.error(
|
|
101
|
-
`Could not rename file "${file}" from: ${fromPath} to: ${toPath}`,
|
|
102
|
-
err
|
|
103
|
-
);
|
|
104
|
-
console.info(`Renamed file "${file}" from: ${fromPath} to: ${toPath}`);
|
|
105
|
-
});
|
|
106
|
-
} else {
|
|
107
|
-
console.error(
|
|
108
|
-
`Could not rename file "${file}" from: ${fromPath} to: ${toPath}
|
|
109
|
-
File does not exist!`
|
|
110
|
-
);
|
|
111
|
-
}
|
|
112
|
-
};
|
|
113
|
-
const checkDir = (filePath) => {
|
|
114
|
-
const directoryPath = import_path.default.dirname(appPath(filePath));
|
|
115
|
-
if (!import_fs.default.existsSync(directoryPath))
|
|
116
|
-
import_fs.default.mkdirSync(directoryPath, { recursive: true });
|
|
117
|
-
};
|
|
118
|
-
const appPath = (filePath) => import_path.default.isAbsolute(filePath) ? filePath : import_path.default.join(appRootDir, filePath);
|
|
119
|
-
const cwdPath = (filePath) => import_path.default.isAbsolute(filePath) ? filePath : import_path.default.join(process.cwd(), filePath);
|
|
120
|
-
const joinPath = import_path.default.join;
|
|
121
|
-
const addExecutePermission = (filePath) => (
|
|
122
|
-
// Fails in windows with `TypeError [ERR_INVALID_ARG_TYPE]: The "mode" argument must be of type number. Received undefined`
|
|
123
|
-
import_fs.default.chmodSync(filePath, import_fs.default.constants.S_IRWXU)
|
|
124
|
-
);
|
|
125
|
-
const detectFileType = (fromFile) => {
|
|
126
|
-
const fileData = readFile(fromFile);
|
|
127
|
-
if (!fileData) throw new Error(`Unable to read file at ${fromFile}`);
|
|
128
|
-
try {
|
|
129
|
-
if (fileData.startsWith("<")) return { contents: fileData, type: "xml" };
|
|
130
|
-
const jsonData = (0, import_assert.tryParse)(fileData);
|
|
131
|
-
if (jsonData) return { contents: jsonData, type: "json" };
|
|
132
|
-
const csv = (0, import_csv.detectCsv)(fileData);
|
|
133
|
-
if (csv) return { contents: fileData, type: "csv" };
|
|
134
|
-
} catch (ex) {
|
|
135
|
-
import_logger.Logger.error(`Problem detecting file type ${fromFile}`, ex);
|
|
136
|
-
}
|
|
137
|
-
};
|
|
138
|
-
const readFileAsJSON = async (fromFile) => {
|
|
139
|
-
const detectedFile = detectFileType(cwdPath(fromFile));
|
|
140
|
-
if (!detectedFile) return void 0;
|
|
141
|
-
try {
|
|
142
|
-
switch (detectedFile.type) {
|
|
143
|
-
case "csv": {
|
|
144
|
-
const flatJson = await (0, import_csv.csvToJson)(detectedFile.contents);
|
|
145
|
-
const unflattenedJson = flatJson.map((record) => (0, import_json.unflattenObject)(record));
|
|
146
|
-
return unflattenedJson;
|
|
147
|
-
}
|
|
148
|
-
case "xml":
|
|
149
|
-
return await (0, import_xml.xmlToJson)(detectedFile.contents);
|
|
150
|
-
case "json":
|
|
151
|
-
default:
|
|
152
|
-
return detectedFile.contents;
|
|
153
|
-
}
|
|
154
|
-
} catch (ex) {
|
|
155
|
-
import_logger.Logger.error(`Problem converting file from ${detectedFile.type}`, ex);
|
|
156
|
-
}
|
|
157
|
-
};
|
|
158
|
-
// Annotate the CommonJS export names for ESM import in node:
|
|
159
|
-
0 && (module.exports = {
|
|
160
|
-
addExecutePermission,
|
|
161
|
-
appPath,
|
|
162
|
-
appRootDir,
|
|
163
|
-
checkDir,
|
|
164
|
-
cwdPath,
|
|
165
|
-
joinPath,
|
|
166
|
-
moveFile,
|
|
167
|
-
readFile,
|
|
168
|
-
readFileAsJSON,
|
|
169
|
-
readFiles,
|
|
170
|
-
removeDirectory,
|
|
171
|
-
removeFile,
|
|
172
|
-
writeFile
|
|
173
|
-
});
|
|
174
|
-
//# sourceMappingURL=file-provider.js.map
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
{
|
|
2
|
-
"version": 3,
|
|
3
|
-
"sources": ["../../src/providers/file-provider.ts"],
|
|
4
|
-
"sourcesContent": ["import fs from 'fs';\nimport { homedir } from 'os';\nimport path from 'path';\nimport { tryParse } from '~/util/assert';\nimport { csvToJson, detectCsv } from '~/util/csv.formatter';\nimport { unflattenObject } from '~/util/json.formatter';\nimport { Logger } from '~/util/logger';\nimport { xmlToJson } from '~/util/xml.formatter';\n\nconst userHomeDir = homedir();\n\nexport const appRootDir =\n process.env.CONTAINER_CONTEXT === 'true'\n ? process.cwd()\n : path.join(userHomeDir, '.contensis/');\n\nexport const readFile = (filePath: string) => {\n const directoryPath = appPath(filePath);\n if (fs.existsSync(directoryPath)) {\n const file = fs.readFileSync(directoryPath, 'utf8');\n return file;\n } else {\n return undefined;\n }\n};\n\nexport const readFiles = (directory: string, createDirectory = true) => {\n const directoryPath = appPath(directory);\n if (fs.existsSync(directoryPath)) {\n const files = fs.readdirSync(directoryPath);\n return files;\n } else if (createDirectory) {\n fs.mkdirSync(directoryPath, { recursive: true });\n return [];\n } else {\n throw new Error(`ENOENT: Directory does not exist ${directoryPath}`);\n // return undefined;\n }\n};\n\nexport const writeFile = (filePath: string, content: string) => {\n const directoryPath = appPath(filePath);\n fs.writeFileSync(directoryPath, content, { encoding: 'utf-8' });\n};\n\nexport const removeFile = (filePath: string) => {\n const directoryPath = appPath(filePath);\n if (fs.existsSync(directoryPath)) {\n fs.rmSync(directoryPath);\n }\n};\n\nexport const removeDirectory = (filePath: string) => {\n const directoryPath = appPath(filePath);\n if (fs.existsSync(directoryPath)) {\n fs.rmSync(directoryPath, { force: true, recursive: true });\n }\n};\n\nexport const moveFile = (file: string, fromPath: string, toPath: string) => {\n const from = path.join(appRootDir, `${fromPath}${file}`);\n const to = path.join(appRootDir, `${toPath}${file}`);\n if (fs.existsSync(from)) {\n checkDir(toPath);\n // if (!fs.existsSync(toPath)) fs.mkdirSync(toPath, { recursive: true });\n\n fs.rename(from, to, err => {\n if (err)\n console.error(\n `Could not rename file \"${file}\" from: ${fromPath} to: ${toPath}`,\n err\n );\n console.info(`Renamed file \"${file}\" from: ${fromPath} to: ${toPath}`);\n });\n } else {\n console.error(\n `Could not rename file \"${file}\" from: ${fromPath} to: ${toPath}\\nFile does not exist!`\n );\n }\n};\n\nexport const checkDir = (filePath: string) => {\n const directoryPath = path.dirname(appPath(filePath));\n if (!fs.existsSync(directoryPath))\n fs.mkdirSync(directoryPath, { recursive: true });\n};\n\nexport const appPath = (filePath: string) =>\n path.isAbsolute(filePath) ? filePath : path.join(appRootDir, filePath);\n\nexport const cwdPath = (filePath: string) =>\n path.isAbsolute(filePath) ? filePath : path.join(process.cwd(), filePath);\n\nexport const joinPath = path.join;\n\nexport const addExecutePermission = (filePath: string) =>\n // Fails in windows with `TypeError [ERR_INVALID_ARG_TYPE]: The \"mode\" argument must be of type number. Received undefined`\n fs.chmodSync(filePath, fs.constants.S_IRWXU);\n\ntype DetectedFileType =\n | { type: 'json'; contents: any }\n | { type: 'xml' | 'csv'; contents: string };\n\nconst detectFileType = (fromFile: string): DetectedFileType | undefined => {\n const fileData = readFile(fromFile);\n if (!fileData) throw new Error(`Unable to read file at ${fromFile}`);\n try {\n // if XML\n if (fileData.startsWith('<')) return { contents: fileData, type: 'xml' };\n\n // if JSON\n const jsonData = tryParse(fileData);\n if (jsonData) return { contents: jsonData, type: 'json' };\n\n // if CSV\n const csv = detectCsv(fileData);\n if (csv) return { contents: fileData, type: 'csv' };\n } catch (ex) {\n Logger.error(`Problem detecting file type ${fromFile}`, ex);\n }\n};\n\nexport const readFileAsJSON = async <T = any>(\n fromFile: string\n): Promise<T | undefined> => {\n const detectedFile = detectFileType(cwdPath(fromFile));\n if (!detectedFile) return undefined;\n try {\n switch (detectedFile.type) {\n case 'csv': {\n const flatJson = await csvToJson(detectedFile.contents);\n const unflattenedJson = flatJson.map(record => unflattenObject(record));\n return unflattenedJson as T;\n }\n case 'xml':\n return (await xmlToJson(detectedFile.contents)) as T;\n\n case 'json':\n default:\n return detectedFile.contents;\n }\n } catch (ex) {\n Logger.error(`Problem converting file from ${detectedFile.type}`, ex);\n }\n};\n"],
|
|
5
|
-
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,gBAAe;AACf,gBAAwB;AACxB,kBAAiB;AACjB,oBAAyB;AACzB,iBAAqC;AACrC,kBAAgC;AAChC,oBAAuB;AACvB,iBAA0B;AAE1B,MAAM,kBAAc,mBAAQ;AAErB,MAAM,aACX,QAAQ,IAAI,sBAAsB,SAC9B,QAAQ,IAAI,IACZ,YAAAA,QAAK,KAAK,aAAa,aAAa;AAEnC,MAAM,WAAW,CAAC,aAAqB;AAC5C,QAAM,gBAAgB,QAAQ,QAAQ;AACtC,MAAI,UAAAC,QAAG,WAAW,aAAa,GAAG;AAChC,UAAM,OAAO,UAAAA,QAAG,aAAa,eAAe,MAAM;AAClD,WAAO;AAAA,EACT,OAAO;AACL,WAAO;AAAA,EACT;AACF;AAEO,MAAM,YAAY,CAAC,WAAmB,kBAAkB,SAAS;AACtE,QAAM,gBAAgB,QAAQ,SAAS;AACvC,MAAI,UAAAA,QAAG,WAAW,aAAa,GAAG;AAChC,UAAM,QAAQ,UAAAA,QAAG,YAAY,aAAa;AAC1C,WAAO;AAAA,EACT,WAAW,iBAAiB;AAC1B,cAAAA,QAAG,UAAU,eAAe,EAAE,WAAW,KAAK,CAAC;AAC/C,WAAO,CAAC;AAAA,EACV,OAAO;AACL,UAAM,IAAI,MAAM,oCAAoC,aAAa,EAAE;AAAA,EAErE;AACF;AAEO,MAAM,YAAY,CAAC,UAAkB,YAAoB;AAC9D,QAAM,gBAAgB,QAAQ,QAAQ;AACtC,YAAAA,QAAG,cAAc,eAAe,SAAS,EAAE,UAAU,QAAQ,CAAC;AAChE;AAEO,MAAM,aAAa,CAAC,aAAqB;AAC9C,QAAM,gBAAgB,QAAQ,QAAQ;AACtC,MAAI,UAAAA,QAAG,WAAW,aAAa,GAAG;AAChC,cAAAA,QAAG,OAAO,aAAa;AAAA,EACzB;AACF;AAEO,MAAM,kBAAkB,CAAC,aAAqB;AACnD,QAAM,gBAAgB,QAAQ,QAAQ;AACtC,MAAI,UAAAA,QAAG,WAAW,aAAa,GAAG;AAChC,cAAAA,QAAG,OAAO,eAAe,EAAE,OAAO,MAAM,WAAW,KAAK,CAAC;AAAA,EAC3D;AACF;AAEO,MAAM,WAAW,CAAC,MAAc,UAAkB,WAAmB;AAC1E,QAAM,OAAO,YAAAD,QAAK,KAAK,YAAY,GAAG,QAAQ,GAAG,IAAI,EAAE;AACvD,QAAM,KAAK,YAAAA,QAAK,KAAK,YAAY,GAAG,MAAM,GAAG,IAAI,EAAE;AACnD,MAAI,UAAAC,QAAG,WAAW,IAAI,GAAG;AACvB,aAAS,MAAM;AAGf,cAAAA,QAAG,OAAO,MAAM,IAAI,SAAO;AACzB,UAAI;AACF,gBAAQ;AAAA,UACN,0BAA0B,IAAI,WAAW,QAAQ,QAAQ,MAAM;AAAA,UAC/D;AAAA,QACF;AACF,cAAQ,KAAK,iBAAiB,IAAI,WAAW,QAAQ,QAAQ,MAAM,EAAE;AAAA,IACvE,CAAC;AAAA,EACH,OAAO;AACL,YAAQ;AAAA,MACN,0BAA0B,IAAI,WAAW,QAAQ,QAAQ,MAAM;AAAA;AAAA,IACjE;AAAA,EACF;AACF;AAEO,MAAM,WAAW,CAAC,aAAqB;AAC5C,QAAM,gBAAgB,YAAAD,QAAK,QAAQ,QAAQ,QAAQ,CAAC;AACpD,MAAI,CAAC,UAAAC,QAAG,WAAW,aAAa;AAC9B,cAAAA,QAAG,UAAU,eAAe,EAAE,WAAW,KAAK,CAAC;AACnD;AAEO,MAAM,UAAU,CAAC,aACtB,YAAAD,QAAK,WAAW,QAAQ,IAAI,WAAW,YAAAA,QAAK,KAAK,YAAY,QAAQ;AAEhE,MAAM,UAAU,CAAC,aACtB,YAAAA,QAAK,WAAW,QAAQ,IAAI,WAAW,YAAAA,QAAK,KAAK,QAAQ,IAAI,GAAG,QAAQ;AAEnE,MAAM,WAAW,YAAAA,QAAK;AAEtB,MAAM,uBAAuB,CAAC;AAAA;AAAA,EAEnC,UAAAC,QAAG,UAAU,UAAU,UAAAA,QAAG,UAAU,OAAO;AAAA;AAM7C,MAAM,iBAAiB,CAAC,aAAmD;AACzE,QAAM,WAAW,SAAS,QAAQ;AAClC,MAAI,CAAC,SAAU,OAAM,IAAI,MAAM,0BAA0B,QAAQ,EAAE;AACnE,MAAI;AAEF,QAAI,SAAS,WAAW,GAAG,EAAG,QAAO,EAAE,UAAU,UAAU,MAAM,MAAM;AAGvE,UAAM,eAAW,wBAAS,QAAQ;AAClC,QAAI,SAAU,QAAO,EAAE,UAAU,UAAU,MAAM,OAAO;AAGxD,UAAM,UAAM,sBAAU,QAAQ;AAC9B,QAAI,IAAK,QAAO,EAAE,UAAU,UAAU,MAAM,MAAM;AAAA,EACpD,SAAS,IAAI;AACX,yBAAO,MAAM,+BAA+B,QAAQ,IAAI,EAAE;AAAA,EAC5D;AACF;AAEO,MAAM,iBAAiB,OAC5B,aAC2B;AAC3B,QAAM,eAAe,eAAe,QAAQ,QAAQ,CAAC;AACrD,MAAI,CAAC,aAAc,QAAO;AAC1B,MAAI;AACF,YAAQ,aAAa,MAAM;AAAA,MACzB,KAAK,OAAO;AACV,cAAM,WAAW,UAAM,sBAAU,aAAa,QAAQ;AACtD,cAAM,kBAAkB,SAAS,IAAI,gBAAU,6BAAgB,MAAM,CAAC;AACtE,eAAO;AAAA,MACT;AAAA,MACA,KAAK;AACH,eAAQ,UAAM,sBAAU,aAAa,QAAQ;AAAA,MAE/C,KAAK;AAAA,MACL;AACE,eAAO,aAAa;AAAA,IACxB;AAAA,EACF,SAAS,IAAI;AACX,yBAAO,MAAM,gCAAgC,aAAa,IAAI,IAAI,EAAE;AAAA,EACtE;AACF;",
|
|
6
|
-
"names": ["path", "fs"]
|
|
7
|
-
}
|