@willbooster/shared-lib-node 2.4.0 → 2.5.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cjs/env.cjs +1 -1
- package/dist/cjs/env.cjs.map +1 -1
- package/dist/cjs/hash.cjs +2 -0
- package/dist/cjs/hash.cjs.map +1 -0
- package/dist/cjs/hash.d.cts +3 -0
- package/dist/cjs/index.cjs +1 -1
- package/dist/cjs/index.d.cts +1 -0
- package/dist/esm/env.js +1 -1
- package/dist/esm/env.js.map +1 -1
- package/dist/esm/hash.d.ts +3 -0
- package/dist/esm/hash.js +2 -0
- package/dist/esm/hash.js.map +1 -0
- package/dist/esm/index.d.ts +1 -0
- package/dist/esm/index.js +1 -1
- package/package.json +9 -9
package/dist/cjs/env.cjs
CHANGED
|
@@ -1,2 +1,2 @@
|
|
|
1
|
-
"use strict";var e=require("node:path"),t=require("dotenv");exports.loadEnvironmentVariables=function(n,o){let r=(n.env??[]).map((e=>e.toString()));const s=n.cascadeNodeEnv?process.env.NODE_ENV:n.cascadeEnv;"string"==typeof s&&(0===r.length&&r.push(".env"),r=r.flatMap((e=>s?[`${e}.${s}.local`,`${e}.local`,`${e}.${s}`,e]:[`${e}.local`,e]))),n.verbose&&console.info("Loading env files:",r);let a={};for(const n of r)a={...t.config({path:e.join(o,n)}).parsed,...a};return a},exports.removeNpmAndYarnEnvironmentVariables=function(e){for(const t of Object.keys(e)){const n=t.toLowerCase();(n.startsWith("npm_")||n.startsWith("yarn_")||n.startsWith("berry_"))&&delete e[t]}};
|
|
1
|
+
"use strict";var e=require("node:path"),t=require("dotenv");exports.loadEnvironmentVariables=function(n,o){let r=(n.env??[]).map((e=>e.toString()));const s=n.cascadeNodeEnv?process.env.NODE_ENV??"":n.cascadeEnv;"string"==typeof s&&(0===r.length&&r.push(".env"),r=r.flatMap((e=>s?[`${e}.${s}.local`,`${e}.local`,`${e}.${s}`,e]:[`${e}.local`,e]))),n.verbose&&console.info("Loading env files:",r);let a={};for(const n of r)a={...t.config({path:e.join(o,n)}).parsed,...a};return a},exports.removeNpmAndYarnEnvironmentVariables=function(e){for(const t of Object.keys(e)){const n=t.toLowerCase();(n.startsWith("npm_")||n.startsWith("yarn_")||n.startsWith("berry_"))&&delete e[t]}};
|
|
2
2
|
//# sourceMappingURL=env.cjs.map
|
package/dist/cjs/env.cjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"env.cjs","sources":["../../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n verbose?: boolean;\n}\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => envPath.toString());\n const cascade = argv.cascadeNodeEnv ? process.env.NODE_ENV : argv.cascadeEnv;\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push('.env');\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars = {};\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, unknown>): void {\n // Remove npm & yarn environment variables from process.env\n for (const key of Object.keys(envVars)) {\n const lowerKey = key.toLowerCase();\n if (lowerKey.startsWith('npm_') || lowerKey.startsWith('yarn_') || lowerKey.startsWith('berry_')) {\n delete envVars[key];\n }\n }\n}\n"],"names":["argv","cwd","envPaths","env","map","envPath","toString","cascade","cascadeNodeEnv","process","NODE_ENV","cascadeEnv","length","push","flatMap","verbose","console","info","envVars","config","path","join","parsed","key","Object","keys","lowerKey","toLowerCase","startsWith"],"mappings":"6FAcO,SAAkCA,EAAeC,GACtD,IAAIC,GAAYF,EAAKG,KAAO,IAAIC,KAAKC,GAAYA,EAAQC,aACzD,MAAMC,EAAUP,EAAKQ,eAAiBC,QAAQN,IAAIO,
|
|
1
|
+
{"version":3,"file":"env.cjs","sources":["../../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n verbose?: boolean;\n}\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => envPath.toString());\n const cascade = argv.cascadeNodeEnv ? process.env.NODE_ENV ?? '' : argv.cascadeEnv;\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push('.env');\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars = {};\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, unknown>): void {\n // Remove npm & yarn environment variables from process.env\n for (const key of Object.keys(envVars)) {\n const lowerKey = key.toLowerCase();\n if (lowerKey.startsWith('npm_') || lowerKey.startsWith('yarn_') || lowerKey.startsWith('berry_')) {\n delete envVars[key];\n }\n }\n}\n"],"names":["argv","cwd","envPaths","env","map","envPath","toString","cascade","cascadeNodeEnv","process","NODE_ENV","cascadeEnv","length","push","flatMap","verbose","console","info","envVars","config","path","join","parsed","key","Object","keys","lowerKey","toLowerCase","startsWith"],"mappings":"6FAcO,SAAkCA,EAAeC,GACtD,IAAIC,GAAYF,EAAKG,KAAO,IAAIC,KAAKC,GAAYA,EAAQC,aACzD,MAAMC,EAAUP,EAAKQ,eAAiBC,QAAQN,IAAIO,UAAY,GAAKV,EAAKW,WACjD,iBAAZJ,IACe,IAApBL,EAASU,QAAcV,EAASW,KAAK,QACzCX,EAAWA,EAASY,SAAST,GAC3BE,EACI,CAAE,GAAEF,KAAWE,UAAkB,GAAEF,UAAkB,GAAEA,KAAWE,IAAWF,GAC7E,CAAE,GAAEA,UAAiBA,MAGzBL,EAAKe,SACPC,QAAQC,KAAK,qBAAsBf,GAGrC,IAAIgB,EAAU,CAAA,EACd,IAAK,MAAMb,KAAWH,EACpBgB,EAAU,IAAKC,SAAO,CAAEC,KAAMA,EAAKC,KAAKpB,EAAKI,KAAYiB,UAAWJ,GAEtE,OAAOA,CACT,+CAKO,SAA8CA,GAEnD,IAAK,MAAMK,KAAOC,OAAOC,KAAKP,GAAU,CACtC,MAAMQ,EAAWH,EAAII,eACjBD,EAASE,WAAW,SAAWF,EAASE,WAAW,UAAYF,EAASE,WAAW,mBAC9EV,EAAQK,EAEnB,CACF"}
|
|
@@ -0,0 +1,2 @@
|
|
|
1
|
+
"use strict";var e=require("node:crypto"),s=require("node:fs"),t=require("node:path");async function i(i){const r=e.createHash("sha512");for(const e of i.sort()){const i=await s.promises.stat(e);if(i.isDirectory()){const i=await s.promises.readdir(e,{withFileTypes:!0,recursive:!0});for(const a of i.sort(((e,s)=>e.name.localeCompare(s.name))))a.isFile()&&r.update(await s.promises.readFile(t.join(e,a.name),"utf8"))}else i.isFile()&&r.update(await s.promises.readFile(e,"utf8"))}return r.digest("hex")}exports.calculateHashFromFiles=i,exports.updateHashFromFiles=async function(e,t){const r=await s.promises.readFile(e,"utf8"),a=await i(t);return r!==a&&(await s.promises.writeFile(e,a,"utf8"),!0)};
|
|
2
|
+
//# sourceMappingURL=hash.cjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"hash.cjs","sources":["../../src/hash.ts"],"sourcesContent":["import crypto from 'node:crypto';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\nexport async function calculateHashFromFiles(paths: string[]): Promise<string> {\n const hash = crypto.createHash('sha512');\n for (const fileOrDirPath of paths.sort()) {\n const stat = await fs.promises.stat(fileOrDirPath);\n if (stat.isDirectory()) {\n // Get all files in the directory\n const dirents = await fs.promises.readdir(fileOrDirPath, { withFileTypes: true, recursive: true });\n for (const dirent of dirents.sort((d1, d2) => d1.name.localeCompare(d2.name))) {\n if (dirent.isFile()) {\n hash.update(await fs.promises.readFile(path.join(fileOrDirPath, dirent.name), 'utf8'));\n }\n }\n } else if (stat.isFile()) {\n hash.update(await fs.promises.readFile(fileOrDirPath, 'utf8'));\n }\n }\n return hash.digest('hex');\n}\n\nexport async function updateHashFromFiles(hashFilePath: string, paths: string[]): Promise<boolean> {\n const oldHash = await fs.promises.readFile(hashFilePath, 'utf8');\n const newHash = await calculateHashFromFiles(paths);\n if (oldHash === newHash) return false;\n\n await fs.promises.writeFile(hashFilePath, newHash, 'utf8');\n return true;\n}\n"],"names":["async","calculateHashFromFiles","paths","hash","crypto","createHash","fileOrDirPath","sort","stat","fs","promises","isDirectory","dirents","readdir","withFileTypes","recursive","dirent","d1","d2","name","localeCompare","isFile","update","readFile","path","join","digest","hashFilePath","oldHash","newHash","writeFile"],"mappings":"sFAIOA,eAAeC,EAAuBC,GAC3C,MAAMC,EAAOC,EAAOC,WAAW,UAC/B,IAAK,MAAMC,KAAiBJ,EAAMK,OAAQ,CACxC,MAAMC,QAAaC,EAAGC,SAASF,KAAKF,GACpC,GAAIE,EAAKG,cAAe,CAEtB,MAAMC,QAAgBH,EAAGC,SAASG,QAAQP,EAAe,CAAEQ,eAAe,EAAMC,WAAW,IAC3F,IAAK,MAAMC,KAAUJ,EAAQL,MAAK,CAACU,EAAIC,IAAOD,EAAGE,KAAKC,cAAcF,EAAGC,QACjEH,EAAOK,UACTlB,EAAKmB,aAAab,EAAGC,SAASa,SAASC,EAAKC,KAAKnB,EAAeU,EAAOG,MAAO,QAGpF,MAAWX,EAAKa,UACdlB,EAAKmB,aAAab,EAAGC,SAASa,SAASjB,EAAe,QAE1D,CACA,OAAOH,EAAKuB,OAAO,MACrB,8DAEO1B,eAAmC2B,EAAsBzB,GAC9D,MAAM0B,QAAgBnB,EAAGC,SAASa,SAASI,EAAc,QACnDE,QAAgB5B,EAAuBC,GAC7C,OAAI0B,IAAYC,UAEVpB,EAAGC,SAASoB,UAAUH,EAAcE,EAAS,SAC5C,EACT"}
|
package/dist/cjs/index.cjs
CHANGED
|
@@ -1,2 +1,2 @@
|
|
|
1
|
-
"use strict";var e=require("./env.cjs"),s=require("./exists.cjs"),
|
|
1
|
+
"use strict";var e=require("./env.cjs"),s=require("./exists.cjs"),r=require("./hash.cjs"),a=require("./spawn.cjs");exports.loadEnvironmentVariables=e.loadEnvironmentVariables,exports.removeNpmAndYarnEnvironmentVariables=e.removeNpmAndYarnEnvironmentVariables,exports.existsAsync=s.existsAsync,exports.calculateHashFromFiles=r.calculateHashFromFiles,exports.updateHashFromFiles=r.updateHashFromFiles,exports.spawnAsync=a.spawnAsync;
|
|
2
2
|
//# sourceMappingURL=index.cjs.map
|
package/dist/cjs/index.d.cts
CHANGED
package/dist/esm/env.js
CHANGED
|
@@ -1,2 +1,2 @@
|
|
|
1
|
-
import t from"node:path";import{config as o}from"dotenv";function e(e,n){let s=(e.env??[]).map((t=>t.toString()));const r=e.cascadeNodeEnv?process.env.NODE_ENV:e.cascadeEnv;"string"==typeof r&&(0===s.length&&s.push(".env"),s=s.flatMap((t=>r?[`${t}.${r}.local`,`${t}.local`,`${t}.${r}`,t]:[`${t}.local`,t]))),e.verbose&&console.info("Loading env files:",s);let a={};for(const e of s)a={...o({path:t.join(n,e)}).parsed,...a};return a}function n(t){for(const o of Object.keys(t)){const e=o.toLowerCase();(e.startsWith("npm_")||e.startsWith("yarn_")||e.startsWith("berry_"))&&delete t[o]}}export{e as loadEnvironmentVariables,n as removeNpmAndYarnEnvironmentVariables};
|
|
1
|
+
import t from"node:path";import{config as o}from"dotenv";function e(e,n){let s=(e.env??[]).map((t=>t.toString()));const r=e.cascadeNodeEnv?process.env.NODE_ENV??"":e.cascadeEnv;"string"==typeof r&&(0===s.length&&s.push(".env"),s=s.flatMap((t=>r?[`${t}.${r}.local`,`${t}.local`,`${t}.${r}`,t]:[`${t}.local`,t]))),e.verbose&&console.info("Loading env files:",s);let a={};for(const e of s)a={...o({path:t.join(n,e)}).parsed,...a};return a}function n(t){for(const o of Object.keys(t)){const e=o.toLowerCase();(e.startsWith("npm_")||e.startsWith("yarn_")||e.startsWith("berry_"))&&delete t[o]}}export{e as loadEnvironmentVariables,n as removeNpmAndYarnEnvironmentVariables};
|
|
2
2
|
//# sourceMappingURL=env.js.map
|
package/dist/esm/env.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"env.js","sources":["../../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n verbose?: boolean;\n}\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => envPath.toString());\n const cascade = argv.cascadeNodeEnv ? process.env.NODE_ENV : argv.cascadeEnv;\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push('.env');\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars = {};\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, unknown>): void {\n // Remove npm & yarn environment variables from process.env\n for (const key of Object.keys(envVars)) {\n const lowerKey = key.toLowerCase();\n if (lowerKey.startsWith('npm_') || lowerKey.startsWith('yarn_') || lowerKey.startsWith('berry_')) {\n delete envVars[key];\n }\n }\n}\n"],"names":["loadEnvironmentVariables","argv","cwd","envPaths","env","map","envPath","toString","cascade","cascadeNodeEnv","process","NODE_ENV","cascadeEnv","length","push","flatMap","verbose","console","info","envVars","config","path","join","parsed","removeNpmAndYarnEnvironmentVariables","key","Object","keys","lowerKey","toLowerCase","startsWith"],"mappings":"yDAcO,SAASA,EAAyBC,EAAeC,GACtD,IAAIC,GAAYF,EAAKG,KAAO,IAAIC,KAAKC,GAAYA,EAAQC,aACzD,MAAMC,EAAUP,EAAKQ,eAAiBC,QAAQN,IAAIO,
|
|
1
|
+
{"version":3,"file":"env.js","sources":["../../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n verbose?: boolean;\n}\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => envPath.toString());\n const cascade = argv.cascadeNodeEnv ? process.env.NODE_ENV ?? '' : argv.cascadeEnv;\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push('.env');\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars = {};\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, unknown>): void {\n // Remove npm & yarn environment variables from process.env\n for (const key of Object.keys(envVars)) {\n const lowerKey = key.toLowerCase();\n if (lowerKey.startsWith('npm_') || lowerKey.startsWith('yarn_') || lowerKey.startsWith('berry_')) {\n delete envVars[key];\n }\n }\n}\n"],"names":["loadEnvironmentVariables","argv","cwd","envPaths","env","map","envPath","toString","cascade","cascadeNodeEnv","process","NODE_ENV","cascadeEnv","length","push","flatMap","verbose","console","info","envVars","config","path","join","parsed","removeNpmAndYarnEnvironmentVariables","key","Object","keys","lowerKey","toLowerCase","startsWith"],"mappings":"yDAcO,SAASA,EAAyBC,EAAeC,GACtD,IAAIC,GAAYF,EAAKG,KAAO,IAAIC,KAAKC,GAAYA,EAAQC,aACzD,MAAMC,EAAUP,EAAKQ,eAAiBC,QAAQN,IAAIO,UAAY,GAAKV,EAAKW,WACjD,iBAAZJ,IACe,IAApBL,EAASU,QAAcV,EAASW,KAAK,QACzCX,EAAWA,EAASY,SAAST,GAC3BE,EACI,CAAE,GAAEF,KAAWE,UAAkB,GAAEF,UAAkB,GAAEA,KAAWE,IAAWF,GAC7E,CAAE,GAAEA,UAAiBA,MAGzBL,EAAKe,SACPC,QAAQC,KAAK,qBAAsBf,GAGrC,IAAIgB,EAAU,CAAA,EACd,IAAK,MAAMb,KAAWH,EACpBgB,EAAU,IAAKC,EAAO,CAAEC,KAAMA,EAAKC,KAAKpB,EAAKI,KAAYiB,UAAWJ,GAEtE,OAAOA,CACT,CAKO,SAASK,EAAqCL,GAEnD,IAAK,MAAMM,KAAOC,OAAOC,KAAKR,GAAU,CACtC,MAAMS,EAAWH,EAAII,eACjBD,EAASE,WAAW,SAAWF,EAASE,WAAW,UAAYF,EAASE,WAAW,mBAC9EX,EAAQM,EAEnB,CACF"}
|
package/dist/esm/hash.js
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
1
|
+
import e from"node:crypto";import t from"node:fs";import i from"node:path";async function o(o){const r=e.createHash("sha512");for(const e of o.sort()){const o=await t.promises.stat(e);if(o.isDirectory()){const o=await t.promises.readdir(e,{withFileTypes:!0,recursive:!0});for(const s of o.sort(((e,t)=>e.name.localeCompare(t.name))))s.isFile()&&r.update(await t.promises.readFile(i.join(e,s.name),"utf8"))}else o.isFile()&&r.update(await t.promises.readFile(e,"utf8"))}return r.digest("hex")}async function r(e,i){const r=await t.promises.readFile(e,"utf8"),s=await o(i);return r!==s&&(await t.promises.writeFile(e,s,"utf8"),!0)}export{o as calculateHashFromFiles,r as updateHashFromFiles};
|
|
2
|
+
//# sourceMappingURL=hash.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"hash.js","sources":["../../src/hash.ts"],"sourcesContent":["import crypto from 'node:crypto';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\nexport async function calculateHashFromFiles(paths: string[]): Promise<string> {\n const hash = crypto.createHash('sha512');\n for (const fileOrDirPath of paths.sort()) {\n const stat = await fs.promises.stat(fileOrDirPath);\n if (stat.isDirectory()) {\n // Get all files in the directory\n const dirents = await fs.promises.readdir(fileOrDirPath, { withFileTypes: true, recursive: true });\n for (const dirent of dirents.sort((d1, d2) => d1.name.localeCompare(d2.name))) {\n if (dirent.isFile()) {\n hash.update(await fs.promises.readFile(path.join(fileOrDirPath, dirent.name), 'utf8'));\n }\n }\n } else if (stat.isFile()) {\n hash.update(await fs.promises.readFile(fileOrDirPath, 'utf8'));\n }\n }\n return hash.digest('hex');\n}\n\nexport async function updateHashFromFiles(hashFilePath: string, paths: string[]): Promise<boolean> {\n const oldHash = await fs.promises.readFile(hashFilePath, 'utf8');\n const newHash = await calculateHashFromFiles(paths);\n if (oldHash === newHash) return false;\n\n await fs.promises.writeFile(hashFilePath, newHash, 'utf8');\n return true;\n}\n"],"names":["async","calculateHashFromFiles","paths","hash","crypto","createHash","fileOrDirPath","sort","stat","fs","promises","isDirectory","dirents","readdir","withFileTypes","recursive","dirent","d1","d2","name","localeCompare","isFile","update","readFile","path","join","digest","updateHashFromFiles","hashFilePath","oldHash","newHash","writeFile"],"mappings":"2EAIOA,eAAeC,EAAuBC,GAC3C,MAAMC,EAAOC,EAAOC,WAAW,UAC/B,IAAK,MAAMC,KAAiBJ,EAAMK,OAAQ,CACxC,MAAMC,QAAaC,EAAGC,SAASF,KAAKF,GACpC,GAAIE,EAAKG,cAAe,CAEtB,MAAMC,QAAgBH,EAAGC,SAASG,QAAQP,EAAe,CAAEQ,eAAe,EAAMC,WAAW,IAC3F,IAAK,MAAMC,KAAUJ,EAAQL,MAAK,CAACU,EAAIC,IAAOD,EAAGE,KAAKC,cAAcF,EAAGC,QACjEH,EAAOK,UACTlB,EAAKmB,aAAab,EAAGC,SAASa,SAASC,EAAKC,KAAKnB,EAAeU,EAAOG,MAAO,QAGpF,MAAWX,EAAKa,UACdlB,EAAKmB,aAAab,EAAGC,SAASa,SAASjB,EAAe,QAE1D,CACA,OAAOH,EAAKuB,OAAO,MACrB,CAEO1B,eAAe2B,EAAoBC,EAAsB1B,GAC9D,MAAM2B,QAAgBpB,EAAGC,SAASa,SAASK,EAAc,QACnDE,QAAgB7B,EAAuBC,GAC7C,OAAI2B,IAAYC,UAEVrB,EAAGC,SAASqB,UAAUH,EAAcE,EAAS,SAC5C,EACT"}
|
package/dist/esm/index.d.ts
CHANGED
package/dist/esm/index.js
CHANGED
|
@@ -1,2 +1,2 @@
|
|
|
1
|
-
export{loadEnvironmentVariables,removeNpmAndYarnEnvironmentVariables}from"./env.js";export{existsAsync}from"./exists.js";export{spawnAsync}from"./spawn.js";
|
|
1
|
+
export{loadEnvironmentVariables,removeNpmAndYarnEnvironmentVariables}from"./env.js";export{existsAsync}from"./exists.js";export{calculateHashFromFiles,updateHashFromFiles}from"./hash.js";export{spawnAsync}from"./spawn.js";
|
|
2
2
|
//# sourceMappingURL=index.js.map
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@willbooster/shared-lib-node",
|
|
3
|
-
"version": "2.
|
|
3
|
+
"version": "2.5.0",
|
|
4
4
|
"license": "Apache-2.0",
|
|
5
5
|
"author": "WillBooster Inc.",
|
|
6
6
|
"sideEffects": false,
|
|
@@ -40,21 +40,21 @@
|
|
|
40
40
|
"tree-kill": "1.2.2"
|
|
41
41
|
},
|
|
42
42
|
"devDependencies": {
|
|
43
|
-
"@types/eslint": "8.44.
|
|
43
|
+
"@types/eslint": "8.44.1",
|
|
44
44
|
"@types/micromatch": "4.0.2",
|
|
45
45
|
"@types/prettier": "2.7.3",
|
|
46
|
-
"@typescript-eslint/eslint-plugin": "6.
|
|
47
|
-
"@typescript-eslint/parser": "6.
|
|
46
|
+
"@typescript-eslint/eslint-plugin": "6.2.0",
|
|
47
|
+
"@typescript-eslint/parser": "6.2.0",
|
|
48
48
|
"@willbooster/eslint-config-ts": "10.4.0",
|
|
49
49
|
"@willbooster/prettier-config": "9.1.1",
|
|
50
|
-
"build-ts": "7.0.
|
|
51
|
-
"eslint": "8.
|
|
52
|
-
"eslint-config-prettier": "8.
|
|
50
|
+
"build-ts": "7.0.2",
|
|
51
|
+
"eslint": "8.46.0",
|
|
52
|
+
"eslint-config-prettier": "8.9.0",
|
|
53
53
|
"eslint-import-resolver-typescript": "3.5.5",
|
|
54
|
-
"eslint-plugin-import": "2.
|
|
54
|
+
"eslint-plugin-import": "2.28.0",
|
|
55
55
|
"eslint-plugin-sort-class-members": "1.18.0",
|
|
56
56
|
"eslint-plugin-sort-destructure-keys": "1.5.0",
|
|
57
|
-
"eslint-plugin-unicorn": "48.0.
|
|
57
|
+
"eslint-plugin-unicorn": "48.0.1",
|
|
58
58
|
"lint-staged": "13.2.3",
|
|
59
59
|
"micromatch": "4.0.5",
|
|
60
60
|
"prettier": "3.0.0",
|