@willbooster/shared-lib-node 3.2.0 → 3.2.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/env.cjs CHANGED
@@ -1,2 +1,2 @@
1
- "use strict";var e=require("node:path"),n=require("dotenv");exports.loadEnvironmentVariables=function(o,t,a){let r=(o.env??[]).map((n=>e.resolve(a??t,n.toString())));const s=o.cascadeEnv??(o.cascadeNodeEnv?process.env.NODE_ENV||"development":o.autoCascadeEnv?process.env.WB_ENV||process.env.APP_ENV||process.env.NODE_ENV||"development":void 0);"string"==typeof s&&(0===r.length&&r.push(e.join(t,".env")),r=r.flatMap((e=>s?[`${e}.${s}.local`,`${e}.local`,`${e}.${s}`,e]:[`${e}.local`,e]))),r=r.map((n=>e.relative(t,n))),o.verbose&&console.info("Loading env files:",r);let i={};const c={...process.env};for(const o of r){i={...n.config({path:e.join(t,o)}).parsed,...i};let a=0;for(const[e,n]of Object.entries(i))c[e]!==n&&(c[e]=n,a++);a>0&&console.info(`Updated ${a} environment variables:`,o)}if(o.checkEnv){const a=Object.keys(n.config({path:e.join(t,o.checkEnv)}).parsed||{}).filter((e=>!(e in i)));if(a.length>0)throw new Error(`Missing environment variables in [${r.join(", ")}]: [${a.join(", ")}]`)}return i},exports.removeNpmAndYarnEnvironmentVariables=function(e){e.PATH&&e.BERRY_BIN_FOLDER&&(e.PATH=e.PATH.replace(`${e.BERRY_BIN_FOLDER}:`,"").replaceAll(/\/private\/var\/folders\/[^:]+:/g,"").replaceAll(/\/var\/tmp\/[^:]+:/g,"").replaceAll(/\/tmp\/[^:]+:/g,""));for(const n of Object.keys(e)){const o=n.toUpperCase();(o.startsWith("NPM_")||o.startsWith("YARN_")||o.startsWith("BERRY_")||"PROJECT_CWD"===o||"INIT_CWD"===o)&&delete e[n]}},exports.yargsOptionsBuilderForEnv={env:{description:".env files to be loaded.",type:"array"},"cascade-env":{description:"Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.",type:"string"},"cascade-node-env":{description:'Same with --cascade-env=<NODE_ENV>. If NODE_ENV is falsy, "development" is applied. Preferred over `auto-cascade-env`.',type:"boolean"},"auto-cascade-env":{description:'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV>. If they are falsy, "development" is applied.',type:"boolean",default:!0},"check-env":{description:"Check whether the keys of the loaded .env files are same with the given .env file.",type:"string",default:".env.example"}};
1
+ "use strict";var e=require("node:path"),n=require("dotenv");exports.loadEnvironmentVariables=function(o,t,a){let r=(o.env??[]).map((n=>e.resolve(a??t,n.toString())));const s=o.cascadeEnv??(o.cascadeNodeEnv?process.env.NODE_ENV||"development":o.autoCascadeEnv?process.env.WB_ENV||process.env.APP_ENV||process.env.NODE_ENV||"development":void 0);"string"==typeof s&&(0===r.length&&r.push(e.join(t,".env")),r=r.flatMap((e=>s?[`${e}.${s}.local`,`${e}.local`,`${e}.${s}`,e]:[`${e}.local`,e]))),r=r.map((n=>e.relative(t,n))),o.verbose&&console.info("Loading env files:",r);let c={};const i={...process.env};for(const o of r){c={...n.config({path:e.join(t,o)}).parsed,...c};let a=0;for(const[e,n]of Object.entries(c))i[e]!==n&&(i[e]=n,a++);a>0&&console.info(`Updated ${a} environment variables:`,o)}if(o.checkEnv){const a=Object.keys(n.config({path:e.join(t,o.checkEnv)}).parsed||{}).filter((e=>!(e in c)));if(a.length>0)throw new Error(`Missing environment variables in [${r.join(", ")}]: [${a.join(", ")}]`)}return c},exports.removeNpmAndYarnEnvironmentVariables=function(e){e.PATH&&e.BERRY_BIN_FOLDER&&(e.PATH=e.PATH.replace(`${e.BERRY_BIN_FOLDER}:`,"").replaceAll(/\/private\/var\/folders\/[^:]+:/g,"").replaceAll(/\/var\/tmp\/[^:]+:/g,"").replaceAll(/\/tmp\/[^:]+:/g,""));for(const n of Object.keys(e)){const o=n.toUpperCase();(o.startsWith("NPM_")||o.startsWith("YARN_")||o.startsWith("BERRY_")||"PROJECT_CWD"===o||"INIT_CWD"===o)&&delete e[n]}},exports.yargsOptionsBuilderForEnv={env:{description:".env files to be loaded.",type:"array"},"cascade-env":{description:"Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.",type:"string"},"cascade-node-env":{description:'Same with --cascade-env=<NODE_ENV || "development">. Preferred over `auto-cascade-env`.',type:"boolean"},"auto-cascade-env":{description:'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV || "">.',type:"boolean",default:!0},"check-env":{description:"Check whether the keys of the loaded .env files are same with the given .env file.",type:"string",default:".env.example"}};
2
2
  //# sourceMappingURL=env.cjs.map
package/dist/env.cjs.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"env.cjs","sources":["../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n autoCascadeEnv?: boolean;\n checkEnv?: string;\n verbose?: boolean;\n}\n\nexport const yargsOptionsBuilderForEnv = {\n env: {\n description: '.env files to be loaded.',\n type: 'array',\n },\n 'cascade-env': {\n description:\n 'Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.',\n type: 'string',\n },\n 'cascade-node-env': {\n description:\n 'Same with --cascade-env=<NODE_ENV>. If NODE_ENV is falsy, \"development\" is applied. Preferred over `auto-cascade-env`.',\n type: 'boolean',\n },\n 'auto-cascade-env': {\n description:\n 'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV>. If they are falsy, \"development\" is applied.',\n type: 'boolean',\n default: true,\n },\n 'check-env': {\n description: 'Check whether the keys of the loaded .env files are same with the given .env file.',\n type: 'string',\n default: '.env.example',\n },\n} as const;\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string, orgCwd?: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => path.resolve(orgCwd ?? cwd, envPath.toString()));\n const cascade =\n argv.cascadeEnv ??\n (argv.cascadeNodeEnv\n ? process.env.NODE_ENV || 'development'\n : argv.autoCascadeEnv\n ? process.env.WB_ENV || process.env.APP_ENV || process.env.NODE_ENV || 'development'\n : undefined);\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push(path.join(cwd, '.env'));\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n envPaths = envPaths.map((envPath) => path.relative(cwd, envPath));\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars: Record<string, string> = {};\n const orgEnvVars = { ...process.env };\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n let count = 0;\n for (const [key, value] of Object.entries(envVars)) {\n if (orgEnvVars[key] !== value) {\n orgEnvVars[key] = value;\n count++;\n }\n }\n if (count > 0) {\n console.info(`Updated ${count} environment variables:`, envPath);\n }\n }\n\n if (argv.checkEnv) {\n const exampleKeys = Object.keys(config({ path: path.join(cwd, argv.checkEnv) }).parsed || {});\n const missingKeys = exampleKeys.filter((key) => !(key in envVars));\n if (missingKeys.length > 0) {\n throw new Error(`Missing environment variables in [${envPaths.join(', ')}]: [${missingKeys.join(', ')}]`);\n }\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, string | undefined>): void {\n // Remove npm & yarn environment variables from process.env\n if (envVars.PATH && envVars.BERRY_BIN_FOLDER) {\n envVars.PATH = envVars.PATH.replace(`${envVars.BERRY_BIN_FOLDER}:`, '')\n // Temporary directory in macOS\n .replaceAll(/\\/private\\/var\\/folders\\/[^:]+:/g, '')\n // Temporary directories in Linux\n .replaceAll(/\\/var\\/tmp\\/[^:]+:/g, '')\n .replaceAll(/\\/tmp\\/[^:]+:/g, '');\n }\n for (const key of Object.keys(envVars)) {\n const upperKey = key.toUpperCase();\n if (\n upperKey.startsWith('NPM_') ||\n upperKey.startsWith('YARN_') ||\n upperKey.startsWith('BERRY_') ||\n upperKey === 'PROJECT_CWD' ||\n upperKey === 'INIT_CWD'\n ) {\n delete envVars[key];\n }\n }\n}\n"],"names":["argv","cwd","orgCwd","envPaths","env","map","envPath","path","resolve","toString","cascade","cascadeEnv","cascadeNodeEnv","process","NODE_ENV","autoCascadeEnv","WB_ENV","APP_ENV","undefined","length","push","join","flatMap","relative","verbose","console","info","envVars","orgEnvVars","config","parsed","count","key","value","Object","entries","checkEnv","missingKeys","keys","filter","Error","PATH","BERRY_BIN_FOLDER","replace","replaceAll","upperKey","toUpperCase","startsWith","description","type","default"],"mappings":"6FA4CO,SAAkCA,EAAeC,EAAaC,GACnE,IAAIC,GAAYH,EAAKI,KAAO,IAAIC,KAAKC,GAAYC,EAAKC,QAAQN,GAAUD,EAAKK,EAAQG,cACrF,MAAMC,EACJV,EAAKW,aACJX,EAAKY,eACFC,QAAQT,IAAIU,UAAY,cACxBd,EAAKe,eACLF,QAAQT,IAAIY,QAAUH,QAAQT,IAAIa,SAAWJ,QAAQT,IAAIU,UAAY,mBACrEI,GACiB,iBAAZR,IACe,IAApBP,EAASgB,QAAchB,EAASiB,KAAKb,EAAKc,KAAKpB,EAAK,SACxDE,EAAWA,EAASmB,SAAShB,GAC3BI,EACI,CAAE,GAAEJ,KAAWI,UAAkB,GAAEJ,UAAkB,GAAEA,KAAWI,IAAWJ,GAC7E,CAAE,GAAEA,UAAiBA,MAG7BH,EAAWA,EAASE,KAAKC,GAAYC,EAAKgB,SAAStB,EAAKK,KACpDN,EAAKwB,SACPC,QAAQC,KAAK,qBAAsBvB,GAGrC,IAAIwB,EAAkC,CAAA,EACtC,MAAMC,EAAa,IAAKf,QAAQT,KAChC,IAAK,MAAME,KAAWH,EAAU,CAC9BwB,EAAU,IAAKE,SAAO,CAAEtB,KAAMA,EAAKc,KAAKpB,EAAKK,KAAYwB,UAAWH,GACpE,IAAII,EAAQ,EACZ,IAAK,MAAOC,EAAKC,KAAUC,OAAOC,QAAQR,GACpCC,EAAWI,KAASC,IACtBL,EAAWI,GAAOC,EAClBF,KAGAA,EAAQ,GACVN,QAAQC,KAAM,WAAUK,2BAAgCzB,EAE5D,CAEA,GAAIN,EAAKoC,SAAU,CACjB,MACMC,EADcH,OAAOI,KAAKT,SAAO,CAAEtB,KAAMA,EAAKc,KAAKpB,EAAKD,EAAKoC,YAAaN,QAAU,CAAA,GAC1DS,QAAQP,KAAUA,KAAOL,KACzD,GAAIU,EAAYlB,OAAS,EACvB,MAAM,IAAIqB,MAAO,qCAAoCrC,EAASkB,KAAK,YAAYgB,EAAYhB,KAAK,SAEpG,CACA,OAAOM,CACT,+CAKO,SAA8CA,GAE/CA,EAAQc,MAAQd,EAAQe,mBAC1Bf,EAAQc,KAAOd,EAAQc,KAAKE,QAAS,GAAEhB,EAAQe,oBAAqB,IAEjEE,WAAW,mCAAoC,IAE/CA,WAAW,sBAAuB,IAClCA,WAAW,iBAAkB,KAElC,IAAK,MAAMZ,KAAOE,OAAOI,KAAKX,GAAU,CACtC,MAAMkB,EAAWb,EAAIc,eAEnBD,EAASE,WAAW,SACpBF,EAASE,WAAW,UACpBF,EAASE,WAAW,WACP,gBAAbF,GACa,aAAbA,WAEOlB,EAAQK,EAEnB,CACF,oCAxGyC,CACvC5B,IAAK,CACH4C,YAAa,2BACbC,KAAM,SAER,cAAe,CACbD,YACE,wLACFC,KAAM,UAER,mBAAoB,CAClBD,YACE,yHACFC,KAAM,WAER,mBAAoB,CAClBD,YACE,wGACFC,KAAM,UACNC,SAAS,GAEX,YAAa,CACXF,YAAa,qFACbC,KAAM,SACNC,QAAS"}
1
+ {"version":3,"file":"env.cjs","sources":["../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n autoCascadeEnv?: boolean;\n checkEnv?: string;\n verbose?: boolean;\n}\n\nexport const yargsOptionsBuilderForEnv = {\n env: {\n description: '.env files to be loaded.',\n type: 'array',\n },\n 'cascade-env': {\n description:\n 'Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.',\n type: 'string',\n },\n 'cascade-node-env': {\n description: 'Same with --cascade-env=<NODE_ENV || \"development\">. Preferred over `auto-cascade-env`.',\n type: 'boolean',\n },\n 'auto-cascade-env': {\n description: 'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV || \"\">.',\n type: 'boolean',\n default: true,\n },\n 'check-env': {\n description: 'Check whether the keys of the loaded .env files are same with the given .env file.',\n type: 'string',\n default: '.env.example',\n },\n} as const;\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string, orgCwd?: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => path.resolve(orgCwd ?? cwd, envPath.toString()));\n const cascade =\n argv.cascadeEnv ??\n (argv.cascadeNodeEnv\n ? process.env.NODE_ENV || 'development'\n : argv.autoCascadeEnv\n ? process.env.WB_ENV || process.env.APP_ENV || process.env.NODE_ENV || 'development'\n : undefined);\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push(path.join(cwd, '.env'));\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n envPaths = envPaths.map((envPath) => path.relative(cwd, envPath));\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars: Record<string, string> = {};\n const orgEnvVars = { ...process.env };\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n let count = 0;\n for (const [key, value] of Object.entries(envVars)) {\n if (orgEnvVars[key] !== value) {\n orgEnvVars[key] = value;\n count++;\n }\n }\n if (count > 0) {\n console.info(`Updated ${count} environment variables:`, envPath);\n }\n }\n\n if (argv.checkEnv) {\n const exampleKeys = Object.keys(config({ path: path.join(cwd, argv.checkEnv) }).parsed || {});\n const missingKeys = exampleKeys.filter((key) => !(key in envVars));\n if (missingKeys.length > 0) {\n throw new Error(`Missing environment variables in [${envPaths.join(', ')}]: [${missingKeys.join(', ')}]`);\n }\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, string | undefined>): void {\n // Remove npm & yarn environment variables from process.env\n if (envVars.PATH && envVars.BERRY_BIN_FOLDER) {\n envVars.PATH = envVars.PATH.replace(`${envVars.BERRY_BIN_FOLDER}:`, '')\n // Temporary directory in macOS\n .replaceAll(/\\/private\\/var\\/folders\\/[^:]+:/g, '')\n // Temporary directories in Linux\n .replaceAll(/\\/var\\/tmp\\/[^:]+:/g, '')\n .replaceAll(/\\/tmp\\/[^:]+:/g, '');\n }\n for (const key of Object.keys(envVars)) {\n const upperKey = key.toUpperCase();\n if (\n upperKey.startsWith('NPM_') ||\n upperKey.startsWith('YARN_') ||\n upperKey.startsWith('BERRY_') ||\n upperKey === 'PROJECT_CWD' ||\n upperKey === 'INIT_CWD'\n ) {\n delete envVars[key];\n }\n }\n}\n"],"names":["argv","cwd","orgCwd","envPaths","env","map","envPath","path","resolve","toString","cascade","cascadeEnv","cascadeNodeEnv","process","NODE_ENV","autoCascadeEnv","WB_ENV","APP_ENV","undefined","length","push","join","flatMap","relative","verbose","console","info","envVars","orgEnvVars","config","parsed","count","key","value","Object","entries","checkEnv","missingKeys","keys","filter","Error","PATH","BERRY_BIN_FOLDER","replace","replaceAll","upperKey","toUpperCase","startsWith","description","type","default"],"mappings":"6FA0CO,SAAkCA,EAAeC,EAAaC,GACnE,IAAIC,GAAYH,EAAKI,KAAO,IAAIC,KAAKC,GAAYC,EAAKC,QAAQN,GAAUD,EAAKK,EAAQG,cACrF,MAAMC,EACJV,EAAKW,aACJX,EAAKY,eACFC,QAAQT,IAAIU,UAAY,cACxBd,EAAKe,eACLF,QAAQT,IAAIY,QAAUH,QAAQT,IAAIa,SAAWJ,QAAQT,IAAIU,UAAY,mBACrEI,GACiB,iBAAZR,IACe,IAApBP,EAASgB,QAAchB,EAASiB,KAAKb,EAAKc,KAAKpB,EAAK,SACxDE,EAAWA,EAASmB,SAAShB,GAC3BI,EACI,CAAE,GAAEJ,KAAWI,UAAkB,GAAEJ,UAAkB,GAAEA,KAAWI,IAAWJ,GAC7E,CAAE,GAAEA,UAAiBA,MAG7BH,EAAWA,EAASE,KAAKC,GAAYC,EAAKgB,SAAStB,EAAKK,KACpDN,EAAKwB,SACPC,QAAQC,KAAK,qBAAsBvB,GAGrC,IAAIwB,EAAkC,CAAA,EACtC,MAAMC,EAAa,IAAKf,QAAQT,KAChC,IAAK,MAAME,KAAWH,EAAU,CAC9BwB,EAAU,IAAKE,SAAO,CAAEtB,KAAMA,EAAKc,KAAKpB,EAAKK,KAAYwB,UAAWH,GACpE,IAAII,EAAQ,EACZ,IAAK,MAAOC,EAAKC,KAAUC,OAAOC,QAAQR,GACpCC,EAAWI,KAASC,IACtBL,EAAWI,GAAOC,EAClBF,KAGAA,EAAQ,GACVN,QAAQC,KAAM,WAAUK,2BAAgCzB,EAE5D,CAEA,GAAIN,EAAKoC,SAAU,CACjB,MACMC,EADcH,OAAOI,KAAKT,SAAO,CAAEtB,KAAMA,EAAKc,KAAKpB,EAAKD,EAAKoC,YAAaN,QAAU,CAAA,GAC1DS,QAAQP,KAAUA,KAAOL,KACzD,GAAIU,EAAYlB,OAAS,EACvB,MAAM,IAAIqB,MAAO,qCAAoCrC,EAASkB,KAAK,YAAYgB,EAAYhB,KAAK,SAEpG,CACA,OAAOM,CACT,+CAKO,SAA8CA,GAE/CA,EAAQc,MAAQd,EAAQe,mBAC1Bf,EAAQc,KAAOd,EAAQc,KAAKE,QAAS,GAAEhB,EAAQe,oBAAqB,IAEjEE,WAAW,mCAAoC,IAE/CA,WAAW,sBAAuB,IAClCA,WAAW,iBAAkB,KAElC,IAAK,MAAMZ,KAAOE,OAAOI,KAAKX,GAAU,CACtC,MAAMkB,EAAWb,EAAIc,eAEnBD,EAASE,WAAW,SACpBF,EAASE,WAAW,UACpBF,EAASE,WAAW,WACP,gBAAbF,GACa,aAAbA,WAEOlB,EAAQK,EAEnB,CACF,oCAtGyC,CACvC5B,IAAK,CACH4C,YAAa,2BACbC,KAAM,SAER,cAAe,CACbD,YACE,wLACFC,KAAM,UAER,mBAAoB,CAClBD,YAAa,0FACbC,KAAM,WAER,mBAAoB,CAClBD,YAAa,iEACbC,KAAM,UACNC,SAAS,GAEX,YAAa,CACXF,YAAa,qFACbC,KAAM,SACNC,QAAS"}
package/dist/env.d.ts CHANGED
@@ -16,11 +16,11 @@ export declare const yargsOptionsBuilderForEnv: {
16
16
  readonly type: "string";
17
17
  };
18
18
  readonly 'cascade-node-env': {
19
- readonly description: "Same with --cascade-env=<NODE_ENV>. If NODE_ENV is falsy, \"development\" is applied. Preferred over `auto-cascade-env`.";
19
+ readonly description: "Same with --cascade-env=<NODE_ENV || \"development\">. Preferred over `auto-cascade-env`.";
20
20
  readonly type: "boolean";
21
21
  };
22
22
  readonly 'auto-cascade-env': {
23
- readonly description: "Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV>. If they are falsy, \"development\" is applied.";
23
+ readonly description: "Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV || \"\">.";
24
24
  readonly type: "boolean";
25
25
  readonly default: true;
26
26
  };
package/dist/env.js CHANGED
@@ -1,2 +1,2 @@
1
- import e from"node:path";import{config as n}from"dotenv";const o={env:{description:".env files to be loaded.",type:"array"},"cascade-env":{description:"Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.",type:"string"},"cascade-node-env":{description:'Same with --cascade-env=<NODE_ENV>. If NODE_ENV is falsy, "development" is applied. Preferred over `auto-cascade-env`.',type:"boolean"},"auto-cascade-env":{description:'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV>. If they are falsy, "development" is applied.',type:"boolean",default:!0},"check-env":{description:"Check whether the keys of the loaded .env files are same with the given .env file.",type:"string",default:".env.example"}};function t(o,t,a){let s=(o.env??[]).map((n=>e.resolve(a??t,n.toString())));const r=o.cascadeEnv??(o.cascadeNodeEnv?process.env.NODE_ENV||"development":o.autoCascadeEnv?process.env.WB_ENV||process.env.APP_ENV||process.env.NODE_ENV||"development":void 0);"string"==typeof r&&(0===s.length&&s.push(e.join(t,".env")),s=s.flatMap((e=>r?[`${e}.${r}.local`,`${e}.local`,`${e}.${r}`,e]:[`${e}.local`,e]))),s=s.map((n=>e.relative(t,n))),o.verbose&&console.info("Loading env files:",s);let c={};const i={...process.env};for(const o of s){c={...n({path:e.join(t,o)}).parsed,...c};let a=0;for(const[e,n]of Object.entries(c))i[e]!==n&&(i[e]=n,a++);a>0&&console.info(`Updated ${a} environment variables:`,o)}if(o.checkEnv){const a=Object.keys(n({path:e.join(t,o.checkEnv)}).parsed||{}).filter((e=>!(e in c)));if(a.length>0)throw new Error(`Missing environment variables in [${s.join(", ")}]: [${a.join(", ")}]`)}return c}function a(e){e.PATH&&e.BERRY_BIN_FOLDER&&(e.PATH=e.PATH.replace(`${e.BERRY_BIN_FOLDER}:`,"").replaceAll(/\/private\/var\/folders\/[^:]+:/g,"").replaceAll(/\/var\/tmp\/[^:]+:/g,"").replaceAll(/\/tmp\/[^:]+:/g,""));for(const n of Object.keys(e)){const o=n.toUpperCase();(o.startsWith("NPM_")||o.startsWith("YARN_")||o.startsWith("BERRY_")||"PROJECT_CWD"===o||"INIT_CWD"===o)&&delete e[n]}}export{t as loadEnvironmentVariables,a as removeNpmAndYarnEnvironmentVariables,o as yargsOptionsBuilderForEnv};
1
+ import e from"node:path";import{config as n}from"dotenv";const o={env:{description:".env files to be loaded.",type:"array"},"cascade-env":{description:"Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.",type:"string"},"cascade-node-env":{description:'Same with --cascade-env=<NODE_ENV || "development">. Preferred over `auto-cascade-env`.',type:"boolean"},"auto-cascade-env":{description:'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV || "">.',type:"boolean",default:!0},"check-env":{description:"Check whether the keys of the loaded .env files are same with the given .env file.",type:"string",default:".env.example"}};function t(o,t,a){let r=(o.env??[]).map((n=>e.resolve(a??t,n.toString())));const s=o.cascadeEnv??(o.cascadeNodeEnv?process.env.NODE_ENV||"development":o.autoCascadeEnv?process.env.WB_ENV||process.env.APP_ENV||process.env.NODE_ENV||"development":void 0);"string"==typeof s&&(0===r.length&&r.push(e.join(t,".env")),r=r.flatMap((e=>s?[`${e}.${s}.local`,`${e}.local`,`${e}.${s}`,e]:[`${e}.local`,e]))),r=r.map((n=>e.relative(t,n))),o.verbose&&console.info("Loading env files:",r);let c={};const i={...process.env};for(const o of r){c={...n({path:e.join(t,o)}).parsed,...c};let a=0;for(const[e,n]of Object.entries(c))i[e]!==n&&(i[e]=n,a++);a>0&&console.info(`Updated ${a} environment variables:`,o)}if(o.checkEnv){const a=Object.keys(n({path:e.join(t,o.checkEnv)}).parsed||{}).filter((e=>!(e in c)));if(a.length>0)throw new Error(`Missing environment variables in [${r.join(", ")}]: [${a.join(", ")}]`)}return c}function a(e){e.PATH&&e.BERRY_BIN_FOLDER&&(e.PATH=e.PATH.replace(`${e.BERRY_BIN_FOLDER}:`,"").replaceAll(/\/private\/var\/folders\/[^:]+:/g,"").replaceAll(/\/var\/tmp\/[^:]+:/g,"").replaceAll(/\/tmp\/[^:]+:/g,""));for(const n of Object.keys(e)){const o=n.toUpperCase();(o.startsWith("NPM_")||o.startsWith("YARN_")||o.startsWith("BERRY_")||"PROJECT_CWD"===o||"INIT_CWD"===o)&&delete e[n]}}export{t as loadEnvironmentVariables,a as removeNpmAndYarnEnvironmentVariables,o as yargsOptionsBuilderForEnv};
2
2
  //# sourceMappingURL=env.js.map
package/dist/env.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"env.js","sources":["../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n autoCascadeEnv?: boolean;\n checkEnv?: string;\n verbose?: boolean;\n}\n\nexport const yargsOptionsBuilderForEnv = {\n env: {\n description: '.env files to be loaded.',\n type: 'array',\n },\n 'cascade-env': {\n description:\n 'Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.',\n type: 'string',\n },\n 'cascade-node-env': {\n description:\n 'Same with --cascade-env=<NODE_ENV>. If NODE_ENV is falsy, \"development\" is applied. Preferred over `auto-cascade-env`.',\n type: 'boolean',\n },\n 'auto-cascade-env': {\n description:\n 'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV>. If they are falsy, \"development\" is applied.',\n type: 'boolean',\n default: true,\n },\n 'check-env': {\n description: 'Check whether the keys of the loaded .env files are same with the given .env file.',\n type: 'string',\n default: '.env.example',\n },\n} as const;\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string, orgCwd?: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => path.resolve(orgCwd ?? cwd, envPath.toString()));\n const cascade =\n argv.cascadeEnv ??\n (argv.cascadeNodeEnv\n ? process.env.NODE_ENV || 'development'\n : argv.autoCascadeEnv\n ? process.env.WB_ENV || process.env.APP_ENV || process.env.NODE_ENV || 'development'\n : undefined);\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push(path.join(cwd, '.env'));\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n envPaths = envPaths.map((envPath) => path.relative(cwd, envPath));\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars: Record<string, string> = {};\n const orgEnvVars = { ...process.env };\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n let count = 0;\n for (const [key, value] of Object.entries(envVars)) {\n if (orgEnvVars[key] !== value) {\n orgEnvVars[key] = value;\n count++;\n }\n }\n if (count > 0) {\n console.info(`Updated ${count} environment variables:`, envPath);\n }\n }\n\n if (argv.checkEnv) {\n const exampleKeys = Object.keys(config({ path: path.join(cwd, argv.checkEnv) }).parsed || {});\n const missingKeys = exampleKeys.filter((key) => !(key in envVars));\n if (missingKeys.length > 0) {\n throw new Error(`Missing environment variables in [${envPaths.join(', ')}]: [${missingKeys.join(', ')}]`);\n }\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, string | undefined>): void {\n // Remove npm & yarn environment variables from process.env\n if (envVars.PATH && envVars.BERRY_BIN_FOLDER) {\n envVars.PATH = envVars.PATH.replace(`${envVars.BERRY_BIN_FOLDER}:`, '')\n // Temporary directory in macOS\n .replaceAll(/\\/private\\/var\\/folders\\/[^:]+:/g, '')\n // Temporary directories in Linux\n .replaceAll(/\\/var\\/tmp\\/[^:]+:/g, '')\n .replaceAll(/\\/tmp\\/[^:]+:/g, '');\n }\n for (const key of Object.keys(envVars)) {\n const upperKey = key.toUpperCase();\n if (\n upperKey.startsWith('NPM_') ||\n upperKey.startsWith('YARN_') ||\n upperKey.startsWith('BERRY_') ||\n upperKey === 'PROJECT_CWD' ||\n upperKey === 'INIT_CWD'\n ) {\n delete envVars[key];\n }\n }\n}\n"],"names":["yargsOptionsBuilderForEnv","env","description","type","default","loadEnvironmentVariables","argv","cwd","orgCwd","envPaths","map","envPath","path","resolve","toString","cascade","cascadeEnv","cascadeNodeEnv","process","NODE_ENV","autoCascadeEnv","WB_ENV","APP_ENV","undefined","length","push","join","flatMap","relative","verbose","console","info","envVars","orgEnvVars","config","parsed","count","key","value","Object","entries","checkEnv","missingKeys","keys","filter","Error","removeNpmAndYarnEnvironmentVariables","PATH","BERRY_BIN_FOLDER","replace","replaceAll","upperKey","toUpperCase","startsWith"],"mappings":"yDAaO,MAAMA,EAA4B,CACvCC,IAAK,CACHC,YAAa,2BACbC,KAAM,SAER,cAAe,CACbD,YACE,wLACFC,KAAM,UAER,mBAAoB,CAClBD,YACE,yHACFC,KAAM,WAER,mBAAoB,CAClBD,YACE,wGACFC,KAAM,UACNC,SAAS,GAEX,YAAa,CACXF,YAAa,qFACbC,KAAM,SACNC,QAAS,iBAON,SAASC,EAAyBC,EAAeC,EAAaC,GACnE,IAAIC,GAAYH,EAAKL,KAAO,IAAIS,KAAKC,GAAYC,EAAKC,QAAQL,GAAUD,EAAKI,EAAQG,cACrF,MAAMC,EACJT,EAAKU,aACJV,EAAKW,eACFC,QAAQjB,IAAIkB,UAAY,cACxBb,EAAKc,eACLF,QAAQjB,IAAIoB,QAAUH,QAAQjB,IAAIqB,SAAWJ,QAAQjB,IAAIkB,UAAY,mBACrEI,GACiB,iBAAZR,IACe,IAApBN,EAASe,QAAcf,EAASgB,KAAKb,EAAKc,KAAKnB,EAAK,SACxDE,EAAWA,EAASkB,SAAShB,GAC3BI,EACI,CAAE,GAAEJ,KAAWI,UAAkB,GAAEJ,UAAkB,GAAEA,KAAWI,IAAWJ,GAC7E,CAAE,GAAEA,UAAiBA,MAG7BF,EAAWA,EAASC,KAAKC,GAAYC,EAAKgB,SAASrB,EAAKI,KACpDL,EAAKuB,SACPC,QAAQC,KAAK,qBAAsBtB,GAGrC,IAAIuB,EAAkC,CAAA,EACtC,MAAMC,EAAa,IAAKf,QAAQjB,KAChC,IAAK,MAAMU,KAAWF,EAAU,CAC9BuB,EAAU,IAAKE,EAAO,CAAEtB,KAAMA,EAAKc,KAAKnB,EAAKI,KAAYwB,UAAWH,GACpE,IAAII,EAAQ,EACZ,IAAK,MAAOC,EAAKC,KAAUC,OAAOC,QAAQR,GACpCC,EAAWI,KAASC,IACtBL,EAAWI,GAAOC,EAClBF,KAGAA,EAAQ,GACVN,QAAQC,KAAM,WAAUK,2BAAgCzB,EAE5D,CAEA,GAAIL,EAAKmC,SAAU,CACjB,MACMC,EADcH,OAAOI,KAAKT,EAAO,CAAEtB,KAAMA,EAAKc,KAAKnB,EAAKD,EAAKmC,YAAaN,QAAU,CAAA,GAC1DS,QAAQP,KAAUA,KAAOL,KACzD,GAAIU,EAAYlB,OAAS,EACvB,MAAM,IAAIqB,MAAO,qCAAoCpC,EAASiB,KAAK,YAAYgB,EAAYhB,KAAK,SAEpG,CACA,OAAOM,CACT,CAKO,SAASc,EAAqCd,GAE/CA,EAAQe,MAAQf,EAAQgB,mBAC1BhB,EAAQe,KAAOf,EAAQe,KAAKE,QAAS,GAAEjB,EAAQgB,oBAAqB,IAEjEE,WAAW,mCAAoC,IAE/CA,WAAW,sBAAuB,IAClCA,WAAW,iBAAkB,KAElC,IAAK,MAAMb,KAAOE,OAAOI,KAAKX,GAAU,CACtC,MAAMmB,EAAWd,EAAIe,eAEnBD,EAASE,WAAW,SACpBF,EAASE,WAAW,UACpBF,EAASE,WAAW,WACP,gBAAbF,GACa,aAAbA,WAEOnB,EAAQK,EAEnB,CACF"}
1
+ {"version":3,"file":"env.js","sources":["../src/env.ts"],"sourcesContent":["import path from 'node:path';\n\nimport { config } from 'dotenv';\n\ninterface Options {\n env?: (string | number)[];\n cascadeEnv?: string;\n cascadeNodeEnv?: boolean;\n autoCascadeEnv?: boolean;\n checkEnv?: string;\n verbose?: boolean;\n}\n\nexport const yargsOptionsBuilderForEnv = {\n env: {\n description: '.env files to be loaded.',\n type: 'array',\n },\n 'cascade-env': {\n description:\n 'Environment to load cascading .env files (e.g., `.env`, `.env.<environment>`, `.env.local` and `.env.<environment>.local`). Preferred over `cascade-node-env` and `auto-cascade-env`.',\n type: 'string',\n },\n 'cascade-node-env': {\n description: 'Same with --cascade-env=<NODE_ENV || \"development\">. Preferred over `auto-cascade-env`.',\n type: 'boolean',\n },\n 'auto-cascade-env': {\n description: 'Same with --cascade-env=<WB_ENV || APP_ENV || NODE_ENV || \"\">.',\n type: 'boolean',\n default: true,\n },\n 'check-env': {\n description: 'Check whether the keys of the loaded .env files are same with the given .env file.',\n type: 'string',\n default: '.env.example',\n },\n} as const;\n\n/**\n * This function loads environment variables from `.env` files.\n * */\nexport function loadEnvironmentVariables(argv: Options, cwd: string, orgCwd?: string): Record<string, string> {\n let envPaths = (argv.env ?? []).map((envPath) => path.resolve(orgCwd ?? cwd, envPath.toString()));\n const cascade =\n argv.cascadeEnv ??\n (argv.cascadeNodeEnv\n ? process.env.NODE_ENV || 'development'\n : argv.autoCascadeEnv\n ? process.env.WB_ENV || process.env.APP_ENV || process.env.NODE_ENV || 'development'\n : undefined);\n if (typeof cascade === 'string') {\n if (envPaths.length === 0) envPaths.push(path.join(cwd, '.env'));\n envPaths = envPaths.flatMap((envPath) =>\n cascade\n ? [`${envPath}.${cascade}.local`, `${envPath}.local`, `${envPath}.${cascade}`, envPath]\n : [`${envPath}.local`, envPath]\n );\n }\n envPaths = envPaths.map((envPath) => path.relative(cwd, envPath));\n if (argv.verbose) {\n console.info('Loading env files:', envPaths);\n }\n\n let envVars: Record<string, string> = {};\n const orgEnvVars = { ...process.env };\n for (const envPath of envPaths) {\n envVars = { ...config({ path: path.join(cwd, envPath) }).parsed, ...envVars };\n let count = 0;\n for (const [key, value] of Object.entries(envVars)) {\n if (orgEnvVars[key] !== value) {\n orgEnvVars[key] = value;\n count++;\n }\n }\n if (count > 0) {\n console.info(`Updated ${count} environment variables:`, envPath);\n }\n }\n\n if (argv.checkEnv) {\n const exampleKeys = Object.keys(config({ path: path.join(cwd, argv.checkEnv) }).parsed || {});\n const missingKeys = exampleKeys.filter((key) => !(key in envVars));\n if (missingKeys.length > 0) {\n throw new Error(`Missing environment variables in [${envPaths.join(', ')}]: [${missingKeys.join(', ')}]`);\n }\n }\n return envVars;\n}\n\n/**\n * This function removes environment variables related to npm and yarn from the given environment variables.\n * */\nexport function removeNpmAndYarnEnvironmentVariables(envVars: Record<string, string | undefined>): void {\n // Remove npm & yarn environment variables from process.env\n if (envVars.PATH && envVars.BERRY_BIN_FOLDER) {\n envVars.PATH = envVars.PATH.replace(`${envVars.BERRY_BIN_FOLDER}:`, '')\n // Temporary directory in macOS\n .replaceAll(/\\/private\\/var\\/folders\\/[^:]+:/g, '')\n // Temporary directories in Linux\n .replaceAll(/\\/var\\/tmp\\/[^:]+:/g, '')\n .replaceAll(/\\/tmp\\/[^:]+:/g, '');\n }\n for (const key of Object.keys(envVars)) {\n const upperKey = key.toUpperCase();\n if (\n upperKey.startsWith('NPM_') ||\n upperKey.startsWith('YARN_') ||\n upperKey.startsWith('BERRY_') ||\n upperKey === 'PROJECT_CWD' ||\n upperKey === 'INIT_CWD'\n ) {\n delete envVars[key];\n }\n }\n}\n"],"names":["yargsOptionsBuilderForEnv","env","description","type","default","loadEnvironmentVariables","argv","cwd","orgCwd","envPaths","map","envPath","path","resolve","toString","cascade","cascadeEnv","cascadeNodeEnv","process","NODE_ENV","autoCascadeEnv","WB_ENV","APP_ENV","undefined","length","push","join","flatMap","relative","verbose","console","info","envVars","orgEnvVars","config","parsed","count","key","value","Object","entries","checkEnv","missingKeys","keys","filter","Error","removeNpmAndYarnEnvironmentVariables","PATH","BERRY_BIN_FOLDER","replace","replaceAll","upperKey","toUpperCase","startsWith"],"mappings":"yDAaO,MAAMA,EAA4B,CACvCC,IAAK,CACHC,YAAa,2BACbC,KAAM,SAER,cAAe,CACbD,YACE,wLACFC,KAAM,UAER,mBAAoB,CAClBD,YAAa,0FACbC,KAAM,WAER,mBAAoB,CAClBD,YAAa,iEACbC,KAAM,UACNC,SAAS,GAEX,YAAa,CACXF,YAAa,qFACbC,KAAM,SACNC,QAAS,iBAON,SAASC,EAAyBC,EAAeC,EAAaC,GACnE,IAAIC,GAAYH,EAAKL,KAAO,IAAIS,KAAKC,GAAYC,EAAKC,QAAQL,GAAUD,EAAKI,EAAQG,cACrF,MAAMC,EACJT,EAAKU,aACJV,EAAKW,eACFC,QAAQjB,IAAIkB,UAAY,cACxBb,EAAKc,eACLF,QAAQjB,IAAIoB,QAAUH,QAAQjB,IAAIqB,SAAWJ,QAAQjB,IAAIkB,UAAY,mBACrEI,GACiB,iBAAZR,IACe,IAApBN,EAASe,QAAcf,EAASgB,KAAKb,EAAKc,KAAKnB,EAAK,SACxDE,EAAWA,EAASkB,SAAShB,GAC3BI,EACI,CAAE,GAAEJ,KAAWI,UAAkB,GAAEJ,UAAkB,GAAEA,KAAWI,IAAWJ,GAC7E,CAAE,GAAEA,UAAiBA,MAG7BF,EAAWA,EAASC,KAAKC,GAAYC,EAAKgB,SAASrB,EAAKI,KACpDL,EAAKuB,SACPC,QAAQC,KAAK,qBAAsBtB,GAGrC,IAAIuB,EAAkC,CAAA,EACtC,MAAMC,EAAa,IAAKf,QAAQjB,KAChC,IAAK,MAAMU,KAAWF,EAAU,CAC9BuB,EAAU,IAAKE,EAAO,CAAEtB,KAAMA,EAAKc,KAAKnB,EAAKI,KAAYwB,UAAWH,GACpE,IAAII,EAAQ,EACZ,IAAK,MAAOC,EAAKC,KAAUC,OAAOC,QAAQR,GACpCC,EAAWI,KAASC,IACtBL,EAAWI,GAAOC,EAClBF,KAGAA,EAAQ,GACVN,QAAQC,KAAM,WAAUK,2BAAgCzB,EAE5D,CAEA,GAAIL,EAAKmC,SAAU,CACjB,MACMC,EADcH,OAAOI,KAAKT,EAAO,CAAEtB,KAAMA,EAAKc,KAAKnB,EAAKD,EAAKmC,YAAaN,QAAU,CAAA,GAC1DS,QAAQP,KAAUA,KAAOL,KACzD,GAAIU,EAAYlB,OAAS,EACvB,MAAM,IAAIqB,MAAO,qCAAoCpC,EAASiB,KAAK,YAAYgB,EAAYhB,KAAK,SAEpG,CACA,OAAOM,CACT,CAKO,SAASc,EAAqCd,GAE/CA,EAAQe,MAAQf,EAAQgB,mBAC1BhB,EAAQe,KAAOf,EAAQe,KAAKE,QAAS,GAAEjB,EAAQgB,oBAAqB,IAEjEE,WAAW,mCAAoC,IAE/CA,WAAW,sBAAuB,IAClCA,WAAW,iBAAkB,KAElC,IAAK,MAAMb,KAAOE,OAAOI,KAAKX,GAAU,CACtC,MAAMmB,EAAWd,EAAIe,eAEnBD,EAASE,WAAW,SACpBF,EAASE,WAAW,UACpBF,EAASE,WAAW,WACP,gBAAbF,GACa,aAAbA,WAEOnB,EAAQK,EAEnB,CACF"}
package/dist/hash.cjs CHANGED
@@ -1,2 +1,2 @@
1
- "use strict";var e=require("node:crypto"),t=require("node:fs"),s=require("node:path");async function r(e,...s){let r="";try{r=await t.promises.readFile(e,"utf8")}catch{}const i=await a(...s);return r!==i&&(await t.promises.writeFile(e,i,"utf8"),!0)}async function a(...r){const a=e.createHash("sha512");for(const e of r.sort()){const r=await t.promises.stat(e);if(r.isDirectory()){const r=await t.promises.readdir(e,{withFileTypes:!0,recursive:!0});for(const e of r.sort(((e,t)=>e.name.localeCompare(t.name))))e.isFile()&&a.update(await t.promises.readFile(s.join(e.path,e.name),"utf8"))}else r.isFile()&&a.update(await t.promises.readFile(e,"utf8"))}return a.digest("hex")}exports.calculateHashFromFiles=a,exports.canSkipSeed=async function(e,...t){return("1"===process.env.ALLOW_TO_SKIP_SEED||"true"===process.env.ALLOW_TO_SKIP_SEED)&&!await r(e,...t)},exports.updateHashFromFiles=r;
1
+ "use strict";var e=require("node:crypto"),r=require("node:fs"),s=require("node:path");async function t(e,...t){let i="";try{i=await r.promises.readFile(e,"utf8")}catch{}const o=await a(...t);return i!==o&&(await r.promises.mkdir(s.dirname(e),{recursive:!0}),await r.promises.writeFile(e,o,"utf8"),!0)}async function a(...t){const a=e.createHash("sha512");for(const e of t.sort()){const t=await r.promises.stat(e);if(t.isDirectory()){const t=await r.promises.readdir(e,{withFileTypes:!0,recursive:!0});for(const e of t.sort(((e,r)=>e.name.localeCompare(r.name))))e.isFile()&&a.update(await r.promises.readFile(s.join(e.path,e.name),"utf8"))}else t.isFile()&&a.update(await r.promises.readFile(e,"utf8"))}return a.digest("hex")}exports.calculateHashFromFiles=a,exports.canSkipSeed=async function(e,...r){return("1"===process.env.ALLOW_TO_SKIP_SEED||"true"===process.env.ALLOW_TO_SKIP_SEED)&&!await t(e,...r)},exports.updateHashFromFiles=t;
2
2
  //# sourceMappingURL=hash.cjs.map
package/dist/hash.cjs.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"hash.cjs","sources":["../src/hash.ts"],"sourcesContent":["import crypto from 'node:crypto';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\n/**\n * Check whether seed command can be skipped or not and update hash file if needed.\n * Note that process.env.ALLOW_TO_SKIP_SEED should be set to '1' or 'true' to skip seed.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether seed command can be skipped.\n */\nexport async function canSkipSeed(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n if (process.env.ALLOW_TO_SKIP_SEED !== '1' && process.env.ALLOW_TO_SKIP_SEED !== 'true') return false;\n return !(await updateHashFromFiles(hashFilePath, ...paths));\n}\n\n/**\n * Update hash file if the hash is different from the current one.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether the hash file was updated.\n */\nexport async function updateHashFromFiles(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n let oldHash = '';\n try {\n oldHash = await fs.promises.readFile(hashFilePath, 'utf8');\n } catch {\n // do nothing\n }\n const newHash = await calculateHashFromFiles(...paths);\n if (oldHash === newHash) return false;\n\n await fs.promises.writeFile(hashFilePath, newHash, 'utf8');\n return true;\n}\n\n/**\n * Calculate hash from files.\n * @param paths Paths to the files or directories.\n * @returns Hash string.\n */\nexport async function calculateHashFromFiles(...paths: string[]): Promise<string> {\n const hash = crypto.createHash('sha512');\n for (const fileOrDirPath of paths.sort()) {\n const stat = await fs.promises.stat(fileOrDirPath);\n if (stat.isDirectory()) {\n // Get all files in the directory\n const dirents = await fs.promises.readdir(fileOrDirPath, { withFileTypes: true, recursive: true });\n for (const dirent of dirents.sort((d1, d2) => d1.name.localeCompare(d2.name))) {\n if (dirent.isFile()) {\n // Node.js 18.17.0 or later has `dirent.path`\n hash.update(\n await fs.promises.readFile(\n path.join((dirent as unknown as Record<'path', string>).path, dirent.name),\n 'utf8'\n )\n );\n }\n }\n } else if (stat.isFile()) {\n hash.update(await fs.promises.readFile(fileOrDirPath, 'utf8'));\n }\n }\n return hash.digest('hex');\n}\n"],"names":["async","updateHashFromFiles","hashFilePath","paths","oldHash","fs","promises","readFile","newHash","calculateHashFromFiles","writeFile","hash","crypto","createHash","fileOrDirPath","sort","stat","isDirectory","dirents","readdir","withFileTypes","recursive","dirent","d1","d2","name","localeCompare","isFile","update","path","join","digest","process","env","ALLOW_TO_SKIP_SEED"],"mappings":"sFAsBOA,eAAeC,EAAoBC,KAAyBC,GACjE,IAAIC,EAAU,GACd,IACEA,QAAgBC,EAAGC,SAASC,SAASL,EAAc,OACrD,CAAE,MACA,CAEF,MAAMM,QAAgBC,KAA0BN,GAChD,OAAIC,IAAYI,UAEVH,EAAGC,SAASI,UAAUR,EAAcM,EAAS,SAC5C,EACT,CAOOR,eAAeS,KAA0BN,GAC9C,MAAMQ,EAAOC,EAAOC,WAAW,UAC/B,IAAK,MAAMC,KAAiBX,EAAMY,OAAQ,CACxC,MAAMC,QAAaX,EAAGC,SAASU,KAAKF,GACpC,GAAIE,EAAKC,cAAe,CAEtB,MAAMC,QAAgBb,EAAGC,SAASa,QAAQL,EAAe,CAAEM,eAAe,EAAMC,WAAW,IAC3F,IAAK,MAAMC,KAAUJ,EAAQH,MAAK,CAACQ,EAAIC,IAAOD,EAAGE,KAAKC,cAAcF,EAAGC,QACjEH,EAAOK,UAEThB,EAAKiB,aACGvB,EAAGC,SAASC,SAChBsB,EAAKC,KAAMR,EAA6CO,KAAMP,EAAOG,MACrE,QAKV,MAAWT,EAAKW,UACdhB,EAAKiB,aAAavB,EAAGC,SAASC,SAASO,EAAe,QAE1D,CACA,OAAOH,EAAKoB,OAAO,MACrB,sDArDO/B,eAA2BE,KAAyBC,GACzD,OAAuC,MAAnC6B,QAAQC,IAAIC,oBAAiE,SAAnCF,QAAQC,IAAIC,4BAC3CjC,EAAoBC,KAAiBC,EACtD"}
1
+ {"version":3,"file":"hash.cjs","sources":["../src/hash.ts"],"sourcesContent":["import crypto from 'node:crypto';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\n/**\n * Check whether seed command can be skipped or not and update hash file if needed.\n * Note that process.env.ALLOW_TO_SKIP_SEED should be set to '1' or 'true' to skip seed.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether seed command can be skipped.\n */\nexport async function canSkipSeed(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n if (process.env.ALLOW_TO_SKIP_SEED !== '1' && process.env.ALLOW_TO_SKIP_SEED !== 'true') return false;\n return !(await updateHashFromFiles(hashFilePath, ...paths));\n}\n\n/**\n * Update hash file if the hash is different from the current one.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether the hash file was updated.\n */\nexport async function updateHashFromFiles(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n let oldHash = '';\n try {\n oldHash = await fs.promises.readFile(hashFilePath, 'utf8');\n } catch {\n // do nothing\n }\n const newHash = await calculateHashFromFiles(...paths);\n if (oldHash === newHash) return false;\n\n await fs.promises.mkdir(path.dirname(hashFilePath), { recursive: true });\n await fs.promises.writeFile(hashFilePath, newHash, 'utf8');\n return true;\n}\n\n/**\n * Calculate hash from files.\n * @param paths Paths to the files or directories.\n * @returns Hash string.\n */\nexport async function calculateHashFromFiles(...paths: string[]): Promise<string> {\n const hash = crypto.createHash('sha512');\n for (const fileOrDirPath of paths.sort()) {\n const stat = await fs.promises.stat(fileOrDirPath);\n if (stat.isDirectory()) {\n // Get all files in the directory\n const dirents = await fs.promises.readdir(fileOrDirPath, { withFileTypes: true, recursive: true });\n for (const dirent of dirents.sort((d1, d2) => d1.name.localeCompare(d2.name))) {\n if (dirent.isFile()) {\n // Node.js 18.17.0 or later has `dirent.path`\n hash.update(\n await fs.promises.readFile(\n path.join((dirent as unknown as Record<'path', string>).path, dirent.name),\n 'utf8'\n )\n );\n }\n }\n } else if (stat.isFile()) {\n hash.update(await fs.promises.readFile(fileOrDirPath, 'utf8'));\n }\n }\n return hash.digest('hex');\n}\n"],"names":["async","updateHashFromFiles","hashFilePath","paths","oldHash","fs","promises","readFile","newHash","calculateHashFromFiles","mkdir","path","dirname","recursive","writeFile","hash","crypto","createHash","fileOrDirPath","sort","stat","isDirectory","dirents","readdir","withFileTypes","dirent","d1","d2","name","localeCompare","isFile","update","join","digest","process","env","ALLOW_TO_SKIP_SEED"],"mappings":"sFAsBOA,eAAeC,EAAoBC,KAAyBC,GACjE,IAAIC,EAAU,GACd,IACEA,QAAgBC,EAAGC,SAASC,SAASL,EAAc,OACrD,CAAE,MACA,CAEF,MAAMM,QAAgBC,KAA0BN,GAChD,OAAIC,IAAYI,UAEVH,EAAGC,SAASI,MAAMC,EAAKC,QAAQV,GAAe,CAAEW,WAAW,UAC3DR,EAAGC,SAASQ,UAAUZ,EAAcM,EAAS,SAC5C,EACT,CAOOR,eAAeS,KAA0BN,GAC9C,MAAMY,EAAOC,EAAOC,WAAW,UAC/B,IAAK,MAAMC,KAAiBf,EAAMgB,OAAQ,CACxC,MAAMC,QAAaf,EAAGC,SAASc,KAAKF,GACpC,GAAIE,EAAKC,cAAe,CAEtB,MAAMC,QAAgBjB,EAAGC,SAASiB,QAAQL,EAAe,CAAEM,eAAe,EAAMX,WAAW,IAC3F,IAAK,MAAMY,KAAUH,EAAQH,MAAK,CAACO,EAAIC,IAAOD,EAAGE,KAAKC,cAAcF,EAAGC,QACjEH,EAAOK,UAETf,EAAKgB,aACG1B,EAAGC,SAASC,SAChBI,EAAKqB,KAAMP,EAA6Cd,KAAMc,EAAOG,MACrE,QAKV,MAAWR,EAAKU,UACdf,EAAKgB,aAAa1B,EAAGC,SAASC,SAASW,EAAe,QAE1D,CACA,OAAOH,EAAKkB,OAAO,MACrB,sDAtDOjC,eAA2BE,KAAyBC,GACzD,OAAuC,MAAnC+B,QAAQC,IAAIC,oBAAiE,SAAnCF,QAAQC,IAAIC,4BAC3CnC,EAAoBC,KAAiBC,EACtD"}
package/dist/hash.js CHANGED
@@ -1,2 +1,2 @@
1
- import e from"node:crypto";import t from"node:fs";import r from"node:path";async function i(e,...t){return("1"===process.env.ALLOW_TO_SKIP_SEED||"true"===process.env.ALLOW_TO_SKIP_SEED)&&!await o(e,...t)}async function o(e,...r){let i="";try{i=await t.promises.readFile(e,"utf8")}catch{}const o=await s(...r);return i!==o&&(await t.promises.writeFile(e,o,"utf8"),!0)}async function s(...i){const o=e.createHash("sha512");for(const e of i.sort()){const i=await t.promises.stat(e);if(i.isDirectory()){const i=await t.promises.readdir(e,{withFileTypes:!0,recursive:!0});for(const e of i.sort(((e,t)=>e.name.localeCompare(t.name))))e.isFile()&&o.update(await t.promises.readFile(r.join(e.path,e.name),"utf8"))}else i.isFile()&&o.update(await t.promises.readFile(e,"utf8"))}return o.digest("hex")}export{s as calculateHashFromFiles,i as canSkipSeed,o as updateHashFromFiles};
1
+ import e from"node:crypto";import t from"node:fs";import r from"node:path";async function i(e,...t){return("1"===process.env.ALLOW_TO_SKIP_SEED||"true"===process.env.ALLOW_TO_SKIP_SEED)&&!await s(e,...t)}async function s(e,...i){let s="";try{s=await t.promises.readFile(e,"utf8")}catch{}const o=await a(...i);return s!==o&&(await t.promises.mkdir(r.dirname(e),{recursive:!0}),await t.promises.writeFile(e,o,"utf8"),!0)}async function a(...i){const s=e.createHash("sha512");for(const e of i.sort()){const i=await t.promises.stat(e);if(i.isDirectory()){const i=await t.promises.readdir(e,{withFileTypes:!0,recursive:!0});for(const e of i.sort(((e,t)=>e.name.localeCompare(t.name))))e.isFile()&&s.update(await t.promises.readFile(r.join(e.path,e.name),"utf8"))}else i.isFile()&&s.update(await t.promises.readFile(e,"utf8"))}return s.digest("hex")}export{a as calculateHashFromFiles,i as canSkipSeed,s as updateHashFromFiles};
2
2
  //# sourceMappingURL=hash.js.map
package/dist/hash.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"file":"hash.js","sources":["../src/hash.ts"],"sourcesContent":["import crypto from 'node:crypto';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\n/**\n * Check whether seed command can be skipped or not and update hash file if needed.\n * Note that process.env.ALLOW_TO_SKIP_SEED should be set to '1' or 'true' to skip seed.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether seed command can be skipped.\n */\nexport async function canSkipSeed(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n if (process.env.ALLOW_TO_SKIP_SEED !== '1' && process.env.ALLOW_TO_SKIP_SEED !== 'true') return false;\n return !(await updateHashFromFiles(hashFilePath, ...paths));\n}\n\n/**\n * Update hash file if the hash is different from the current one.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether the hash file was updated.\n */\nexport async function updateHashFromFiles(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n let oldHash = '';\n try {\n oldHash = await fs.promises.readFile(hashFilePath, 'utf8');\n } catch {\n // do nothing\n }\n const newHash = await calculateHashFromFiles(...paths);\n if (oldHash === newHash) return false;\n\n await fs.promises.writeFile(hashFilePath, newHash, 'utf8');\n return true;\n}\n\n/**\n * Calculate hash from files.\n * @param paths Paths to the files or directories.\n * @returns Hash string.\n */\nexport async function calculateHashFromFiles(...paths: string[]): Promise<string> {\n const hash = crypto.createHash('sha512');\n for (const fileOrDirPath of paths.sort()) {\n const stat = await fs.promises.stat(fileOrDirPath);\n if (stat.isDirectory()) {\n // Get all files in the directory\n const dirents = await fs.promises.readdir(fileOrDirPath, { withFileTypes: true, recursive: true });\n for (const dirent of dirents.sort((d1, d2) => d1.name.localeCompare(d2.name))) {\n if (dirent.isFile()) {\n // Node.js 18.17.0 or later has `dirent.path`\n hash.update(\n await fs.promises.readFile(\n path.join((dirent as unknown as Record<'path', string>).path, dirent.name),\n 'utf8'\n )\n );\n }\n }\n } else if (stat.isFile()) {\n hash.update(await fs.promises.readFile(fileOrDirPath, 'utf8'));\n }\n }\n return hash.digest('hex');\n}\n"],"names":["async","canSkipSeed","hashFilePath","paths","process","env","ALLOW_TO_SKIP_SEED","updateHashFromFiles","oldHash","fs","promises","readFile","newHash","calculateHashFromFiles","writeFile","hash","crypto","createHash","fileOrDirPath","sort","stat","isDirectory","dirents","readdir","withFileTypes","recursive","dirent","d1","d2","name","localeCompare","isFile","update","path","join","digest"],"mappings":"2EAWOA,eAAeC,EAAYC,KAAyBC,GACzD,OAAuC,MAAnCC,QAAQC,IAAIC,oBAAiE,SAAnCF,QAAQC,IAAIC,4BAC3CC,EAAoBL,KAAiBC,EACtD,CAQOH,eAAeO,EAAoBL,KAAyBC,GACjE,IAAIK,EAAU,GACd,IACEA,QAAgBC,EAAGC,SAASC,SAAST,EAAc,OACrD,CAAE,MACA,CAEF,MAAMU,QAAgBC,KAA0BV,GAChD,OAAIK,IAAYI,UAEVH,EAAGC,SAASI,UAAUZ,EAAcU,EAAS,SAC5C,EACT,CAOOZ,eAAea,KAA0BV,GAC9C,MAAMY,EAAOC,EAAOC,WAAW,UAC/B,IAAK,MAAMC,KAAiBf,EAAMgB,OAAQ,CACxC,MAAMC,QAAaX,EAAGC,SAASU,KAAKF,GACpC,GAAIE,EAAKC,cAAe,CAEtB,MAAMC,QAAgBb,EAAGC,SAASa,QAAQL,EAAe,CAAEM,eAAe,EAAMC,WAAW,IAC3F,IAAK,MAAMC,KAAUJ,EAAQH,MAAK,CAACQ,EAAIC,IAAOD,EAAGE,KAAKC,cAAcF,EAAGC,QACjEH,EAAOK,UAEThB,EAAKiB,aACGvB,EAAGC,SAASC,SAChBsB,EAAKC,KAAMR,EAA6CO,KAAMP,EAAOG,MACrE,QAKV,MAAWT,EAAKW,UACdhB,EAAKiB,aAAavB,EAAGC,SAASC,SAASO,EAAe,QAE1D,CACA,OAAOH,EAAKoB,OAAO,MACrB"}
1
+ {"version":3,"file":"hash.js","sources":["../src/hash.ts"],"sourcesContent":["import crypto from 'node:crypto';\nimport fs from 'node:fs';\nimport path from 'node:path';\n\n/**\n * Check whether seed command can be skipped or not and update hash file if needed.\n * Note that process.env.ALLOW_TO_SKIP_SEED should be set to '1' or 'true' to skip seed.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether seed command can be skipped.\n */\nexport async function canSkipSeed(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n if (process.env.ALLOW_TO_SKIP_SEED !== '1' && process.env.ALLOW_TO_SKIP_SEED !== 'true') return false;\n return !(await updateHashFromFiles(hashFilePath, ...paths));\n}\n\n/**\n * Update hash file if the hash is different from the current one.\n * @param hashFilePath Path to the hash file.\n * @param paths Paths to the files or directories.\n * @returns Whether the hash file was updated.\n */\nexport async function updateHashFromFiles(hashFilePath: string, ...paths: string[]): Promise<boolean> {\n let oldHash = '';\n try {\n oldHash = await fs.promises.readFile(hashFilePath, 'utf8');\n } catch {\n // do nothing\n }\n const newHash = await calculateHashFromFiles(...paths);\n if (oldHash === newHash) return false;\n\n await fs.promises.mkdir(path.dirname(hashFilePath), { recursive: true });\n await fs.promises.writeFile(hashFilePath, newHash, 'utf8');\n return true;\n}\n\n/**\n * Calculate hash from files.\n * @param paths Paths to the files or directories.\n * @returns Hash string.\n */\nexport async function calculateHashFromFiles(...paths: string[]): Promise<string> {\n const hash = crypto.createHash('sha512');\n for (const fileOrDirPath of paths.sort()) {\n const stat = await fs.promises.stat(fileOrDirPath);\n if (stat.isDirectory()) {\n // Get all files in the directory\n const dirents = await fs.promises.readdir(fileOrDirPath, { withFileTypes: true, recursive: true });\n for (const dirent of dirents.sort((d1, d2) => d1.name.localeCompare(d2.name))) {\n if (dirent.isFile()) {\n // Node.js 18.17.0 or later has `dirent.path`\n hash.update(\n await fs.promises.readFile(\n path.join((dirent as unknown as Record<'path', string>).path, dirent.name),\n 'utf8'\n )\n );\n }\n }\n } else if (stat.isFile()) {\n hash.update(await fs.promises.readFile(fileOrDirPath, 'utf8'));\n }\n }\n return hash.digest('hex');\n}\n"],"names":["async","canSkipSeed","hashFilePath","paths","process","env","ALLOW_TO_SKIP_SEED","updateHashFromFiles","oldHash","fs","promises","readFile","newHash","calculateHashFromFiles","mkdir","path","dirname","recursive","writeFile","hash","crypto","createHash","fileOrDirPath","sort","stat","isDirectory","dirents","readdir","withFileTypes","dirent","d1","d2","name","localeCompare","isFile","update","join","digest"],"mappings":"2EAWOA,eAAeC,EAAYC,KAAyBC,GACzD,OAAuC,MAAnCC,QAAQC,IAAIC,oBAAiE,SAAnCF,QAAQC,IAAIC,4BAC3CC,EAAoBL,KAAiBC,EACtD,CAQOH,eAAeO,EAAoBL,KAAyBC,GACjE,IAAIK,EAAU,GACd,IACEA,QAAgBC,EAAGC,SAASC,SAAST,EAAc,OACrD,CAAE,MACA,CAEF,MAAMU,QAAgBC,KAA0BV,GAChD,OAAIK,IAAYI,UAEVH,EAAGC,SAASI,MAAMC,EAAKC,QAAQd,GAAe,CAAEe,WAAW,UAC3DR,EAAGC,SAASQ,UAAUhB,EAAcU,EAAS,SAC5C,EACT,CAOOZ,eAAea,KAA0BV,GAC9C,MAAMgB,EAAOC,EAAOC,WAAW,UAC/B,IAAK,MAAMC,KAAiBnB,EAAMoB,OAAQ,CACxC,MAAMC,QAAaf,EAAGC,SAASc,KAAKF,GACpC,GAAIE,EAAKC,cAAe,CAEtB,MAAMC,QAAgBjB,EAAGC,SAASiB,QAAQL,EAAe,CAAEM,eAAe,EAAMX,WAAW,IAC3F,IAAK,MAAMY,KAAUH,EAAQH,MAAK,CAACO,EAAIC,IAAOD,EAAGE,KAAKC,cAAcF,EAAGC,QACjEH,EAAOK,UAETf,EAAKgB,aACG1B,EAAGC,SAASC,SAChBI,EAAKqB,KAAMP,EAA6Cd,KAAMc,EAAOG,MACrE,QAKV,MAAWR,EAAKU,UACdf,EAAKgB,aAAa1B,EAAGC,SAASC,SAASW,EAAe,QAE1D,CACA,OAAOH,EAAKkB,OAAO,MACrB"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@willbooster/shared-lib-node",
3
- "version": "3.2.0",
3
+ "version": "3.2.2",
4
4
  "license": "Apache-2.0",
5
5
  "author": "WillBooster Inc.",
6
6
  "sideEffects": false,