@commercetools-frontend/application-cli 2.3.2 → 2.5.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -18,7 +18,7 @@ var _mapInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instan
18
18
  var _Object$entries = require('@babel/runtime-corejs3/core-js-stable/object/entries');
19
19
  var _Set = require('@babel/runtime-corejs3/core-js-stable/set');
20
20
  var _flatMapInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/flat-map');
21
- var storageBucketsConfig = require('../../dist/storage-buckets-config-8e7abeff.cjs.dev.js');
21
+ var storageBucketsConfig = require('../../dist/storage-buckets-config-cf50df38.cjs.dev.js');
22
22
  var fs = require('node:fs');
23
23
  var path$1 = require('node:path');
24
24
  var listr2 = require('listr2');
@@ -41,9 +41,17 @@ var _JSON$stringify = require('@babel/runtime-corejs3/core-js-stable/json/string
41
41
  var applicationConfig = require('@commercetools-frontend/application-config');
42
42
  var l10n = require('@commercetools-frontend/l10n');
43
43
  var _sliceInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/slice');
44
+ var _startsWithInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/starts-with');
45
+ var _trimInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/trim');
46
+ var _someInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/some');
47
+ var _everyInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/every');
48
+ var _includesInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/includes');
49
+ var micromatch = require('micromatch');
50
+ var snakeCase = require('lodash/snakeCase');
44
51
  var jsonschema = require('jsonschema');
45
52
  require('cosmiconfig');
46
53
  require('ts-deepmerge');
54
+ require('lodash');
47
55
 
48
56
  function _interopDefault (e) { return e && e.__esModule ? e : { 'default': e }; }
49
57
 
@@ -70,6 +78,13 @@ var fs__default$1 = /*#__PURE__*/_interopDefault(fs$1);
70
78
  var _findInstanceProperty__default = /*#__PURE__*/_interopDefault(_findInstanceProperty);
71
79
  var _JSON$stringify__default = /*#__PURE__*/_interopDefault(_JSON$stringify);
72
80
  var _sliceInstanceProperty__default = /*#__PURE__*/_interopDefault(_sliceInstanceProperty);
81
+ var _startsWithInstanceProperty__default = /*#__PURE__*/_interopDefault(_startsWithInstanceProperty);
82
+ var _trimInstanceProperty__default = /*#__PURE__*/_interopDefault(_trimInstanceProperty);
83
+ var _someInstanceProperty__default = /*#__PURE__*/_interopDefault(_someInstanceProperty);
84
+ var _everyInstanceProperty__default = /*#__PURE__*/_interopDefault(_everyInstanceProperty);
85
+ var _includesInstanceProperty__default = /*#__PURE__*/_interopDefault(_includesInstanceProperty);
86
+ var micromatch__default = /*#__PURE__*/_interopDefault(micromatch);
87
+ var snakeCase__default = /*#__PURE__*/_interopDefault(snakeCase);
73
88
 
74
89
  function getApplicationDirectory(cwd) {
75
90
  return fs__default["default"].realpathSync(cwd);
@@ -85,7 +100,7 @@ function isCI() {
85
100
  }
86
101
 
87
102
  function ownKeys$4(e, r) { var t = _Object$keys__default["default"](e); if (_Object$getOwnPropertySymbols__default["default"]) { var o = _Object$getOwnPropertySymbols__default["default"](e); r && (o = _filterInstanceProperty__default["default"](o).call(o, function (r) { return _Object$getOwnPropertyDescriptor__default["default"](e, r).enumerable; })), t.push.apply(t, o); } return t; }
88
- function _objectSpread$4(e) { for (var r = 1; r < arguments.length; r++) { var _context5, _context6; var t = null != arguments[r] ? arguments[r] : {}; r % 2 ? _forEachInstanceProperty__default["default"](_context5 = ownKeys$4(Object(t), !0)).call(_context5, function (r) { _defineProperty(e, r, t[r]); }) : _Object$getOwnPropertyDescriptors__default["default"] ? _Object$defineProperties__default["default"](e, _Object$getOwnPropertyDescriptors__default["default"](t)) : _forEachInstanceProperty__default["default"](_context6 = ownKeys$4(Object(t))).call(_context6, function (r) { _Object$defineProperty__default["default"](e, r, _Object$getOwnPropertyDescriptor__default["default"](t, r)); }); } return e; }
103
+ function _objectSpread$4(e) { for (var r = 1; r < arguments.length; r++) { var _context6, _context7; var t = null != arguments[r] ? arguments[r] : {}; r % 2 ? _forEachInstanceProperty__default["default"](_context6 = ownKeys$4(Object(t), !0)).call(_context6, function (r) { _defineProperty(e, r, t[r]); }) : _Object$getOwnPropertyDescriptors__default["default"] ? _Object$defineProperties__default["default"](e, _Object$getOwnPropertyDescriptors__default["default"](t)) : _forEachInstanceProperty__default["default"](_context7 = ownKeys$4(Object(t))).call(_context7, function (r) { _Object$defineProperty__default["default"](e, r, _Object$getOwnPropertyDescriptor__default["default"](t, r)); }); } return e; }
89
104
  function _callSuper(_this, derived, args) {
90
105
  function isNativeReflectConstruct() {
91
106
  if (typeof Reflect === "undefined" || !_Reflect$construct__default["default"]) return false;
@@ -190,6 +205,10 @@ let StorageProvider = /*#__PURE__*/function () {
190
205
  }
191
206
  const storageProvider = storageBucketsConfig.storageProviders[tag];
192
207
  const publicBaseUrl = (_storageProvider$urls3 = storageProvider.urls.public[_classPrivateFieldLooseBase(this, _bucketEnvironment$1)[_bucketEnvironment$1]]) !== null && _storageProvider$urls3 !== void 0 ? _storageProvider$urls3 : storageProvider.urls.public.default;
208
+ if (!publicBaseUrl) {
209
+ var _context5;
210
+ throw new Error(_concatInstanceProperty__default["default"](_context5 = "'publicBaseUrl' is not defined for '".concat(tag, "' storage provider for ")).call(_context5, _classPrivateFieldLooseBase(this, _bucketEnvironment$1)[_bucketEnvironment$1], " or as default."));
211
+ }
193
212
  return publicBaseUrl;
194
213
  }
195
214
  }]);
@@ -418,7 +437,7 @@ let AwsStorageUploadScriptsGenerator = /*#__PURE__*/function () {
418
437
  buildRevision = _ref3.buildRevision,
419
438
  buildNumber = _ref3.buildNumber,
420
439
  applicationIndexOutFile = _ref3.applicationIndexOutFile;
421
- return _concatInstanceProperty__default["default"](_context19 = _concatInstanceProperty__default["default"](_context20 = _concatInstanceProperty__default["default"](_context21 = _concatInstanceProperty__default["default"](_context22 = _concatInstanceProperty__default["default"](_context23 = _concatInstanceProperty__default["default"](_context24 = _concatInstanceProperty__default["default"](_context25 = _concatInstanceProperty__default["default"](_context26 = _concatInstanceProperty__default["default"](_context27 = _concatInstanceProperty__default["default"](_context28 = _concatInstanceProperty__default["default"](_context29 = "\n #!/usr/bin/env bash\n\n echo \"Uploading static assets to Amazon S3 bucket ".concat(bucketUrl, "\"\n\n set -e\n\n aws s3 cp \"$(dirname \"$0\")/")).call(_context29, applicationIndexOutFile, "\" \\\n \"")).call(_context28, bucketUrl, "/\" \\\n --content-type=\"text/html\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context27, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n\n echo \"Creating version.json and uploading it to bucket ")).call(_context26, bucketUrl, "\"\n\n NODE_ENV=production ")).call(_context25, packageManagerName, " application-cli create-version \\\n --version-url=")).call(_context24, cdnUrl, "/")).call(_context23, _classPrivateFieldLooseBase(this, _bucketEnvironment2)[_bucketEnvironment2], "/version.json \\\n --build-revision=")).call(_context22, buildRevision, " \\\n --build-number=")).call(_context21, buildNumber, " \\\n --out-file=$(dirname \"$0\")/version.json\n\n aws s3 cp \"$(dirname \"$0\")\" \\\n \"")).call(_context20, bucketUrl, "/\" \\\n --exclude \"*\" \\\n --include \"version.json\" \\\n --content-type=\"application/json\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context19, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n");
440
+ return _concatInstanceProperty__default["default"](_context19 = _concatInstanceProperty__default["default"](_context20 = _concatInstanceProperty__default["default"](_context21 = _concatInstanceProperty__default["default"](_context22 = _concatInstanceProperty__default["default"](_context23 = _concatInstanceProperty__default["default"](_context24 = _concatInstanceProperty__default["default"](_context25 = _concatInstanceProperty__default["default"](_context26 = _concatInstanceProperty__default["default"](_context27 = _concatInstanceProperty__default["default"](_context28 = _concatInstanceProperty__default["default"](_context29 = "\n #!/usr/bin/env bash\n\n echo \"Uploading static assets to Amazon S3 bucket ".concat(bucketUrl, "\"\n\n set -e\n\n aws s3 cp \"$(dirname \"$0\")/")).call(_context29, applicationIndexOutFile, "\" \\\n \"")).call(_context28, bucketUrl, "/\" \\\n --content-type=\"text/html\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context27, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n\n echo \"Creating version.json and uploading it to bucket ")).call(_context26, bucketUrl, "\"\n\n NODE_ENV=production ")).call(_context25, packageManagerName, " application-cli create-version \\\n --version-url=")).call(_context24, cdnUrl, "/")).call(_context23, _classPrivateFieldLooseBase(this, _bucketEnvironment2)[_bucketEnvironment2], "/version.json \\\n --build-revision=")).call(_context22, buildRevision, " \\\n --build-number=")).call(_context21, buildNumber, " \\\n --out-file=$(dirname \"$0\")/version.json\n\n aws s3 cp \"$(dirname \"$0\")/version.json\" \\\n \"")).call(_context20, bucketUrl, "/\" \\\n --content-type=\"application/json\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context19, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n");
422
441
  }
423
442
  }, {
424
443
  key: "getProductionBundlesUploadScript",
@@ -607,7 +626,7 @@ async function compileEnvironmentApplicationIndexes(_ref3) {
607
626
  throw new Error(moveResult.stderr);
608
627
  }
609
628
  }
610
- async function command$3(cliFlags, cwd) {
629
+ async function command$4(cliFlags, cwd) {
611
630
  var _context3;
612
631
  const storageBucketConfig = await storageBucketsConfig.loadStorageBucketsConfig();
613
632
  const applicationDirectory = getApplicationDirectory(cwd);
@@ -780,7 +799,7 @@ const mapApplicationMenuConfigToGraqhQLMenuJson = config => {
780
799
  shouldRenderDivider: (_menuLinks$shouldRend = menuLinks.shouldRenderDivider) !== null && _menuLinks$shouldRend !== void 0 ? _menuLinks$shouldRend : false
781
800
  };
782
801
  };
783
- async function command$2(cliFlags, cwd) {
802
+ async function command$3(cliFlags, cwd) {
784
803
  const applicationDirectory = getApplicationDirectory(cwd);
785
804
  const monorepoRoot = findRoot.findRootSync(cwd);
786
805
  const dotenvPath = cliFlags.dotenvFolder && path__default["default"].join(monorepoRoot.rootDir, cliFlags.dotenvFolder);
@@ -811,7 +830,7 @@ async function command$2(cliFlags, cwd) {
811
830
  });
812
831
  }
813
832
 
814
- async function command$1(cliFlags) {
833
+ async function command$2(cliFlags) {
815
834
  const numberOfRollbacks = cliFlags.rollbacks - 1;
816
835
  let nextRollbacks;
817
836
  try {
@@ -847,6 +866,158 @@ async function command$1(cliFlags) {
847
866
  }
848
867
  }
849
868
 
869
+ /**
870
+ * This is heavily inspired by https://circleci.com/developer/orbs/orb/circleci/path-filtering.
871
+ *
872
+ * It detects changed files between `HEAD` and a base revision.
873
+ * To match them against configured RegEx tr
874
+ * All matched triggers will be written as a dotenv file.
875
+ * The dotenv file is read in a CircleCI step and be evaluated.
876
+ */
877
+ const git = {
878
+ // https://git-scm.com/docs/git-merge-base
879
+ base: (baseBranch, headRevision) => {
880
+ var _context;
881
+ return _concatInstanceProperty__default["default"](_context = "git merge-base ".concat(baseBranch, " ")).call(_context, headRevision);
882
+ },
883
+ // https://git-scm.com/docs/git-diff
884
+ changedFiles: (mergeRevision, headRevision) => {
885
+ var _context2;
886
+ return _concatInstanceProperty__default["default"](_context2 = "git diff --name-only ".concat(mergeRevision, " ")).call(_context2, headRevision);
887
+ },
888
+ commitMessage: headRevision => "git log --format=oneline -n 1 ".concat(headRevision)
889
+ };
890
+ const helpers = {
891
+ async writeOutDotEnvFile(cliFlags, cwd, matchingTriggers) {
892
+ var _context3;
893
+ // If desired read the env file and write out the matching triggers.
894
+ if (!cliFlags.outEnvFile) {
895
+ return;
896
+ }
897
+ const filePath = path__default$1["default"].join(fs__default["default"].realpathSync(cwd), cliFlags.outEnvFile);
898
+ const fileContents = _mapInstanceProperty__default["default"](_context3 = _Object$entries__default["default"](matchingTriggers)).call(_context3, _ref => {
899
+ var _context5;
900
+ let _ref2 = _slicedToArray(_ref, 2),
901
+ triggerName = _ref2[0],
902
+ triggerValue = _ref2[1];
903
+ const triggerNameForEnvFile = "".concat(snakeCase__default["default"](triggerName).toUpperCase());
904
+
905
+ // General pipeline optimization hints are not transformed
906
+ if (_startsWithInstanceProperty__default["default"](triggerName).call(triggerName, 'allowPipelineOptimizations')) {
907
+ var _context4;
908
+ return _concatInstanceProperty__default["default"](_context4 = "".concat(triggerNameForEnvFile, "=")).call(_context4, triggerValue);
909
+ }
910
+ return _concatInstanceProperty__default["default"](_context5 = "DID_".concat(triggerNameForEnvFile, "_CHANGE=")).call(_context5, triggerValue);
911
+ }).join('\n');
912
+ await fs__default["default"].promises.writeFile(filePath, fileContents);
913
+ if (!cliFlags.silent) {
914
+ console.log("\uD83D\uDCDD Wrote out file to '".concat(filePath, "' with contents:"));
915
+ console.log(fileContents);
916
+ }
917
+ },
918
+ async getChangedFiles(cliFlags) {
919
+ var _context6, _context7;
920
+ const baseCmdResult = await execa.command(git.base(cliFlags.baseBranch, cliFlags.headRevision));
921
+ const mergeRevision = baseCmdResult.stdout;
922
+ const changedFilesCmdResult = await execa.command(git.changedFiles(mergeRevision, cliFlags.headRevision));
923
+ const changedFiles = _filterInstanceProperty__default["default"](_context6 = _mapInstanceProperty__default["default"](_context7 = changedFilesCmdResult.stdout.split('\n')).call(_context7, filePath => _trimInstanceProperty__default["default"](filePath).call(filePath))).call(_context6, filePath => filePath.length > 0);
924
+ return changedFiles;
925
+ },
926
+ async matchTriggersAgainstChangedFiles(cliFlags, config, changedFiles) {
927
+ const matchedTriggers = {};
928
+
929
+ // Evaluate each trigger against each file.
930
+ _forEachInstanceProperty__default["default"](config).call(config, async trigger => {
931
+ var _trigger$exclude;
932
+ const hasTriggerBeenInitialized = typeof matchedTriggers[trigger.name] === 'number';
933
+
934
+ // Given the trigger with this name was never evaluated it has to be defaulted to 0.
935
+ // As without any matches we should indicate nothing changed.
936
+ if (!hasTriggerBeenInitialized) {
937
+ matchedTriggers[trigger.name] = 0;
938
+ }
939
+ // Given the trigger was already evaluated to be positive we can skip this evaluation.
940
+ if (matchedTriggers[trigger.name] === 1) {
941
+ return matchedTriggers;
942
+ }
943
+
944
+ // In any other case we evaluate this trigger.
945
+ const anyFileChangedForTrigger = _someInstanceProperty__default["default"](micromatch__default["default"]).call(micromatch__default["default"], changedFiles, trigger.include, {
946
+ ignore: trigger.ignore
947
+ });
948
+ if (!cliFlags.silent && anyFileChangedForTrigger) {
949
+ console.log("\u2139\uFE0F Files for trigger ".concat(trigger.name, " changed."));
950
+ }
951
+ let onlyExcludedFilesChangedForTrigger = false;
952
+ if (((_trigger$exclude = trigger.exclude) === null || _trigger$exclude === void 0 ? void 0 : _trigger$exclude.length) > 0) {
953
+ // NOTE: `micromatch.every` evaluates if every file matches
954
+ // every pattern.
955
+ // We need to evaluate if every file matches some pattern.
956
+ onlyExcludedFilesChangedForTrigger = _everyInstanceProperty__default["default"](changedFiles).call(changedFiles, changedFile => {
957
+ return micromatch__default["default"].isMatch(changedFile, trigger.exclude, {
958
+ ignore: trigger.ignore
959
+ });
960
+ });
961
+ }
962
+ if (!cliFlags.silent && onlyExcludedFilesChangedForTrigger) {
963
+ console.log("\u2139\uFE0F Only excluded files for trigger ".concat(trigger.name, " changed."));
964
+ }
965
+ if (onlyExcludedFilesChangedForTrigger) {
966
+ matchedTriggers[trigger.name] = 0;
967
+ } else {
968
+ matchedTriggers[trigger.name] = Number(anyFileChangedForTrigger);
969
+ }
970
+ return matchedTriggers;
971
+ });
972
+ return matchedTriggers;
973
+ }
974
+ };
975
+ async function command$1(cliFlags, config, cwd) {
976
+ const enablePipelineOptimizations = process.env.ENABLE_PIPELINE_OPTIMIZATIONS === '1';
977
+ const isDevelopmentBranch = cliFlags.branch !== cliFlags.baseBranch;
978
+ const triggersContainingSharedFiles = _filterInstanceProperty__default["default"](config).call(config, trigger => trigger.containsSharedFiles);
979
+ if (!cliFlags.silent) {
980
+ var _context8;
981
+ console.log("\u2139\uFE0F Pipeline optimizations are ".concat(enablePipelineOptimizations ? 'enabled' : 'disabled', "."));
982
+ console.log("\u2139\uFE0F Changes have been commited to the ".concat(isDevelopmentBranch ? 'a development' : 'the main', " branch."));
983
+ console.log(_concatInstanceProperty__default["default"](_context8 = "\uD83D\uDEA7 Comparing '".concat(cliFlags.baseBranch, "' against '")).call(_context8, cliFlags.headRevision, "' to determine changed files."));
984
+ }
985
+
986
+ // Collect and parse changed files from git comparing base and head revision.
987
+ const changedFiles = await helpers.getChangedFiles(cliFlags);
988
+ if (!cliFlags.silent) {
989
+ if (changedFiles.length === 0) {
990
+ console.log("\u2139\uFE0F No changes found.");
991
+ } else {
992
+ console.log("\u2139\uFE0F ".concat(changedFiles.length, " changes found."));
993
+ }
994
+ }
995
+
996
+ // Read the trigger file to match the changed files against.
997
+ const matchedTriggers = await helpers.matchTriggersAgainstChangedFiles(cliFlags, config, changedFiles);
998
+ const commitMessageCmdResult = await execa.command(git.commitMessage(cliFlags.headRevision));
999
+ const commitMessage = commitMessageCmdResult.stdout;
1000
+ const hasCommitMessageTrigger = commitMessage && _includesInstanceProperty__default["default"](commitMessage).call(commitMessage, '[ci all]');
1001
+ const doesSharedTriggerMatch = _someInstanceProperty__default["default"](triggersContainingSharedFiles).call(triggersContainingSharedFiles, triggerContainingSharedFiles => matchedTriggers[triggerContainingSharedFiles.name] === 1);
1002
+ if (!cliFlags.silent) {
1003
+ console.log("\u2139\uFE0F The git commit message ".concat(hasCommitMessageTrigger ? 'does' : 'does not', " contain a [ci all] trigger."));
1004
+ }
1005
+ const doesPackageFolderTriggerMatch = matchedTriggers[cliFlags.triggerName] === 1;
1006
+ if (enablePipelineOptimizations && isDevelopmentBranch && !hasCommitMessageTrigger && !doesSharedTriggerMatch && !doesPackageFolderTriggerMatch) {
1007
+ if (!cliFlags.silent) {
1008
+ console.log("\u2139\uFE0F No relevant changes found for ".concat(cliFlags.triggerName, "."));
1009
+ }
1010
+ matchedTriggers['allowPipelineOptimizationsForTrigger'] = 1;
1011
+ } else {
1012
+ if (!cliFlags.silent) {
1013
+ console.log("\u2139\uFE0F Relevant changes found for ".concat(cliFlags.triggerName, "."));
1014
+ }
1015
+ matchedTriggers['allowPipelineOptimizationsForTrigger'] = 0;
1016
+ }
1017
+ await helpers.writeOutDotEnvFile(cliFlags, cwd, matchedTriggers);
1018
+ return matchedTriggers;
1019
+ }
1020
+
850
1021
  function ownKeys(e, r) { var t = _Object$keys__default["default"](e); if (_Object$getOwnPropertySymbols__default["default"]) { var o = _Object$getOwnPropertySymbols__default["default"](e); r && (o = _filterInstanceProperty__default["default"](o).call(o, function (r) { return _Object$getOwnPropertyDescriptor__default["default"](e, r).enumerable; })), t.push.apply(t, o); } return t; }
851
1022
  function _objectSpread(e) { for (var r = 1; r < arguments.length; r++) { var _context, _context2; var t = null != arguments[r] ? arguments[r] : {}; r % 2 ? _forEachInstanceProperty__default["default"](_context = ownKeys(Object(t), !0)).call(_context, function (r) { _defineProperty(e, r, t[r]); }) : _Object$getOwnPropertyDescriptors__default["default"] ? _Object$defineProperties__default["default"](e, _Object$getOwnPropertyDescriptors__default["default"](t)) : _forEachInstanceProperty__default["default"](_context2 = ownKeys(Object(t))).call(_context2, function (r) { _Object$defineProperty__default["default"](e, r, _Object$getOwnPropertyDescriptor__default["default"](t, r)); }); } return e; }
852
1023
  const baseMenuProperties = {
@@ -967,7 +1138,7 @@ async function command(cliFlags) {
967
1138
 
968
1139
  var pkgJson = {
969
1140
  name: "@commercetools-frontend/application-cli",
970
- version: "2.3.2",
1141
+ version: "2.5.0",
971
1142
  description: "Internal CLI to manage Merchant Center application deployments across various environments.",
972
1143
  keywords: [
973
1144
  "commercetools",
@@ -1002,11 +1173,15 @@ var pkgJson = {
1002
1173
  execa: "5.1.1",
1003
1174
  jsonschema: "^1.4.1",
1004
1175
  listr2: "8.2.0",
1176
+ lodash: "4.17.21",
1177
+ micromatch: "4.0.5",
1005
1178
  "node-fetch": "2.7.0",
1006
1179
  "ts-deepmerge": "7.0.0"
1007
1180
  },
1008
1181
  devDependencies: {
1009
1182
  "@tsconfig/node20": "20.1.4",
1183
+ "@types/lodash": "^4.14.198",
1184
+ "@types/micromatch": "4.0.6",
1010
1185
  "@types/node": "20.12.7",
1011
1186
  typescript: "5.2.2"
1012
1187
  },
@@ -1048,11 +1223,11 @@ const run = async () => {
1048
1223
  }).option('--ci-assets-root-path [path]', '(optional) A replacement value for the scripts root path only used on CI (e.g. "--ci-assets-root-path=/root/") used in generated scripts.').option('--skip-menu', '(optional) If provided, it will skip uploading the `menu.json`.', {
1049
1224
  default: false
1050
1225
  }).action(async options => {
1051
- await command$3(options, cwd);
1226
+ await command$4(options, cwd);
1052
1227
  });
1053
1228
  const usageCompileMenu = 'Compile the menu links of an application into a `menu.json`. This is only required for internal applications';
1054
1229
  cli.command('compile-menu', usageCompileMenu).usage("compile-menu \n\n ".concat(usageCompileMenu)).option('--dotenv-folder [string]', '(optional) The path to a folder containing a dotenv file `.env.production` and a cloud-environment specific dotenv file (for example `.env.gcp-production-eu`). Those values are parsed and merged together to be used by the application config.').action(async options => {
1055
- await command$2(options, cwd);
1230
+ await command$3(options, cwd);
1056
1231
  });
1057
1232
  const usageValidateMenu = 'Validate compiled `menu.json` file';
1058
1233
  cli.command('validate-menu', usageValidateMenu).usage("validate-menu \n\n ".concat(usageValidateMenu)).option('--input-file <path>', '(required) The path to the `menu.json` file to be validated.').option('--navigation [string]', '(optional) Location of the menu navigation. Possible values are `top`.').action(async options => {
@@ -1062,7 +1237,20 @@ const run = async () => {
1062
1237
  cli.command('create-version', usageCreateVersion).usage("create-version \n\n ".concat(usageCreateVersion)).option('--version-url <url>', "(required) The path of an application's current `version.json` within the storage bucket.").option('--rollbacks [int]', '(optional) The number of max rollbacks to keep', {
1063
1238
  default: 15
1064
1239
  }).option('--out-file [path]', '(optional) The path to the file where to write the JSON. If not specified, the JSON is printed to stdout.').action(async options => {
1065
- await command$1(options);
1240
+ await command$2(options);
1241
+ });
1242
+
1243
+ // Command: Evaluate change triggers
1244
+ const usageEvaluateChangeTriggers = 'Evaluates changed files against a base and evaluates them against defined triggers.';
1245
+ cli.command('evaluate-change-triggers', usageEvaluateChangeTriggers).usage("evaluate-change-triggers \n\n ".concat(usageEvaluateChangeTriggers)).option('--branch <string>', 'The branch of the pull request', {
1246
+ default: process.env.CIRCLE_BRANCH
1247
+ }).option('--base-branch <string>', 'The base revision of the git commit compare against (e.g. "main")').option('--head-revision <string>', 'The revision of the git head to compare with', {
1248
+ default: process.env.CIRCLE_SHA1
1249
+ }).option('--trigger-name <string>', 'The trigger to evaluate for.').option('--silent', '(optional) Disable logging', {
1250
+ default: false
1251
+ }).option('--out-env-file [string]', '(optional) A file path where the matched triggers are written as a dotenv file.').action(async options => {
1252
+ const config = await storageBucketsConfig.loadConfig('circleci-change-triggers', []);
1253
+ await command$1(options, config, cwd);
1066
1254
  });
1067
1255
  cli.help();
1068
1256
  cli.version(pkgJson.version);