@commercetools-frontend/application-cli 2.3.2 → 2.5.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/cli/dist/commercetools-frontend-application-cli-cli.cjs.dev.js +198 -10
- package/cli/dist/commercetools-frontend-application-cli-cli.cjs.prod.js +198 -10
- package/cli/dist/commercetools-frontend-application-cli-cli.esm.js +192 -11
- package/dist/commercetools-frontend-application-cli.cjs.dev.js +3 -3
- package/dist/commercetools-frontend-application-cli.cjs.prod.js +3 -3
- package/dist/commercetools-frontend-application-cli.esm.js +3 -1
- package/dist/declarations/src/constants.d.ts +2 -8
- package/dist/declarations/src/storage-buckets-config.d.ts +2 -1
- package/dist/declarations/src/types.d.ts +18 -1
- package/dist/{storage-buckets-config-8e7abeff.cjs.dev.js → storage-buckets-config-0a380c12.cjs.prod.js} +39 -18
- package/dist/{storage-buckets-config-f41d07bc.esm.js → storage-buckets-config-896b4064.esm.js} +37 -17
- package/dist/{storage-buckets-config-e092f613.cjs.prod.js → storage-buckets-config-cf50df38.cjs.dev.js} +39 -18
- package/package.json +5 -1
|
@@ -18,7 +18,7 @@ var _mapInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instan
|
|
|
18
18
|
var _Object$entries = require('@babel/runtime-corejs3/core-js-stable/object/entries');
|
|
19
19
|
var _Set = require('@babel/runtime-corejs3/core-js-stable/set');
|
|
20
20
|
var _flatMapInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/flat-map');
|
|
21
|
-
var storageBucketsConfig = require('../../dist/storage-buckets-config-
|
|
21
|
+
var storageBucketsConfig = require('../../dist/storage-buckets-config-cf50df38.cjs.dev.js');
|
|
22
22
|
var fs = require('node:fs');
|
|
23
23
|
var path$1 = require('node:path');
|
|
24
24
|
var listr2 = require('listr2');
|
|
@@ -41,9 +41,17 @@ var _JSON$stringify = require('@babel/runtime-corejs3/core-js-stable/json/string
|
|
|
41
41
|
var applicationConfig = require('@commercetools-frontend/application-config');
|
|
42
42
|
var l10n = require('@commercetools-frontend/l10n');
|
|
43
43
|
var _sliceInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/slice');
|
|
44
|
+
var _startsWithInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/starts-with');
|
|
45
|
+
var _trimInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/trim');
|
|
46
|
+
var _someInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/some');
|
|
47
|
+
var _everyInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/every');
|
|
48
|
+
var _includesInstanceProperty = require('@babel/runtime-corejs3/core-js-stable/instance/includes');
|
|
49
|
+
var micromatch = require('micromatch');
|
|
50
|
+
var snakeCase = require('lodash/snakeCase');
|
|
44
51
|
var jsonschema = require('jsonschema');
|
|
45
52
|
require('cosmiconfig');
|
|
46
53
|
require('ts-deepmerge');
|
|
54
|
+
require('lodash');
|
|
47
55
|
|
|
48
56
|
function _interopDefault (e) { return e && e.__esModule ? e : { 'default': e }; }
|
|
49
57
|
|
|
@@ -70,6 +78,13 @@ var fs__default$1 = /*#__PURE__*/_interopDefault(fs$1);
|
|
|
70
78
|
var _findInstanceProperty__default = /*#__PURE__*/_interopDefault(_findInstanceProperty);
|
|
71
79
|
var _JSON$stringify__default = /*#__PURE__*/_interopDefault(_JSON$stringify);
|
|
72
80
|
var _sliceInstanceProperty__default = /*#__PURE__*/_interopDefault(_sliceInstanceProperty);
|
|
81
|
+
var _startsWithInstanceProperty__default = /*#__PURE__*/_interopDefault(_startsWithInstanceProperty);
|
|
82
|
+
var _trimInstanceProperty__default = /*#__PURE__*/_interopDefault(_trimInstanceProperty);
|
|
83
|
+
var _someInstanceProperty__default = /*#__PURE__*/_interopDefault(_someInstanceProperty);
|
|
84
|
+
var _everyInstanceProperty__default = /*#__PURE__*/_interopDefault(_everyInstanceProperty);
|
|
85
|
+
var _includesInstanceProperty__default = /*#__PURE__*/_interopDefault(_includesInstanceProperty);
|
|
86
|
+
var micromatch__default = /*#__PURE__*/_interopDefault(micromatch);
|
|
87
|
+
var snakeCase__default = /*#__PURE__*/_interopDefault(snakeCase);
|
|
73
88
|
|
|
74
89
|
function getApplicationDirectory(cwd) {
|
|
75
90
|
return fs__default["default"].realpathSync(cwd);
|
|
@@ -85,7 +100,7 @@ function isCI() {
|
|
|
85
100
|
}
|
|
86
101
|
|
|
87
102
|
function ownKeys$4(e, r) { var t = _Object$keys__default["default"](e); if (_Object$getOwnPropertySymbols__default["default"]) { var o = _Object$getOwnPropertySymbols__default["default"](e); r && (o = _filterInstanceProperty__default["default"](o).call(o, function (r) { return _Object$getOwnPropertyDescriptor__default["default"](e, r).enumerable; })), t.push.apply(t, o); } return t; }
|
|
88
|
-
function _objectSpread$4(e) { for (var r = 1; r < arguments.length; r++) { var
|
|
103
|
+
function _objectSpread$4(e) { for (var r = 1; r < arguments.length; r++) { var _context6, _context7; var t = null != arguments[r] ? arguments[r] : {}; r % 2 ? _forEachInstanceProperty__default["default"](_context6 = ownKeys$4(Object(t), !0)).call(_context6, function (r) { _defineProperty(e, r, t[r]); }) : _Object$getOwnPropertyDescriptors__default["default"] ? _Object$defineProperties__default["default"](e, _Object$getOwnPropertyDescriptors__default["default"](t)) : _forEachInstanceProperty__default["default"](_context7 = ownKeys$4(Object(t))).call(_context7, function (r) { _Object$defineProperty__default["default"](e, r, _Object$getOwnPropertyDescriptor__default["default"](t, r)); }); } return e; }
|
|
89
104
|
function _callSuper(_this, derived, args) {
|
|
90
105
|
function isNativeReflectConstruct() {
|
|
91
106
|
if (typeof Reflect === "undefined" || !_Reflect$construct__default["default"]) return false;
|
|
@@ -190,6 +205,10 @@ let StorageProvider = /*#__PURE__*/function () {
|
|
|
190
205
|
}
|
|
191
206
|
const storageProvider = storageBucketsConfig.storageProviders[tag];
|
|
192
207
|
const publicBaseUrl = (_storageProvider$urls3 = storageProvider.urls.public[_classPrivateFieldLooseBase(this, _bucketEnvironment$1)[_bucketEnvironment$1]]) !== null && _storageProvider$urls3 !== void 0 ? _storageProvider$urls3 : storageProvider.urls.public.default;
|
|
208
|
+
if (!publicBaseUrl) {
|
|
209
|
+
var _context5;
|
|
210
|
+
throw new Error(_concatInstanceProperty__default["default"](_context5 = "'publicBaseUrl' is not defined for '".concat(tag, "' storage provider for ")).call(_context5, _classPrivateFieldLooseBase(this, _bucketEnvironment$1)[_bucketEnvironment$1], " or as default."));
|
|
211
|
+
}
|
|
193
212
|
return publicBaseUrl;
|
|
194
213
|
}
|
|
195
214
|
}]);
|
|
@@ -418,7 +437,7 @@ let AwsStorageUploadScriptsGenerator = /*#__PURE__*/function () {
|
|
|
418
437
|
buildRevision = _ref3.buildRevision,
|
|
419
438
|
buildNumber = _ref3.buildNumber,
|
|
420
439
|
applicationIndexOutFile = _ref3.applicationIndexOutFile;
|
|
421
|
-
return _concatInstanceProperty__default["default"](_context19 = _concatInstanceProperty__default["default"](_context20 = _concatInstanceProperty__default["default"](_context21 = _concatInstanceProperty__default["default"](_context22 = _concatInstanceProperty__default["default"](_context23 = _concatInstanceProperty__default["default"](_context24 = _concatInstanceProperty__default["default"](_context25 = _concatInstanceProperty__default["default"](_context26 = _concatInstanceProperty__default["default"](_context27 = _concatInstanceProperty__default["default"](_context28 = _concatInstanceProperty__default["default"](_context29 = "\n #!/usr/bin/env bash\n\n echo \"Uploading static assets to Amazon S3 bucket ".concat(bucketUrl, "\"\n\n set -e\n\n aws s3 cp \"$(dirname \"$0\")/")).call(_context29, applicationIndexOutFile, "\" \\\n \"")).call(_context28, bucketUrl, "/\" \\\n --content-type=\"text/html\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context27, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n\n echo \"Creating version.json and uploading it to bucket ")).call(_context26, bucketUrl, "\"\n\n NODE_ENV=production ")).call(_context25, packageManagerName, " application-cli create-version \\\n --version-url=")).call(_context24, cdnUrl, "/")).call(_context23, _classPrivateFieldLooseBase(this, _bucketEnvironment2)[_bucketEnvironment2], "/version.json \\\n --build-revision=")).call(_context22, buildRevision, " \\\n --build-number=")).call(_context21, buildNumber, " \\\n --out-file=$(dirname \"$0\")/version.json\n\n aws s3 cp \"$(dirname \"$0\")\" \\\n \"")).call(_context20, bucketUrl, "/\" \\\n --
|
|
440
|
+
return _concatInstanceProperty__default["default"](_context19 = _concatInstanceProperty__default["default"](_context20 = _concatInstanceProperty__default["default"](_context21 = _concatInstanceProperty__default["default"](_context22 = _concatInstanceProperty__default["default"](_context23 = _concatInstanceProperty__default["default"](_context24 = _concatInstanceProperty__default["default"](_context25 = _concatInstanceProperty__default["default"](_context26 = _concatInstanceProperty__default["default"](_context27 = _concatInstanceProperty__default["default"](_context28 = _concatInstanceProperty__default["default"](_context29 = "\n #!/usr/bin/env bash\n\n echo \"Uploading static assets to Amazon S3 bucket ".concat(bucketUrl, "\"\n\n set -e\n\n aws s3 cp \"$(dirname \"$0\")/")).call(_context29, applicationIndexOutFile, "\" \\\n \"")).call(_context28, bucketUrl, "/\" \\\n --content-type=\"text/html\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context27, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n\n echo \"Creating version.json and uploading it to bucket ")).call(_context26, bucketUrl, "\"\n\n NODE_ENV=production ")).call(_context25, packageManagerName, " application-cli create-version \\\n --version-url=")).call(_context24, cdnUrl, "/")).call(_context23, _classPrivateFieldLooseBase(this, _bucketEnvironment2)[_bucketEnvironment2], "/version.json \\\n --build-revision=")).call(_context22, buildRevision, " \\\n --build-number=")).call(_context21, buildNumber, " \\\n --out-file=$(dirname \"$0\")/version.json\n\n aws s3 cp \"$(dirname \"$0\")/version.json\" \\\n \"")).call(_context20, bucketUrl, "/\" \\\n --content-type=\"application/json\" \\\n --cache-control=\"public,max-age=0,no-transform\" \\\n --profile ")).call(_context19, _classPrivateFieldLooseBase(this, _bucketRegion2)[_bucketRegion2], "\n");
|
|
422
441
|
}
|
|
423
442
|
}, {
|
|
424
443
|
key: "getProductionBundlesUploadScript",
|
|
@@ -607,7 +626,7 @@ async function compileEnvironmentApplicationIndexes(_ref3) {
|
|
|
607
626
|
throw new Error(moveResult.stderr);
|
|
608
627
|
}
|
|
609
628
|
}
|
|
610
|
-
async function command$
|
|
629
|
+
async function command$4(cliFlags, cwd) {
|
|
611
630
|
var _context3;
|
|
612
631
|
const storageBucketConfig = await storageBucketsConfig.loadStorageBucketsConfig();
|
|
613
632
|
const applicationDirectory = getApplicationDirectory(cwd);
|
|
@@ -780,7 +799,7 @@ const mapApplicationMenuConfigToGraqhQLMenuJson = config => {
|
|
|
780
799
|
shouldRenderDivider: (_menuLinks$shouldRend = menuLinks.shouldRenderDivider) !== null && _menuLinks$shouldRend !== void 0 ? _menuLinks$shouldRend : false
|
|
781
800
|
};
|
|
782
801
|
};
|
|
783
|
-
async function command$
|
|
802
|
+
async function command$3(cliFlags, cwd) {
|
|
784
803
|
const applicationDirectory = getApplicationDirectory(cwd);
|
|
785
804
|
const monorepoRoot = findRoot.findRootSync(cwd);
|
|
786
805
|
const dotenvPath = cliFlags.dotenvFolder && path__default["default"].join(monorepoRoot.rootDir, cliFlags.dotenvFolder);
|
|
@@ -811,7 +830,7 @@ async function command$2(cliFlags, cwd) {
|
|
|
811
830
|
});
|
|
812
831
|
}
|
|
813
832
|
|
|
814
|
-
async function command$
|
|
833
|
+
async function command$2(cliFlags) {
|
|
815
834
|
const numberOfRollbacks = cliFlags.rollbacks - 1;
|
|
816
835
|
let nextRollbacks;
|
|
817
836
|
try {
|
|
@@ -847,6 +866,158 @@ async function command$1(cliFlags) {
|
|
|
847
866
|
}
|
|
848
867
|
}
|
|
849
868
|
|
|
869
|
+
/**
|
|
870
|
+
* This is heavily inspired by https://circleci.com/developer/orbs/orb/circleci/path-filtering.
|
|
871
|
+
*
|
|
872
|
+
* It detects changed files between `HEAD` and a base revision.
|
|
873
|
+
* To match them against configured RegEx tr
|
|
874
|
+
* All matched triggers will be written as a dotenv file.
|
|
875
|
+
* The dotenv file is read in a CircleCI step and be evaluated.
|
|
876
|
+
*/
|
|
877
|
+
const git = {
|
|
878
|
+
// https://git-scm.com/docs/git-merge-base
|
|
879
|
+
base: (baseBranch, headRevision) => {
|
|
880
|
+
var _context;
|
|
881
|
+
return _concatInstanceProperty__default["default"](_context = "git merge-base ".concat(baseBranch, " ")).call(_context, headRevision);
|
|
882
|
+
},
|
|
883
|
+
// https://git-scm.com/docs/git-diff
|
|
884
|
+
changedFiles: (mergeRevision, headRevision) => {
|
|
885
|
+
var _context2;
|
|
886
|
+
return _concatInstanceProperty__default["default"](_context2 = "git diff --name-only ".concat(mergeRevision, " ")).call(_context2, headRevision);
|
|
887
|
+
},
|
|
888
|
+
commitMessage: headRevision => "git log --format=oneline -n 1 ".concat(headRevision)
|
|
889
|
+
};
|
|
890
|
+
const helpers = {
|
|
891
|
+
async writeOutDotEnvFile(cliFlags, cwd, matchingTriggers) {
|
|
892
|
+
var _context3;
|
|
893
|
+
// If desired read the env file and write out the matching triggers.
|
|
894
|
+
if (!cliFlags.outEnvFile) {
|
|
895
|
+
return;
|
|
896
|
+
}
|
|
897
|
+
const filePath = path__default$1["default"].join(fs__default["default"].realpathSync(cwd), cliFlags.outEnvFile);
|
|
898
|
+
const fileContents = _mapInstanceProperty__default["default"](_context3 = _Object$entries__default["default"](matchingTriggers)).call(_context3, _ref => {
|
|
899
|
+
var _context5;
|
|
900
|
+
let _ref2 = _slicedToArray(_ref, 2),
|
|
901
|
+
triggerName = _ref2[0],
|
|
902
|
+
triggerValue = _ref2[1];
|
|
903
|
+
const triggerNameForEnvFile = "".concat(snakeCase__default["default"](triggerName).toUpperCase());
|
|
904
|
+
|
|
905
|
+
// General pipeline optimization hints are not transformed
|
|
906
|
+
if (_startsWithInstanceProperty__default["default"](triggerName).call(triggerName, 'allowPipelineOptimizations')) {
|
|
907
|
+
var _context4;
|
|
908
|
+
return _concatInstanceProperty__default["default"](_context4 = "".concat(triggerNameForEnvFile, "=")).call(_context4, triggerValue);
|
|
909
|
+
}
|
|
910
|
+
return _concatInstanceProperty__default["default"](_context5 = "DID_".concat(triggerNameForEnvFile, "_CHANGE=")).call(_context5, triggerValue);
|
|
911
|
+
}).join('\n');
|
|
912
|
+
await fs__default["default"].promises.writeFile(filePath, fileContents);
|
|
913
|
+
if (!cliFlags.silent) {
|
|
914
|
+
console.log("\uD83D\uDCDD Wrote out file to '".concat(filePath, "' with contents:"));
|
|
915
|
+
console.log(fileContents);
|
|
916
|
+
}
|
|
917
|
+
},
|
|
918
|
+
async getChangedFiles(cliFlags) {
|
|
919
|
+
var _context6, _context7;
|
|
920
|
+
const baseCmdResult = await execa.command(git.base(cliFlags.baseBranch, cliFlags.headRevision));
|
|
921
|
+
const mergeRevision = baseCmdResult.stdout;
|
|
922
|
+
const changedFilesCmdResult = await execa.command(git.changedFiles(mergeRevision, cliFlags.headRevision));
|
|
923
|
+
const changedFiles = _filterInstanceProperty__default["default"](_context6 = _mapInstanceProperty__default["default"](_context7 = changedFilesCmdResult.stdout.split('\n')).call(_context7, filePath => _trimInstanceProperty__default["default"](filePath).call(filePath))).call(_context6, filePath => filePath.length > 0);
|
|
924
|
+
return changedFiles;
|
|
925
|
+
},
|
|
926
|
+
async matchTriggersAgainstChangedFiles(cliFlags, config, changedFiles) {
|
|
927
|
+
const matchedTriggers = {};
|
|
928
|
+
|
|
929
|
+
// Evaluate each trigger against each file.
|
|
930
|
+
_forEachInstanceProperty__default["default"](config).call(config, async trigger => {
|
|
931
|
+
var _trigger$exclude;
|
|
932
|
+
const hasTriggerBeenInitialized = typeof matchedTriggers[trigger.name] === 'number';
|
|
933
|
+
|
|
934
|
+
// Given the trigger with this name was never evaluated it has to be defaulted to 0.
|
|
935
|
+
// As without any matches we should indicate nothing changed.
|
|
936
|
+
if (!hasTriggerBeenInitialized) {
|
|
937
|
+
matchedTriggers[trigger.name] = 0;
|
|
938
|
+
}
|
|
939
|
+
// Given the trigger was already evaluated to be positive we can skip this evaluation.
|
|
940
|
+
if (matchedTriggers[trigger.name] === 1) {
|
|
941
|
+
return matchedTriggers;
|
|
942
|
+
}
|
|
943
|
+
|
|
944
|
+
// In any other case we evaluate this trigger.
|
|
945
|
+
const anyFileChangedForTrigger = _someInstanceProperty__default["default"](micromatch__default["default"]).call(micromatch__default["default"], changedFiles, trigger.include, {
|
|
946
|
+
ignore: trigger.ignore
|
|
947
|
+
});
|
|
948
|
+
if (!cliFlags.silent && anyFileChangedForTrigger) {
|
|
949
|
+
console.log("\u2139\uFE0F Files for trigger ".concat(trigger.name, " changed."));
|
|
950
|
+
}
|
|
951
|
+
let onlyExcludedFilesChangedForTrigger = false;
|
|
952
|
+
if (((_trigger$exclude = trigger.exclude) === null || _trigger$exclude === void 0 ? void 0 : _trigger$exclude.length) > 0) {
|
|
953
|
+
// NOTE: `micromatch.every` evaluates if every file matches
|
|
954
|
+
// every pattern.
|
|
955
|
+
// We need to evaluate if every file matches some pattern.
|
|
956
|
+
onlyExcludedFilesChangedForTrigger = _everyInstanceProperty__default["default"](changedFiles).call(changedFiles, changedFile => {
|
|
957
|
+
return micromatch__default["default"].isMatch(changedFile, trigger.exclude, {
|
|
958
|
+
ignore: trigger.ignore
|
|
959
|
+
});
|
|
960
|
+
});
|
|
961
|
+
}
|
|
962
|
+
if (!cliFlags.silent && onlyExcludedFilesChangedForTrigger) {
|
|
963
|
+
console.log("\u2139\uFE0F Only excluded files for trigger ".concat(trigger.name, " changed."));
|
|
964
|
+
}
|
|
965
|
+
if (onlyExcludedFilesChangedForTrigger) {
|
|
966
|
+
matchedTriggers[trigger.name] = 0;
|
|
967
|
+
} else {
|
|
968
|
+
matchedTriggers[trigger.name] = Number(anyFileChangedForTrigger);
|
|
969
|
+
}
|
|
970
|
+
return matchedTriggers;
|
|
971
|
+
});
|
|
972
|
+
return matchedTriggers;
|
|
973
|
+
}
|
|
974
|
+
};
|
|
975
|
+
async function command$1(cliFlags, config, cwd) {
|
|
976
|
+
const enablePipelineOptimizations = process.env.ENABLE_PIPELINE_OPTIMIZATIONS === '1';
|
|
977
|
+
const isDevelopmentBranch = cliFlags.branch !== cliFlags.baseBranch;
|
|
978
|
+
const triggersContainingSharedFiles = _filterInstanceProperty__default["default"](config).call(config, trigger => trigger.containsSharedFiles);
|
|
979
|
+
if (!cliFlags.silent) {
|
|
980
|
+
var _context8;
|
|
981
|
+
console.log("\u2139\uFE0F Pipeline optimizations are ".concat(enablePipelineOptimizations ? 'enabled' : 'disabled', "."));
|
|
982
|
+
console.log("\u2139\uFE0F Changes have been commited to the ".concat(isDevelopmentBranch ? 'a development' : 'the main', " branch."));
|
|
983
|
+
console.log(_concatInstanceProperty__default["default"](_context8 = "\uD83D\uDEA7 Comparing '".concat(cliFlags.baseBranch, "' against '")).call(_context8, cliFlags.headRevision, "' to determine changed files."));
|
|
984
|
+
}
|
|
985
|
+
|
|
986
|
+
// Collect and parse changed files from git comparing base and head revision.
|
|
987
|
+
const changedFiles = await helpers.getChangedFiles(cliFlags);
|
|
988
|
+
if (!cliFlags.silent) {
|
|
989
|
+
if (changedFiles.length === 0) {
|
|
990
|
+
console.log("\u2139\uFE0F No changes found.");
|
|
991
|
+
} else {
|
|
992
|
+
console.log("\u2139\uFE0F ".concat(changedFiles.length, " changes found."));
|
|
993
|
+
}
|
|
994
|
+
}
|
|
995
|
+
|
|
996
|
+
// Read the trigger file to match the changed files against.
|
|
997
|
+
const matchedTriggers = await helpers.matchTriggersAgainstChangedFiles(cliFlags, config, changedFiles);
|
|
998
|
+
const commitMessageCmdResult = await execa.command(git.commitMessage(cliFlags.headRevision));
|
|
999
|
+
const commitMessage = commitMessageCmdResult.stdout;
|
|
1000
|
+
const hasCommitMessageTrigger = commitMessage && _includesInstanceProperty__default["default"](commitMessage).call(commitMessage, '[ci all]');
|
|
1001
|
+
const doesSharedTriggerMatch = _someInstanceProperty__default["default"](triggersContainingSharedFiles).call(triggersContainingSharedFiles, triggerContainingSharedFiles => matchedTriggers[triggerContainingSharedFiles.name] === 1);
|
|
1002
|
+
if (!cliFlags.silent) {
|
|
1003
|
+
console.log("\u2139\uFE0F The git commit message ".concat(hasCommitMessageTrigger ? 'does' : 'does not', " contain a [ci all] trigger."));
|
|
1004
|
+
}
|
|
1005
|
+
const doesPackageFolderTriggerMatch = matchedTriggers[cliFlags.triggerName] === 1;
|
|
1006
|
+
if (enablePipelineOptimizations && isDevelopmentBranch && !hasCommitMessageTrigger && !doesSharedTriggerMatch && !doesPackageFolderTriggerMatch) {
|
|
1007
|
+
if (!cliFlags.silent) {
|
|
1008
|
+
console.log("\u2139\uFE0F No relevant changes found for ".concat(cliFlags.triggerName, "."));
|
|
1009
|
+
}
|
|
1010
|
+
matchedTriggers['allowPipelineOptimizationsForTrigger'] = 1;
|
|
1011
|
+
} else {
|
|
1012
|
+
if (!cliFlags.silent) {
|
|
1013
|
+
console.log("\u2139\uFE0F Relevant changes found for ".concat(cliFlags.triggerName, "."));
|
|
1014
|
+
}
|
|
1015
|
+
matchedTriggers['allowPipelineOptimizationsForTrigger'] = 0;
|
|
1016
|
+
}
|
|
1017
|
+
await helpers.writeOutDotEnvFile(cliFlags, cwd, matchedTriggers);
|
|
1018
|
+
return matchedTriggers;
|
|
1019
|
+
}
|
|
1020
|
+
|
|
850
1021
|
function ownKeys(e, r) { var t = _Object$keys__default["default"](e); if (_Object$getOwnPropertySymbols__default["default"]) { var o = _Object$getOwnPropertySymbols__default["default"](e); r && (o = _filterInstanceProperty__default["default"](o).call(o, function (r) { return _Object$getOwnPropertyDescriptor__default["default"](e, r).enumerable; })), t.push.apply(t, o); } return t; }
|
|
851
1022
|
function _objectSpread(e) { for (var r = 1; r < arguments.length; r++) { var _context, _context2; var t = null != arguments[r] ? arguments[r] : {}; r % 2 ? _forEachInstanceProperty__default["default"](_context = ownKeys(Object(t), !0)).call(_context, function (r) { _defineProperty(e, r, t[r]); }) : _Object$getOwnPropertyDescriptors__default["default"] ? _Object$defineProperties__default["default"](e, _Object$getOwnPropertyDescriptors__default["default"](t)) : _forEachInstanceProperty__default["default"](_context2 = ownKeys(Object(t))).call(_context2, function (r) { _Object$defineProperty__default["default"](e, r, _Object$getOwnPropertyDescriptor__default["default"](t, r)); }); } return e; }
|
|
852
1023
|
const baseMenuProperties = {
|
|
@@ -967,7 +1138,7 @@ async function command(cliFlags) {
|
|
|
967
1138
|
|
|
968
1139
|
var pkgJson = {
|
|
969
1140
|
name: "@commercetools-frontend/application-cli",
|
|
970
|
-
version: "2.
|
|
1141
|
+
version: "2.5.0",
|
|
971
1142
|
description: "Internal CLI to manage Merchant Center application deployments across various environments.",
|
|
972
1143
|
keywords: [
|
|
973
1144
|
"commercetools",
|
|
@@ -1002,11 +1173,15 @@ var pkgJson = {
|
|
|
1002
1173
|
execa: "5.1.1",
|
|
1003
1174
|
jsonschema: "^1.4.1",
|
|
1004
1175
|
listr2: "8.2.0",
|
|
1176
|
+
lodash: "4.17.21",
|
|
1177
|
+
micromatch: "4.0.5",
|
|
1005
1178
|
"node-fetch": "2.7.0",
|
|
1006
1179
|
"ts-deepmerge": "7.0.0"
|
|
1007
1180
|
},
|
|
1008
1181
|
devDependencies: {
|
|
1009
1182
|
"@tsconfig/node20": "20.1.4",
|
|
1183
|
+
"@types/lodash": "^4.14.198",
|
|
1184
|
+
"@types/micromatch": "4.0.6",
|
|
1010
1185
|
"@types/node": "20.12.7",
|
|
1011
1186
|
typescript: "5.2.2"
|
|
1012
1187
|
},
|
|
@@ -1048,11 +1223,11 @@ const run = async () => {
|
|
|
1048
1223
|
}).option('--ci-assets-root-path [path]', '(optional) A replacement value for the scripts root path only used on CI (e.g. "--ci-assets-root-path=/root/") used in generated scripts.').option('--skip-menu', '(optional) If provided, it will skip uploading the `menu.json`.', {
|
|
1049
1224
|
default: false
|
|
1050
1225
|
}).action(async options => {
|
|
1051
|
-
await command$
|
|
1226
|
+
await command$4(options, cwd);
|
|
1052
1227
|
});
|
|
1053
1228
|
const usageCompileMenu = 'Compile the menu links of an application into a `menu.json`. This is only required for internal applications';
|
|
1054
1229
|
cli.command('compile-menu', usageCompileMenu).usage("compile-menu \n\n ".concat(usageCompileMenu)).option('--dotenv-folder [string]', '(optional) The path to a folder containing a dotenv file `.env.production` and a cloud-environment specific dotenv file (for example `.env.gcp-production-eu`). Those values are parsed and merged together to be used by the application config.').action(async options => {
|
|
1055
|
-
await command$
|
|
1230
|
+
await command$3(options, cwd);
|
|
1056
1231
|
});
|
|
1057
1232
|
const usageValidateMenu = 'Validate compiled `menu.json` file';
|
|
1058
1233
|
cli.command('validate-menu', usageValidateMenu).usage("validate-menu \n\n ".concat(usageValidateMenu)).option('--input-file <path>', '(required) The path to the `menu.json` file to be validated.').option('--navigation [string]', '(optional) Location of the menu navigation. Possible values are `top`.').action(async options => {
|
|
@@ -1062,7 +1237,20 @@ const run = async () => {
|
|
|
1062
1237
|
cli.command('create-version', usageCreateVersion).usage("create-version \n\n ".concat(usageCreateVersion)).option('--version-url <url>', "(required) The path of an application's current `version.json` within the storage bucket.").option('--rollbacks [int]', '(optional) The number of max rollbacks to keep', {
|
|
1063
1238
|
default: 15
|
|
1064
1239
|
}).option('--out-file [path]', '(optional) The path to the file where to write the JSON. If not specified, the JSON is printed to stdout.').action(async options => {
|
|
1065
|
-
await command$
|
|
1240
|
+
await command$2(options);
|
|
1241
|
+
});
|
|
1242
|
+
|
|
1243
|
+
// Command: Evaluate change triggers
|
|
1244
|
+
const usageEvaluateChangeTriggers = 'Evaluates changed files against a base and evaluates them against defined triggers.';
|
|
1245
|
+
cli.command('evaluate-change-triggers', usageEvaluateChangeTriggers).usage("evaluate-change-triggers \n\n ".concat(usageEvaluateChangeTriggers)).option('--branch <string>', 'The branch of the pull request', {
|
|
1246
|
+
default: process.env.CIRCLE_BRANCH
|
|
1247
|
+
}).option('--base-branch <string>', 'The base revision of the git commit compare against (e.g. "main")').option('--head-revision <string>', 'The revision of the git head to compare with', {
|
|
1248
|
+
default: process.env.CIRCLE_SHA1
|
|
1249
|
+
}).option('--trigger-name <string>', 'The trigger to evaluate for.').option('--silent', '(optional) Disable logging', {
|
|
1250
|
+
default: false
|
|
1251
|
+
}).option('--out-env-file [string]', '(optional) A file path where the matched triggers are written as a dotenv file.').action(async options => {
|
|
1252
|
+
const config = await storageBucketsConfig.loadConfig('circleci-change-triggers', []);
|
|
1253
|
+
await command$1(options, config, cwd);
|
|
1066
1254
|
});
|
|
1067
1255
|
cli.help();
|
|
1068
1256
|
cli.version(pkgJson.version);
|