@webiny/api-page-builder-import-export 0.0.0-mt-3 → 0.0.0-unstable.1e66d121db

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/{importPages/client.d.ts → client.d.ts} +3 -2
  2. package/client.js +47 -0
  3. package/client.js.map +1 -0
  4. package/exportPages/combine/index.d.ts +7 -11
  5. package/exportPages/combine/index.js +37 -11
  6. package/exportPages/combine/index.js.map +1 -0
  7. package/exportPages/process/index.d.ts +10 -14
  8. package/exportPages/process/index.js +32 -28
  9. package/exportPages/process/index.js.map +1 -0
  10. package/exportPages/s3Stream.d.ts +2 -0
  11. package/exportPages/s3Stream.js +14 -7
  12. package/exportPages/s3Stream.js.map +1 -0
  13. package/exportPages/utils.d.ts +1 -1
  14. package/exportPages/utils.js +2 -2
  15. package/exportPages/utils.js.map +1 -0
  16. package/exportPages/zipper.d.ts +1 -0
  17. package/exportPages/zipper.js +3 -5
  18. package/exportPages/zipper.js.map +1 -0
  19. package/graphql/crud/pageImportExportTasks.crud.d.ts +1 -1
  20. package/graphql/crud/pageImportExportTasks.crud.js +47 -30
  21. package/graphql/crud/pageImportExportTasks.crud.js.map +1 -0
  22. package/graphql/crud/pages.crud.d.ts +1 -1
  23. package/graphql/crud/pages.crud.js +29 -21
  24. package/graphql/crud/pages.crud.js.map +1 -0
  25. package/graphql/crud.d.ts +1 -1
  26. package/graphql/crud.js +1 -1
  27. package/graphql/crud.js.map +1 -0
  28. package/graphql/graphql/pageImportExportTasks.gql.js +1 -1
  29. package/graphql/graphql/pageImportExportTasks.gql.js.map +1 -0
  30. package/graphql/graphql/pages.gql.js +8 -6
  31. package/graphql/graphql/pages.gql.js.map +1 -0
  32. package/graphql/graphql/utils/resolve.d.ts +1 -1
  33. package/graphql/graphql/utils/resolve.js.map +1 -0
  34. package/graphql/graphql.d.ts +1 -1
  35. package/graphql/graphql.js +1 -1
  36. package/graphql/graphql.js.map +1 -0
  37. package/graphql/index.d.ts +1 -1
  38. package/graphql/index.js +1 -1
  39. package/graphql/index.js.map +1 -0
  40. package/graphql/types.d.ts +5 -6
  41. package/graphql/types.js.map +1 -0
  42. package/importPages/create/index.d.ts +11 -16
  43. package/importPages/create/index.js +39 -30
  44. package/importPages/create/index.js.map +1 -0
  45. package/importPages/process/index.d.ts +10 -17
  46. package/importPages/process/index.js +25 -23
  47. package/importPages/process/index.js.map +1 -0
  48. package/importPages/utils.d.ts +17 -10
  49. package/importPages/utils.js +90 -59
  50. package/importPages/utils.js.map +1 -0
  51. package/mockSecurity.js.map +1 -0
  52. package/package.json +36 -34
  53. package/types.d.ts +8 -7
  54. package/types.js.map +1 -0
  55. package/importPages/client.js +0 -40
@@ -1,6 +1,6 @@
1
1
  "use strict";
2
2
 
3
- var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
3
+ var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault").default;
4
4
 
5
5
  Object.defineProperty(exports, "__esModule", {
6
6
  value: true
@@ -8,7 +8,7 @@ Object.defineProperty(exports, "__esModule", {
8
8
  exports.importPage = importPage;
9
9
  exports.initialStats = initialStats;
10
10
  exports.readExtractAndUploadZipFileContents = readExtractAndUploadZipFileContents;
11
- exports.zeroPad = exports.uploadPageAssets = void 0;
11
+ exports.uploadPageAssets = void 0;
12
12
 
13
13
  var _uniqid = _interopRequireDefault(require("uniqid"));
14
14
 
@@ -46,25 +46,29 @@ const INSTALL_DIR = "/tmp";
46
46
  const INSTALL_EXTRACT_DIR = _path.default.join(INSTALL_DIR, "apiPageBuilderImportPage");
47
47
 
48
48
  const FILES_COUNT_IN_EACH_BATCH = 15;
49
- const ZIP_CONTENT_TYPE = "application/zip";
50
49
 
51
- function updateImageInPageSettings({
52
- settings,
53
- fileIdToKeyMap,
54
- srcPrefix
55
- }) {
50
+ function updateImageInPageSettings(params) {
51
+ const {
52
+ settings,
53
+ fileIdToKeyMap,
54
+ srcPrefix
55
+ } = params;
56
56
  let newSettings = settings;
57
57
  const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith("/") ? srcPrefix.slice(0, -1) : srcPrefix;
58
58
 
59
59
  if (_dotPropImmutable.default.get(newSettings, "general.image.src")) {
60
- newSettings = _dotPropImmutable.default.set(newSettings, "general.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(settings.general.image.id)}`);
60
+ var _settings$general, _settings$general$ima;
61
+
62
+ newSettings = _dotPropImmutable.default.set(newSettings, "general.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(((_settings$general = settings.general) === null || _settings$general === void 0 ? void 0 : (_settings$general$ima = _settings$general.image) === null || _settings$general$ima === void 0 ? void 0 : _settings$general$ima.id) || "")}`);
61
63
  }
62
64
 
63
65
  if (_dotPropImmutable.default.get(newSettings, "social.image.src")) {
64
- newSettings = _dotPropImmutable.default.set(newSettings, "social.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(settings.social.image.id)}`);
66
+ var _settings$social, _settings$social$imag;
67
+
68
+ newSettings = _dotPropImmutable.default.set(newSettings, "social.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(((_settings$social = settings.social) === null || _settings$social === void 0 ? void 0 : (_settings$social$imag = _settings$social.image) === null || _settings$social$imag === void 0 ? void 0 : _settings$social$imag.id) || "")}`);
65
69
  }
66
70
 
67
- return settings;
71
+ return newSettings;
68
72
  }
69
73
 
70
74
  function updateFilesInPageData({
@@ -111,23 +115,27 @@ function updateFilesInPageData({
111
115
  }
112
116
  }
113
117
 
114
- const uploadPageAssets = async ({
115
- context,
116
- filesData,
117
- fileUploadsData
118
- }) => {
118
+ const uploadPageAssets = async params => {
119
+ const {
120
+ context,
121
+ filesData,
122
+ fileUploadsData
123
+ } = params; // Save uploaded file key against static id for later use.
124
+
125
+ const fileIdToKeyMap = new Map();
119
126
  /**
120
127
  * This function contains logic of file download from S3.
121
128
  * Current we're not mocking zip file download from S3 in tests at the moment.
122
129
  * So, we're manually mocking it in case of test just by returning an empty object.
123
130
  */
131
+
124
132
  if (process.env.NODE_ENV === "test") {
125
- return {};
133
+ return {
134
+ fileIdToKeyMap
135
+ };
126
136
  }
127
137
 
128
- console.log("INSIDE uploadPageAssets"); // Save uploaded file key against static id for later use.
129
-
130
- const fileIdToKeyMap = new Map(); // Save files meta data against old key for later use.
138
+ console.log("INSIDE uploadPageAssets"); // Save files meta data against old key for later use.
131
139
 
132
140
  const fileKeyToFileMap = new Map(); // Initialize maps.
133
141
 
@@ -145,7 +153,12 @@ const uploadPageAssets = async ({
145
153
 
146
154
  const createFilesInput = fileUploadResults.map(uploadResult => {
147
155
  const newKey = uploadResult.Key;
148
- const file = fileKeyToFileMap.get(getOldFileKey(newKey)); // Update the file map with newly uploaded file.
156
+ const file = fileKeyToFileMap.get(getOldFileKey(newKey));
157
+
158
+ if (!file) {
159
+ return null;
160
+ } // Update the file map with newly uploaded file.
161
+
149
162
 
150
163
  fileIdToKeyMap.set(file.id, newKey);
151
164
  return {
@@ -156,7 +169,7 @@ const uploadPageAssets = async ({
156
169
  meta: file.meta,
157
170
  tags: file.tags
158
171
  };
159
- });
172
+ }).filter(Boolean);
160
173
  const createFilesPromises = []; // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).
161
174
 
162
175
  const createFilesInputChunks = (0, _chunk.default)(createFilesInput, FILES_COUNT_IN_EACH_BATCH);
@@ -206,25 +219,31 @@ async function importPage({
206
219
  files
207
220
  } = await (0, _loadJsonFile.default)(PAGE_DATA_FILE_PATH); // Only update page data if there are files.
208
221
 
209
- if (Array.isArray(files) && files.length) {
222
+ if (files && Array.isArray(files) && files.length > 0) {
210
223
  // Upload page assets.
211
224
  const {
212
225
  fileIdToKeyMap
213
226
  } = await uploadPageAssets({
214
227
  context,
228
+
229
+ /**
230
+ * TODO @ts-refactor @ashutosh figure out correct types.
231
+ */
232
+ // @ts-ignore
215
233
  filesData: files,
216
234
  fileUploadsData
217
235
  });
236
+ const settings = await context.fileManager.settings.getSettings();
218
237
  const {
219
- srcPrefix
220
- } = await context.fileManager.settings.getSettings();
238
+ srcPrefix = ""
239
+ } = settings || {};
221
240
  updateFilesInPageData({
222
- data: page.content,
241
+ data: page.content || {},
223
242
  fileIdToKeyMap,
224
243
  srcPrefix
225
244
  });
226
245
  page.settings = updateImageInPageSettings({
227
- settings: page.settings,
246
+ settings: page.settings || {},
228
247
  fileIdToKeyMap,
229
248
  srcPrefix
230
249
  });
@@ -270,14 +289,6 @@ async function uploadFilesFromS3({
270
289
  return Promise.all(promises);
271
290
  }
272
291
 
273
- async function getObjectMetaFromS3(Key) {
274
- const meta = await _s3Stream.s3Stream.getObjectHead(Key);
275
-
276
- if (meta.ContentType !== ZIP_CONTENT_TYPE) {
277
- throw new _error.default(`Unsupported file type: "${meta.ContentType}"`, "UNSUPPORTED_FILE");
278
- }
279
- }
280
-
281
292
  function getOldFileKey(key) {
282
293
  /*
283
294
  * Because we know the naming convention, we can extract the old key from new key.
@@ -298,32 +309,23 @@ function getFileNameWithoutExt(fileName) {
298
309
 
299
310
  /**
300
311
  * Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.
301
- * @param zipFileKey
312
+ * @param zipFileUrl
302
313
  * @return PageImportData S3 file keys for all uploaded assets group by page.
303
314
  */
304
- async function readExtractAndUploadZipFileContents(zipFileKey) {
315
+ async function readExtractAndUploadZipFileContents(zipFileUrl) {
305
316
  const log = console.log;
306
317
  const pageImportDataList = [];
307
- let readStream; // Check whether it is a URL
308
318
 
309
- if (zipFileKey.startsWith("http")) {
310
- const response = await (0, _nodeFetch.default)(zipFileKey);
319
+ const zipFileName = _path.default.basename(zipFileUrl).split("?")[0];
311
320
 
312
- if (!response.ok) {
313
- throw new _error.default(`Unable to downloading file: "${zipFileKey}"`, response.statusText);
314
- }
321
+ const response = await (0, _nodeFetch.default)(zipFileUrl);
315
322
 
316
- readStream = response.body;
317
- } else {
318
- // We're first retrieving object's meta data, just to check whether the file is available at the given Key
319
- await getObjectMetaFromS3(zipFileKey);
320
- readStream = _s3Stream.s3Stream.readStream(zipFileKey);
323
+ if (!response.ok) {
324
+ throw new _error.default(`Unable to downloading file: "${zipFileUrl}"`, response.statusText);
321
325
  }
322
326
 
323
- const uniquePath = (0, _uniqid.default)("IMPORT_PAGES/");
324
-
325
- const zipFileName = _path.default.basename(zipFileKey); // Read export file and download it in the disk
326
-
327
+ const readStream = response.body;
328
+ const uniquePath = (0, _uniqid.default)("IMPORT_PAGES/"); // Read export file and download it in the disk
327
329
 
328
330
  const ZIP_FILE_PATH = _path.default.join(INSTALL_DIR, zipFileName);
329
331
 
@@ -332,7 +334,7 @@ async function readExtractAndUploadZipFileContents(zipFileKey) {
332
334
  log(`Downloaded file "${zipFileName}" at ${ZIP_FILE_PATH}`); // Extract the downloaded zip file
333
335
 
334
336
  const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);
335
- log(`Removing ZIP file "${zipFileKey}" from ${ZIP_FILE_PATH}`);
337
+ log(`Removing ZIP file "${zipFileUrl}" from ${ZIP_FILE_PATH}`);
336
338
  await (0, _downloadInstallFiles.deleteFile)(ZIP_FILE_PATH); // Extract each page zip and upload their content's to S3
337
339
 
338
340
  for (let i = 0; i < zipFilePaths.length; i++) {
@@ -382,16 +384,13 @@ async function deleteS3Folder(key) {
382
384
  }
383
385
 
384
386
  const response = await _s3Stream.s3Stream.listObject(key);
385
- const keys = response.Contents.map(c => c.Key);
387
+ const keys = (response.Contents || []).map(c => c.Key).filter(Boolean);
386
388
  console.log(`Found ${keys.length} files.`);
387
389
  const deleteFilePromises = keys.map(key => _s3Stream.s3Stream.deleteObject(key));
388
390
  await Promise.all(deleteFilePromises);
389
391
  console.log(`Successfully deleted ${deleteFilePromises.length} files.`);
390
- }
391
-
392
- const zeroPad = version => `${version}`.padStart(5, "0");
392
+ } // export const zeroPad = version => `${version}`.padStart(5, "0");
393
393
 
394
- exports.zeroPad = zeroPad;
395
394
 
396
395
  function initialStats(total) {
397
396
  return {
@@ -419,6 +418,13 @@ function extractZipToDisk(exportFileZipPath) {
419
418
  if (err) {
420
419
  console.warn("ERROR: Failed to extract zip: ", exportFileZipPath, err);
421
420
  reject(err);
421
+ return;
422
+ }
423
+
424
+ if (!zipFile) {
425
+ console.log("ERROR: Missing zip file resource for path: " + exportFileZipPath);
426
+ reject("Missing Zip File Resource.");
427
+ return;
422
428
  }
423
429
 
424
430
  console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);
@@ -445,6 +451,13 @@ function extractZipToDisk(exportFileZipPath) {
445
451
  if (err) {
446
452
  console.warn("ERROR: Failed to openReadStream for file: ", entry.fileName, err);
447
453
  reject(err);
454
+ return;
455
+ }
456
+
457
+ if (!readStream) {
458
+ console.log("ERROR: Missing Read Stream Resource when extracting to disk.");
459
+ reject("Missing Read Stream Resource.");
460
+ return;
448
461
  }
449
462
 
450
463
  const filePath = _path.default.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);
@@ -453,7 +466,9 @@ function extractZipToDisk(exportFileZipPath) {
453
466
  pageZipFilePaths.push(filePath);
454
467
  zipFile.readEntry();
455
468
  });
456
- streamPipeline(readStream, (0, _fs.createWriteStream)(filePath));
469
+ streamPipeline(readStream, (0, _fs.createWriteStream)(filePath)).catch(error => {
470
+ reject(error);
471
+ });
457
472
  });
458
473
  }
459
474
  });
@@ -478,6 +493,13 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
478
493
  if (err) {
479
494
  console.warn("ERROR: Failed to extract zip: ", pageDataZipFilePath, err);
480
495
  reject(err);
496
+ return;
497
+ }
498
+
499
+ if (!zipFile) {
500
+ console.log("ERROR: Probably failed to extract zip: " + pageDataZipFilePath);
501
+ reject("Missing Zip File Resource.");
502
+ return;
481
503
  }
482
504
 
483
505
  console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);
@@ -509,6 +531,13 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
509
531
  if (err) {
510
532
  console.warn("ERROR: Failed while performing [openReadStream] for file: ", entry.fileName, err);
511
533
  reject(err);
534
+ return;
535
+ }
536
+
537
+ if (!readStream) {
538
+ console.log("ERROR: Missing Read Stream while importing pages.");
539
+ reject("Missing Read Strea Resource.");
540
+ return;
512
541
  }
513
542
 
514
543
  readStream.on("end", function () {
@@ -530,6 +559,8 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
530
559
 
531
560
  streamPipeline(readStream, streamPassThrough).then(() => {
532
561
  fileUploadPromises.push(promise);
562
+ }).catch(error => {
563
+ reject(error);
533
564
  });
534
565
  });
535
566
  }
@@ -0,0 +1 @@
1
+ {"version":3,"names":["streamPipeline","promisify","pipeline","INSTALL_DIR","INSTALL_EXTRACT_DIR","path","join","FILES_COUNT_IN_EACH_BATCH","updateImageInPageSettings","params","settings","fileIdToKeyMap","srcPrefix","newSettings","srcPrefixWithoutTrailingSlash","endsWith","slice","dotProp","get","set","general","image","id","social","updateFilesInPageData","data","Array","isArray","i","length","element","tuple","Object","entries","key","value","has","name","src","uploadPageAssets","context","filesData","fileUploadsData","Map","process","env","NODE_ENV","console","log","fileKeyToFileMap","file","type","fileUploadResults","uploadFilesFromS3","oldKeyToNewKeyMap","assets","createFilesInput","map","uploadResult","newKey","Key","getOldFileKey","size","meta","tags","filter","Boolean","createFilesPromises","createFilesInputChunks","chunk","createFilesInputChunk","push","fileManager","files","createFilesInBatch","Promise","all","importPage","pageKey","PAGE_EXTRACT_DIR","ensureDirSync","pageDataFileKey","PAGE_DATA_FILE_PATH","basename","resolve","reject","s3Stream","readStream","on","pipe","createWriteStream","page","loadJson","getSettings","content","deleteFile","deleteS3Folder","dirname","oldKeysForAssets","keys","promises","oldKey","tempNewKey","fileMetaData","uniqueId","streamPassThrough","streamPassThroughUploadPromise","promise","writeStream","rest","split","e","FILE_CONTENT_TYPE","getFileNameWithoutExt","fileName","replace","extname","readExtractAndUploadZipFileContents","zipFileUrl","pageImportDataList","zipFileName","response","fetch","ok","WebinyError","statusText","body","uniquePath","ZIP_FILE_PATH","zipFilePaths","extractZipToDisk","currentPath","dataMap","extractZipAndUploadToS3","ASSETS_DIR_NAME","preparePageDataDirMap","filePath","isAsset","listObject","Contents","c","deleteFilePromises","deleteObject","initialStats","total","PageImportExportTaskStatus","PENDING","PROCESSING","COMPLETED","FAILED","exportFileZipPath","pageZipFilePaths","uniqueFolderNameForExport","EXPORT_FILE_EXTRACTION_PATH","yauzl","open","lazyEntries","err","zipFile","warn","info","entryCount","readEntry","entry","test","openReadStream","catch","error","pageDataZipFilePath","filePaths","fileUploadPromises","uniquePageKey","then","res","forEach","r"],"sources":["utils.ts"],"sourcesContent":["import uniqueId from \"uniqid\";\nimport S3 from \"aws-sdk/clients/s3\";\nimport dotProp from \"dot-prop-immutable\";\nimport { createWriteStream } from \"fs\";\nimport { ensureDirSync } from \"fs-extra\";\nimport { promisify } from \"util\";\nimport { pipeline } from \"stream\";\nimport fetch from \"node-fetch\";\nimport path from \"path\";\nimport yauzl from \"yauzl\";\nimport chunk from \"lodash/chunk\";\nimport loadJson from \"load-json-file\";\nimport { FileInput } from \"@webiny/api-file-manager/types\";\nimport WebinyError from \"@webiny/error\";\nimport { deleteFile } from \"@webiny/api-page-builder/graphql/crud/install/utils/downloadInstallFiles\";\nimport { File, PageImportExportTaskStatus } from \"~/types\";\nimport { PbPageImportExportContext } from \"~/graphql/types\";\nimport { s3Stream } from \"~/exportPages/s3Stream\";\nimport { ExportedPageData } from \"~/exportPages/utils\";\nimport { PageSettings } from \"@webiny/api-page-builder/types\";\n\ninterface FileItem extends File {\n key: string;\n type: string;\n name: string;\n size: number;\n meta: Record<string, any>;\n tags: string[];\n}\n\nconst streamPipeline = promisify(pipeline);\n\nconst INSTALL_DIR = \"/tmp\";\nconst INSTALL_EXTRACT_DIR = path.join(INSTALL_DIR, \"apiPageBuilderImportPage\");\nconst FILES_COUNT_IN_EACH_BATCH = 15;\n\ninterface UpdateFilesInPageDataParams {\n data: Record<string, any>;\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n}\n\ninterface UpdateImageInPageSettingsParams {\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n settings: PageSettings;\n}\n\nfunction updateImageInPageSettings(\n params: UpdateImageInPageSettingsParams\n): UpdateImageInPageSettingsParams[\"settings\"] {\n const { settings, fileIdToKeyMap, srcPrefix } = params;\n let newSettings = settings;\n\n const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith(\"/\")\n ? srcPrefix.slice(0, -1)\n : srcPrefix;\n\n if (dotProp.get(newSettings, \"general.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"general.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.general?.image?.id || \"\"\n )}`\n );\n }\n if (dotProp.get(newSettings, \"social.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"social.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.social?.image?.id || \"\"\n )}`\n );\n }\n\n return newSettings;\n}\n\nfunction updateFilesInPageData({ data, fileIdToKeyMap, srcPrefix }: UpdateFilesInPageDataParams) {\n // BASE CASE: Termination point\n if (!data || typeof data !== \"object\") {\n return;\n }\n // Recursively call function if data is array\n if (Array.isArray(data)) {\n for (let i = 0; i < data.length; i++) {\n const element = data[i];\n updateFilesInPageData({ data: element, fileIdToKeyMap, srcPrefix });\n }\n return;\n }\n // Main logic\n const tuple = Object.entries(data);\n for (let i = 0; i < tuple.length; i++) {\n const [key, value] = tuple[i];\n\n if (key === \"file\" && value && fileIdToKeyMap.has(value.id)) {\n value.key = fileIdToKeyMap.get(value.id);\n value.name = fileIdToKeyMap.get(value.id);\n value.src = `${srcPrefix}${srcPrefix.endsWith(\"/\") ? \"\" : \"/\"}${fileIdToKeyMap.get(\n value.id\n )}`;\n } else {\n updateFilesInPageData({ data: value, srcPrefix, fileIdToKeyMap });\n }\n }\n}\n\ninterface UploadPageAssetsParams {\n context: PbPageImportExportContext;\n filesData: FileItem[];\n fileUploadsData: FileUploadsData;\n}\n\ninterface UploadPageAssetsReturnType {\n fileIdToKeyMap: Map<string, string>;\n}\n\nexport const uploadPageAssets = async (\n params: UploadPageAssetsParams\n): Promise<UploadPageAssetsReturnType> => {\n const { context, filesData, fileUploadsData } = params;\n // Save uploaded file key against static id for later use.\n const fileIdToKeyMap = new Map<string, string>();\n /**\n * This function contains logic of file download from S3.\n * Current we're not mocking zip file download from S3 in tests at the moment.\n * So, we're manually mocking it in case of test just by returning an empty object.\n */\n if (process.env.NODE_ENV === \"test\") {\n return {\n fileIdToKeyMap\n };\n }\n console.log(\"INSIDE uploadPageAssets\");\n\n // Save files meta data against old key for later use.\n const fileKeyToFileMap = new Map<string, FileItem>();\n // Initialize maps.\n for (let i = 0; i < filesData.length; i++) {\n const file = filesData[i];\n fileKeyToFileMap.set(file.key, file);\n\n // Initialize the value\n fileIdToKeyMap.set(file.id, file.type);\n }\n\n const fileUploadResults = await uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap: fileUploadsData.assets\n });\n\n // Create files in File Manager\n const createFilesInput = fileUploadResults\n .map((uploadResult): FileInput | null => {\n const newKey = uploadResult.Key;\n const file = fileKeyToFileMap.get(getOldFileKey(newKey));\n if (!file) {\n return null;\n }\n\n // Update the file map with newly uploaded file.\n fileIdToKeyMap.set(file.id, newKey);\n\n return {\n key: newKey,\n name: file.name,\n size: file.size,\n type: file.type,\n meta: file.meta,\n tags: file.tags\n };\n })\n .filter(Boolean) as FileInput[];\n\n const createFilesPromises = [];\n // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).\n const createFilesInputChunks = chunk(createFilesInput, FILES_COUNT_IN_EACH_BATCH);\n for (let i = 0; i < createFilesInputChunks.length; i++) {\n const createFilesInputChunk = createFilesInputChunks[i];\n createFilesPromises.push(\n /*\n * We need to break down files into chunks because\n * `createFilesInBatch` operation has a limit on number of files it can handle at once.\n */\n context.fileManager.files.createFilesInBatch(createFilesInputChunk)\n );\n }\n\n await Promise.all(createFilesPromises);\n\n return {\n fileIdToKeyMap\n };\n};\n\ninterface FileUploadsData {\n data: string;\n assets: Record<string, string>;\n}\n\ninterface ImportPageParams {\n key: string;\n pageKey: string;\n context: PbPageImportExportContext;\n fileUploadsData: FileUploadsData;\n}\n\nexport async function importPage({\n pageKey,\n context,\n fileUploadsData\n}: ImportPageParams): Promise<ExportedPageData[\"page\"]> {\n const log = console.log;\n\n // Making Directory for page in which we're going to extract the page data file.\n const PAGE_EXTRACT_DIR = path.join(INSTALL_EXTRACT_DIR, pageKey);\n ensureDirSync(PAGE_EXTRACT_DIR);\n\n const pageDataFileKey = dotProp.get(fileUploadsData, `data`);\n const PAGE_DATA_FILE_PATH = path.join(PAGE_EXTRACT_DIR, path.basename(pageDataFileKey));\n\n log(`Downloading Page data file: ${pageDataFileKey} at \"${PAGE_DATA_FILE_PATH}\"`);\n // Download and save page data file in disk.\n await new Promise((resolve, reject) => {\n s3Stream\n .readStream(pageDataFileKey)\n .on(\"error\", reject)\n .pipe(createWriteStream(PAGE_DATA_FILE_PATH))\n .on(\"error\", reject)\n .on(\"finish\", resolve);\n });\n\n // Load the page data file from disk.\n log(`Load file ${pageDataFileKey}`);\n const { page, files } = await loadJson<ExportedPageData>(PAGE_DATA_FILE_PATH);\n\n // Only update page data if there are files.\n if (files && Array.isArray(files) && files.length > 0) {\n // Upload page assets.\n const { fileIdToKeyMap } = await uploadPageAssets({\n context,\n /**\n * TODO @ts-refactor @ashutosh figure out correct types.\n */\n // @ts-ignore\n filesData: files,\n fileUploadsData\n });\n\n const settings = await context.fileManager.settings.getSettings();\n\n const { srcPrefix = \"\" } = settings || {};\n updateFilesInPageData({\n data: page.content || {},\n fileIdToKeyMap,\n srcPrefix\n });\n\n page.settings = updateImageInPageSettings({\n settings: page.settings || {},\n fileIdToKeyMap,\n srcPrefix\n });\n }\n\n log(\"Removing Directory for page...\");\n await deleteFile(pageKey);\n\n log(`Remove page contents from S3...`);\n await deleteS3Folder(path.dirname(fileUploadsData.data));\n\n return page;\n}\n\ninterface UploadFilesFromZipParams {\n fileKeyToFileMap: Map<string, any>;\n oldKeyToNewKeyMap: Record<string, string>;\n}\n\nasync function uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap\n}: UploadFilesFromZipParams): Promise<S3.ManagedUpload.SendData[]> {\n const oldKeysForAssets = Object.keys(oldKeyToNewKeyMap);\n\n const promises = [];\n // Upload all assets.\n for (let i = 0; i < oldKeysForAssets.length; i++) {\n const oldKey = oldKeysForAssets[i];\n const tempNewKey = oldKeyToNewKeyMap[oldKey];\n\n // Read file.\n const readStream = s3Stream.readStream(tempNewKey);\n // Get file meta data.\n const fileMetaData = fileKeyToFileMap.get(oldKey);\n\n if (fileMetaData) {\n const newKey = uniqueId(\"\", `-${fileMetaData.key}`);\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, fileMetaData.type);\n readStream.pipe(streamPassThrough);\n promises.push(promise);\n\n console.log(`Successfully queued file \"${newKey}\"`);\n }\n }\n\n return Promise.all(promises);\n}\n\nfunction getOldFileKey(key: string) {\n /*\n * Because we know the naming convention, we can extract the old key from new key.\n */\n try {\n const [, ...rest] = key.split(\"-\");\n return rest.join(\"-\");\n } catch (e) {\n return key;\n }\n}\n\nconst FILE_CONTENT_TYPE = \"application/octet-stream\";\n\nfunction getFileNameWithoutExt(fileName: string): string {\n return path.basename(fileName).replace(path.extname(fileName), \"\");\n}\n\ninterface PageImportData {\n assets: Record<string, string>;\n data: string;\n key: string;\n}\n\n/**\n * Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.\n * @param zipFileUrl\n * @return PageImportData S3 file keys for all uploaded assets group by page.\n */\nexport async function readExtractAndUploadZipFileContents(\n zipFileUrl: string\n): Promise<PageImportData[]> {\n const log = console.log;\n const pageImportDataList = [];\n\n const zipFileName = path.basename(zipFileUrl).split(\"?\")[0];\n\n const response = await fetch(zipFileUrl);\n if (!response.ok) {\n throw new WebinyError(`Unable to downloading file: \"${zipFileUrl}\"`, response.statusText);\n }\n\n const readStream = response.body;\n\n const uniquePath = uniqueId(\"IMPORT_PAGES/\");\n // Read export file and download it in the disk\n const ZIP_FILE_PATH = path.join(INSTALL_DIR, zipFileName);\n\n const writeStream = createWriteStream(ZIP_FILE_PATH);\n await streamPipeline(readStream, writeStream);\n log(`Downloaded file \"${zipFileName}\" at ${ZIP_FILE_PATH}`);\n\n // Extract the downloaded zip file\n const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);\n\n log(`Removing ZIP file \"${zipFileUrl}\" from ${ZIP_FILE_PATH}`);\n await deleteFile(ZIP_FILE_PATH);\n\n // Extract each page zip and upload their content's to S3\n for (let i = 0; i < zipFilePaths.length; i++) {\n const currentPath = zipFilePaths[i];\n const dataMap = await extractZipAndUploadToS3(currentPath, uniquePath);\n pageImportDataList.push(dataMap);\n }\n log(\"Removing all ZIP files located at \", path.dirname(zipFilePaths[0]));\n await deleteFile(path.dirname(zipFilePaths[0]));\n\n return pageImportDataList;\n}\n\nconst ASSETS_DIR_NAME = \"/assets\";\n\nfunction preparePageDataDirMap({\n map,\n filePath,\n newKey\n}: {\n map: PageImportData;\n filePath: string;\n newKey: string;\n}): PageImportData {\n const dirname = path.dirname(filePath);\n const fileName = path.basename(filePath);\n /*\n * We want to use dot (.) as part of object key rather than creating nested object(s).\n * Also, the file name might contain dots in it beside the extension, so, we are escaping them all.\n */\n const oldKey = fileName.replace(/\\./g, \"\\\\.\");\n\n const isAsset = dirname.endsWith(ASSETS_DIR_NAME);\n\n if (isAsset) {\n map = dotProp.set(map, `assets.${oldKey}`, newKey);\n } else {\n // We only need to know the newKey for data file.\n map = dotProp.set(map, `data`, newKey);\n }\n\n return map;\n}\n\nasync function deleteS3Folder(key: string): Promise<void> {\n // Append trailing slash i.e \"/\" to key to make sure we only delete a specific folder.\n if (!key.endsWith(\"/\")) {\n key = `${key}/`;\n }\n\n const response = await s3Stream.listObject(key);\n const keys = (response.Contents || []).map(c => c.Key).filter(Boolean) as string[];\n console.log(`Found ${keys.length} files.`);\n\n const deleteFilePromises = keys.map(key => s3Stream.deleteObject(key));\n\n await Promise.all(deleteFilePromises);\n console.log(`Successfully deleted ${deleteFilePromises.length} files.`);\n}\n\n// export const zeroPad = version => `${version}`.padStart(5, \"0\");\n\nexport function initialStats(total: number) {\n return {\n [PageImportExportTaskStatus.PENDING]: total,\n [PageImportExportTaskStatus.PROCESSING]: 0,\n [PageImportExportTaskStatus.COMPLETED]: 0,\n [PageImportExportTaskStatus.FAILED]: 0,\n total\n };\n}\n\nfunction extractZipToDisk(exportFileZipPath: string): Promise<string[]> {\n return new Promise((resolve, reject) => {\n const pageZipFilePaths: string[] = [];\n const uniqueFolderNameForExport = getFileNameWithoutExt(exportFileZipPath);\n const EXPORT_FILE_EXTRACTION_PATH = path.join(INSTALL_DIR, uniqueFolderNameForExport);\n // Make sure DIR exists\n ensureDirSync(EXPORT_FILE_EXTRACTION_PATH);\n\n yauzl.open(exportFileZipPath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", exportFileZipPath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Missing zip file resource for path: \" + exportFileZipPath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn(\"ERROR: Failed on END event for file: \", exportFileZipPath, err);\n reject(err);\n }\n resolve(pageZipFilePaths);\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed to openReadStream for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\n \"ERROR: Missing Read Stream Resource when extracting to disk.\"\n );\n reject(\"Missing Read Stream Resource.\");\n return;\n }\n\n const filePath = path.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);\n\n readStream.on(\"end\", function () {\n pageZipFilePaths.push(filePath);\n zipFile.readEntry();\n });\n\n streamPipeline(readStream, createWriteStream(filePath)).catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n\nfunction extractZipAndUploadToS3(\n pageDataZipFilePath: string,\n uniquePath: string\n): Promise<PageImportData> {\n return new Promise((resolve, reject) => {\n const filePaths = [];\n const fileUploadPromises: Promise<S3.ManagedUpload.SendData>[] = [];\n const uniquePageKey = getFileNameWithoutExt(pageDataZipFilePath);\n let dataMap: PageImportData = {\n key: uniquePageKey,\n assets: {},\n data: \"\"\n };\n yauzl.open(pageDataZipFilePath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", pageDataZipFilePath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Probably failed to extract zip: \" + pageDataZipFilePath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn('ERROR: Failed on \"END\" for file: ', pageDataZipFilePath, err);\n reject(err);\n }\n\n Promise.all(fileUploadPromises).then(res => {\n res.forEach(r => {\n console.info(\"Done uploading... \", r);\n });\n resolve(dataMap);\n });\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed while performing [openReadStream] for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\"ERROR: Missing Read Stream while importing pages.\");\n reject(\"Missing Read Strea Resource.\");\n return;\n }\n readStream.on(\"end\", function () {\n filePaths.push(entry.fileName);\n zipFile.readEntry();\n });\n\n const newKey = `${uniquePath}/${uniquePageKey}/${entry.fileName}`;\n // Modify in place\n dataMap = preparePageDataDirMap({\n map: dataMap,\n filePath: entry.fileName,\n newKey\n });\n\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, FILE_CONTENT_TYPE);\n\n streamPipeline(readStream, streamPassThrough)\n .then(() => {\n fileUploadPromises.push(promise);\n })\n .catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n"],"mappings":";;;;;;;;;;;;AAAA;;AAEA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AAEA;;AACA;;AACA;;AAEA;;AAaA,MAAMA,cAAc,GAAG,IAAAC,eAAA,EAAUC,gBAAV,CAAvB;AAEA,MAAMC,WAAW,GAAG,MAApB;;AACA,MAAMC,mBAAmB,GAAGC,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuB,0BAAvB,CAA5B;;AACA,MAAMI,yBAAyB,GAAG,EAAlC;;AAcA,SAASC,yBAAT,CACIC,MADJ,EAE+C;EAC3C,MAAM;IAAEC,QAAF;IAAYC,cAAZ;IAA4BC;EAA5B,IAA0CH,MAAhD;EACA,IAAII,WAAW,GAAGH,QAAlB;EAEA,MAAMI,6BAA6B,GAAGF,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAChCH,SAAS,CAACI,KAAV,CAAgB,CAAhB,EAAmB,CAAC,CAApB,CADgC,GAEhCJ,SAFN;;EAIA,IAAIK,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,mBAAzB,CAAJ,EAAmD;IAAA;;IAC/CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,mBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,sBAAAR,QAAQ,CAACU,OAAT,iGAAkBC,KAAlB,gFAAyBC,EAAzB,KAA+B,EADC,CAElC,EALQ,CAAd;EAOH;;EACD,IAAIL,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,kBAAzB,CAAJ,EAAkD;IAAA;;IAC9CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,kBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,qBAAAR,QAAQ,CAACa,MAAT,+FAAiBF,KAAjB,gFAAwBC,EAAxB,KAA8B,EADE,CAElC,EALQ,CAAd;EAOH;;EAED,OAAOT,WAAP;AACH;;AAED,SAASW,qBAAT,CAA+B;EAAEC,IAAF;EAAQd,cAAR;EAAwBC;AAAxB,CAA/B,EAAiG;EAC7F;EACA,IAAI,CAACa,IAAD,IAAS,OAAOA,IAAP,KAAgB,QAA7B,EAAuC;IACnC;EACH,CAJ4F,CAK7F;;;EACA,IAAIC,KAAK,CAACC,OAAN,CAAcF,IAAd,CAAJ,EAAyB;IACrB,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGH,IAAI,CAACI,MAAzB,EAAiCD,CAAC,EAAlC,EAAsC;MAClC,MAAME,OAAO,GAAGL,IAAI,CAACG,CAAD,CAApB;MACAJ,qBAAqB,CAAC;QAAEC,IAAI,EAAEK,OAAR;QAAiBnB,cAAjB;QAAiCC;MAAjC,CAAD,CAArB;IACH;;IACD;EACH,CAZ4F,CAa7F;;;EACA,MAAMmB,KAAK,GAAGC,MAAM,CAACC,OAAP,CAAeR,IAAf,CAAd;;EACA,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGG,KAAK,CAACF,MAA1B,EAAkCD,CAAC,EAAnC,EAAuC;IACnC,MAAM,CAACM,GAAD,EAAMC,KAAN,IAAeJ,KAAK,CAACH,CAAD,CAA1B;;IAEA,IAAIM,GAAG,KAAK,MAAR,IAAkBC,KAAlB,IAA2BxB,cAAc,CAACyB,GAAf,CAAmBD,KAAK,CAACb,EAAzB,CAA/B,EAA6D;MACzDa,KAAK,CAACD,GAAN,GAAYvB,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAZ;MACAa,KAAK,CAACE,IAAN,GAAa1B,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAb;MACAa,KAAK,CAACG,GAAN,GAAa,GAAE1B,SAAU,GAAEA,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAA0B,EAA1B,GAA+B,GAAI,GAAEJ,cAAc,CAACO,GAAf,CAC5DiB,KAAK,CAACb,EADsD,CAE9D,EAFF;IAGH,CAND,MAMO;MACHE,qBAAqB,CAAC;QAAEC,IAAI,EAAEU,KAAR;QAAevB,SAAf;QAA0BD;MAA1B,CAAD,CAArB;IACH;EACJ;AACJ;;AAYM,MAAM4B,gBAAgB,GAAG,MAC5B9B,MAD4B,IAEU;EACtC,MAAM;IAAE+B,OAAF;IAAWC,SAAX;IAAsBC;EAAtB,IAA0CjC,MAAhD,CADsC,CAEtC;;EACA,MAAME,cAAc,GAAG,IAAIgC,GAAJ,EAAvB;EACA;AACJ;AACA;AACA;AACA;;EACI,IAAIC,OAAO,CAACC,GAAR,CAAYC,QAAZ,KAAyB,MAA7B,EAAqC;IACjC,OAAO;MACHnC;IADG,CAAP;EAGH;;EACDoC,OAAO,CAACC,GAAR,CAAY,yBAAZ,EAdsC,CAgBtC;;EACA,MAAMC,gBAAgB,GAAG,IAAIN,GAAJ,EAAzB,CAjBsC,CAkBtC;;EACA,KAAK,IAAIf,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGa,SAAS,CAACZ,MAA9B,EAAsCD,CAAC,EAAvC,EAA2C;IACvC,MAAMsB,IAAI,GAAGT,SAAS,CAACb,CAAD,CAAtB;IACAqB,gBAAgB,CAAC9B,GAAjB,CAAqB+B,IAAI,CAAChB,GAA1B,EAA+BgB,IAA/B,EAFuC,CAIvC;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4B4B,IAAI,CAACC,IAAjC;EACH;;EAED,MAAMC,iBAAiB,GAAG,MAAMC,iBAAiB,CAAC;IAC9CJ,gBAD8C;IAE9CK,iBAAiB,EAAEZ,eAAe,CAACa;EAFW,CAAD,CAAjD,CA3BsC,CAgCtC;;EACA,MAAMC,gBAAgB,GAAGJ,iBAAiB,CACrCK,GADoB,CACfC,YAAD,IAAoC;IACrC,MAAMC,MAAM,GAAGD,YAAY,CAACE,GAA5B;IACA,MAAMV,IAAI,GAAGD,gBAAgB,CAAC/B,GAAjB,CAAqB2C,aAAa,CAACF,MAAD,CAAlC,CAAb;;IACA,IAAI,CAACT,IAAL,EAAW;MACP,OAAO,IAAP;IACH,CALoC,CAOrC;;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4BqC,MAA5B;IAEA,OAAO;MACHzB,GAAG,EAAEyB,MADF;MAEHtB,IAAI,EAAEa,IAAI,CAACb,IAFR;MAGHyB,IAAI,EAAEZ,IAAI,CAACY,IAHR;MAIHX,IAAI,EAAED,IAAI,CAACC,IAJR;MAKHY,IAAI,EAAEb,IAAI,CAACa,IALR;MAMHC,IAAI,EAAEd,IAAI,CAACc;IANR,CAAP;EAQH,CAnBoB,EAoBpBC,MApBoB,CAoBbC,OApBa,CAAzB;EAsBA,MAAMC,mBAAmB,GAAG,EAA5B,CAvDsC,CAwDtC;;EACA,MAAMC,sBAAsB,GAAG,IAAAC,cAAA,EAAMb,gBAAN,EAAwBjD,yBAAxB,CAA/B;;EACA,KAAK,IAAIqB,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGwC,sBAAsB,CAACvC,MAA3C,EAAmDD,CAAC,EAApD,EAAwD;IACpD,MAAM0C,qBAAqB,GAAGF,sBAAsB,CAACxC,CAAD,CAApD;IACAuC,mBAAmB,CAACI,IAApB;IACI;AACZ;AACA;AACA;IACY/B,OAAO,CAACgC,WAAR,CAAoBC,KAApB,CAA0BC,kBAA1B,CAA6CJ,qBAA7C,CALJ;EAOH;;EAED,MAAMK,OAAO,CAACC,GAAR,CAAYT,mBAAZ,CAAN;EAEA,OAAO;IACHxD;EADG,CAAP;AAGH,CA5EM;;;;AA0FA,eAAekE,UAAf,CAA0B;EAC7BC,OAD6B;EAE7BtC,OAF6B;EAG7BE;AAH6B,CAA1B,EAIiD;EACpD,MAAMM,GAAG,GAAGD,OAAO,CAACC,GAApB,CADoD,CAGpD;;EACA,MAAM+B,gBAAgB,GAAG1E,aAAA,CAAKC,IAAL,CAAUF,mBAAV,EAA+B0E,OAA/B,CAAzB;;EACA,IAAAE,sBAAA,EAAcD,gBAAd;;EAEA,MAAME,eAAe,GAAGhE,yBAAA,CAAQC,GAAR,CAAYwB,eAAZ,EAA8B,MAA9B,CAAxB;;EACA,MAAMwC,mBAAmB,GAAG7E,aAAA,CAAKC,IAAL,CAAUyE,gBAAV,EAA4B1E,aAAA,CAAK8E,QAAL,CAAcF,eAAd,CAA5B,CAA5B;;EAEAjC,GAAG,CAAE,+BAA8BiC,eAAgB,QAAOC,mBAAoB,GAA3E,CAAH,CAVoD,CAWpD;;EACA,MAAM,IAAIP,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACnCC,kBAAA,CACKC,UADL,CACgBN,eADhB,EAEKO,EAFL,CAEQ,OAFR,EAEiBH,MAFjB,EAGKI,IAHL,CAGU,IAAAC,qBAAA,EAAkBR,mBAAlB,CAHV,EAIKM,EAJL,CAIQ,OAJR,EAIiBH,MAJjB,EAKKG,EALL,CAKQ,QALR,EAKkBJ,OALlB;EAMH,CAPK,CAAN,CAZoD,CAqBpD;;EACApC,GAAG,CAAE,aAAYiC,eAAgB,EAA9B,CAAH;EACA,MAAM;IAAEU,IAAF;IAAQlB;EAAR,IAAkB,MAAM,IAAAmB,qBAAA,EAA2BV,mBAA3B,CAA9B,CAvBoD,CAyBpD;;EACA,IAAIT,KAAK,IAAI/C,KAAK,CAACC,OAAN,CAAc8C,KAAd,CAAT,IAAiCA,KAAK,CAAC5C,MAAN,GAAe,CAApD,EAAuD;IACnD;IACA,MAAM;MAAElB;IAAF,IAAqB,MAAM4B,gBAAgB,CAAC;MAC9CC,OAD8C;;MAE9C;AACZ;AACA;MACY;MACAC,SAAS,EAAEgC,KANmC;MAO9C/B;IAP8C,CAAD,CAAjD;IAUA,MAAMhC,QAAQ,GAAG,MAAM8B,OAAO,CAACgC,WAAR,CAAoB9D,QAApB,CAA6BmF,WAA7B,EAAvB;IAEA,MAAM;MAAEjF,SAAS,GAAG;IAAd,IAAqBF,QAAQ,IAAI,EAAvC;IACAc,qBAAqB,CAAC;MAClBC,IAAI,EAAEkE,IAAI,CAACG,OAAL,IAAgB,EADJ;MAElBnF,cAFkB;MAGlBC;IAHkB,CAAD,CAArB;IAMA+E,IAAI,CAACjF,QAAL,GAAgBF,yBAAyB,CAAC;MACtCE,QAAQ,EAAEiF,IAAI,CAACjF,QAAL,IAAiB,EADW;MAEtCC,cAFsC;MAGtCC;IAHsC,CAAD,CAAzC;EAKH;;EAEDoC,GAAG,CAAC,gCAAD,CAAH;EACA,MAAM,IAAA+C,gCAAA,EAAWjB,OAAX,CAAN;EAEA9B,GAAG,CAAE,iCAAF,CAAH;EACA,MAAMgD,cAAc,CAAC3F,aAAA,CAAK4F,OAAL,CAAavD,eAAe,CAACjB,IAA7B,CAAD,CAApB;EAEA,OAAOkE,IAAP;AACH;;AAOD,eAAetC,iBAAf,CAAiC;EAC7BJ,gBAD6B;EAE7BK;AAF6B,CAAjC,EAGmE;EAC/D,MAAM4C,gBAAgB,GAAGlE,MAAM,CAACmE,IAAP,CAAY7C,iBAAZ,CAAzB;EAEA,MAAM8C,QAAQ,GAAG,EAAjB,CAH+D,CAI/D;;EACA,KAAK,IAAIxE,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGsE,gBAAgB,CAACrE,MAArC,EAA6CD,CAAC,EAA9C,EAAkD;IAC9C,MAAMyE,MAAM,GAAGH,gBAAgB,CAACtE,CAAD,CAA/B;IACA,MAAM0E,UAAU,GAAGhD,iBAAiB,CAAC+C,MAAD,CAApC,CAF8C,CAI9C;;IACA,MAAMd,UAAU,GAAGD,kBAAA,CAASC,UAAT,CAAoBe,UAApB,CAAnB,CAL8C,CAM9C;;;IACA,MAAMC,YAAY,GAAGtD,gBAAgB,CAAC/B,GAAjB,CAAqBmF,MAArB,CAArB;;IAEA,IAAIE,YAAJ,EAAkB;MACd,MAAM5C,MAAM,GAAG,IAAA6C,eAAA,EAAS,EAAT,EAAc,IAAGD,YAAY,CAACrE,GAAI,EAAlC,CAAf;;MACA,MAAM;QAAEuE,iBAAF;QAAqBC,8BAA8B,EAAEC;MAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6B4C,YAAY,CAACpD,IAA1C,CADJ;;MAEAoC,UAAU,CAACE,IAAX,CAAgBgB,iBAAhB;MACAL,QAAQ,CAAC7B,IAAT,CAAcoC,OAAd;MAEA5D,OAAO,CAACC,GAAR,CAAa,6BAA4BW,MAAO,GAAhD;IACH;EACJ;;EAED,OAAOgB,OAAO,CAACC,GAAR,CAAYwB,QAAZ,CAAP;AACH;;AAED,SAASvC,aAAT,CAAuB3B,GAAvB,EAAoC;EAChC;AACJ;AACA;EACI,IAAI;IACA,MAAM,GAAG,GAAG2E,IAAN,IAAc3E,GAAG,CAAC4E,KAAJ,CAAU,GAAV,CAApB;IACA,OAAOD,IAAI,CAACvG,IAAL,CAAU,GAAV,CAAP;EACH,CAHD,CAGE,OAAOyG,CAAP,EAAU;IACR,OAAO7E,GAAP;EACH;AACJ;;AAED,MAAM8E,iBAAiB,GAAG,0BAA1B;;AAEA,SAASC,qBAAT,CAA+BC,QAA/B,EAAyD;EACrD,OAAO7G,aAAA,CAAK8E,QAAL,CAAc+B,QAAd,EAAwBC,OAAxB,CAAgC9G,aAAA,CAAK+G,OAAL,CAAaF,QAAb,CAAhC,EAAwD,EAAxD,CAAP;AACH;;AAQD;AACA;AACA;AACA;AACA;AACO,eAAeG,mCAAf,CACHC,UADG,EAEsB;EACzB,MAAMtE,GAAG,GAAGD,OAAO,CAACC,GAApB;EACA,MAAMuE,kBAAkB,GAAG,EAA3B;;EAEA,MAAMC,WAAW,GAAGnH,aAAA,CAAK8E,QAAL,CAAcmC,UAAd,EAA0BR,KAA1B,CAAgC,GAAhC,EAAqC,CAArC,CAApB;;EAEA,MAAMW,QAAQ,GAAG,MAAM,IAAAC,kBAAA,EAAMJ,UAAN,CAAvB;;EACA,IAAI,CAACG,QAAQ,CAACE,EAAd,EAAkB;IACd,MAAM,IAAIC,cAAJ,CAAiB,gCAA+BN,UAAW,GAA3D,EAA+DG,QAAQ,CAACI,UAAxE,CAAN;EACH;;EAED,MAAMtC,UAAU,GAAGkC,QAAQ,CAACK,IAA5B;EAEA,MAAMC,UAAU,GAAG,IAAAvB,eAAA,EAAS,eAAT,CAAnB,CAbyB,CAczB;;EACA,MAAMwB,aAAa,GAAG3H,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBqH,WAAvB,CAAtB;;EAEA,MAAMZ,WAAW,GAAG,IAAAlB,qBAAA,EAAkBsC,aAAlB,CAApB;EACA,MAAMhI,cAAc,CAACuF,UAAD,EAAaqB,WAAb,CAApB;EACA5D,GAAG,CAAE,oBAAmBwE,WAAY,QAAOQ,aAAc,EAAtD,CAAH,CAnByB,CAqBzB;;EACA,MAAMC,YAAY,GAAG,MAAMC,gBAAgB,CAACF,aAAD,CAA3C;EAEAhF,GAAG,CAAE,sBAAqBsE,UAAW,UAASU,aAAc,EAAzD,CAAH;EACA,MAAM,IAAAjC,gCAAA,EAAWiC,aAAX,CAAN,CAzByB,CA2BzB;;EACA,KAAK,IAAIpG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGqG,YAAY,CAACpG,MAAjC,EAAyCD,CAAC,EAA1C,EAA8C;IAC1C,MAAMuG,WAAW,GAAGF,YAAY,CAACrG,CAAD,CAAhC;IACA,MAAMwG,OAAO,GAAG,MAAMC,uBAAuB,CAACF,WAAD,EAAcJ,UAAd,CAA7C;IACAR,kBAAkB,CAAChD,IAAnB,CAAwB6D,OAAxB;EACH;;EACDpF,GAAG,CAAC,oCAAD,EAAuC3C,aAAA,CAAK4F,OAAL,CAAagC,YAAY,CAAC,CAAD,CAAzB,CAAvC,CAAH;EACA,MAAM,IAAAlC,gCAAA,EAAW1F,aAAA,CAAK4F,OAAL,CAAagC,YAAY,CAAC,CAAD,CAAzB,CAAX,CAAN;EAEA,OAAOV,kBAAP;AACH;;AAED,MAAMe,eAAe,GAAG,SAAxB;;AAEA,SAASC,qBAAT,CAA+B;EAC3B9E,GAD2B;EAE3B+E,QAF2B;EAG3B7E;AAH2B,CAA/B,EAQmB;EACf,MAAMsC,OAAO,GAAG5F,aAAA,CAAK4F,OAAL,CAAauC,QAAb,CAAhB;;EACA,MAAMtB,QAAQ,GAAG7G,aAAA,CAAK8E,QAAL,CAAcqD,QAAd,CAAjB;EACA;AACJ;AACA;AACA;;;EACI,MAAMnC,MAAM,GAAGa,QAAQ,CAACC,OAAT,CAAiB,KAAjB,EAAwB,KAAxB,CAAf;EAEA,MAAMsB,OAAO,GAAGxC,OAAO,CAAClF,QAAR,CAAiBuH,eAAjB,CAAhB;;EAEA,IAAIG,OAAJ,EAAa;IACThF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,UAAS4C,MAAO,EAAlC,EAAqC1C,MAArC,CAAN;EACH,CAFD,MAEO;IACH;IACAF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,MAAlB,EAAyBE,MAAzB,CAAN;EACH;;EAED,OAAOF,GAAP;AACH;;AAED,eAAeuC,cAAf,CAA8B9D,GAA9B,EAA0D;EACtD;EACA,IAAI,CAACA,GAAG,CAACnB,QAAJ,CAAa,GAAb,CAAL,EAAwB;IACpBmB,GAAG,GAAI,GAAEA,GAAI,GAAb;EACH;;EAED,MAAMuF,QAAQ,GAAG,MAAMnC,kBAAA,CAASoD,UAAT,CAAoBxG,GAApB,CAAvB;EACA,MAAMiE,IAAI,GAAG,CAACsB,QAAQ,CAACkB,QAAT,IAAqB,EAAtB,EAA0BlF,GAA1B,CAA8BmF,CAAC,IAAIA,CAAC,CAAChF,GAArC,EAA0CK,MAA1C,CAAiDC,OAAjD,CAAb;EACAnB,OAAO,CAACC,GAAR,CAAa,SAAQmD,IAAI,CAACtE,MAAO,SAAjC;EAEA,MAAMgH,kBAAkB,GAAG1C,IAAI,CAAC1C,GAAL,CAASvB,GAAG,IAAIoD,kBAAA,CAASwD,YAAT,CAAsB5G,GAAtB,CAAhB,CAA3B;EAEA,MAAMyC,OAAO,CAACC,GAAR,CAAYiE,kBAAZ,CAAN;EACA9F,OAAO,CAACC,GAAR,CAAa,wBAAuB6F,kBAAkB,CAAChH,MAAO,SAA9D;AACH,C,CAED;;;AAEO,SAASkH,YAAT,CAAsBC,KAAtB,EAAqC;EACxC,OAAO;IACH,CAACC,iCAAA,CAA2BC,OAA5B,GAAsCF,KADnC;IAEH,CAACC,iCAAA,CAA2BE,UAA5B,GAAyC,CAFtC;IAGH,CAACF,iCAAA,CAA2BG,SAA5B,GAAwC,CAHrC;IAIH,CAACH,iCAAA,CAA2BI,MAA5B,GAAqC,CAJlC;IAKHL;EALG,CAAP;AAOH;;AAED,SAASd,gBAAT,CAA0BoB,iBAA1B,EAAwE;EACpE,OAAO,IAAI3E,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMkE,gBAA0B,GAAG,EAAnC;IACA,MAAMC,yBAAyB,GAAGvC,qBAAqB,CAACqC,iBAAD,CAAvD;;IACA,MAAMG,2BAA2B,GAAGpJ,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBqJ,yBAAvB,CAApC,CAHoC,CAIpC;;;IACA,IAAAxE,sBAAA,EAAcyE,2BAAd;;IAEAC,cAAA,CAAMC,IAAN,CAAWL,iBAAX,EAA8B;MAAEM,WAAW,EAAE;IAAf,CAA9B,EAAqD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MACzE,IAAID,GAAJ,EAAS;QACL9G,OAAO,CAACgH,IAAR,CAAa,gCAAb,EAA+CT,iBAA/C,EAAkEO,GAAlE;QACAxE,MAAM,CAACwE,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACV/G,OAAO,CAACC,GAAR,CAAY,gDAAgDsG,iBAA5D;QACAjE,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MAEDtC,OAAO,CAACiH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MAEAH,OAAO,CAACtE,EAAR,CAAW,KAAX,EAAkB,UAAUqE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACL9G,OAAO,CAACgH,IAAR,CAAa,uCAAb,EAAsDT,iBAAtD,EAAyEO,GAAzE;UACAxE,MAAM,CAACwE,GAAD,CAAN;QACH;;QACDzE,OAAO,CAACmE,gBAAD,CAAP;MACH,CAND;MAQAO,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAACtE,EAAR,CAAW,OAAX,EAAoB,UAAU2E,KAAV,EAAiB;QACjCpH,OAAO,CAACiH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAetE,UAAf,EAA2B;YACrD,IAAIsE,GAAJ,EAAS;cACL9G,OAAO,CAACgH,IAAR,CACI,4CADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKAxE,MAAM,CAACwE,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAACtE,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CACI,8DADJ;cAGAqC,MAAM,CAAC,+BAAD,CAAN;cACA;YACH;;YAED,MAAMmD,QAAQ,GAAGnI,aAAA,CAAKC,IAAL,CAAUmJ,2BAAV,EAAuCU,KAAK,CAACjD,QAA7C,CAAjB;;YAEA3B,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7B+D,gBAAgB,CAAChF,IAAjB,CAAsBiE,QAAtB;cACAsB,OAAO,CAACI,SAAR;YACH,CAHD;YAKAlK,cAAc,CAACuF,UAAD,EAAa,IAAAG,qBAAA,EAAkB8C,QAAlB,CAAb,CAAd,CAAwD8B,KAAxD,CAA8DC,KAAK,IAAI;cACnElF,MAAM,CAACkF,KAAD,CAAN;YACH,CAFD;UAGH,CA5BD;QA6BH;MACJ,CAvCD;IAwCH,CAhED;EAiEH,CAxEM,CAAP;AAyEH;;AAED,SAASlC,uBAAT,CACImC,mBADJ,EAEIzC,UAFJ,EAG2B;EACvB,OAAO,IAAIpD,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMoF,SAAS,GAAG,EAAlB;IACA,MAAMC,kBAAwD,GAAG,EAAjE;IACA,MAAMC,aAAa,GAAG1D,qBAAqB,CAACuD,mBAAD,CAA3C;IACA,IAAIpC,OAAuB,GAAG;MAC1BlG,GAAG,EAAEyI,aADqB;MAE1BpH,MAAM,EAAE,EAFkB;MAG1B9B,IAAI,EAAE;IAHoB,CAA9B;;IAKAiI,cAAA,CAAMC,IAAN,CAAWa,mBAAX,EAAgC;MAAEZ,WAAW,EAAE;IAAf,CAAhC,EAAuD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MAC3E,IAAID,GAAJ,EAAS;QACL9G,OAAO,CAACgH,IAAR,CAAa,gCAAb,EAA+CS,mBAA/C,EAAoEX,GAApE;QACAxE,MAAM,CAACwE,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACV/G,OAAO,CAACC,GAAR,CAAY,4CAA4CwH,mBAAxD;QACAnF,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MACDtC,OAAO,CAACiH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MACAH,OAAO,CAACtE,EAAR,CAAW,KAAX,EAAkB,UAAUqE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACL9G,OAAO,CAACgH,IAAR,CAAa,mCAAb,EAAkDS,mBAAlD,EAAuEX,GAAvE;UACAxE,MAAM,CAACwE,GAAD,CAAN;QACH;;QAEDlF,OAAO,CAACC,GAAR,CAAY8F,kBAAZ,EAAgCE,IAAhC,CAAqCC,GAAG,IAAI;UACxCA,GAAG,CAACC,OAAJ,CAAYC,CAAC,IAAI;YACbhI,OAAO,CAACiH,IAAR,CAAa,oBAAb,EAAmCe,CAAnC;UACH,CAFD;UAGA3F,OAAO,CAACgD,OAAD,CAAP;QACH,CALD;MAMH,CAZD;MAcA0B,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAACtE,EAAR,CAAW,OAAX,EAAoB,UAAU2E,KAAV,EAAiB;QACjCpH,OAAO,CAACiH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAetE,UAAf,EAA2B;YACrD,IAAIsE,GAAJ,EAAS;cACL9G,OAAO,CAACgH,IAAR,CACI,4DADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKAxE,MAAM,CAACwE,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAACtE,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CAAY,mDAAZ;cACAqC,MAAM,CAAC,8BAAD,CAAN;cACA;YACH;;YACDE,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7BiF,SAAS,CAAClG,IAAV,CAAe4F,KAAK,CAACjD,QAArB;cACA4C,OAAO,CAACI,SAAR;YACH,CAHD;YAKA,MAAMvG,MAAM,GAAI,GAAEoE,UAAW,IAAG4C,aAAc,IAAGR,KAAK,CAACjD,QAAS,EAAhE,CApBqD,CAqBrD;;YACAkB,OAAO,GAAGG,qBAAqB,CAAC;cAC5B9E,GAAG,EAAE2E,OADuB;cAE5BI,QAAQ,EAAE2B,KAAK,CAACjD,QAFY;cAG5BvD;YAH4B,CAAD,CAA/B;;YAMA,MAAM;cAAE8C,iBAAF;cAAqBC,8BAA8B,EAAEC;YAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6BqD,iBAA7B,CADJ;;YAGAhH,cAAc,CAACuF,UAAD,EAAakB,iBAAb,CAAd,CACKmE,IADL,CACU,MAAM;cACRF,kBAAkB,CAACnG,IAAnB,CAAwBoC,OAAxB;YACH,CAHL,EAIK2D,KAJL,CAIWC,KAAK,IAAI;cACZlF,MAAM,CAACkF,KAAD,CAAN;YACH,CANL;UAOH,CAtCD;QAuCH;MACJ,CAjDD;IAkDH,CA9ED;EA+EH,CAxFM,CAAP;AAyFH"}
@@ -0,0 +1 @@
1
+ {"version":3,"names":["mockSecurity","identity","context","security","disableAuthorization","setIdentity"],"sources":["mockSecurity.ts"],"sourcesContent":["import { SecurityContext, SecurityIdentity } from \"@webiny/api-security/types\";\n\nexport const mockSecurity = (identity: SecurityIdentity, context: SecurityContext) => {\n context.security.disableAuthorization();\n context.security.setIdentity(identity);\n};\n"],"mappings":";;;;;;;AAEO,MAAMA,YAAY,GAAG,CAACC,QAAD,EAA6BC,OAA7B,KAA0D;EAClFA,OAAO,CAACC,QAAR,CAAiBC,oBAAjB;EACAF,OAAO,CAACC,QAAR,CAAiBE,WAAjB,CAA6BJ,QAA7B;AACH,CAHM"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@webiny/api-page-builder-import-export",
3
- "version": "0.0.0-mt-3",
3
+ "version": "0.0.0-unstable.1e66d121db",
4
4
  "main": "index.js",
5
5
  "keywords": [
6
6
  "pbie:base"
@@ -14,52 +14,54 @@
14
14
  "author": "Webiny Ltd",
15
15
  "license": "MIT",
16
16
  "dependencies": {
17
- "@babel/runtime": "7.15.4",
17
+ "@babel/runtime": "7.19.0",
18
18
  "@commodo/fields": "1.1.2-beta.20",
19
- "@webiny/api-file-manager": "0.0.0-mt-3",
20
- "@webiny/api-page-builder": "0.0.0-mt-3",
21
- "@webiny/api-security": "0.0.0-mt-3",
22
- "@webiny/error": "0.0.0-mt-3",
23
- "@webiny/handler": "0.0.0-mt-3",
24
- "@webiny/handler-args": "0.0.0-mt-3",
25
- "@webiny/handler-aws": "0.0.0-mt-3",
26
- "@webiny/handler-graphql": "0.0.0-mt-3",
27
- "@webiny/validation": "0.0.0-mt-3",
28
- "archiver": "5.3.0",
19
+ "@webiny/api": "0.0.0-unstable.1e66d121db",
20
+ "@webiny/api-file-manager": "0.0.0-unstable.1e66d121db",
21
+ "@webiny/api-page-builder": "0.0.0-unstable.1e66d121db",
22
+ "@webiny/api-security": "0.0.0-unstable.1e66d121db",
23
+ "@webiny/error": "0.0.0-unstable.1e66d121db",
24
+ "@webiny/handler": "0.0.0-unstable.1e66d121db",
25
+ "@webiny/handler-aws": "0.0.0-unstable.1e66d121db",
26
+ "@webiny/handler-graphql": "0.0.0-unstable.1e66d121db",
27
+ "@webiny/utils": "0.0.0-unstable.1e66d121db",
28
+ "@webiny/validation": "0.0.0-unstable.1e66d121db",
29
+ "archiver": "5.3.1",
29
30
  "commodo-fields-object": "1.0.6",
30
31
  "dot-prop-immutable": "2.1.1",
31
- "fs-extra": "7.0.1",
32
+ "fs-extra": "9.1.0",
32
33
  "load-json-file": "6.2.0",
33
34
  "lodash": "4.17.21",
34
35
  "mdbid": "1.0.0",
35
- "node-fetch": "2.6.5",
36
+ "node-fetch": "2.6.7",
36
37
  "stream": "0.0.2",
37
38
  "uniqid": "5.4.0",
38
39
  "yauzl": "2.10.0"
39
40
  },
40
41
  "devDependencies": {
41
- "@babel/cli": "^7.5.5",
42
- "@babel/core": "^7.5.5",
43
- "@babel/plugin-proposal-export-default-from": "^7.5.2",
44
- "@babel/preset-env": "^7.5.5",
45
- "@babel/preset-typescript": "^7.8.3",
46
- "@elastic/elasticsearch": "7.12.0",
47
- "@shelf/jest-elasticsearch": "^1.0.0",
48
- "@types/puppeteer": "^5.4.2",
49
- "@webiny/api-dynamodb-to-elasticsearch": "^0.0.0-mt-3",
50
- "@webiny/api-file-manager-ddb-es": "^0.0.0-mt-3",
51
- "@webiny/api-i18n-ddb": "^0.0.0-mt-3",
52
- "@webiny/api-security-so-ddb": "^0.0.0-mt-3",
53
- "@webiny/api-tenancy": "^0.0.0-mt-3",
54
- "@webiny/api-tenancy-so-ddb": "^0.0.0-mt-3",
55
- "@webiny/cli": "^0.0.0-mt-3",
56
- "@webiny/db": "^0.0.0-mt-3",
57
- "@webiny/project-utils": "^0.0.0-mt-3",
58
- "jest": "^26.6.3",
42
+ "@babel/cli": "^7.19.3",
43
+ "@babel/core": "^7.19.3",
44
+ "@babel/plugin-proposal-export-default-from": "^7.16.0",
45
+ "@babel/preset-env": "^7.19.4",
46
+ "@babel/preset-typescript": "^7.18.6",
47
+ "@types/archiver": "^5.3.1",
48
+ "@types/node-fetch": "^2.6.1",
49
+ "@types/yauzl": "^2.9.2",
50
+ "@webiny/api-dynamodb-to-elasticsearch": "^0.0.0-unstable.1e66d121db",
51
+ "@webiny/api-file-manager-ddb-es": "^0.0.0-unstable.1e66d121db",
52
+ "@webiny/api-i18n-ddb": "^0.0.0-unstable.1e66d121db",
53
+ "@webiny/api-security-so-ddb": "^0.0.0-unstable.1e66d121db",
54
+ "@webiny/api-tenancy": "^0.0.0-unstable.1e66d121db",
55
+ "@webiny/api-tenancy-so-ddb": "^0.0.0-unstable.1e66d121db",
56
+ "@webiny/api-wcp": "^0.0.0-unstable.1e66d121db",
57
+ "@webiny/cli": "^0.0.0-unstable.1e66d121db",
58
+ "@webiny/db": "^0.0.0-unstable.1e66d121db",
59
+ "@webiny/project-utils": "^0.0.0-unstable.1e66d121db",
60
+ "jest": "^28.1.0",
59
61
  "jest-dynalite": "^3.2.0",
60
62
  "rimraf": "^3.0.2",
61
63
  "ttypescript": "^1.5.12",
62
- "typescript": "^4.1.3"
64
+ "typescript": "4.7.4"
63
65
  },
64
66
  "publishConfig": {
65
67
  "access": "public",
@@ -76,5 +78,5 @@
76
78
  ]
77
79
  }
78
80
  },
79
- "gitHead": "ebea815be2be99404591cba465cc1fe88355bd48"
81
+ "gitHead": "b670bf27c5039de1a2b0be764a09ba4cb94ad5e2"
80
82
  }
package/types.d.ts CHANGED
@@ -16,6 +16,11 @@ export interface PageImportExportTaskStats {
16
16
  [PageImportExportTaskStatus.FAILED]: number;
17
17
  total: number;
18
18
  }
19
+ interface CreatedBy {
20
+ id: string;
21
+ type: string;
22
+ displayName: string | null;
23
+ }
19
24
  export interface PageImportExportTask {
20
25
  id: string;
21
26
  parent: string;
@@ -25,18 +30,14 @@ export interface PageImportExportTask {
25
30
  error: Record<string, any>;
26
31
  input: Record<string, any>;
27
32
  createdOn: string;
28
- createdBy: {
29
- type: string;
30
- id: string;
31
- displayName: string;
32
- };
33
+ createdBy: CreatedBy;
33
34
  tenant: string;
34
35
  locale: string;
35
36
  }
36
- export declare type File = {
37
+ export interface File {
37
38
  id: string;
38
39
  src: string;
39
- };
40
+ }
40
41
  export interface MetaResponse {
41
42
  cursor: string | null;
42
43
  totalCount: number;
package/types.js.map ADDED
@@ -0,0 +1 @@
1
+ {"version":3,"names":["PageExportRevisionType","PageImportExportTaskStatus"],"sources":["types.ts"],"sourcesContent":["export * from \"./graphql/types\";\n\n// Entities.\nexport enum PageExportRevisionType {\n PUBLISHED = \"published\",\n LATEST = \"latest\"\n}\n\nexport enum PageImportExportTaskStatus {\n PENDING = \"pending\",\n PROCESSING = \"processing\",\n COMPLETED = \"completed\",\n FAILED = \"failed\"\n}\n\nexport interface PageImportExportTaskStats {\n [PageImportExportTaskStatus.PENDING]: number;\n [PageImportExportTaskStatus.PROCESSING]: number;\n [PageImportExportTaskStatus.COMPLETED]: number;\n [PageImportExportTaskStatus.FAILED]: number;\n total: number;\n}\n\ninterface CreatedBy {\n id: string;\n type: string;\n displayName: string | null;\n}\n\nexport interface PageImportExportTask {\n id: string;\n parent: string;\n status: PageImportExportTaskStatus;\n data: Record<string, any>;\n stats: PageImportExportTaskStats;\n error: Record<string, any>;\n input: Record<string, any>;\n createdOn: string;\n createdBy: CreatedBy;\n tenant: string;\n locale: string;\n}\n\nexport interface File {\n id: string;\n src: string;\n}\n\nexport interface MetaResponse {\n cursor: string | null;\n totalCount: number;\n hasMoreItems: boolean;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsGetParams {\n where: {\n id: string;\n tenant: string;\n locale: string;\n };\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsListParams {\n where: {\n tenant: string;\n locale: string;\n };\n sort?: string[];\n limit?: number;\n after?: string | null;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport type PageImportExportTaskStorageOperationsListResponse = [\n PageImportExportTask[],\n MetaResponse\n];\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsCreateParams {\n input: Record<string, any>;\n task: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsUpdateParams {\n input: Record<string, any>;\n original: PageImportExportTask;\n task: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsDeleteParams {\n task: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsGetSubTaskParams {\n where: {\n id: string;\n parent: string;\n tenant: string;\n locale: string;\n };\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsListSubTaskParams {\n where: {\n tenant: string;\n locale: string;\n parent: string;\n status: PageImportExportTaskStatus;\n createdBy?: string;\n };\n sort?: string[];\n limit?: number;\n after?: string | null;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport type PageImportExportTaskStorageOperationsListSubTaskResponse = [\n PageImportExportTask[],\n MetaResponse\n];\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsCreateSubTaskParams {\n input: Record<string, any>;\n subTask: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsUpdateSubTaskParams {\n input: Record<string, any>;\n original: PageImportExportTask;\n subTask: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsUpdateTaskStatsParams {\n input: {\n prevStatus: PageImportExportTaskStatus;\n nextStatus: PageImportExportTaskStatus;\n };\n original: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperations {\n /**\n * Get a single page import export task item by given params.\n */\n getTask(\n params: PageImportExportTaskStorageOperationsGetParams\n ): Promise<PageImportExportTask | null>;\n\n /**\n * Get all page import export tasks by given params.\n */\n listTasks(\n params: PageImportExportTaskStorageOperationsListParams\n ): Promise<PageImportExportTaskStorageOperationsListResponse>;\n\n createTask(\n params: PageImportExportTaskStorageOperationsCreateParams\n ): Promise<PageImportExportTask>;\n\n updateTask(\n params: PageImportExportTaskStorageOperationsUpdateParams\n ): Promise<PageImportExportTask>;\n\n deleteTask(\n params: PageImportExportTaskStorageOperationsDeleteParams\n ): Promise<PageImportExportTask>;\n\n updateTaskStats(\n params: PageImportExportTaskStorageOperationsUpdateTaskStatsParams\n ): Promise<PageImportExportTask>;\n\n /**\n * Get a single page import export sub-task item by given params.\n */\n getSubTask(\n params: PageImportExportTaskStorageOperationsGetSubTaskParams\n ): Promise<PageImportExportTask | null>;\n\n /**\n * Get all page import export sub-tasks by given params.\n */\n listSubTasks(\n params: PageImportExportTaskStorageOperationsListSubTaskParams\n ): Promise<PageImportExportTaskStorageOperationsListSubTaskResponse>;\n\n createSubTask(\n params: PageImportExportTaskStorageOperationsCreateSubTaskParams\n ): Promise<PageImportExportTask>;\n\n updateSubTask(\n params: PageImportExportTaskStorageOperationsUpdateSubTaskParams\n ): Promise<PageImportExportTask>;\n}\n"],"mappings":";;;;;;;;;;;AAAA;;AAAA;EAAA;EAAA;EAAA;EAAA;IAAA;IAAA;MAAA;IAAA;EAAA;AAAA;AAEA;IACYA,sB;;;WAAAA,sB;EAAAA,sB;EAAAA,sB;GAAAA,sB,sCAAAA,sB;;IAKAC,0B;;;WAAAA,0B;EAAAA,0B;EAAAA,0B;EAAAA,0B;EAAAA,0B;GAAAA,0B,0CAAAA,0B"}
@@ -1,40 +0,0 @@
1
- "use strict";
2
-
3
- var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
4
-
5
- Object.defineProperty(exports, "__esModule", {
6
- value: true
7
- });
8
- exports.invokeHandlerClient = invokeHandlerClient;
9
-
10
- var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/defineProperty"));
11
-
12
- function ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); if (enumerableOnly) { symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; }); } keys.push.apply(keys, symbols); } return keys; }
13
-
14
- function _objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = arguments[i] != null ? arguments[i] : {}; if (i % 2) { ownKeys(Object(source), true).forEach(function (key) { (0, _defineProperty2.default)(target, key, source[key]); }); } else if (Object.getOwnPropertyDescriptors) { Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)); } else { ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } } return target; }
15
-
16
- async function invokeHandlerClient({
17
- context,
18
- name,
19
- payload
20
- }) {
21
- /*
22
- * Prepare "invocationArgs", we're hacking our wat here.
23
- * They are necessary to setup the "context.pageBuilder" object among other things in IMPORT_PAGE_FUNCTION
24
- */
25
- const {
26
- request
27
- } = context.http;
28
- const invocationArgs = {
29
- httpMethod: request.method,
30
- body: request.body,
31
- headers: request.headers,
32
- cookies: request.cookies
33
- }; // Invoke handler
34
-
35
- await context.handlerClient.invoke({
36
- name: name,
37
- payload: _objectSpread(_objectSpread({}, payload), invocationArgs),
38
- await: false
39
- });
40
- }