@webiny/api-page-builder-import-export 0.0.0-mt-3 → 0.0.0-unstable.1e66d121db
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/{importPages/client.d.ts → client.d.ts} +3 -2
- package/client.js +47 -0
- package/client.js.map +1 -0
- package/exportPages/combine/index.d.ts +7 -11
- package/exportPages/combine/index.js +37 -11
- package/exportPages/combine/index.js.map +1 -0
- package/exportPages/process/index.d.ts +10 -14
- package/exportPages/process/index.js +32 -28
- package/exportPages/process/index.js.map +1 -0
- package/exportPages/s3Stream.d.ts +2 -0
- package/exportPages/s3Stream.js +14 -7
- package/exportPages/s3Stream.js.map +1 -0
- package/exportPages/utils.d.ts +1 -1
- package/exportPages/utils.js +2 -2
- package/exportPages/utils.js.map +1 -0
- package/exportPages/zipper.d.ts +1 -0
- package/exportPages/zipper.js +3 -5
- package/exportPages/zipper.js.map +1 -0
- package/graphql/crud/pageImportExportTasks.crud.d.ts +1 -1
- package/graphql/crud/pageImportExportTasks.crud.js +47 -30
- package/graphql/crud/pageImportExportTasks.crud.js.map +1 -0
- package/graphql/crud/pages.crud.d.ts +1 -1
- package/graphql/crud/pages.crud.js +29 -21
- package/graphql/crud/pages.crud.js.map +1 -0
- package/graphql/crud.d.ts +1 -1
- package/graphql/crud.js +1 -1
- package/graphql/crud.js.map +1 -0
- package/graphql/graphql/pageImportExportTasks.gql.js +1 -1
- package/graphql/graphql/pageImportExportTasks.gql.js.map +1 -0
- package/graphql/graphql/pages.gql.js +8 -6
- package/graphql/graphql/pages.gql.js.map +1 -0
- package/graphql/graphql/utils/resolve.d.ts +1 -1
- package/graphql/graphql/utils/resolve.js.map +1 -0
- package/graphql/graphql.d.ts +1 -1
- package/graphql/graphql.js +1 -1
- package/graphql/graphql.js.map +1 -0
- package/graphql/index.d.ts +1 -1
- package/graphql/index.js +1 -1
- package/graphql/index.js.map +1 -0
- package/graphql/types.d.ts +5 -6
- package/graphql/types.js.map +1 -0
- package/importPages/create/index.d.ts +11 -16
- package/importPages/create/index.js +39 -30
- package/importPages/create/index.js.map +1 -0
- package/importPages/process/index.d.ts +10 -17
- package/importPages/process/index.js +25 -23
- package/importPages/process/index.js.map +1 -0
- package/importPages/utils.d.ts +17 -10
- package/importPages/utils.js +90 -59
- package/importPages/utils.js.map +1 -0
- package/mockSecurity.js.map +1 -0
- package/package.json +36 -34
- package/types.d.ts +8 -7
- package/types.js.map +1 -0
- package/importPages/client.js +0 -40
package/importPages/utils.js
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
"use strict";
|
2
2
|
|
3
|
-
var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
|
3
|
+
var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault").default;
|
4
4
|
|
5
5
|
Object.defineProperty(exports, "__esModule", {
|
6
6
|
value: true
|
@@ -8,7 +8,7 @@ Object.defineProperty(exports, "__esModule", {
|
|
8
8
|
exports.importPage = importPage;
|
9
9
|
exports.initialStats = initialStats;
|
10
10
|
exports.readExtractAndUploadZipFileContents = readExtractAndUploadZipFileContents;
|
11
|
-
exports.
|
11
|
+
exports.uploadPageAssets = void 0;
|
12
12
|
|
13
13
|
var _uniqid = _interopRequireDefault(require("uniqid"));
|
14
14
|
|
@@ -46,25 +46,29 @@ const INSTALL_DIR = "/tmp";
|
|
46
46
|
const INSTALL_EXTRACT_DIR = _path.default.join(INSTALL_DIR, "apiPageBuilderImportPage");
|
47
47
|
|
48
48
|
const FILES_COUNT_IN_EACH_BATCH = 15;
|
49
|
-
const ZIP_CONTENT_TYPE = "application/zip";
|
50
49
|
|
51
|
-
function updateImageInPageSettings({
|
52
|
-
|
53
|
-
|
54
|
-
|
55
|
-
|
50
|
+
function updateImageInPageSettings(params) {
|
51
|
+
const {
|
52
|
+
settings,
|
53
|
+
fileIdToKeyMap,
|
54
|
+
srcPrefix
|
55
|
+
} = params;
|
56
56
|
let newSettings = settings;
|
57
57
|
const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith("/") ? srcPrefix.slice(0, -1) : srcPrefix;
|
58
58
|
|
59
59
|
if (_dotPropImmutable.default.get(newSettings, "general.image.src")) {
|
60
|
-
|
60
|
+
var _settings$general, _settings$general$ima;
|
61
|
+
|
62
|
+
newSettings = _dotPropImmutable.default.set(newSettings, "general.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(((_settings$general = settings.general) === null || _settings$general === void 0 ? void 0 : (_settings$general$ima = _settings$general.image) === null || _settings$general$ima === void 0 ? void 0 : _settings$general$ima.id) || "")}`);
|
61
63
|
}
|
62
64
|
|
63
65
|
if (_dotPropImmutable.default.get(newSettings, "social.image.src")) {
|
64
|
-
|
66
|
+
var _settings$social, _settings$social$imag;
|
67
|
+
|
68
|
+
newSettings = _dotPropImmutable.default.set(newSettings, "social.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(((_settings$social = settings.social) === null || _settings$social === void 0 ? void 0 : (_settings$social$imag = _settings$social.image) === null || _settings$social$imag === void 0 ? void 0 : _settings$social$imag.id) || "")}`);
|
65
69
|
}
|
66
70
|
|
67
|
-
return
|
71
|
+
return newSettings;
|
68
72
|
}
|
69
73
|
|
70
74
|
function updateFilesInPageData({
|
@@ -111,23 +115,27 @@ function updateFilesInPageData({
|
|
111
115
|
}
|
112
116
|
}
|
113
117
|
|
114
|
-
const uploadPageAssets = async
|
115
|
-
|
116
|
-
|
117
|
-
|
118
|
-
|
118
|
+
const uploadPageAssets = async params => {
|
119
|
+
const {
|
120
|
+
context,
|
121
|
+
filesData,
|
122
|
+
fileUploadsData
|
123
|
+
} = params; // Save uploaded file key against static id for later use.
|
124
|
+
|
125
|
+
const fileIdToKeyMap = new Map();
|
119
126
|
/**
|
120
127
|
* This function contains logic of file download from S3.
|
121
128
|
* Current we're not mocking zip file download from S3 in tests at the moment.
|
122
129
|
* So, we're manually mocking it in case of test just by returning an empty object.
|
123
130
|
*/
|
131
|
+
|
124
132
|
if (process.env.NODE_ENV === "test") {
|
125
|
-
return {
|
133
|
+
return {
|
134
|
+
fileIdToKeyMap
|
135
|
+
};
|
126
136
|
}
|
127
137
|
|
128
|
-
console.log("INSIDE uploadPageAssets"); // Save
|
129
|
-
|
130
|
-
const fileIdToKeyMap = new Map(); // Save files meta data against old key for later use.
|
138
|
+
console.log("INSIDE uploadPageAssets"); // Save files meta data against old key for later use.
|
131
139
|
|
132
140
|
const fileKeyToFileMap = new Map(); // Initialize maps.
|
133
141
|
|
@@ -145,7 +153,12 @@ const uploadPageAssets = async ({
|
|
145
153
|
|
146
154
|
const createFilesInput = fileUploadResults.map(uploadResult => {
|
147
155
|
const newKey = uploadResult.Key;
|
148
|
-
const file = fileKeyToFileMap.get(getOldFileKey(newKey));
|
156
|
+
const file = fileKeyToFileMap.get(getOldFileKey(newKey));
|
157
|
+
|
158
|
+
if (!file) {
|
159
|
+
return null;
|
160
|
+
} // Update the file map with newly uploaded file.
|
161
|
+
|
149
162
|
|
150
163
|
fileIdToKeyMap.set(file.id, newKey);
|
151
164
|
return {
|
@@ -156,7 +169,7 @@ const uploadPageAssets = async ({
|
|
156
169
|
meta: file.meta,
|
157
170
|
tags: file.tags
|
158
171
|
};
|
159
|
-
});
|
172
|
+
}).filter(Boolean);
|
160
173
|
const createFilesPromises = []; // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).
|
161
174
|
|
162
175
|
const createFilesInputChunks = (0, _chunk.default)(createFilesInput, FILES_COUNT_IN_EACH_BATCH);
|
@@ -206,25 +219,31 @@ async function importPage({
|
|
206
219
|
files
|
207
220
|
} = await (0, _loadJsonFile.default)(PAGE_DATA_FILE_PATH); // Only update page data if there are files.
|
208
221
|
|
209
|
-
if (Array.isArray(files) && files.length) {
|
222
|
+
if (files && Array.isArray(files) && files.length > 0) {
|
210
223
|
// Upload page assets.
|
211
224
|
const {
|
212
225
|
fileIdToKeyMap
|
213
226
|
} = await uploadPageAssets({
|
214
227
|
context,
|
228
|
+
|
229
|
+
/**
|
230
|
+
* TODO @ts-refactor @ashutosh figure out correct types.
|
231
|
+
*/
|
232
|
+
// @ts-ignore
|
215
233
|
filesData: files,
|
216
234
|
fileUploadsData
|
217
235
|
});
|
236
|
+
const settings = await context.fileManager.settings.getSettings();
|
218
237
|
const {
|
219
|
-
srcPrefix
|
220
|
-
} =
|
238
|
+
srcPrefix = ""
|
239
|
+
} = settings || {};
|
221
240
|
updateFilesInPageData({
|
222
|
-
data: page.content,
|
241
|
+
data: page.content || {},
|
223
242
|
fileIdToKeyMap,
|
224
243
|
srcPrefix
|
225
244
|
});
|
226
245
|
page.settings = updateImageInPageSettings({
|
227
|
-
settings: page.settings,
|
246
|
+
settings: page.settings || {},
|
228
247
|
fileIdToKeyMap,
|
229
248
|
srcPrefix
|
230
249
|
});
|
@@ -270,14 +289,6 @@ async function uploadFilesFromS3({
|
|
270
289
|
return Promise.all(promises);
|
271
290
|
}
|
272
291
|
|
273
|
-
async function getObjectMetaFromS3(Key) {
|
274
|
-
const meta = await _s3Stream.s3Stream.getObjectHead(Key);
|
275
|
-
|
276
|
-
if (meta.ContentType !== ZIP_CONTENT_TYPE) {
|
277
|
-
throw new _error.default(`Unsupported file type: "${meta.ContentType}"`, "UNSUPPORTED_FILE");
|
278
|
-
}
|
279
|
-
}
|
280
|
-
|
281
292
|
function getOldFileKey(key) {
|
282
293
|
/*
|
283
294
|
* Because we know the naming convention, we can extract the old key from new key.
|
@@ -298,32 +309,23 @@ function getFileNameWithoutExt(fileName) {
|
|
298
309
|
|
299
310
|
/**
|
300
311
|
* Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.
|
301
|
-
* @param
|
312
|
+
* @param zipFileUrl
|
302
313
|
* @return PageImportData S3 file keys for all uploaded assets group by page.
|
303
314
|
*/
|
304
|
-
async function readExtractAndUploadZipFileContents(
|
315
|
+
async function readExtractAndUploadZipFileContents(zipFileUrl) {
|
305
316
|
const log = console.log;
|
306
317
|
const pageImportDataList = [];
|
307
|
-
let readStream; // Check whether it is a URL
|
308
318
|
|
309
|
-
|
310
|
-
const response = await (0, _nodeFetch.default)(zipFileKey);
|
319
|
+
const zipFileName = _path.default.basename(zipFileUrl).split("?")[0];
|
311
320
|
|
312
|
-
|
313
|
-
throw new _error.default(`Unable to downloading file: "${zipFileKey}"`, response.statusText);
|
314
|
-
}
|
321
|
+
const response = await (0, _nodeFetch.default)(zipFileUrl);
|
315
322
|
|
316
|
-
|
317
|
-
|
318
|
-
// We're first retrieving object's meta data, just to check whether the file is available at the given Key
|
319
|
-
await getObjectMetaFromS3(zipFileKey);
|
320
|
-
readStream = _s3Stream.s3Stream.readStream(zipFileKey);
|
323
|
+
if (!response.ok) {
|
324
|
+
throw new _error.default(`Unable to downloading file: "${zipFileUrl}"`, response.statusText);
|
321
325
|
}
|
322
326
|
|
323
|
-
const
|
324
|
-
|
325
|
-
const zipFileName = _path.default.basename(zipFileKey); // Read export file and download it in the disk
|
326
|
-
|
327
|
+
const readStream = response.body;
|
328
|
+
const uniquePath = (0, _uniqid.default)("IMPORT_PAGES/"); // Read export file and download it in the disk
|
327
329
|
|
328
330
|
const ZIP_FILE_PATH = _path.default.join(INSTALL_DIR, zipFileName);
|
329
331
|
|
@@ -332,7 +334,7 @@ async function readExtractAndUploadZipFileContents(zipFileKey) {
|
|
332
334
|
log(`Downloaded file "${zipFileName}" at ${ZIP_FILE_PATH}`); // Extract the downloaded zip file
|
333
335
|
|
334
336
|
const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);
|
335
|
-
log(`Removing ZIP file "${
|
337
|
+
log(`Removing ZIP file "${zipFileUrl}" from ${ZIP_FILE_PATH}`);
|
336
338
|
await (0, _downloadInstallFiles.deleteFile)(ZIP_FILE_PATH); // Extract each page zip and upload their content's to S3
|
337
339
|
|
338
340
|
for (let i = 0; i < zipFilePaths.length; i++) {
|
@@ -382,16 +384,13 @@ async function deleteS3Folder(key) {
|
|
382
384
|
}
|
383
385
|
|
384
386
|
const response = await _s3Stream.s3Stream.listObject(key);
|
385
|
-
const keys = response.Contents.map(c => c.Key);
|
387
|
+
const keys = (response.Contents || []).map(c => c.Key).filter(Boolean);
|
386
388
|
console.log(`Found ${keys.length} files.`);
|
387
389
|
const deleteFilePromises = keys.map(key => _s3Stream.s3Stream.deleteObject(key));
|
388
390
|
await Promise.all(deleteFilePromises);
|
389
391
|
console.log(`Successfully deleted ${deleteFilePromises.length} files.`);
|
390
|
-
}
|
391
|
-
|
392
|
-
const zeroPad = version => `${version}`.padStart(5, "0");
|
392
|
+
} // export const zeroPad = version => `${version}`.padStart(5, "0");
|
393
393
|
|
394
|
-
exports.zeroPad = zeroPad;
|
395
394
|
|
396
395
|
function initialStats(total) {
|
397
396
|
return {
|
@@ -419,6 +418,13 @@ function extractZipToDisk(exportFileZipPath) {
|
|
419
418
|
if (err) {
|
420
419
|
console.warn("ERROR: Failed to extract zip: ", exportFileZipPath, err);
|
421
420
|
reject(err);
|
421
|
+
return;
|
422
|
+
}
|
423
|
+
|
424
|
+
if (!zipFile) {
|
425
|
+
console.log("ERROR: Missing zip file resource for path: " + exportFileZipPath);
|
426
|
+
reject("Missing Zip File Resource.");
|
427
|
+
return;
|
422
428
|
}
|
423
429
|
|
424
430
|
console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);
|
@@ -445,6 +451,13 @@ function extractZipToDisk(exportFileZipPath) {
|
|
445
451
|
if (err) {
|
446
452
|
console.warn("ERROR: Failed to openReadStream for file: ", entry.fileName, err);
|
447
453
|
reject(err);
|
454
|
+
return;
|
455
|
+
}
|
456
|
+
|
457
|
+
if (!readStream) {
|
458
|
+
console.log("ERROR: Missing Read Stream Resource when extracting to disk.");
|
459
|
+
reject("Missing Read Stream Resource.");
|
460
|
+
return;
|
448
461
|
}
|
449
462
|
|
450
463
|
const filePath = _path.default.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);
|
@@ -453,7 +466,9 @@ function extractZipToDisk(exportFileZipPath) {
|
|
453
466
|
pageZipFilePaths.push(filePath);
|
454
467
|
zipFile.readEntry();
|
455
468
|
});
|
456
|
-
streamPipeline(readStream, (0, _fs.createWriteStream)(filePath))
|
469
|
+
streamPipeline(readStream, (0, _fs.createWriteStream)(filePath)).catch(error => {
|
470
|
+
reject(error);
|
471
|
+
});
|
457
472
|
});
|
458
473
|
}
|
459
474
|
});
|
@@ -478,6 +493,13 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
|
|
478
493
|
if (err) {
|
479
494
|
console.warn("ERROR: Failed to extract zip: ", pageDataZipFilePath, err);
|
480
495
|
reject(err);
|
496
|
+
return;
|
497
|
+
}
|
498
|
+
|
499
|
+
if (!zipFile) {
|
500
|
+
console.log("ERROR: Probably failed to extract zip: " + pageDataZipFilePath);
|
501
|
+
reject("Missing Zip File Resource.");
|
502
|
+
return;
|
481
503
|
}
|
482
504
|
|
483
505
|
console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);
|
@@ -509,6 +531,13 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
|
|
509
531
|
if (err) {
|
510
532
|
console.warn("ERROR: Failed while performing [openReadStream] for file: ", entry.fileName, err);
|
511
533
|
reject(err);
|
534
|
+
return;
|
535
|
+
}
|
536
|
+
|
537
|
+
if (!readStream) {
|
538
|
+
console.log("ERROR: Missing Read Stream while importing pages.");
|
539
|
+
reject("Missing Read Strea Resource.");
|
540
|
+
return;
|
512
541
|
}
|
513
542
|
|
514
543
|
readStream.on("end", function () {
|
@@ -530,6 +559,8 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
|
|
530
559
|
|
531
560
|
streamPipeline(readStream, streamPassThrough).then(() => {
|
532
561
|
fileUploadPromises.push(promise);
|
562
|
+
}).catch(error => {
|
563
|
+
reject(error);
|
533
564
|
});
|
534
565
|
});
|
535
566
|
}
|
@@ -0,0 +1 @@
|
|
1
|
+
{"version":3,"names":["streamPipeline","promisify","pipeline","INSTALL_DIR","INSTALL_EXTRACT_DIR","path","join","FILES_COUNT_IN_EACH_BATCH","updateImageInPageSettings","params","settings","fileIdToKeyMap","srcPrefix","newSettings","srcPrefixWithoutTrailingSlash","endsWith","slice","dotProp","get","set","general","image","id","social","updateFilesInPageData","data","Array","isArray","i","length","element","tuple","Object","entries","key","value","has","name","src","uploadPageAssets","context","filesData","fileUploadsData","Map","process","env","NODE_ENV","console","log","fileKeyToFileMap","file","type","fileUploadResults","uploadFilesFromS3","oldKeyToNewKeyMap","assets","createFilesInput","map","uploadResult","newKey","Key","getOldFileKey","size","meta","tags","filter","Boolean","createFilesPromises","createFilesInputChunks","chunk","createFilesInputChunk","push","fileManager","files","createFilesInBatch","Promise","all","importPage","pageKey","PAGE_EXTRACT_DIR","ensureDirSync","pageDataFileKey","PAGE_DATA_FILE_PATH","basename","resolve","reject","s3Stream","readStream","on","pipe","createWriteStream","page","loadJson","getSettings","content","deleteFile","deleteS3Folder","dirname","oldKeysForAssets","keys","promises","oldKey","tempNewKey","fileMetaData","uniqueId","streamPassThrough","streamPassThroughUploadPromise","promise","writeStream","rest","split","e","FILE_CONTENT_TYPE","getFileNameWithoutExt","fileName","replace","extname","readExtractAndUploadZipFileContents","zipFileUrl","pageImportDataList","zipFileName","response","fetch","ok","WebinyError","statusText","body","uniquePath","ZIP_FILE_PATH","zipFilePaths","extractZipToDisk","currentPath","dataMap","extractZipAndUploadToS3","ASSETS_DIR_NAME","preparePageDataDirMap","filePath","isAsset","listObject","Contents","c","deleteFilePromises","deleteObject","initialStats","total","PageImportExportTaskStatus","PENDING","PROCESSING","COMPLETED","FAILED","exportFileZipPath","pageZipFilePaths","uniqueFolderNameForExport","EXPORT_FILE_EXTRACTION_PATH","yauzl","open","lazyEntries","err","zipFile","warn","info","entryCount","readEntry","entry","test","openReadStream","catch","error","pageDataZipFilePath","filePaths","fileUploadPromises","uniquePageKey","then","res","forEach","r"],"sources":["utils.ts"],"sourcesContent":["import uniqueId from \"uniqid\";\nimport S3 from \"aws-sdk/clients/s3\";\nimport dotProp from \"dot-prop-immutable\";\nimport { createWriteStream } from \"fs\";\nimport { ensureDirSync } from \"fs-extra\";\nimport { promisify } from \"util\";\nimport { pipeline } from \"stream\";\nimport fetch from \"node-fetch\";\nimport path from \"path\";\nimport yauzl from \"yauzl\";\nimport chunk from \"lodash/chunk\";\nimport loadJson from \"load-json-file\";\nimport { FileInput } from \"@webiny/api-file-manager/types\";\nimport WebinyError from \"@webiny/error\";\nimport { deleteFile } from \"@webiny/api-page-builder/graphql/crud/install/utils/downloadInstallFiles\";\nimport { File, PageImportExportTaskStatus } from \"~/types\";\nimport { PbPageImportExportContext } from \"~/graphql/types\";\nimport { s3Stream } from \"~/exportPages/s3Stream\";\nimport { ExportedPageData } from \"~/exportPages/utils\";\nimport { PageSettings } from \"@webiny/api-page-builder/types\";\n\ninterface FileItem extends File {\n key: string;\n type: string;\n name: string;\n size: number;\n meta: Record<string, any>;\n tags: string[];\n}\n\nconst streamPipeline = promisify(pipeline);\n\nconst INSTALL_DIR = \"/tmp\";\nconst INSTALL_EXTRACT_DIR = path.join(INSTALL_DIR, \"apiPageBuilderImportPage\");\nconst FILES_COUNT_IN_EACH_BATCH = 15;\n\ninterface UpdateFilesInPageDataParams {\n data: Record<string, any>;\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n}\n\ninterface UpdateImageInPageSettingsParams {\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n settings: PageSettings;\n}\n\nfunction updateImageInPageSettings(\n params: UpdateImageInPageSettingsParams\n): UpdateImageInPageSettingsParams[\"settings\"] {\n const { settings, fileIdToKeyMap, srcPrefix } = params;\n let newSettings = settings;\n\n const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith(\"/\")\n ? srcPrefix.slice(0, -1)\n : srcPrefix;\n\n if (dotProp.get(newSettings, \"general.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"general.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.general?.image?.id || \"\"\n )}`\n );\n }\n if (dotProp.get(newSettings, \"social.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"social.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.social?.image?.id || \"\"\n )}`\n );\n }\n\n return newSettings;\n}\n\nfunction updateFilesInPageData({ data, fileIdToKeyMap, srcPrefix }: UpdateFilesInPageDataParams) {\n // BASE CASE: Termination point\n if (!data || typeof data !== \"object\") {\n return;\n }\n // Recursively call function if data is array\n if (Array.isArray(data)) {\n for (let i = 0; i < data.length; i++) {\n const element = data[i];\n updateFilesInPageData({ data: element, fileIdToKeyMap, srcPrefix });\n }\n return;\n }\n // Main logic\n const tuple = Object.entries(data);\n for (let i = 0; i < tuple.length; i++) {\n const [key, value] = tuple[i];\n\n if (key === \"file\" && value && fileIdToKeyMap.has(value.id)) {\n value.key = fileIdToKeyMap.get(value.id);\n value.name = fileIdToKeyMap.get(value.id);\n value.src = `${srcPrefix}${srcPrefix.endsWith(\"/\") ? \"\" : \"/\"}${fileIdToKeyMap.get(\n value.id\n )}`;\n } else {\n updateFilesInPageData({ data: value, srcPrefix, fileIdToKeyMap });\n }\n }\n}\n\ninterface UploadPageAssetsParams {\n context: PbPageImportExportContext;\n filesData: FileItem[];\n fileUploadsData: FileUploadsData;\n}\n\ninterface UploadPageAssetsReturnType {\n fileIdToKeyMap: Map<string, string>;\n}\n\nexport const uploadPageAssets = async (\n params: UploadPageAssetsParams\n): Promise<UploadPageAssetsReturnType> => {\n const { context, filesData, fileUploadsData } = params;\n // Save uploaded file key against static id for later use.\n const fileIdToKeyMap = new Map<string, string>();\n /**\n * This function contains logic of file download from S3.\n * Current we're not mocking zip file download from S3 in tests at the moment.\n * So, we're manually mocking it in case of test just by returning an empty object.\n */\n if (process.env.NODE_ENV === \"test\") {\n return {\n fileIdToKeyMap\n };\n }\n console.log(\"INSIDE uploadPageAssets\");\n\n // Save files meta data against old key for later use.\n const fileKeyToFileMap = new Map<string, FileItem>();\n // Initialize maps.\n for (let i = 0; i < filesData.length; i++) {\n const file = filesData[i];\n fileKeyToFileMap.set(file.key, file);\n\n // Initialize the value\n fileIdToKeyMap.set(file.id, file.type);\n }\n\n const fileUploadResults = await uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap: fileUploadsData.assets\n });\n\n // Create files in File Manager\n const createFilesInput = fileUploadResults\n .map((uploadResult): FileInput | null => {\n const newKey = uploadResult.Key;\n const file = fileKeyToFileMap.get(getOldFileKey(newKey));\n if (!file) {\n return null;\n }\n\n // Update the file map with newly uploaded file.\n fileIdToKeyMap.set(file.id, newKey);\n\n return {\n key: newKey,\n name: file.name,\n size: file.size,\n type: file.type,\n meta: file.meta,\n tags: file.tags\n };\n })\n .filter(Boolean) as FileInput[];\n\n const createFilesPromises = [];\n // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).\n const createFilesInputChunks = chunk(createFilesInput, FILES_COUNT_IN_EACH_BATCH);\n for (let i = 0; i < createFilesInputChunks.length; i++) {\n const createFilesInputChunk = createFilesInputChunks[i];\n createFilesPromises.push(\n /*\n * We need to break down files into chunks because\n * `createFilesInBatch` operation has a limit on number of files it can handle at once.\n */\n context.fileManager.files.createFilesInBatch(createFilesInputChunk)\n );\n }\n\n await Promise.all(createFilesPromises);\n\n return {\n fileIdToKeyMap\n };\n};\n\ninterface FileUploadsData {\n data: string;\n assets: Record<string, string>;\n}\n\ninterface ImportPageParams {\n key: string;\n pageKey: string;\n context: PbPageImportExportContext;\n fileUploadsData: FileUploadsData;\n}\n\nexport async function importPage({\n pageKey,\n context,\n fileUploadsData\n}: ImportPageParams): Promise<ExportedPageData[\"page\"]> {\n const log = console.log;\n\n // Making Directory for page in which we're going to extract the page data file.\n const PAGE_EXTRACT_DIR = path.join(INSTALL_EXTRACT_DIR, pageKey);\n ensureDirSync(PAGE_EXTRACT_DIR);\n\n const pageDataFileKey = dotProp.get(fileUploadsData, `data`);\n const PAGE_DATA_FILE_PATH = path.join(PAGE_EXTRACT_DIR, path.basename(pageDataFileKey));\n\n log(`Downloading Page data file: ${pageDataFileKey} at \"${PAGE_DATA_FILE_PATH}\"`);\n // Download and save page data file in disk.\n await new Promise((resolve, reject) => {\n s3Stream\n .readStream(pageDataFileKey)\n .on(\"error\", reject)\n .pipe(createWriteStream(PAGE_DATA_FILE_PATH))\n .on(\"error\", reject)\n .on(\"finish\", resolve);\n });\n\n // Load the page data file from disk.\n log(`Load file ${pageDataFileKey}`);\n const { page, files } = await loadJson<ExportedPageData>(PAGE_DATA_FILE_PATH);\n\n // Only update page data if there are files.\n if (files && Array.isArray(files) && files.length > 0) {\n // Upload page assets.\n const { fileIdToKeyMap } = await uploadPageAssets({\n context,\n /**\n * TODO @ts-refactor @ashutosh figure out correct types.\n */\n // @ts-ignore\n filesData: files,\n fileUploadsData\n });\n\n const settings = await context.fileManager.settings.getSettings();\n\n const { srcPrefix = \"\" } = settings || {};\n updateFilesInPageData({\n data: page.content || {},\n fileIdToKeyMap,\n srcPrefix\n });\n\n page.settings = updateImageInPageSettings({\n settings: page.settings || {},\n fileIdToKeyMap,\n srcPrefix\n });\n }\n\n log(\"Removing Directory for page...\");\n await deleteFile(pageKey);\n\n log(`Remove page contents from S3...`);\n await deleteS3Folder(path.dirname(fileUploadsData.data));\n\n return page;\n}\n\ninterface UploadFilesFromZipParams {\n fileKeyToFileMap: Map<string, any>;\n oldKeyToNewKeyMap: Record<string, string>;\n}\n\nasync function uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap\n}: UploadFilesFromZipParams): Promise<S3.ManagedUpload.SendData[]> {\n const oldKeysForAssets = Object.keys(oldKeyToNewKeyMap);\n\n const promises = [];\n // Upload all assets.\n for (let i = 0; i < oldKeysForAssets.length; i++) {\n const oldKey = oldKeysForAssets[i];\n const tempNewKey = oldKeyToNewKeyMap[oldKey];\n\n // Read file.\n const readStream = s3Stream.readStream(tempNewKey);\n // Get file meta data.\n const fileMetaData = fileKeyToFileMap.get(oldKey);\n\n if (fileMetaData) {\n const newKey = uniqueId(\"\", `-${fileMetaData.key}`);\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, fileMetaData.type);\n readStream.pipe(streamPassThrough);\n promises.push(promise);\n\n console.log(`Successfully queued file \"${newKey}\"`);\n }\n }\n\n return Promise.all(promises);\n}\n\nfunction getOldFileKey(key: string) {\n /*\n * Because we know the naming convention, we can extract the old key from new key.\n */\n try {\n const [, ...rest] = key.split(\"-\");\n return rest.join(\"-\");\n } catch (e) {\n return key;\n }\n}\n\nconst FILE_CONTENT_TYPE = \"application/octet-stream\";\n\nfunction getFileNameWithoutExt(fileName: string): string {\n return path.basename(fileName).replace(path.extname(fileName), \"\");\n}\n\ninterface PageImportData {\n assets: Record<string, string>;\n data: string;\n key: string;\n}\n\n/**\n * Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.\n * @param zipFileUrl\n * @return PageImportData S3 file keys for all uploaded assets group by page.\n */\nexport async function readExtractAndUploadZipFileContents(\n zipFileUrl: string\n): Promise<PageImportData[]> {\n const log = console.log;\n const pageImportDataList = [];\n\n const zipFileName = path.basename(zipFileUrl).split(\"?\")[0];\n\n const response = await fetch(zipFileUrl);\n if (!response.ok) {\n throw new WebinyError(`Unable to downloading file: \"${zipFileUrl}\"`, response.statusText);\n }\n\n const readStream = response.body;\n\n const uniquePath = uniqueId(\"IMPORT_PAGES/\");\n // Read export file and download it in the disk\n const ZIP_FILE_PATH = path.join(INSTALL_DIR, zipFileName);\n\n const writeStream = createWriteStream(ZIP_FILE_PATH);\n await streamPipeline(readStream, writeStream);\n log(`Downloaded file \"${zipFileName}\" at ${ZIP_FILE_PATH}`);\n\n // Extract the downloaded zip file\n const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);\n\n log(`Removing ZIP file \"${zipFileUrl}\" from ${ZIP_FILE_PATH}`);\n await deleteFile(ZIP_FILE_PATH);\n\n // Extract each page zip and upload their content's to S3\n for (let i = 0; i < zipFilePaths.length; i++) {\n const currentPath = zipFilePaths[i];\n const dataMap = await extractZipAndUploadToS3(currentPath, uniquePath);\n pageImportDataList.push(dataMap);\n }\n log(\"Removing all ZIP files located at \", path.dirname(zipFilePaths[0]));\n await deleteFile(path.dirname(zipFilePaths[0]));\n\n return pageImportDataList;\n}\n\nconst ASSETS_DIR_NAME = \"/assets\";\n\nfunction preparePageDataDirMap({\n map,\n filePath,\n newKey\n}: {\n map: PageImportData;\n filePath: string;\n newKey: string;\n}): PageImportData {\n const dirname = path.dirname(filePath);\n const fileName = path.basename(filePath);\n /*\n * We want to use dot (.) as part of object key rather than creating nested object(s).\n * Also, the file name might contain dots in it beside the extension, so, we are escaping them all.\n */\n const oldKey = fileName.replace(/\\./g, \"\\\\.\");\n\n const isAsset = dirname.endsWith(ASSETS_DIR_NAME);\n\n if (isAsset) {\n map = dotProp.set(map, `assets.${oldKey}`, newKey);\n } else {\n // We only need to know the newKey for data file.\n map = dotProp.set(map, `data`, newKey);\n }\n\n return map;\n}\n\nasync function deleteS3Folder(key: string): Promise<void> {\n // Append trailing slash i.e \"/\" to key to make sure we only delete a specific folder.\n if (!key.endsWith(\"/\")) {\n key = `${key}/`;\n }\n\n const response = await s3Stream.listObject(key);\n const keys = (response.Contents || []).map(c => c.Key).filter(Boolean) as string[];\n console.log(`Found ${keys.length} files.`);\n\n const deleteFilePromises = keys.map(key => s3Stream.deleteObject(key));\n\n await Promise.all(deleteFilePromises);\n console.log(`Successfully deleted ${deleteFilePromises.length} files.`);\n}\n\n// export const zeroPad = version => `${version}`.padStart(5, \"0\");\n\nexport function initialStats(total: number) {\n return {\n [PageImportExportTaskStatus.PENDING]: total,\n [PageImportExportTaskStatus.PROCESSING]: 0,\n [PageImportExportTaskStatus.COMPLETED]: 0,\n [PageImportExportTaskStatus.FAILED]: 0,\n total\n };\n}\n\nfunction extractZipToDisk(exportFileZipPath: string): Promise<string[]> {\n return new Promise((resolve, reject) => {\n const pageZipFilePaths: string[] = [];\n const uniqueFolderNameForExport = getFileNameWithoutExt(exportFileZipPath);\n const EXPORT_FILE_EXTRACTION_PATH = path.join(INSTALL_DIR, uniqueFolderNameForExport);\n // Make sure DIR exists\n ensureDirSync(EXPORT_FILE_EXTRACTION_PATH);\n\n yauzl.open(exportFileZipPath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", exportFileZipPath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Missing zip file resource for path: \" + exportFileZipPath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn(\"ERROR: Failed on END event for file: \", exportFileZipPath, err);\n reject(err);\n }\n resolve(pageZipFilePaths);\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed to openReadStream for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\n \"ERROR: Missing Read Stream Resource when extracting to disk.\"\n );\n reject(\"Missing Read Stream Resource.\");\n return;\n }\n\n const filePath = path.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);\n\n readStream.on(\"end\", function () {\n pageZipFilePaths.push(filePath);\n zipFile.readEntry();\n });\n\n streamPipeline(readStream, createWriteStream(filePath)).catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n\nfunction extractZipAndUploadToS3(\n pageDataZipFilePath: string,\n uniquePath: string\n): Promise<PageImportData> {\n return new Promise((resolve, reject) => {\n const filePaths = [];\n const fileUploadPromises: Promise<S3.ManagedUpload.SendData>[] = [];\n const uniquePageKey = getFileNameWithoutExt(pageDataZipFilePath);\n let dataMap: PageImportData = {\n key: uniquePageKey,\n assets: {},\n data: \"\"\n };\n yauzl.open(pageDataZipFilePath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", pageDataZipFilePath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Probably failed to extract zip: \" + pageDataZipFilePath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn('ERROR: Failed on \"END\" for file: ', pageDataZipFilePath, err);\n reject(err);\n }\n\n Promise.all(fileUploadPromises).then(res => {\n res.forEach(r => {\n console.info(\"Done uploading... \", r);\n });\n resolve(dataMap);\n });\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed while performing [openReadStream] for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\"ERROR: Missing Read Stream while importing pages.\");\n reject(\"Missing Read Strea Resource.\");\n return;\n }\n readStream.on(\"end\", function () {\n filePaths.push(entry.fileName);\n zipFile.readEntry();\n });\n\n const newKey = `${uniquePath}/${uniquePageKey}/${entry.fileName}`;\n // Modify in place\n dataMap = preparePageDataDirMap({\n map: dataMap,\n filePath: entry.fileName,\n newKey\n });\n\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, FILE_CONTENT_TYPE);\n\n streamPipeline(readStream, streamPassThrough)\n .then(() => {\n fileUploadPromises.push(promise);\n })\n .catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n"],"mappings":";;;;;;;;;;;;AAAA;;AAEA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AAEA;;AACA;;AACA;;AAEA;;AAaA,MAAMA,cAAc,GAAG,IAAAC,eAAA,EAAUC,gBAAV,CAAvB;AAEA,MAAMC,WAAW,GAAG,MAApB;;AACA,MAAMC,mBAAmB,GAAGC,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuB,0BAAvB,CAA5B;;AACA,MAAMI,yBAAyB,GAAG,EAAlC;;AAcA,SAASC,yBAAT,CACIC,MADJ,EAE+C;EAC3C,MAAM;IAAEC,QAAF;IAAYC,cAAZ;IAA4BC;EAA5B,IAA0CH,MAAhD;EACA,IAAII,WAAW,GAAGH,QAAlB;EAEA,MAAMI,6BAA6B,GAAGF,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAChCH,SAAS,CAACI,KAAV,CAAgB,CAAhB,EAAmB,CAAC,CAApB,CADgC,GAEhCJ,SAFN;;EAIA,IAAIK,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,mBAAzB,CAAJ,EAAmD;IAAA;;IAC/CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,mBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,sBAAAR,QAAQ,CAACU,OAAT,iGAAkBC,KAAlB,gFAAyBC,EAAzB,KAA+B,EADC,CAElC,EALQ,CAAd;EAOH;;EACD,IAAIL,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,kBAAzB,CAAJ,EAAkD;IAAA;;IAC9CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,kBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,qBAAAR,QAAQ,CAACa,MAAT,+FAAiBF,KAAjB,gFAAwBC,EAAxB,KAA8B,EADE,CAElC,EALQ,CAAd;EAOH;;EAED,OAAOT,WAAP;AACH;;AAED,SAASW,qBAAT,CAA+B;EAAEC,IAAF;EAAQd,cAAR;EAAwBC;AAAxB,CAA/B,EAAiG;EAC7F;EACA,IAAI,CAACa,IAAD,IAAS,OAAOA,IAAP,KAAgB,QAA7B,EAAuC;IACnC;EACH,CAJ4F,CAK7F;;;EACA,IAAIC,KAAK,CAACC,OAAN,CAAcF,IAAd,CAAJ,EAAyB;IACrB,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGH,IAAI,CAACI,MAAzB,EAAiCD,CAAC,EAAlC,EAAsC;MAClC,MAAME,OAAO,GAAGL,IAAI,CAACG,CAAD,CAApB;MACAJ,qBAAqB,CAAC;QAAEC,IAAI,EAAEK,OAAR;QAAiBnB,cAAjB;QAAiCC;MAAjC,CAAD,CAArB;IACH;;IACD;EACH,CAZ4F,CAa7F;;;EACA,MAAMmB,KAAK,GAAGC,MAAM,CAACC,OAAP,CAAeR,IAAf,CAAd;;EACA,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGG,KAAK,CAACF,MAA1B,EAAkCD,CAAC,EAAnC,EAAuC;IACnC,MAAM,CAACM,GAAD,EAAMC,KAAN,IAAeJ,KAAK,CAACH,CAAD,CAA1B;;IAEA,IAAIM,GAAG,KAAK,MAAR,IAAkBC,KAAlB,IAA2BxB,cAAc,CAACyB,GAAf,CAAmBD,KAAK,CAACb,EAAzB,CAA/B,EAA6D;MACzDa,KAAK,CAACD,GAAN,GAAYvB,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAZ;MACAa,KAAK,CAACE,IAAN,GAAa1B,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAb;MACAa,KAAK,CAACG,GAAN,GAAa,GAAE1B,SAAU,GAAEA,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAA0B,EAA1B,GAA+B,GAAI,GAAEJ,cAAc,CAACO,GAAf,CAC5DiB,KAAK,CAACb,EADsD,CAE9D,EAFF;IAGH,CAND,MAMO;MACHE,qBAAqB,CAAC;QAAEC,IAAI,EAAEU,KAAR;QAAevB,SAAf;QAA0BD;MAA1B,CAAD,CAArB;IACH;EACJ;AACJ;;AAYM,MAAM4B,gBAAgB,GAAG,MAC5B9B,MAD4B,IAEU;EACtC,MAAM;IAAE+B,OAAF;IAAWC,SAAX;IAAsBC;EAAtB,IAA0CjC,MAAhD,CADsC,CAEtC;;EACA,MAAME,cAAc,GAAG,IAAIgC,GAAJ,EAAvB;EACA;AACJ;AACA;AACA;AACA;;EACI,IAAIC,OAAO,CAACC,GAAR,CAAYC,QAAZ,KAAyB,MAA7B,EAAqC;IACjC,OAAO;MACHnC;IADG,CAAP;EAGH;;EACDoC,OAAO,CAACC,GAAR,CAAY,yBAAZ,EAdsC,CAgBtC;;EACA,MAAMC,gBAAgB,GAAG,IAAIN,GAAJ,EAAzB,CAjBsC,CAkBtC;;EACA,KAAK,IAAIf,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGa,SAAS,CAACZ,MAA9B,EAAsCD,CAAC,EAAvC,EAA2C;IACvC,MAAMsB,IAAI,GAAGT,SAAS,CAACb,CAAD,CAAtB;IACAqB,gBAAgB,CAAC9B,GAAjB,CAAqB+B,IAAI,CAAChB,GAA1B,EAA+BgB,IAA/B,EAFuC,CAIvC;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4B4B,IAAI,CAACC,IAAjC;EACH;;EAED,MAAMC,iBAAiB,GAAG,MAAMC,iBAAiB,CAAC;IAC9CJ,gBAD8C;IAE9CK,iBAAiB,EAAEZ,eAAe,CAACa;EAFW,CAAD,CAAjD,CA3BsC,CAgCtC;;EACA,MAAMC,gBAAgB,GAAGJ,iBAAiB,CACrCK,GADoB,CACfC,YAAD,IAAoC;IACrC,MAAMC,MAAM,GAAGD,YAAY,CAACE,GAA5B;IACA,MAAMV,IAAI,GAAGD,gBAAgB,CAAC/B,GAAjB,CAAqB2C,aAAa,CAACF,MAAD,CAAlC,CAAb;;IACA,IAAI,CAACT,IAAL,EAAW;MACP,OAAO,IAAP;IACH,CALoC,CAOrC;;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4BqC,MAA5B;IAEA,OAAO;MACHzB,GAAG,EAAEyB,MADF;MAEHtB,IAAI,EAAEa,IAAI,CAACb,IAFR;MAGHyB,IAAI,EAAEZ,IAAI,CAACY,IAHR;MAIHX,IAAI,EAAED,IAAI,CAACC,IAJR;MAKHY,IAAI,EAAEb,IAAI,CAACa,IALR;MAMHC,IAAI,EAAEd,IAAI,CAACc;IANR,CAAP;EAQH,CAnBoB,EAoBpBC,MApBoB,CAoBbC,OApBa,CAAzB;EAsBA,MAAMC,mBAAmB,GAAG,EAA5B,CAvDsC,CAwDtC;;EACA,MAAMC,sBAAsB,GAAG,IAAAC,cAAA,EAAMb,gBAAN,EAAwBjD,yBAAxB,CAA/B;;EACA,KAAK,IAAIqB,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGwC,sBAAsB,CAACvC,MAA3C,EAAmDD,CAAC,EAApD,EAAwD;IACpD,MAAM0C,qBAAqB,GAAGF,sBAAsB,CAACxC,CAAD,CAApD;IACAuC,mBAAmB,CAACI,IAApB;IACI;AACZ;AACA;AACA;IACY/B,OAAO,CAACgC,WAAR,CAAoBC,KAApB,CAA0BC,kBAA1B,CAA6CJ,qBAA7C,CALJ;EAOH;;EAED,MAAMK,OAAO,CAACC,GAAR,CAAYT,mBAAZ,CAAN;EAEA,OAAO;IACHxD;EADG,CAAP;AAGH,CA5EM;;;;AA0FA,eAAekE,UAAf,CAA0B;EAC7BC,OAD6B;EAE7BtC,OAF6B;EAG7BE;AAH6B,CAA1B,EAIiD;EACpD,MAAMM,GAAG,GAAGD,OAAO,CAACC,GAApB,CADoD,CAGpD;;EACA,MAAM+B,gBAAgB,GAAG1E,aAAA,CAAKC,IAAL,CAAUF,mBAAV,EAA+B0E,OAA/B,CAAzB;;EACA,IAAAE,sBAAA,EAAcD,gBAAd;;EAEA,MAAME,eAAe,GAAGhE,yBAAA,CAAQC,GAAR,CAAYwB,eAAZ,EAA8B,MAA9B,CAAxB;;EACA,MAAMwC,mBAAmB,GAAG7E,aAAA,CAAKC,IAAL,CAAUyE,gBAAV,EAA4B1E,aAAA,CAAK8E,QAAL,CAAcF,eAAd,CAA5B,CAA5B;;EAEAjC,GAAG,CAAE,+BAA8BiC,eAAgB,QAAOC,mBAAoB,GAA3E,CAAH,CAVoD,CAWpD;;EACA,MAAM,IAAIP,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACnCC,kBAAA,CACKC,UADL,CACgBN,eADhB,EAEKO,EAFL,CAEQ,OAFR,EAEiBH,MAFjB,EAGKI,IAHL,CAGU,IAAAC,qBAAA,EAAkBR,mBAAlB,CAHV,EAIKM,EAJL,CAIQ,OAJR,EAIiBH,MAJjB,EAKKG,EALL,CAKQ,QALR,EAKkBJ,OALlB;EAMH,CAPK,CAAN,CAZoD,CAqBpD;;EACApC,GAAG,CAAE,aAAYiC,eAAgB,EAA9B,CAAH;EACA,MAAM;IAAEU,IAAF;IAAQlB;EAAR,IAAkB,MAAM,IAAAmB,qBAAA,EAA2BV,mBAA3B,CAA9B,CAvBoD,CAyBpD;;EACA,IAAIT,KAAK,IAAI/C,KAAK,CAACC,OAAN,CAAc8C,KAAd,CAAT,IAAiCA,KAAK,CAAC5C,MAAN,GAAe,CAApD,EAAuD;IACnD;IACA,MAAM;MAAElB;IAAF,IAAqB,MAAM4B,gBAAgB,CAAC;MAC9CC,OAD8C;;MAE9C;AACZ;AACA;MACY;MACAC,SAAS,EAAEgC,KANmC;MAO9C/B;IAP8C,CAAD,CAAjD;IAUA,MAAMhC,QAAQ,GAAG,MAAM8B,OAAO,CAACgC,WAAR,CAAoB9D,QAApB,CAA6BmF,WAA7B,EAAvB;IAEA,MAAM;MAAEjF,SAAS,GAAG;IAAd,IAAqBF,QAAQ,IAAI,EAAvC;IACAc,qBAAqB,CAAC;MAClBC,IAAI,EAAEkE,IAAI,CAACG,OAAL,IAAgB,EADJ;MAElBnF,cAFkB;MAGlBC;IAHkB,CAAD,CAArB;IAMA+E,IAAI,CAACjF,QAAL,GAAgBF,yBAAyB,CAAC;MACtCE,QAAQ,EAAEiF,IAAI,CAACjF,QAAL,IAAiB,EADW;MAEtCC,cAFsC;MAGtCC;IAHsC,CAAD,CAAzC;EAKH;;EAEDoC,GAAG,CAAC,gCAAD,CAAH;EACA,MAAM,IAAA+C,gCAAA,EAAWjB,OAAX,CAAN;EAEA9B,GAAG,CAAE,iCAAF,CAAH;EACA,MAAMgD,cAAc,CAAC3F,aAAA,CAAK4F,OAAL,CAAavD,eAAe,CAACjB,IAA7B,CAAD,CAApB;EAEA,OAAOkE,IAAP;AACH;;AAOD,eAAetC,iBAAf,CAAiC;EAC7BJ,gBAD6B;EAE7BK;AAF6B,CAAjC,EAGmE;EAC/D,MAAM4C,gBAAgB,GAAGlE,MAAM,CAACmE,IAAP,CAAY7C,iBAAZ,CAAzB;EAEA,MAAM8C,QAAQ,GAAG,EAAjB,CAH+D,CAI/D;;EACA,KAAK,IAAIxE,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGsE,gBAAgB,CAACrE,MAArC,EAA6CD,CAAC,EAA9C,EAAkD;IAC9C,MAAMyE,MAAM,GAAGH,gBAAgB,CAACtE,CAAD,CAA/B;IACA,MAAM0E,UAAU,GAAGhD,iBAAiB,CAAC+C,MAAD,CAApC,CAF8C,CAI9C;;IACA,MAAMd,UAAU,GAAGD,kBAAA,CAASC,UAAT,CAAoBe,UAApB,CAAnB,CAL8C,CAM9C;;;IACA,MAAMC,YAAY,GAAGtD,gBAAgB,CAAC/B,GAAjB,CAAqBmF,MAArB,CAArB;;IAEA,IAAIE,YAAJ,EAAkB;MACd,MAAM5C,MAAM,GAAG,IAAA6C,eAAA,EAAS,EAAT,EAAc,IAAGD,YAAY,CAACrE,GAAI,EAAlC,CAAf;;MACA,MAAM;QAAEuE,iBAAF;QAAqBC,8BAA8B,EAAEC;MAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6B4C,YAAY,CAACpD,IAA1C,CADJ;;MAEAoC,UAAU,CAACE,IAAX,CAAgBgB,iBAAhB;MACAL,QAAQ,CAAC7B,IAAT,CAAcoC,OAAd;MAEA5D,OAAO,CAACC,GAAR,CAAa,6BAA4BW,MAAO,GAAhD;IACH;EACJ;;EAED,OAAOgB,OAAO,CAACC,GAAR,CAAYwB,QAAZ,CAAP;AACH;;AAED,SAASvC,aAAT,CAAuB3B,GAAvB,EAAoC;EAChC;AACJ;AACA;EACI,IAAI;IACA,MAAM,GAAG,GAAG2E,IAAN,IAAc3E,GAAG,CAAC4E,KAAJ,CAAU,GAAV,CAApB;IACA,OAAOD,IAAI,CAACvG,IAAL,CAAU,GAAV,CAAP;EACH,CAHD,CAGE,OAAOyG,CAAP,EAAU;IACR,OAAO7E,GAAP;EACH;AACJ;;AAED,MAAM8E,iBAAiB,GAAG,0BAA1B;;AAEA,SAASC,qBAAT,CAA+BC,QAA/B,EAAyD;EACrD,OAAO7G,aAAA,CAAK8E,QAAL,CAAc+B,QAAd,EAAwBC,OAAxB,CAAgC9G,aAAA,CAAK+G,OAAL,CAAaF,QAAb,CAAhC,EAAwD,EAAxD,CAAP;AACH;;AAQD;AACA;AACA;AACA;AACA;AACO,eAAeG,mCAAf,CACHC,UADG,EAEsB;EACzB,MAAMtE,GAAG,GAAGD,OAAO,CAACC,GAApB;EACA,MAAMuE,kBAAkB,GAAG,EAA3B;;EAEA,MAAMC,WAAW,GAAGnH,aAAA,CAAK8E,QAAL,CAAcmC,UAAd,EAA0BR,KAA1B,CAAgC,GAAhC,EAAqC,CAArC,CAApB;;EAEA,MAAMW,QAAQ,GAAG,MAAM,IAAAC,kBAAA,EAAMJ,UAAN,CAAvB;;EACA,IAAI,CAACG,QAAQ,CAACE,EAAd,EAAkB;IACd,MAAM,IAAIC,cAAJ,CAAiB,gCAA+BN,UAAW,GAA3D,EAA+DG,QAAQ,CAACI,UAAxE,CAAN;EACH;;EAED,MAAMtC,UAAU,GAAGkC,QAAQ,CAACK,IAA5B;EAEA,MAAMC,UAAU,GAAG,IAAAvB,eAAA,EAAS,eAAT,CAAnB,CAbyB,CAczB;;EACA,MAAMwB,aAAa,GAAG3H,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBqH,WAAvB,CAAtB;;EAEA,MAAMZ,WAAW,GAAG,IAAAlB,qBAAA,EAAkBsC,aAAlB,CAApB;EACA,MAAMhI,cAAc,CAACuF,UAAD,EAAaqB,WAAb,CAApB;EACA5D,GAAG,CAAE,oBAAmBwE,WAAY,QAAOQ,aAAc,EAAtD,CAAH,CAnByB,CAqBzB;;EACA,MAAMC,YAAY,GAAG,MAAMC,gBAAgB,CAACF,aAAD,CAA3C;EAEAhF,GAAG,CAAE,sBAAqBsE,UAAW,UAASU,aAAc,EAAzD,CAAH;EACA,MAAM,IAAAjC,gCAAA,EAAWiC,aAAX,CAAN,CAzByB,CA2BzB;;EACA,KAAK,IAAIpG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGqG,YAAY,CAACpG,MAAjC,EAAyCD,CAAC,EAA1C,EAA8C;IAC1C,MAAMuG,WAAW,GAAGF,YAAY,CAACrG,CAAD,CAAhC;IACA,MAAMwG,OAAO,GAAG,MAAMC,uBAAuB,CAACF,WAAD,EAAcJ,UAAd,CAA7C;IACAR,kBAAkB,CAAChD,IAAnB,CAAwB6D,OAAxB;EACH;;EACDpF,GAAG,CAAC,oCAAD,EAAuC3C,aAAA,CAAK4F,OAAL,CAAagC,YAAY,CAAC,CAAD,CAAzB,CAAvC,CAAH;EACA,MAAM,IAAAlC,gCAAA,EAAW1F,aAAA,CAAK4F,OAAL,CAAagC,YAAY,CAAC,CAAD,CAAzB,CAAX,CAAN;EAEA,OAAOV,kBAAP;AACH;;AAED,MAAMe,eAAe,GAAG,SAAxB;;AAEA,SAASC,qBAAT,CAA+B;EAC3B9E,GAD2B;EAE3B+E,QAF2B;EAG3B7E;AAH2B,CAA/B,EAQmB;EACf,MAAMsC,OAAO,GAAG5F,aAAA,CAAK4F,OAAL,CAAauC,QAAb,CAAhB;;EACA,MAAMtB,QAAQ,GAAG7G,aAAA,CAAK8E,QAAL,CAAcqD,QAAd,CAAjB;EACA;AACJ;AACA;AACA;;;EACI,MAAMnC,MAAM,GAAGa,QAAQ,CAACC,OAAT,CAAiB,KAAjB,EAAwB,KAAxB,CAAf;EAEA,MAAMsB,OAAO,GAAGxC,OAAO,CAAClF,QAAR,CAAiBuH,eAAjB,CAAhB;;EAEA,IAAIG,OAAJ,EAAa;IACThF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,UAAS4C,MAAO,EAAlC,EAAqC1C,MAArC,CAAN;EACH,CAFD,MAEO;IACH;IACAF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,MAAlB,EAAyBE,MAAzB,CAAN;EACH;;EAED,OAAOF,GAAP;AACH;;AAED,eAAeuC,cAAf,CAA8B9D,GAA9B,EAA0D;EACtD;EACA,IAAI,CAACA,GAAG,CAACnB,QAAJ,CAAa,GAAb,CAAL,EAAwB;IACpBmB,GAAG,GAAI,GAAEA,GAAI,GAAb;EACH;;EAED,MAAMuF,QAAQ,GAAG,MAAMnC,kBAAA,CAASoD,UAAT,CAAoBxG,GAApB,CAAvB;EACA,MAAMiE,IAAI,GAAG,CAACsB,QAAQ,CAACkB,QAAT,IAAqB,EAAtB,EAA0BlF,GAA1B,CAA8BmF,CAAC,IAAIA,CAAC,CAAChF,GAArC,EAA0CK,MAA1C,CAAiDC,OAAjD,CAAb;EACAnB,OAAO,CAACC,GAAR,CAAa,SAAQmD,IAAI,CAACtE,MAAO,SAAjC;EAEA,MAAMgH,kBAAkB,GAAG1C,IAAI,CAAC1C,GAAL,CAASvB,GAAG,IAAIoD,kBAAA,CAASwD,YAAT,CAAsB5G,GAAtB,CAAhB,CAA3B;EAEA,MAAMyC,OAAO,CAACC,GAAR,CAAYiE,kBAAZ,CAAN;EACA9F,OAAO,CAACC,GAAR,CAAa,wBAAuB6F,kBAAkB,CAAChH,MAAO,SAA9D;AACH,C,CAED;;;AAEO,SAASkH,YAAT,CAAsBC,KAAtB,EAAqC;EACxC,OAAO;IACH,CAACC,iCAAA,CAA2BC,OAA5B,GAAsCF,KADnC;IAEH,CAACC,iCAAA,CAA2BE,UAA5B,GAAyC,CAFtC;IAGH,CAACF,iCAAA,CAA2BG,SAA5B,GAAwC,CAHrC;IAIH,CAACH,iCAAA,CAA2BI,MAA5B,GAAqC,CAJlC;IAKHL;EALG,CAAP;AAOH;;AAED,SAASd,gBAAT,CAA0BoB,iBAA1B,EAAwE;EACpE,OAAO,IAAI3E,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMkE,gBAA0B,GAAG,EAAnC;IACA,MAAMC,yBAAyB,GAAGvC,qBAAqB,CAACqC,iBAAD,CAAvD;;IACA,MAAMG,2BAA2B,GAAGpJ,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBqJ,yBAAvB,CAApC,CAHoC,CAIpC;;;IACA,IAAAxE,sBAAA,EAAcyE,2BAAd;;IAEAC,cAAA,CAAMC,IAAN,CAAWL,iBAAX,EAA8B;MAAEM,WAAW,EAAE;IAAf,CAA9B,EAAqD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MACzE,IAAID,GAAJ,EAAS;QACL9G,OAAO,CAACgH,IAAR,CAAa,gCAAb,EAA+CT,iBAA/C,EAAkEO,GAAlE;QACAxE,MAAM,CAACwE,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACV/G,OAAO,CAACC,GAAR,CAAY,gDAAgDsG,iBAA5D;QACAjE,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MAEDtC,OAAO,CAACiH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MAEAH,OAAO,CAACtE,EAAR,CAAW,KAAX,EAAkB,UAAUqE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACL9G,OAAO,CAACgH,IAAR,CAAa,uCAAb,EAAsDT,iBAAtD,EAAyEO,GAAzE;UACAxE,MAAM,CAACwE,GAAD,CAAN;QACH;;QACDzE,OAAO,CAACmE,gBAAD,CAAP;MACH,CAND;MAQAO,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAACtE,EAAR,CAAW,OAAX,EAAoB,UAAU2E,KAAV,EAAiB;QACjCpH,OAAO,CAACiH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAetE,UAAf,EAA2B;YACrD,IAAIsE,GAAJ,EAAS;cACL9G,OAAO,CAACgH,IAAR,CACI,4CADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKAxE,MAAM,CAACwE,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAACtE,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CACI,8DADJ;cAGAqC,MAAM,CAAC,+BAAD,CAAN;cACA;YACH;;YAED,MAAMmD,QAAQ,GAAGnI,aAAA,CAAKC,IAAL,CAAUmJ,2BAAV,EAAuCU,KAAK,CAACjD,QAA7C,CAAjB;;YAEA3B,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7B+D,gBAAgB,CAAChF,IAAjB,CAAsBiE,QAAtB;cACAsB,OAAO,CAACI,SAAR;YACH,CAHD;YAKAlK,cAAc,CAACuF,UAAD,EAAa,IAAAG,qBAAA,EAAkB8C,QAAlB,CAAb,CAAd,CAAwD8B,KAAxD,CAA8DC,KAAK,IAAI;cACnElF,MAAM,CAACkF,KAAD,CAAN;YACH,CAFD;UAGH,CA5BD;QA6BH;MACJ,CAvCD;IAwCH,CAhED;EAiEH,CAxEM,CAAP;AAyEH;;AAED,SAASlC,uBAAT,CACImC,mBADJ,EAEIzC,UAFJ,EAG2B;EACvB,OAAO,IAAIpD,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMoF,SAAS,GAAG,EAAlB;IACA,MAAMC,kBAAwD,GAAG,EAAjE;IACA,MAAMC,aAAa,GAAG1D,qBAAqB,CAACuD,mBAAD,CAA3C;IACA,IAAIpC,OAAuB,GAAG;MAC1BlG,GAAG,EAAEyI,aADqB;MAE1BpH,MAAM,EAAE,EAFkB;MAG1B9B,IAAI,EAAE;IAHoB,CAA9B;;IAKAiI,cAAA,CAAMC,IAAN,CAAWa,mBAAX,EAAgC;MAAEZ,WAAW,EAAE;IAAf,CAAhC,EAAuD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MAC3E,IAAID,GAAJ,EAAS;QACL9G,OAAO,CAACgH,IAAR,CAAa,gCAAb,EAA+CS,mBAA/C,EAAoEX,GAApE;QACAxE,MAAM,CAACwE,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACV/G,OAAO,CAACC,GAAR,CAAY,4CAA4CwH,mBAAxD;QACAnF,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MACDtC,OAAO,CAACiH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MACAH,OAAO,CAACtE,EAAR,CAAW,KAAX,EAAkB,UAAUqE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACL9G,OAAO,CAACgH,IAAR,CAAa,mCAAb,EAAkDS,mBAAlD,EAAuEX,GAAvE;UACAxE,MAAM,CAACwE,GAAD,CAAN;QACH;;QAEDlF,OAAO,CAACC,GAAR,CAAY8F,kBAAZ,EAAgCE,IAAhC,CAAqCC,GAAG,IAAI;UACxCA,GAAG,CAACC,OAAJ,CAAYC,CAAC,IAAI;YACbhI,OAAO,CAACiH,IAAR,CAAa,oBAAb,EAAmCe,CAAnC;UACH,CAFD;UAGA3F,OAAO,CAACgD,OAAD,CAAP;QACH,CALD;MAMH,CAZD;MAcA0B,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAACtE,EAAR,CAAW,OAAX,EAAoB,UAAU2E,KAAV,EAAiB;QACjCpH,OAAO,CAACiH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAetE,UAAf,EAA2B;YACrD,IAAIsE,GAAJ,EAAS;cACL9G,OAAO,CAACgH,IAAR,CACI,4DADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKAxE,MAAM,CAACwE,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAACtE,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CAAY,mDAAZ;cACAqC,MAAM,CAAC,8BAAD,CAAN;cACA;YACH;;YACDE,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7BiF,SAAS,CAAClG,IAAV,CAAe4F,KAAK,CAACjD,QAArB;cACA4C,OAAO,CAACI,SAAR;YACH,CAHD;YAKA,MAAMvG,MAAM,GAAI,GAAEoE,UAAW,IAAG4C,aAAc,IAAGR,KAAK,CAACjD,QAAS,EAAhE,CApBqD,CAqBrD;;YACAkB,OAAO,GAAGG,qBAAqB,CAAC;cAC5B9E,GAAG,EAAE2E,OADuB;cAE5BI,QAAQ,EAAE2B,KAAK,CAACjD,QAFY;cAG5BvD;YAH4B,CAAD,CAA/B;;YAMA,MAAM;cAAE8C,iBAAF;cAAqBC,8BAA8B,EAAEC;YAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6BqD,iBAA7B,CADJ;;YAGAhH,cAAc,CAACuF,UAAD,EAAakB,iBAAb,CAAd,CACKmE,IADL,CACU,MAAM;cACRF,kBAAkB,CAACnG,IAAnB,CAAwBoC,OAAxB;YACH,CAHL,EAIK2D,KAJL,CAIWC,KAAK,IAAI;cACZlF,MAAM,CAACkF,KAAD,CAAN;YACH,CANL;UAOH,CAtCD;QAuCH;MACJ,CAjDD;IAkDH,CA9ED;EA+EH,CAxFM,CAAP;AAyFH"}
|
@@ -0,0 +1 @@
|
|
1
|
+
{"version":3,"names":["mockSecurity","identity","context","security","disableAuthorization","setIdentity"],"sources":["mockSecurity.ts"],"sourcesContent":["import { SecurityContext, SecurityIdentity } from \"@webiny/api-security/types\";\n\nexport const mockSecurity = (identity: SecurityIdentity, context: SecurityContext) => {\n context.security.disableAuthorization();\n context.security.setIdentity(identity);\n};\n"],"mappings":";;;;;;;AAEO,MAAMA,YAAY,GAAG,CAACC,QAAD,EAA6BC,OAA7B,KAA0D;EAClFA,OAAO,CAACC,QAAR,CAAiBC,oBAAjB;EACAF,OAAO,CAACC,QAAR,CAAiBE,WAAjB,CAA6BJ,QAA7B;AACH,CAHM"}
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@webiny/api-page-builder-import-export",
|
3
|
-
"version": "0.0.0-
|
3
|
+
"version": "0.0.0-unstable.1e66d121db",
|
4
4
|
"main": "index.js",
|
5
5
|
"keywords": [
|
6
6
|
"pbie:base"
|
@@ -14,52 +14,54 @@
|
|
14
14
|
"author": "Webiny Ltd",
|
15
15
|
"license": "MIT",
|
16
16
|
"dependencies": {
|
17
|
-
"@babel/runtime": "7.
|
17
|
+
"@babel/runtime": "7.19.0",
|
18
18
|
"@commodo/fields": "1.1.2-beta.20",
|
19
|
-
"@webiny/api
|
20
|
-
"@webiny/api-
|
21
|
-
"@webiny/api-
|
22
|
-
"@webiny/
|
23
|
-
"@webiny/
|
24
|
-
"@webiny/handler
|
25
|
-
"@webiny/handler-aws": "0.0.0-
|
26
|
-
"@webiny/handler-graphql": "0.0.0-
|
27
|
-
"@webiny/
|
28
|
-
"
|
19
|
+
"@webiny/api": "0.0.0-unstable.1e66d121db",
|
20
|
+
"@webiny/api-file-manager": "0.0.0-unstable.1e66d121db",
|
21
|
+
"@webiny/api-page-builder": "0.0.0-unstable.1e66d121db",
|
22
|
+
"@webiny/api-security": "0.0.0-unstable.1e66d121db",
|
23
|
+
"@webiny/error": "0.0.0-unstable.1e66d121db",
|
24
|
+
"@webiny/handler": "0.0.0-unstable.1e66d121db",
|
25
|
+
"@webiny/handler-aws": "0.0.0-unstable.1e66d121db",
|
26
|
+
"@webiny/handler-graphql": "0.0.0-unstable.1e66d121db",
|
27
|
+
"@webiny/utils": "0.0.0-unstable.1e66d121db",
|
28
|
+
"@webiny/validation": "0.0.0-unstable.1e66d121db",
|
29
|
+
"archiver": "5.3.1",
|
29
30
|
"commodo-fields-object": "1.0.6",
|
30
31
|
"dot-prop-immutable": "2.1.1",
|
31
|
-
"fs-extra": "
|
32
|
+
"fs-extra": "9.1.0",
|
32
33
|
"load-json-file": "6.2.0",
|
33
34
|
"lodash": "4.17.21",
|
34
35
|
"mdbid": "1.0.0",
|
35
|
-
"node-fetch": "2.6.
|
36
|
+
"node-fetch": "2.6.7",
|
36
37
|
"stream": "0.0.2",
|
37
38
|
"uniqid": "5.4.0",
|
38
39
|
"yauzl": "2.10.0"
|
39
40
|
},
|
40
41
|
"devDependencies": {
|
41
|
-
"@babel/cli": "^7.
|
42
|
-
"@babel/core": "^7.
|
43
|
-
"@babel/plugin-proposal-export-default-from": "^7.
|
44
|
-
"@babel/preset-env": "^7.
|
45
|
-
"@babel/preset-typescript": "^7.
|
46
|
-
"@
|
47
|
-
"@
|
48
|
-
"@types/
|
49
|
-
"@webiny/api-dynamodb-to-elasticsearch": "^0.0.0-
|
50
|
-
"@webiny/api-file-manager-ddb-es": "^0.0.0-
|
51
|
-
"@webiny/api-i18n-ddb": "^0.0.0-
|
52
|
-
"@webiny/api-security-so-ddb": "^0.0.0-
|
53
|
-
"@webiny/api-tenancy": "^0.0.0-
|
54
|
-
"@webiny/api-tenancy-so-ddb": "^0.0.0-
|
55
|
-
"@webiny/
|
56
|
-
"@webiny/
|
57
|
-
"@webiny/
|
58
|
-
"
|
42
|
+
"@babel/cli": "^7.19.3",
|
43
|
+
"@babel/core": "^7.19.3",
|
44
|
+
"@babel/plugin-proposal-export-default-from": "^7.16.0",
|
45
|
+
"@babel/preset-env": "^7.19.4",
|
46
|
+
"@babel/preset-typescript": "^7.18.6",
|
47
|
+
"@types/archiver": "^5.3.1",
|
48
|
+
"@types/node-fetch": "^2.6.1",
|
49
|
+
"@types/yauzl": "^2.9.2",
|
50
|
+
"@webiny/api-dynamodb-to-elasticsearch": "^0.0.0-unstable.1e66d121db",
|
51
|
+
"@webiny/api-file-manager-ddb-es": "^0.0.0-unstable.1e66d121db",
|
52
|
+
"@webiny/api-i18n-ddb": "^0.0.0-unstable.1e66d121db",
|
53
|
+
"@webiny/api-security-so-ddb": "^0.0.0-unstable.1e66d121db",
|
54
|
+
"@webiny/api-tenancy": "^0.0.0-unstable.1e66d121db",
|
55
|
+
"@webiny/api-tenancy-so-ddb": "^0.0.0-unstable.1e66d121db",
|
56
|
+
"@webiny/api-wcp": "^0.0.0-unstable.1e66d121db",
|
57
|
+
"@webiny/cli": "^0.0.0-unstable.1e66d121db",
|
58
|
+
"@webiny/db": "^0.0.0-unstable.1e66d121db",
|
59
|
+
"@webiny/project-utils": "^0.0.0-unstable.1e66d121db",
|
60
|
+
"jest": "^28.1.0",
|
59
61
|
"jest-dynalite": "^3.2.0",
|
60
62
|
"rimraf": "^3.0.2",
|
61
63
|
"ttypescript": "^1.5.12",
|
62
|
-
"typescript": "
|
64
|
+
"typescript": "4.7.4"
|
63
65
|
},
|
64
66
|
"publishConfig": {
|
65
67
|
"access": "public",
|
@@ -76,5 +78,5 @@
|
|
76
78
|
]
|
77
79
|
}
|
78
80
|
},
|
79
|
-
"gitHead": "
|
81
|
+
"gitHead": "b670bf27c5039de1a2b0be764a09ba4cb94ad5e2"
|
80
82
|
}
|
package/types.d.ts
CHANGED
@@ -16,6 +16,11 @@ export interface PageImportExportTaskStats {
|
|
16
16
|
[PageImportExportTaskStatus.FAILED]: number;
|
17
17
|
total: number;
|
18
18
|
}
|
19
|
+
interface CreatedBy {
|
20
|
+
id: string;
|
21
|
+
type: string;
|
22
|
+
displayName: string | null;
|
23
|
+
}
|
19
24
|
export interface PageImportExportTask {
|
20
25
|
id: string;
|
21
26
|
parent: string;
|
@@ -25,18 +30,14 @@ export interface PageImportExportTask {
|
|
25
30
|
error: Record<string, any>;
|
26
31
|
input: Record<string, any>;
|
27
32
|
createdOn: string;
|
28
|
-
createdBy:
|
29
|
-
type: string;
|
30
|
-
id: string;
|
31
|
-
displayName: string;
|
32
|
-
};
|
33
|
+
createdBy: CreatedBy;
|
33
34
|
tenant: string;
|
34
35
|
locale: string;
|
35
36
|
}
|
36
|
-
export
|
37
|
+
export interface File {
|
37
38
|
id: string;
|
38
39
|
src: string;
|
39
|
-
}
|
40
|
+
}
|
40
41
|
export interface MetaResponse {
|
41
42
|
cursor: string | null;
|
42
43
|
totalCount: number;
|
package/types.js.map
ADDED
@@ -0,0 +1 @@
|
|
1
|
+
{"version":3,"names":["PageExportRevisionType","PageImportExportTaskStatus"],"sources":["types.ts"],"sourcesContent":["export * from \"./graphql/types\";\n\n// Entities.\nexport enum PageExportRevisionType {\n PUBLISHED = \"published\",\n LATEST = \"latest\"\n}\n\nexport enum PageImportExportTaskStatus {\n PENDING = \"pending\",\n PROCESSING = \"processing\",\n COMPLETED = \"completed\",\n FAILED = \"failed\"\n}\n\nexport interface PageImportExportTaskStats {\n [PageImportExportTaskStatus.PENDING]: number;\n [PageImportExportTaskStatus.PROCESSING]: number;\n [PageImportExportTaskStatus.COMPLETED]: number;\n [PageImportExportTaskStatus.FAILED]: number;\n total: number;\n}\n\ninterface CreatedBy {\n id: string;\n type: string;\n displayName: string | null;\n}\n\nexport interface PageImportExportTask {\n id: string;\n parent: string;\n status: PageImportExportTaskStatus;\n data: Record<string, any>;\n stats: PageImportExportTaskStats;\n error: Record<string, any>;\n input: Record<string, any>;\n createdOn: string;\n createdBy: CreatedBy;\n tenant: string;\n locale: string;\n}\n\nexport interface File {\n id: string;\n src: string;\n}\n\nexport interface MetaResponse {\n cursor: string | null;\n totalCount: number;\n hasMoreItems: boolean;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsGetParams {\n where: {\n id: string;\n tenant: string;\n locale: string;\n };\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsListParams {\n where: {\n tenant: string;\n locale: string;\n };\n sort?: string[];\n limit?: number;\n after?: string | null;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport type PageImportExportTaskStorageOperationsListResponse = [\n PageImportExportTask[],\n MetaResponse\n];\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsCreateParams {\n input: Record<string, any>;\n task: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsUpdateParams {\n input: Record<string, any>;\n original: PageImportExportTask;\n task: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsDeleteParams {\n task: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsGetSubTaskParams {\n where: {\n id: string;\n parent: string;\n tenant: string;\n locale: string;\n };\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsListSubTaskParams {\n where: {\n tenant: string;\n locale: string;\n parent: string;\n status: PageImportExportTaskStatus;\n createdBy?: string;\n };\n sort?: string[];\n limit?: number;\n after?: string | null;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport type PageImportExportTaskStorageOperationsListSubTaskResponse = [\n PageImportExportTask[],\n MetaResponse\n];\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsCreateSubTaskParams {\n input: Record<string, any>;\n subTask: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsUpdateSubTaskParams {\n input: Record<string, any>;\n original: PageImportExportTask;\n subTask: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperationsUpdateTaskStatsParams {\n input: {\n prevStatus: PageImportExportTaskStatus;\n nextStatus: PageImportExportTaskStatus;\n };\n original: PageImportExportTask;\n}\n\n/**\n * @category StorageOperations\n * @category PageImportExportTaskStorageOperations\n */\nexport interface PageImportExportTaskStorageOperations {\n /**\n * Get a single page import export task item by given params.\n */\n getTask(\n params: PageImportExportTaskStorageOperationsGetParams\n ): Promise<PageImportExportTask | null>;\n\n /**\n * Get all page import export tasks by given params.\n */\n listTasks(\n params: PageImportExportTaskStorageOperationsListParams\n ): Promise<PageImportExportTaskStorageOperationsListResponse>;\n\n createTask(\n params: PageImportExportTaskStorageOperationsCreateParams\n ): Promise<PageImportExportTask>;\n\n updateTask(\n params: PageImportExportTaskStorageOperationsUpdateParams\n ): Promise<PageImportExportTask>;\n\n deleteTask(\n params: PageImportExportTaskStorageOperationsDeleteParams\n ): Promise<PageImportExportTask>;\n\n updateTaskStats(\n params: PageImportExportTaskStorageOperationsUpdateTaskStatsParams\n ): Promise<PageImportExportTask>;\n\n /**\n * Get a single page import export sub-task item by given params.\n */\n getSubTask(\n params: PageImportExportTaskStorageOperationsGetSubTaskParams\n ): Promise<PageImportExportTask | null>;\n\n /**\n * Get all page import export sub-tasks by given params.\n */\n listSubTasks(\n params: PageImportExportTaskStorageOperationsListSubTaskParams\n ): Promise<PageImportExportTaskStorageOperationsListSubTaskResponse>;\n\n createSubTask(\n params: PageImportExportTaskStorageOperationsCreateSubTaskParams\n ): Promise<PageImportExportTask>;\n\n updateSubTask(\n params: PageImportExportTaskStorageOperationsUpdateSubTaskParams\n ): Promise<PageImportExportTask>;\n}\n"],"mappings":";;;;;;;;;;;AAAA;;AAAA;EAAA;EAAA;EAAA;EAAA;IAAA;IAAA;MAAA;IAAA;EAAA;AAAA;AAEA;IACYA,sB;;;WAAAA,sB;EAAAA,sB;EAAAA,sB;GAAAA,sB,sCAAAA,sB;;IAKAC,0B;;;WAAAA,0B;EAAAA,0B;EAAAA,0B;EAAAA,0B;EAAAA,0B;GAAAA,0B,0CAAAA,0B"}
|
package/importPages/client.js
DELETED
@@ -1,40 +0,0 @@
|
|
1
|
-
"use strict";
|
2
|
-
|
3
|
-
var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
|
4
|
-
|
5
|
-
Object.defineProperty(exports, "__esModule", {
|
6
|
-
value: true
|
7
|
-
});
|
8
|
-
exports.invokeHandlerClient = invokeHandlerClient;
|
9
|
-
|
10
|
-
var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/defineProperty"));
|
11
|
-
|
12
|
-
function ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); if (enumerableOnly) { symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; }); } keys.push.apply(keys, symbols); } return keys; }
|
13
|
-
|
14
|
-
function _objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = arguments[i] != null ? arguments[i] : {}; if (i % 2) { ownKeys(Object(source), true).forEach(function (key) { (0, _defineProperty2.default)(target, key, source[key]); }); } else if (Object.getOwnPropertyDescriptors) { Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)); } else { ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } } return target; }
|
15
|
-
|
16
|
-
async function invokeHandlerClient({
|
17
|
-
context,
|
18
|
-
name,
|
19
|
-
payload
|
20
|
-
}) {
|
21
|
-
/*
|
22
|
-
* Prepare "invocationArgs", we're hacking our wat here.
|
23
|
-
* They are necessary to setup the "context.pageBuilder" object among other things in IMPORT_PAGE_FUNCTION
|
24
|
-
*/
|
25
|
-
const {
|
26
|
-
request
|
27
|
-
} = context.http;
|
28
|
-
const invocationArgs = {
|
29
|
-
httpMethod: request.method,
|
30
|
-
body: request.body,
|
31
|
-
headers: request.headers,
|
32
|
-
cookies: request.cookies
|
33
|
-
}; // Invoke handler
|
34
|
-
|
35
|
-
await context.handlerClient.invoke({
|
36
|
-
name: name,
|
37
|
-
payload: _objectSpread(_objectSpread({}, payload), invocationArgs),
|
38
|
-
await: false
|
39
|
-
});
|
40
|
-
}
|