@webiny/api-page-builder-import-export 5.23.1 → 5.25.0-beta.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (41) hide show
  1. package/exportPages/combine/index.d.ts +9 -8
  2. package/exportPages/combine/index.js +21 -1
  3. package/exportPages/combine/index.js.map +1 -0
  4. package/exportPages/process/index.d.ts +9 -8
  5. package/exportPages/process/index.js +5 -5
  6. package/exportPages/process/index.js.map +1 -0
  7. package/exportPages/s3Stream.js.map +1 -0
  8. package/exportPages/utils.d.ts +1 -1
  9. package/exportPages/utils.js +1 -1
  10. package/exportPages/utils.js.map +1 -0
  11. package/exportPages/zipper.js.map +1 -0
  12. package/graphql/crud/pageImportExportTasks.crud.js +19 -9
  13. package/graphql/crud/pageImportExportTasks.crud.js.map +1 -0
  14. package/graphql/crud/pages.crud.js +6 -3
  15. package/graphql/crud/pages.crud.js.map +1 -0
  16. package/graphql/crud.d.ts +1 -1
  17. package/graphql/crud.js.map +1 -0
  18. package/graphql/graphql/pageImportExportTasks.gql.js.map +1 -0
  19. package/graphql/graphql/pages.gql.js +6 -0
  20. package/graphql/graphql/pages.gql.js.map +1 -0
  21. package/graphql/graphql/utils/resolve.d.ts +1 -1
  22. package/graphql/graphql/utils/resolve.js.map +1 -0
  23. package/graphql/graphql.js.map +1 -0
  24. package/graphql/index.d.ts +1 -1
  25. package/graphql/index.js.map +1 -0
  26. package/graphql/types.d.ts +2 -2
  27. package/graphql/types.js.map +1 -0
  28. package/importPages/client.js.map +1 -0
  29. package/importPages/create/index.d.ts +9 -8
  30. package/importPages/create/index.js +13 -1
  31. package/importPages/create/index.js.map +1 -0
  32. package/importPages/process/index.d.ts +9 -8
  33. package/importPages/process/index.js +3 -1
  34. package/importPages/process/index.js.map +1 -0
  35. package/importPages/utils.d.ts +15 -8
  36. package/importPages/utils.js +75 -30
  37. package/importPages/utils.js.map +1 -0
  38. package/mockSecurity.js.map +1 -0
  39. package/package.json +25 -22
  40. package/types.d.ts +8 -7
  41. package/types.js.map +1 -0
@@ -2,19 +2,20 @@ import { HandlerPlugin } from "@webiny/handler/types";
2
2
  import { ArgsContext } from "@webiny/handler-args/types";
3
3
  import { PageImportExportTask, PbPageImportExportContext } from "../../types";
4
4
  import { SecurityIdentity } from "@webiny/api-security/types";
5
- export declare type HandlerArgs = {
5
+ export interface HandlerArgs {
6
6
  category: string;
7
7
  zipFileKey?: string;
8
8
  zipFileUrl?: string;
9
9
  task: PageImportExportTask;
10
10
  identity: SecurityIdentity;
11
- };
12
- export declare type HandlerResponse = {
13
- data: string;
14
- error: {
15
- message: string;
16
- };
17
- };
11
+ }
12
+ interface HandlerResponseError {
13
+ message: string;
14
+ }
15
+ export interface HandlerResponse {
16
+ data: string | null;
17
+ error: HandlerResponseError | null;
18
+ }
18
19
  interface Configuration {
19
20
  handlers: {
20
21
  process: string;
@@ -13,6 +13,8 @@ var _client = require("../client");
13
13
 
14
14
  var _mockSecurity = require("../../mockSecurity");
15
15
 
16
+ var _utils2 = require("@webiny/utils");
17
+
16
18
  /**
17
19
  * Handles the import page workflow.
18
20
  */
@@ -35,6 +37,16 @@ var _default = configuration => ({
35
37
  zipFileUrl,
36
38
  identity
37
39
  } = args;
40
+
41
+ if (!zipFileKey && !zipFileUrl) {
42
+ return {
43
+ data: null,
44
+ error: {
45
+ message: "Missing zipFileKey and zipFileUrl parameters. One must be defined."
46
+ }
47
+ };
48
+ }
49
+
38
50
  (0, _mockSecurity.mockSecurity)(identity, context); // Step 1: Read the zip file
39
51
 
40
52
  const pageImportDataList = await (0, _utils.readExtractAndUploadZipFileContents)(zipFileKey || zipFileUrl); // Once we have map we can start processing each page
@@ -43,7 +55,7 @@ var _default = configuration => ({
43
55
  for (let i = 0; i < pageImportDataList.length; i++) {
44
56
  const pagesDirMap = pageImportDataList[i]; // Create sub task
45
57
 
46
- const subtask = await pageBuilder.pageImportExportTask.createSubTask(task.id, (0, _utils.zeroPad)(i + 1), {
58
+ const subtask = await pageBuilder.pageImportExportTask.createSubTask(task.id, (0, _utils2.zeroPad)(i + 1, 5), {
47
59
  status: _types.PageImportExportTaskStatus.PENDING,
48
60
  data: {
49
61
  pageKey: pagesDirMap.key,
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["index.ts"],"names":["configuration","type","handle","context","log","console","invocationArgs","args","pageBuilder","task","category","zipFileKey","zipFileUrl","identity","data","error","message","pageImportDataList","i","length","pagesDirMap","subtask","pageImportExportTask","createSubTask","id","status","PageImportExportTaskStatus","PENDING","pageKey","key","input","fileUploadsData","updateTask","PROCESSING","stats","name","handlers","process","payload","taskId","subTaskIndex","security","getIdentity","e","FAILED","code"],"mappings":";;;;;;;AAEA;;AAKA;;AACA;;AAGA;;AACA;;AAwBA;AACA;AACA;eAEIA,aADW,KAE2D;AACtEC,EAAAA,IAAI,EAAE,SADgE;;AAEtE,QAAMC,MAAN,CAAaC,OAAb,EAAgD;AAC5C,UAAMC,GAAG,GAAGC,OAAO,CAACD,GAApB;;AAEA,QAAI;AACAA,MAAAA,GAAG,CAAC,6BAAD,CAAH;AACA,YAAM;AAAEE,QAAAA,cAAc,EAAEC,IAAlB;AAAwBC,QAAAA;AAAxB,UAAwCL,OAA9C;AACA,YAAM;AAAEM,QAAAA,IAAF;AAAQC,QAAAA,QAAR;AAAkBC,QAAAA,UAAlB;AAA8BC,QAAAA,UAA9B;AAA0CC,QAAAA;AAA1C,UAAuDN,IAA7D;;AACA,UAAI,CAACI,UAAD,IAAe,CAACC,UAApB,EAAgC;AAC5B,eAAO;AACHE,UAAAA,IAAI,EAAE,IADH;AAEHC,UAAAA,KAAK,EAAE;AACHC,YAAAA,OAAO,EACH;AAFD;AAFJ,SAAP;AAOH;;AACD,sCAAaH,QAAb,EAAuBV,OAAvB,EAbA,CAcA;;AACA,YAAMc,kBAAkB,GAAG,MAAM,gDAC7BN,UAAU,IAAKC,UADc,CAAjC,CAfA,CAkBA;AAEA;;AACA,WAAK,IAAIM,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGD,kBAAkB,CAACE,MAAvC,EAA+CD,CAAC,EAAhD,EAAoD;AAChD,cAAME,WAAW,GAAGH,kBAAkB,CAACC,CAAD,CAAtC,CADgD,CAEhD;;AACA,cAAMG,OAAO,GAAG,MAAMb,WAAW,CAACc,oBAAZ,CAAiCC,aAAjC,CAClBd,IAAI,CAACe,EADa,EAElB,qBAAQN,CAAC,GAAG,CAAZ,EAAe,CAAf,CAFkB,EAGlB;AACIO,UAAAA,MAAM,EAAEC,kCAA2BC,OADvC;AAEIb,UAAAA,IAAI,EAAE;AACFc,YAAAA,OAAO,EAAER,WAAW,CAACS,GADnB;AAEFnB,YAAAA,QAFE;AAGFC,YAAAA,UAHE;AAIFC,YAAAA,UAJE;AAKFkB,YAAAA,KAAK,EAAE;AACHC,cAAAA,eAAe,EAAEX;AADd;AALL;AAFV,SAHkB,CAAtB;AAgBAhB,QAAAA,GAAG,CAAE,mBAAkBiB,OAAO,CAACG,EAAG,aAA/B,CAAH;AACH,OAzCD,CA0CA;;;AACA,YAAMhB,WAAW,CAACc,oBAAZ,CAAiCU,UAAjC,CAA4CvB,IAAI,CAACe,EAAjD,EAAqD;AACvDC,QAAAA,MAAM,EAAEC,kCAA2BO,UADoB;AAEvDC,QAAAA,KAAK,EAAE,yBAAajB,kBAAkB,CAACE,MAAhC;AAFgD,OAArD,CAAN;AAKA,YAAM,iCAAwC;AAC1ChB,QAAAA,OAD0C;AAE1CgC,QAAAA,IAAI,EAAEnC,aAAa,CAACoC,QAAd,CAAuBC,OAFa;AAG1CC,QAAAA,OAAO,EAAE;AACLC,UAAAA,MAAM,EAAE9B,IAAI,CAACe,EADR;AAEL;AACAgB,UAAAA,YAAY,EAAE,CAHT;AAIL3B,UAAAA,QAAQ,EAAEV,OAAO,CAACsC,QAAR,CAAiBC,WAAjB;AAJL;AAHiC,OAAxC,CAAN;AAUH,KA1DD,CA0DE,OAAOC,CAAP,EAAU;AACRvC,MAAAA,GAAG,CAAC,iCAAD,EAAoCuC,CAApC,CAAH;AAEA;AACZ;AACA;AACA;;AACY,YAAM;AAAErC,QAAAA,cAAc,EAAEC,IAAlB;AAAwBC,QAAAA;AAAxB,UAAwCL,OAA9C;AACA,YAAM;AAAEM,QAAAA;AAAF,UAAWF,IAAjB;AAEA,YAAMC,WAAW,CAACc,oBAAZ,CAAiCU,UAAjC,CAA4CvB,IAAI,CAACe,EAAjD,EAAqD;AACvDC,QAAAA,MAAM,EAAEC,kCAA2BkB,MADoB;AAEvD7B,QAAAA,KAAK,EAAE;AACHoB,UAAAA,IAAI,EAAEQ,CAAC,CAACR,IADL;AAEHnB,UAAAA,OAAO,EAAE2B,CAAC,CAAC3B,OAFR;AAGH6B,UAAAA,IAAI,EAAEF,CAAC,CAACE,IAAF,IAAU;AAHb;AAFgD,OAArD,CAAN;AASA,aAAO;AACH/B,QAAAA,IAAI,EAAE,IADH;AAEHC,QAAAA,KAAK,EAAE;AACHC,UAAAA,OAAO,EAAE2B,CAAC,CAAC3B;AADR;AAFJ,OAAP;AAMH;;AAED,WAAO;AACHF,MAAAA,IAAI,EAAE,EADH;AAEHC,MAAAA,KAAK,EAAE;AAFJ,KAAP;AAIH;;AA9FqE,CAF3D,C","sourcesContent":["import { HandlerPlugin } from \"@webiny/handler/types\";\nimport { ArgsContext } from \"@webiny/handler-args/types\";\nimport {\n PageImportExportTask,\n PageImportExportTaskStatus,\n PbPageImportExportContext\n} from \"~/types\";\nimport { initialStats, readExtractAndUploadZipFileContents } from \"~/importPages/utils\";\nimport { invokeHandlerClient } from \"~/importPages/client\";\nimport { HandlerArgs as ProcessHandlerArgs } from \"../process\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { zeroPad } from \"@webiny/utils\";\n\nexport interface HandlerArgs {\n category: string;\n zipFileKey?: string;\n zipFileUrl?: string;\n task: PageImportExportTask;\n identity: SecurityIdentity;\n}\n\ninterface HandlerResponseError {\n message: string;\n}\nexport interface HandlerResponse {\n data: string | null;\n error: HandlerResponseError | null;\n}\n\ninterface Configuration {\n handlers: {\n process: string;\n };\n}\n\n/**\n * Handles the import page workflow.\n */\nexport default (\n configuration: Configuration\n): HandlerPlugin<PbPageImportExportContext, ArgsContext<HandlerArgs>> => ({\n type: \"handler\",\n async handle(context): Promise<HandlerResponse> {\n const log = console.log;\n\n try {\n log(\"RUNNING Import Pages Create\");\n const { invocationArgs: args, pageBuilder } = context;\n const { task, category, zipFileKey, zipFileUrl, identity } = args;\n if (!zipFileKey && !zipFileUrl) {\n return {\n data: null,\n error: {\n message:\n \"Missing zipFileKey and zipFileUrl parameters. One must be defined.\"\n }\n };\n }\n mockSecurity(identity, context);\n // Step 1: Read the zip file\n const pageImportDataList = await readExtractAndUploadZipFileContents(\n zipFileKey || (zipFileUrl as string)\n );\n // Once we have map we can start processing each page\n\n // For each page create a sub task and invoke the process handler\n for (let i = 0; i < pageImportDataList.length; i++) {\n const pagesDirMap = pageImportDataList[i];\n // Create sub task\n const subtask = await pageBuilder.pageImportExportTask.createSubTask(\n task.id,\n zeroPad(i + 1, 5),\n {\n status: PageImportExportTaskStatus.PENDING,\n data: {\n pageKey: pagesDirMap.key,\n category,\n zipFileKey,\n zipFileUrl,\n input: {\n fileUploadsData: pagesDirMap\n }\n }\n }\n );\n log(`Added SUB_TASK \"${subtask.id}\" to queue.`);\n }\n // Update main task status\n await pageBuilder.pageImportExportTask.updateTask(task.id, {\n status: PageImportExportTaskStatus.PROCESSING,\n stats: initialStats(pageImportDataList.length)\n });\n\n await invokeHandlerClient<ProcessHandlerArgs>({\n context,\n name: configuration.handlers.process,\n payload: {\n taskId: task.id,\n // Execute \"Process\" for the first sub task.\n subTaskIndex: 1,\n identity: context.security.getIdentity()\n }\n });\n } catch (e) {\n log(\"[IMPORT_PAGES_CREATE] Error => \", e);\n\n /**\n * In case of error, we'll update the task status to \"failed\",\n * so that, client can show notify the user appropriately.\n */\n const { invocationArgs: args, pageBuilder } = context;\n const { task } = args;\n\n await pageBuilder.pageImportExportTask.updateTask(task.id, {\n status: PageImportExportTaskStatus.FAILED,\n error: {\n name: e.name,\n message: e.message,\n code: e.code || \"EXPORT_FAILED\"\n }\n });\n\n return {\n data: null,\n error: {\n message: e.message\n }\n };\n }\n\n return {\n data: \"\",\n error: null\n };\n }\n});\n"]}
@@ -2,17 +2,18 @@ import { HandlerPlugin } from "@webiny/handler/types";
2
2
  import { ArgsContext } from "@webiny/handler-args/types";
3
3
  import { PbPageImportExportContext } from "../../types";
4
4
  import { SecurityIdentity } from "@webiny/api-security/types";
5
- export declare type HandlerArgs = {
5
+ export interface HandlerArgs {
6
6
  taskId: string;
7
7
  subTaskIndex: number;
8
8
  identity: SecurityIdentity;
9
- };
10
- export declare type HandlerResponse = {
11
- data: string;
12
- error: {
13
- message: string;
14
- };
15
- };
9
+ }
10
+ interface HandlerResponseError {
11
+ message: string;
12
+ }
13
+ export interface HandlerResponse {
14
+ data: string | null;
15
+ error: HandlerResponseError | null;
16
+ }
16
17
  interface Configuration {
17
18
  handlers: {
18
19
  process: string;
@@ -13,6 +13,8 @@ var _client = require("../client");
13
13
 
14
14
  var _mockSecurity = require("../../mockSecurity");
15
15
 
16
+ var _utils2 = require("@webiny/utils");
17
+
16
18
  /**
17
19
  * Handles the import page workflow.
18
20
  */
@@ -43,7 +45,7 @@ var _default = configuration => ({
43
45
  * Note: We're not going to DB for getting next sub-task to process,
44
46
  * because the data might be out of sync due to GSI eventual consistency.
45
47
  */
46
- subTask = await pageBuilder.pageImportExportTask.getSubTask(taskId, (0, _utils.zeroPad)(subTaskIndex));
48
+ subTask = await pageBuilder.pageImportExportTask.getSubTask(taskId, (0, _utils2.zeroPad)(subTaskIndex, 5));
47
49
  /**
48
50
  * Base condition!!
49
51
  * Bail out early, if task not found or task's status is not "pending".
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["index.ts"],"names":["configuration","type","handle","context","log","console","subTask","noPendingTask","prevStatusOfSubTask","PageImportExportTaskStatus","PENDING","invocationArgs","args","pageBuilder","taskId","subTaskIndex","identity","pageImportExportTask","getSubTask","status","data","error","id","pageKey","category","zipFileKey","input","fileUploadsData","updateSubTask","PROCESSING","updateStats","prevStatus","nextStatus","page","key","pbPage","createPage","updatePage","content","title","path","settings","COMPLETED","message","version","e","FAILED","name","stack","code","updateTask","handlers","process","payload","security","getIdentity"],"mappings":";;;;;;;AAEA;;AACA;;AACA;;AAEA;;AACA;;AAsBA;AACA;AACA;eAEIA,aADW,KAE2D;AACtEC,EAAAA,IAAI,EAAE,SADgE;;AAEtE,QAAMC,MAAN,CAAaC,OAAb,EAAgD;AAC5C,UAAMC,GAAG,GAAGC,OAAO,CAACD,GAApB;AACA,QAAIE,OAAJ;AACA,QAAIC,aAAa,GAAG,IAApB;AACA,QAAIC,mBAAmB,GAAGC,kCAA2BC,OAArD;AAEAN,IAAAA,GAAG,CAAC,mCAAD,CAAH;AACA,UAAM;AAAEO,MAAAA,cAAc,EAAEC,IAAlB;AAAwBC,MAAAA;AAAxB,QAAwCV,OAA9C;AACA,UAAM;AAAEW,MAAAA,MAAF;AAAUC,MAAAA,YAAV;AAAwBC,MAAAA;AAAxB,QAAqCJ,IAA3C,CAR4C,CAS5C;AACA;;AACA,oCAAaI,QAAb,EAAuBb,OAAvB;;AAEA,QAAI;AACA;AACZ;AACA;AACA;AAEYG,MAAAA,OAAO,GAAG,MAAMO,WAAW,CAACI,oBAAZ,CAAiCC,UAAjC,CACZJ,MADY,EAEZ,qBAAQC,YAAR,EAAsB,CAAtB,CAFY,CAAhB;AAKA;AACZ;AACA;AACA;;AACY,UAAI,CAACT,OAAD,IAAYA,OAAO,CAACa,MAAR,KAAmBV,kCAA2BC,OAA9D,EAAuE;AACnEH,QAAAA,aAAa,GAAG,IAAhB;AACA,eAAO;AACHa,UAAAA,IAAI,EAAE,EADH;AAEHC,UAAAA,KAAK,EAAE;AAFJ,SAAP;AAIH,OAND,MAMO;AACHd,QAAAA,aAAa,GAAG,KAAhB;AACH;;AACDC,MAAAA,mBAAmB,GAAGF,OAAO,CAACa,MAA9B;AAEAf,MAAAA,GAAG,CAAE,uBAAsBE,OAAO,CAACgB,EAAG,EAAnC,CAAH;AAEA,YAAM;AAAEC,QAAAA,OAAF;AAAWC,QAAAA,QAAX;AAAqBC,QAAAA,UAArB;AAAiCC,QAAAA;AAAjC,UAA2CpB,OAAO,CAACc,IAAzD;AACA,YAAM;AAAEO,QAAAA;AAAF,UAAsBD,KAA5B;AAEAtB,MAAAA,GAAG,CAAE,wBAAuBmB,OAAQ,GAAjC,CAAH,CA/BA,CAiCA;;AACAjB,MAAAA,OAAO,GAAG,MAAMO,WAAW,CAACI,oBAAZ,CAAiCW,aAAjC,CAA+Cd,MAA/C,EAAuDR,OAAO,CAACgB,EAA/D,EAAmE;AAC/EH,QAAAA,MAAM,EAAEV,kCAA2BoB;AAD4C,OAAnE,CAAhB,CAlCA,CAqCA;;AACA,YAAMhB,WAAW,CAACI,oBAAZ,CAAiCa,WAAjC,CAA6ChB,MAA7C,EAAqD;AACvDiB,QAAAA,UAAU,EAAEvB,mBAD2C;AAEvDwB,QAAAA,UAAU,EAAEvB,kCAA2BoB;AAFgB,OAArD,CAAN;AAIArB,MAAAA,mBAAmB,GAAGF,OAAO,CAACa,MAA9B,CA1CA,CA4CA;;AACA,YAAMc,IAAI,GAAG,MAAM,uBAAW;AAC1B9B,QAAAA,OAD0B;AAE1BoB,QAAAA,OAF0B;AAG1BW,QAAAA,GAAG,EAAET,UAHqB;AAI1BE,QAAAA;AAJ0B,OAAX,CAAnB,CA7CA,CAoDA;;AACA,UAAIQ,MAAM,GAAG,MAAMhC,OAAO,CAACU,WAAR,CAAoBuB,UAApB,CAA+BZ,QAA/B,CAAnB,CArDA,CAuDA;;AACAW,MAAAA,MAAM,GAAG,MAAMhC,OAAO,CAACU,WAAR,CAAoBwB,UAApB,CAA+BF,MAAM,CAACb,EAAtC,EAA0C;AACrDgB,QAAAA,OAAO,EAAEL,IAAI,CAACK,OADuC;AAErDC,QAAAA,KAAK,EAAEN,IAAI,CAACM,KAFyC;AAGrDC,QAAAA,IAAI,EAAEP,IAAI,CAACO,IAH0C;AAIrDC,QAAAA,QAAQ,EAAER,IAAI,CAACQ;AAJsC,OAA1C,CAAf,CAxDA,CA+DA;AAEA;;AACAnC,MAAAA,OAAO,GAAG,MAAMO,WAAW,CAACI,oBAAZ,CAAiCW,aAAjC,CAA+Cd,MAA/C,EAAuDR,OAAO,CAACgB,EAA/D,EAAmE;AAC/EH,QAAAA,MAAM,EAAEV,kCAA2BiC,SAD4C;AAE/EtB,QAAAA,IAAI,EAAE;AACFuB,UAAAA,OAAO,EAAE,MADP;AAEFV,UAAAA,IAAI,EAAE;AACFX,YAAAA,EAAE,EAAEa,MAAM,CAACb,EADT;AAEFiB,YAAAA,KAAK,EAAEJ,MAAM,CAACI,KAFZ;AAGFK,YAAAA,OAAO,EAAET,MAAM,CAACS,OAHd;AAIFzB,YAAAA,MAAM,EAAEgB,MAAM,CAAChB;AAJb;AAFJ;AAFyE,OAAnE,CAAhB,CAlEA,CA8EA;;AACA,YAAMN,WAAW,CAACI,oBAAZ,CAAiCa,WAAjC,CAA6ChB,MAA7C,EAAqD;AACvDiB,QAAAA,UAAU,EAAEvB,mBAD2C;AAEvDwB,QAAAA,UAAU,EAAEvB,kCAA2BiC;AAFgB,OAArD,CAAN;AAIAlC,MAAAA,mBAAmB,GAAGF,OAAO,CAACa,MAA9B;AACH,KApFD,CAoFE,OAAO0B,CAAP,EAAU;AACRzC,MAAAA,GAAG,CAAC,kCAAD,EAAqCyC,CAArC,CAAH;;AAEA,UAAIvC,OAAO,IAAIA,OAAO,CAACgB,EAAvB,EAA2B;AACvB;AAChB;AACA;AACA;AACgB,cAAM;AAAEX,UAAAA,cAAc,EAAEC,IAAlB;AAAwBC,UAAAA;AAAxB,YAAwCV,OAA9C;AACA,cAAM;AAAEW,UAAAA;AAAF,YAAaF,IAAnB;AAEAN,QAAAA,OAAO,GAAG,MAAMO,WAAW,CAACI,oBAAZ,CAAiCW,aAAjC,CAA+Cd,MAA/C,EAAuDR,OAAO,CAACgB,EAA/D,EAAmE;AAC/EH,UAAAA,MAAM,EAAEV,kCAA2BqC,MAD4C;AAE/EzB,UAAAA,KAAK,EAAE;AACH0B,YAAAA,IAAI,EAAEF,CAAC,CAACE,IADL;AAEHJ,YAAAA,OAAO,EAAEE,CAAC,CAACF,OAFR;AAGHK,YAAAA,KAAK,EAAEH,CAAC,CAACG,KAHN;AAIHC,YAAAA,IAAI,EAAE;AAJH;AAFwE,SAAnE,CAAhB,CARuB,CAkBvB;;AACA,cAAMpC,WAAW,CAACI,oBAAZ,CAAiCa,WAAjC,CAA6ChB,MAA7C,EAAqD;AACvDiB,UAAAA,UAAU,EAAEvB,mBAD2C;AAEvDwB,UAAAA,UAAU,EAAEvB,kCAA2BqC;AAFgB,SAArD,CAAN;AAIAtC,QAAAA,mBAAmB,GAAGF,OAAO,CAACa,MAA9B;AACH;;AAED,aAAO;AACHC,QAAAA,IAAI,EAAE,IADH;AAEHC,QAAAA,KAAK,EAAE;AACHsB,UAAAA,OAAO,EAAEE,CAAC,CAACF;AADR;AAFJ,OAAP;AAMH,KAvHD,SAuHU;AACN;AACA,UAAIpC,aAAJ,EAAmB;AACfH,QAAAA,GAAG,CAAE,gCAA+BU,MAAO,EAAxC,CAAH;AAEA,cAAMD,WAAW,CAACI,oBAAZ,CAAiCiC,UAAjC,CAA4CpC,MAA5C,EAAoD;AACtDK,UAAAA,MAAM,EAAEV,kCAA2BiC,SADmB;AAEtDtB,UAAAA,IAAI,EAAE;AACFuB,YAAAA,OAAO,EAAG;AADR;AAFgD,SAApD,CAAN;AAMH,OATD,MASO;AACHvC,QAAAA,GAAG,CAAE,8BAA6BW,YAAY,GAAG,CAAE,GAAhD,CAAH,CADG,CAEH;;AACA,cAAM,iCAAiC;AACnCZ,UAAAA,OADmC;AAEnC4C,UAAAA,IAAI,EAAE/C,aAAa,CAACmD,QAAd,CAAuBC,OAFM;AAGnCC,UAAAA,OAAO,EAAE;AACLvC,YAAAA,MADK;AAELC,YAAAA,YAAY,EAAEA,YAAY,GAAG,CAFxB;AAGLC,YAAAA,QAAQ,EAAEb,OAAO,CAACmD,QAAR,CAAiBC,WAAjB;AAHL;AAH0B,SAAjC,CAAN;AASH;AACJ;;AACD,WAAO;AACHnC,MAAAA,IAAI,EAAE,EADH;AAEHC,MAAAA,KAAK,EAAE;AAFJ,KAAP;AAIH;;AAnKqE,CAF3D,C","sourcesContent":["import { HandlerPlugin } from \"@webiny/handler/types\";\nimport { ArgsContext } from \"@webiny/handler-args/types\";\nimport { PageImportExportTaskStatus, PbPageImportExportContext } from \"~/types\";\nimport { importPage } from \"~/importPages/utils\";\nimport { invokeHandlerClient } from \"~/importPages/client\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { zeroPad } from \"@webiny/utils\";\n\nexport interface HandlerArgs {\n taskId: string;\n subTaskIndex: number;\n identity: SecurityIdentity;\n}\n\ninterface HandlerResponseError {\n message: string;\n}\nexport interface HandlerResponse {\n data: string | null;\n error: HandlerResponseError | null;\n}\n\ninterface Configuration {\n handlers: {\n process: string;\n };\n}\n\n/**\n * Handles the import page workflow.\n */\nexport default (\n configuration: Configuration\n): HandlerPlugin<PbPageImportExportContext, ArgsContext<HandlerArgs>> => ({\n type: \"handler\",\n async handle(context): Promise<HandlerResponse> {\n const log = console.log;\n let subTask;\n let noPendingTask = true;\n let prevStatusOfSubTask = PageImportExportTaskStatus.PENDING;\n\n log(\"RUNNING Import Page Queue Process\");\n const { invocationArgs: args, pageBuilder } = context;\n const { taskId, subTaskIndex, identity } = args;\n // Disable authorization; this is necessary because we call Page Builder CRUD methods which include authorization checks\n // and this Lambda is invoked internally, without credentials.\n mockSecurity(identity, context);\n\n try {\n /*\n * Note: We're not going to DB for getting next sub-task to process,\n * because the data might be out of sync due to GSI eventual consistency.\n */\n\n subTask = await pageBuilder.pageImportExportTask.getSubTask(\n taskId,\n zeroPad(subTaskIndex, 5)\n );\n\n /**\n * Base condition!!\n * Bail out early, if task not found or task's status is not \"pending\".\n */\n if (!subTask || subTask.status !== PageImportExportTaskStatus.PENDING) {\n noPendingTask = true;\n return {\n data: \"\",\n error: null\n };\n } else {\n noPendingTask = false;\n }\n prevStatusOfSubTask = subTask.status;\n\n log(`Fetched sub task => ${subTask.id}`);\n\n const { pageKey, category, zipFileKey, input } = subTask.data;\n const { fileUploadsData } = input;\n\n log(`Processing page key \"${pageKey}\"`);\n\n // Mark task status as PROCESSING\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.PROCESSING\n });\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.PROCESSING\n });\n prevStatusOfSubTask = subTask.status;\n\n // Real job\n const page = await importPage({\n context,\n pageKey,\n key: zipFileKey,\n fileUploadsData\n });\n\n // Create a page\n let pbPage = await context.pageBuilder.createPage(category);\n\n // Update page with data\n pbPage = await context.pageBuilder.updatePage(pbPage.id, {\n content: page.content,\n title: page.title,\n path: page.path,\n settings: page.settings\n });\n\n // TODO: Publish page\n\n // Update task record in DB\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.COMPLETED,\n data: {\n message: \"Done\",\n page: {\n id: pbPage.id,\n title: pbPage.title,\n version: pbPage.version,\n status: pbPage.status\n }\n }\n });\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.COMPLETED\n });\n prevStatusOfSubTask = subTask.status;\n } catch (e) {\n log(\"[IMPORT_PAGES_PROCESS] Error => \", e);\n\n if (subTask && subTask.id) {\n /**\n * In case of error, we'll update the task status to \"failed\",\n * so that, client can show notify the user appropriately.\n */\n const { invocationArgs: args, pageBuilder } = context;\n const { taskId } = args;\n\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.FAILED,\n error: {\n name: e.name,\n message: e.message,\n stack: e.stack,\n code: \"IMPORT_FAILED\"\n }\n });\n\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.FAILED\n });\n prevStatusOfSubTask = subTask.status;\n }\n\n return {\n data: null,\n error: {\n message: e.message\n }\n };\n } finally {\n // Base condition!\n if (noPendingTask) {\n log(`No pending sub-task for task ${taskId}`);\n\n await pageBuilder.pageImportExportTask.updateTask(taskId, {\n status: PageImportExportTaskStatus.COMPLETED,\n data: {\n message: `Finish importing pages.`\n }\n });\n } else {\n log(`Invoking PROCESS for task \"${subTaskIndex + 1}\"`);\n // We want to continue with Self invocation no matter if current page error out.\n await invokeHandlerClient<HandlerArgs>({\n context,\n name: configuration.handlers.process,\n payload: {\n taskId,\n subTaskIndex: subTaskIndex + 1,\n identity: context.security.getIdentity()\n }\n });\n }\n }\n return {\n data: \"\",\n error: null\n };\n }\n});\n"]}
@@ -1,15 +1,23 @@
1
- import { PageImportExportTaskStatus } from "../types";
1
+ import { File } from "../types";
2
2
  import { PbPageImportExportContext } from "../graphql/types";
3
3
  import { ExportedPageData } from "../exportPages/utils";
4
+ interface FileItem extends File {
5
+ key: string;
6
+ type: string;
7
+ name: string;
8
+ size: number;
9
+ meta: Record<string, any>;
10
+ tags: string[];
11
+ }
4
12
  interface UploadPageAssetsParams {
5
13
  context: PbPageImportExportContext;
6
- filesData: Record<string, any>[];
14
+ filesData: FileItem[];
7
15
  fileUploadsData: FileUploadsData;
8
16
  }
9
17
  interface UploadPageAssetsReturnType {
10
- fileIdToKeyMap?: Map<string, string>;
18
+ fileIdToKeyMap: Map<string, string>;
11
19
  }
12
- export declare const uploadPageAssets: ({ context, filesData, fileUploadsData }: UploadPageAssetsParams) => Promise<UploadPageAssetsReturnType>;
20
+ export declare const uploadPageAssets: (params: UploadPageAssetsParams) => Promise<UploadPageAssetsReturnType>;
13
21
  interface FileUploadsData {
14
22
  data: string;
15
23
  assets: Record<string, string>;
@@ -32,12 +40,11 @@ interface PageImportData {
32
40
  * @return PageImportData S3 file keys for all uploaded assets group by page.
33
41
  */
34
42
  export declare function readExtractAndUploadZipFileContents(zipFileKey: string): Promise<PageImportData[]>;
35
- export declare const zeroPad: (version: any) => string;
36
- export declare function initialStats(total: any): {
37
- pending: any;
43
+ export declare function initialStats(total: number): {
44
+ pending: number;
38
45
  processing: number;
39
46
  completed: number;
40
47
  failed: number;
41
- total: any;
48
+ total: number;
42
49
  };
43
50
  export {};
@@ -8,7 +8,7 @@ Object.defineProperty(exports, "__esModule", {
8
8
  exports.importPage = importPage;
9
9
  exports.initialStats = initialStats;
10
10
  exports.readExtractAndUploadZipFileContents = readExtractAndUploadZipFileContents;
11
- exports.zeroPad = exports.uploadPageAssets = void 0;
11
+ exports.uploadPageAssets = void 0;
12
12
 
13
13
  var _uniqid = _interopRequireDefault(require("uniqid"));
14
14
 
@@ -48,23 +48,28 @@ const INSTALL_EXTRACT_DIR = _path.default.join(INSTALL_DIR, "apiPageBuilderImpor
48
48
  const FILES_COUNT_IN_EACH_BATCH = 15;
49
49
  const ZIP_CONTENT_TYPE = "application/zip";
50
50
 
51
- function updateImageInPageSettings({
52
- settings,
53
- fileIdToKeyMap,
54
- srcPrefix
55
- }) {
51
+ function updateImageInPageSettings(params) {
52
+ const {
53
+ settings,
54
+ fileIdToKeyMap,
55
+ srcPrefix
56
+ } = params;
56
57
  let newSettings = settings;
57
58
  const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith("/") ? srcPrefix.slice(0, -1) : srcPrefix;
58
59
 
59
60
  if (_dotPropImmutable.default.get(newSettings, "general.image.src")) {
60
- newSettings = _dotPropImmutable.default.set(newSettings, "general.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(settings.general.image.id)}`);
61
+ var _settings$general, _settings$general$ima;
62
+
63
+ newSettings = _dotPropImmutable.default.set(newSettings, "general.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(((_settings$general = settings.general) === null || _settings$general === void 0 ? void 0 : (_settings$general$ima = _settings$general.image) === null || _settings$general$ima === void 0 ? void 0 : _settings$general$ima.id) || "")}`);
61
64
  }
62
65
 
63
66
  if (_dotPropImmutable.default.get(newSettings, "social.image.src")) {
64
- newSettings = _dotPropImmutable.default.set(newSettings, "social.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(settings.social.image.id)}`);
67
+ var _settings$social, _settings$social$imag;
68
+
69
+ newSettings = _dotPropImmutable.default.set(newSettings, "social.image.src", `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(((_settings$social = settings.social) === null || _settings$social === void 0 ? void 0 : (_settings$social$imag = _settings$social.image) === null || _settings$social$imag === void 0 ? void 0 : _settings$social$imag.id) || "")}`);
65
70
  }
66
71
 
67
- return settings;
72
+ return newSettings;
68
73
  }
69
74
 
70
75
  function updateFilesInPageData({
@@ -111,23 +116,27 @@ function updateFilesInPageData({
111
116
  }
112
117
  }
113
118
 
114
- const uploadPageAssets = async ({
115
- context,
116
- filesData,
117
- fileUploadsData
118
- }) => {
119
+ const uploadPageAssets = async params => {
120
+ const {
121
+ context,
122
+ filesData,
123
+ fileUploadsData
124
+ } = params; // Save uploaded file key against static id for later use.
125
+
126
+ const fileIdToKeyMap = new Map();
119
127
  /**
120
128
  * This function contains logic of file download from S3.
121
129
  * Current we're not mocking zip file download from S3 in tests at the moment.
122
130
  * So, we're manually mocking it in case of test just by returning an empty object.
123
131
  */
132
+
124
133
  if (process.env.NODE_ENV === "test") {
125
- return {};
134
+ return {
135
+ fileIdToKeyMap
136
+ };
126
137
  }
127
138
 
128
- console.log("INSIDE uploadPageAssets"); // Save uploaded file key against static id for later use.
129
-
130
- const fileIdToKeyMap = new Map(); // Save files meta data against old key for later use.
139
+ console.log("INSIDE uploadPageAssets"); // Save files meta data against old key for later use.
131
140
 
132
141
  const fileKeyToFileMap = new Map(); // Initialize maps.
133
142
 
@@ -145,7 +154,12 @@ const uploadPageAssets = async ({
145
154
 
146
155
  const createFilesInput = fileUploadResults.map(uploadResult => {
147
156
  const newKey = uploadResult.Key;
148
- const file = fileKeyToFileMap.get(getOldFileKey(newKey)); // Update the file map with newly uploaded file.
157
+ const file = fileKeyToFileMap.get(getOldFileKey(newKey));
158
+
159
+ if (!file) {
160
+ return null;
161
+ } // Update the file map with newly uploaded file.
162
+
149
163
 
150
164
  fileIdToKeyMap.set(file.id, newKey);
151
165
  return {
@@ -156,7 +170,7 @@ const uploadPageAssets = async ({
156
170
  meta: file.meta,
157
171
  tags: file.tags
158
172
  };
159
- });
173
+ }).filter(Boolean);
160
174
  const createFilesPromises = []; // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).
161
175
 
162
176
  const createFilesInputChunks = (0, _chunk.default)(createFilesInput, FILES_COUNT_IN_EACH_BATCH);
@@ -206,25 +220,31 @@ async function importPage({
206
220
  files
207
221
  } = await (0, _loadJsonFile.default)(PAGE_DATA_FILE_PATH); // Only update page data if there are files.
208
222
 
209
- if (Array.isArray(files) && files.length) {
223
+ if (files && Array.isArray(files) && files.length > 0) {
210
224
  // Upload page assets.
211
225
  const {
212
226
  fileIdToKeyMap
213
227
  } = await uploadPageAssets({
214
228
  context,
229
+
230
+ /**
231
+ * TODO @ts-refactor @ashutosh figure out correct types.
232
+ */
233
+ // @ts-ignore
215
234
  filesData: files,
216
235
  fileUploadsData
217
236
  });
237
+ const settings = await context.fileManager.settings.getSettings();
218
238
  const {
219
- srcPrefix
220
- } = await context.fileManager.settings.getSettings();
239
+ srcPrefix = ""
240
+ } = settings || {};
221
241
  updateFilesInPageData({
222
- data: page.content,
242
+ data: page.content || {},
223
243
  fileIdToKeyMap,
224
244
  srcPrefix
225
245
  });
226
246
  page.settings = updateImageInPageSettings({
227
- settings: page.settings,
247
+ settings: page.settings || {},
228
248
  fileIdToKeyMap,
229
249
  srcPrefix
230
250
  });
@@ -382,16 +402,13 @@ async function deleteS3Folder(key) {
382
402
  }
383
403
 
384
404
  const response = await _s3Stream.s3Stream.listObject(key);
385
- const keys = response.Contents.map(c => c.Key);
405
+ const keys = (response.Contents || []).map(c => c.Key).filter(Boolean);
386
406
  console.log(`Found ${keys.length} files.`);
387
407
  const deleteFilePromises = keys.map(key => _s3Stream.s3Stream.deleteObject(key));
388
408
  await Promise.all(deleteFilePromises);
389
409
  console.log(`Successfully deleted ${deleteFilePromises.length} files.`);
390
- }
410
+ } // export const zeroPad = version => `${version}`.padStart(5, "0");
391
411
 
392
- const zeroPad = version => `${version}`.padStart(5, "0");
393
-
394
- exports.zeroPad = zeroPad;
395
412
 
396
413
  function initialStats(total) {
397
414
  return {
@@ -419,6 +436,13 @@ function extractZipToDisk(exportFileZipPath) {
419
436
  if (err) {
420
437
  console.warn("ERROR: Failed to extract zip: ", exportFileZipPath, err);
421
438
  reject(err);
439
+ return;
440
+ }
441
+
442
+ if (!zipFile) {
443
+ console.log("ERROR: Missing zip file resource for path: " + exportFileZipPath);
444
+ reject("Missing Zip File Resource.");
445
+ return;
422
446
  }
423
447
 
424
448
  console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);
@@ -445,6 +469,13 @@ function extractZipToDisk(exportFileZipPath) {
445
469
  if (err) {
446
470
  console.warn("ERROR: Failed to openReadStream for file: ", entry.fileName, err);
447
471
  reject(err);
472
+ return;
473
+ }
474
+
475
+ if (!readStream) {
476
+ console.log("ERROR: Missing Read Stream Resource when extracting to disk.");
477
+ reject("Missing Read Stream Resource.");
478
+ return;
448
479
  }
449
480
 
450
481
  const filePath = _path.default.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);
@@ -480,6 +511,13 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
480
511
  if (err) {
481
512
  console.warn("ERROR: Failed to extract zip: ", pageDataZipFilePath, err);
482
513
  reject(err);
514
+ return;
515
+ }
516
+
517
+ if (!zipFile) {
518
+ console.log("ERROR: Probably failed to extract zip: " + pageDataZipFilePath);
519
+ reject("Missing Zip File Resource.");
520
+ return;
483
521
  }
484
522
 
485
523
  console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);
@@ -511,6 +549,13 @@ function extractZipAndUploadToS3(pageDataZipFilePath, uniquePath) {
511
549
  if (err) {
512
550
  console.warn("ERROR: Failed while performing [openReadStream] for file: ", entry.fileName, err);
513
551
  reject(err);
552
+ return;
553
+ }
554
+
555
+ if (!readStream) {
556
+ console.log("ERROR: Missing Read Stream while importing pages.");
557
+ reject("Missing Read Strea Resource.");
558
+ return;
514
559
  }
515
560
 
516
561
  readStream.on("end", function () {
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["utils.ts"],"names":["streamPipeline","pipeline","INSTALL_DIR","INSTALL_EXTRACT_DIR","path","join","FILES_COUNT_IN_EACH_BATCH","ZIP_CONTENT_TYPE","updateImageInPageSettings","params","settings","fileIdToKeyMap","srcPrefix","newSettings","srcPrefixWithoutTrailingSlash","endsWith","slice","dotProp","get","set","general","image","id","social","updateFilesInPageData","data","Array","isArray","i","length","element","tuple","Object","entries","key","value","has","name","src","uploadPageAssets","context","filesData","fileUploadsData","Map","process","env","NODE_ENV","console","log","fileKeyToFileMap","file","type","fileUploadResults","uploadFilesFromS3","oldKeyToNewKeyMap","assets","createFilesInput","map","uploadResult","newKey","Key","getOldFileKey","size","meta","tags","filter","Boolean","createFilesPromises","createFilesInputChunks","createFilesInputChunk","push","fileManager","files","createFilesInBatch","Promise","all","importPage","pageKey","PAGE_EXTRACT_DIR","pageDataFileKey","PAGE_DATA_FILE_PATH","basename","resolve","reject","s3Stream","readStream","on","pipe","page","getSettings","content","deleteS3Folder","dirname","oldKeysForAssets","keys","promises","oldKey","tempNewKey","fileMetaData","streamPassThrough","streamPassThroughUploadPromise","promise","writeStream","getObjectMetaFromS3","getObjectHead","ContentType","WebinyError","rest","split","e","FILE_CONTENT_TYPE","getFileNameWithoutExt","fileName","replace","extname","readExtractAndUploadZipFileContents","zipFileKey","pageImportDataList","startsWith","response","ok","statusText","body","uniquePath","zipFileName","ZIP_FILE_PATH","zipFilePaths","extractZipToDisk","currentPath","dataMap","extractZipAndUploadToS3","ASSETS_DIR_NAME","preparePageDataDirMap","filePath","isAsset","listObject","Contents","c","deleteFilePromises","deleteObject","initialStats","total","PageImportExportTaskStatus","PENDING","PROCESSING","COMPLETED","FAILED","exportFileZipPath","pageZipFilePaths","uniqueFolderNameForExport","EXPORT_FILE_EXTRACTION_PATH","yauzl","open","lazyEntries","err","zipFile","warn","info","entryCount","readEntry","entry","test","openReadStream","catch","error","pageDataZipFilePath","filePaths","fileUploadPromises","uniquePageKey","then","res","forEach","r"],"mappings":";;;;;;;;;;;;AAAA;;AAEA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AAEA;;AACA;;AACA;;AAEA;;AAaA,MAAMA,cAAc,GAAG,qBAAUC,gBAAV,CAAvB;AAEA,MAAMC,WAAW,GAAG,MAApB;;AACA,MAAMC,mBAAmB,GAAGC,cAAKC,IAAL,CAAUH,WAAV,EAAuB,0BAAvB,CAA5B;;AACA,MAAMI,yBAAyB,GAAG,EAAlC;AACA,MAAMC,gBAAgB,GAAG,iBAAzB;;AAcA,SAASC,yBAAT,CACIC,MADJ,EAE+C;AAC3C,QAAM;AAAEC,IAAAA,QAAF;AAAYC,IAAAA,cAAZ;AAA4BC,IAAAA;AAA5B,MAA0CH,MAAhD;AACA,MAAII,WAAW,GAAGH,QAAlB;AAEA,QAAMI,6BAA6B,GAAGF,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAChCH,SAAS,CAACI,KAAV,CAAgB,CAAhB,EAAmB,CAAC,CAApB,CADgC,GAEhCJ,SAFN;;AAIA,MAAIK,0BAAQC,GAAR,CAAYL,WAAZ,EAAyB,mBAAzB,CAAJ,EAAmD;AAAA;;AAC/CA,IAAAA,WAAW,GAAGI,0BAAQE,GAAR,CACVN,WADU,EAEV,mBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,sBAAAR,QAAQ,CAACU,OAAT,iGAAkBC,KAAlB,gFAAyBC,EAAzB,KAA+B,EADC,CAElC,EALQ,CAAd;AAOH;;AACD,MAAIL,0BAAQC,GAAR,CAAYL,WAAZ,EAAyB,kBAAzB,CAAJ,EAAkD;AAAA;;AAC9CA,IAAAA,WAAW,GAAGI,0BAAQE,GAAR,CACVN,WADU,EAEV,kBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,qBAAAR,QAAQ,CAACa,MAAT,+FAAiBF,KAAjB,gFAAwBC,EAAxB,KAA8B,EADE,CAElC,EALQ,CAAd;AAOH;;AAED,SAAOT,WAAP;AACH;;AAED,SAASW,qBAAT,CAA+B;AAAEC,EAAAA,IAAF;AAAQd,EAAAA,cAAR;AAAwBC,EAAAA;AAAxB,CAA/B,EAAiG;AAC7F;AACA,MAAI,CAACa,IAAD,IAAS,OAAOA,IAAP,KAAgB,QAA7B,EAAuC;AACnC;AACH,GAJ4F,CAK7F;;;AACA,MAAIC,KAAK,CAACC,OAAN,CAAcF,IAAd,CAAJ,EAAyB;AACrB,SAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGH,IAAI,CAACI,MAAzB,EAAiCD,CAAC,EAAlC,EAAsC;AAClC,YAAME,OAAO,GAAGL,IAAI,CAACG,CAAD,CAApB;AACAJ,MAAAA,qBAAqB,CAAC;AAAEC,QAAAA,IAAI,EAAEK,OAAR;AAAiBnB,QAAAA,cAAjB;AAAiCC,QAAAA;AAAjC,OAAD,CAArB;AACH;;AACD;AACH,GAZ4F,CAa7F;;;AACA,QAAMmB,KAAK,GAAGC,MAAM,CAACC,OAAP,CAAeR,IAAf,CAAd;;AACA,OAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGG,KAAK,CAACF,MAA1B,EAAkCD,CAAC,EAAnC,EAAuC;AACnC,UAAM,CAACM,GAAD,EAAMC,KAAN,IAAeJ,KAAK,CAACH,CAAD,CAA1B;;AAEA,QAAIM,GAAG,KAAK,MAAR,IAAkBC,KAAlB,IAA2BxB,cAAc,CAACyB,GAAf,CAAmBD,KAAK,CAACb,EAAzB,CAA/B,EAA6D;AACzDa,MAAAA,KAAK,CAACD,GAAN,GAAYvB,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAZ;AACAa,MAAAA,KAAK,CAACE,IAAN,GAAa1B,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAb;AACAa,MAAAA,KAAK,CAACG,GAAN,GAAa,GAAE1B,SAAU,GAAEA,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAA0B,EAA1B,GAA+B,GAAI,GAAEJ,cAAc,CAACO,GAAf,CAC5DiB,KAAK,CAACb,EADsD,CAE9D,EAFF;AAGH,KAND,MAMO;AACHE,MAAAA,qBAAqB,CAAC;AAAEC,QAAAA,IAAI,EAAEU,KAAR;AAAevB,QAAAA,SAAf;AAA0BD,QAAAA;AAA1B,OAAD,CAArB;AACH;AACJ;AACJ;;AAYM,MAAM4B,gBAAgB,GAAG,MAC5B9B,MAD4B,IAEU;AACtC,QAAM;AAAE+B,IAAAA,OAAF;AAAWC,IAAAA,SAAX;AAAsBC,IAAAA;AAAtB,MAA0CjC,MAAhD,CADsC,CAEtC;;AACA,QAAME,cAAc,GAAG,IAAIgC,GAAJ,EAAvB;AACA;AACJ;AACA;AACA;AACA;;AACI,MAAIC,OAAO,CAACC,GAAR,CAAYC,QAAZ,KAAyB,MAA7B,EAAqC;AACjC,WAAO;AACHnC,MAAAA;AADG,KAAP;AAGH;;AACDoC,EAAAA,OAAO,CAACC,GAAR,CAAY,yBAAZ,EAdsC,CAgBtC;;AACA,QAAMC,gBAAgB,GAAG,IAAIN,GAAJ,EAAzB,CAjBsC,CAkBtC;;AACA,OAAK,IAAIf,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGa,SAAS,CAACZ,MAA9B,EAAsCD,CAAC,EAAvC,EAA2C;AACvC,UAAMsB,IAAI,GAAGT,SAAS,CAACb,CAAD,CAAtB;AACAqB,IAAAA,gBAAgB,CAAC9B,GAAjB,CAAqB+B,IAAI,CAAChB,GAA1B,EAA+BgB,IAA/B,EAFuC,CAIvC;;AACAvC,IAAAA,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4B4B,IAAI,CAACC,IAAjC;AACH;;AAED,QAAMC,iBAAiB,GAAG,MAAMC,iBAAiB,CAAC;AAC9CJ,IAAAA,gBAD8C;AAE9CK,IAAAA,iBAAiB,EAAEZ,eAAe,CAACa;AAFW,GAAD,CAAjD,CA3BsC,CAgCtC;;AACA,QAAMC,gBAAgB,GAAGJ,iBAAiB,CACrCK,GADoB,CACfC,YAAD,IAAoC;AACrC,UAAMC,MAAM,GAAGD,YAAY,CAACE,GAA5B;AACA,UAAMV,IAAI,GAAGD,gBAAgB,CAAC/B,GAAjB,CAAqB2C,aAAa,CAACF,MAAD,CAAlC,CAAb;;AACA,QAAI,CAACT,IAAL,EAAW;AACP,aAAO,IAAP;AACH,KALoC,CAOrC;;;AACAvC,IAAAA,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4BqC,MAA5B;AAEA,WAAO;AACHzB,MAAAA,GAAG,EAAEyB,MADF;AAEHtB,MAAAA,IAAI,EAAEa,IAAI,CAACb,IAFR;AAGHyB,MAAAA,IAAI,EAAEZ,IAAI,CAACY,IAHR;AAIHX,MAAAA,IAAI,EAAED,IAAI,CAACC,IAJR;AAKHY,MAAAA,IAAI,EAAEb,IAAI,CAACa,IALR;AAMHC,MAAAA,IAAI,EAAEd,IAAI,CAACc;AANR,KAAP;AAQH,GAnBoB,EAoBpBC,MApBoB,CAoBbC,OApBa,CAAzB;AAsBA,QAAMC,mBAAmB,GAAG,EAA5B,CAvDsC,CAwDtC;;AACA,QAAMC,sBAAsB,GAAG,oBAAMZ,gBAAN,EAAwBlD,yBAAxB,CAA/B;;AACA,OAAK,IAAIsB,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGwC,sBAAsB,CAACvC,MAA3C,EAAmDD,CAAC,EAApD,EAAwD;AACpD,UAAMyC,qBAAqB,GAAGD,sBAAsB,CAACxC,CAAD,CAApD;AACAuC,IAAAA,mBAAmB,CAACG,IAApB;AACI;AACZ;AACA;AACA;AACY9B,IAAAA,OAAO,CAAC+B,WAAR,CAAoBC,KAApB,CAA0BC,kBAA1B,CAA6CJ,qBAA7C,CALJ;AAOH;;AAED,QAAMK,OAAO,CAACC,GAAR,CAAYR,mBAAZ,CAAN;AAEA,SAAO;AACHxD,IAAAA;AADG,GAAP;AAGH,CA5EM;;;;AA0FA,eAAeiE,UAAf,CAA0B;AAC7BC,EAAAA,OAD6B;AAE7BrC,EAAAA,OAF6B;AAG7BE,EAAAA;AAH6B,CAA1B,EAIiD;AACpD,QAAMM,GAAG,GAAGD,OAAO,CAACC,GAApB,CADoD,CAGpD;;AACA,QAAM8B,gBAAgB,GAAG1E,cAAKC,IAAL,CAAUF,mBAAV,EAA+B0E,OAA/B,CAAzB;;AACA,8BAAcC,gBAAd;;AAEA,QAAMC,eAAe,GAAG9D,0BAAQC,GAAR,CAAYwB,eAAZ,EAA8B,MAA9B,CAAxB;;AACA,QAAMsC,mBAAmB,GAAG5E,cAAKC,IAAL,CAAUyE,gBAAV,EAA4B1E,cAAK6E,QAAL,CAAcF,eAAd,CAA5B,CAA5B;;AAEA/B,EAAAA,GAAG,CAAE,+BAA8B+B,eAAgB,QAAOC,mBAAoB,GAA3E,CAAH,CAVoD,CAWpD;;AACA,QAAM,IAAIN,OAAJ,CAAY,CAACQ,OAAD,EAAUC,MAAV,KAAqB;AACnCC,uBACKC,UADL,CACgBN,eADhB,EAEKO,EAFL,CAEQ,OAFR,EAEiBH,MAFjB,EAGKI,IAHL,CAGU,2BAAkBP,mBAAlB,CAHV,EAIKM,EAJL,CAIQ,OAJR,EAIiBH,MAJjB,EAKKG,EALL,CAKQ,QALR,EAKkBJ,OALlB;AAMH,GAPK,CAAN,CAZoD,CAqBpD;;AACAlC,EAAAA,GAAG,CAAE,aAAY+B,eAAgB,EAA9B,CAAH;AACA,QAAM;AAAES,IAAAA,IAAF;AAAQhB,IAAAA;AAAR,MAAkB,MAAM,2BAA2BQ,mBAA3B,CAA9B,CAvBoD,CAyBpD;;AACA,MAAIR,KAAK,IAAI9C,KAAK,CAACC,OAAN,CAAc6C,KAAd,CAAT,IAAiCA,KAAK,CAAC3C,MAAN,GAAe,CAApD,EAAuD;AACnD;AACA,UAAM;AAAElB,MAAAA;AAAF,QAAqB,MAAM4B,gBAAgB,CAAC;AAC9CC,MAAAA,OAD8C;;AAE9C;AACZ;AACA;AACY;AACAC,MAAAA,SAAS,EAAE+B,KANmC;AAO9C9B,MAAAA;AAP8C,KAAD,CAAjD;AAUA,UAAMhC,QAAQ,GAAG,MAAM8B,OAAO,CAAC+B,WAAR,CAAoB7D,QAApB,CAA6B+E,WAA7B,EAAvB;AAEA,UAAM;AAAE7E,MAAAA,SAAS,GAAG;AAAd,QAAqBF,QAAQ,IAAI,EAAvC;AACAc,IAAAA,qBAAqB,CAAC;AAClBC,MAAAA,IAAI,EAAE+D,IAAI,CAACE,OAAL,IAAgB,EADJ;AAElB/E,MAAAA,cAFkB;AAGlBC,MAAAA;AAHkB,KAAD,CAArB;AAMA4E,IAAAA,IAAI,CAAC9E,QAAL,GAAgBF,yBAAyB,CAAC;AACtCE,MAAAA,QAAQ,EAAE8E,IAAI,CAAC9E,QAAL,IAAiB,EADW;AAEtCC,MAAAA,cAFsC;AAGtCC,MAAAA;AAHsC,KAAD,CAAzC;AAKH;;AAEDoC,EAAAA,GAAG,CAAC,gCAAD,CAAH;AACA,QAAM,sCAAW6B,OAAX,CAAN;AAEA7B,EAAAA,GAAG,CAAE,iCAAF,CAAH;AACA,QAAM2C,cAAc,CAACvF,cAAKwF,OAAL,CAAalD,eAAe,CAACjB,IAA7B,CAAD,CAApB;AAEA,SAAO+D,IAAP;AACH;;AAOD,eAAenC,iBAAf,CAAiC;AAC7BJ,EAAAA,gBAD6B;AAE7BK,EAAAA;AAF6B,CAAjC,EAGmE;AAC/D,QAAMuC,gBAAgB,GAAG7D,MAAM,CAAC8D,IAAP,CAAYxC,iBAAZ,CAAzB;AAEA,QAAMyC,QAAQ,GAAG,EAAjB,CAH+D,CAI/D;;AACA,OAAK,IAAInE,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGiE,gBAAgB,CAAChE,MAArC,EAA6CD,CAAC,EAA9C,EAAkD;AAC9C,UAAMoE,MAAM,GAAGH,gBAAgB,CAACjE,CAAD,CAA/B;AACA,UAAMqE,UAAU,GAAG3C,iBAAiB,CAAC0C,MAAD,CAApC,CAF8C,CAI9C;;AACA,UAAMX,UAAU,GAAGD,mBAASC,UAAT,CAAoBY,UAApB,CAAnB,CAL8C,CAM9C;;;AACA,UAAMC,YAAY,GAAGjD,gBAAgB,CAAC/B,GAAjB,CAAqB8E,MAArB,CAArB;;AAEA,QAAIE,YAAJ,EAAkB;AACd,YAAMvC,MAAM,GAAG,qBAAS,EAAT,EAAc,IAAGuC,YAAY,CAAChE,GAAI,EAAlC,CAAf;;AACA,YAAM;AAAEiE,QAAAA,iBAAF;AAAqBC,QAAAA,8BAA8B,EAAEC;AAArD,UACFjB,mBAASkB,WAAT,CAAqB3C,MAArB,EAA6BuC,YAAY,CAAC/C,IAA1C,CADJ;;AAEAkC,MAAAA,UAAU,CAACE,IAAX,CAAgBY,iBAAhB;AACAJ,MAAAA,QAAQ,CAACzB,IAAT,CAAc+B,OAAd;AAEAtD,MAAAA,OAAO,CAACC,GAAR,CAAa,6BAA4BW,MAAO,GAAhD;AACH;AACJ;;AAED,SAAOe,OAAO,CAACC,GAAR,CAAYoB,QAAZ,CAAP;AACH;;AAED,eAAeQ,mBAAf,CAAmC3C,GAAnC,EAAgD;AAC5C,QAAMG,IAAI,GAAG,MAAMqB,mBAASoB,aAAT,CAAuB5C,GAAvB,CAAnB;;AAEA,MAAIG,IAAI,CAAC0C,WAAL,KAAqBlG,gBAAzB,EAA2C;AACvC,UAAM,IAAImG,cAAJ,CAAiB,2BAA0B3C,IAAI,CAAC0C,WAAY,GAA5D,EAAgE,kBAAhE,CAAN;AACH;AACJ;;AAED,SAAS5C,aAAT,CAAuB3B,GAAvB,EAAoC;AAChC;AACJ;AACA;AACI,MAAI;AACA,UAAM,GAAG,GAAGyE,IAAN,IAAczE,GAAG,CAAC0E,KAAJ,CAAU,GAAV,CAApB;AACA,WAAOD,IAAI,CAACtG,IAAL,CAAU,GAAV,CAAP;AACH,GAHD,CAGE,OAAOwG,CAAP,EAAU;AACR,WAAO3E,GAAP;AACH;AACJ;;AAED,MAAM4E,iBAAiB,GAAG,0BAA1B;;AAEA,SAASC,qBAAT,CAA+BC,QAA/B,EAAyD;AACrD,SAAO5G,cAAK6E,QAAL,CAAc+B,QAAd,EAAwBC,OAAxB,CAAgC7G,cAAK8G,OAAL,CAAaF,QAAb,CAAhC,EAAwD,EAAxD,CAAP;AACH;;AAQD;AACA;AACA;AACA;AACA;AACO,eAAeG,mCAAf,CACHC,UADG,EAEsB;AACzB,QAAMpE,GAAG,GAAGD,OAAO,CAACC,GAApB;AACA,QAAMqE,kBAAkB,GAAG,EAA3B;AACA,MAAIhC,UAAJ,CAHyB,CAIzB;;AACA,MAAI+B,UAAU,CAACE,UAAX,CAAsB,MAAtB,CAAJ,EAAmC;AAC/B,UAAMC,QAAQ,GAAG,MAAM,wBAAMH,UAAN,CAAvB;;AACA,QAAI,CAACG,QAAQ,CAACC,EAAd,EAAkB;AACd,YAAM,IAAId,cAAJ,CACD,gCAA+BU,UAAW,GADzC,EAEFG,QAAQ,CAACE,UAFP,CAAN;AAIH;;AAEDpC,IAAAA,UAAU,GAAGkC,QAAQ,CAACG,IAAtB;AACH,GAVD,MAUO;AACH;AACA,UAAMnB,mBAAmB,CAACa,UAAD,CAAzB;AAEA/B,IAAAA,UAAU,GAAGD,mBAASC,UAAT,CAAoB+B,UAApB,CAAb;AACH;;AAED,QAAMO,UAAU,GAAG,qBAAS,eAAT,CAAnB;;AACA,QAAMC,WAAW,GAAGxH,cAAK6E,QAAL,CAAcmC,UAAd,CAApB,CAvByB,CAwBzB;;;AACA,QAAMS,aAAa,GAAGzH,cAAKC,IAAL,CAAUH,WAAV,EAAuB0H,WAAvB,CAAtB;;AAEA,QAAMtB,WAAW,GAAG,2BAAkBuB,aAAlB,CAApB;AACA,QAAM7H,cAAc,CAACqF,UAAD,EAAaiB,WAAb,CAApB;AACAtD,EAAAA,GAAG,CAAE,oBAAmB4E,WAAY,QAAOC,aAAc,EAAtD,CAAH,CA7ByB,CA+BzB;;AACA,QAAMC,YAAY,GAAG,MAAMC,gBAAgB,CAACF,aAAD,CAA3C;AAEA7E,EAAAA,GAAG,CAAE,sBAAqBoE,UAAW,UAASS,aAAc,EAAzD,CAAH;AACA,QAAM,sCAAWA,aAAX,CAAN,CAnCyB,CAqCzB;;AACA,OAAK,IAAIjG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGkG,YAAY,CAACjG,MAAjC,EAAyCD,CAAC,EAA1C,EAA8C;AAC1C,UAAMoG,WAAW,GAAGF,YAAY,CAAClG,CAAD,CAAhC;AACA,UAAMqG,OAAO,GAAG,MAAMC,uBAAuB,CAACF,WAAD,EAAcL,UAAd,CAA7C;AACAN,IAAAA,kBAAkB,CAAC/C,IAAnB,CAAwB2D,OAAxB;AACH;;AACDjF,EAAAA,GAAG,CAAC,oCAAD,EAAuC5C,cAAKwF,OAAL,CAAakC,YAAY,CAAC,CAAD,CAAzB,CAAvC,CAAH;AACA,QAAM,sCAAW1H,cAAKwF,OAAL,CAAakC,YAAY,CAAC,CAAD,CAAzB,CAAX,CAAN;AAEA,SAAOT,kBAAP;AACH;;AAED,MAAMc,eAAe,GAAG,SAAxB;;AAEA,SAASC,qBAAT,CAA+B;AAC3B3E,EAAAA,GAD2B;AAE3B4E,EAAAA,QAF2B;AAG3B1E,EAAAA;AAH2B,CAA/B,EAQmB;AACf,QAAMiC,OAAO,GAAGxF,cAAKwF,OAAL,CAAayC,QAAb,CAAhB;;AACA,QAAMrB,QAAQ,GAAG5G,cAAK6E,QAAL,CAAcoD,QAAd,CAAjB;AACA;AACJ;AACA;AACA;;;AACI,QAAMrC,MAAM,GAAGgB,QAAQ,CAACC,OAAT,CAAiB,KAAjB,EAAwB,KAAxB,CAAf;AAEA,QAAMqB,OAAO,GAAG1C,OAAO,CAAC7E,QAAR,CAAiBoH,eAAjB,CAAhB;;AAEA,MAAIG,OAAJ,EAAa;AACT7E,IAAAA,GAAG,GAAGxC,0BAAQE,GAAR,CAAYsC,GAAZ,EAAkB,UAASuC,MAAO,EAAlC,EAAqCrC,MAArC,CAAN;AACH,GAFD,MAEO;AACH;AACAF,IAAAA,GAAG,GAAGxC,0BAAQE,GAAR,CAAYsC,GAAZ,EAAkB,MAAlB,EAAyBE,MAAzB,CAAN;AACH;;AAED,SAAOF,GAAP;AACH;;AAED,eAAekC,cAAf,CAA8BzD,GAA9B,EAA0D;AACtD;AACA,MAAI,CAACA,GAAG,CAACnB,QAAJ,CAAa,GAAb,CAAL,EAAwB;AACpBmB,IAAAA,GAAG,GAAI,GAAEA,GAAI,GAAb;AACH;;AAED,QAAMqF,QAAQ,GAAG,MAAMnC,mBAASmD,UAAT,CAAoBrG,GAApB,CAAvB;AACA,QAAM4D,IAAI,GAAG,CAACyB,QAAQ,CAACiB,QAAT,IAAqB,EAAtB,EAA0B/E,GAA1B,CAA8BgF,CAAC,IAAIA,CAAC,CAAC7E,GAArC,EAA0CK,MAA1C,CAAiDC,OAAjD,CAAb;AACAnB,EAAAA,OAAO,CAACC,GAAR,CAAa,SAAQ8C,IAAI,CAACjE,MAAO,SAAjC;AAEA,QAAM6G,kBAAkB,GAAG5C,IAAI,CAACrC,GAAL,CAASvB,GAAG,IAAIkD,mBAASuD,YAAT,CAAsBzG,GAAtB,CAAhB,CAA3B;AAEA,QAAMwC,OAAO,CAACC,GAAR,CAAY+D,kBAAZ,CAAN;AACA3F,EAAAA,OAAO,CAACC,GAAR,CAAa,wBAAuB0F,kBAAkB,CAAC7G,MAAO,SAA9D;AACH,C,CAED;;;AAEO,SAAS+G,YAAT,CAAsBC,KAAtB,EAAqC;AACxC,SAAO;AACH,KAACC,kCAA2BC,OAA5B,GAAsCF,KADnC;AAEH,KAACC,kCAA2BE,UAA5B,GAAyC,CAFtC;AAGH,KAACF,kCAA2BG,SAA5B,GAAwC,CAHrC;AAIH,KAACH,kCAA2BI,MAA5B,GAAqC,CAJlC;AAKHL,IAAAA;AALG,GAAP;AAOH;;AAED,SAASd,gBAAT,CAA0BoB,iBAA1B,EAAwE;AACpE,SAAO,IAAIzE,OAAJ,CAAY,CAACQ,OAAD,EAAUC,MAAV,KAAqB;AACpC,UAAMiE,gBAA0B,GAAG,EAAnC;AACA,UAAMC,yBAAyB,GAAGtC,qBAAqB,CAACoC,iBAAD,CAAvD;;AACA,UAAMG,2BAA2B,GAAGlJ,cAAKC,IAAL,CAAUH,WAAV,EAAuBmJ,yBAAvB,CAApC,CAHoC,CAIpC;;;AACA,gCAAcC,2BAAd;;AAEAC,mBAAMC,IAAN,CAAWL,iBAAX,EAA8B;AAAEM,MAAAA,WAAW,EAAE;AAAf,KAA9B,EAAqD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;AACzE,UAAID,GAAJ,EAAS;AACL3G,QAAAA,OAAO,CAAC6G,IAAR,CAAa,gCAAb,EAA+CT,iBAA/C,EAAkEO,GAAlE;AACAvE,QAAAA,MAAM,CAACuE,GAAD,CAAN;AACA;AACH;;AACD,UAAI,CAACC,OAAL,EAAc;AACV5G,QAAAA,OAAO,CAACC,GAAR,CAAY,gDAAgDmG,iBAA5D;AACAhE,QAAAA,MAAM,CAAC,4BAAD,CAAN;AACA;AACH;;AAEDpC,MAAAA,OAAO,CAAC8G,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;AAEAH,MAAAA,OAAO,CAACrE,EAAR,CAAW,KAAX,EAAkB,UAAUoE,GAAV,EAAe;AAC7B,YAAIA,GAAJ,EAAS;AACL3G,UAAAA,OAAO,CAAC6G,IAAR,CAAa,uCAAb,EAAsDT,iBAAtD,EAAyEO,GAAzE;AACAvE,UAAAA,MAAM,CAACuE,GAAD,CAAN;AACH;;AACDxE,QAAAA,OAAO,CAACkE,gBAAD,CAAP;AACH,OAND;AAQAO,MAAAA,OAAO,CAACI,SAAR;AAEAJ,MAAAA,OAAO,CAACrE,EAAR,CAAW,OAAX,EAAoB,UAAU0E,KAAV,EAAiB;AACjCjH,QAAAA,OAAO,CAAC8G,IAAR,CAAc,sBAAqBG,KAAK,CAAChD,QAAS,GAAlD;;AACA,YAAI,MAAMiD,IAAN,CAAWD,KAAK,CAAChD,QAAjB,CAAJ,EAAgC;AAC5B;AACA;AACA;AACA2C,UAAAA,OAAO,CAACI,SAAR;AACH,SALD,MAKO;AACH;AACAJ,UAAAA,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAerE,UAAf,EAA2B;AACrD,gBAAIqE,GAAJ,EAAS;AACL3G,cAAAA,OAAO,CAAC6G,IAAR,CACI,4CADJ,EAEII,KAAK,CAAChD,QAFV,EAGI0C,GAHJ;AAKAvE,cAAAA,MAAM,CAACuE,GAAD,CAAN;AACA;AACH;;AACD,gBAAI,CAACrE,UAAL,EAAiB;AACbtC,cAAAA,OAAO,CAACC,GAAR,CACI,8DADJ;AAGAmC,cAAAA,MAAM,CAAC,+BAAD,CAAN;AACA;AACH;;AAED,kBAAMkD,QAAQ,GAAGjI,cAAKC,IAAL,CAAUiJ,2BAAV,EAAuCU,KAAK,CAAChD,QAA7C,CAAjB;;AAEA3B,YAAAA,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;AAC7B8D,cAAAA,gBAAgB,CAAC9E,IAAjB,CAAsB+D,QAAtB;AACAsB,cAAAA,OAAO,CAACI,SAAR;AACH,aAHD;AAKA/J,YAAAA,cAAc,CAACqF,UAAD,EAAa,2BAAkBgD,QAAlB,CAAb,CAAd,CAAwD8B,KAAxD,CAA8DC,KAAK,IAAI;AACnEjF,cAAAA,MAAM,CAACiF,KAAD,CAAN;AACH,aAFD;AAGH,WA5BD;AA6BH;AACJ,OAvCD;AAwCH,KAhED;AAiEH,GAxEM,CAAP;AAyEH;;AAED,SAASlC,uBAAT,CACImC,mBADJ,EAEI1C,UAFJ,EAG2B;AACvB,SAAO,IAAIjD,OAAJ,CAAY,CAACQ,OAAD,EAAUC,MAAV,KAAqB;AACpC,UAAMmF,SAAS,GAAG,EAAlB;AACA,UAAMC,kBAAwD,GAAG,EAAjE;AACA,UAAMC,aAAa,GAAGzD,qBAAqB,CAACsD,mBAAD,CAA3C;AACA,QAAIpC,OAAuB,GAAG;AAC1B/F,MAAAA,GAAG,EAAEsI,aADqB;AAE1BjH,MAAAA,MAAM,EAAE,EAFkB;AAG1B9B,MAAAA,IAAI,EAAE;AAHoB,KAA9B;;AAKA8H,mBAAMC,IAAN,CAAWa,mBAAX,EAAgC;AAAEZ,MAAAA,WAAW,EAAE;AAAf,KAAhC,EAAuD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;AAC3E,UAAID,GAAJ,EAAS;AACL3G,QAAAA,OAAO,CAAC6G,IAAR,CAAa,gCAAb,EAA+CS,mBAA/C,EAAoEX,GAApE;AACAvE,QAAAA,MAAM,CAACuE,GAAD,CAAN;AACA;AACH;;AACD,UAAI,CAACC,OAAL,EAAc;AACV5G,QAAAA,OAAO,CAACC,GAAR,CAAY,4CAA4CqH,mBAAxD;AACAlF,QAAAA,MAAM,CAAC,4BAAD,CAAN;AACA;AACH;;AACDpC,MAAAA,OAAO,CAAC8G,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;AACAH,MAAAA,OAAO,CAACrE,EAAR,CAAW,KAAX,EAAkB,UAAUoE,GAAV,EAAe;AAC7B,YAAIA,GAAJ,EAAS;AACL3G,UAAAA,OAAO,CAAC6G,IAAR,CAAa,mCAAb,EAAkDS,mBAAlD,EAAuEX,GAAvE;AACAvE,UAAAA,MAAM,CAACuE,GAAD,CAAN;AACH;;AAEDhF,QAAAA,OAAO,CAACC,GAAR,CAAY4F,kBAAZ,EAAgCE,IAAhC,CAAqCC,GAAG,IAAI;AACxCA,UAAAA,GAAG,CAACC,OAAJ,CAAYC,CAAC,IAAI;AACb7H,YAAAA,OAAO,CAAC8G,IAAR,CAAa,oBAAb,EAAmCe,CAAnC;AACH,WAFD;AAGA1F,UAAAA,OAAO,CAAC+C,OAAD,CAAP;AACH,SALD;AAMH,OAZD;AAcA0B,MAAAA,OAAO,CAACI,SAAR;AAEAJ,MAAAA,OAAO,CAACrE,EAAR,CAAW,OAAX,EAAoB,UAAU0E,KAAV,EAAiB;AACjCjH,QAAAA,OAAO,CAAC8G,IAAR,CAAc,sBAAqBG,KAAK,CAAChD,QAAS,GAAlD;;AACA,YAAI,MAAMiD,IAAN,CAAWD,KAAK,CAAChD,QAAjB,CAAJ,EAAgC;AAC5B;AACA;AACA;AACA2C,UAAAA,OAAO,CAACI,SAAR;AACH,SALD,MAKO;AACH;AACAJ,UAAAA,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAerE,UAAf,EAA2B;AACrD,gBAAIqE,GAAJ,EAAS;AACL3G,cAAAA,OAAO,CAAC6G,IAAR,CACI,4DADJ,EAEII,KAAK,CAAChD,QAFV,EAGI0C,GAHJ;AAKAvE,cAAAA,MAAM,CAACuE,GAAD,CAAN;AACA;AACH;;AACD,gBAAI,CAACrE,UAAL,EAAiB;AACbtC,cAAAA,OAAO,CAACC,GAAR,CAAY,mDAAZ;AACAmC,cAAAA,MAAM,CAAC,8BAAD,CAAN;AACA;AACH;;AACDE,YAAAA,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;AAC7BgF,cAAAA,SAAS,CAAChG,IAAV,CAAe0F,KAAK,CAAChD,QAArB;AACA2C,cAAAA,OAAO,CAACI,SAAR;AACH,aAHD;AAKA,kBAAMpG,MAAM,GAAI,GAAEgE,UAAW,IAAG6C,aAAc,IAAGR,KAAK,CAAChD,QAAS,EAAhE,CApBqD,CAqBrD;;AACAiB,YAAAA,OAAO,GAAGG,qBAAqB,CAAC;AAC5B3E,cAAAA,GAAG,EAAEwE,OADuB;AAE5BI,cAAAA,QAAQ,EAAE2B,KAAK,CAAChD,QAFY;AAG5BrD,cAAAA;AAH4B,aAAD,CAA/B;;AAMA,kBAAM;AAAEwC,cAAAA,iBAAF;AAAqBC,cAAAA,8BAA8B,EAAEC;AAArD,gBACFjB,mBAASkB,WAAT,CAAqB3C,MAArB,EAA6BmD,iBAA7B,CADJ;;AAGA9G,YAAAA,cAAc,CAACqF,UAAD,EAAac,iBAAb,CAAd,CACKsE,IADL,CACU,MAAM;AACRF,cAAAA,kBAAkB,CAACjG,IAAnB,CAAwB+B,OAAxB;AACH,aAHL,EAIK8D,KAJL,CAIWC,KAAK,IAAI;AACZjF,cAAAA,MAAM,CAACiF,KAAD,CAAN;AACH,aANL;AAOH,WAtCD;AAuCH;AACJ,OAjDD;AAkDH,KA9ED;AA+EH,GAxFM,CAAP;AAyFH","sourcesContent":["import uniqueId from \"uniqid\";\nimport S3 from \"aws-sdk/clients/s3\";\nimport dotProp from \"dot-prop-immutable\";\nimport { createWriteStream } from \"fs\";\nimport { ensureDirSync } from \"fs-extra\";\nimport { promisify } from \"util\";\nimport { pipeline } from \"stream\";\nimport fetch from \"node-fetch\";\nimport path from \"path\";\nimport yauzl from \"yauzl\";\nimport chunk from \"lodash/chunk\";\nimport loadJson from \"load-json-file\";\nimport { FileInput } from \"@webiny/api-file-manager/types\";\nimport WebinyError from \"@webiny/error\";\nimport { deleteFile } from \"@webiny/api-page-builder/graphql/crud/install/utils/downloadInstallFiles\";\nimport { File, PageImportExportTaskStatus } from \"~/types\";\nimport { PbPageImportExportContext } from \"~/graphql/types\";\nimport { s3Stream } from \"~/exportPages/s3Stream\";\nimport { ExportedPageData } from \"~/exportPages/utils\";\nimport { PageSettings } from \"@webiny/api-page-builder/types\";\n\ninterface FileItem extends File {\n key: string;\n type: string;\n name: string;\n size: number;\n meta: Record<string, any>;\n tags: string[];\n}\n\nconst streamPipeline = promisify(pipeline);\n\nconst INSTALL_DIR = \"/tmp\";\nconst INSTALL_EXTRACT_DIR = path.join(INSTALL_DIR, \"apiPageBuilderImportPage\");\nconst FILES_COUNT_IN_EACH_BATCH = 15;\nconst ZIP_CONTENT_TYPE = \"application/zip\";\n\ninterface UpdateFilesInPageDataParams {\n data: Record<string, any>;\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n}\n\ninterface UpdateImageInPageSettingsParams {\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n settings: PageSettings;\n}\n\nfunction updateImageInPageSettings(\n params: UpdateImageInPageSettingsParams\n): UpdateImageInPageSettingsParams[\"settings\"] {\n const { settings, fileIdToKeyMap, srcPrefix } = params;\n let newSettings = settings;\n\n const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith(\"/\")\n ? srcPrefix.slice(0, -1)\n : srcPrefix;\n\n if (dotProp.get(newSettings, \"general.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"general.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.general?.image?.id || \"\"\n )}`\n );\n }\n if (dotProp.get(newSettings, \"social.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"social.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.social?.image?.id || \"\"\n )}`\n );\n }\n\n return newSettings;\n}\n\nfunction updateFilesInPageData({ data, fileIdToKeyMap, srcPrefix }: UpdateFilesInPageDataParams) {\n // BASE CASE: Termination point\n if (!data || typeof data !== \"object\") {\n return;\n }\n // Recursively call function if data is array\n if (Array.isArray(data)) {\n for (let i = 0; i < data.length; i++) {\n const element = data[i];\n updateFilesInPageData({ data: element, fileIdToKeyMap, srcPrefix });\n }\n return;\n }\n // Main logic\n const tuple = Object.entries(data);\n for (let i = 0; i < tuple.length; i++) {\n const [key, value] = tuple[i];\n\n if (key === \"file\" && value && fileIdToKeyMap.has(value.id)) {\n value.key = fileIdToKeyMap.get(value.id);\n value.name = fileIdToKeyMap.get(value.id);\n value.src = `${srcPrefix}${srcPrefix.endsWith(\"/\") ? \"\" : \"/\"}${fileIdToKeyMap.get(\n value.id\n )}`;\n } else {\n updateFilesInPageData({ data: value, srcPrefix, fileIdToKeyMap });\n }\n }\n}\n\ninterface UploadPageAssetsParams {\n context: PbPageImportExportContext;\n filesData: FileItem[];\n fileUploadsData: FileUploadsData;\n}\n\ninterface UploadPageAssetsReturnType {\n fileIdToKeyMap: Map<string, string>;\n}\n\nexport const uploadPageAssets = async (\n params: UploadPageAssetsParams\n): Promise<UploadPageAssetsReturnType> => {\n const { context, filesData, fileUploadsData } = params;\n // Save uploaded file key against static id for later use.\n const fileIdToKeyMap = new Map<string, string>();\n /**\n * This function contains logic of file download from S3.\n * Current we're not mocking zip file download from S3 in tests at the moment.\n * So, we're manually mocking it in case of test just by returning an empty object.\n */\n if (process.env.NODE_ENV === \"test\") {\n return {\n fileIdToKeyMap\n };\n }\n console.log(\"INSIDE uploadPageAssets\");\n\n // Save files meta data against old key for later use.\n const fileKeyToFileMap = new Map<string, FileItem>();\n // Initialize maps.\n for (let i = 0; i < filesData.length; i++) {\n const file = filesData[i];\n fileKeyToFileMap.set(file.key, file);\n\n // Initialize the value\n fileIdToKeyMap.set(file.id, file.type);\n }\n\n const fileUploadResults = await uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap: fileUploadsData.assets\n });\n\n // Create files in File Manager\n const createFilesInput = fileUploadResults\n .map((uploadResult): FileInput | null => {\n const newKey = uploadResult.Key;\n const file = fileKeyToFileMap.get(getOldFileKey(newKey));\n if (!file) {\n return null;\n }\n\n // Update the file map with newly uploaded file.\n fileIdToKeyMap.set(file.id, newKey);\n\n return {\n key: newKey,\n name: file.name,\n size: file.size,\n type: file.type,\n meta: file.meta,\n tags: file.tags\n };\n })\n .filter(Boolean) as FileInput[];\n\n const createFilesPromises = [];\n // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).\n const createFilesInputChunks = chunk(createFilesInput, FILES_COUNT_IN_EACH_BATCH);\n for (let i = 0; i < createFilesInputChunks.length; i++) {\n const createFilesInputChunk = createFilesInputChunks[i];\n createFilesPromises.push(\n /*\n * We need to break down files into chunks because\n * `createFilesInBatch` operation has a limit on number of files it can handle at once.\n */\n context.fileManager.files.createFilesInBatch(createFilesInputChunk)\n );\n }\n\n await Promise.all(createFilesPromises);\n\n return {\n fileIdToKeyMap\n };\n};\n\ninterface FileUploadsData {\n data: string;\n assets: Record<string, string>;\n}\n\ninterface ImportPageParams {\n key: string;\n pageKey: string;\n context: PbPageImportExportContext;\n fileUploadsData: FileUploadsData;\n}\n\nexport async function importPage({\n pageKey,\n context,\n fileUploadsData\n}: ImportPageParams): Promise<ExportedPageData[\"page\"]> {\n const log = console.log;\n\n // Making Directory for page in which we're going to extract the page data file.\n const PAGE_EXTRACT_DIR = path.join(INSTALL_EXTRACT_DIR, pageKey);\n ensureDirSync(PAGE_EXTRACT_DIR);\n\n const pageDataFileKey = dotProp.get(fileUploadsData, `data`);\n const PAGE_DATA_FILE_PATH = path.join(PAGE_EXTRACT_DIR, path.basename(pageDataFileKey));\n\n log(`Downloading Page data file: ${pageDataFileKey} at \"${PAGE_DATA_FILE_PATH}\"`);\n // Download and save page data file in disk.\n await new Promise((resolve, reject) => {\n s3Stream\n .readStream(pageDataFileKey)\n .on(\"error\", reject)\n .pipe(createWriteStream(PAGE_DATA_FILE_PATH))\n .on(\"error\", reject)\n .on(\"finish\", resolve);\n });\n\n // Load the page data file from disk.\n log(`Load file ${pageDataFileKey}`);\n const { page, files } = await loadJson<ExportedPageData>(PAGE_DATA_FILE_PATH);\n\n // Only update page data if there are files.\n if (files && Array.isArray(files) && files.length > 0) {\n // Upload page assets.\n const { fileIdToKeyMap } = await uploadPageAssets({\n context,\n /**\n * TODO @ts-refactor @ashutosh figure out correct types.\n */\n // @ts-ignore\n filesData: files,\n fileUploadsData\n });\n\n const settings = await context.fileManager.settings.getSettings();\n\n const { srcPrefix = \"\" } = settings || {};\n updateFilesInPageData({\n data: page.content || {},\n fileIdToKeyMap,\n srcPrefix\n });\n\n page.settings = updateImageInPageSettings({\n settings: page.settings || {},\n fileIdToKeyMap,\n srcPrefix\n });\n }\n\n log(\"Removing Directory for page...\");\n await deleteFile(pageKey);\n\n log(`Remove page contents from S3...`);\n await deleteS3Folder(path.dirname(fileUploadsData.data));\n\n return page;\n}\n\ninterface UploadFilesFromZipParams {\n fileKeyToFileMap: Map<string, any>;\n oldKeyToNewKeyMap: Record<string, string>;\n}\n\nasync function uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap\n}: UploadFilesFromZipParams): Promise<S3.ManagedUpload.SendData[]> {\n const oldKeysForAssets = Object.keys(oldKeyToNewKeyMap);\n\n const promises = [];\n // Upload all assets.\n for (let i = 0; i < oldKeysForAssets.length; i++) {\n const oldKey = oldKeysForAssets[i];\n const tempNewKey = oldKeyToNewKeyMap[oldKey];\n\n // Read file.\n const readStream = s3Stream.readStream(tempNewKey);\n // Get file meta data.\n const fileMetaData = fileKeyToFileMap.get(oldKey);\n\n if (fileMetaData) {\n const newKey = uniqueId(\"\", `-${fileMetaData.key}`);\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, fileMetaData.type);\n readStream.pipe(streamPassThrough);\n promises.push(promise);\n\n console.log(`Successfully queued file \"${newKey}\"`);\n }\n }\n\n return Promise.all(promises);\n}\n\nasync function getObjectMetaFromS3(Key: string) {\n const meta = await s3Stream.getObjectHead(Key);\n\n if (meta.ContentType !== ZIP_CONTENT_TYPE) {\n throw new WebinyError(`Unsupported file type: \"${meta.ContentType}\"`, \"UNSUPPORTED_FILE\");\n }\n}\n\nfunction getOldFileKey(key: string) {\n /*\n * Because we know the naming convention, we can extract the old key from new key.\n */\n try {\n const [, ...rest] = key.split(\"-\");\n return rest.join(\"-\");\n } catch (e) {\n return key;\n }\n}\n\nconst FILE_CONTENT_TYPE = \"application/octet-stream\";\n\nfunction getFileNameWithoutExt(fileName: string): string {\n return path.basename(fileName).replace(path.extname(fileName), \"\");\n}\n\ninterface PageImportData {\n assets: Record<string, string>;\n data: string;\n key: string;\n}\n\n/**\n * Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.\n * @param zipFileKey\n * @return PageImportData S3 file keys for all uploaded assets group by page.\n */\nexport async function readExtractAndUploadZipFileContents(\n zipFileKey: string\n): Promise<PageImportData[]> {\n const log = console.log;\n const pageImportDataList = [];\n let readStream;\n // Check whether it is a URL\n if (zipFileKey.startsWith(\"http\")) {\n const response = await fetch(zipFileKey);\n if (!response.ok) {\n throw new WebinyError(\n `Unable to downloading file: \"${zipFileKey}\"`,\n response.statusText\n );\n }\n\n readStream = response.body;\n } else {\n // We're first retrieving object's meta data, just to check whether the file is available at the given Key\n await getObjectMetaFromS3(zipFileKey);\n\n readStream = s3Stream.readStream(zipFileKey);\n }\n\n const uniquePath = uniqueId(\"IMPORT_PAGES/\");\n const zipFileName = path.basename(zipFileKey);\n // Read export file and download it in the disk\n const ZIP_FILE_PATH = path.join(INSTALL_DIR, zipFileName);\n\n const writeStream = createWriteStream(ZIP_FILE_PATH);\n await streamPipeline(readStream, writeStream);\n log(`Downloaded file \"${zipFileName}\" at ${ZIP_FILE_PATH}`);\n\n // Extract the downloaded zip file\n const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);\n\n log(`Removing ZIP file \"${zipFileKey}\" from ${ZIP_FILE_PATH}`);\n await deleteFile(ZIP_FILE_PATH);\n\n // Extract each page zip and upload their content's to S3\n for (let i = 0; i < zipFilePaths.length; i++) {\n const currentPath = zipFilePaths[i];\n const dataMap = await extractZipAndUploadToS3(currentPath, uniquePath);\n pageImportDataList.push(dataMap);\n }\n log(\"Removing all ZIP files located at \", path.dirname(zipFilePaths[0]));\n await deleteFile(path.dirname(zipFilePaths[0]));\n\n return pageImportDataList;\n}\n\nconst ASSETS_DIR_NAME = \"/assets\";\n\nfunction preparePageDataDirMap({\n map,\n filePath,\n newKey\n}: {\n map: PageImportData;\n filePath: string;\n newKey: string;\n}): PageImportData {\n const dirname = path.dirname(filePath);\n const fileName = path.basename(filePath);\n /*\n * We want to use dot (.) as part of object key rather than creating nested object(s).\n * Also, the file name might contain dots in it beside the extension, so, we are escaping them all.\n */\n const oldKey = fileName.replace(/\\./g, \"\\\\.\");\n\n const isAsset = dirname.endsWith(ASSETS_DIR_NAME);\n\n if (isAsset) {\n map = dotProp.set(map, `assets.${oldKey}`, newKey);\n } else {\n // We only need to know the newKey for data file.\n map = dotProp.set(map, `data`, newKey);\n }\n\n return map;\n}\n\nasync function deleteS3Folder(key: string): Promise<void> {\n // Append trailing slash i.e \"/\" to key to make sure we only delete a specific folder.\n if (!key.endsWith(\"/\")) {\n key = `${key}/`;\n }\n\n const response = await s3Stream.listObject(key);\n const keys = (response.Contents || []).map(c => c.Key).filter(Boolean) as string[];\n console.log(`Found ${keys.length} files.`);\n\n const deleteFilePromises = keys.map(key => s3Stream.deleteObject(key));\n\n await Promise.all(deleteFilePromises);\n console.log(`Successfully deleted ${deleteFilePromises.length} files.`);\n}\n\n// export const zeroPad = version => `${version}`.padStart(5, \"0\");\n\nexport function initialStats(total: number) {\n return {\n [PageImportExportTaskStatus.PENDING]: total,\n [PageImportExportTaskStatus.PROCESSING]: 0,\n [PageImportExportTaskStatus.COMPLETED]: 0,\n [PageImportExportTaskStatus.FAILED]: 0,\n total\n };\n}\n\nfunction extractZipToDisk(exportFileZipPath: string): Promise<string[]> {\n return new Promise((resolve, reject) => {\n const pageZipFilePaths: string[] = [];\n const uniqueFolderNameForExport = getFileNameWithoutExt(exportFileZipPath);\n const EXPORT_FILE_EXTRACTION_PATH = path.join(INSTALL_DIR, uniqueFolderNameForExport);\n // Make sure DIR exists\n ensureDirSync(EXPORT_FILE_EXTRACTION_PATH);\n\n yauzl.open(exportFileZipPath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", exportFileZipPath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Missing zip file resource for path: \" + exportFileZipPath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn(\"ERROR: Failed on END event for file: \", exportFileZipPath, err);\n reject(err);\n }\n resolve(pageZipFilePaths);\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed to openReadStream for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\n \"ERROR: Missing Read Stream Resource when extracting to disk.\"\n );\n reject(\"Missing Read Stream Resource.\");\n return;\n }\n\n const filePath = path.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);\n\n readStream.on(\"end\", function () {\n pageZipFilePaths.push(filePath);\n zipFile.readEntry();\n });\n\n streamPipeline(readStream, createWriteStream(filePath)).catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n\nfunction extractZipAndUploadToS3(\n pageDataZipFilePath: string,\n uniquePath: string\n): Promise<PageImportData> {\n return new Promise((resolve, reject) => {\n const filePaths = [];\n const fileUploadPromises: Promise<S3.ManagedUpload.SendData>[] = [];\n const uniquePageKey = getFileNameWithoutExt(pageDataZipFilePath);\n let dataMap: PageImportData = {\n key: uniquePageKey,\n assets: {},\n data: \"\"\n };\n yauzl.open(pageDataZipFilePath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", pageDataZipFilePath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Probably failed to extract zip: \" + pageDataZipFilePath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn('ERROR: Failed on \"END\" for file: ', pageDataZipFilePath, err);\n reject(err);\n }\n\n Promise.all(fileUploadPromises).then(res => {\n res.forEach(r => {\n console.info(\"Done uploading... \", r);\n });\n resolve(dataMap);\n });\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed while performing [openReadStream] for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\"ERROR: Missing Read Stream while importing pages.\");\n reject(\"Missing Read Strea Resource.\");\n return;\n }\n readStream.on(\"end\", function () {\n filePaths.push(entry.fileName);\n zipFile.readEntry();\n });\n\n const newKey = `${uniquePath}/${uniquePageKey}/${entry.fileName}`;\n // Modify in place\n dataMap = preparePageDataDirMap({\n map: dataMap,\n filePath: entry.fileName,\n newKey\n });\n\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, FILE_CONTENT_TYPE);\n\n streamPipeline(readStream, streamPassThrough)\n .then(() => {\n fileUploadPromises.push(promise);\n })\n .catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n"]}
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["mockSecurity.ts"],"names":["mockSecurity","identity","context","security","disableAuthorization","setIdentity"],"mappings":";;;;;;;AAEO,MAAMA,YAAY,GAAG,CAACC,QAAD,EAA6BC,OAA7B,KAA0D;AAClFA,EAAAA,OAAO,CAACC,QAAR,CAAiBC,oBAAjB;AACAF,EAAAA,OAAO,CAACC,QAAR,CAAiBE,WAAjB,CAA6BJ,QAA7B;AACH,CAHM","sourcesContent":["import { SecurityContext, SecurityIdentity } from \"@webiny/api-security/types\";\n\nexport const mockSecurity = (identity: SecurityIdentity, context: SecurityContext) => {\n context.security.disableAuthorization();\n context.security.setIdentity(identity);\n};\n"]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@webiny/api-page-builder-import-export",
3
- "version": "5.23.1",
3
+ "version": "5.25.0-beta.0",
4
4
  "main": "index.js",
5
5
  "keywords": [
6
6
  "pbie:base"
@@ -14,17 +14,18 @@
14
14
  "author": "Webiny Ltd",
15
15
  "license": "MIT",
16
16
  "dependencies": {
17
- "@babel/runtime": "7.16.7",
17
+ "@babel/runtime": "7.17.2",
18
18
  "@commodo/fields": "1.1.2-beta.20",
19
- "@webiny/api-file-manager": "5.23.1",
20
- "@webiny/api-page-builder": "5.23.1",
21
- "@webiny/api-security": "5.23.1",
22
- "@webiny/error": "5.23.1",
23
- "@webiny/handler": "5.23.1",
24
- "@webiny/handler-args": "5.23.1",
25
- "@webiny/handler-aws": "5.23.1",
26
- "@webiny/handler-graphql": "5.23.1",
27
- "@webiny/validation": "5.23.1",
19
+ "@webiny/api-file-manager": "5.25.0-beta.0",
20
+ "@webiny/api-page-builder": "5.25.0-beta.0",
21
+ "@webiny/api-security": "5.25.0-beta.0",
22
+ "@webiny/error": "5.25.0-beta.0",
23
+ "@webiny/handler": "5.25.0-beta.0",
24
+ "@webiny/handler-args": "5.25.0-beta.0",
25
+ "@webiny/handler-aws": "5.25.0-beta.0",
26
+ "@webiny/handler-graphql": "5.25.0-beta.0",
27
+ "@webiny/utils": "5.25.0-beta.0",
28
+ "@webiny/validation": "5.25.0-beta.0",
28
29
  "archiver": "5.3.0",
29
30
  "commodo-fields-object": "1.0.6",
30
31
  "dot-prop-immutable": "2.1.1",
@@ -45,21 +46,23 @@
45
46
  "@babel/preset-typescript": "^7.16.0",
46
47
  "@elastic/elasticsearch": "7.12.0",
47
48
  "@shelf/jest-elasticsearch": "^1.0.0",
49
+ "@types/archiver": "^5.3.1",
48
50
  "@types/puppeteer": "^5.4.2",
49
- "@webiny/api-dynamodb-to-elasticsearch": "^5.23.1",
50
- "@webiny/api-file-manager-ddb-es": "^5.23.1",
51
- "@webiny/api-i18n-ddb": "^5.23.1",
52
- "@webiny/api-security-so-ddb": "^5.23.1",
53
- "@webiny/api-tenancy": "^5.23.1",
54
- "@webiny/api-tenancy-so-ddb": "^5.23.1",
55
- "@webiny/cli": "^5.23.1",
56
- "@webiny/db": "^5.23.1",
57
- "@webiny/project-utils": "^5.23.1",
51
+ "@types/yauzl": "^2.9.2",
52
+ "@webiny/api-dynamodb-to-elasticsearch": "^5.25.0-beta.0",
53
+ "@webiny/api-file-manager-ddb-es": "^5.25.0-beta.0",
54
+ "@webiny/api-i18n-ddb": "^5.25.0-beta.0",
55
+ "@webiny/api-security-so-ddb": "^5.25.0-beta.0",
56
+ "@webiny/api-tenancy": "^5.25.0-beta.0",
57
+ "@webiny/api-tenancy-so-ddb": "^5.25.0-beta.0",
58
+ "@webiny/cli": "^5.25.0-beta.0",
59
+ "@webiny/db": "^5.25.0-beta.0",
60
+ "@webiny/project-utils": "^5.25.0-beta.0",
58
61
  "jest": "^26.6.3",
59
62
  "jest-dynalite": "^3.2.0",
60
63
  "rimraf": "^3.0.2",
61
64
  "ttypescript": "^1.5.12",
62
- "typescript": "^4.1.3"
65
+ "typescript": "4.5.5"
63
66
  },
64
67
  "publishConfig": {
65
68
  "access": "public",
@@ -76,5 +79,5 @@
76
79
  ]
77
80
  }
78
81
  },
79
- "gitHead": "a726d09d2647d13e5a4f376cef23463564ef7ca0"
82
+ "gitHead": "2d3e7833575e88fde77d84e5490e746933a5ec28"
80
83
  }
package/types.d.ts CHANGED
@@ -16,6 +16,11 @@ export interface PageImportExportTaskStats {
16
16
  [PageImportExportTaskStatus.FAILED]: number;
17
17
  total: number;
18
18
  }
19
+ interface CreatedBy {
20
+ id: string;
21
+ type: string;
22
+ displayName: string | null;
23
+ }
19
24
  export interface PageImportExportTask {
20
25
  id: string;
21
26
  parent: string;
@@ -25,18 +30,14 @@ export interface PageImportExportTask {
25
30
  error: Record<string, any>;
26
31
  input: Record<string, any>;
27
32
  createdOn: string;
28
- createdBy: {
29
- type: string;
30
- id: string;
31
- displayName: string;
32
- };
33
+ createdBy: CreatedBy;
33
34
  tenant: string;
34
35
  locale: string;
35
36
  }
36
- export declare type File = {
37
+ export interface File {
37
38
  id: string;
38
39
  src: string;
39
- };
40
+ }
40
41
  export interface MetaResponse {
41
42
  cursor: string | null;
42
43
  totalCount: number;