@webiny/api-page-builder-import-export 5.32.0-beta.0 → 5.33.0-beta.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/exportPages/combine/index.js +2 -2
- package/exportPages/combine/index.js.map +1 -1
- package/exportPages/process/index.js +1 -1
- package/exportPages/process/index.js.map +1 -1
- package/exportPages/s3Stream.d.ts +1 -0
- package/exportPages/s3Stream.js +10 -1
- package/exportPages/s3Stream.js.map +1 -1
- package/graphql/crud/pages.crud.js +0 -3
- package/graphql/crud/pages.crud.js.map +1 -1
- package/graphql/graphql/pages.gql.js +1 -5
- package/graphql/graphql/pages.gql.js.map +1 -1
- package/graphql/types.d.ts +1 -2
- package/graphql/types.js.map +1 -1
- package/importPages/create/index.d.ts +1 -2
- package/importPages/create/index.js +3 -6
- package/importPages/create/index.js.map +1 -1
- package/importPages/utils.d.ts +2 -2
- package/importPages/utils.js +9 -27
- package/importPages/utils.js.map +1 -1
- package/package.json +22 -22
@@ -73,7 +73,7 @@ var _default = () => {
|
|
73
73
|
data: {
|
74
74
|
message: `Finish uploading page export.`,
|
75
75
|
key: pageExportUpload.Key,
|
76
|
-
url: pageExportUpload.
|
76
|
+
url: _s3Stream.s3Stream.getPresignedUrl(pageExportUpload.Key)
|
77
77
|
}
|
78
78
|
}); // Remove individual zip files from storage
|
79
79
|
|
@@ -93,7 +93,7 @@ var _default = () => {
|
|
93
93
|
name: e.name,
|
94
94
|
message: e.message,
|
95
95
|
stack: e.stack,
|
96
|
-
code: "
|
96
|
+
code: "EXPORT_FAILED"
|
97
97
|
}
|
98
98
|
});
|
99
99
|
return {
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["createRawEventHandler","payload","context","log","console","pageBuilder","taskId","identity","mockSecurity","task","pageImportExportTask","getTask","data","error","message","exportPagesDataKey","input","listObjectResponse","s3Stream","listObject","Contents","zipFileKeys","filter","file","Key","map","Boolean","zipOfZip","ZipOfZip","pageExportUpload","process","Location","updateTask","status","PageImportExportTaskStatus","COMPLETED","key","url","deleteFilePromises","deleteObject","Promise","all","length","e","FAILED","name","stack","code"],"sources":["index.ts"],"sourcesContent":["import { PageImportExportTaskStatus, PbPageImportExportContext } from \"~/types\";\nimport { s3Stream } from \"../s3Stream\";\nimport { ZipOfZip } from \"../zipper\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { createRawEventHandler } from \"@webiny/handler-aws\";\n\nexport interface Payload {\n taskId: string;\n identity: SecurityIdentity;\n}\n\nexport interface Response {\n data: string | null;\n error: Partial<Error> | null;\n}\n\n/**\n * Handles the export pages combine workflow.\n */\nexport default () => {\n return createRawEventHandler<Payload, PbPageImportExportContext, Response>(\n async ({ payload, context }) => {\n const log = console.log;\n\n log(\"RUNNING Export Pages Combine Handler\");\n const { pageBuilder } = context;\n const { taskId, identity } = payload;\n\n mockSecurity(identity, context);\n\n try {\n const task = await pageBuilder.pageImportExportTask.getTask(taskId);\n if (!task) {\n return {\n data: null,\n error: {\n message: `There is no task with ID \"${taskId}\".`\n }\n };\n }\n\n const { exportPagesDataKey } = task.input;\n\n // Get all files (zip) from given key\n const listObjectResponse = await s3Stream.listObject(exportPagesDataKey);\n if (!listObjectResponse.Contents) {\n return {\n data: null,\n error: {\n message:\n \"There is no Contents defined on S3 Stream while combining pages.\"\n }\n };\n }\n\n const zipFileKeys = listObjectResponse.Contents.filter(\n file => file.Key !== exportPagesDataKey\n )\n .map(file => file.Key)\n .filter(Boolean) as string[];\n\n // Prepare zip of all zips\n const zipOfZip = new ZipOfZip(zipFileKeys);\n\n // Upload\n const pageExportUpload = await zipOfZip.process();\n log(`Done uploading... File is located at ${pageExportUpload.Location} `);\n\n // Update task status and save export page data key\n await pageBuilder.pageImportExportTask.updateTask(taskId, {\n status: PageImportExportTaskStatus.COMPLETED,\n data: {\n message: `Finish uploading page export.`,\n key: pageExportUpload.Key,\n url: pageExportUpload.
|
1
|
+
{"version":3,"names":["createRawEventHandler","payload","context","log","console","pageBuilder","taskId","identity","mockSecurity","task","pageImportExportTask","getTask","data","error","message","exportPagesDataKey","input","listObjectResponse","s3Stream","listObject","Contents","zipFileKeys","filter","file","Key","map","Boolean","zipOfZip","ZipOfZip","pageExportUpload","process","Location","updateTask","status","PageImportExportTaskStatus","COMPLETED","key","url","getPresignedUrl","deleteFilePromises","deleteObject","Promise","all","length","e","FAILED","name","stack","code"],"sources":["index.ts"],"sourcesContent":["import { PageImportExportTaskStatus, PbPageImportExportContext } from \"~/types\";\nimport { s3Stream } from \"../s3Stream\";\nimport { ZipOfZip } from \"../zipper\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { createRawEventHandler } from \"@webiny/handler-aws\";\n\nexport interface Payload {\n taskId: string;\n identity: SecurityIdentity;\n}\n\nexport interface Response {\n data: string | null;\n error: Partial<Error> | null;\n}\n\n/**\n * Handles the export pages combine workflow.\n */\nexport default () => {\n return createRawEventHandler<Payload, PbPageImportExportContext, Response>(\n async ({ payload, context }) => {\n const log = console.log;\n\n log(\"RUNNING Export Pages Combine Handler\");\n const { pageBuilder } = context;\n const { taskId, identity } = payload;\n\n mockSecurity(identity, context);\n\n try {\n const task = await pageBuilder.pageImportExportTask.getTask(taskId);\n if (!task) {\n return {\n data: null,\n error: {\n message: `There is no task with ID \"${taskId}\".`\n }\n };\n }\n\n const { exportPagesDataKey } = task.input;\n\n // Get all files (zip) from given key\n const listObjectResponse = await s3Stream.listObject(exportPagesDataKey);\n if (!listObjectResponse.Contents) {\n return {\n data: null,\n error: {\n message:\n \"There is no Contents defined on S3 Stream while combining pages.\"\n }\n };\n }\n\n const zipFileKeys = listObjectResponse.Contents.filter(\n file => file.Key !== exportPagesDataKey\n )\n .map(file => file.Key)\n .filter(Boolean) as string[];\n\n // Prepare zip of all zips\n const zipOfZip = new ZipOfZip(zipFileKeys);\n\n // Upload\n const pageExportUpload = await zipOfZip.process();\n log(`Done uploading... File is located at ${pageExportUpload.Location} `);\n\n // Update task status and save export page data key\n await pageBuilder.pageImportExportTask.updateTask(taskId, {\n status: PageImportExportTaskStatus.COMPLETED,\n data: {\n message: `Finish uploading page export.`,\n key: pageExportUpload.Key,\n url: s3Stream.getPresignedUrl(pageExportUpload.Key)\n }\n });\n\n // Remove individual zip files from storage\n const deleteFilePromises = zipFileKeys.map(key => s3Stream.deleteObject(key));\n await Promise.all(deleteFilePromises);\n log(`Successfully deleted ${deleteFilePromises.length} zip files.`);\n } catch (e) {\n log(\"[EXPORT_PAGES_COMBINE] Error => \", e);\n\n /**\n * In case of error, we'll update the task status to \"failed\",\n * so that, client can show notify the user appropriately.\n */\n await pageBuilder.pageImportExportTask.updateTask(taskId, {\n status: PageImportExportTaskStatus.FAILED,\n error: {\n name: e.name,\n message: e.message,\n stack: e.stack,\n code: \"EXPORT_FAILED\"\n }\n });\n\n return {\n data: null,\n error: {\n message: e.message\n }\n };\n }\n return {\n data: \"\",\n error: null\n };\n }\n );\n};\n"],"mappings":";;;;;;;AAAA;;AACA;;AACA;;AACA;;AAEA;;AAYA;AACA;AACA;eACe,MAAM;EACjB,OAAO,IAAAA,iCAAA,EACH,OAAO;IAAEC,OAAF;IAAWC;EAAX,CAAP,KAAgC;IAC5B,MAAMC,GAAG,GAAGC,OAAO,CAACD,GAApB;IAEAA,GAAG,CAAC,sCAAD,CAAH;IACA,MAAM;MAAEE;IAAF,IAAkBH,OAAxB;IACA,MAAM;MAAEI,MAAF;MAAUC;IAAV,IAAuBN,OAA7B;IAEA,IAAAO,0BAAA,EAAaD,QAAb,EAAuBL,OAAvB;;IAEA,IAAI;MACA,MAAMO,IAAI,GAAG,MAAMJ,WAAW,CAACK,oBAAZ,CAAiCC,OAAjC,CAAyCL,MAAzC,CAAnB;;MACA,IAAI,CAACG,IAAL,EAAW;QACP,OAAO;UACHG,IAAI,EAAE,IADH;UAEHC,KAAK,EAAE;YACHC,OAAO,EAAG,6BAA4BR,MAAO;UAD1C;QAFJ,CAAP;MAMH;;MAED,MAAM;QAAES;MAAF,IAAyBN,IAAI,CAACO,KAApC,CAXA,CAaA;;MACA,MAAMC,kBAAkB,GAAG,MAAMC,kBAAA,CAASC,UAAT,CAAoBJ,kBAApB,CAAjC;;MACA,IAAI,CAACE,kBAAkB,CAACG,QAAxB,EAAkC;QAC9B,OAAO;UACHR,IAAI,EAAE,IADH;UAEHC,KAAK,EAAE;YACHC,OAAO,EACH;UAFD;QAFJ,CAAP;MAOH;;MAED,MAAMO,WAAW,GAAGJ,kBAAkB,CAACG,QAAnB,CAA4BE,MAA5B,CAChBC,IAAI,IAAIA,IAAI,CAACC,GAAL,KAAaT,kBADL,EAGfU,GAHe,CAGXF,IAAI,IAAIA,IAAI,CAACC,GAHF,EAIfF,MAJe,CAIRI,OAJQ,CAApB,CAzBA,CA+BA;;MACA,MAAMC,QAAQ,GAAG,IAAIC,gBAAJ,CAAaP,WAAb,CAAjB,CAhCA,CAkCA;;MACA,MAAMQ,gBAAgB,GAAG,MAAMF,QAAQ,CAACG,OAAT,EAA/B;MACA3B,GAAG,CAAE,wCAAuC0B,gBAAgB,CAACE,QAAS,GAAnE,CAAH,CApCA,CAsCA;;MACA,MAAM1B,WAAW,CAACK,oBAAZ,CAAiCsB,UAAjC,CAA4C1B,MAA5C,EAAoD;QACtD2B,MAAM,EAAEC,iCAAA,CAA2BC,SADmB;QAEtDvB,IAAI,EAAE;UACFE,OAAO,EAAG,+BADR;UAEFsB,GAAG,EAAEP,gBAAgB,CAACL,GAFpB;UAGFa,GAAG,EAAEnB,kBAAA,CAASoB,eAAT,CAAyBT,gBAAgB,CAACL,GAA1C;QAHH;MAFgD,CAApD,CAAN,CAvCA,CAgDA;;MACA,MAAMe,kBAAkB,GAAGlB,WAAW,CAACI,GAAZ,CAAgBW,GAAG,IAAIlB,kBAAA,CAASsB,YAAT,CAAsBJ,GAAtB,CAAvB,CAA3B;MACA,MAAMK,OAAO,CAACC,GAAR,CAAYH,kBAAZ,CAAN;MACApC,GAAG,CAAE,wBAAuBoC,kBAAkB,CAACI,MAAO,aAAnD,CAAH;IACH,CApDD,CAoDE,OAAOC,CAAP,EAAU;MACRzC,GAAG,CAAC,kCAAD,EAAqCyC,CAArC,CAAH;MAEA;AAChB;AACA;AACA;;MACgB,MAAMvC,WAAW,CAACK,oBAAZ,CAAiCsB,UAAjC,CAA4C1B,MAA5C,EAAoD;QACtD2B,MAAM,EAAEC,iCAAA,CAA2BW,MADmB;QAEtDhC,KAAK,EAAE;UACHiC,IAAI,EAAEF,CAAC,CAACE,IADL;UAEHhC,OAAO,EAAE8B,CAAC,CAAC9B,OAFR;UAGHiC,KAAK,EAAEH,CAAC,CAACG,KAHN;UAIHC,IAAI,EAAE;QAJH;MAF+C,CAApD,CAAN;MAUA,OAAO;QACHpC,IAAI,EAAE,IADH;QAEHC,KAAK,EAAE;UACHC,OAAO,EAAE8B,CAAC,CAAC9B;QADR;MAFJ,CAAP;IAMH;;IACD,OAAO;MACHF,IAAI,EAAE,EADH;MAEHC,KAAK,EAAE;IAFJ,CAAP;EAIH,CA1FE,CAAP;AA4FH,C"}
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["configuration","createRawEventHandler","payload","context","log","console","subTask","noPendingTask","prevStatusOfSubTask","PageImportExportTaskStatus","PENDING","pageBuilder","fileManager","taskId","subTaskIndex","identity","mockSecurity","pageImportExportTask","getSubTask","zeroPad","status","data","error","id","input","pageId","exportPagesDataKey","revisionType","page","PageExportRevisionType","PUBLISHED","getPublishedPageById","getPage","e","NotFoundError","version","updateSubTask","PROCESSING","updateStats","prevStatus","nextStatus","pageDataZip","exportPage","COMPLETED","message","key","Key","FAILED","name","stack","code","invokeHandlerClient","handlers","combine","security","getIdentity","description","process"],"sources":["index.ts"],"sourcesContent":["import {\n PageExportRevisionType,\n PageImportExportTaskStatus,\n PbPageImportExportContext\n} from \"~/types\";\nimport { invokeHandlerClient } from \"~/client\";\nimport { NotFoundError } from \"@webiny/handler-graphql\";\nimport { exportPage } from \"~/exportPages/utils\";\nimport { Payload as ExtractPayload } from \"../combine\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { zeroPad } from \"@webiny/utils\";\nimport { createRawEventHandler } from \"@webiny/handler-aws\";\n\ninterface Configuration {\n handlers: {\n process: string;\n combine: string;\n };\n}\n\nexport interface Payload {\n taskId: string;\n subTaskIndex: number;\n identity?: SecurityIdentity;\n}\n\nexport interface Response {\n data: string | null;\n error: Partial<Error> | null;\n}\n\n/**\n * Handles the export pages process workflow.\n */\nexport default (configuration: Configuration) => {\n return createRawEventHandler<Payload, PbPageImportExportContext, Response>(\n async ({ payload, context }) => {\n const log = console.log;\n let subTask;\n let noPendingTask = true;\n let prevStatusOfSubTask = PageImportExportTaskStatus.PENDING;\n\n log(\"RUNNING Export Pages Process Handler\");\n const { pageBuilder, fileManager } = context;\n const { taskId, subTaskIndex, identity } = payload;\n // Disable authorization; this is necessary because we call Page Builder CRUD methods which include authorization checks\n // and this Lambda is invoked internally, without credentials.\n mockSecurity(identity as SecurityIdentity, context);\n\n try {\n /*\n * Note: We're not going to DB for finding the next sub-task to process,\n * because the data might be out of sync due to GSI eventual consistency.\n */\n subTask = await pageBuilder.pageImportExportTask.getSubTask(\n taskId,\n zeroPad(subTaskIndex, 5)\n );\n /**\n * Base condition!!\n * Bail out early, if task not found or task's status is not \"pending\".\n */\n if (!subTask || subTask.status !== PageImportExportTaskStatus.PENDING) {\n noPendingTask = true;\n return {\n data: \"\",\n error: null\n };\n } else {\n noPendingTask = false;\n }\n\n log(`Fetched sub task => ${subTask.id}`);\n\n const { input } = subTask;\n const { pageId, exportPagesDataKey, revisionType } = input;\n\n /**\n * At the moment, we only export a single revision of the page.\n * It could be \"published\" or \"latest\" depending upon user input.\n *\n * Note: In case of no \"published\" revision available, we use the latest revision.\n */\n let page;\n try {\n if (revisionType === PageExportRevisionType.PUBLISHED) {\n // Get \"published\" page.\n page = await pageBuilder.getPublishedPageById({ id: pageId });\n } else {\n // Get \"latest\" page.\n page = await pageBuilder.getPage(pageId);\n }\n } catch (e) {\n // If we're looking for \"published\" page and doesn't found it, get latest page.\n if (\n revisionType === PageExportRevisionType.PUBLISHED &&\n e instanceof NotFoundError\n ) {\n page = await pageBuilder.getPage(pageId);\n } else {\n throw e;\n }\n }\n\n if (!page) {\n log(`Unable to load page \"${pageId}\"`);\n throw new NotFoundError(`Unable to load page \"${pageId}\"`);\n }\n\n log(`Processing page key \"${pageId}\" | version ${page.version} | ${page.status}`);\n\n // Mark task status as PROCESSING\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.PROCESSING\n });\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.PROCESSING\n });\n prevStatusOfSubTask = subTask.status;\n\n log(`Extracting page data and uploading to storage...`);\n // Extract Page\n const pageDataZip = await exportPage(page, exportPagesDataKey, fileManager);\n log(`Finish uploading zip...`);\n // Update task record in DB\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.COMPLETED,\n data: {\n message: `Finish uploading data for page \"${page.id}\" v${page.version} (${page.status}).`,\n key: pageDataZip.Key\n }\n });\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.COMPLETED\n });\n prevStatusOfSubTask = subTask.status;\n } catch (e) {\n log(\"[EXPORT_PAGES_PROCESS] Error => \", e);\n\n if (subTask && subTask.id) {\n /**\n * In case of error, we'll update the task status to \"failed\",\n * so that, client can show notify the user appropriately.\n */\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(\n taskId,\n subTask.id,\n {\n status: PageImportExportTaskStatus.FAILED,\n error: {\n name: e.name,\n message: e.message,\n stack: e.stack,\n code: \"IMPORT_FAILED\"\n }\n }\n );\n\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.FAILED\n });\n prevStatusOfSubTask = subTask.status;\n }\n\n return {\n data: null,\n error: {\n message: e.message\n }\n };\n } finally {\n // Base condition!\n if (noPendingTask) {\n log(`No pending sub-task for task ${taskId}`);\n // Combine individual page zip files.\n await invokeHandlerClient<ExtractPayload>({\n context,\n name: configuration.handlers.combine,\n payload: {\n taskId,\n identity: context.security.getIdentity()\n },\n description: \"Export pages - combine\"\n });\n } else {\n console.log(`Invoking PROCESS for task \"${subTaskIndex + 1}\"`);\n // We want to continue with Self invocation no matter if current page error out.\n await invokeHandlerClient<Payload>({\n context,\n name: configuration.handlers.process,\n payload: {\n taskId,\n subTaskIndex: subTaskIndex + 1,\n identity: context.security.getIdentity()\n },\n description: \"Export pages - process - subtask\"\n });\n }\n }\n return {\n data: \"\",\n error: null\n };\n }\n );\n};\n"],"mappings":";;;;;;;AAAA;;AAKA;;AACA;;AACA;;AAEA;;AAEA;;AACA;;AAoBA;AACA;AACA;eACgBA,aAAD,IAAkC;EAC7C,OAAO,IAAAC,iCAAA,EACH,OAAO;IAAEC,OAAF;IAAWC;EAAX,CAAP,KAAgC;IAC5B,MAAMC,GAAG,GAAGC,OAAO,CAACD,GAApB;IACA,IAAIE,OAAJ;IACA,IAAIC,aAAa,GAAG,IAApB;IACA,IAAIC,mBAAmB,GAAGC,iCAAA,CAA2BC,OAArD;IAEAN,GAAG,CAAC,sCAAD,CAAH;IACA,MAAM;MAAEO,WAAF;MAAeC;IAAf,IAA+BT,OAArC;IACA,MAAM;MAAEU,MAAF;MAAUC,YAAV;MAAwBC;IAAxB,IAAqCb,OAA3C,CAR4B,CAS5B;IACA;;IACA,IAAAc,0BAAA,EAAaD,QAAb,EAA2CZ,OAA3C;;IAEA,IAAI;MACA;AAChB;AACA;AACA;MACgBG,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCC,UAAjC,CACZL,MADY,EAEZ,IAAAM,eAAA,EAAQL,YAAR,EAAsB,CAAtB,CAFY,CAAhB;MAIA;AAChB;AACA;AACA;;MACgB,IAAI,CAACR,OAAD,IAAYA,OAAO,CAACc,MAAR,KAAmBX,iCAAA,CAA2BC,OAA9D,EAAuE;QACnEH,aAAa,GAAG,IAAhB;QACA,OAAO;UACHc,IAAI,EAAE,EADH;UAEHC,KAAK,EAAE;QAFJ,CAAP;MAIH,CAND,MAMO;QACHf,aAAa,GAAG,KAAhB;MACH;;MAEDH,GAAG,CAAE,uBAAsBE,OAAO,CAACiB,EAAG,EAAnC,CAAH;MAEA,MAAM;QAAEC;MAAF,IAAYlB,OAAlB;MACA,MAAM;QAAEmB,MAAF;QAAUC,kBAAV;QAA8BC;MAA9B,IAA+CH,KAArD;MAEA;AAChB;AACA;AACA;AACA;AACA;;MACgB,IAAII,IAAJ;;MACA,IAAI;QACA,IAAID,YAAY,KAAKE,6BAAA,CAAuBC,SAA5C,EAAuD;UACnD;UACAF,IAAI,GAAG,MAAMjB,WAAW,CAACoB,oBAAZ,CAAiC;YAAER,EAAE,EAAEE;UAAN,CAAjC,CAAb;QACH,CAHD,MAGO;UACH;UACAG,IAAI,GAAG,MAAMjB,WAAW,CAACqB,OAAZ,CAAoBP,MAApB,CAAb;QACH;MACJ,CARD,CAQE,OAAOQ,CAAP,EAAU;QACR;QACA,IACIN,YAAY,KAAKE,6BAAA,CAAuBC,SAAxC,IACAG,CAAC,YAAYC,6BAFjB,EAGE;UACEN,IAAI,GAAG,MAAMjB,WAAW,CAACqB,OAAZ,CAAoBP,MAApB,CAAb;QACH,CALD,MAKO;UACH,MAAMQ,CAAN;QACH;MACJ;;MAED,IAAI,CAACL,IAAL,EAAW;QACPxB,GAAG,CAAE,wBAAuBqB,MAAO,GAAhC,CAAH;QACA,MAAM,IAAIS,6BAAJ,CAAmB,wBAAuBT,MAAO,GAAjD,CAAN;MACH;;MAEDrB,GAAG,CAAE,wBAAuBqB,MAAO,eAAcG,IAAI,CAACO,OAAQ,MAAKP,IAAI,CAACR,MAAO,EAA5E,CAAH,CA5DA,CA8DA;;MACAd,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCmB,aAAjC,CAA+CvB,MAA/C,EAAuDP,OAAO,CAACiB,EAA/D,EAAmE;QAC/EH,MAAM,EAAEX,iCAAA,CAA2B4B;MAD4C,CAAnE,CAAhB,CA/DA,CAkEA;;MACA,MAAM1B,WAAW,CAACM,oBAAZ,CAAiCqB,WAAjC,CAA6CzB,MAA7C,EAAqD;QACvD0B,UAAU,EAAE/B,mBAD2C;QAEvDgC,UAAU,EAAE/B,iCAAA,CAA2B4B;MAFgB,CAArD,CAAN;MAIA7B,mBAAmB,GAAGF,OAAO,CAACc,MAA9B;MAEAhB,GAAG,CAAE,kDAAF,CAAH,CAzEA,CA0EA;;MACA,MAAMqC,WAAW,GAAG,MAAM,IAAAC,iBAAA,EAAWd,IAAX,EAAiBF,kBAAjB,EAAqCd,WAArC,CAA1B;MACAR,GAAG,CAAE,yBAAF,CAAH,CA5EA,CA6EA;;MACAE,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCmB,aAAjC,CAA+CvB,MAA/C,EAAuDP,OAAO,CAACiB,EAA/D,EAAmE;QAC/EH,MAAM,EAAEX,iCAAA,CAA2BkC,SAD4C;QAE/EtB,IAAI,EAAE;UACFuB,OAAO,EAAG,mCAAkChB,IAAI,CAACL,EAAG,MAAKK,IAAI,CAACO,OAAQ,KAAIP,IAAI,CAACR,MAAO,IADpF;UAEFyB,GAAG,EAAEJ,WAAW,CAACK;QAFf;MAFyE,CAAnE,CAAhB,CA9EA,CAqFA;;MACA,MAAMnC,WAAW,CAACM,oBAAZ,CAAiCqB,WAAjC,CAA6CzB,MAA7C,EAAqD;QACvD0B,UAAU,EAAE/B,mBAD2C;QAEvDgC,UAAU,EAAE/B,iCAAA,CAA2BkC;MAFgB,CAArD,CAAN;MAIAnC,mBAAmB,GAAGF,OAAO,CAACc,MAA9B;IACH,CA3FD,CA2FE,OAAOa,CAAP,EAAU;MACR7B,GAAG,CAAC,kCAAD,EAAqC6B,CAArC,CAAH;;MAEA,IAAI3B,OAAO,IAAIA,OAAO,CAACiB,EAAvB,EAA2B;QACvB;AACpB;AACA;AACA;QACoBjB,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCmB,aAAjC,CACZvB,MADY,EAEZP,OAAO,CAACiB,EAFI,EAGZ;UACIH,MAAM,EAAEX,iCAAA,CAA2BsC,MADvC;UAEIzB,KAAK,EAAE;YACH0B,IAAI,EAAEf,CAAC,CAACe,IADL;YAEHJ,OAAO,EAAEX,CAAC,CAACW,OAFR;YAGHK,KAAK,EAAEhB,CAAC,CAACgB,KAHN;YAIHC,IAAI,EAAE;UAJH;QAFX,CAHY,CAAhB,CALuB,CAmBvB;;QACA,MAAMvC,WAAW,CAACM,oBAAZ,CAAiCqB,WAAjC,CAA6CzB,MAA7C,EAAqD;UACvD0B,UAAU,EAAE/B,mBAD2C;UAEvDgC,UAAU,EAAE/B,iCAAA,CAA2BsC;QAFgB,CAArD,CAAN;QAIAvC,mBAAmB,GAAGF,OAAO,CAACc,MAA9B;MACH;;MAED,OAAO;QACHC,IAAI,EAAE,IADH;QAEHC,KAAK,EAAE;UACHsB,OAAO,EAAEX,CAAC,CAACW;QADR;MAFJ,CAAP;IAMH,CA/HD,SA+HU;MACN;MACA,IAAIrC,aAAJ,EAAmB;QACfH,GAAG,CAAE,gCAA+BS,MAAO,EAAxC,CAAH,CADe,CAEf;;QACA,MAAM,IAAAsC,2BAAA,EAAoC;UACtChD,OADsC;UAEtC6C,IAAI,EAAEhD,aAAa,CAACoD,QAAd,CAAuBC,OAFS;UAGtCnD,OAAO,EAAE;YACLW,MADK;YAELE,QAAQ,EAAEZ,OAAO,CAACmD,QAAR,CAAiBC,WAAjB;UAFL,CAH6B;UAOtCC,WAAW,EAAE;QAPyB,CAApC,CAAN;MASH,CAZD,MAYO;QACHnD,OAAO,CAACD,GAAR,CAAa,8BAA6BU,YAAY,GAAG,CAAE,GAA3D,EADG,CAEH;;QACA,MAAM,IAAAqC,2BAAA,EAA6B;UAC/BhD,OAD+B;UAE/B6C,IAAI,EAAEhD,aAAa,CAACoD,QAAd,CAAuBK,OAFE;UAG/BvD,OAAO,EAAE;YACLW,MADK;YAELC,YAAY,EAAEA,YAAY,GAAG,CAFxB;YAGLC,QAAQ,EAAEZ,OAAO,CAACmD,QAAR,CAAiBC,WAAjB;UAHL,CAHsB;UAQ/BC,WAAW,EAAE;QARkB,CAA7B,CAAN;MAUH;IACJ;;IACD,OAAO;MACHnC,IAAI,EAAE,EADH;MAEHC,KAAK,EAAE;IAFJ,CAAP;EAIH,CA9KE,CAAP;AAgLH,C"}
|
1
|
+
{"version":3,"names":["configuration","createRawEventHandler","payload","context","log","console","subTask","noPendingTask","prevStatusOfSubTask","PageImportExportTaskStatus","PENDING","pageBuilder","fileManager","taskId","subTaskIndex","identity","mockSecurity","pageImportExportTask","getSubTask","zeroPad","status","data","error","id","input","pageId","exportPagesDataKey","revisionType","page","PageExportRevisionType","PUBLISHED","getPublishedPageById","getPage","e","NotFoundError","version","updateSubTask","PROCESSING","updateStats","prevStatus","nextStatus","pageDataZip","exportPage","COMPLETED","message","key","Key","FAILED","name","stack","code","invokeHandlerClient","handlers","combine","security","getIdentity","description","process"],"sources":["index.ts"],"sourcesContent":["import {\n PageExportRevisionType,\n PageImportExportTaskStatus,\n PbPageImportExportContext\n} from \"~/types\";\nimport { invokeHandlerClient } from \"~/client\";\nimport { NotFoundError } from \"@webiny/handler-graphql\";\nimport { exportPage } from \"~/exportPages/utils\";\nimport { Payload as ExtractPayload } from \"../combine\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { zeroPad } from \"@webiny/utils\";\nimport { createRawEventHandler } from \"@webiny/handler-aws\";\n\ninterface Configuration {\n handlers: {\n process: string;\n combine: string;\n };\n}\n\nexport interface Payload {\n taskId: string;\n subTaskIndex: number;\n identity?: SecurityIdentity;\n}\n\nexport interface Response {\n data: string | null;\n error: Partial<Error> | null;\n}\n\n/**\n * Handles the export pages process workflow.\n */\nexport default (configuration: Configuration) => {\n return createRawEventHandler<Payload, PbPageImportExportContext, Response>(\n async ({ payload, context }) => {\n const log = console.log;\n let subTask;\n let noPendingTask = true;\n let prevStatusOfSubTask = PageImportExportTaskStatus.PENDING;\n\n log(\"RUNNING Export Pages Process Handler\");\n const { pageBuilder, fileManager } = context;\n const { taskId, subTaskIndex, identity } = payload;\n // Disable authorization; this is necessary because we call Page Builder CRUD methods which include authorization checks\n // and this Lambda is invoked internally, without credentials.\n mockSecurity(identity as SecurityIdentity, context);\n\n try {\n /*\n * Note: We're not going to DB for finding the next sub-task to process,\n * because the data might be out of sync due to GSI eventual consistency.\n */\n subTask = await pageBuilder.pageImportExportTask.getSubTask(\n taskId,\n zeroPad(subTaskIndex, 5)\n );\n /**\n * Base condition!!\n * Bail out early, if task not found or task's status is not \"pending\".\n */\n if (!subTask || subTask.status !== PageImportExportTaskStatus.PENDING) {\n noPendingTask = true;\n return {\n data: \"\",\n error: null\n };\n } else {\n noPendingTask = false;\n }\n\n log(`Fetched sub task => ${subTask.id}`);\n\n const { input } = subTask;\n const { pageId, exportPagesDataKey, revisionType } = input;\n\n /**\n * At the moment, we only export a single revision of the page.\n * It could be \"published\" or \"latest\" depending upon user input.\n *\n * Note: In case of no \"published\" revision available, we use the latest revision.\n */\n let page;\n try {\n if (revisionType === PageExportRevisionType.PUBLISHED) {\n // Get \"published\" page.\n page = await pageBuilder.getPublishedPageById({ id: pageId });\n } else {\n // Get \"latest\" page.\n page = await pageBuilder.getPage(pageId);\n }\n } catch (e) {\n // If we're looking for \"published\" page and doesn't found it, get latest page.\n if (\n revisionType === PageExportRevisionType.PUBLISHED &&\n e instanceof NotFoundError\n ) {\n page = await pageBuilder.getPage(pageId);\n } else {\n throw e;\n }\n }\n\n if (!page) {\n log(`Unable to load page \"${pageId}\"`);\n throw new NotFoundError(`Unable to load page \"${pageId}\"`);\n }\n\n log(`Processing page key \"${pageId}\" | version ${page.version} | ${page.status}`);\n\n // Mark task status as PROCESSING\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.PROCESSING\n });\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.PROCESSING\n });\n prevStatusOfSubTask = subTask.status;\n\n log(`Extracting page data and uploading to storage...`);\n // Extract Page\n const pageDataZip = await exportPage(page, exportPagesDataKey, fileManager);\n log(`Finish uploading zip...`);\n // Update task record in DB\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(taskId, subTask.id, {\n status: PageImportExportTaskStatus.COMPLETED,\n data: {\n message: `Finish uploading data for page \"${page.id}\" v${page.version} (${page.status}).`,\n key: pageDataZip.Key\n }\n });\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.COMPLETED\n });\n prevStatusOfSubTask = subTask.status;\n } catch (e) {\n log(\"[EXPORT_PAGES_PROCESS] Error => \", e);\n\n if (subTask && subTask.id) {\n /**\n * In case of error, we'll update the task status to \"failed\",\n * so that, client can show notify the user appropriately.\n */\n subTask = await pageBuilder.pageImportExportTask.updateSubTask(\n taskId,\n subTask.id,\n {\n status: PageImportExportTaskStatus.FAILED,\n error: {\n name: e.name,\n message: e.message,\n stack: e.stack,\n code: \"EXPORT_FAILED\"\n }\n }\n );\n\n // Update stats in main task\n await pageBuilder.pageImportExportTask.updateStats(taskId, {\n prevStatus: prevStatusOfSubTask,\n nextStatus: PageImportExportTaskStatus.FAILED\n });\n prevStatusOfSubTask = subTask.status;\n }\n\n return {\n data: null,\n error: {\n message: e.message\n }\n };\n } finally {\n // Base condition!\n if (noPendingTask) {\n log(`No pending sub-task for task ${taskId}`);\n // Combine individual page zip files.\n await invokeHandlerClient<ExtractPayload>({\n context,\n name: configuration.handlers.combine,\n payload: {\n taskId,\n identity: context.security.getIdentity()\n },\n description: \"Export pages - combine\"\n });\n } else {\n console.log(`Invoking PROCESS for task \"${subTaskIndex + 1}\"`);\n // We want to continue with Self invocation no matter if current page error out.\n await invokeHandlerClient<Payload>({\n context,\n name: configuration.handlers.process,\n payload: {\n taskId,\n subTaskIndex: subTaskIndex + 1,\n identity: context.security.getIdentity()\n },\n description: \"Export pages - process - subtask\"\n });\n }\n }\n return {\n data: \"\",\n error: null\n };\n }\n );\n};\n"],"mappings":";;;;;;;AAAA;;AAKA;;AACA;;AACA;;AAEA;;AAEA;;AACA;;AAoBA;AACA;AACA;eACgBA,aAAD,IAAkC;EAC7C,OAAO,IAAAC,iCAAA,EACH,OAAO;IAAEC,OAAF;IAAWC;EAAX,CAAP,KAAgC;IAC5B,MAAMC,GAAG,GAAGC,OAAO,CAACD,GAApB;IACA,IAAIE,OAAJ;IACA,IAAIC,aAAa,GAAG,IAApB;IACA,IAAIC,mBAAmB,GAAGC,iCAAA,CAA2BC,OAArD;IAEAN,GAAG,CAAC,sCAAD,CAAH;IACA,MAAM;MAAEO,WAAF;MAAeC;IAAf,IAA+BT,OAArC;IACA,MAAM;MAAEU,MAAF;MAAUC,YAAV;MAAwBC;IAAxB,IAAqCb,OAA3C,CAR4B,CAS5B;IACA;;IACA,IAAAc,0BAAA,EAAaD,QAAb,EAA2CZ,OAA3C;;IAEA,IAAI;MACA;AAChB;AACA;AACA;MACgBG,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCC,UAAjC,CACZL,MADY,EAEZ,IAAAM,eAAA,EAAQL,YAAR,EAAsB,CAAtB,CAFY,CAAhB;MAIA;AAChB;AACA;AACA;;MACgB,IAAI,CAACR,OAAD,IAAYA,OAAO,CAACc,MAAR,KAAmBX,iCAAA,CAA2BC,OAA9D,EAAuE;QACnEH,aAAa,GAAG,IAAhB;QACA,OAAO;UACHc,IAAI,EAAE,EADH;UAEHC,KAAK,EAAE;QAFJ,CAAP;MAIH,CAND,MAMO;QACHf,aAAa,GAAG,KAAhB;MACH;;MAEDH,GAAG,CAAE,uBAAsBE,OAAO,CAACiB,EAAG,EAAnC,CAAH;MAEA,MAAM;QAAEC;MAAF,IAAYlB,OAAlB;MACA,MAAM;QAAEmB,MAAF;QAAUC,kBAAV;QAA8BC;MAA9B,IAA+CH,KAArD;MAEA;AAChB;AACA;AACA;AACA;AACA;;MACgB,IAAII,IAAJ;;MACA,IAAI;QACA,IAAID,YAAY,KAAKE,6BAAA,CAAuBC,SAA5C,EAAuD;UACnD;UACAF,IAAI,GAAG,MAAMjB,WAAW,CAACoB,oBAAZ,CAAiC;YAAER,EAAE,EAAEE;UAAN,CAAjC,CAAb;QACH,CAHD,MAGO;UACH;UACAG,IAAI,GAAG,MAAMjB,WAAW,CAACqB,OAAZ,CAAoBP,MAApB,CAAb;QACH;MACJ,CARD,CAQE,OAAOQ,CAAP,EAAU;QACR;QACA,IACIN,YAAY,KAAKE,6BAAA,CAAuBC,SAAxC,IACAG,CAAC,YAAYC,6BAFjB,EAGE;UACEN,IAAI,GAAG,MAAMjB,WAAW,CAACqB,OAAZ,CAAoBP,MAApB,CAAb;QACH,CALD,MAKO;UACH,MAAMQ,CAAN;QACH;MACJ;;MAED,IAAI,CAACL,IAAL,EAAW;QACPxB,GAAG,CAAE,wBAAuBqB,MAAO,GAAhC,CAAH;QACA,MAAM,IAAIS,6BAAJ,CAAmB,wBAAuBT,MAAO,GAAjD,CAAN;MACH;;MAEDrB,GAAG,CAAE,wBAAuBqB,MAAO,eAAcG,IAAI,CAACO,OAAQ,MAAKP,IAAI,CAACR,MAAO,EAA5E,CAAH,CA5DA,CA8DA;;MACAd,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCmB,aAAjC,CAA+CvB,MAA/C,EAAuDP,OAAO,CAACiB,EAA/D,EAAmE;QAC/EH,MAAM,EAAEX,iCAAA,CAA2B4B;MAD4C,CAAnE,CAAhB,CA/DA,CAkEA;;MACA,MAAM1B,WAAW,CAACM,oBAAZ,CAAiCqB,WAAjC,CAA6CzB,MAA7C,EAAqD;QACvD0B,UAAU,EAAE/B,mBAD2C;QAEvDgC,UAAU,EAAE/B,iCAAA,CAA2B4B;MAFgB,CAArD,CAAN;MAIA7B,mBAAmB,GAAGF,OAAO,CAACc,MAA9B;MAEAhB,GAAG,CAAE,kDAAF,CAAH,CAzEA,CA0EA;;MACA,MAAMqC,WAAW,GAAG,MAAM,IAAAC,iBAAA,EAAWd,IAAX,EAAiBF,kBAAjB,EAAqCd,WAArC,CAA1B;MACAR,GAAG,CAAE,yBAAF,CAAH,CA5EA,CA6EA;;MACAE,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCmB,aAAjC,CAA+CvB,MAA/C,EAAuDP,OAAO,CAACiB,EAA/D,EAAmE;QAC/EH,MAAM,EAAEX,iCAAA,CAA2BkC,SAD4C;QAE/EtB,IAAI,EAAE;UACFuB,OAAO,EAAG,mCAAkChB,IAAI,CAACL,EAAG,MAAKK,IAAI,CAACO,OAAQ,KAAIP,IAAI,CAACR,MAAO,IADpF;UAEFyB,GAAG,EAAEJ,WAAW,CAACK;QAFf;MAFyE,CAAnE,CAAhB,CA9EA,CAqFA;;MACA,MAAMnC,WAAW,CAACM,oBAAZ,CAAiCqB,WAAjC,CAA6CzB,MAA7C,EAAqD;QACvD0B,UAAU,EAAE/B,mBAD2C;QAEvDgC,UAAU,EAAE/B,iCAAA,CAA2BkC;MAFgB,CAArD,CAAN;MAIAnC,mBAAmB,GAAGF,OAAO,CAACc,MAA9B;IACH,CA3FD,CA2FE,OAAOa,CAAP,EAAU;MACR7B,GAAG,CAAC,kCAAD,EAAqC6B,CAArC,CAAH;;MAEA,IAAI3B,OAAO,IAAIA,OAAO,CAACiB,EAAvB,EAA2B;QACvB;AACpB;AACA;AACA;QACoBjB,OAAO,GAAG,MAAMK,WAAW,CAACM,oBAAZ,CAAiCmB,aAAjC,CACZvB,MADY,EAEZP,OAAO,CAACiB,EAFI,EAGZ;UACIH,MAAM,EAAEX,iCAAA,CAA2BsC,MADvC;UAEIzB,KAAK,EAAE;YACH0B,IAAI,EAAEf,CAAC,CAACe,IADL;YAEHJ,OAAO,EAAEX,CAAC,CAACW,OAFR;YAGHK,KAAK,EAAEhB,CAAC,CAACgB,KAHN;YAIHC,IAAI,EAAE;UAJH;QAFX,CAHY,CAAhB,CALuB,CAmBvB;;QACA,MAAMvC,WAAW,CAACM,oBAAZ,CAAiCqB,WAAjC,CAA6CzB,MAA7C,EAAqD;UACvD0B,UAAU,EAAE/B,mBAD2C;UAEvDgC,UAAU,EAAE/B,iCAAA,CAA2BsC;QAFgB,CAArD,CAAN;QAIAvC,mBAAmB,GAAGF,OAAO,CAACc,MAA9B;MACH;;MAED,OAAO;QACHC,IAAI,EAAE,IADH;QAEHC,KAAK,EAAE;UACHsB,OAAO,EAAEX,CAAC,CAACW;QADR;MAFJ,CAAP;IAMH,CA/HD,SA+HU;MACN;MACA,IAAIrC,aAAJ,EAAmB;QACfH,GAAG,CAAE,gCAA+BS,MAAO,EAAxC,CAAH,CADe,CAEf;;QACA,MAAM,IAAAsC,2BAAA,EAAoC;UACtChD,OADsC;UAEtC6C,IAAI,EAAEhD,aAAa,CAACoD,QAAd,CAAuBC,OAFS;UAGtCnD,OAAO,EAAE;YACLW,MADK;YAELE,QAAQ,EAAEZ,OAAO,CAACmD,QAAR,CAAiBC,WAAjB;UAFL,CAH6B;UAOtCC,WAAW,EAAE;QAPyB,CAApC,CAAN;MASH,CAZD,MAYO;QACHnD,OAAO,CAACD,GAAR,CAAa,8BAA6BU,YAAY,GAAG,CAAE,GAA3D,EADG,CAEH;;QACA,MAAM,IAAAqC,2BAAA,EAA6B;UAC/BhD,OAD+B;UAE/B6C,IAAI,EAAEhD,aAAa,CAACoD,QAAd,CAAuBK,OAFE;UAG/BvD,OAAO,EAAE;YACLW,MADK;YAELC,YAAY,EAAEA,YAAY,GAAG,CAFxB;YAGLC,QAAQ,EAAEZ,OAAO,CAACmD,QAAR,CAAiBC,WAAjB;UAHL,CAHsB;UAQ/BC,WAAW,EAAE;QARkB,CAA7B,CAAN;MAUH;IACJ;;IACD,OAAO;MACHnC,IAAI,EAAE,EADH;MAEHC,KAAK,EAAE;IAFJ,CAAP;EAIH,CA9KE,CAAP;AAgLH,C"}
|
@@ -6,6 +6,7 @@ declare class S3Stream {
|
|
6
6
|
s3: S3;
|
7
7
|
bucket: string;
|
8
8
|
constructor();
|
9
|
+
getPresignedUrl(key: string): string;
|
9
10
|
/**
|
10
11
|
* We're checking if the file is accessible on S3 by getting object meta data.
|
11
12
|
* It help us to filter files that we need to download as part of export data.
|
package/exportPages/s3Stream.js
CHANGED
@@ -28,6 +28,15 @@ class S3Stream {
|
|
28
28
|
});
|
29
29
|
this.bucket = process.env.S3_BUCKET;
|
30
30
|
}
|
31
|
+
|
32
|
+
getPresignedUrl(key) {
|
33
|
+
return this.s3.getSignedUrl("getObject", {
|
34
|
+
Bucket: this.bucket,
|
35
|
+
Key: key,
|
36
|
+
Expires: 604800 // 1 week
|
37
|
+
|
38
|
+
});
|
39
|
+
}
|
31
40
|
/**
|
32
41
|
* We're checking if the file is accessible on S3 by getting object meta data.
|
33
42
|
* It help us to filter files that we need to download as part of export data.
|
@@ -63,7 +72,7 @@ class S3Stream {
|
|
63
72
|
writeStream(Key, contentType = ARCHIVE_CONTENT_TYPE) {
|
64
73
|
const streamPassThrough = new _stream.Stream.PassThrough();
|
65
74
|
const params = {
|
66
|
-
ACL: "
|
75
|
+
ACL: "private",
|
67
76
|
Body: streamPassThrough,
|
68
77
|
Bucket: this.bucket,
|
69
78
|
ContentType: contentType,
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["ARCHIVE_CONTENT_TYPE","S3Stream","constructor","s3","S3","region","process","env","AWS_REGION","bucket","S3_BUCKET","
|
1
|
+
{"version":3,"names":["ARCHIVE_CONTENT_TYPE","S3Stream","constructor","s3","S3","region","process","env","AWS_REGION","bucket","S3_BUCKET","getPresignedUrl","key","getSignedUrl","Bucket","Key","Expires","isFileAccessible","getObjectHead","error","console","warn","log","headObject","promise","readStream","getObject","createReadStream","writeStream","contentType","streamPassThrough","Stream","PassThrough","params","ACL","Body","ContentType","streamPassThroughUploadPromise","upload","listObject","prefix","listObjects","Prefix","deleteObject","s3Stream"],"sources":["s3Stream.ts"],"sourcesContent":["import { Stream, PassThrough, Readable } from \"stream\";\nimport S3 from \"aws-sdk/clients/s3\";\n\nconst ARCHIVE_CONTENT_TYPE = \"application/zip\";\n\nclass S3Stream {\n s3: S3;\n bucket: string;\n\n constructor() {\n this.s3 = new S3({\n region: process.env.AWS_REGION as string\n });\n this.bucket = process.env.S3_BUCKET as string;\n }\n\n getPresignedUrl(key: string) {\n return this.s3.getSignedUrl(\"getObject\", {\n Bucket: this.bucket,\n Key: key,\n Expires: 604800 // 1 week\n });\n }\n\n /**\n * We're checking if the file is accessible on S3 by getting object meta data.\n * It help us to filter files that we need to download as part of export data.\n * @param Key {string}\n */\n async isFileAccessible(Key: string): Promise<boolean> {\n try {\n await this.getObjectHead(Key);\n return true;\n } catch (error) {\n console.warn(`Error while fetching meta data for file \"${Key}\"`);\n console.log(error);\n return false;\n }\n }\n\n getObjectHead(Key: string): Promise<S3.HeadObjectOutput> {\n return this.s3.headObject({ Bucket: this.bucket, Key }).promise();\n }\n\n readStream(Key: string): Readable {\n return this.s3.getObject({ Bucket: this.bucket, Key }).createReadStream();\n }\n\n writeStream(\n Key: string,\n contentType: string = ARCHIVE_CONTENT_TYPE\n ): {\n streamPassThrough: PassThrough;\n streamPassThroughUploadPromise: Promise<S3.ManagedUpload.SendData>;\n } {\n const streamPassThrough = new Stream.PassThrough();\n\n const params: S3.PutObjectRequest = {\n ACL: \"private\",\n Body: streamPassThrough,\n Bucket: this.bucket,\n ContentType: contentType,\n Key\n };\n\n return {\n streamPassThrough: streamPassThrough,\n /**\n * We're not using the `FileManager` storage plugin here because it currently doesn't support streams.\n */\n streamPassThroughUploadPromise: this.s3.upload(params).promise()\n };\n }\n\n upload(params: {\n Key: string;\n ContentType: string;\n Body: Buffer;\n }): Promise<S3.ManagedUpload.SendData> {\n return this.s3\n .upload({\n ACL: \"private\",\n Bucket: this.bucket,\n ...params\n })\n .promise();\n }\n\n listObject(prefix: string): Promise<S3.ListObjectsOutput> {\n return this.s3\n .listObjects({\n Bucket: this.bucket,\n Prefix: prefix\n })\n .promise();\n }\n\n deleteObject(key: string): Promise<S3.DeleteObjectOutput> {\n return this.s3.deleteObject({ Key: key, Bucket: this.bucket }).promise();\n }\n}\n\nexport const s3Stream = new S3Stream();\n"],"mappings":";;;;;;;;;;;AAAA;;AACA;;;;;;AAEA,MAAMA,oBAAoB,GAAG,iBAA7B;;AAEA,MAAMC,QAAN,CAAe;EAIXC,WAAW,GAAG;IAAA;IAAA;IACV,KAAKC,EAAL,GAAU,IAAIC,UAAJ,CAAO;MACbC,MAAM,EAAEC,OAAO,CAACC,GAAR,CAAYC;IADP,CAAP,CAAV;IAGA,KAAKC,MAAL,GAAcH,OAAO,CAACC,GAAR,CAAYG,SAA1B;EACH;;EAEDC,eAAe,CAACC,GAAD,EAAc;IACzB,OAAO,KAAKT,EAAL,CAAQU,YAAR,CAAqB,WAArB,EAAkC;MACrCC,MAAM,EAAE,KAAKL,MADwB;MAErCM,GAAG,EAAEH,GAFgC;MAGrCI,OAAO,EAAE,MAH4B,CAGrB;;IAHqB,CAAlC,CAAP;EAKH;EAED;AACJ;AACA;AACA;AACA;;;EAC0B,MAAhBC,gBAAgB,CAACF,GAAD,EAAgC;IAClD,IAAI;MACA,MAAM,KAAKG,aAAL,CAAmBH,GAAnB,CAAN;MACA,OAAO,IAAP;IACH,CAHD,CAGE,OAAOI,KAAP,EAAc;MACZC,OAAO,CAACC,IAAR,CAAc,4CAA2CN,GAAI,GAA7D;MACAK,OAAO,CAACE,GAAR,CAAYH,KAAZ;MACA,OAAO,KAAP;IACH;EACJ;;EAEDD,aAAa,CAACH,GAAD,EAA4C;IACrD,OAAO,KAAKZ,EAAL,CAAQoB,UAAR,CAAmB;MAAET,MAAM,EAAE,KAAKL,MAAf;MAAuBM;IAAvB,CAAnB,EAAiDS,OAAjD,EAAP;EACH;;EAEDC,UAAU,CAACV,GAAD,EAAwB;IAC9B,OAAO,KAAKZ,EAAL,CAAQuB,SAAR,CAAkB;MAAEZ,MAAM,EAAE,KAAKL,MAAf;MAAuBM;IAAvB,CAAlB,EAAgDY,gBAAhD,EAAP;EACH;;EAEDC,WAAW,CACPb,GADO,EAEPc,WAAmB,GAAG7B,oBAFf,EAMT;IACE,MAAM8B,iBAAiB,GAAG,IAAIC,cAAA,CAAOC,WAAX,EAA1B;IAEA,MAAMC,MAA2B,GAAG;MAChCC,GAAG,EAAE,SAD2B;MAEhCC,IAAI,EAAEL,iBAF0B;MAGhChB,MAAM,EAAE,KAAKL,MAHmB;MAIhC2B,WAAW,EAAEP,WAJmB;MAKhCd;IALgC,CAApC;IAQA,OAAO;MACHe,iBAAiB,EAAEA,iBADhB;;MAEH;AACZ;AACA;MACYO,8BAA8B,EAAE,KAAKlC,EAAL,CAAQmC,MAAR,CAAeL,MAAf,EAAuBT,OAAvB;IAL7B,CAAP;EAOH;;EAEDc,MAAM,CAACL,MAAD,EAIiC;IACnC,OAAO,KAAK9B,EAAL,CACFmC,MADE;MAECJ,GAAG,EAAE,SAFN;MAGCpB,MAAM,EAAE,KAAKL;IAHd,GAIIwB,MAJJ,GAMFT,OANE,EAAP;EAOH;;EAEDe,UAAU,CAACC,MAAD,EAAgD;IACtD,OAAO,KAAKrC,EAAL,CACFsC,WADE,CACU;MACT3B,MAAM,EAAE,KAAKL,MADJ;MAETiC,MAAM,EAAEF;IAFC,CADV,EAKFhB,OALE,EAAP;EAMH;;EAEDmB,YAAY,CAAC/B,GAAD,EAA8C;IACtD,OAAO,KAAKT,EAAL,CAAQwC,YAAR,CAAqB;MAAE5B,GAAG,EAAEH,GAAP;MAAYE,MAAM,EAAE,KAAKL;IAAzB,CAArB,EAAwDe,OAAxD,EAAP;EACH;;AA9FU;;AAiGR,MAAMoB,QAAQ,GAAG,IAAI3C,QAAJ,EAAjB"}
|
@@ -33,7 +33,6 @@ var _default = new _api.ContextPlugin(context => {
|
|
33
33
|
const importExportCrud = {
|
34
34
|
async importPages({
|
35
35
|
category: categorySlug,
|
36
|
-
zipFileKey,
|
37
36
|
zipFileUrl
|
38
37
|
}) {
|
39
38
|
await (0, _checkBasePermissions.default)(context, PERMISSION_NAME, {
|
@@ -51,7 +50,6 @@ var _default = new _api.ContextPlugin(context => {
|
|
51
50
|
status: _types.PageImportExportTaskStatus.PENDING,
|
52
51
|
input: {
|
53
52
|
category: categorySlug,
|
54
|
-
zipFileKey,
|
55
53
|
zipFileUrl
|
56
54
|
}
|
57
55
|
});
|
@@ -66,7 +64,6 @@ var _default = new _api.ContextPlugin(context => {
|
|
66
64
|
name: IMPORT_PAGES_CREATE_HANDLER,
|
67
65
|
payload: {
|
68
66
|
category: categorySlug,
|
69
|
-
zipFileKey,
|
70
67
|
zipFileUrl,
|
71
68
|
task,
|
72
69
|
identity: context.security.getIdentity()
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["PERMISSION_NAME","EXPORT_PAGES_PROCESS_HANDLER","process","env","IMPORT_PAGES_CREATE_HANDLER","ContextPlugin","context","importExportCrud","importPages","category","categorySlug","zipFileKey","zipFileUrl","checkBasePermissions","rwd","pageBuilder","getCategory","NotFoundError","task","pageImportExportTask","createTask","status","PageImportExportTaskStatus","PENDING","input","invokeHandlerClient","name","payload","identity","security","getIdentity","description","exportPages","ids","initialPageIds","revisionType","where","sort","search","pageIds","Array","isArray","length","pages","meta","hasMoreItems","cursor","totalCount","listLatestPages","after","forEach","page","push","id","WebinyError","exportPagesDataKey","EXPORT_PAGES_FOLDER_KEY","i","pageId","createSubTask","zeroPad","updateTask","PROCESSING","stats","initialStats","taskId","subTaskIndex"],"sources":["pages.crud.ts"],"sourcesContent":["import WebinyError from \"@webiny/error\";\nimport { NotFoundError } from \"@webiny/handler-graphql\";\nimport { ContextPlugin } from \"@webiny/api\";\nimport checkBasePermissions from \"@webiny/api-page-builder/graphql/crud/utils/checkBasePermissions\";\nimport {\n PageImportExportTaskStatus,\n PagesImportExportCrud,\n PbPageImportExportContext\n} from \"~/types\";\nimport { invokeHandlerClient } from \"~/client\";\nimport { Payload as CreateHandlerPayload } from \"~/importPages/create\";\nimport { initialStats } from \"~/importPages/utils\";\nimport { Payload as ExportPagesProcessHandlerPayload } from \"~/exportPages/process\";\nimport { EXPORT_PAGES_FOLDER_KEY } from \"~/exportPages/utils\";\nimport { MetaResponse } from \"@webiny/api-page-builder/types\";\nimport { zeroPad } from \"@webiny/utils\";\n\nconst PERMISSION_NAME = \"pb.page\";\nconst EXPORT_PAGES_PROCESS_HANDLER = process.env.EXPORT_PAGES_PROCESS_HANDLER as string;\nconst IMPORT_PAGES_CREATE_HANDLER = process.env.IMPORT_PAGES_CREATE_HANDLER as string;\n\nexport default new ContextPlugin<PbPageImportExportContext>(context => {\n const importExportCrud: PagesImportExportCrud = {\n async importPages({ category: categorySlug, zipFileKey, zipFileUrl }) {\n await checkBasePermissions(context, PERMISSION_NAME, {\n rwd: \"w\"\n });\n\n // Bail out early if category not found\n const category = await context.pageBuilder.getCategory(categorySlug);\n if (!category) {\n throw new NotFoundError(`Category with slug \"${categorySlug}\" not found.`);\n }\n\n // Create a task for import page\n const task = await context.pageBuilder.pageImportExportTask.createTask({\n status: PageImportExportTaskStatus.PENDING,\n input: {\n category: categorySlug,\n zipFileKey,\n zipFileUrl\n }\n });\n /**\n * Import Pages\n * ImportPages\n * importPages\n */\n await invokeHandlerClient<CreateHandlerPayload>({\n context,\n name: IMPORT_PAGES_CREATE_HANDLER,\n payload: {\n category: categorySlug,\n zipFileKey,\n zipFileUrl,\n task,\n identity: context.security.getIdentity()\n },\n description: \"Import Pages - create\"\n });\n\n return {\n task\n };\n },\n\n async exportPages({ ids: initialPageIds, revisionType, where, sort, search }) {\n await checkBasePermissions(context, PERMISSION_NAME, {\n rwd: \"w\"\n });\n let pageIds: string[] = initialPageIds || [];\n // If no ids are provided then it means we want to export all pages\n if (!initialPageIds || (Array.isArray(initialPageIds) && initialPageIds.length === 0)) {\n pageIds = [];\n let pages = [];\n let meta: MetaResponse = {\n hasMoreItems: true,\n cursor: null,\n totalCount: 0\n };\n // Paginate pages\n while (meta.hasMoreItems) {\n [pages, meta] = await context.pageBuilder.listLatestPages({\n after: meta.cursor,\n where: where,\n sort: sort,\n search: search\n });\n // Save page ids\n pages.forEach(page => pageIds.push(page.id));\n }\n }\n\n if (pageIds.length === 0) {\n throw new WebinyError(\n \"Cannot export pages - no pages found for provided inputs.\",\n \"EMPTY_EXPORT_NO_PAGES_FOUND\"\n );\n }\n\n // Create the main task for page export.\n const task = await context.pageBuilder.pageImportExportTask.createTask({\n status: PageImportExportTaskStatus.PENDING\n });\n const exportPagesDataKey = `${EXPORT_PAGES_FOLDER_KEY}/${task.id}`;\n // For each page create a sub task and invoke the process handler.\n for (let i = 0; i < pageIds.length; i++) {\n const pageId = pageIds[i];\n // Create sub task.\n await context.pageBuilder.pageImportExportTask.createSubTask(\n task.id,\n zeroPad(i + 1, 5),\n {\n status: PageImportExportTaskStatus.PENDING,\n input: {\n pageId,\n exportPagesDataKey,\n revisionType\n }\n }\n );\n }\n // Update main task status.\n await context.pageBuilder.pageImportExportTask.updateTask(task.id, {\n status: PageImportExportTaskStatus.PROCESSING,\n stats: initialStats(pageIds.length),\n input: {\n exportPagesDataKey,\n revisionType\n }\n });\n\n /**\n * Export Pages\n * ExportPages\n * exportPages\n */\n // Invoke handler.\n await invokeHandlerClient<ExportPagesProcessHandlerPayload>({\n context,\n name: EXPORT_PAGES_PROCESS_HANDLER,\n payload: {\n taskId: task.id,\n subTaskIndex: 1,\n identity: context.security.getIdentity()\n },\n description: \"Export pages - process\"\n });\n\n return { task };\n }\n };\n // Modify context\n context.pageBuilder.pages = importExportCrud;\n});\n"],"mappings":";;;;;;;;;AAAA;;AACA;;AACA;;AACA;;AACA;;AAKA;;AAEA;;AAEA;;AAEA;;AAEA,MAAMA,eAAe,GAAG,SAAxB;AACA,MAAMC,4BAA4B,GAAGC,OAAO,CAACC,GAAR,CAAYF,4BAAjD;AACA,MAAMG,2BAA2B,GAAGF,OAAO,CAACC,GAAR,CAAYC,2BAAhD;;eAEe,IAAIC,kBAAJ,CAA6CC,OAAO,IAAI;EACnE,MAAMC,gBAAuC,GAAG;IAC5C,MAAMC,WAAN,CAAkB;MAAEC,QAAQ,EAAEC,YAAZ;MAA0BC,UAA1B;MAAsCC;IAAtC,CAAlB,EAAsE;MAClE,MAAM,IAAAC,6BAAA,EAAqBP,OAArB,EAA8BN,eAA9B,EAA+C;QACjDc,GAAG,EAAE;MAD4C,CAA/C,CAAN,CADkE,CAKlE;;MACA,MAAML,QAAQ,GAAG,MAAMH,OAAO,CAACS,WAAR,CAAoBC,WAApB,CAAgCN,YAAhC,CAAvB;;MACA,IAAI,CAACD,QAAL,EAAe;QACX,MAAM,IAAIQ,6BAAJ,CAAmB,uBAAsBP,YAAa,cAAtD,CAAN;MACH,CATiE,CAWlE;;;MACA,MAAMQ,IAAI,GAAG,MAAMZ,OAAO,CAACS,WAAR,CAAoBI,oBAApB,CAAyCC,UAAzC,CAAoD;QACnEC,MAAM,EAAEC,iCAAA,CAA2BC,OADgC;QAEnEC,KAAK,EAAE;UACHf,QAAQ,EAAEC,YADP;UAEHC,UAFG;UAGHC;QAHG;MAF4D,CAApD,CAAnB;MAQA;AACZ;AACA;AACA;AACA;;MACY,MAAM,IAAAa,2BAAA,EAA0C;QAC5CnB,OAD4C;QAE5CoB,IAAI,EAAEtB,2BAFsC;QAG5CuB,OAAO,EAAE;UACLlB,QAAQ,EAAEC,YADL;UAELC,UAFK;UAGLC,UAHK;UAILM,IAJK;UAKLU,QAAQ,EAAEtB,OAAO,CAACuB,QAAR,CAAiBC,WAAjB;QALL,CAHmC;QAU5CC,WAAW,EAAE;MAV+B,CAA1C,CAAN;MAaA,OAAO;QACHb;MADG,CAAP;IAGH,CA1C2C;;IA4C5C,MAAMc,WAAN,CAAkB;MAAEC,GAAG,EAAEC,cAAP;MAAuBC,YAAvB;MAAqCC,KAArC;MAA4CC,IAA5C;MAAkDC;IAAlD,CAAlB,EAA8E;MAC1E,MAAM,IAAAzB,6BAAA,EAAqBP,OAArB,EAA8BN,eAA9B,EAA+C;QACjDc,GAAG,EAAE;MAD4C,CAA/C,CAAN;MAGA,IAAIyB,OAAiB,GAAGL,cAAc,IAAI,EAA1C,CAJ0E,CAK1E;;MACA,IAAI,CAACA,cAAD,IAAoBM,KAAK,CAACC,OAAN,CAAcP,cAAd,KAAiCA,cAAc,CAACQ,MAAf,KAA0B,CAAnF,EAAuF;QACnFH,OAAO,GAAG,EAAV;QACA,IAAII,KAAK,GAAG,EAAZ;QACA,IAAIC,IAAkB,GAAG;UACrBC,YAAY,EAAE,IADO;UAErBC,MAAM,EAAE,IAFa;UAGrBC,UAAU,EAAE;QAHS,CAAzB,CAHmF,CAQnF;;QACA,OAAOH,IAAI,CAACC,YAAZ,EAA0B;UACtB,CAACF,KAAD,EAAQC,IAAR,IAAgB,MAAMtC,OAAO,CAACS,WAAR,CAAoBiC,eAApB,CAAoC;YACtDC,KAAK,EAAEL,IAAI,CAACE,MAD0C;YAEtDV,KAAK,EAAEA,KAF+C;YAGtDC,IAAI,EAAEA,IAHgD;YAItDC,MAAM,EAAEA;UAJ8C,CAApC,CAAtB,CADsB,CAOtB;;UACAK,KAAK,CAACO,OAAN,CAAcC,IAAI,IAAIZ,OAAO,CAACa,IAAR,CAAaD,IAAI,CAACE,EAAlB,CAAtB;QACH;MACJ;;MAED,IAAId,OAAO,CAACG,MAAR,KAAmB,CAAvB,EAA0B;QACtB,MAAM,IAAIY,cAAJ,CACF,2DADE,EAEF,6BAFE,CAAN;MAIH,CAhCyE,CAkC1E;;;MACA,MAAMpC,IAAI,GAAG,MAAMZ,OAAO,CAACS,WAAR,CAAoBI,oBAApB,CAAyCC,UAAzC,CAAoD;QACnEC,MAAM,EAAEC,iCAAA,CAA2BC;MADgC,CAApD,CAAnB;MAGA,MAAMgC,kBAAkB,GAAI,GAAEC,+BAAwB,IAAGtC,IAAI,CAACmC,EAAG,EAAjE,CAtC0E,CAuC1E;;MACA,KAAK,IAAII,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGlB,OAAO,CAACG,MAA5B,EAAoCe,CAAC,EAArC,EAAyC;QACrC,MAAMC,MAAM,GAAGnB,OAAO,CAACkB,CAAD,CAAtB,CADqC,CAErC;;QACA,MAAMnD,OAAO,CAACS,WAAR,CAAoBI,oBAApB,CAAyCwC,aAAzC,CACFzC,IAAI,CAACmC,EADH,EAEF,IAAAO,eAAA,EAAQH,CAAC,GAAG,CAAZ,EAAe,CAAf,CAFE,EAGF;UACIpC,MAAM,EAAEC,iCAAA,CAA2BC,OADvC;UAEIC,KAAK,EAAE;YACHkC,MADG;YAEHH,kBAFG;YAGHpB;UAHG;QAFX,CAHE,CAAN;MAYH,CAvDyE,CAwD1E;;;MACA,MAAM7B,OAAO,CAACS,WAAR,CAAoBI,oBAApB,CAAyC0C,UAAzC,CAAoD3C,IAAI,CAACmC,EAAzD,EAA6D;QAC/DhC,MAAM,EAAEC,iCAAA,CAA2BwC,UAD4B;QAE/DC,KAAK,EAAE,IAAAC,mBAAA,EAAazB,OAAO,CAACG,MAArB,CAFwD;QAG/DlB,KAAK,EAAE;UACH+B,kBADG;UAEHpB;QAFG;MAHwD,CAA7D,CAAN;MASA;AACZ;AACA;AACA;AACA;MACY;;MACA,MAAM,IAAAV,2BAAA,EAAsD;QACxDnB,OADwD;QAExDoB,IAAI,EAAEzB,4BAFkD;QAGxD0B,OAAO,EAAE;UACLsC,MAAM,EAAE/C,IAAI,CAACmC,EADR;UAELa,YAAY,EAAE,CAFT;UAGLtC,QAAQ,EAAEtB,OAAO,CAACuB,QAAR,CAAiBC,WAAjB;QAHL,CAH+C;QAQxDC,WAAW,EAAE;MAR2C,CAAtD,CAAN;MAWA,OAAO;QAAEb;MAAF,CAAP;IACH;;EAhI2C,CAAhD,CADmE,CAmInE;;EACAZ,OAAO,CAACS,WAAR,CAAoB4B,KAApB,GAA4BpC,gBAA5B;AACH,CArIc,C"}
|
1
|
+
{"version":3,"names":["PERMISSION_NAME","EXPORT_PAGES_PROCESS_HANDLER","process","env","IMPORT_PAGES_CREATE_HANDLER","ContextPlugin","context","importExportCrud","importPages","category","categorySlug","zipFileUrl","checkBasePermissions","rwd","pageBuilder","getCategory","NotFoundError","task","pageImportExportTask","createTask","status","PageImportExportTaskStatus","PENDING","input","invokeHandlerClient","name","payload","identity","security","getIdentity","description","exportPages","ids","initialPageIds","revisionType","where","sort","search","pageIds","Array","isArray","length","pages","meta","hasMoreItems","cursor","totalCount","listLatestPages","after","forEach","page","push","id","WebinyError","exportPagesDataKey","EXPORT_PAGES_FOLDER_KEY","i","pageId","createSubTask","zeroPad","updateTask","PROCESSING","stats","initialStats","taskId","subTaskIndex"],"sources":["pages.crud.ts"],"sourcesContent":["import WebinyError from \"@webiny/error\";\nimport { NotFoundError } from \"@webiny/handler-graphql\";\nimport { ContextPlugin } from \"@webiny/api\";\nimport checkBasePermissions from \"@webiny/api-page-builder/graphql/crud/utils/checkBasePermissions\";\nimport {\n PageImportExportTaskStatus,\n PagesImportExportCrud,\n PbPageImportExportContext\n} from \"~/types\";\nimport { invokeHandlerClient } from \"~/client\";\nimport { Payload as CreateHandlerPayload } from \"~/importPages/create\";\nimport { initialStats } from \"~/importPages/utils\";\nimport { Payload as ExportPagesProcessHandlerPayload } from \"~/exportPages/process\";\nimport { EXPORT_PAGES_FOLDER_KEY } from \"~/exportPages/utils\";\nimport { MetaResponse } from \"@webiny/api-page-builder/types\";\nimport { zeroPad } from \"@webiny/utils\";\n\nconst PERMISSION_NAME = \"pb.page\";\nconst EXPORT_PAGES_PROCESS_HANDLER = process.env.EXPORT_PAGES_PROCESS_HANDLER as string;\nconst IMPORT_PAGES_CREATE_HANDLER = process.env.IMPORT_PAGES_CREATE_HANDLER as string;\n\nexport default new ContextPlugin<PbPageImportExportContext>(context => {\n const importExportCrud: PagesImportExportCrud = {\n async importPages({ category: categorySlug, zipFileUrl }) {\n await checkBasePermissions(context, PERMISSION_NAME, {\n rwd: \"w\"\n });\n\n // Bail out early if category not found\n const category = await context.pageBuilder.getCategory(categorySlug);\n if (!category) {\n throw new NotFoundError(`Category with slug \"${categorySlug}\" not found.`);\n }\n\n // Create a task for import page\n const task = await context.pageBuilder.pageImportExportTask.createTask({\n status: PageImportExportTaskStatus.PENDING,\n input: {\n category: categorySlug,\n zipFileUrl\n }\n });\n /**\n * Import Pages\n * ImportPages\n * importPages\n */\n await invokeHandlerClient<CreateHandlerPayload>({\n context,\n name: IMPORT_PAGES_CREATE_HANDLER,\n payload: {\n category: categorySlug,\n zipFileUrl,\n task,\n identity: context.security.getIdentity()\n },\n description: \"Import Pages - create\"\n });\n\n return {\n task\n };\n },\n\n async exportPages({ ids: initialPageIds, revisionType, where, sort, search }) {\n await checkBasePermissions(context, PERMISSION_NAME, {\n rwd: \"w\"\n });\n let pageIds: string[] = initialPageIds || [];\n // If no ids are provided then it means we want to export all pages\n if (!initialPageIds || (Array.isArray(initialPageIds) && initialPageIds.length === 0)) {\n pageIds = [];\n let pages = [];\n let meta: MetaResponse = {\n hasMoreItems: true,\n cursor: null,\n totalCount: 0\n };\n // Paginate pages\n while (meta.hasMoreItems) {\n [pages, meta] = await context.pageBuilder.listLatestPages({\n after: meta.cursor,\n where: where,\n sort: sort,\n search: search\n });\n // Save page ids\n pages.forEach(page => pageIds.push(page.id));\n }\n }\n\n if (pageIds.length === 0) {\n throw new WebinyError(\n \"Cannot export pages - no pages found for provided inputs.\",\n \"EMPTY_EXPORT_NO_PAGES_FOUND\"\n );\n }\n\n // Create the main task for page export.\n const task = await context.pageBuilder.pageImportExportTask.createTask({\n status: PageImportExportTaskStatus.PENDING\n });\n const exportPagesDataKey = `${EXPORT_PAGES_FOLDER_KEY}/${task.id}`;\n // For each page create a sub task and invoke the process handler.\n for (let i = 0; i < pageIds.length; i++) {\n const pageId = pageIds[i];\n // Create sub task.\n await context.pageBuilder.pageImportExportTask.createSubTask(\n task.id,\n zeroPad(i + 1, 5),\n {\n status: PageImportExportTaskStatus.PENDING,\n input: {\n pageId,\n exportPagesDataKey,\n revisionType\n }\n }\n );\n }\n // Update main task status.\n await context.pageBuilder.pageImportExportTask.updateTask(task.id, {\n status: PageImportExportTaskStatus.PROCESSING,\n stats: initialStats(pageIds.length),\n input: {\n exportPagesDataKey,\n revisionType\n }\n });\n\n /**\n * Export Pages\n * ExportPages\n * exportPages\n */\n // Invoke handler.\n await invokeHandlerClient<ExportPagesProcessHandlerPayload>({\n context,\n name: EXPORT_PAGES_PROCESS_HANDLER,\n payload: {\n taskId: task.id,\n subTaskIndex: 1,\n identity: context.security.getIdentity()\n },\n description: \"Export pages - process\"\n });\n\n return { task };\n }\n };\n // Modify context\n context.pageBuilder.pages = importExportCrud;\n});\n"],"mappings":";;;;;;;;;AAAA;;AACA;;AACA;;AACA;;AACA;;AAKA;;AAEA;;AAEA;;AAEA;;AAEA,MAAMA,eAAe,GAAG,SAAxB;AACA,MAAMC,4BAA4B,GAAGC,OAAO,CAACC,GAAR,CAAYF,4BAAjD;AACA,MAAMG,2BAA2B,GAAGF,OAAO,CAACC,GAAR,CAAYC,2BAAhD;;eAEe,IAAIC,kBAAJ,CAA6CC,OAAO,IAAI;EACnE,MAAMC,gBAAuC,GAAG;IAC5C,MAAMC,WAAN,CAAkB;MAAEC,QAAQ,EAAEC,YAAZ;MAA0BC;IAA1B,CAAlB,EAA0D;MACtD,MAAM,IAAAC,6BAAA,EAAqBN,OAArB,EAA8BN,eAA9B,EAA+C;QACjDa,GAAG,EAAE;MAD4C,CAA/C,CAAN,CADsD,CAKtD;;MACA,MAAMJ,QAAQ,GAAG,MAAMH,OAAO,CAACQ,WAAR,CAAoBC,WAApB,CAAgCL,YAAhC,CAAvB;;MACA,IAAI,CAACD,QAAL,EAAe;QACX,MAAM,IAAIO,6BAAJ,CAAmB,uBAAsBN,YAAa,cAAtD,CAAN;MACH,CATqD,CAWtD;;;MACA,MAAMO,IAAI,GAAG,MAAMX,OAAO,CAACQ,WAAR,CAAoBI,oBAApB,CAAyCC,UAAzC,CAAoD;QACnEC,MAAM,EAAEC,iCAAA,CAA2BC,OADgC;QAEnEC,KAAK,EAAE;UACHd,QAAQ,EAAEC,YADP;UAEHC;QAFG;MAF4D,CAApD,CAAnB;MAOA;AACZ;AACA;AACA;AACA;;MACY,MAAM,IAAAa,2BAAA,EAA0C;QAC5ClB,OAD4C;QAE5CmB,IAAI,EAAErB,2BAFsC;QAG5CsB,OAAO,EAAE;UACLjB,QAAQ,EAAEC,YADL;UAELC,UAFK;UAGLM,IAHK;UAILU,QAAQ,EAAErB,OAAO,CAACsB,QAAR,CAAiBC,WAAjB;QAJL,CAHmC;QAS5CC,WAAW,EAAE;MAT+B,CAA1C,CAAN;MAYA,OAAO;QACHb;MADG,CAAP;IAGH,CAxC2C;;IA0C5C,MAAMc,WAAN,CAAkB;MAAEC,GAAG,EAAEC,cAAP;MAAuBC,YAAvB;MAAqCC,KAArC;MAA4CC,IAA5C;MAAkDC;IAAlD,CAAlB,EAA8E;MAC1E,MAAM,IAAAzB,6BAAA,EAAqBN,OAArB,EAA8BN,eAA9B,EAA+C;QACjDa,GAAG,EAAE;MAD4C,CAA/C,CAAN;MAGA,IAAIyB,OAAiB,GAAGL,cAAc,IAAI,EAA1C,CAJ0E,CAK1E;;MACA,IAAI,CAACA,cAAD,IAAoBM,KAAK,CAACC,OAAN,CAAcP,cAAd,KAAiCA,cAAc,CAACQ,MAAf,KAA0B,CAAnF,EAAuF;QACnFH,OAAO,GAAG,EAAV;QACA,IAAII,KAAK,GAAG,EAAZ;QACA,IAAIC,IAAkB,GAAG;UACrBC,YAAY,EAAE,IADO;UAErBC,MAAM,EAAE,IAFa;UAGrBC,UAAU,EAAE;QAHS,CAAzB,CAHmF,CAQnF;;QACA,OAAOH,IAAI,CAACC,YAAZ,EAA0B;UACtB,CAACF,KAAD,EAAQC,IAAR,IAAgB,MAAMrC,OAAO,CAACQ,WAAR,CAAoBiC,eAApB,CAAoC;YACtDC,KAAK,EAAEL,IAAI,CAACE,MAD0C;YAEtDV,KAAK,EAAEA,KAF+C;YAGtDC,IAAI,EAAEA,IAHgD;YAItDC,MAAM,EAAEA;UAJ8C,CAApC,CAAtB,CADsB,CAOtB;;UACAK,KAAK,CAACO,OAAN,CAAcC,IAAI,IAAIZ,OAAO,CAACa,IAAR,CAAaD,IAAI,CAACE,EAAlB,CAAtB;QACH;MACJ;;MAED,IAAId,OAAO,CAACG,MAAR,KAAmB,CAAvB,EAA0B;QACtB,MAAM,IAAIY,cAAJ,CACF,2DADE,EAEF,6BAFE,CAAN;MAIH,CAhCyE,CAkC1E;;;MACA,MAAMpC,IAAI,GAAG,MAAMX,OAAO,CAACQ,WAAR,CAAoBI,oBAApB,CAAyCC,UAAzC,CAAoD;QACnEC,MAAM,EAAEC,iCAAA,CAA2BC;MADgC,CAApD,CAAnB;MAGA,MAAMgC,kBAAkB,GAAI,GAAEC,+BAAwB,IAAGtC,IAAI,CAACmC,EAAG,EAAjE,CAtC0E,CAuC1E;;MACA,KAAK,IAAII,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGlB,OAAO,CAACG,MAA5B,EAAoCe,CAAC,EAArC,EAAyC;QACrC,MAAMC,MAAM,GAAGnB,OAAO,CAACkB,CAAD,CAAtB,CADqC,CAErC;;QACA,MAAMlD,OAAO,CAACQ,WAAR,CAAoBI,oBAApB,CAAyCwC,aAAzC,CACFzC,IAAI,CAACmC,EADH,EAEF,IAAAO,eAAA,EAAQH,CAAC,GAAG,CAAZ,EAAe,CAAf,CAFE,EAGF;UACIpC,MAAM,EAAEC,iCAAA,CAA2BC,OADvC;UAEIC,KAAK,EAAE;YACHkC,MADG;YAEHH,kBAFG;YAGHpB;UAHG;QAFX,CAHE,CAAN;MAYH,CAvDyE,CAwD1E;;;MACA,MAAM5B,OAAO,CAACQ,WAAR,CAAoBI,oBAApB,CAAyC0C,UAAzC,CAAoD3C,IAAI,CAACmC,EAAzD,EAA6D;QAC/DhC,MAAM,EAAEC,iCAAA,CAA2BwC,UAD4B;QAE/DC,KAAK,EAAE,IAAAC,mBAAA,EAAazB,OAAO,CAACG,MAArB,CAFwD;QAG/DlB,KAAK,EAAE;UACH+B,kBADG;UAEHpB;QAFG;MAHwD,CAA7D,CAAN;MASA;AACZ;AACA;AACA;AACA;MACY;;MACA,MAAM,IAAAV,2BAAA,EAAsD;QACxDlB,OADwD;QAExDmB,IAAI,EAAExB,4BAFkD;QAGxDyB,OAAO,EAAE;UACLsC,MAAM,EAAE/C,IAAI,CAACmC,EADR;UAELa,YAAY,EAAE,CAFT;UAGLtC,QAAQ,EAAErB,OAAO,CAACsB,QAAR,CAAiBC,WAAjB;QAHL,CAH+C;QAQxDC,WAAW,EAAE;MAR2C,CAAtD,CAAN;MAWA,OAAO;QAAEb;MAAF,CAAP;IACH;;EA9H2C,CAAhD,CADmE,CAiInE;;EACAX,OAAO,CAACQ,WAAR,CAAoB4B,KAApB,GAA4BnC,gBAA5B;AACH,CAnIc,C"}
|
@@ -49,11 +49,7 @@ const plugin = {
|
|
49
49
|
): PbExportPageResponse
|
50
50
|
|
51
51
|
# Import pages
|
52
|
-
importPages(
|
53
|
-
category: String!
|
54
|
-
zipFileKey: String
|
55
|
-
zipFileUrl: String
|
56
|
-
): PbImportPageResponse
|
52
|
+
importPages(category: String!, zipFileUrl: String): PbImportPageResponse
|
57
53
|
}
|
58
54
|
`,
|
59
55
|
resolvers: {
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["plugin","type","schema","typeDefs","resolvers","PbMutation","exportPages","_","args","context","resolve","pageBuilder","pages","importPages"],"sources":["pages.gql.ts"],"sourcesContent":["import { GraphQLSchemaPlugin } from \"@webiny/handler-graphql/types\";\nimport { ExportPagesParams, ImportPagesParams } from \"~/types\";\nimport { PbPageImportExportContext } from \"../types\";\nimport resolve from \"./utils/resolve\";\n\nconst plugin: GraphQLSchemaPlugin<PbPageImportExportContext> = {\n type: \"graphql-schema\",\n schema: {\n typeDefs: /* GraphQL */ `\n type PbExportPageData {\n task: PbPageImportExportTask\n }\n\n type PbExportPageResponse {\n data: PbExportPageData\n error: PbError\n }\n\n type PbImportPageData {\n task: PbPageImportExportTask\n }\n\n type PbImportPageResponse {\n data: PbImportPageData\n error: PbError\n }\n\n enum PbExportPageRevisionType {\n published\n latest\n }\n\n extend type PbMutation {\n # Export pages\n exportPages(\n ids: [ID!]\n revisionType: PbExportPageRevisionType!\n where: PbListPagesWhereInput\n sort: [PbListPagesSort!]\n search: PbListPagesSearchInput\n ): PbExportPageResponse\n\n # Import pages\n importPages(
|
1
|
+
{"version":3,"names":["plugin","type","schema","typeDefs","resolvers","PbMutation","exportPages","_","args","context","resolve","pageBuilder","pages","importPages"],"sources":["pages.gql.ts"],"sourcesContent":["import { GraphQLSchemaPlugin } from \"@webiny/handler-graphql/types\";\nimport { ExportPagesParams, ImportPagesParams } from \"~/types\";\nimport { PbPageImportExportContext } from \"../types\";\nimport resolve from \"./utils/resolve\";\n\nconst plugin: GraphQLSchemaPlugin<PbPageImportExportContext> = {\n type: \"graphql-schema\",\n schema: {\n typeDefs: /* GraphQL */ `\n type PbExportPageData {\n task: PbPageImportExportTask\n }\n\n type PbExportPageResponse {\n data: PbExportPageData\n error: PbError\n }\n\n type PbImportPageData {\n task: PbPageImportExportTask\n }\n\n type PbImportPageResponse {\n data: PbImportPageData\n error: PbError\n }\n\n enum PbExportPageRevisionType {\n published\n latest\n }\n\n extend type PbMutation {\n # Export pages\n exportPages(\n ids: [ID!]\n revisionType: PbExportPageRevisionType!\n where: PbListPagesWhereInput\n sort: [PbListPagesSort!]\n search: PbListPagesSearchInput\n ): PbExportPageResponse\n\n # Import pages\n importPages(category: String!, zipFileUrl: String): PbImportPageResponse\n }\n `,\n resolvers: {\n PbMutation: {\n exportPages: async (_, args: any, context) => {\n /**\n * We know that args is ExportPagesParams.\n */\n return resolve(() =>\n context.pageBuilder.pages.exportPages(args as unknown as ExportPagesParams)\n );\n },\n\n importPages: async (_, args: any, context) => {\n /**\n * We know that args is ExportPagesParams.\n */\n return resolve(() =>\n context.pageBuilder.pages.importPages(args as unknown as ImportPagesParams)\n );\n }\n }\n }\n }\n};\n\nexport default plugin;\n"],"mappings":";;;;;;;;;AAGA;;AAEA,MAAMA,MAAsD,GAAG;EAC3DC,IAAI,EAAE,gBADqD;EAE3DC,MAAM,EAAE;IACJC,QAAQ;IAAE;IAAe;AACjC;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA,SAtCY;IAuCJC,SAAS,EAAE;MACPC,UAAU,EAAE;QACRC,WAAW,EAAE,OAAOC,CAAP,EAAUC,IAAV,EAAqBC,OAArB,KAAiC;UAC1C;AACpB;AACA;UACoB,OAAO,IAAAC,gBAAA,EAAQ,MACXD,OAAO,CAACE,WAAR,CAAoBC,KAApB,CAA0BN,WAA1B,CAAsCE,IAAtC,CADG,CAAP;QAGH,CARO;QAURK,WAAW,EAAE,OAAON,CAAP,EAAUC,IAAV,EAAqBC,OAArB,KAAiC;UAC1C;AACpB;AACA;UACoB,OAAO,IAAAC,gBAAA,EAAQ,MACXD,OAAO,CAACE,WAAR,CAAoBC,KAApB,CAA0BC,WAA1B,CAAsCL,IAAtC,CADG,CAAP;QAGH;MAjBO;IADL;EAvCP;AAFmD,CAA/D;eAiEeR,M"}
|
package/graphql/types.d.ts
CHANGED
@@ -19,8 +19,7 @@ export interface ExportPagesParams {
|
|
19
19
|
}
|
20
20
|
export interface ImportPagesParams {
|
21
21
|
category: string;
|
22
|
-
|
23
|
-
zipFileUrl?: string;
|
22
|
+
zipFileUrl: string;
|
24
23
|
}
|
25
24
|
export declare type PagesImportExportCrud = {
|
26
25
|
exportPages(params: ExportPagesParams): Promise<{
|
package/graphql/types.js.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":[],"sources":["types.ts"],"sourcesContent":["import { PbContext } from \"@webiny/api-page-builder/types\";\nimport {\n PageExportRevisionType,\n PageImportExportTask,\n PageImportExportTaskStatus,\n PageImportExportTaskStorageOperations,\n PageImportExportTaskStorageOperationsListParams\n} from \"~/types\";\n\nexport interface ExportPagesParams {\n ids?: string[];\n revisionType: PageExportRevisionType;\n where?: {\n category?: string;\n status?: string;\n tags?: { query: string[]; rule?: \"any\" | \"all\" };\n [key: string]: any;\n };\n search?: { query?: string };\n sort?: string[];\n}\n\nexport interface ImportPagesParams {\n category: string;\n
|
1
|
+
{"version":3,"names":[],"sources":["types.ts"],"sourcesContent":["import { PbContext } from \"@webiny/api-page-builder/types\";\nimport {\n PageExportRevisionType,\n PageImportExportTask,\n PageImportExportTaskStatus,\n PageImportExportTaskStorageOperations,\n PageImportExportTaskStorageOperationsListParams\n} from \"~/types\";\n\nexport interface ExportPagesParams {\n ids?: string[];\n revisionType: PageExportRevisionType;\n where?: {\n category?: string;\n status?: string;\n tags?: { query: string[]; rule?: \"any\" | \"all\" };\n [key: string]: any;\n };\n search?: { query?: string };\n sort?: string[];\n}\n\nexport interface ImportPagesParams {\n category: string;\n zipFileUrl: string;\n}\n\nexport type PagesImportExportCrud = {\n exportPages(params: ExportPagesParams): Promise<{ task: PageImportExportTask }>;\n importPages(params: ImportPagesParams): Promise<{ task: PageImportExportTask }>;\n};\n\ntype PageImportExportTaskCreateData = Omit<PageImportExportTask, \"id\" | \"createdOn\" | \"createdBy\">;\n\nexport type PageImportExportTaskCrud = {\n /**\n * To be used internally in our code.\n * @internal\n */\n storageOperations: PageImportExportTaskStorageOperations;\n\n getTask(id: string): Promise<PageImportExportTask | null>;\n listTasks(\n params?: PageImportExportTaskStorageOperationsListParams\n ): Promise<PageImportExportTask[]>;\n createTask(data: Partial<PageImportExportTaskCreateData>): Promise<PageImportExportTask>;\n updateTask(\n id: string,\n data: Partial<PageImportExportTaskCreateData>\n ): Promise<PageImportExportTask>;\n updateStats(\n id: string,\n data: {\n prevStatus: PageImportExportTaskStatus;\n nextStatus: PageImportExportTaskStatus;\n }\n ): Promise<PageImportExportTask>;\n deleteTask(id: string): Promise<PageImportExportTask>;\n getSubTask(id: string, subtaskId: string): Promise<PageImportExportTask | null>;\n listSubTasks(\n id: string,\n status: PageImportExportTaskStatus,\n limit: number\n ): Promise<PageImportExportTask[]>;\n createSubTask(\n id: string,\n subTaskId: string,\n data: Partial<PageImportExportTaskCreateData>\n ): Promise<PageImportExportTask>;\n updateSubTask(\n id: string,\n subTaskId: string,\n data: Partial<PageImportExportTaskCreateData>\n ): Promise<PageImportExportTask>;\n};\n\nexport interface PbPageImportExportContext extends PbContext {\n pageBuilder: PbContext[\"pageBuilder\"] & {\n pages: PagesImportExportCrud;\n pageImportExportTask: PageImportExportTaskCrud;\n };\n}\n\nexport interface PageImportExportPluginsParams {\n storageOperations: PageImportExportTaskStorageOperations;\n}\n"],"mappings":""}
|
@@ -32,7 +32,6 @@ var _default = configuration => {
|
|
32
32
|
const {
|
33
33
|
task,
|
34
34
|
category,
|
35
|
-
zipFileKey,
|
36
35
|
zipFileUrl,
|
37
36
|
identity
|
38
37
|
} = payload;
|
@@ -40,19 +39,18 @@ var _default = configuration => {
|
|
40
39
|
try {
|
41
40
|
log("RUNNING Import Pages Create");
|
42
41
|
|
43
|
-
if (!
|
42
|
+
if (!zipFileUrl) {
|
44
43
|
return {
|
45
44
|
data: null,
|
46
45
|
error: {
|
47
|
-
message:
|
46
|
+
message: `Missing "zipFileUrl"!`
|
48
47
|
}
|
49
48
|
};
|
50
49
|
}
|
51
50
|
|
52
51
|
(0, _mockSecurity.mockSecurity)(identity, context); // Step 1: Read the zip file
|
53
52
|
|
54
|
-
const pageImportDataList = await (0, _utils.readExtractAndUploadZipFileContents)(
|
55
|
-
// For each page create a sub task and invoke the process handler
|
53
|
+
const pageImportDataList = await (0, _utils.readExtractAndUploadZipFileContents)(zipFileUrl); // For each page create a subtask and invoke the process handler
|
56
54
|
|
57
55
|
for (let i = 0; i < pageImportDataList.length; i++) {
|
58
56
|
const pagesDirMap = pageImportDataList[i]; // Create sub task
|
@@ -62,7 +60,6 @@ var _default = configuration => {
|
|
62
60
|
data: {
|
63
61
|
pageKey: pagesDirMap.key,
|
64
62
|
category,
|
65
|
-
zipFileKey,
|
66
63
|
zipFileUrl,
|
67
64
|
input: {
|
68
65
|
fileUploadsData: pagesDirMap
|
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["configuration","createRawEventHandler","payload","context","log","console","pageBuilder","task","category","
|
1
|
+
{"version":3,"names":["configuration","createRawEventHandler","payload","context","log","console","pageBuilder","task","category","zipFileUrl","identity","data","error","message","mockSecurity","pageImportDataList","readExtractAndUploadZipFileContents","i","length","pagesDirMap","subtask","pageImportExportTask","createSubTask","id","zeroPad","status","PageImportExportTaskStatus","PENDING","pageKey","key","input","fileUploadsData","updateTask","PROCESSING","stats","initialStats","invokeHandlerClient","name","handlers","process","taskId","subTaskIndex","security","getIdentity","description","e","FAILED","code"],"sources":["index.ts"],"sourcesContent":["import {\n PageImportExportTask,\n PageImportExportTaskStatus,\n PbPageImportExportContext\n} from \"~/types\";\nimport { initialStats, readExtractAndUploadZipFileContents } from \"~/importPages/utils\";\nimport { invokeHandlerClient } from \"~/client\";\nimport { Payload as ProcessPayload } from \"../process\";\nimport { SecurityIdentity } from \"@webiny/api-security/types\";\nimport { mockSecurity } from \"~/mockSecurity\";\nimport { zeroPad } from \"@webiny/utils\";\nimport { createRawEventHandler } from \"@webiny/handler-aws\";\n\ninterface Configuration {\n handlers: {\n process: string;\n };\n}\n\nexport interface Payload {\n category: string;\n zipFileUrl: string;\n task: PageImportExportTask;\n identity: SecurityIdentity;\n}\nexport interface Response {\n data: string | null;\n error: Partial<Error> | null;\n}\n\n/**\n * Handles the import page workflow.\n */\nexport default (configuration: Configuration) => {\n return createRawEventHandler<Payload, PbPageImportExportContext, Response>(\n async ({ payload, context }) => {\n const log = console.log;\n\n const { pageBuilder } = context;\n const { task, category, zipFileUrl, identity } = payload;\n try {\n log(\"RUNNING Import Pages Create\");\n if (!zipFileUrl) {\n return {\n data: null,\n error: {\n message: `Missing \"zipFileUrl\"!`\n }\n };\n }\n mockSecurity(identity, context);\n // Step 1: Read the zip file\n const pageImportDataList = await readExtractAndUploadZipFileContents(zipFileUrl);\n\n // For each page create a subtask and invoke the process handler\n for (let i = 0; i < pageImportDataList.length; i++) {\n const pagesDirMap = pageImportDataList[i];\n // Create sub task\n const subtask = await pageBuilder.pageImportExportTask.createSubTask(\n task.id,\n zeroPad(i + 1, 5),\n {\n status: PageImportExportTaskStatus.PENDING,\n data: {\n pageKey: pagesDirMap.key,\n category,\n zipFileUrl,\n input: {\n fileUploadsData: pagesDirMap\n }\n }\n }\n );\n log(`Added SUB_TASK \"${subtask.id}\" to queue.`);\n }\n // Update main task status\n await pageBuilder.pageImportExportTask.updateTask(task.id, {\n status: PageImportExportTaskStatus.PROCESSING,\n stats: initialStats(pageImportDataList.length)\n });\n\n await invokeHandlerClient<ProcessPayload>({\n context,\n name: configuration.handlers.process,\n payload: {\n taskId: task.id,\n // Execute \"Process\" for the first sub task.\n subTaskIndex: 1,\n identity: context.security.getIdentity()\n },\n description: \"Import pages - process - first\"\n });\n } catch (e) {\n log(\"[IMPORT_PAGES_CREATE] Error => \", e);\n\n /**\n * In case of error, we'll update the task status to \"failed\",\n * so that, client can show notify the user appropriately.\n */\n\n await pageBuilder.pageImportExportTask.updateTask(task.id, {\n status: PageImportExportTaskStatus.FAILED,\n error: {\n name: e.name,\n message: e.message,\n code: e.code || \"EXPORT_FAILED\"\n }\n });\n\n return {\n data: null,\n error: {\n message: e.message\n }\n };\n }\n\n return {\n data: \"\",\n error: null\n };\n }\n );\n};\n"],"mappings":";;;;;;;AAAA;;AAKA;;AACA;;AAGA;;AACA;;AACA;;AAmBA;AACA;AACA;eACgBA,aAAD,IAAkC;EAC7C,OAAO,IAAAC,iCAAA,EACH,OAAO;IAAEC,OAAF;IAAWC;EAAX,CAAP,KAAgC;IAC5B,MAAMC,GAAG,GAAGC,OAAO,CAACD,GAApB;IAEA,MAAM;MAAEE;IAAF,IAAkBH,OAAxB;IACA,MAAM;MAAEI,IAAF;MAAQC,QAAR;MAAkBC,UAAlB;MAA8BC;IAA9B,IAA2CR,OAAjD;;IACA,IAAI;MACAE,GAAG,CAAC,6BAAD,CAAH;;MACA,IAAI,CAACK,UAAL,EAAiB;QACb,OAAO;UACHE,IAAI,EAAE,IADH;UAEHC,KAAK,EAAE;YACHC,OAAO,EAAG;UADP;QAFJ,CAAP;MAMH;;MACD,IAAAC,0BAAA,EAAaJ,QAAb,EAAuBP,OAAvB,EAVA,CAWA;;MACA,MAAMY,kBAAkB,GAAG,MAAM,IAAAC,0CAAA,EAAoCP,UAApC,CAAjC,CAZA,CAcA;;MACA,KAAK,IAAIQ,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGF,kBAAkB,CAACG,MAAvC,EAA+CD,CAAC,EAAhD,EAAoD;QAChD,MAAME,WAAW,GAAGJ,kBAAkB,CAACE,CAAD,CAAtC,CADgD,CAEhD;;QACA,MAAMG,OAAO,GAAG,MAAMd,WAAW,CAACe,oBAAZ,CAAiCC,aAAjC,CAClBf,IAAI,CAACgB,EADa,EAElB,IAAAC,eAAA,EAAQP,CAAC,GAAG,CAAZ,EAAe,CAAf,CAFkB,EAGlB;UACIQ,MAAM,EAAEC,iCAAA,CAA2BC,OADvC;UAEIhB,IAAI,EAAE;YACFiB,OAAO,EAAET,WAAW,CAACU,GADnB;YAEFrB,QAFE;YAGFC,UAHE;YAIFqB,KAAK,EAAE;cACHC,eAAe,EAAEZ;YADd;UAJL;QAFV,CAHkB,CAAtB;QAeAf,GAAG,CAAE,mBAAkBgB,OAAO,CAACG,EAAG,aAA/B,CAAH;MACH,CAlCD,CAmCA;;;MACA,MAAMjB,WAAW,CAACe,oBAAZ,CAAiCW,UAAjC,CAA4CzB,IAAI,CAACgB,EAAjD,EAAqD;QACvDE,MAAM,EAAEC,iCAAA,CAA2BO,UADoB;QAEvDC,KAAK,EAAE,IAAAC,mBAAA,EAAapB,kBAAkB,CAACG,MAAhC;MAFgD,CAArD,CAAN;MAKA,MAAM,IAAAkB,2BAAA,EAAoC;QACtCjC,OADsC;QAEtCkC,IAAI,EAAErC,aAAa,CAACsC,QAAd,CAAuBC,OAFS;QAGtCrC,OAAO,EAAE;UACLsC,MAAM,EAAEjC,IAAI,CAACgB,EADR;UAEL;UACAkB,YAAY,EAAE,CAHT;UAIL/B,QAAQ,EAAEP,OAAO,CAACuC,QAAR,CAAiBC,WAAjB;QAJL,CAH6B;QAStCC,WAAW,EAAE;MATyB,CAApC,CAAN;IAWH,CApDD,CAoDE,OAAOC,CAAP,EAAU;MACRzC,GAAG,CAAC,iCAAD,EAAoCyC,CAApC,CAAH;MAEA;AAChB;AACA;AACA;;MAEgB,MAAMvC,WAAW,CAACe,oBAAZ,CAAiCW,UAAjC,CAA4CzB,IAAI,CAACgB,EAAjD,EAAqD;QACvDE,MAAM,EAAEC,iCAAA,CAA2BoB,MADoB;QAEvDlC,KAAK,EAAE;UACHyB,IAAI,EAAEQ,CAAC,CAACR,IADL;UAEHxB,OAAO,EAAEgC,CAAC,CAAChC,OAFR;UAGHkC,IAAI,EAAEF,CAAC,CAACE,IAAF,IAAU;QAHb;MAFgD,CAArD,CAAN;MASA,OAAO;QACHpC,IAAI,EAAE,IADH;QAEHC,KAAK,EAAE;UACHC,OAAO,EAAEgC,CAAC,CAAChC;QADR;MAFJ,CAAP;IAMH;;IAED,OAAO;MACHF,IAAI,EAAE,EADH;MAEHC,KAAK,EAAE;IAFJ,CAAP;EAIH,CAvFE,CAAP;AAyFH,C"}
|
package/importPages/utils.d.ts
CHANGED
@@ -36,10 +36,10 @@ interface PageImportData {
|
|
36
36
|
}
|
37
37
|
/**
|
38
38
|
* Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.
|
39
|
-
* @param
|
39
|
+
* @param zipFileUrl
|
40
40
|
* @return PageImportData S3 file keys for all uploaded assets group by page.
|
41
41
|
*/
|
42
|
-
export declare function readExtractAndUploadZipFileContents(
|
42
|
+
export declare function readExtractAndUploadZipFileContents(zipFileUrl: string): Promise<PageImportData[]>;
|
43
43
|
export declare function initialStats(total: number): {
|
44
44
|
pending: number;
|
45
45
|
processing: number;
|
package/importPages/utils.js
CHANGED
@@ -46,7 +46,6 @@ const INSTALL_DIR = "/tmp";
|
|
46
46
|
const INSTALL_EXTRACT_DIR = _path.default.join(INSTALL_DIR, "apiPageBuilderImportPage");
|
47
47
|
|
48
48
|
const FILES_COUNT_IN_EACH_BATCH = 15;
|
49
|
-
const ZIP_CONTENT_TYPE = "application/zip";
|
50
49
|
|
51
50
|
function updateImageInPageSettings(params) {
|
52
51
|
const {
|
@@ -290,14 +289,6 @@ async function uploadFilesFromS3({
|
|
290
289
|
return Promise.all(promises);
|
291
290
|
}
|
292
291
|
|
293
|
-
async function getObjectMetaFromS3(Key) {
|
294
|
-
const meta = await _s3Stream.s3Stream.getObjectHead(Key);
|
295
|
-
|
296
|
-
if (meta.ContentType !== ZIP_CONTENT_TYPE) {
|
297
|
-
throw new _error.default(`Unsupported file type: "${meta.ContentType}"`, "UNSUPPORTED_FILE");
|
298
|
-
}
|
299
|
-
}
|
300
|
-
|
301
292
|
function getOldFileKey(key) {
|
302
293
|
/*
|
303
294
|
* Because we know the naming convention, we can extract the old key from new key.
|
@@ -318,32 +309,23 @@ function getFileNameWithoutExt(fileName) {
|
|
318
309
|
|
319
310
|
/**
|
320
311
|
* Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.
|
321
|
-
* @param
|
312
|
+
* @param zipFileUrl
|
322
313
|
* @return PageImportData S3 file keys for all uploaded assets group by page.
|
323
314
|
*/
|
324
|
-
async function readExtractAndUploadZipFileContents(
|
315
|
+
async function readExtractAndUploadZipFileContents(zipFileUrl) {
|
325
316
|
const log = console.log;
|
326
317
|
const pageImportDataList = [];
|
327
|
-
let readStream; // Check whether it is a URL
|
328
318
|
|
329
|
-
|
330
|
-
const response = await (0, _nodeFetch.default)(zipFileKey);
|
319
|
+
const zipFileName = _path.default.basename(zipFileUrl).split("?")[0];
|
331
320
|
|
332
|
-
|
333
|
-
throw new _error.default(`Unable to downloading file: "${zipFileKey}"`, response.statusText);
|
334
|
-
}
|
321
|
+
const response = await (0, _nodeFetch.default)(zipFileUrl);
|
335
322
|
|
336
|
-
|
337
|
-
|
338
|
-
// We're first retrieving object's meta data, just to check whether the file is available at the given Key
|
339
|
-
await getObjectMetaFromS3(zipFileKey);
|
340
|
-
readStream = _s3Stream.s3Stream.readStream(zipFileKey);
|
323
|
+
if (!response.ok) {
|
324
|
+
throw new _error.default(`Unable to downloading file: "${zipFileUrl}"`, response.statusText);
|
341
325
|
}
|
342
326
|
|
343
|
-
const
|
344
|
-
|
345
|
-
const zipFileName = _path.default.basename(zipFileKey); // Read export file and download it in the disk
|
346
|
-
|
327
|
+
const readStream = response.body;
|
328
|
+
const uniquePath = (0, _uniqid.default)("IMPORT_PAGES/"); // Read export file and download it in the disk
|
347
329
|
|
348
330
|
const ZIP_FILE_PATH = _path.default.join(INSTALL_DIR, zipFileName);
|
349
331
|
|
@@ -352,7 +334,7 @@ async function readExtractAndUploadZipFileContents(zipFileKey) {
|
|
352
334
|
log(`Downloaded file "${zipFileName}" at ${ZIP_FILE_PATH}`); // Extract the downloaded zip file
|
353
335
|
|
354
336
|
const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);
|
355
|
-
log(`Removing ZIP file "${
|
337
|
+
log(`Removing ZIP file "${zipFileUrl}" from ${ZIP_FILE_PATH}`);
|
356
338
|
await (0, _downloadInstallFiles.deleteFile)(ZIP_FILE_PATH); // Extract each page zip and upload their content's to S3
|
357
339
|
|
358
340
|
for (let i = 0; i < zipFilePaths.length; i++) {
|
package/importPages/utils.js.map
CHANGED
@@ -1 +1 @@
|
|
1
|
-
{"version":3,"names":["streamPipeline","promisify","pipeline","INSTALL_DIR","INSTALL_EXTRACT_DIR","path","join","FILES_COUNT_IN_EACH_BATCH","ZIP_CONTENT_TYPE","updateImageInPageSettings","params","settings","fileIdToKeyMap","srcPrefix","newSettings","srcPrefixWithoutTrailingSlash","endsWith","slice","dotProp","get","set","general","image","id","social","updateFilesInPageData","data","Array","isArray","i","length","element","tuple","Object","entries","key","value","has","name","src","uploadPageAssets","context","filesData","fileUploadsData","Map","process","env","NODE_ENV","console","log","fileKeyToFileMap","file","type","fileUploadResults","uploadFilesFromS3","oldKeyToNewKeyMap","assets","createFilesInput","map","uploadResult","newKey","Key","getOldFileKey","size","meta","tags","filter","Boolean","createFilesPromises","createFilesInputChunks","chunk","createFilesInputChunk","push","fileManager","files","createFilesInBatch","Promise","all","importPage","pageKey","PAGE_EXTRACT_DIR","ensureDirSync","pageDataFileKey","PAGE_DATA_FILE_PATH","basename","resolve","reject","s3Stream","readStream","on","pipe","createWriteStream","page","loadJson","getSettings","content","deleteFile","deleteS3Folder","dirname","oldKeysForAssets","keys","promises","oldKey","tempNewKey","fileMetaData","uniqueId","streamPassThrough","streamPassThroughUploadPromise","promise","writeStream","getObjectMetaFromS3","getObjectHead","ContentType","WebinyError","rest","split","e","FILE_CONTENT_TYPE","getFileNameWithoutExt","fileName","replace","extname","readExtractAndUploadZipFileContents","zipFileKey","pageImportDataList","startsWith","response","fetch","ok","statusText","body","uniquePath","zipFileName","ZIP_FILE_PATH","zipFilePaths","extractZipToDisk","currentPath","dataMap","extractZipAndUploadToS3","ASSETS_DIR_NAME","preparePageDataDirMap","filePath","isAsset","listObject","Contents","c","deleteFilePromises","deleteObject","initialStats","total","PageImportExportTaskStatus","PENDING","PROCESSING","COMPLETED","FAILED","exportFileZipPath","pageZipFilePaths","uniqueFolderNameForExport","EXPORT_FILE_EXTRACTION_PATH","yauzl","open","lazyEntries","err","zipFile","warn","info","entryCount","readEntry","entry","test","openReadStream","catch","error","pageDataZipFilePath","filePaths","fileUploadPromises","uniquePageKey","then","res","forEach","r"],"sources":["utils.ts"],"sourcesContent":["import uniqueId from \"uniqid\";\nimport S3 from \"aws-sdk/clients/s3\";\nimport dotProp from \"dot-prop-immutable\";\nimport { createWriteStream } from \"fs\";\nimport { ensureDirSync } from \"fs-extra\";\nimport { promisify } from \"util\";\nimport { pipeline } from \"stream\";\nimport fetch from \"node-fetch\";\nimport path from \"path\";\nimport yauzl from \"yauzl\";\nimport chunk from \"lodash/chunk\";\nimport loadJson from \"load-json-file\";\nimport { FileInput } from \"@webiny/api-file-manager/types\";\nimport WebinyError from \"@webiny/error\";\nimport { deleteFile } from \"@webiny/api-page-builder/graphql/crud/install/utils/downloadInstallFiles\";\nimport { File, PageImportExportTaskStatus } from \"~/types\";\nimport { PbPageImportExportContext } from \"~/graphql/types\";\nimport { s3Stream } from \"~/exportPages/s3Stream\";\nimport { ExportedPageData } from \"~/exportPages/utils\";\nimport { PageSettings } from \"@webiny/api-page-builder/types\";\n\ninterface FileItem extends File {\n key: string;\n type: string;\n name: string;\n size: number;\n meta: Record<string, any>;\n tags: string[];\n}\n\nconst streamPipeline = promisify(pipeline);\n\nconst INSTALL_DIR = \"/tmp\";\nconst INSTALL_EXTRACT_DIR = path.join(INSTALL_DIR, \"apiPageBuilderImportPage\");\nconst FILES_COUNT_IN_EACH_BATCH = 15;\nconst ZIP_CONTENT_TYPE = \"application/zip\";\n\ninterface UpdateFilesInPageDataParams {\n data: Record<string, any>;\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n}\n\ninterface UpdateImageInPageSettingsParams {\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n settings: PageSettings;\n}\n\nfunction updateImageInPageSettings(\n params: UpdateImageInPageSettingsParams\n): UpdateImageInPageSettingsParams[\"settings\"] {\n const { settings, fileIdToKeyMap, srcPrefix } = params;\n let newSettings = settings;\n\n const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith(\"/\")\n ? srcPrefix.slice(0, -1)\n : srcPrefix;\n\n if (dotProp.get(newSettings, \"general.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"general.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.general?.image?.id || \"\"\n )}`\n );\n }\n if (dotProp.get(newSettings, \"social.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"social.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.social?.image?.id || \"\"\n )}`\n );\n }\n\n return newSettings;\n}\n\nfunction updateFilesInPageData({ data, fileIdToKeyMap, srcPrefix }: UpdateFilesInPageDataParams) {\n // BASE CASE: Termination point\n if (!data || typeof data !== \"object\") {\n return;\n }\n // Recursively call function if data is array\n if (Array.isArray(data)) {\n for (let i = 0; i < data.length; i++) {\n const element = data[i];\n updateFilesInPageData({ data: element, fileIdToKeyMap, srcPrefix });\n }\n return;\n }\n // Main logic\n const tuple = Object.entries(data);\n for (let i = 0; i < tuple.length; i++) {\n const [key, value] = tuple[i];\n\n if (key === \"file\" && value && fileIdToKeyMap.has(value.id)) {\n value.key = fileIdToKeyMap.get(value.id);\n value.name = fileIdToKeyMap.get(value.id);\n value.src = `${srcPrefix}${srcPrefix.endsWith(\"/\") ? \"\" : \"/\"}${fileIdToKeyMap.get(\n value.id\n )}`;\n } else {\n updateFilesInPageData({ data: value, srcPrefix, fileIdToKeyMap });\n }\n }\n}\n\ninterface UploadPageAssetsParams {\n context: PbPageImportExportContext;\n filesData: FileItem[];\n fileUploadsData: FileUploadsData;\n}\n\ninterface UploadPageAssetsReturnType {\n fileIdToKeyMap: Map<string, string>;\n}\n\nexport const uploadPageAssets = async (\n params: UploadPageAssetsParams\n): Promise<UploadPageAssetsReturnType> => {\n const { context, filesData, fileUploadsData } = params;\n // Save uploaded file key against static id for later use.\n const fileIdToKeyMap = new Map<string, string>();\n /**\n * This function contains logic of file download from S3.\n * Current we're not mocking zip file download from S3 in tests at the moment.\n * So, we're manually mocking it in case of test just by returning an empty object.\n */\n if (process.env.NODE_ENV === \"test\") {\n return {\n fileIdToKeyMap\n };\n }\n console.log(\"INSIDE uploadPageAssets\");\n\n // Save files meta data against old key for later use.\n const fileKeyToFileMap = new Map<string, FileItem>();\n // Initialize maps.\n for (let i = 0; i < filesData.length; i++) {\n const file = filesData[i];\n fileKeyToFileMap.set(file.key, file);\n\n // Initialize the value\n fileIdToKeyMap.set(file.id, file.type);\n }\n\n const fileUploadResults = await uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap: fileUploadsData.assets\n });\n\n // Create files in File Manager\n const createFilesInput = fileUploadResults\n .map((uploadResult): FileInput | null => {\n const newKey = uploadResult.Key;\n const file = fileKeyToFileMap.get(getOldFileKey(newKey));\n if (!file) {\n return null;\n }\n\n // Update the file map with newly uploaded file.\n fileIdToKeyMap.set(file.id, newKey);\n\n return {\n key: newKey,\n name: file.name,\n size: file.size,\n type: file.type,\n meta: file.meta,\n tags: file.tags\n };\n })\n .filter(Boolean) as FileInput[];\n\n const createFilesPromises = [];\n // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).\n const createFilesInputChunks = chunk(createFilesInput, FILES_COUNT_IN_EACH_BATCH);\n for (let i = 0; i < createFilesInputChunks.length; i++) {\n const createFilesInputChunk = createFilesInputChunks[i];\n createFilesPromises.push(\n /*\n * We need to break down files into chunks because\n * `createFilesInBatch` operation has a limit on number of files it can handle at once.\n */\n context.fileManager.files.createFilesInBatch(createFilesInputChunk)\n );\n }\n\n await Promise.all(createFilesPromises);\n\n return {\n fileIdToKeyMap\n };\n};\n\ninterface FileUploadsData {\n data: string;\n assets: Record<string, string>;\n}\n\ninterface ImportPageParams {\n key: string;\n pageKey: string;\n context: PbPageImportExportContext;\n fileUploadsData: FileUploadsData;\n}\n\nexport async function importPage({\n pageKey,\n context,\n fileUploadsData\n}: ImportPageParams): Promise<ExportedPageData[\"page\"]> {\n const log = console.log;\n\n // Making Directory for page in which we're going to extract the page data file.\n const PAGE_EXTRACT_DIR = path.join(INSTALL_EXTRACT_DIR, pageKey);\n ensureDirSync(PAGE_EXTRACT_DIR);\n\n const pageDataFileKey = dotProp.get(fileUploadsData, `data`);\n const PAGE_DATA_FILE_PATH = path.join(PAGE_EXTRACT_DIR, path.basename(pageDataFileKey));\n\n log(`Downloading Page data file: ${pageDataFileKey} at \"${PAGE_DATA_FILE_PATH}\"`);\n // Download and save page data file in disk.\n await new Promise((resolve, reject) => {\n s3Stream\n .readStream(pageDataFileKey)\n .on(\"error\", reject)\n .pipe(createWriteStream(PAGE_DATA_FILE_PATH))\n .on(\"error\", reject)\n .on(\"finish\", resolve);\n });\n\n // Load the page data file from disk.\n log(`Load file ${pageDataFileKey}`);\n const { page, files } = await loadJson<ExportedPageData>(PAGE_DATA_FILE_PATH);\n\n // Only update page data if there are files.\n if (files && Array.isArray(files) && files.length > 0) {\n // Upload page assets.\n const { fileIdToKeyMap } = await uploadPageAssets({\n context,\n /**\n * TODO @ts-refactor @ashutosh figure out correct types.\n */\n // @ts-ignore\n filesData: files,\n fileUploadsData\n });\n\n const settings = await context.fileManager.settings.getSettings();\n\n const { srcPrefix = \"\" } = settings || {};\n updateFilesInPageData({\n data: page.content || {},\n fileIdToKeyMap,\n srcPrefix\n });\n\n page.settings = updateImageInPageSettings({\n settings: page.settings || {},\n fileIdToKeyMap,\n srcPrefix\n });\n }\n\n log(\"Removing Directory for page...\");\n await deleteFile(pageKey);\n\n log(`Remove page contents from S3...`);\n await deleteS3Folder(path.dirname(fileUploadsData.data));\n\n return page;\n}\n\ninterface UploadFilesFromZipParams {\n fileKeyToFileMap: Map<string, any>;\n oldKeyToNewKeyMap: Record<string, string>;\n}\n\nasync function uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap\n}: UploadFilesFromZipParams): Promise<S3.ManagedUpload.SendData[]> {\n const oldKeysForAssets = Object.keys(oldKeyToNewKeyMap);\n\n const promises = [];\n // Upload all assets.\n for (let i = 0; i < oldKeysForAssets.length; i++) {\n const oldKey = oldKeysForAssets[i];\n const tempNewKey = oldKeyToNewKeyMap[oldKey];\n\n // Read file.\n const readStream = s3Stream.readStream(tempNewKey);\n // Get file meta data.\n const fileMetaData = fileKeyToFileMap.get(oldKey);\n\n if (fileMetaData) {\n const newKey = uniqueId(\"\", `-${fileMetaData.key}`);\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, fileMetaData.type);\n readStream.pipe(streamPassThrough);\n promises.push(promise);\n\n console.log(`Successfully queued file \"${newKey}\"`);\n }\n }\n\n return Promise.all(promises);\n}\n\nasync function getObjectMetaFromS3(Key: string) {\n const meta = await s3Stream.getObjectHead(Key);\n\n if (meta.ContentType !== ZIP_CONTENT_TYPE) {\n throw new WebinyError(`Unsupported file type: \"${meta.ContentType}\"`, \"UNSUPPORTED_FILE\");\n }\n}\n\nfunction getOldFileKey(key: string) {\n /*\n * Because we know the naming convention, we can extract the old key from new key.\n */\n try {\n const [, ...rest] = key.split(\"-\");\n return rest.join(\"-\");\n } catch (e) {\n return key;\n }\n}\n\nconst FILE_CONTENT_TYPE = \"application/octet-stream\";\n\nfunction getFileNameWithoutExt(fileName: string): string {\n return path.basename(fileName).replace(path.extname(fileName), \"\");\n}\n\ninterface PageImportData {\n assets: Record<string, string>;\n data: string;\n key: string;\n}\n\n/**\n * Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.\n * @param zipFileKey\n * @return PageImportData S3 file keys for all uploaded assets group by page.\n */\nexport async function readExtractAndUploadZipFileContents(\n zipFileKey: string\n): Promise<PageImportData[]> {\n const log = console.log;\n const pageImportDataList = [];\n let readStream;\n // Check whether it is a URL\n if (zipFileKey.startsWith(\"http\")) {\n const response = await fetch(zipFileKey);\n if (!response.ok) {\n throw new WebinyError(\n `Unable to downloading file: \"${zipFileKey}\"`,\n response.statusText\n );\n }\n\n readStream = response.body;\n } else {\n // We're first retrieving object's meta data, just to check whether the file is available at the given Key\n await getObjectMetaFromS3(zipFileKey);\n\n readStream = s3Stream.readStream(zipFileKey);\n }\n\n const uniquePath = uniqueId(\"IMPORT_PAGES/\");\n const zipFileName = path.basename(zipFileKey);\n // Read export file and download it in the disk\n const ZIP_FILE_PATH = path.join(INSTALL_DIR, zipFileName);\n\n const writeStream = createWriteStream(ZIP_FILE_PATH);\n await streamPipeline(readStream, writeStream);\n log(`Downloaded file \"${zipFileName}\" at ${ZIP_FILE_PATH}`);\n\n // Extract the downloaded zip file\n const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);\n\n log(`Removing ZIP file \"${zipFileKey}\" from ${ZIP_FILE_PATH}`);\n await deleteFile(ZIP_FILE_PATH);\n\n // Extract each page zip and upload their content's to S3\n for (let i = 0; i < zipFilePaths.length; i++) {\n const currentPath = zipFilePaths[i];\n const dataMap = await extractZipAndUploadToS3(currentPath, uniquePath);\n pageImportDataList.push(dataMap);\n }\n log(\"Removing all ZIP files located at \", path.dirname(zipFilePaths[0]));\n await deleteFile(path.dirname(zipFilePaths[0]));\n\n return pageImportDataList;\n}\n\nconst ASSETS_DIR_NAME = \"/assets\";\n\nfunction preparePageDataDirMap({\n map,\n filePath,\n newKey\n}: {\n map: PageImportData;\n filePath: string;\n newKey: string;\n}): PageImportData {\n const dirname = path.dirname(filePath);\n const fileName = path.basename(filePath);\n /*\n * We want to use dot (.) as part of object key rather than creating nested object(s).\n * Also, the file name might contain dots in it beside the extension, so, we are escaping them all.\n */\n const oldKey = fileName.replace(/\\./g, \"\\\\.\");\n\n const isAsset = dirname.endsWith(ASSETS_DIR_NAME);\n\n if (isAsset) {\n map = dotProp.set(map, `assets.${oldKey}`, newKey);\n } else {\n // We only need to know the newKey for data file.\n map = dotProp.set(map, `data`, newKey);\n }\n\n return map;\n}\n\nasync function deleteS3Folder(key: string): Promise<void> {\n // Append trailing slash i.e \"/\" to key to make sure we only delete a specific folder.\n if (!key.endsWith(\"/\")) {\n key = `${key}/`;\n }\n\n const response = await s3Stream.listObject(key);\n const keys = (response.Contents || []).map(c => c.Key).filter(Boolean) as string[];\n console.log(`Found ${keys.length} files.`);\n\n const deleteFilePromises = keys.map(key => s3Stream.deleteObject(key));\n\n await Promise.all(deleteFilePromises);\n console.log(`Successfully deleted ${deleteFilePromises.length} files.`);\n}\n\n// export const zeroPad = version => `${version}`.padStart(5, \"0\");\n\nexport function initialStats(total: number) {\n return {\n [PageImportExportTaskStatus.PENDING]: total,\n [PageImportExportTaskStatus.PROCESSING]: 0,\n [PageImportExportTaskStatus.COMPLETED]: 0,\n [PageImportExportTaskStatus.FAILED]: 0,\n total\n };\n}\n\nfunction extractZipToDisk(exportFileZipPath: string): Promise<string[]> {\n return new Promise((resolve, reject) => {\n const pageZipFilePaths: string[] = [];\n const uniqueFolderNameForExport = getFileNameWithoutExt(exportFileZipPath);\n const EXPORT_FILE_EXTRACTION_PATH = path.join(INSTALL_DIR, uniqueFolderNameForExport);\n // Make sure DIR exists\n ensureDirSync(EXPORT_FILE_EXTRACTION_PATH);\n\n yauzl.open(exportFileZipPath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", exportFileZipPath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Missing zip file resource for path: \" + exportFileZipPath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn(\"ERROR: Failed on END event for file: \", exportFileZipPath, err);\n reject(err);\n }\n resolve(pageZipFilePaths);\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed to openReadStream for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\n \"ERROR: Missing Read Stream Resource when extracting to disk.\"\n );\n reject(\"Missing Read Stream Resource.\");\n return;\n }\n\n const filePath = path.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);\n\n readStream.on(\"end\", function () {\n pageZipFilePaths.push(filePath);\n zipFile.readEntry();\n });\n\n streamPipeline(readStream, createWriteStream(filePath)).catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n\nfunction extractZipAndUploadToS3(\n pageDataZipFilePath: string,\n uniquePath: string\n): Promise<PageImportData> {\n return new Promise((resolve, reject) => {\n const filePaths = [];\n const fileUploadPromises: Promise<S3.ManagedUpload.SendData>[] = [];\n const uniquePageKey = getFileNameWithoutExt(pageDataZipFilePath);\n let dataMap: PageImportData = {\n key: uniquePageKey,\n assets: {},\n data: \"\"\n };\n yauzl.open(pageDataZipFilePath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", pageDataZipFilePath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Probably failed to extract zip: \" + pageDataZipFilePath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn('ERROR: Failed on \"END\" for file: ', pageDataZipFilePath, err);\n reject(err);\n }\n\n Promise.all(fileUploadPromises).then(res => {\n res.forEach(r => {\n console.info(\"Done uploading... \", r);\n });\n resolve(dataMap);\n });\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed while performing [openReadStream] for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\"ERROR: Missing Read Stream while importing pages.\");\n reject(\"Missing Read Strea Resource.\");\n return;\n }\n readStream.on(\"end\", function () {\n filePaths.push(entry.fileName);\n zipFile.readEntry();\n });\n\n const newKey = `${uniquePath}/${uniquePageKey}/${entry.fileName}`;\n // Modify in place\n dataMap = preparePageDataDirMap({\n map: dataMap,\n filePath: entry.fileName,\n newKey\n });\n\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, FILE_CONTENT_TYPE);\n\n streamPipeline(readStream, streamPassThrough)\n .then(() => {\n fileUploadPromises.push(promise);\n })\n .catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n"],"mappings":";;;;;;;;;;;;AAAA;;AAEA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AAEA;;AACA;;AACA;;AAEA;;AAaA,MAAMA,cAAc,GAAG,IAAAC,eAAA,EAAUC,gBAAV,CAAvB;AAEA,MAAMC,WAAW,GAAG,MAApB;;AACA,MAAMC,mBAAmB,GAAGC,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuB,0BAAvB,CAA5B;;AACA,MAAMI,yBAAyB,GAAG,EAAlC;AACA,MAAMC,gBAAgB,GAAG,iBAAzB;;AAcA,SAASC,yBAAT,CACIC,MADJ,EAE+C;EAC3C,MAAM;IAAEC,QAAF;IAAYC,cAAZ;IAA4BC;EAA5B,IAA0CH,MAAhD;EACA,IAAII,WAAW,GAAGH,QAAlB;EAEA,MAAMI,6BAA6B,GAAGF,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAChCH,SAAS,CAACI,KAAV,CAAgB,CAAhB,EAAmB,CAAC,CAApB,CADgC,GAEhCJ,SAFN;;EAIA,IAAIK,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,mBAAzB,CAAJ,EAAmD;IAAA;;IAC/CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,mBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,sBAAAR,QAAQ,CAACU,OAAT,iGAAkBC,KAAlB,gFAAyBC,EAAzB,KAA+B,EADC,CAElC,EALQ,CAAd;EAOH;;EACD,IAAIL,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,kBAAzB,CAAJ,EAAkD;IAAA;;IAC9CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,kBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,qBAAAR,QAAQ,CAACa,MAAT,+FAAiBF,KAAjB,gFAAwBC,EAAxB,KAA8B,EADE,CAElC,EALQ,CAAd;EAOH;;EAED,OAAOT,WAAP;AACH;;AAED,SAASW,qBAAT,CAA+B;EAAEC,IAAF;EAAQd,cAAR;EAAwBC;AAAxB,CAA/B,EAAiG;EAC7F;EACA,IAAI,CAACa,IAAD,IAAS,OAAOA,IAAP,KAAgB,QAA7B,EAAuC;IACnC;EACH,CAJ4F,CAK7F;;;EACA,IAAIC,KAAK,CAACC,OAAN,CAAcF,IAAd,CAAJ,EAAyB;IACrB,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGH,IAAI,CAACI,MAAzB,EAAiCD,CAAC,EAAlC,EAAsC;MAClC,MAAME,OAAO,GAAGL,IAAI,CAACG,CAAD,CAApB;MACAJ,qBAAqB,CAAC;QAAEC,IAAI,EAAEK,OAAR;QAAiBnB,cAAjB;QAAiCC;MAAjC,CAAD,CAArB;IACH;;IACD;EACH,CAZ4F,CAa7F;;;EACA,MAAMmB,KAAK,GAAGC,MAAM,CAACC,OAAP,CAAeR,IAAf,CAAd;;EACA,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGG,KAAK,CAACF,MAA1B,EAAkCD,CAAC,EAAnC,EAAuC;IACnC,MAAM,CAACM,GAAD,EAAMC,KAAN,IAAeJ,KAAK,CAACH,CAAD,CAA1B;;IAEA,IAAIM,GAAG,KAAK,MAAR,IAAkBC,KAAlB,IAA2BxB,cAAc,CAACyB,GAAf,CAAmBD,KAAK,CAACb,EAAzB,CAA/B,EAA6D;MACzDa,KAAK,CAACD,GAAN,GAAYvB,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAZ;MACAa,KAAK,CAACE,IAAN,GAAa1B,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAb;MACAa,KAAK,CAACG,GAAN,GAAa,GAAE1B,SAAU,GAAEA,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAA0B,EAA1B,GAA+B,GAAI,GAAEJ,cAAc,CAACO,GAAf,CAC5DiB,KAAK,CAACb,EADsD,CAE9D,EAFF;IAGH,CAND,MAMO;MACHE,qBAAqB,CAAC;QAAEC,IAAI,EAAEU,KAAR;QAAevB,SAAf;QAA0BD;MAA1B,CAAD,CAArB;IACH;EACJ;AACJ;;AAYM,MAAM4B,gBAAgB,GAAG,MAC5B9B,MAD4B,IAEU;EACtC,MAAM;IAAE+B,OAAF;IAAWC,SAAX;IAAsBC;EAAtB,IAA0CjC,MAAhD,CADsC,CAEtC;;EACA,MAAME,cAAc,GAAG,IAAIgC,GAAJ,EAAvB;EACA;AACJ;AACA;AACA;AACA;;EACI,IAAIC,OAAO,CAACC,GAAR,CAAYC,QAAZ,KAAyB,MAA7B,EAAqC;IACjC,OAAO;MACHnC;IADG,CAAP;EAGH;;EACDoC,OAAO,CAACC,GAAR,CAAY,yBAAZ,EAdsC,CAgBtC;;EACA,MAAMC,gBAAgB,GAAG,IAAIN,GAAJ,EAAzB,CAjBsC,CAkBtC;;EACA,KAAK,IAAIf,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGa,SAAS,CAACZ,MAA9B,EAAsCD,CAAC,EAAvC,EAA2C;IACvC,MAAMsB,IAAI,GAAGT,SAAS,CAACb,CAAD,CAAtB;IACAqB,gBAAgB,CAAC9B,GAAjB,CAAqB+B,IAAI,CAAChB,GAA1B,EAA+BgB,IAA/B,EAFuC,CAIvC;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4B4B,IAAI,CAACC,IAAjC;EACH;;EAED,MAAMC,iBAAiB,GAAG,MAAMC,iBAAiB,CAAC;IAC9CJ,gBAD8C;IAE9CK,iBAAiB,EAAEZ,eAAe,CAACa;EAFW,CAAD,CAAjD,CA3BsC,CAgCtC;;EACA,MAAMC,gBAAgB,GAAGJ,iBAAiB,CACrCK,GADoB,CACfC,YAAD,IAAoC;IACrC,MAAMC,MAAM,GAAGD,YAAY,CAACE,GAA5B;IACA,MAAMV,IAAI,GAAGD,gBAAgB,CAAC/B,GAAjB,CAAqB2C,aAAa,CAACF,MAAD,CAAlC,CAAb;;IACA,IAAI,CAACT,IAAL,EAAW;MACP,OAAO,IAAP;IACH,CALoC,CAOrC;;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4BqC,MAA5B;IAEA,OAAO;MACHzB,GAAG,EAAEyB,MADF;MAEHtB,IAAI,EAAEa,IAAI,CAACb,IAFR;MAGHyB,IAAI,EAAEZ,IAAI,CAACY,IAHR;MAIHX,IAAI,EAAED,IAAI,CAACC,IAJR;MAKHY,IAAI,EAAEb,IAAI,CAACa,IALR;MAMHC,IAAI,EAAEd,IAAI,CAACc;IANR,CAAP;EAQH,CAnBoB,EAoBpBC,MApBoB,CAoBbC,OApBa,CAAzB;EAsBA,MAAMC,mBAAmB,GAAG,EAA5B,CAvDsC,CAwDtC;;EACA,MAAMC,sBAAsB,GAAG,IAAAC,cAAA,EAAMb,gBAAN,EAAwBlD,yBAAxB,CAA/B;;EACA,KAAK,IAAIsB,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGwC,sBAAsB,CAACvC,MAA3C,EAAmDD,CAAC,EAApD,EAAwD;IACpD,MAAM0C,qBAAqB,GAAGF,sBAAsB,CAACxC,CAAD,CAApD;IACAuC,mBAAmB,CAACI,IAApB;IACI;AACZ;AACA;AACA;IACY/B,OAAO,CAACgC,WAAR,CAAoBC,KAApB,CAA0BC,kBAA1B,CAA6CJ,qBAA7C,CALJ;EAOH;;EAED,MAAMK,OAAO,CAACC,GAAR,CAAYT,mBAAZ,CAAN;EAEA,OAAO;IACHxD;EADG,CAAP;AAGH,CA5EM;;;;AA0FA,eAAekE,UAAf,CAA0B;EAC7BC,OAD6B;EAE7BtC,OAF6B;EAG7BE;AAH6B,CAA1B,EAIiD;EACpD,MAAMM,GAAG,GAAGD,OAAO,CAACC,GAApB,CADoD,CAGpD;;EACA,MAAM+B,gBAAgB,GAAG3E,aAAA,CAAKC,IAAL,CAAUF,mBAAV,EAA+B2E,OAA/B,CAAzB;;EACA,IAAAE,sBAAA,EAAcD,gBAAd;;EAEA,MAAME,eAAe,GAAGhE,yBAAA,CAAQC,GAAR,CAAYwB,eAAZ,EAA8B,MAA9B,CAAxB;;EACA,MAAMwC,mBAAmB,GAAG9E,aAAA,CAAKC,IAAL,CAAU0E,gBAAV,EAA4B3E,aAAA,CAAK+E,QAAL,CAAcF,eAAd,CAA5B,CAA5B;;EAEAjC,GAAG,CAAE,+BAA8BiC,eAAgB,QAAOC,mBAAoB,GAA3E,CAAH,CAVoD,CAWpD;;EACA,MAAM,IAAIP,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACnCC,kBAAA,CACKC,UADL,CACgBN,eADhB,EAEKO,EAFL,CAEQ,OAFR,EAEiBH,MAFjB,EAGKI,IAHL,CAGU,IAAAC,qBAAA,EAAkBR,mBAAlB,CAHV,EAIKM,EAJL,CAIQ,OAJR,EAIiBH,MAJjB,EAKKG,EALL,CAKQ,QALR,EAKkBJ,OALlB;EAMH,CAPK,CAAN,CAZoD,CAqBpD;;EACApC,GAAG,CAAE,aAAYiC,eAAgB,EAA9B,CAAH;EACA,MAAM;IAAEU,IAAF;IAAQlB;EAAR,IAAkB,MAAM,IAAAmB,qBAAA,EAA2BV,mBAA3B,CAA9B,CAvBoD,CAyBpD;;EACA,IAAIT,KAAK,IAAI/C,KAAK,CAACC,OAAN,CAAc8C,KAAd,CAAT,IAAiCA,KAAK,CAAC5C,MAAN,GAAe,CAApD,EAAuD;IACnD;IACA,MAAM;MAAElB;IAAF,IAAqB,MAAM4B,gBAAgB,CAAC;MAC9CC,OAD8C;;MAE9C;AACZ;AACA;MACY;MACAC,SAAS,EAAEgC,KANmC;MAO9C/B;IAP8C,CAAD,CAAjD;IAUA,MAAMhC,QAAQ,GAAG,MAAM8B,OAAO,CAACgC,WAAR,CAAoB9D,QAApB,CAA6BmF,WAA7B,EAAvB;IAEA,MAAM;MAAEjF,SAAS,GAAG;IAAd,IAAqBF,QAAQ,IAAI,EAAvC;IACAc,qBAAqB,CAAC;MAClBC,IAAI,EAAEkE,IAAI,CAACG,OAAL,IAAgB,EADJ;MAElBnF,cAFkB;MAGlBC;IAHkB,CAAD,CAArB;IAMA+E,IAAI,CAACjF,QAAL,GAAgBF,yBAAyB,CAAC;MACtCE,QAAQ,EAAEiF,IAAI,CAACjF,QAAL,IAAiB,EADW;MAEtCC,cAFsC;MAGtCC;IAHsC,CAAD,CAAzC;EAKH;;EAEDoC,GAAG,CAAC,gCAAD,CAAH;EACA,MAAM,IAAA+C,gCAAA,EAAWjB,OAAX,CAAN;EAEA9B,GAAG,CAAE,iCAAF,CAAH;EACA,MAAMgD,cAAc,CAAC5F,aAAA,CAAK6F,OAAL,CAAavD,eAAe,CAACjB,IAA7B,CAAD,CAApB;EAEA,OAAOkE,IAAP;AACH;;AAOD,eAAetC,iBAAf,CAAiC;EAC7BJ,gBAD6B;EAE7BK;AAF6B,CAAjC,EAGmE;EAC/D,MAAM4C,gBAAgB,GAAGlE,MAAM,CAACmE,IAAP,CAAY7C,iBAAZ,CAAzB;EAEA,MAAM8C,QAAQ,GAAG,EAAjB,CAH+D,CAI/D;;EACA,KAAK,IAAIxE,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGsE,gBAAgB,CAACrE,MAArC,EAA6CD,CAAC,EAA9C,EAAkD;IAC9C,MAAMyE,MAAM,GAAGH,gBAAgB,CAACtE,CAAD,CAA/B;IACA,MAAM0E,UAAU,GAAGhD,iBAAiB,CAAC+C,MAAD,CAApC,CAF8C,CAI9C;;IACA,MAAMd,UAAU,GAAGD,kBAAA,CAASC,UAAT,CAAoBe,UAApB,CAAnB,CAL8C,CAM9C;;;IACA,MAAMC,YAAY,GAAGtD,gBAAgB,CAAC/B,GAAjB,CAAqBmF,MAArB,CAArB;;IAEA,IAAIE,YAAJ,EAAkB;MACd,MAAM5C,MAAM,GAAG,IAAA6C,eAAA,EAAS,EAAT,EAAc,IAAGD,YAAY,CAACrE,GAAI,EAAlC,CAAf;;MACA,MAAM;QAAEuE,iBAAF;QAAqBC,8BAA8B,EAAEC;MAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6B4C,YAAY,CAACpD,IAA1C,CADJ;;MAEAoC,UAAU,CAACE,IAAX,CAAgBgB,iBAAhB;MACAL,QAAQ,CAAC7B,IAAT,CAAcoC,OAAd;MAEA5D,OAAO,CAACC,GAAR,CAAa,6BAA4BW,MAAO,GAAhD;IACH;EACJ;;EAED,OAAOgB,OAAO,CAACC,GAAR,CAAYwB,QAAZ,CAAP;AACH;;AAED,eAAeS,mBAAf,CAAmCjD,GAAnC,EAAgD;EAC5C,MAAMG,IAAI,GAAG,MAAMuB,kBAAA,CAASwB,aAAT,CAAuBlD,GAAvB,CAAnB;;EAEA,IAAIG,IAAI,CAACgD,WAAL,KAAqBxG,gBAAzB,EAA2C;IACvC,MAAM,IAAIyG,cAAJ,CAAiB,2BAA0BjD,IAAI,CAACgD,WAAY,GAA5D,EAAgE,kBAAhE,CAAN;EACH;AACJ;;AAED,SAASlD,aAAT,CAAuB3B,GAAvB,EAAoC;EAChC;AACJ;AACA;EACI,IAAI;IACA,MAAM,GAAG,GAAG+E,IAAN,IAAc/E,GAAG,CAACgF,KAAJ,CAAU,GAAV,CAApB;IACA,OAAOD,IAAI,CAAC5G,IAAL,CAAU,GAAV,CAAP;EACH,CAHD,CAGE,OAAO8G,CAAP,EAAU;IACR,OAAOjF,GAAP;EACH;AACJ;;AAED,MAAMkF,iBAAiB,GAAG,0BAA1B;;AAEA,SAASC,qBAAT,CAA+BC,QAA/B,EAAyD;EACrD,OAAOlH,aAAA,CAAK+E,QAAL,CAAcmC,QAAd,EAAwBC,OAAxB,CAAgCnH,aAAA,CAAKoH,OAAL,CAAaF,QAAb,CAAhC,EAAwD,EAAxD,CAAP;AACH;;AAQD;AACA;AACA;AACA;AACA;AACO,eAAeG,mCAAf,CACHC,UADG,EAEsB;EACzB,MAAM1E,GAAG,GAAGD,OAAO,CAACC,GAApB;EACA,MAAM2E,kBAAkB,GAAG,EAA3B;EACA,IAAIpC,UAAJ,CAHyB,CAIzB;;EACA,IAAImC,UAAU,CAACE,UAAX,CAAsB,MAAtB,CAAJ,EAAmC;IAC/B,MAAMC,QAAQ,GAAG,MAAM,IAAAC,kBAAA,EAAMJ,UAAN,CAAvB;;IACA,IAAI,CAACG,QAAQ,CAACE,EAAd,EAAkB;MACd,MAAM,IAAIf,cAAJ,CACD,gCAA+BU,UAAW,GADzC,EAEFG,QAAQ,CAACG,UAFP,CAAN;IAIH;;IAEDzC,UAAU,GAAGsC,QAAQ,CAACI,IAAtB;EACH,CAVD,MAUO;IACH;IACA,MAAMpB,mBAAmB,CAACa,UAAD,CAAzB;IAEAnC,UAAU,GAAGD,kBAAA,CAASC,UAAT,CAAoBmC,UAApB,CAAb;EACH;;EAED,MAAMQ,UAAU,GAAG,IAAA1B,eAAA,EAAS,eAAT,CAAnB;;EACA,MAAM2B,WAAW,GAAG/H,aAAA,CAAK+E,QAAL,CAAcuC,UAAd,CAApB,CAvByB,CAwBzB;;;EACA,MAAMU,aAAa,GAAGhI,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBiI,WAAvB,CAAtB;;EAEA,MAAMvB,WAAW,GAAG,IAAAlB,qBAAA,EAAkB0C,aAAlB,CAApB;EACA,MAAMrI,cAAc,CAACwF,UAAD,EAAaqB,WAAb,CAApB;EACA5D,GAAG,CAAE,oBAAmBmF,WAAY,QAAOC,aAAc,EAAtD,CAAH,CA7ByB,CA+BzB;;EACA,MAAMC,YAAY,GAAG,MAAMC,gBAAgB,CAACF,aAAD,CAA3C;EAEApF,GAAG,CAAE,sBAAqB0E,UAAW,UAASU,aAAc,EAAzD,CAAH;EACA,MAAM,IAAArC,gCAAA,EAAWqC,aAAX,CAAN,CAnCyB,CAqCzB;;EACA,KAAK,IAAIxG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGyG,YAAY,CAACxG,MAAjC,EAAyCD,CAAC,EAA1C,EAA8C;IAC1C,MAAM2G,WAAW,GAAGF,YAAY,CAACzG,CAAD,CAAhC;IACA,MAAM4G,OAAO,GAAG,MAAMC,uBAAuB,CAACF,WAAD,EAAcL,UAAd,CAA7C;IACAP,kBAAkB,CAACpD,IAAnB,CAAwBiE,OAAxB;EACH;;EACDxF,GAAG,CAAC,oCAAD,EAAuC5C,aAAA,CAAK6F,OAAL,CAAaoC,YAAY,CAAC,CAAD,CAAzB,CAAvC,CAAH;EACA,MAAM,IAAAtC,gCAAA,EAAW3F,aAAA,CAAK6F,OAAL,CAAaoC,YAAY,CAAC,CAAD,CAAzB,CAAX,CAAN;EAEA,OAAOV,kBAAP;AACH;;AAED,MAAMe,eAAe,GAAG,SAAxB;;AAEA,SAASC,qBAAT,CAA+B;EAC3BlF,GAD2B;EAE3BmF,QAF2B;EAG3BjF;AAH2B,CAA/B,EAQmB;EACf,MAAMsC,OAAO,GAAG7F,aAAA,CAAK6F,OAAL,CAAa2C,QAAb,CAAhB;;EACA,MAAMtB,QAAQ,GAAGlH,aAAA,CAAK+E,QAAL,CAAcyD,QAAd,CAAjB;EACA;AACJ;AACA;AACA;;;EACI,MAAMvC,MAAM,GAAGiB,QAAQ,CAACC,OAAT,CAAiB,KAAjB,EAAwB,KAAxB,CAAf;EAEA,MAAMsB,OAAO,GAAG5C,OAAO,CAAClF,QAAR,CAAiB2H,eAAjB,CAAhB;;EAEA,IAAIG,OAAJ,EAAa;IACTpF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,UAAS4C,MAAO,EAAlC,EAAqC1C,MAArC,CAAN;EACH,CAFD,MAEO;IACH;IACAF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,MAAlB,EAAyBE,MAAzB,CAAN;EACH;;EAED,OAAOF,GAAP;AACH;;AAED,eAAeuC,cAAf,CAA8B9D,GAA9B,EAA0D;EACtD;EACA,IAAI,CAACA,GAAG,CAACnB,QAAJ,CAAa,GAAb,CAAL,EAAwB;IACpBmB,GAAG,GAAI,GAAEA,GAAI,GAAb;EACH;;EAED,MAAM2F,QAAQ,GAAG,MAAMvC,kBAAA,CAASwD,UAAT,CAAoB5G,GAApB,CAAvB;EACA,MAAMiE,IAAI,GAAG,CAAC0B,QAAQ,CAACkB,QAAT,IAAqB,EAAtB,EAA0BtF,GAA1B,CAA8BuF,CAAC,IAAIA,CAAC,CAACpF,GAArC,EAA0CK,MAA1C,CAAiDC,OAAjD,CAAb;EACAnB,OAAO,CAACC,GAAR,CAAa,SAAQmD,IAAI,CAACtE,MAAO,SAAjC;EAEA,MAAMoH,kBAAkB,GAAG9C,IAAI,CAAC1C,GAAL,CAASvB,GAAG,IAAIoD,kBAAA,CAAS4D,YAAT,CAAsBhH,GAAtB,CAAhB,CAA3B;EAEA,MAAMyC,OAAO,CAACC,GAAR,CAAYqE,kBAAZ,CAAN;EACAlG,OAAO,CAACC,GAAR,CAAa,wBAAuBiG,kBAAkB,CAACpH,MAAO,SAA9D;AACH,C,CAED;;;AAEO,SAASsH,YAAT,CAAsBC,KAAtB,EAAqC;EACxC,OAAO;IACH,CAACC,iCAAA,CAA2BC,OAA5B,GAAsCF,KADnC;IAEH,CAACC,iCAAA,CAA2BE,UAA5B,GAAyC,CAFtC;IAGH,CAACF,iCAAA,CAA2BG,SAA5B,GAAwC,CAHrC;IAIH,CAACH,iCAAA,CAA2BI,MAA5B,GAAqC,CAJlC;IAKHL;EALG,CAAP;AAOH;;AAED,SAASd,gBAAT,CAA0BoB,iBAA1B,EAAwE;EACpE,OAAO,IAAI/E,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMsE,gBAA0B,GAAG,EAAnC;IACA,MAAMC,yBAAyB,GAAGvC,qBAAqB,CAACqC,iBAAD,CAAvD;;IACA,MAAMG,2BAA2B,GAAGzJ,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuB0J,yBAAvB,CAApC,CAHoC,CAIpC;;;IACA,IAAA5E,sBAAA,EAAc6E,2BAAd;;IAEAC,cAAA,CAAMC,IAAN,CAAWL,iBAAX,EAA8B;MAAEM,WAAW,EAAE;IAAf,CAA9B,EAAqD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MACzE,IAAID,GAAJ,EAAS;QACLlH,OAAO,CAACoH,IAAR,CAAa,gCAAb,EAA+CT,iBAA/C,EAAkEO,GAAlE;QACA5E,MAAM,CAAC4E,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACVnH,OAAO,CAACC,GAAR,CAAY,gDAAgD0G,iBAA5D;QACArE,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MAEDtC,OAAO,CAACqH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MAEAH,OAAO,CAAC1E,EAAR,CAAW,KAAX,EAAkB,UAAUyE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACLlH,OAAO,CAACoH,IAAR,CAAa,uCAAb,EAAsDT,iBAAtD,EAAyEO,GAAzE;UACA5E,MAAM,CAAC4E,GAAD,CAAN;QACH;;QACD7E,OAAO,CAACuE,gBAAD,CAAP;MACH,CAND;MAQAO,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAAC1E,EAAR,CAAW,OAAX,EAAoB,UAAU+E,KAAV,EAAiB;QACjCxH,OAAO,CAACqH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAe1E,UAAf,EAA2B;YACrD,IAAI0E,GAAJ,EAAS;cACLlH,OAAO,CAACoH,IAAR,CACI,4CADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKA5E,MAAM,CAAC4E,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAAC1E,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CACI,8DADJ;cAGAqC,MAAM,CAAC,+BAAD,CAAN;cACA;YACH;;YAED,MAAMuD,QAAQ,GAAGxI,aAAA,CAAKC,IAAL,CAAUwJ,2BAAV,EAAuCU,KAAK,CAACjD,QAA7C,CAAjB;;YAEA/B,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7BmE,gBAAgB,CAACpF,IAAjB,CAAsBqE,QAAtB;cACAsB,OAAO,CAACI,SAAR;YACH,CAHD;YAKAvK,cAAc,CAACwF,UAAD,EAAa,IAAAG,qBAAA,EAAkBkD,QAAlB,CAAb,CAAd,CAAwD8B,KAAxD,CAA8DC,KAAK,IAAI;cACnEtF,MAAM,CAACsF,KAAD,CAAN;YACH,CAFD;UAGH,CA5BD;QA6BH;MACJ,CAvCD;IAwCH,CAhED;EAiEH,CAxEM,CAAP;AAyEH;;AAED,SAASlC,uBAAT,CACImC,mBADJ,EAEI1C,UAFJ,EAG2B;EACvB,OAAO,IAAIvD,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMwF,SAAS,GAAG,EAAlB;IACA,MAAMC,kBAAwD,GAAG,EAAjE;IACA,MAAMC,aAAa,GAAG1D,qBAAqB,CAACuD,mBAAD,CAA3C;IACA,IAAIpC,OAAuB,GAAG;MAC1BtG,GAAG,EAAE6I,aADqB;MAE1BxH,MAAM,EAAE,EAFkB;MAG1B9B,IAAI,EAAE;IAHoB,CAA9B;;IAKAqI,cAAA,CAAMC,IAAN,CAAWa,mBAAX,EAAgC;MAAEZ,WAAW,EAAE;IAAf,CAAhC,EAAuD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MAC3E,IAAID,GAAJ,EAAS;QACLlH,OAAO,CAACoH,IAAR,CAAa,gCAAb,EAA+CS,mBAA/C,EAAoEX,GAApE;QACA5E,MAAM,CAAC4E,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACVnH,OAAO,CAACC,GAAR,CAAY,4CAA4C4H,mBAAxD;QACAvF,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MACDtC,OAAO,CAACqH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MACAH,OAAO,CAAC1E,EAAR,CAAW,KAAX,EAAkB,UAAUyE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACLlH,OAAO,CAACoH,IAAR,CAAa,mCAAb,EAAkDS,mBAAlD,EAAuEX,GAAvE;UACA5E,MAAM,CAAC4E,GAAD,CAAN;QACH;;QAEDtF,OAAO,CAACC,GAAR,CAAYkG,kBAAZ,EAAgCE,IAAhC,CAAqCC,GAAG,IAAI;UACxCA,GAAG,CAACC,OAAJ,CAAYC,CAAC,IAAI;YACbpI,OAAO,CAACqH,IAAR,CAAa,oBAAb,EAAmCe,CAAnC;UACH,CAFD;UAGA/F,OAAO,CAACoD,OAAD,CAAP;QACH,CALD;MAMH,CAZD;MAcA0B,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAAC1E,EAAR,CAAW,OAAX,EAAoB,UAAU+E,KAAV,EAAiB;QACjCxH,OAAO,CAACqH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAe1E,UAAf,EAA2B;YACrD,IAAI0E,GAAJ,EAAS;cACLlH,OAAO,CAACoH,IAAR,CACI,4DADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKA5E,MAAM,CAAC4E,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAAC1E,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CAAY,mDAAZ;cACAqC,MAAM,CAAC,8BAAD,CAAN;cACA;YACH;;YACDE,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7BqF,SAAS,CAACtG,IAAV,CAAegG,KAAK,CAACjD,QAArB;cACA4C,OAAO,CAACI,SAAR;YACH,CAHD;YAKA,MAAM3G,MAAM,GAAI,GAAEuE,UAAW,IAAG6C,aAAc,IAAGR,KAAK,CAACjD,QAAS,EAAhE,CApBqD,CAqBrD;;YACAkB,OAAO,GAAGG,qBAAqB,CAAC;cAC5BlF,GAAG,EAAE+E,OADuB;cAE5BI,QAAQ,EAAE2B,KAAK,CAACjD,QAFY;cAG5B3D;YAH4B,CAAD,CAA/B;;YAMA,MAAM;cAAE8C,iBAAF;cAAqBC,8BAA8B,EAAEC;YAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6ByD,iBAA7B,CADJ;;YAGArH,cAAc,CAACwF,UAAD,EAAakB,iBAAb,CAAd,CACKuE,IADL,CACU,MAAM;cACRF,kBAAkB,CAACvG,IAAnB,CAAwBoC,OAAxB;YACH,CAHL,EAIK+D,KAJL,CAIWC,KAAK,IAAI;cACZtF,MAAM,CAACsF,KAAD,CAAN;YACH,CANL;UAOH,CAtCD;QAuCH;MACJ,CAjDD;IAkDH,CA9ED;EA+EH,CAxFM,CAAP;AAyFH"}
|
1
|
+
{"version":3,"names":["streamPipeline","promisify","pipeline","INSTALL_DIR","INSTALL_EXTRACT_DIR","path","join","FILES_COUNT_IN_EACH_BATCH","updateImageInPageSettings","params","settings","fileIdToKeyMap","srcPrefix","newSettings","srcPrefixWithoutTrailingSlash","endsWith","slice","dotProp","get","set","general","image","id","social","updateFilesInPageData","data","Array","isArray","i","length","element","tuple","Object","entries","key","value","has","name","src","uploadPageAssets","context","filesData","fileUploadsData","Map","process","env","NODE_ENV","console","log","fileKeyToFileMap","file","type","fileUploadResults","uploadFilesFromS3","oldKeyToNewKeyMap","assets","createFilesInput","map","uploadResult","newKey","Key","getOldFileKey","size","meta","tags","filter","Boolean","createFilesPromises","createFilesInputChunks","chunk","createFilesInputChunk","push","fileManager","files","createFilesInBatch","Promise","all","importPage","pageKey","PAGE_EXTRACT_DIR","ensureDirSync","pageDataFileKey","PAGE_DATA_FILE_PATH","basename","resolve","reject","s3Stream","readStream","on","pipe","createWriteStream","page","loadJson","getSettings","content","deleteFile","deleteS3Folder","dirname","oldKeysForAssets","keys","promises","oldKey","tempNewKey","fileMetaData","uniqueId","streamPassThrough","streamPassThroughUploadPromise","promise","writeStream","rest","split","e","FILE_CONTENT_TYPE","getFileNameWithoutExt","fileName","replace","extname","readExtractAndUploadZipFileContents","zipFileUrl","pageImportDataList","zipFileName","response","fetch","ok","WebinyError","statusText","body","uniquePath","ZIP_FILE_PATH","zipFilePaths","extractZipToDisk","currentPath","dataMap","extractZipAndUploadToS3","ASSETS_DIR_NAME","preparePageDataDirMap","filePath","isAsset","listObject","Contents","c","deleteFilePromises","deleteObject","initialStats","total","PageImportExportTaskStatus","PENDING","PROCESSING","COMPLETED","FAILED","exportFileZipPath","pageZipFilePaths","uniqueFolderNameForExport","EXPORT_FILE_EXTRACTION_PATH","yauzl","open","lazyEntries","err","zipFile","warn","info","entryCount","readEntry","entry","test","openReadStream","catch","error","pageDataZipFilePath","filePaths","fileUploadPromises","uniquePageKey","then","res","forEach","r"],"sources":["utils.ts"],"sourcesContent":["import uniqueId from \"uniqid\";\nimport S3 from \"aws-sdk/clients/s3\";\nimport dotProp from \"dot-prop-immutable\";\nimport { createWriteStream } from \"fs\";\nimport { ensureDirSync } from \"fs-extra\";\nimport { promisify } from \"util\";\nimport { pipeline } from \"stream\";\nimport fetch from \"node-fetch\";\nimport path from \"path\";\nimport yauzl from \"yauzl\";\nimport chunk from \"lodash/chunk\";\nimport loadJson from \"load-json-file\";\nimport { FileInput } from \"@webiny/api-file-manager/types\";\nimport WebinyError from \"@webiny/error\";\nimport { deleteFile } from \"@webiny/api-page-builder/graphql/crud/install/utils/downloadInstallFiles\";\nimport { File, PageImportExportTaskStatus } from \"~/types\";\nimport { PbPageImportExportContext } from \"~/graphql/types\";\nimport { s3Stream } from \"~/exportPages/s3Stream\";\nimport { ExportedPageData } from \"~/exportPages/utils\";\nimport { PageSettings } from \"@webiny/api-page-builder/types\";\n\ninterface FileItem extends File {\n key: string;\n type: string;\n name: string;\n size: number;\n meta: Record<string, any>;\n tags: string[];\n}\n\nconst streamPipeline = promisify(pipeline);\n\nconst INSTALL_DIR = \"/tmp\";\nconst INSTALL_EXTRACT_DIR = path.join(INSTALL_DIR, \"apiPageBuilderImportPage\");\nconst FILES_COUNT_IN_EACH_BATCH = 15;\n\ninterface UpdateFilesInPageDataParams {\n data: Record<string, any>;\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n}\n\ninterface UpdateImageInPageSettingsParams {\n fileIdToKeyMap: Map<string, string>;\n srcPrefix: string;\n settings: PageSettings;\n}\n\nfunction updateImageInPageSettings(\n params: UpdateImageInPageSettingsParams\n): UpdateImageInPageSettingsParams[\"settings\"] {\n const { settings, fileIdToKeyMap, srcPrefix } = params;\n let newSettings = settings;\n\n const srcPrefixWithoutTrailingSlash = srcPrefix.endsWith(\"/\")\n ? srcPrefix.slice(0, -1)\n : srcPrefix;\n\n if (dotProp.get(newSettings, \"general.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"general.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.general?.image?.id || \"\"\n )}`\n );\n }\n if (dotProp.get(newSettings, \"social.image.src\")) {\n newSettings = dotProp.set(\n newSettings,\n \"social.image.src\",\n `${srcPrefixWithoutTrailingSlash}/${fileIdToKeyMap.get(\n settings.social?.image?.id || \"\"\n )}`\n );\n }\n\n return newSettings;\n}\n\nfunction updateFilesInPageData({ data, fileIdToKeyMap, srcPrefix }: UpdateFilesInPageDataParams) {\n // BASE CASE: Termination point\n if (!data || typeof data !== \"object\") {\n return;\n }\n // Recursively call function if data is array\n if (Array.isArray(data)) {\n for (let i = 0; i < data.length; i++) {\n const element = data[i];\n updateFilesInPageData({ data: element, fileIdToKeyMap, srcPrefix });\n }\n return;\n }\n // Main logic\n const tuple = Object.entries(data);\n for (let i = 0; i < tuple.length; i++) {\n const [key, value] = tuple[i];\n\n if (key === \"file\" && value && fileIdToKeyMap.has(value.id)) {\n value.key = fileIdToKeyMap.get(value.id);\n value.name = fileIdToKeyMap.get(value.id);\n value.src = `${srcPrefix}${srcPrefix.endsWith(\"/\") ? \"\" : \"/\"}${fileIdToKeyMap.get(\n value.id\n )}`;\n } else {\n updateFilesInPageData({ data: value, srcPrefix, fileIdToKeyMap });\n }\n }\n}\n\ninterface UploadPageAssetsParams {\n context: PbPageImportExportContext;\n filesData: FileItem[];\n fileUploadsData: FileUploadsData;\n}\n\ninterface UploadPageAssetsReturnType {\n fileIdToKeyMap: Map<string, string>;\n}\n\nexport const uploadPageAssets = async (\n params: UploadPageAssetsParams\n): Promise<UploadPageAssetsReturnType> => {\n const { context, filesData, fileUploadsData } = params;\n // Save uploaded file key against static id for later use.\n const fileIdToKeyMap = new Map<string, string>();\n /**\n * This function contains logic of file download from S3.\n * Current we're not mocking zip file download from S3 in tests at the moment.\n * So, we're manually mocking it in case of test just by returning an empty object.\n */\n if (process.env.NODE_ENV === \"test\") {\n return {\n fileIdToKeyMap\n };\n }\n console.log(\"INSIDE uploadPageAssets\");\n\n // Save files meta data against old key for later use.\n const fileKeyToFileMap = new Map<string, FileItem>();\n // Initialize maps.\n for (let i = 0; i < filesData.length; i++) {\n const file = filesData[i];\n fileKeyToFileMap.set(file.key, file);\n\n // Initialize the value\n fileIdToKeyMap.set(file.id, file.type);\n }\n\n const fileUploadResults = await uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap: fileUploadsData.assets\n });\n\n // Create files in File Manager\n const createFilesInput = fileUploadResults\n .map((uploadResult): FileInput | null => {\n const newKey = uploadResult.Key;\n const file = fileKeyToFileMap.get(getOldFileKey(newKey));\n if (!file) {\n return null;\n }\n\n // Update the file map with newly uploaded file.\n fileIdToKeyMap.set(file.id, newKey);\n\n return {\n key: newKey,\n name: file.name,\n size: file.size,\n type: file.type,\n meta: file.meta,\n tags: file.tags\n };\n })\n .filter(Boolean) as FileInput[];\n\n const createFilesPromises = [];\n // Gives an array of chunks (each consists of FILES_COUNT_IN_EACH_BATCH items).\n const createFilesInputChunks = chunk(createFilesInput, FILES_COUNT_IN_EACH_BATCH);\n for (let i = 0; i < createFilesInputChunks.length; i++) {\n const createFilesInputChunk = createFilesInputChunks[i];\n createFilesPromises.push(\n /*\n * We need to break down files into chunks because\n * `createFilesInBatch` operation has a limit on number of files it can handle at once.\n */\n context.fileManager.files.createFilesInBatch(createFilesInputChunk)\n );\n }\n\n await Promise.all(createFilesPromises);\n\n return {\n fileIdToKeyMap\n };\n};\n\ninterface FileUploadsData {\n data: string;\n assets: Record<string, string>;\n}\n\ninterface ImportPageParams {\n key: string;\n pageKey: string;\n context: PbPageImportExportContext;\n fileUploadsData: FileUploadsData;\n}\n\nexport async function importPage({\n pageKey,\n context,\n fileUploadsData\n}: ImportPageParams): Promise<ExportedPageData[\"page\"]> {\n const log = console.log;\n\n // Making Directory for page in which we're going to extract the page data file.\n const PAGE_EXTRACT_DIR = path.join(INSTALL_EXTRACT_DIR, pageKey);\n ensureDirSync(PAGE_EXTRACT_DIR);\n\n const pageDataFileKey = dotProp.get(fileUploadsData, `data`);\n const PAGE_DATA_FILE_PATH = path.join(PAGE_EXTRACT_DIR, path.basename(pageDataFileKey));\n\n log(`Downloading Page data file: ${pageDataFileKey} at \"${PAGE_DATA_FILE_PATH}\"`);\n // Download and save page data file in disk.\n await new Promise((resolve, reject) => {\n s3Stream\n .readStream(pageDataFileKey)\n .on(\"error\", reject)\n .pipe(createWriteStream(PAGE_DATA_FILE_PATH))\n .on(\"error\", reject)\n .on(\"finish\", resolve);\n });\n\n // Load the page data file from disk.\n log(`Load file ${pageDataFileKey}`);\n const { page, files } = await loadJson<ExportedPageData>(PAGE_DATA_FILE_PATH);\n\n // Only update page data if there are files.\n if (files && Array.isArray(files) && files.length > 0) {\n // Upload page assets.\n const { fileIdToKeyMap } = await uploadPageAssets({\n context,\n /**\n * TODO @ts-refactor @ashutosh figure out correct types.\n */\n // @ts-ignore\n filesData: files,\n fileUploadsData\n });\n\n const settings = await context.fileManager.settings.getSettings();\n\n const { srcPrefix = \"\" } = settings || {};\n updateFilesInPageData({\n data: page.content || {},\n fileIdToKeyMap,\n srcPrefix\n });\n\n page.settings = updateImageInPageSettings({\n settings: page.settings || {},\n fileIdToKeyMap,\n srcPrefix\n });\n }\n\n log(\"Removing Directory for page...\");\n await deleteFile(pageKey);\n\n log(`Remove page contents from S3...`);\n await deleteS3Folder(path.dirname(fileUploadsData.data));\n\n return page;\n}\n\ninterface UploadFilesFromZipParams {\n fileKeyToFileMap: Map<string, any>;\n oldKeyToNewKeyMap: Record<string, string>;\n}\n\nasync function uploadFilesFromS3({\n fileKeyToFileMap,\n oldKeyToNewKeyMap\n}: UploadFilesFromZipParams): Promise<S3.ManagedUpload.SendData[]> {\n const oldKeysForAssets = Object.keys(oldKeyToNewKeyMap);\n\n const promises = [];\n // Upload all assets.\n for (let i = 0; i < oldKeysForAssets.length; i++) {\n const oldKey = oldKeysForAssets[i];\n const tempNewKey = oldKeyToNewKeyMap[oldKey];\n\n // Read file.\n const readStream = s3Stream.readStream(tempNewKey);\n // Get file meta data.\n const fileMetaData = fileKeyToFileMap.get(oldKey);\n\n if (fileMetaData) {\n const newKey = uniqueId(\"\", `-${fileMetaData.key}`);\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, fileMetaData.type);\n readStream.pipe(streamPassThrough);\n promises.push(promise);\n\n console.log(`Successfully queued file \"${newKey}\"`);\n }\n }\n\n return Promise.all(promises);\n}\n\nfunction getOldFileKey(key: string) {\n /*\n * Because we know the naming convention, we can extract the old key from new key.\n */\n try {\n const [, ...rest] = key.split(\"-\");\n return rest.join(\"-\");\n } catch (e) {\n return key;\n }\n}\n\nconst FILE_CONTENT_TYPE = \"application/octet-stream\";\n\nfunction getFileNameWithoutExt(fileName: string): string {\n return path.basename(fileName).replace(path.extname(fileName), \"\");\n}\n\ninterface PageImportData {\n assets: Record<string, string>;\n data: string;\n key: string;\n}\n\n/**\n * Function will read the given zip file from S3 via stream, extract its content and upload it to S3 bucket.\n * @param zipFileUrl\n * @return PageImportData S3 file keys for all uploaded assets group by page.\n */\nexport async function readExtractAndUploadZipFileContents(\n zipFileUrl: string\n): Promise<PageImportData[]> {\n const log = console.log;\n const pageImportDataList = [];\n\n const zipFileName = path.basename(zipFileUrl).split(\"?\")[0];\n\n const response = await fetch(zipFileUrl);\n if (!response.ok) {\n throw new WebinyError(`Unable to downloading file: \"${zipFileUrl}\"`, response.statusText);\n }\n\n const readStream = response.body;\n\n const uniquePath = uniqueId(\"IMPORT_PAGES/\");\n // Read export file and download it in the disk\n const ZIP_FILE_PATH = path.join(INSTALL_DIR, zipFileName);\n\n const writeStream = createWriteStream(ZIP_FILE_PATH);\n await streamPipeline(readStream, writeStream);\n log(`Downloaded file \"${zipFileName}\" at ${ZIP_FILE_PATH}`);\n\n // Extract the downloaded zip file\n const zipFilePaths = await extractZipToDisk(ZIP_FILE_PATH);\n\n log(`Removing ZIP file \"${zipFileUrl}\" from ${ZIP_FILE_PATH}`);\n await deleteFile(ZIP_FILE_PATH);\n\n // Extract each page zip and upload their content's to S3\n for (let i = 0; i < zipFilePaths.length; i++) {\n const currentPath = zipFilePaths[i];\n const dataMap = await extractZipAndUploadToS3(currentPath, uniquePath);\n pageImportDataList.push(dataMap);\n }\n log(\"Removing all ZIP files located at \", path.dirname(zipFilePaths[0]));\n await deleteFile(path.dirname(zipFilePaths[0]));\n\n return pageImportDataList;\n}\n\nconst ASSETS_DIR_NAME = \"/assets\";\n\nfunction preparePageDataDirMap({\n map,\n filePath,\n newKey\n}: {\n map: PageImportData;\n filePath: string;\n newKey: string;\n}): PageImportData {\n const dirname = path.dirname(filePath);\n const fileName = path.basename(filePath);\n /*\n * We want to use dot (.) as part of object key rather than creating nested object(s).\n * Also, the file name might contain dots in it beside the extension, so, we are escaping them all.\n */\n const oldKey = fileName.replace(/\\./g, \"\\\\.\");\n\n const isAsset = dirname.endsWith(ASSETS_DIR_NAME);\n\n if (isAsset) {\n map = dotProp.set(map, `assets.${oldKey}`, newKey);\n } else {\n // We only need to know the newKey for data file.\n map = dotProp.set(map, `data`, newKey);\n }\n\n return map;\n}\n\nasync function deleteS3Folder(key: string): Promise<void> {\n // Append trailing slash i.e \"/\" to key to make sure we only delete a specific folder.\n if (!key.endsWith(\"/\")) {\n key = `${key}/`;\n }\n\n const response = await s3Stream.listObject(key);\n const keys = (response.Contents || []).map(c => c.Key).filter(Boolean) as string[];\n console.log(`Found ${keys.length} files.`);\n\n const deleteFilePromises = keys.map(key => s3Stream.deleteObject(key));\n\n await Promise.all(deleteFilePromises);\n console.log(`Successfully deleted ${deleteFilePromises.length} files.`);\n}\n\n// export const zeroPad = version => `${version}`.padStart(5, \"0\");\n\nexport function initialStats(total: number) {\n return {\n [PageImportExportTaskStatus.PENDING]: total,\n [PageImportExportTaskStatus.PROCESSING]: 0,\n [PageImportExportTaskStatus.COMPLETED]: 0,\n [PageImportExportTaskStatus.FAILED]: 0,\n total\n };\n}\n\nfunction extractZipToDisk(exportFileZipPath: string): Promise<string[]> {\n return new Promise((resolve, reject) => {\n const pageZipFilePaths: string[] = [];\n const uniqueFolderNameForExport = getFileNameWithoutExt(exportFileZipPath);\n const EXPORT_FILE_EXTRACTION_PATH = path.join(INSTALL_DIR, uniqueFolderNameForExport);\n // Make sure DIR exists\n ensureDirSync(EXPORT_FILE_EXTRACTION_PATH);\n\n yauzl.open(exportFileZipPath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", exportFileZipPath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Missing zip file resource for path: \" + exportFileZipPath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn(\"ERROR: Failed on END event for file: \", exportFileZipPath, err);\n reject(err);\n }\n resolve(pageZipFilePaths);\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed to openReadStream for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\n \"ERROR: Missing Read Stream Resource when extracting to disk.\"\n );\n reject(\"Missing Read Stream Resource.\");\n return;\n }\n\n const filePath = path.join(EXPORT_FILE_EXTRACTION_PATH, entry.fileName);\n\n readStream.on(\"end\", function () {\n pageZipFilePaths.push(filePath);\n zipFile.readEntry();\n });\n\n streamPipeline(readStream, createWriteStream(filePath)).catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n\nfunction extractZipAndUploadToS3(\n pageDataZipFilePath: string,\n uniquePath: string\n): Promise<PageImportData> {\n return new Promise((resolve, reject) => {\n const filePaths = [];\n const fileUploadPromises: Promise<S3.ManagedUpload.SendData>[] = [];\n const uniquePageKey = getFileNameWithoutExt(pageDataZipFilePath);\n let dataMap: PageImportData = {\n key: uniquePageKey,\n assets: {},\n data: \"\"\n };\n yauzl.open(pageDataZipFilePath, { lazyEntries: true }, function (err, zipFile) {\n if (err) {\n console.warn(\"ERROR: Failed to extract zip: \", pageDataZipFilePath, err);\n reject(err);\n return;\n }\n if (!zipFile) {\n console.log(\"ERROR: Probably failed to extract zip: \" + pageDataZipFilePath);\n reject(\"Missing Zip File Resource.\");\n return;\n }\n console.info(`The ZIP file contains ${zipFile.entryCount} entries.`);\n zipFile.on(\"end\", function (err) {\n if (err) {\n console.warn('ERROR: Failed on \"END\" for file: ', pageDataZipFilePath, err);\n reject(err);\n }\n\n Promise.all(fileUploadPromises).then(res => {\n res.forEach(r => {\n console.info(\"Done uploading... \", r);\n });\n resolve(dataMap);\n });\n });\n\n zipFile.readEntry();\n\n zipFile.on(\"entry\", function (entry) {\n console.info(`Processing entry: \"${entry.fileName}\"`);\n if (/\\/$/.test(entry.fileName)) {\n // Directory file names end with '/'.\n // Note that entries for directories themselves are optional.\n // An entry's fileName implicitly requires its parent directories to exist.\n zipFile.readEntry();\n } else {\n // file entry\n zipFile.openReadStream(entry, function (err, readStream) {\n if (err) {\n console.warn(\n \"ERROR: Failed while performing [openReadStream] for file: \",\n entry.fileName,\n err\n );\n reject(err);\n return;\n }\n if (!readStream) {\n console.log(\"ERROR: Missing Read Stream while importing pages.\");\n reject(\"Missing Read Strea Resource.\");\n return;\n }\n readStream.on(\"end\", function () {\n filePaths.push(entry.fileName);\n zipFile.readEntry();\n });\n\n const newKey = `${uniquePath}/${uniquePageKey}/${entry.fileName}`;\n // Modify in place\n dataMap = preparePageDataDirMap({\n map: dataMap,\n filePath: entry.fileName,\n newKey\n });\n\n const { streamPassThrough, streamPassThroughUploadPromise: promise } =\n s3Stream.writeStream(newKey, FILE_CONTENT_TYPE);\n\n streamPipeline(readStream, streamPassThrough)\n .then(() => {\n fileUploadPromises.push(promise);\n })\n .catch(error => {\n reject(error);\n });\n });\n }\n });\n });\n });\n}\n"],"mappings":";;;;;;;;;;;;AAAA;;AAEA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AACA;;AAEA;;AACA;;AACA;;AAEA;;AAaA,MAAMA,cAAc,GAAG,IAAAC,eAAA,EAAUC,gBAAV,CAAvB;AAEA,MAAMC,WAAW,GAAG,MAApB;;AACA,MAAMC,mBAAmB,GAAGC,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuB,0BAAvB,CAA5B;;AACA,MAAMI,yBAAyB,GAAG,EAAlC;;AAcA,SAASC,yBAAT,CACIC,MADJ,EAE+C;EAC3C,MAAM;IAAEC,QAAF;IAAYC,cAAZ;IAA4BC;EAA5B,IAA0CH,MAAhD;EACA,IAAII,WAAW,GAAGH,QAAlB;EAEA,MAAMI,6BAA6B,GAAGF,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAChCH,SAAS,CAACI,KAAV,CAAgB,CAAhB,EAAmB,CAAC,CAApB,CADgC,GAEhCJ,SAFN;;EAIA,IAAIK,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,mBAAzB,CAAJ,EAAmD;IAAA;;IAC/CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,mBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,sBAAAR,QAAQ,CAACU,OAAT,iGAAkBC,KAAlB,gFAAyBC,EAAzB,KAA+B,EADC,CAElC,EALQ,CAAd;EAOH;;EACD,IAAIL,yBAAA,CAAQC,GAAR,CAAYL,WAAZ,EAAyB,kBAAzB,CAAJ,EAAkD;IAAA;;IAC9CA,WAAW,GAAGI,yBAAA,CAAQE,GAAR,CACVN,WADU,EAEV,kBAFU,EAGT,GAAEC,6BAA8B,IAAGH,cAAc,CAACO,GAAf,CAChC,qBAAAR,QAAQ,CAACa,MAAT,+FAAiBF,KAAjB,gFAAwBC,EAAxB,KAA8B,EADE,CAElC,EALQ,CAAd;EAOH;;EAED,OAAOT,WAAP;AACH;;AAED,SAASW,qBAAT,CAA+B;EAAEC,IAAF;EAAQd,cAAR;EAAwBC;AAAxB,CAA/B,EAAiG;EAC7F;EACA,IAAI,CAACa,IAAD,IAAS,OAAOA,IAAP,KAAgB,QAA7B,EAAuC;IACnC;EACH,CAJ4F,CAK7F;;;EACA,IAAIC,KAAK,CAACC,OAAN,CAAcF,IAAd,CAAJ,EAAyB;IACrB,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGH,IAAI,CAACI,MAAzB,EAAiCD,CAAC,EAAlC,EAAsC;MAClC,MAAME,OAAO,GAAGL,IAAI,CAACG,CAAD,CAApB;MACAJ,qBAAqB,CAAC;QAAEC,IAAI,EAAEK,OAAR;QAAiBnB,cAAjB;QAAiCC;MAAjC,CAAD,CAArB;IACH;;IACD;EACH,CAZ4F,CAa7F;;;EACA,MAAMmB,KAAK,GAAGC,MAAM,CAACC,OAAP,CAAeR,IAAf,CAAd;;EACA,KAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGG,KAAK,CAACF,MAA1B,EAAkCD,CAAC,EAAnC,EAAuC;IACnC,MAAM,CAACM,GAAD,EAAMC,KAAN,IAAeJ,KAAK,CAACH,CAAD,CAA1B;;IAEA,IAAIM,GAAG,KAAK,MAAR,IAAkBC,KAAlB,IAA2BxB,cAAc,CAACyB,GAAf,CAAmBD,KAAK,CAACb,EAAzB,CAA/B,EAA6D;MACzDa,KAAK,CAACD,GAAN,GAAYvB,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAZ;MACAa,KAAK,CAACE,IAAN,GAAa1B,cAAc,CAACO,GAAf,CAAmBiB,KAAK,CAACb,EAAzB,CAAb;MACAa,KAAK,CAACG,GAAN,GAAa,GAAE1B,SAAU,GAAEA,SAAS,CAACG,QAAV,CAAmB,GAAnB,IAA0B,EAA1B,GAA+B,GAAI,GAAEJ,cAAc,CAACO,GAAf,CAC5DiB,KAAK,CAACb,EADsD,CAE9D,EAFF;IAGH,CAND,MAMO;MACHE,qBAAqB,CAAC;QAAEC,IAAI,EAAEU,KAAR;QAAevB,SAAf;QAA0BD;MAA1B,CAAD,CAArB;IACH;EACJ;AACJ;;AAYM,MAAM4B,gBAAgB,GAAG,MAC5B9B,MAD4B,IAEU;EACtC,MAAM;IAAE+B,OAAF;IAAWC,SAAX;IAAsBC;EAAtB,IAA0CjC,MAAhD,CADsC,CAEtC;;EACA,MAAME,cAAc,GAAG,IAAIgC,GAAJ,EAAvB;EACA;AACJ;AACA;AACA;AACA;;EACI,IAAIC,OAAO,CAACC,GAAR,CAAYC,QAAZ,KAAyB,MAA7B,EAAqC;IACjC,OAAO;MACHnC;IADG,CAAP;EAGH;;EACDoC,OAAO,CAACC,GAAR,CAAY,yBAAZ,EAdsC,CAgBtC;;EACA,MAAMC,gBAAgB,GAAG,IAAIN,GAAJ,EAAzB,CAjBsC,CAkBtC;;EACA,KAAK,IAAIf,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGa,SAAS,CAACZ,MAA9B,EAAsCD,CAAC,EAAvC,EAA2C;IACvC,MAAMsB,IAAI,GAAGT,SAAS,CAACb,CAAD,CAAtB;IACAqB,gBAAgB,CAAC9B,GAAjB,CAAqB+B,IAAI,CAAChB,GAA1B,EAA+BgB,IAA/B,EAFuC,CAIvC;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4B4B,IAAI,CAACC,IAAjC;EACH;;EAED,MAAMC,iBAAiB,GAAG,MAAMC,iBAAiB,CAAC;IAC9CJ,gBAD8C;IAE9CK,iBAAiB,EAAEZ,eAAe,CAACa;EAFW,CAAD,CAAjD,CA3BsC,CAgCtC;;EACA,MAAMC,gBAAgB,GAAGJ,iBAAiB,CACrCK,GADoB,CACfC,YAAD,IAAoC;IACrC,MAAMC,MAAM,GAAGD,YAAY,CAACE,GAA5B;IACA,MAAMV,IAAI,GAAGD,gBAAgB,CAAC/B,GAAjB,CAAqB2C,aAAa,CAACF,MAAD,CAAlC,CAAb;;IACA,IAAI,CAACT,IAAL,EAAW;MACP,OAAO,IAAP;IACH,CALoC,CAOrC;;;IACAvC,cAAc,CAACQ,GAAf,CAAmB+B,IAAI,CAAC5B,EAAxB,EAA4BqC,MAA5B;IAEA,OAAO;MACHzB,GAAG,EAAEyB,MADF;MAEHtB,IAAI,EAAEa,IAAI,CAACb,IAFR;MAGHyB,IAAI,EAAEZ,IAAI,CAACY,IAHR;MAIHX,IAAI,EAAED,IAAI,CAACC,IAJR;MAKHY,IAAI,EAAEb,IAAI,CAACa,IALR;MAMHC,IAAI,EAAEd,IAAI,CAACc;IANR,CAAP;EAQH,CAnBoB,EAoBpBC,MApBoB,CAoBbC,OApBa,CAAzB;EAsBA,MAAMC,mBAAmB,GAAG,EAA5B,CAvDsC,CAwDtC;;EACA,MAAMC,sBAAsB,GAAG,IAAAC,cAAA,EAAMb,gBAAN,EAAwBjD,yBAAxB,CAA/B;;EACA,KAAK,IAAIqB,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGwC,sBAAsB,CAACvC,MAA3C,EAAmDD,CAAC,EAApD,EAAwD;IACpD,MAAM0C,qBAAqB,GAAGF,sBAAsB,CAACxC,CAAD,CAApD;IACAuC,mBAAmB,CAACI,IAApB;IACI;AACZ;AACA;AACA;IACY/B,OAAO,CAACgC,WAAR,CAAoBC,KAApB,CAA0BC,kBAA1B,CAA6CJ,qBAA7C,CALJ;EAOH;;EAED,MAAMK,OAAO,CAACC,GAAR,CAAYT,mBAAZ,CAAN;EAEA,OAAO;IACHxD;EADG,CAAP;AAGH,CA5EM;;;;AA0FA,eAAekE,UAAf,CAA0B;EAC7BC,OAD6B;EAE7BtC,OAF6B;EAG7BE;AAH6B,CAA1B,EAIiD;EACpD,MAAMM,GAAG,GAAGD,OAAO,CAACC,GAApB,CADoD,CAGpD;;EACA,MAAM+B,gBAAgB,GAAG1E,aAAA,CAAKC,IAAL,CAAUF,mBAAV,EAA+B0E,OAA/B,CAAzB;;EACA,IAAAE,sBAAA,EAAcD,gBAAd;;EAEA,MAAME,eAAe,GAAGhE,yBAAA,CAAQC,GAAR,CAAYwB,eAAZ,EAA8B,MAA9B,CAAxB;;EACA,MAAMwC,mBAAmB,GAAG7E,aAAA,CAAKC,IAAL,CAAUyE,gBAAV,EAA4B1E,aAAA,CAAK8E,QAAL,CAAcF,eAAd,CAA5B,CAA5B;;EAEAjC,GAAG,CAAE,+BAA8BiC,eAAgB,QAAOC,mBAAoB,GAA3E,CAAH,CAVoD,CAWpD;;EACA,MAAM,IAAIP,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACnCC,kBAAA,CACKC,UADL,CACgBN,eADhB,EAEKO,EAFL,CAEQ,OAFR,EAEiBH,MAFjB,EAGKI,IAHL,CAGU,IAAAC,qBAAA,EAAkBR,mBAAlB,CAHV,EAIKM,EAJL,CAIQ,OAJR,EAIiBH,MAJjB,EAKKG,EALL,CAKQ,QALR,EAKkBJ,OALlB;EAMH,CAPK,CAAN,CAZoD,CAqBpD;;EACApC,GAAG,CAAE,aAAYiC,eAAgB,EAA9B,CAAH;EACA,MAAM;IAAEU,IAAF;IAAQlB;EAAR,IAAkB,MAAM,IAAAmB,qBAAA,EAA2BV,mBAA3B,CAA9B,CAvBoD,CAyBpD;;EACA,IAAIT,KAAK,IAAI/C,KAAK,CAACC,OAAN,CAAc8C,KAAd,CAAT,IAAiCA,KAAK,CAAC5C,MAAN,GAAe,CAApD,EAAuD;IACnD;IACA,MAAM;MAAElB;IAAF,IAAqB,MAAM4B,gBAAgB,CAAC;MAC9CC,OAD8C;;MAE9C;AACZ;AACA;MACY;MACAC,SAAS,EAAEgC,KANmC;MAO9C/B;IAP8C,CAAD,CAAjD;IAUA,MAAMhC,QAAQ,GAAG,MAAM8B,OAAO,CAACgC,WAAR,CAAoB9D,QAApB,CAA6BmF,WAA7B,EAAvB;IAEA,MAAM;MAAEjF,SAAS,GAAG;IAAd,IAAqBF,QAAQ,IAAI,EAAvC;IACAc,qBAAqB,CAAC;MAClBC,IAAI,EAAEkE,IAAI,CAACG,OAAL,IAAgB,EADJ;MAElBnF,cAFkB;MAGlBC;IAHkB,CAAD,CAArB;IAMA+E,IAAI,CAACjF,QAAL,GAAgBF,yBAAyB,CAAC;MACtCE,QAAQ,EAAEiF,IAAI,CAACjF,QAAL,IAAiB,EADW;MAEtCC,cAFsC;MAGtCC;IAHsC,CAAD,CAAzC;EAKH;;EAEDoC,GAAG,CAAC,gCAAD,CAAH;EACA,MAAM,IAAA+C,gCAAA,EAAWjB,OAAX,CAAN;EAEA9B,GAAG,CAAE,iCAAF,CAAH;EACA,MAAMgD,cAAc,CAAC3F,aAAA,CAAK4F,OAAL,CAAavD,eAAe,CAACjB,IAA7B,CAAD,CAApB;EAEA,OAAOkE,IAAP;AACH;;AAOD,eAAetC,iBAAf,CAAiC;EAC7BJ,gBAD6B;EAE7BK;AAF6B,CAAjC,EAGmE;EAC/D,MAAM4C,gBAAgB,GAAGlE,MAAM,CAACmE,IAAP,CAAY7C,iBAAZ,CAAzB;EAEA,MAAM8C,QAAQ,GAAG,EAAjB,CAH+D,CAI/D;;EACA,KAAK,IAAIxE,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGsE,gBAAgB,CAACrE,MAArC,EAA6CD,CAAC,EAA9C,EAAkD;IAC9C,MAAMyE,MAAM,GAAGH,gBAAgB,CAACtE,CAAD,CAA/B;IACA,MAAM0E,UAAU,GAAGhD,iBAAiB,CAAC+C,MAAD,CAApC,CAF8C,CAI9C;;IACA,MAAMd,UAAU,GAAGD,kBAAA,CAASC,UAAT,CAAoBe,UAApB,CAAnB,CAL8C,CAM9C;;;IACA,MAAMC,YAAY,GAAGtD,gBAAgB,CAAC/B,GAAjB,CAAqBmF,MAArB,CAArB;;IAEA,IAAIE,YAAJ,EAAkB;MACd,MAAM5C,MAAM,GAAG,IAAA6C,eAAA,EAAS,EAAT,EAAc,IAAGD,YAAY,CAACrE,GAAI,EAAlC,CAAf;;MACA,MAAM;QAAEuE,iBAAF;QAAqBC,8BAA8B,EAAEC;MAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6B4C,YAAY,CAACpD,IAA1C,CADJ;;MAEAoC,UAAU,CAACE,IAAX,CAAgBgB,iBAAhB;MACAL,QAAQ,CAAC7B,IAAT,CAAcoC,OAAd;MAEA5D,OAAO,CAACC,GAAR,CAAa,6BAA4BW,MAAO,GAAhD;IACH;EACJ;;EAED,OAAOgB,OAAO,CAACC,GAAR,CAAYwB,QAAZ,CAAP;AACH;;AAED,SAASvC,aAAT,CAAuB3B,GAAvB,EAAoC;EAChC;AACJ;AACA;EACI,IAAI;IACA,MAAM,GAAG,GAAG2E,IAAN,IAAc3E,GAAG,CAAC4E,KAAJ,CAAU,GAAV,CAApB;IACA,OAAOD,IAAI,CAACvG,IAAL,CAAU,GAAV,CAAP;EACH,CAHD,CAGE,OAAOyG,CAAP,EAAU;IACR,OAAO7E,GAAP;EACH;AACJ;;AAED,MAAM8E,iBAAiB,GAAG,0BAA1B;;AAEA,SAASC,qBAAT,CAA+BC,QAA/B,EAAyD;EACrD,OAAO7G,aAAA,CAAK8E,QAAL,CAAc+B,QAAd,EAAwBC,OAAxB,CAAgC9G,aAAA,CAAK+G,OAAL,CAAaF,QAAb,CAAhC,EAAwD,EAAxD,CAAP;AACH;;AAQD;AACA;AACA;AACA;AACA;AACO,eAAeG,mCAAf,CACHC,UADG,EAEsB;EACzB,MAAMtE,GAAG,GAAGD,OAAO,CAACC,GAApB;EACA,MAAMuE,kBAAkB,GAAG,EAA3B;;EAEA,MAAMC,WAAW,GAAGnH,aAAA,CAAK8E,QAAL,CAAcmC,UAAd,EAA0BR,KAA1B,CAAgC,GAAhC,EAAqC,CAArC,CAApB;;EAEA,MAAMW,QAAQ,GAAG,MAAM,IAAAC,kBAAA,EAAMJ,UAAN,CAAvB;;EACA,IAAI,CAACG,QAAQ,CAACE,EAAd,EAAkB;IACd,MAAM,IAAIC,cAAJ,CAAiB,gCAA+BN,UAAW,GAA3D,EAA+DG,QAAQ,CAACI,UAAxE,CAAN;EACH;;EAED,MAAMtC,UAAU,GAAGkC,QAAQ,CAACK,IAA5B;EAEA,MAAMC,UAAU,GAAG,IAAAvB,eAAA,EAAS,eAAT,CAAnB,CAbyB,CAczB;;EACA,MAAMwB,aAAa,GAAG3H,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBqH,WAAvB,CAAtB;;EAEA,MAAMZ,WAAW,GAAG,IAAAlB,qBAAA,EAAkBsC,aAAlB,CAApB;EACA,MAAMhI,cAAc,CAACuF,UAAD,EAAaqB,WAAb,CAApB;EACA5D,GAAG,CAAE,oBAAmBwE,WAAY,QAAOQ,aAAc,EAAtD,CAAH,CAnByB,CAqBzB;;EACA,MAAMC,YAAY,GAAG,MAAMC,gBAAgB,CAACF,aAAD,CAA3C;EAEAhF,GAAG,CAAE,sBAAqBsE,UAAW,UAASU,aAAc,EAAzD,CAAH;EACA,MAAM,IAAAjC,gCAAA,EAAWiC,aAAX,CAAN,CAzByB,CA2BzB;;EACA,KAAK,IAAIpG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGqG,YAAY,CAACpG,MAAjC,EAAyCD,CAAC,EAA1C,EAA8C;IAC1C,MAAMuG,WAAW,GAAGF,YAAY,CAACrG,CAAD,CAAhC;IACA,MAAMwG,OAAO,GAAG,MAAMC,uBAAuB,CAACF,WAAD,EAAcJ,UAAd,CAA7C;IACAR,kBAAkB,CAAChD,IAAnB,CAAwB6D,OAAxB;EACH;;EACDpF,GAAG,CAAC,oCAAD,EAAuC3C,aAAA,CAAK4F,OAAL,CAAagC,YAAY,CAAC,CAAD,CAAzB,CAAvC,CAAH;EACA,MAAM,IAAAlC,gCAAA,EAAW1F,aAAA,CAAK4F,OAAL,CAAagC,YAAY,CAAC,CAAD,CAAzB,CAAX,CAAN;EAEA,OAAOV,kBAAP;AACH;;AAED,MAAMe,eAAe,GAAG,SAAxB;;AAEA,SAASC,qBAAT,CAA+B;EAC3B9E,GAD2B;EAE3B+E,QAF2B;EAG3B7E;AAH2B,CAA/B,EAQmB;EACf,MAAMsC,OAAO,GAAG5F,aAAA,CAAK4F,OAAL,CAAauC,QAAb,CAAhB;;EACA,MAAMtB,QAAQ,GAAG7G,aAAA,CAAK8E,QAAL,CAAcqD,QAAd,CAAjB;EACA;AACJ;AACA;AACA;;;EACI,MAAMnC,MAAM,GAAGa,QAAQ,CAACC,OAAT,CAAiB,KAAjB,EAAwB,KAAxB,CAAf;EAEA,MAAMsB,OAAO,GAAGxC,OAAO,CAAClF,QAAR,CAAiBuH,eAAjB,CAAhB;;EAEA,IAAIG,OAAJ,EAAa;IACThF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,UAAS4C,MAAO,EAAlC,EAAqC1C,MAArC,CAAN;EACH,CAFD,MAEO;IACH;IACAF,GAAG,GAAGxC,yBAAA,CAAQE,GAAR,CAAYsC,GAAZ,EAAkB,MAAlB,EAAyBE,MAAzB,CAAN;EACH;;EAED,OAAOF,GAAP;AACH;;AAED,eAAeuC,cAAf,CAA8B9D,GAA9B,EAA0D;EACtD;EACA,IAAI,CAACA,GAAG,CAACnB,QAAJ,CAAa,GAAb,CAAL,EAAwB;IACpBmB,GAAG,GAAI,GAAEA,GAAI,GAAb;EACH;;EAED,MAAMuF,QAAQ,GAAG,MAAMnC,kBAAA,CAASoD,UAAT,CAAoBxG,GAApB,CAAvB;EACA,MAAMiE,IAAI,GAAG,CAACsB,QAAQ,CAACkB,QAAT,IAAqB,EAAtB,EAA0BlF,GAA1B,CAA8BmF,CAAC,IAAIA,CAAC,CAAChF,GAArC,EAA0CK,MAA1C,CAAiDC,OAAjD,CAAb;EACAnB,OAAO,CAACC,GAAR,CAAa,SAAQmD,IAAI,CAACtE,MAAO,SAAjC;EAEA,MAAMgH,kBAAkB,GAAG1C,IAAI,CAAC1C,GAAL,CAASvB,GAAG,IAAIoD,kBAAA,CAASwD,YAAT,CAAsB5G,GAAtB,CAAhB,CAA3B;EAEA,MAAMyC,OAAO,CAACC,GAAR,CAAYiE,kBAAZ,CAAN;EACA9F,OAAO,CAACC,GAAR,CAAa,wBAAuB6F,kBAAkB,CAAChH,MAAO,SAA9D;AACH,C,CAED;;;AAEO,SAASkH,YAAT,CAAsBC,KAAtB,EAAqC;EACxC,OAAO;IACH,CAACC,iCAAA,CAA2BC,OAA5B,GAAsCF,KADnC;IAEH,CAACC,iCAAA,CAA2BE,UAA5B,GAAyC,CAFtC;IAGH,CAACF,iCAAA,CAA2BG,SAA5B,GAAwC,CAHrC;IAIH,CAACH,iCAAA,CAA2BI,MAA5B,GAAqC,CAJlC;IAKHL;EALG,CAAP;AAOH;;AAED,SAASd,gBAAT,CAA0BoB,iBAA1B,EAAwE;EACpE,OAAO,IAAI3E,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMkE,gBAA0B,GAAG,EAAnC;IACA,MAAMC,yBAAyB,GAAGvC,qBAAqB,CAACqC,iBAAD,CAAvD;;IACA,MAAMG,2BAA2B,GAAGpJ,aAAA,CAAKC,IAAL,CAAUH,WAAV,EAAuBqJ,yBAAvB,CAApC,CAHoC,CAIpC;;;IACA,IAAAxE,sBAAA,EAAcyE,2BAAd;;IAEAC,cAAA,CAAMC,IAAN,CAAWL,iBAAX,EAA8B;MAAEM,WAAW,EAAE;IAAf,CAA9B,EAAqD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MACzE,IAAID,GAAJ,EAAS;QACL9G,OAAO,CAACgH,IAAR,CAAa,gCAAb,EAA+CT,iBAA/C,EAAkEO,GAAlE;QACAxE,MAAM,CAACwE,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACV/G,OAAO,CAACC,GAAR,CAAY,gDAAgDsG,iBAA5D;QACAjE,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MAEDtC,OAAO,CAACiH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MAEAH,OAAO,CAACtE,EAAR,CAAW,KAAX,EAAkB,UAAUqE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACL9G,OAAO,CAACgH,IAAR,CAAa,uCAAb,EAAsDT,iBAAtD,EAAyEO,GAAzE;UACAxE,MAAM,CAACwE,GAAD,CAAN;QACH;;QACDzE,OAAO,CAACmE,gBAAD,CAAP;MACH,CAND;MAQAO,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAACtE,EAAR,CAAW,OAAX,EAAoB,UAAU2E,KAAV,EAAiB;QACjCpH,OAAO,CAACiH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAetE,UAAf,EAA2B;YACrD,IAAIsE,GAAJ,EAAS;cACL9G,OAAO,CAACgH,IAAR,CACI,4CADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKAxE,MAAM,CAACwE,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAACtE,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CACI,8DADJ;cAGAqC,MAAM,CAAC,+BAAD,CAAN;cACA;YACH;;YAED,MAAMmD,QAAQ,GAAGnI,aAAA,CAAKC,IAAL,CAAUmJ,2BAAV,EAAuCU,KAAK,CAACjD,QAA7C,CAAjB;;YAEA3B,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7B+D,gBAAgB,CAAChF,IAAjB,CAAsBiE,QAAtB;cACAsB,OAAO,CAACI,SAAR;YACH,CAHD;YAKAlK,cAAc,CAACuF,UAAD,EAAa,IAAAG,qBAAA,EAAkB8C,QAAlB,CAAb,CAAd,CAAwD8B,KAAxD,CAA8DC,KAAK,IAAI;cACnElF,MAAM,CAACkF,KAAD,CAAN;YACH,CAFD;UAGH,CA5BD;QA6BH;MACJ,CAvCD;IAwCH,CAhED;EAiEH,CAxEM,CAAP;AAyEH;;AAED,SAASlC,uBAAT,CACImC,mBADJ,EAEIzC,UAFJ,EAG2B;EACvB,OAAO,IAAIpD,OAAJ,CAAY,CAACS,OAAD,EAAUC,MAAV,KAAqB;IACpC,MAAMoF,SAAS,GAAG,EAAlB;IACA,MAAMC,kBAAwD,GAAG,EAAjE;IACA,MAAMC,aAAa,GAAG1D,qBAAqB,CAACuD,mBAAD,CAA3C;IACA,IAAIpC,OAAuB,GAAG;MAC1BlG,GAAG,EAAEyI,aADqB;MAE1BpH,MAAM,EAAE,EAFkB;MAG1B9B,IAAI,EAAE;IAHoB,CAA9B;;IAKAiI,cAAA,CAAMC,IAAN,CAAWa,mBAAX,EAAgC;MAAEZ,WAAW,EAAE;IAAf,CAAhC,EAAuD,UAAUC,GAAV,EAAeC,OAAf,EAAwB;MAC3E,IAAID,GAAJ,EAAS;QACL9G,OAAO,CAACgH,IAAR,CAAa,gCAAb,EAA+CS,mBAA/C,EAAoEX,GAApE;QACAxE,MAAM,CAACwE,GAAD,CAAN;QACA;MACH;;MACD,IAAI,CAACC,OAAL,EAAc;QACV/G,OAAO,CAACC,GAAR,CAAY,4CAA4CwH,mBAAxD;QACAnF,MAAM,CAAC,4BAAD,CAAN;QACA;MACH;;MACDtC,OAAO,CAACiH,IAAR,CAAc,yBAAwBF,OAAO,CAACG,UAAW,WAAzD;MACAH,OAAO,CAACtE,EAAR,CAAW,KAAX,EAAkB,UAAUqE,GAAV,EAAe;QAC7B,IAAIA,GAAJ,EAAS;UACL9G,OAAO,CAACgH,IAAR,CAAa,mCAAb,EAAkDS,mBAAlD,EAAuEX,GAAvE;UACAxE,MAAM,CAACwE,GAAD,CAAN;QACH;;QAEDlF,OAAO,CAACC,GAAR,CAAY8F,kBAAZ,EAAgCE,IAAhC,CAAqCC,GAAG,IAAI;UACxCA,GAAG,CAACC,OAAJ,CAAYC,CAAC,IAAI;YACbhI,OAAO,CAACiH,IAAR,CAAa,oBAAb,EAAmCe,CAAnC;UACH,CAFD;UAGA3F,OAAO,CAACgD,OAAD,CAAP;QACH,CALD;MAMH,CAZD;MAcA0B,OAAO,CAACI,SAAR;MAEAJ,OAAO,CAACtE,EAAR,CAAW,OAAX,EAAoB,UAAU2E,KAAV,EAAiB;QACjCpH,OAAO,CAACiH,IAAR,CAAc,sBAAqBG,KAAK,CAACjD,QAAS,GAAlD;;QACA,IAAI,MAAMkD,IAAN,CAAWD,KAAK,CAACjD,QAAjB,CAAJ,EAAgC;UAC5B;UACA;UACA;UACA4C,OAAO,CAACI,SAAR;QACH,CALD,MAKO;UACH;UACAJ,OAAO,CAACO,cAAR,CAAuBF,KAAvB,EAA8B,UAAUN,GAAV,EAAetE,UAAf,EAA2B;YACrD,IAAIsE,GAAJ,EAAS;cACL9G,OAAO,CAACgH,IAAR,CACI,4DADJ,EAEII,KAAK,CAACjD,QAFV,EAGI2C,GAHJ;cAKAxE,MAAM,CAACwE,GAAD,CAAN;cACA;YACH;;YACD,IAAI,CAACtE,UAAL,EAAiB;cACbxC,OAAO,CAACC,GAAR,CAAY,mDAAZ;cACAqC,MAAM,CAAC,8BAAD,CAAN;cACA;YACH;;YACDE,UAAU,CAACC,EAAX,CAAc,KAAd,EAAqB,YAAY;cAC7BiF,SAAS,CAAClG,IAAV,CAAe4F,KAAK,CAACjD,QAArB;cACA4C,OAAO,CAACI,SAAR;YACH,CAHD;YAKA,MAAMvG,MAAM,GAAI,GAAEoE,UAAW,IAAG4C,aAAc,IAAGR,KAAK,CAACjD,QAAS,EAAhE,CApBqD,CAqBrD;;YACAkB,OAAO,GAAGG,qBAAqB,CAAC;cAC5B9E,GAAG,EAAE2E,OADuB;cAE5BI,QAAQ,EAAE2B,KAAK,CAACjD,QAFY;cAG5BvD;YAH4B,CAAD,CAA/B;;YAMA,MAAM;cAAE8C,iBAAF;cAAqBC,8BAA8B,EAAEC;YAArD,IACFrB,kBAAA,CAASsB,WAAT,CAAqBjD,MAArB,EAA6BqD,iBAA7B,CADJ;;YAGAhH,cAAc,CAACuF,UAAD,EAAakB,iBAAb,CAAd,CACKmE,IADL,CACU,MAAM;cACRF,kBAAkB,CAACnG,IAAnB,CAAwBoC,OAAxB;YACH,CAHL,EAIK2D,KAJL,CAIWC,KAAK,IAAI;cACZlF,MAAM,CAACkF,KAAD,CAAN;YACH,CANL;UAOH,CAtCD;QAuCH;MACJ,CAjDD;IAkDH,CA9ED;EA+EH,CAxFM,CAAP;AAyFH"}
|
package/package.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1
1
|
{
|
2
2
|
"name": "@webiny/api-page-builder-import-export",
|
3
|
-
"version": "5.
|
3
|
+
"version": "5.33.0-beta.0",
|
4
4
|
"main": "index.js",
|
5
5
|
"keywords": [
|
6
6
|
"pbie:base"
|
@@ -16,16 +16,16 @@
|
|
16
16
|
"dependencies": {
|
17
17
|
"@babel/runtime": "7.18.9",
|
18
18
|
"@commodo/fields": "1.1.2-beta.20",
|
19
|
-
"@webiny/api": "5.
|
20
|
-
"@webiny/api-file-manager": "5.
|
21
|
-
"@webiny/api-page-builder": "5.
|
22
|
-
"@webiny/api-security": "5.
|
23
|
-
"@webiny/error": "5.
|
24
|
-
"@webiny/handler": "5.
|
25
|
-
"@webiny/handler-aws": "5.
|
26
|
-
"@webiny/handler-graphql": "5.
|
27
|
-
"@webiny/utils": "5.
|
28
|
-
"@webiny/validation": "5.
|
19
|
+
"@webiny/api": "5.33.0-beta.0",
|
20
|
+
"@webiny/api-file-manager": "5.33.0-beta.0",
|
21
|
+
"@webiny/api-page-builder": "5.33.0-beta.0",
|
22
|
+
"@webiny/api-security": "5.33.0-beta.0",
|
23
|
+
"@webiny/error": "5.33.0-beta.0",
|
24
|
+
"@webiny/handler": "5.33.0-beta.0",
|
25
|
+
"@webiny/handler-aws": "5.33.0-beta.0",
|
26
|
+
"@webiny/handler-graphql": "5.33.0-beta.0",
|
27
|
+
"@webiny/utils": "5.33.0-beta.0",
|
28
|
+
"@webiny/validation": "5.33.0-beta.0",
|
29
29
|
"archiver": "5.3.1",
|
30
30
|
"commodo-fields-object": "1.0.6",
|
31
31
|
"dot-prop-immutable": "2.1.1",
|
@@ -47,16 +47,16 @@
|
|
47
47
|
"@types/archiver": "^5.3.1",
|
48
48
|
"@types/node-fetch": "^2.6.1",
|
49
49
|
"@types/yauzl": "^2.9.2",
|
50
|
-
"@webiny/api-dynamodb-to-elasticsearch": "^5.
|
51
|
-
"@webiny/api-file-manager-ddb-es": "^5.
|
52
|
-
"@webiny/api-i18n-ddb": "^5.
|
53
|
-
"@webiny/api-security-so-ddb": "^5.
|
54
|
-
"@webiny/api-tenancy": "^5.
|
55
|
-
"@webiny/api-tenancy-so-ddb": "^5.
|
56
|
-
"@webiny/api-wcp": "^5.
|
57
|
-
"@webiny/cli": "^5.
|
58
|
-
"@webiny/db": "^5.
|
59
|
-
"@webiny/project-utils": "^5.
|
50
|
+
"@webiny/api-dynamodb-to-elasticsearch": "^5.33.0-beta.0",
|
51
|
+
"@webiny/api-file-manager-ddb-es": "^5.33.0-beta.0",
|
52
|
+
"@webiny/api-i18n-ddb": "^5.33.0-beta.0",
|
53
|
+
"@webiny/api-security-so-ddb": "^5.33.0-beta.0",
|
54
|
+
"@webiny/api-tenancy": "^5.33.0-beta.0",
|
55
|
+
"@webiny/api-tenancy-so-ddb": "^5.33.0-beta.0",
|
56
|
+
"@webiny/api-wcp": "^5.33.0-beta.0",
|
57
|
+
"@webiny/cli": "^5.33.0-beta.0",
|
58
|
+
"@webiny/db": "^5.33.0-beta.0",
|
59
|
+
"@webiny/project-utils": "^5.33.0-beta.0",
|
60
60
|
"jest": "^28.1.0",
|
61
61
|
"jest-dynalite": "^3.2.0",
|
62
62
|
"rimraf": "^3.0.2",
|
@@ -78,5 +78,5 @@
|
|
78
78
|
]
|
79
79
|
}
|
80
80
|
},
|
81
|
-
"gitHead": "
|
81
|
+
"gitHead": "8809a70796e5117e92416d6859c1f5dcae2a8c98"
|
82
82
|
}
|