@griddo/cx 11.9.7-rc.2 → 11.9.8-rc.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +78 -193
- package/build/commands/end-render.js +31 -0
- package/build/commands/end-render.js.map +7 -0
- package/build/commands/prepare-assets-directory.js +9 -0
- package/build/commands/prepare-assets-directory.js.map +7 -0
- package/build/commands/prepare-domains-render.js +38 -0
- package/build/commands/prepare-domains-render.js.map +7 -0
- package/build/commands/reset-render.js +31 -0
- package/build/commands/reset-render.js.map +7 -0
- package/build/commands/start-embeddings.js +31 -0
- package/build/commands/start-embeddings.js.map +7 -0
- package/build/commands/start-render.js +66 -0
- package/build/commands/start-render.js.map +7 -0
- package/build/commands/upload-search-content.js +31 -0
- package/build/commands/upload-search-content.js.map +7 -0
- package/build/core/GriddoLog.d.ts +16 -0
- package/build/core/db.d.ts +4 -0
- package/build/core/dist-rollback.d.ts +2 -0
- package/build/{errors/index.d.ts → core/errors.d.ts} +5 -4
- package/build/core/fs.d.ts +69 -0
- package/build/core/logger.d.ts +18 -0
- package/build/index.d.ts +10 -29
- package/build/index.js +406 -73
- package/build/services/auth.d.ts +2 -5
- package/build/services/manage-store.d.ts +48 -0
- package/build/services/render.d.ts +70 -0
- package/build/shared/envs.d.ts +19 -0
- package/build/{errors/errors-data.d.ts → shared/errors.d.ts} +5 -3
- package/build/shared/npm-modules/brush.d.ts +18 -0
- package/build/shared/npm-modules/find-up-simple.d.ts +34 -0
- package/build/shared/npm-modules/pkg-dir.d.ts +7 -0
- package/build/{types → shared/types}/api.d.ts +18 -18
- package/build/{types → shared/types}/global.d.ts +15 -16
- package/build/{types → shared/types}/navigation.d.ts +5 -5
- package/build/{types → shared/types}/pages.d.ts +9 -9
- package/build/shared/types/render.d.ts +54 -0
- package/build/{types → shared/types}/sites.d.ts +18 -19
- package/cli.mjs +239 -0
- package/exporter/build-esbuild.noop +42 -0
- package/exporter/build.sh +17 -28
- package/exporter/commands/README.md +151 -0
- package/exporter/commands/end-render.ts +65 -86
- package/exporter/commands/prepare-assets-directory.ts +34 -0
- package/exporter/commands/prepare-domains-render.ts +143 -35
- package/exporter/commands/reset-render.ts +12 -7
- package/exporter/commands/single-domain-upload-search-content.noop +206 -0
- package/exporter/commands/start-embeddings.ts +29 -0
- package/exporter/commands/start-render.ts +26 -64
- package/exporter/commands/upload-search-content.ts +201 -26
- package/exporter/core/GriddoLog.ts +45 -0
- package/exporter/core/check-env-health.ts +200 -0
- package/exporter/core/db-class.ts +54 -0
- package/exporter/core/db.ts +33 -0
- package/exporter/core/dist-rollback.ts +40 -0
- package/exporter/core/errors.ts +82 -0
- package/exporter/core/fs.ts +385 -0
- package/exporter/{utils → core}/images.ts +1 -6
- package/exporter/{utils → core}/instance.ts +9 -13
- package/exporter/core/life-cycle.ts +73 -0
- package/exporter/core/logger.ts +141 -0
- package/exporter/core/objects.ts +37 -0
- package/exporter/core/print-logos.ts +21 -0
- package/exporter/index.ts +14 -56
- package/exporter/services/api.ts +306 -0
- package/exporter/services/auth.ts +8 -10
- package/exporter/services/domains.ts +23 -8
- package/exporter/services/manage-sites.ts +116 -0
- package/exporter/services/manage-store.ts +235 -0
- package/exporter/services/navigation.ts +12 -18
- package/exporter/{utils → services}/pages.ts +27 -92
- package/exporter/services/reference-fields.ts +14 -32
- package/exporter/services/render-artifacts.ts +44 -0
- package/exporter/services/render.ts +229 -0
- package/exporter/services/robots.ts +33 -61
- package/exporter/services/sitemaps.ts +129 -0
- package/exporter/services/sites.ts +40 -28
- package/exporter/services/store.ts +386 -319
- package/exporter/shared/context.ts +49 -0
- package/exporter/{constants → shared}/endpoints.ts +12 -11
- package/exporter/shared/envs.ts +62 -0
- package/exporter/{errors/errors-data.ts → shared/errors.ts} +24 -14
- package/exporter/shared/npm-modules/README.md +36 -0
- package/exporter/shared/npm-modules/brush.ts +34 -0
- package/exporter/shared/npm-modules/find-up-simple.ts +100 -0
- package/exporter/shared/npm-modules/pkg-dir.ts +17 -0
- package/exporter/shared/npm-modules/xml-parser.ts +57 -0
- package/exporter/{types → shared/types}/api.ts +40 -41
- package/exporter/{types → shared/types}/global.ts +17 -21
- package/exporter/{types → shared/types}/navigation.ts +3 -3
- package/exporter/{types → shared/types}/pages.ts +10 -11
- package/exporter/shared/types/render.ts +63 -0
- package/exporter/{types → shared/types}/sites.ts +18 -19
- package/exporter/ssg-adapters/gatsby/actions/clean.ts +26 -0
- package/exporter/ssg-adapters/gatsby/actions/close.ts +17 -0
- package/exporter/ssg-adapters/gatsby/actions/data.ts +22 -0
- package/exporter/ssg-adapters/gatsby/actions/healthCheck.ts +10 -0
- package/exporter/ssg-adapters/gatsby/actions/init.ts +12 -0
- package/exporter/ssg-adapters/gatsby/actions/logs.ts +10 -0
- package/exporter/ssg-adapters/gatsby/actions/meta.ts +13 -0
- package/exporter/ssg-adapters/gatsby/actions/prepare.ts +9 -0
- package/exporter/ssg-adapters/gatsby/actions/relocation.ts +15 -0
- package/exporter/ssg-adapters/gatsby/actions/restore.ts +21 -0
- package/exporter/ssg-adapters/gatsby/actions/ssg.ts +12 -0
- package/exporter/ssg-adapters/gatsby/actions/sync.ts +65 -0
- package/exporter/ssg-adapters/gatsby/index.ts +114 -0
- package/exporter/ssg-adapters/gatsby/shared/artifacts.ts +16 -0
- package/exporter/ssg-adapters/gatsby/shared/diff-assets.ts +128 -0
- package/exporter/ssg-adapters/gatsby/shared/extract-assets.ts +75 -0
- package/exporter/ssg-adapters/gatsby/shared/gatsby-build.ts +58 -0
- package/exporter/ssg-adapters/gatsby/shared/sync-render.ts +300 -0
- package/exporter/ssg-adapters/gatsby/shared/types.ts +35 -0
- package/exporter/ssg-adapters/gatsby/shared/utils.ts +33 -0
- package/gatsby-browser.tsx +41 -58
- package/gatsby-config.ts +10 -17
- package/gatsby-node.ts +20 -80
- package/gatsby-ssr.tsx +2 -1
- package/package.json +41 -92
- package/plugins/gatsby-plugin-svgr-loader/gatsby-node.js +55 -0
- package/plugins/gatsby-plugin-svgr-loader/package.json +8 -0
- package/react/DynamicScript/index.tsx +33 -0
- package/{exporter/react/Favicon → react/GriddoFavicon}/index.tsx +3 -9
- package/{exporter/react → react}/GriddoIntegrations/index.tsx +17 -23
- package/{exporter/react → react}/GriddoIntegrations/utils.ts +24 -12
- package/react/GriddoOpenGraph/index.tsx +39 -0
- package/src/components/Head.tsx +30 -73
- package/src/components/template.tsx +8 -30
- package/src/gatsby-node-utils.ts +76 -2
- package/src/html.tsx +2 -11
- package/src/types.ts +5 -5
- package/tsconfig.commands.json +36 -0
- package/tsconfig.exporter.json +20 -0
- package/tsconfig.json +5 -3
- package/build/adapters/gatsby/index.d.ts +0 -4
- package/build/adapters/gatsby/utils.d.ts +0 -22
- package/build/artifacts/index.d.ts +0 -6
- package/build/commands/end-render.d.ts +0 -2
- package/build/commands/move-assets.d.ts +0 -1
- package/build/commands/prepare-domains-render.d.ts +0 -1
- package/build/commands/reset-render.d.ts +0 -2
- package/build/commands/start-render.d.ts +0 -2
- package/build/commands/upload-search-content.d.ts +0 -2
- package/build/constants/envs.d.ts +0 -37
- package/build/constants/index.d.ts +0 -57
- package/build/end-render.js +0 -74
- package/build/end-render.js.map +0 -7
- package/build/index.js.map +0 -7
- package/build/prepare-domains-render.js +0 -73
- package/build/prepare-domains-render.js.map +0 -7
- package/build/react/Favicon/index.d.ts +0 -5
- package/build/react/Favicon/utils.d.ts +0 -9
- package/build/react/GriddoIntegrations/index.d.ts +0 -20
- package/build/react/GriddoIntegrations/utils.d.ts +0 -26
- package/build/react/index.d.ts +0 -3
- package/build/react/index.js +0 -3
- package/build/registers/api.d.ts +0 -9
- package/build/registers/gatsby.d.ts +0 -9
- package/build/registers/index.d.ts +0 -3
- package/build/reset-render.js +0 -74
- package/build/reset-render.js.map +0 -7
- package/build/services/domains.d.ts +0 -6
- package/build/services/navigation.d.ts +0 -50
- package/build/services/reference-fields.d.ts +0 -20
- package/build/services/register.d.ts +0 -36
- package/build/services/robots.d.ts +0 -19
- package/build/services/settings.d.ts +0 -4
- package/build/services/sites.d.ts +0 -29
- package/build/services/store.d.ts +0 -6
- package/build/start-render.js +0 -100
- package/build/start-render.js.map +0 -7
- package/build/types/templates.d.ts +0 -8
- package/build/upload-search-content.js +0 -74
- package/build/upload-search-content.js.map +0 -7
- package/build/utils/alerts.d.ts +0 -3
- package/build/utils/api.d.ts +0 -23
- package/build/utils/cache.d.ts +0 -35
- package/build/utils/core-utils.d.ts +0 -107
- package/build/utils/create-build-data.d.ts +0 -8
- package/build/utils/domains.d.ts +0 -13
- package/build/utils/folders.d.ts +0 -53
- package/build/utils/health-checks.d.ts +0 -7
- package/build/utils/images.d.ts +0 -16
- package/build/utils/loggin.d.ts +0 -51
- package/build/utils/pages.d.ts +0 -34
- package/build/utils/render.d.ts +0 -13
- package/build/utils/searches.d.ts +0 -15
- package/build/utils/sites.d.ts +0 -31
- package/build/utils/store.d.ts +0 -81
- package/cx.config.d.ts +0 -5
- package/cx.config.js +0 -36
- package/exporter/adapters/gatsby/index.ts +0 -162
- package/exporter/adapters/gatsby/utils.ts +0 -161
- package/exporter/artifacts/README.md +0 -34
- package/exporter/artifacts/index.ts +0 -33
- package/exporter/commands/move-assets.ts +0 -11
- package/exporter/constants/envs.ts +0 -94
- package/exporter/constants/index.ts +0 -129
- package/exporter/errors/index.ts +0 -40
- package/exporter/react/index.tsx +0 -11
- package/exporter/registers/api.ts +0 -14
- package/exporter/registers/gatsby.ts +0 -14
- package/exporter/registers/index.ts +0 -4
- package/exporter/services/register.ts +0 -113
- package/exporter/services/settings.ts +0 -17
- package/exporter/utils/alerts.ts +0 -29
- package/exporter/utils/api.ts +0 -243
- package/exporter/utils/cache.ts +0 -142
- package/exporter/utils/core-utils.ts +0 -458
- package/exporter/utils/create-build-data.ts +0 -17
- package/exporter/utils/domains.ts +0 -39
- package/exporter/utils/folders.ts +0 -320
- package/exporter/utils/health-checks.ts +0 -64
- package/exporter/utils/loggin.ts +0 -184
- package/exporter/utils/render.ts +0 -71
- package/exporter/utils/searches.ts +0 -156
- package/exporter/utils/sites.ts +0 -312
- package/exporter/utils/store.ts +0 -314
- package/src/README.md +0 -7
- package/start-render.js +0 -7
- /package/build/{utils → core}/instance.d.ts +0 -0
- /package/build/{constants → shared}/endpoints.d.ts +0 -0
- /package/exporter/{types → shared/types}/templates.ts +0 -0
- /package/{exporter/react/Favicon → react/GriddoFavicon}/utils.ts +0 -0
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
{
|
|
2
|
+
"version": 3,
|
|
3
|
+
"sources": ["../../exporter/shared/envs.ts", "../../exporter/shared/npm-modules/brush.ts", "../../exporter/core/GriddoLog.ts", "../../exporter/services/api.ts", "../../exporter/services/render.ts", "../../exporter/core/db.ts", "../../exporter/shared/npm-modules/pkg-dir.ts", "../../exporter/shared/npm-modules/find-up-simple.ts", "../../exporter/core/fs.ts", "../../exporter/shared/errors.ts", "../../exporter/shared/endpoints.ts", "../../exporter/services/auth.ts", "../../exporter/core/errors.ts", "../../exporter/core/logger.ts", "../../exporter/commands/reset-render.ts"],
|
|
4
|
+
"sourcesContent": ["const { env } = process;\n\n/**\n * Returns true/false from string\n */\nfunction envIsTruthy(value?: string): boolean {\n\tif (!value) return false;\n\n\tswitch (value.trim().toLowerCase()) {\n\t\tcase \"1\":\n\t\tcase \"true\":\n\t\tcase \"yes\":\n\t\tcase \"y\":\n\t\tcase \"on\":\n\t\t\treturn true;\n\t\tdefault:\n\t\t\treturn false;\n\t}\n}\n\n// Credentials\nconst GRIDDO_API_URL = env.GRIDDO_API_URL || env.API_URL;\nconst GRIDDO_PUBLIC_API_URL = env.GRIDDO_PUBLIC_API_URL || env.PUBLIC_API_URL;\nconst GRIDDO_BOT_USER = env.botEmail || env.GRIDDO_BOT_USER;\nconst GRIDDO_BOT_PASSWORD = env.botPassword || env.GRIDDO_BOT_PASSWORD;\n\n// Rendering\nconst GRIDDO_API_CONCURRENCY_COUNT = Number.parseInt(env.GRIDDO_API_CONCURRENCY_COUNT || \"10\");\nconst GRIDDO_SKIP_BUILD_CHECKS = envIsTruthy(env.GRIDDO_SKIP_BUILD_CHECKS);\nconst GRIDDO_RENDER_BY_DOMAINS = envIsTruthy(env.GRIDDO_RENDER_BY_DOMAINS);\nconst GRIDDO_BUILD_LOGS = envIsTruthy(env.GRIDDO_BUILD_LOGS);\nconst GRIDDO_BUILD_LOGS_TO_FILE = envIsTruthy(env.GRIDDO_BUILD_LOGS_TO_FILE);\nconst GRIDDO_BUILD_LOGS_BUFFER_SIZE = Number.parseInt(env.GRIDDO_BUILD_LOGS_BUFFER_SIZE || \"500\");\nconst GRIDDO_SSG_VERBOSE_LOGS = envIsTruthy(env.GRIDDO_SSG_VERBOSE_LOGS);\nconst GRIDDO_SEARCH_FEATURE = envIsTruthy(env.GRIDDO_SEARCH_FEATURE);\nconst GRIDDO_ASSET_PREFIX = env.GRIDDO_ASSET_PREFIX || env.ASSET_PREFIX;\nconst GRIDDO_REACT_APP_INSTANCE = env.GRIDDO_REACT_APP_INSTANCE || env.REACT_APP_INSTANCE;\nconst GRIDDO_AI_EMBEDDINGS = envIsTruthy(env.GRIDDO_AI_EMBEDDINGS);\nconst GRIDDO_VERBOSE_LOGS = envIsTruthy(env.GRIDDO_VERBOSE_LOGS);\nconst GRIDDO_USE_DIST_BACKUP = envIsTruthy(env.GRIDDO_USE_DIST_BACKUP);\nconst GRIDDO_SSG_BUNDLE_ANALYZER = envIsTruthy(env.GRIDDO_SSG_BUNDLE_ANALYZER);\n\nexport {\n\tGRIDDO_AI_EMBEDDINGS,\n\tGRIDDO_API_CONCURRENCY_COUNT,\n\tGRIDDO_API_URL,\n\tGRIDDO_ASSET_PREFIX,\n\tGRIDDO_BOT_PASSWORD,\n\tGRIDDO_BOT_USER,\n\tGRIDDO_BUILD_LOGS,\n\tGRIDDO_BUILD_LOGS_BUFFER_SIZE,\n\tGRIDDO_BUILD_LOGS_TO_FILE,\n\tGRIDDO_PUBLIC_API_URL,\n\tGRIDDO_REACT_APP_INSTANCE,\n\tGRIDDO_RENDER_BY_DOMAINS,\n\tGRIDDO_SEARCH_FEATURE,\n\tGRIDDO_SKIP_BUILD_CHECKS,\n\tGRIDDO_SSG_BUNDLE_ANALYZER,\n\tGRIDDO_SSG_VERBOSE_LOGS,\n\tGRIDDO_USE_DIST_BACKUP,\n\tGRIDDO_VERBOSE_LOGS,\n};\n", "//\n// Brush adds color to a string|number, it does not print it!\n// Its simple, no log, no chains, just color in a string|number\n// usage:\n// console.log(brush.green(\"sucess!\"))\n//\n\nconst RESET = \"\\x1b[0m\";\nconst CODES = {\n\tblack: \"\\x1b[30m\",\n\tred: \"\\x1b[31m\",\n\tgreen: \"\\x1b[32m\",\n\tyellow: \"\\x1b[33m\",\n\tblue: \"\\x1b[34m\",\n\tmagenta: \"\\x1b[35m\",\n\tcyan: \"\\x1b[36m\",\n\twhite: \"\\x1b[37m\",\n\tgray: \"\\x1b[90m\",\n\tbold: \"\\x1b[1m\",\n\tdim: \"\\x1b[2m\",\n} as const;\n\ntype ColorFunction = (text: string | number) => string;\ntype ColorName = keyof typeof CODES;\ntype Brush = Record<ColorName, ColorFunction>;\n\nconst brush = {} as Brush;\n\nfor (const color in CODES) {\n\tconst key = color as ColorName;\n\tbrush[key] = (text: string | number) => `${CODES[key]}${text}${RESET}`;\n}\n\nexport { brush };\n", "import { GRIDDO_BUILD_LOGS, GRIDDO_VERBOSE_LOGS } from \"../shared/envs\";\nimport { brush } from \"../shared/npm-modules/brush\";\n\n/**\n * Clase est\u00E1tica para gestionar los logs de la aplicaci\u00F3n.\n * No se puede instanciar, se usa directamente: GriddoLogs.info(\"mensaje\").\n */\nclass GriddoLog {\n\t/** El constructor es privado para prevenir la instanciaci\u00F3n de la clase. */\n\tprivate constructor() {}\n\n\tpublic static verbose(...str: unknown[]): void {\n\t\tif (GRIDDO_VERBOSE_LOGS) {\n\t\t\tconsole.log(brush.yellow(\"verbose\"), brush.dim(str.join(\" \")));\n\t\t}\n\t}\n\n\tpublic static build(...str: unknown[]): void {\n\t\tif (GRIDDO_BUILD_LOGS) {\n\t\t\tGriddoLog.log(...str);\n\t\t}\n\t}\n\n\tpublic static info(...str: unknown[]): void {\n\t\tconsole.log(`${brush.blue(\"info\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static success(...str: unknown[]): void {\n\t\tconsole.log(`${brush.green(\"success\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static error(...str: unknown[]): void {\n\t\tconsole.error(`${brush.red(\"error\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static warn(...str: unknown[]): void {\n\t\tconsole.warn(`${brush.yellow(\"warn\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static log(...args: Parameters<typeof console.log>): void {\n\t\tconsole.log(...args);\n\t}\n}\n\nexport { GriddoLog };\n", "import type {\n\tAPIRequest,\n\tAPIResponses,\n\tGetAPI,\n\tPostAPI,\n\tPutAPI,\n\tShowApiErrorOptions,\n} from \"../shared/types/api\";\nimport type { Petition } from \"../shared/types/global\";\n\nimport crypto from \"node:crypto\";\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { RenderError } from \"../core/errors\";\nimport { pathExists } from \"../core/fs\";\nimport { GriddoLog } from \"../core/GriddoLog\";\nimport { addLogToBuffer } from \"../core/logger\";\nimport { brush } from \"../shared/npm-modules/brush\";\nimport { AuthService } from \"./auth\";\nimport { getRenderPathsHydratedWithDomainFromDB } from \"./render\";\n\n// Envs\nconst { env } = process;\nconst { RETRY_WAIT_SECONDS = \"4\", RETRY_ATTEMPTS = \"4\" } = env;\n\n/**\n * Make a GET/PUT/POST request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns {Promise<T>} A promise that is resolved with the data from the API response.\n *\n * @example\n *\tconst response = await requestAPI<Site>(\n *\t\t{ endpoint: \"...\", cacheKey: \"...\", ... },\n *\t\t\"get\",\n *\t\t\"...\"\n *\t);\n */\nasync function requestAPI<T extends APIResponses>(\n\tprops: APIRequest,\n\tmethod: string,\n\tappendToLog = \"\",\n): Promise<T> {\n\tconst {\n\t\tendpoint,\n\t\tbody,\n\t\tcacheKey = \"\",\n\t\tattempt = 1,\n\t\theaders,\n\t\tuseApiCacheDir = true,\n\t\tlogToFile = true,\n\t} = props;\n\tconst cacheOptions = { endpoint, body, headers, cacheKey };\n\n\t// Cache\n\tif (cacheKey && useApiCacheDir) {\n\t\tconst start = new Date();\n\t\tconst cacheData = await searchCacheData<T>(cacheOptions);\n\n\t\tif (cacheData) {\n\t\t\tif (logToFile) {\n\t\t\t\tconst siteId = getSafeSiteId(cacheData);\n\t\t\t\tconst siteIdMsg = siteId ? `site: ${siteId}` : \"\";\n\t\t\t\tconst duration = msToSec(Date.now() - start.getTime());\n\t\t\t\taddLogToBuffer(`${method} (cache) ${siteIdMsg} ${endpoint} - ${duration}s ${appendToLog}`);\n\t\t\t}\n\t\t\treturn cacheData;\n\t\t}\n\t}\n\n\t// Network\n\ttry {\n\t\tconst start = new Date();\n\n\t\t// Prepare fetch options\n\t\tconst fetchOptions: RequestInit = {\n\t\t\tmethod: method.toUpperCase(),\n\t\t\theaders: Object.assign({}, headers, AuthService.headers) as Record<string, string>,\n\t\t};\n\n\t\t// Add body for non-GET requests\n\t\tif (method.toLowerCase() !== \"get\" && body) {\n\t\t\tfetchOptions.body = JSON.stringify(body);\n\t\t\tif (!fetchOptions.headers) fetchOptions.headers = {};\n\t\t\t(fetchOptions.headers as Record<string, string>)[\"Content-Type\"] = \"application/json\";\n\t\t}\n\n\t\tconst response = await fetch(endpoint, fetchOptions);\n\n\t\t// Handle non-2xx responses\n\t\tif (!response.ok) {\n\t\t\tif (response.status === 404) {\n\t\t\t\t// @ts-expect-error page maybe will be 404\n\t\t\t\treturn null;\n\t\t\t}\n\n\t\t\tthrow new Error(`HTTP ${response.status}: ${response.statusText}`);\n\t\t}\n\n\t\tconst data: T = await response.json();\n\n\t\tif (logToFile) {\n\t\t\tconst siteId = getSafeSiteId(data);\n\t\t\tconst siteIdMsg = siteId ? `site: ${siteId}` : \"\";\n\t\t\tconst duration = msToSec(Date.now() - start.getTime());\n\t\t\taddLogToBuffer(`${method} (fetch) ${siteIdMsg} ${endpoint} - ${duration}s ${appendToLog}`);\n\t\t}\n\n\t\tif (useApiCacheDir) {\n\t\t\tawait saveCache(cacheOptions, data);\n\t\t}\n\n\t\treturn data;\n\t} catch (e) {\n\t\tconst error = e as Error;\n\n\t\tif (attempt > parseInt(RETRY_ATTEMPTS)) {\n\t\t\tGriddoLog.log(`\nMax attempts ${RETRY_ATTEMPTS} reached\n--------------------------------------\n- ${method.toUpperCase()} ${endpoint}\n- BODY: ${JSON.stringify(body)}\n- HEADERS: ${JSON.stringify(headers)}\n- ERROR: ${error.message}\n--------------------------------------\n`);\n\t\t\tthrow new RenderError(error);\n\t\t}\n\n\t\tshowApiError(error, {\n\t\t\tcallInfo: { endpoint, body },\n\t\t});\n\n\t\tGriddoLog.warn(`Waiting for retry: ${method}`, endpoint);\n\n\t\tawait delay(parseInt(RETRY_WAIT_SECONDS) * 1000);\n\n\t\treturn requestAPI<T>(\n\t\t\t{\n\t\t\t\tendpoint,\n\t\t\t\tbody,\n\t\t\t\theaders,\n\t\t\t\tcacheKey,\n\t\t\t\tattempt: attempt + 1,\n\t\t\t},\n\t\t\tmethod,\n\t\t\tappendToLog,\n\t\t);\n\t}\n}\n\n/**\n * Make a GET request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns A promise that is resolved with the data from the API response.\n */\nasync function getApi<T extends APIResponses>(props: GetAPI) {\n\treturn requestAPI<T>(props, \"get\");\n}\n\n/**\n * Make a PUT request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns A promise that is resolved with the data from the API response.\n */\nasync function putApi<T extends APIResponses>(props: PutAPI) {\n\treturn requestAPI<T>(props, \"put\");\n}\n\n/**\n * Make a POST request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns A promise that is resolved with the data from the API response.\n */\nasync function postApi<T extends APIResponses>(props: PostAPI) {\n\tconst { endpoint, body, headers } = props;\n\tconst referenceFieldBodyParams =\n\t\tendpoint.endsWith(\"/distributor\") &&\n\t\t`# ReferenceField body: ${JSON.stringify(body)} lang: ${JSON.stringify(headers?.lang)}`;\n\n\treturn requestAPI<T>(props, \"post\", referenceFieldBodyParams || \"\");\n}\n\n/**\n * Shows an API error through the terminal.\n */\nfunction showApiError(error: Error, options: ShowApiErrorOptions) {\n\tconst { message, stack } = error;\n\tconst { callInfo } = options;\n\tconst callInfoArray = [];\n\n\tfor (const item of Object.keys(callInfo) as (keyof typeof callInfo)[]) {\n\t\tcallInfoArray.push(\n\t\t\t`${item}: ${\n\t\t\t\ttypeof callInfo[item] === \"object\" ? JSON.stringify(callInfo[item]) : callInfo[item]\n\t\t\t}`,\n\t\t);\n\t}\n\n\t// Compose the errors output\n\tconst callInfoStr = callInfoArray.join(\"\\n\");\n\tconst errorDetailsStr = `${message}\\n${stack}`;\n\n\t// Print the error\n\tGriddoLog.warn(\n\t\tbrush.red(`\n=============\n\n{ Call info }\n${callInfoStr}\n\n{ Error details }\n${errorDetailsStr}\n\n=============\n`),\n\t);\n}\n\n/**\n * Return a siteID from a response object if exist\n * @param response A response object\n */\nfunction getSafeSiteId(response: APIResponses) {\n\tif (typeof response !== \"object\" || response === null || Array.isArray(response)) {\n\t\treturn undefined;\n\t}\n\n\treturn \"site\" in response && response.site ? response.site : undefined;\n}\n\n/**\n * Custom delay using the \"promise hack\",\n *\n * @param ms Amount of miliseconds to be delayed\n */\nfunction delay(ms: number): Promise<void> {\n\treturn new Promise((res) => setTimeout(res, ms));\n}\n\n/**\n * Converts milliseconds to seconds with a fixed number of decimals.\n *\n * @param ms The number in milliseconds.\n * @param fixed The amount of fixed decimals.\n * @returns The converted number in seconds with the fixed number of decimals.\n */\nfunction msToSec(ms: number, decimals = 3): number {\n\treturn Number.parseFloat((ms / 1000).toFixed(decimals));\n}\n\n/**\n * Generate a filename with a hash using a Petition object\n * @param petition An object\n */\nasync function generateFilenameWithHash(petition: Petition) {\n\tconst { __root } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst apiCacheDir = path.join(__root, \"apiCache\");\n\n\tconst hashSum = crypto.createHash(\"sha256\");\n\thashSum.update(JSON.stringify(petition));\n\n\treturn `${apiCacheDir}/${hashSum.digest(\"hex\")}`;\n}\n\n/**\n * Save a file using a hash name.\n *\n * @param petition An object.\n * @param content Content to be saved.\n */\nasync function saveCache<T>(petition: Petition, content: T) {\n\tconst stringContent = typeof content === \"string\" ? content : JSON.stringify(content);\n\tconst filename = await generateFilenameWithHash(petition);\n\tconst filepath = path.dirname(filename);\n\n\tif (!(await pathExists(filepath))) {\n\t\tawait fsp.mkdir(filepath, { recursive: true });\n\t}\n\n\tawait fsp.writeFile(filename, stringContent, \"utf8\");\n}\n\n/**\n * Search in the `apiCache` dir for a file using the petition as hash generator.\n * Return the file content if found or null if not.\n *\n * @param petition An object\n */\nasync function searchCacheData<T>(petition: Petition) {\n\ttry {\n\t\tconst file = await generateFilenameWithHash(petition);\n\t\tconst fileContent = await fsp.readFile(file, \"utf8\");\n\t\tconst jsonData = JSON.parse(fileContent) as T;\n\n\t\treturn jsonData;\n\t} catch {\n\t\treturn null;\n\t}\n}\n\nexport { getApi as get, postApi as post, putApi as put };\n", "import type { RenderModeTuple } from \"../shared/types/render\";\n\nimport { execSync } from \"node:child_process\";\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { readDB, writeDB } from \"../core/db\";\nimport { throwError } from \"../core/errors\";\nimport { pathExists } from \"../core/fs\";\nimport { GriddoLog } from \"../core/GriddoLog\";\nimport { RenderUUIDError } from \"../shared/errors\";\nimport { brush } from \"../shared/npm-modules/brush\";\nimport { RENDER_MODE } from \"../shared/types/render\";\nimport { AuthService } from \"./auth\";\nimport { getBuildMetadata } from \"./manage-store\";\n\n/**\n * Creates a sentinel file with the current date and time.\n * This file is used to track later if node_modules/@griddo/cx was cleaned by a\n * npm install coming from a deploy.\n */\nasync function markRenderAsStarted(options: { domain: string; basePath: string }) {\n\tconst { domain } = options;\n\n\tconst db = await readDB();\n\tdb.domains[domain].isRendering = true;\n\tawait writeDB(db);\n\n\t// Creamos un archivo centinela, si al terminar el render este archivo no\n\t// existe es que ha habido un deploy por medio y hay que invalidar el render\n\tconst { __ssg } = await getRenderPathsHydratedWithDomainFromDB();\n\n\tconst renderSentinelFile = path.join(__ssg, `.render-sentinel-${domain}`);\n\tawait fsp.writeFile(renderSentinelFile, new Date().toISOString());\n}\n\nasync function markRenderAsCompleted(domain: string) {\n\tconst db = await readDB();\n\tdb.domains[domain].isRendering = false;\n\tdb.currentRenderingDomain = null;\n\tdb.domains[domain].renderMode = \"COMPLETED\";\n\t// db.domains[domain].shouldBeRendered = false;\n\tawait writeDB(db);\n\n\t// Borramos finalmente el archivo centinela\n\tconst { __ssg } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst renderSentinelFile = path.join(__ssg, `.render-sentinel-${domain}`);\n\tawait fsp.unlink(renderSentinelFile);\n}\n\nasync function assertRenderIsValid(domain: string) {\n\t// Comprobamos que .render-sentinel exista, si no es que un deploy lo borro\n\t// y hay que invalidar el render.\n\tconst { __ssg } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst renderSentinelFile = path.join(__ssg, `.render-sentinel-${domain}`);\n\tif (!(await pathExists(renderSentinelFile))) {\n\t\tthrowError(RenderUUIDError);\n\t}\n}\n\n/**\n * Determines the appropriate render mode for a given domain based on its current state,\n * previous render errors, deployment status, and whether rendering is required.\n *\n * @param options - An object containing:\n * - `domain`: The domain name to resolve the render mode for.\n * - `shouldBeRendered`: A boolean indicating if the domain should be rendered.\n * @returns An object with:\n * - `renderMode`: The resolved render mode (`FROM_SCRATCH`, `INCREMENTAL`, or `IDLE`).\n * - `reason`: A string describing the reason for the chosen render mode.\n *\n * @remarks\n * The function checks for missing exports, previous render errors, new deployments,\n * and whether rendering is necessary to decide the render mode.\n *\n * @todo\n * Improve ifs and reason concatenations...\n */\nasync function resolveDomainRenderMode(options: { domain: string; shouldBeRendered: boolean }) {\n\tconst { domain, shouldBeRendered } = options;\n\n\tconst db = await readDB();\n\n\tconst { __cache, __exports } = await getRenderPathsHydratedWithDomainFromDB({ domain });\n\tconst exportsAreMissing = !(await pathExists(path.join(__exports)));\n\tconst previousRenderFailed = db.domains[domain]?.isRendering;\n\tconst newDeployDetected = await hasNewCommit(__cache);\n\n\tif (exportsAreMissing) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.FROM_SCRATCH,\n\t\t\treason: \"missing exports directory\",\n\t\t};\n\t}\n\n\tif (previousRenderFailed) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.FROM_SCRATCH,\n\t\t\treason: \"error in previous render\",\n\t\t};\n\t}\n\n\tif (newDeployDetected) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.FROM_SCRATCH,\n\t\t\treason: \"new commit hash\",\n\t\t};\n\t}\n\n\tif (!shouldBeRendered) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.IDLE,\n\t\t\treason: \"no activity\",\n\t\t};\n\t}\n\n\treturn {\n\t\trenderMode: RENDER_MODE.INCREMENTAL,\n\t\treason: \"has changes\",\n\t};\n}\n\nasync function hasNewCommit(basePath: string): Promise<boolean> {\n\tconst commitFile = path.join(basePath, \"commit\");\n\tconst currentCommit = execSync(\"git rev-parse HEAD\").toString().trim();\n\n\tif (await pathExists(commitFile)) {\n\t\tconst savedCommit = (await fsp.readFile(commitFile, \"utf-8\")).trim();\n\t\tif (savedCommit === currentCommit) {\n\t\t\treturn false; // No hay nuevo commit\n\t\t}\n\n\t\treturn true;\n\t}\n\n\treturn true;\n}\n\nasync function updateCommitFile(options: { basePath: string }) {\n\tconst { basePath } = options;\n\tconst currentCommit = execSync(\"git rev-parse HEAD\").toString().trim();\n\tawait fsp.writeFile(path.join(basePath, \"commit\"), currentCommit);\n}\n\nasync function getRenderModeFromDB(domain: string): Promise<RenderModeTuple> {\n\tconst db = await readDB();\n\n\tif (!db.domains[domain]) {\n\t\tthrow new Error(brush.red(`[!] Error: Domain ${domain} not found in DB`));\n\t}\n\n\tif (!db.domains[domain].renderMode) {\n\t\tthrow new Error(brush.red(`[!] Error: Render mode not found for domain ${domain}`));\n\t}\n\n\treturn {\n\t\trenderMode: db.domains[domain].renderMode,\n\t\treason: db.domains[domain].renderModeReason,\n\t};\n}\n\nasync function getRenderPathsHydratedWithDomainFromDB(options?: {\n\tdomain?: string;\n\tdbFilePath?: string;\n}) {\n\tconst { domain, dbFilePath } = options || {};\n\n\tconst db = await readDB(dbFilePath);\n\tconst paths = db.paths;\n\n\treturn {\n\t\t__root: paths.root,\n\t\t__cache: path.join(paths.cxCache, domain || \"\"),\n\t\t__components: paths.components,\n\t\t__cx: paths.cx,\n\t\t__sites: paths.exportsDir,\n\t\t__exports: path.join(paths.exportsDir, domain || \"\"),\n\t\t__exports_backup: path.join(paths.exportsDirBackup, domain || \"\"),\n\t\t__ssg: paths.ssg,\n\t\t__exports_dist: path.join(paths.exportsDir, domain || \"\", \"dist\"),\n\t};\n}\n\nasync function getRenderMetadataFromDB() {\n\tconst db = await readDB();\n\treturn {\n\t\tgriddoVersion: db.griddoVersion,\n\t\tbuildReportFileName: db.buildReportFileName,\n\t};\n}\n\n/**\n * Save a file with the end of build process to use as `end-render` signal.\n */\nasync function generateBuildReport(domain: string) {\n\tconst authControl = await AuthService.login();\n\n\tconst { __root } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst { buildReportFileName } = await getRenderMetadataFromDB();\n\tconst { buildProcessData } = await getBuildMetadata(domain);\n\n\tconst buildSitesInfo = Object.keys(buildProcessData).map((siteID) => ({\n\t\t...buildProcessData[siteID],\n\t\tsiteId: Number.parseInt(siteID),\n\t}));\n\n\tconst report = {\n\t\tauthControl,\n\t\tsites: buildSitesInfo,\n\t};\n\n\tconst reportFilePath = path.join(__root, \"current-dist\", buildReportFileName);\n\n\tawait fsp.writeFile(reportFilePath, JSON.stringify(report));\n\n\tGriddoLog.verbose(`build report saved in ${reportFilePath}`);\n}\n\nexport {\n\tassertRenderIsValid,\n\tgenerateBuildReport,\n\tgetRenderMetadataFromDB,\n\tgetRenderModeFromDB,\n\tgetRenderPathsHydratedWithDomainFromDB,\n\tmarkRenderAsCompleted,\n\tmarkRenderAsStarted,\n\tresolveDomainRenderMode,\n\tupdateCommitFile,\n};\n", "import type { RenderDB } from \"../shared/types/render\";\n\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { pkgDirSync } from \"../shared/npm-modules/pkg-dir\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nconst root = pkgDirSync({ cwd: path.resolve(__dirname, \"../../..\") }) || \"\";\nconst cache = path.join(root, \".griddo/cache\");\nconst dbFilePath = path.join(cache, \"db.json\");\n\nasync function readDB(customDBPath = \"\") {\n\tconst file = customDBPath || dbFilePath;\n\ttry {\n\t\treturn JSON.parse(await fsp.readFile(file, \"utf-8\")) as RenderDB;\n\t} catch (error) {\n\t\tGriddoLog.error(`Failed to read DB file at ${file}:`, error);\n\t\tthrow error;\n\t}\n}\n\nasync function writeDB(renderDB: RenderDB, customDBPath = \"\") {\n\tconst file = customDBPath || dbFilePath;\n\ttry {\n\t\tawait fsp.writeFile(file, JSON.stringify(renderDB, null, \"\\t\"));\n\t} catch (error) {\n\t\tGriddoLog.error(`Failed to write DB file at ${file}:`, error);\n\t\tthrow error;\n\t}\n}\n\nexport { readDB, writeDB };\n", "import path from \"node:path\";\n\nimport { findUp, findUpSync } from \"./find-up-simple\";\n\nasync function pkgDir(options?: { readonly cwd?: string }) {\n\tconst { cwd } = options || {};\n\tconst filePath = await findUp(\"package.json\", { cwd });\n\treturn filePath && path.dirname(filePath);\n}\n\nfunction pkgDirSync(options?: { readonly cwd?: string }) {\n\tconst { cwd } = options || {};\n\tconst filePath = findUpSync(\"package.json\", { cwd });\n\treturn filePath && path.dirname(filePath);\n}\n\nexport { pkgDir, pkgDirSync };\n", "import fs from \"node:fs\";\nimport fsPromises from \"node:fs/promises\";\nimport path from \"node:path\";\nimport process from \"node:process\";\nimport { fileURLToPath } from \"node:url\";\n\n/**\n * Opciones para controlar el comportamiento de la b\u00FAsqueda.\n */\nexport type Options = {\n\t/**\n\tEl directorio desde donde empezar a buscar.\n\t@default process.cwd()\n\t*/\n\treadonly cwd?: URL | string;\n\n\t/**\n\tEl tipo de ruta a buscar.\n\t@default 'file'\n\t*/\n\treadonly type?: \"file\" | \"directory\";\n\n\t/**\n\tUn directorio en el que la b\u00FAsqueda se detiene si no se encuentran coincidencias.\n\t@default El directorio ra\u00EDz del sistema\n\t*/\n\treadonly stopAt?: URL | string;\n};\n\n// Funci\u00F3n auxiliar para convertir una URL en una ruta de archivo de tipo string.\nconst toPath = (urlOrPath: URL | string | undefined): string | undefined =>\n\turlOrPath instanceof URL ? fileURLToPath(urlOrPath) : urlOrPath;\n\n/**\n * Encuentra un archivo o directorio de forma as\u00EDncrona subiendo por los directorios padre.\n * @param name - El nombre del archivo o directorio a buscar.\n * @param options - Opciones de b\u00FAsqueda.\n * @returns La ruta encontrada o `undefined` si no se pudo encontrar.\n */\nexport async function findUp(name: string, options: Options = {}): Promise<string | undefined> {\n\tconst { cwd = process.cwd(), type = \"file\", stopAt: stopAtOption } = options;\n\n\tlet directory = path.resolve(toPath(cwd) ?? \"\");\n\tconst { root } = path.parse(directory);\n\tconst stopAt = stopAtOption ? path.resolve(directory, toPath(stopAtOption)!) : root;\n\tconst isAbsoluteName = path.isAbsolute(name);\n\n\twhile (true) {\n\t\tconst filePath = isAbsoluteName ? name : path.join(directory, name);\n\t\ttry {\n\t\t\t// eslint-disable-next-line no-await-in-loop\n\t\t\tconst stats = await fsPromises.stat(filePath);\n\t\t\tif ((type === \"file\" && stats.isFile()) || (type === \"directory\" && stats.isDirectory())) {\n\t\t\t\treturn filePath;\n\t\t\t}\n\t\t} catch {\n\t\t\t// Ignora errores (ej. el archivo no existe) y contin\u00FAa la b\u00FAsqueda.\n\t\t}\n\n\t\tif (directory === stopAt || directory === root) {\n\t\t\tbreak;\n\t\t}\n\n\t\tdirectory = path.dirname(directory);\n\t}\n}\n\n/**\n * Encuentra un archivo o directorio de forma s\u00EDncrona subiendo por los directorios padre.\n * @param name - El nombre del archivo o directorio a buscar.\n * @param options - Opciones de b\u00FAsqueda.\n * @returns La ruta encontrada o `undefined` si no se pudo encontrar.\n */\nexport function findUpSync(name: string, options: Options = {}): string | undefined {\n\tconst { cwd = process.cwd(), type = \"file\", stopAt: stopAtOption } = options;\n\n\tlet directory = path.resolve(toPath(cwd) ?? \"\");\n\tconst { root } = path.parse(directory);\n\tconst stopAt = stopAtOption ? path.resolve(directory, toPath(stopAtOption)!) : root;\n\tconst isAbsoluteName = path.isAbsolute(name);\n\n\twhile (true) {\n\t\tconst filePath = isAbsoluteName ? name : path.join(directory, name);\n\t\ttry {\n\t\t\tconst stats = fs.statSync(filePath, { throwIfNoEntry: false });\n\t\t\tif ((type === \"file\" && stats?.isFile()) || (type === \"directory\" && stats?.isDirectory())) {\n\t\t\t\treturn filePath;\n\t\t\t}\n\t\t} catch {\n\t\t\t// En teor\u00EDa, statSync con `throwIfNoEntry: false` no deber\u00EDa lanzar un error,\n\t\t\t// pero se mantiene por seguridad.\n\t\t}\n\n\t\tif (directory === stopAt || directory === root) {\n\t\t\tbreak;\n\t\t}\n\n\t\tdirectory = path.dirname(directory);\n\t}\n}\n", "import type { MakeDirectoryOptions } from \"node:fs\";\n\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { ArtifactError } from \"../shared/errors\";\nimport { throwError } from \"./errors\";\nimport { GriddoLog } from \"./GriddoLog\";\n\n/**\n * Remove an empty directory from the basePath recursively.\n * If the directory has only .xml files it will handle as empty too (empty site)\n *\n * @param baseDir - The base directory.\n */\nasync function deleteDisposableSiteDirs(baseDir: string) {\n\tif (!(await pathExists(baseDir))) {\n\t\treturn;\n\t}\n\n\tconst sitesDirs = (await fsp.readdir(baseDir, { withFileTypes: true })).filter((file) =>\n\t\tfile.isDirectory(),\n\t);\n\n\tfor (const siteDir of sitesDirs) {\n\t\tconst sitePath = path.join(baseDir, siteDir.name);\n\t\tif (await siteIsEmpty(sitePath)) {\n\t\t\tawait fsp.rm(sitePath, { recursive: true });\n\t\t}\n\t}\n}\n\n/**\n * Creates multiple artifact directories.\n *\n * @param dirs - An array of directory paths.\n * @param options - Same option as `fs.mkdirSync()`\n */\nasync function mkDirs(dirs: string[], options?: MakeDirectoryOptions) {\n\tfor (const dir of dirs) {\n\t\ttry {\n\t\t\tif (!(await pathExists(dir))) {\n\t\t\t\tawait fsp.mkdir(dir, { recursive: true, ...options });\n\t\t\t\tGriddoLog.verbose(`create directory: ${dir}`);\n\t\t\t}\n\t\t} catch (error) {\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\nasync function renamePath(src: string, dst: string) {\n\ttry {\n\t\tif (await pathExists(src)) {\n\t\t\tawait fsp.rename(src, dst);\n\t\t\tGriddoLog.verbose(`rename ${src} to ${dst}`);\n\t\t}\n\t} catch (error) {\n\t\tthrowError(ArtifactError, error);\n\t}\n}\n\n/**\n * Copy multiple directories with backup option.\n *\n * @param src - Source directory.\n * @param dst - Destination directory.\n * @param dirs - Directories to copy.\n * @param options.withBackup - Create a previous backup before copy.\n */\nasync function cpDirs(\n\tsrc: string,\n\tdst: string,\n\tdirs: string[],\n\toptions = {\n\t\twithBackup: false,\n\t},\n) {\n\tconst { withBackup } = options;\n\tfor (const dir of dirs) {\n\t\tconst srcCompose = path.join(src, dir);\n\t\tconst dstCompose = path.join(dst, dir);\n\n\t\t// The dir we want to copy, doesn't exist.\n\t\tif (!(await pathExists(srcCompose))) {\n\t\t\tGriddoLog.info(`(Maybe first render) Source directory does not exist: ${srcCompose}`);\n\t\t\tcontinue;\n\t\t}\n\n\t\t// Create the backup\n\t\tif (withBackup) {\n\t\t\tawait createBackup(dstCompose);\n\t\t\tGriddoLog.verbose(`create backup: ${dstCompose}`);\n\t\t}\n\n\t\t// Copy directory\n\t\ttry {\n\t\t\t// First clean destination\n\t\t\tif (await pathExists(dstCompose)) {\n\t\t\t\tawait fsp.rm(dstCompose, { recursive: true, force: true });\n\t\t\t\tGriddoLog.verbose(`clean destination: ${dstCompose}`);\n\t\t\t}\n\n\t\t\t// Then copy src to dst\n\t\t\tawait fsp.cp(srcCompose, dstCompose, {\n\t\t\t\trecursive: true,\n\t\t\t\tpreserveTimestamps: true,\n\t\t\t});\n\t\t\tGriddoLog.verbose(`copy: ${srcCompose} to ${dstCompose}`);\n\n\t\t\tif (withBackup) {\n\t\t\t\tawait deleteBackup(dstCompose);\n\t\t\t\tGriddoLog.verbose(`delete backup: ${dstCompose}`);\n\t\t\t}\n\t\t} catch (error) {\n\t\t\tif (withBackup) {\n\t\t\t\tawait restoreBackup(dstCompose);\n\t\t\t\tGriddoLog.verbose(\"Backup has been restored.\");\n\t\t\t}\n\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\n/**\n * Move artifacts between cx-paths\n *\n * @param src - Source directory.\n * @param dst - Destination directory.\n * @param dirs - Directories to move.\n * @param options - Options.\n */\nasync function mvDirs(\n\tsrc: string,\n\tdst: string,\n\tdirs: string[],\n\toptions?: { withBackup?: boolean; override?: boolean },\n) {\n\tconst { override, withBackup } = options || {};\n\n\tfor (const dir of dirs) {\n\t\tconst srcCompose = path.join(src, dir);\n\t\tconst dstCompose = path.join(dst, dir);\n\n\t\tif (!(await pathExists(srcCompose))) {\n\t\t\tcontinue;\n\t\t}\n\n\t\tif (withBackup) {\n\t\t\tawait createBackup(dstCompose);\n\t\t}\n\n\t\ttry {\n\t\t\t// Clean destination\n\t\t\tif (override && (await pathExists(dstCompose))) {\n\t\t\t\tawait fsp.rm(dstCompose, { recursive: true, force: true });\n\t\t\t}\n\n\t\t\tawait fsp.rename(srcCompose, dstCompose);\n\t\t\tGriddoLog.verbose(`moved: ${srcCompose} to ${dstCompose}`);\n\n\t\t\tif (withBackup) {\n\t\t\t\tawait deleteBackup(dstCompose);\n\t\t\t}\n\t\t} catch (error) {\n\t\t\tif (withBackup) {\n\t\t\t\tawait restoreBackup(dstCompose);\n\t\t\t\tGriddoLog.info(\"Backup has been restored.\");\n\t\t\t}\n\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\n/**\n * Removes multiple artifact directories.\n *\n * @param dirs - An array of directory paths.\n */\nasync function rmDirs(dirs: string[]) {\n\tfor (const dir of dirs) {\n\t\ttry {\n\t\t\tawait fsp.rm(dir, { recursive: true, force: true });\n\t\t\tGriddoLog.verbose(`artifact removed: ${dir}`);\n\t\t} catch (error) {\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\nasync function restoreBackup(src: string, suffix = \"-BACKUP\") {\n\tconst dst = src + suffix;\n\ttry {\n\t\tawait fsp.rename(dst, src);\n\t\tGriddoLog.info(`Backup ${dst} has been restored`);\n\t} catch (_error) {\n\t\tthrow new Error(`Error while delete ${dst} backup`);\n\t}\n}\n\nasync function deleteBackup(src: string, suffix = \"-BACKUP\") {\n\tconst dst = src + suffix;\n\n\tif (!(await pathExists(dst))) {\n\t\treturn;\n\t}\n\n\ttry {\n\t\tawait fsp.rm(dst, { recursive: true, force: true });\n\t\tGriddoLog.verbose(`Backup ${dst} has been deleted`);\n\t} catch (_error) {\n\t\tthrow new Error(`Error while delete ${dst} backup`);\n\t}\n}\n\nasync function createBackup(src: string, suffix = \"-BACKUP\") {\n\tconst dst = src + suffix;\n\n\tif (!(await pathExists(src))) {\n\t\treturn;\n\t}\n\n\tif (await pathExists(dst)) {\n\t\tGriddoLog.warn(`Destination ${dst} already exists`);\n\t\treturn;\n\t}\n\n\ttry {\n\t\tawait fsp.rename(src, dst);\n\t\tGriddoLog.verbose(`Backup of ${src} has been created in ${dst}`);\n\t} catch (error) {\n\t\tGriddoLog.error(`Error while coping ${src} to ${dst} backup`);\n\t\tthrowError(ArtifactError, error);\n\t}\n}\n\n/**\n * Return true if the site folder is empty or only has xml files. (Recursively)\n */\nasync function siteIsEmpty(sitePath: string) {\n\tconst siteFiles = (\n\t\tawait fsp.readdir(sitePath, {\n\t\t\twithFileTypes: true,\n\t\t\trecursive: true,\n\t\t})\n\t).filter((file) => file.isFile() && !path.basename(file.name).startsWith(\".\"));\n\n\tconst xmlFiles = siteFiles.filter((file) => file.name.endsWith(\".xml\"));\n\n\tif (siteFiles.length === xmlFiles.length) {\n\t\treturn true;\n\t}\n}\n\n/**\n * Delete empty directories from the given directory in a recursive way.\n */\nasync function deleteEmptyDirectories(dirPath: string) {\n\ttry {\n\t\tconst stats = await fsp.stat(dirPath);\n\n\t\t// Si no es un directorio, no hacemos nada\n\t\tif (!stats.isDirectory()) {\n\t\t\treturn;\n\t\t}\n\n\t\tlet filesInDirectory: string[];\n\t\ttry {\n\t\t\tfilesInDirectory = await fsp.readdir(dirPath);\n\t\t} catch (err: any) {\n\t\t\t// Si el directorio no existe o no se puede leer (ej. permisos), lo saltamos.\n\t\t\tif (err.code === \"ENOENT\") {\n\t\t\t\tGriddoLog.warn(`The directory \"${dirPath}\" does not exist, skipping it.`);\n\t\t\t\treturn;\n\t\t\t}\n\t\t\tGriddoLog.error(`Error al leer el directorio \"${dirPath}\":`, err);\n\t\t\tthrow err; // Re-lanza el error para que sea manejado por el llamador\n\t\t}\n\n\t\t// Recorrer los contenidos del directorio\n\t\tfor (const file of filesInDirectory) {\n\t\t\tconst fullPath = path.join(dirPath, file);\n\t\t\tawait deleteEmptyDirectories(fullPath); // Llamada recursiva s\u00EDncrona\n\t\t}\n\n\t\t// Despu\u00E9s de procesar todos los subdirectorios, verifica si el directorio actual est\u00E1 vac\u00EDo\n\t\tconst remainingFiles = await fsp.readdir(dirPath);\n\n\t\tif (remainingFiles.length === 0) {\n\t\t\ttry {\n\t\t\t\tawait fsp.rmdir(dirPath);\n\t\t\t\tGriddoLog.verbose(`Remove empty directory: ${dirPath}`);\n\t\t\t} catch (err: any) {\n\t\t\t\t// Puede que haya habido un problema de concurrencia o permisos\n\t\t\t\tif (err.code === \"ENOENT\") {\n\t\t\t\t\tGriddoLog.warn(\n\t\t\t\t\t\t`El directorio \"${dirPath}\" ya no existe. Posiblemente fue borrado por otra operaci\u00F3n.`,\n\t\t\t\t\t);\n\t\t\t\t} else {\n\t\t\t\t\tGriddoLog.error(`Error al borrar el directorio \"${dirPath}\":`, err);\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} catch (err: any) {\n\t\tif (err.code === \"ENOENT\") {\n\t\t\t// El directorio ya no existe, no es un error para nosotros en este contexto\n\t\t\tGriddoLog.warn(`The directory \"${dirPath}\" does not exist or has already been processed.`);\n\t\t} else {\n\t\t\tGriddoLog.error(`General error general while processing \"${dirPath}\":`, err);\n\t\t\tthrow err;\n\t\t}\n\t}\n}\n\nasync function pathExists(dir: string) {\n\ttry {\n\t\tawait fsp.access(dir);\n\t\treturn true;\n\t} catch {\n\t\treturn false;\n\t}\n}\n\n/**\n * Busca recursivamente archivos que terminen con un sufijo espec\u00EDfico dentro de un directorio.\n * Esta funci\u00F3n es un generador as\u00EDncrono, lo que la hace muy eficiente en uso de memoria.\n *\n * @param dir El directorio base para comenzar la b\u00FAsqueda.\n * @param suffix El sufijo con el que deben terminar los nombres de archivo (ej: 'page-data.json').\n * @returns Un generador as\u00EDncrono que produce la ruta completa de cada archivo encontrado.\n * @throws Si el directorio inicial `dir` no existe o no se puede leer.\n */\nasync function* findFilesBySuffix(dir: string, suffix: string): AsyncGenerator<string> {\n\tconst dirHandle = await fsp.opendir(dir);\n\tfor await (const item of dirHandle) {\n\t\tconst fullPath = path.join(dir, item.name);\n\t\tif (item.isDirectory()) {\n\t\t\t// yield* para encadenar otro generator.\n\t\t\tyield* findFilesBySuffix(fullPath, suffix);\n\t\t} else if (item.isFile() && item.name.endsWith(suffix)) {\n\t\t\tyield fullPath;\n\t\t}\n\t}\n}\n\n/**\n * Walk a directory and returns the JSON file absolute paths with one level of depth.\n * /abs/.../sotre/331/158.json\n * /abs/.../sotre/114/443.json\n * /abs/.../sotre/131/217.json\n * /abs/.../sotre/191/281.json\n */\nasync function* walkStore(storeDir: string): AsyncGenerator<string> {\n\tconst storeDirHandle = await fsp.opendir(storeDir);\n\n\tfor await (const siteDirent of storeDirHandle) {\n\t\tif (siteDirent.isDirectory()) {\n\t\t\tconst siteDirPath = path.join(storeDir, siteDirent.name);\n\t\t\tconst siteDirHandle = await fsp.opendir(siteDirPath);\n\n\t\t\tfor await (const fileDirent of siteDirHandle) {\n\t\t\t\tconst filePath = path.join(siteDirPath, fileDirent.name);\n\n\t\t\t\tif (fileDirent.isFile() && path.extname(filePath) === \".json\") {\n\t\t\t\t\tyield filePath;\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}\n\nexport {\n\tcpDirs,\n\tdeleteDisposableSiteDirs,\n\tdeleteEmptyDirectories,\n\tfindFilesBySuffix,\n\tmkDirs,\n\tmvDirs,\n\tpathExists,\n\trenamePath,\n\trmDirs,\n\twalkStore,\n};\n", "/**\n * Do you want to add a new error to the list?\n *\n * 1 - Add the new error type name to the `ErrorsType` union type.\n * 2 - Export a new ErrorData object (or a function that returns one) in this\n * file by completing the `error` (ErrosType) and `message` (string) properties\n * obligatorily.\n */\n\nimport type { SpawnSyncReturns } from \"node:child_process\";\nimport type { ErrorData } from \"../core/errors\";\n\ntype ErrorsType =\n\t| \"ArtifactError\"\n\t| \"BundlesInconsistencyError\"\n\t| \"CheckHealthError\"\n\t| \"ErrorInSSGBuildProcess\"\n\t| \"LifecycleExecutionError\"\n\t| \"LoginError\"\n\t| \"NoDomainsFoundError\"\n\t| \"NoJSConfigFileFound\"\n\t| \"ReadFromStoreError\"\n\t| \"ReferenceFieldSourcesNotFoundError\"\n\t| \"RenderUUIDError\"\n\t| \"UploadSearchError\"\n\t| \"WriteToStoreError\";\n\nconst ArtifactError: ErrorData = {\n\terror: \"ArtifactError\",\n\tmessage: \"There was a problem with an artifact\",\n\texpected:\n\t\t\"An external process may have has modified or deleted one of the artifacts (files and directories).\",\n\thint: \"Have there been any recent deployments? These can delete directories from the current render.\",\n};\n\nconst ErrorInSSGBuildProcess = (command: SpawnSyncReturns<string>): ErrorData => ({\n\terror: \"ErrorInSSGBuildProcess\",\n\tmessage: `Error in SSG build process: ${JSON.stringify(command)}`,\n\texpected: \"This can happen if there was a problem with the SSG build process.\",\n});\n\nconst LifecycleExecutionError = (attempts: number, name: string): ErrorData => ({\n\terror: \"LifecycleExecutionError\",\n\tmessage: `Exceeded maximum retry attempts (${attempts}) for ${name} LifeCycle`,\n});\n\nconst LoginError: ErrorData = {\n\terror: \"LoginError\",\n\tmessage: \"There was a problem logging in to the API\",\n\texpected: \"This happens if the API is currently not working or the credentials are incorrect.\",\n};\n\nconst NoDomainsFoundError: ErrorData = {\n\terror: \"NoDomainsFoundError\",\n\tmessage: \"No domains were found in this instance. The process cannot continue.\",\n\texpected:\n\t\t\"This may happen if the API is not functioning, or the site is not properly configured, or the domains are not registered.\",\n\thint: \"You can contact the instance administrator.\",\n};\n\nconst NoJSConfigFileFound: ErrorData = {\n\terror: \"NoJSConfigFileFound\",\n\tmessage: \"Could not find jsconfig.json or tsconfig.json\",\n\texpected:\n\t\t\"This can happen if the instance is not properly configured with a jsconfig.json or tsconfig.json file.\",\n};\n\nconst ReadFromStoreError: ErrorData = {\n\terror: \"ReadFromStoreError\",\n\tmessage: \"There was an error reading a file to the Store directory\",\n\thint: \"There may be an issue such as permissions preventing the file from being read.\",\n};\n\nconst ReferenceFieldSourcesNotFoundError: ErrorData = {\n\terror: \"ReferenceFieldSourcesNotFoundError\",\n\tmessage: \"The distributor has no sources defined.\",\n\texpected:\n\t\t\"It is expected to have at least one data source in the `sources` property, even if it is empty.\",\n};\n\nconst RenderUUIDError: ErrorData = {\n\terror: \"RenderUUIDError\",\n\tmessage: `Render sentinel file does not exist.\nThe rendering UUID cannot be read safely.\nThere was probably an instance deployment during the render, and files were deleted.\n\nThe files generated in this render will not be published.`,\n};\n\nconst WriteToStoreError: ErrorData = {\n\terror: \"WriteToStoreError\",\n\tmessage: \"There was an error writing a file to the Store directory\",\n\thint: \"There may be an issue such as lack of space or permissions preventing the file from being written.\",\n};\n\nconst UploadSearchError: ErrorData = {\n\terror: \"UploadSearchError\",\n\tmessage: \"There was an error uploading content to API for search\",\n\thint: \"This happens if the API is currently not working or the credentials are incorrect.\",\n};\n\nconst CheckHealthError: ErrorData = {\n\terror: \"CheckHealthError\",\n\tmessage: \"There was a problem with environment vars configuration.\",\n\texpected: \"Some of the required environment variables are not set correctly or are missing\",\n\thint: \"Are the environment variables correctly set?\",\n};\n\nexport {\n\tArtifactError,\n\tCheckHealthError,\n\tErrorInSSGBuildProcess,\n\tLifecycleExecutionError,\n\tLoginError,\n\tNoDomainsFoundError,\n\tNoJSConfigFileFound,\n\tReadFromStoreError,\n\tReferenceFieldSourcesNotFoundError,\n\tRenderUUIDError,\n\tUploadSearchError,\n\tWriteToStoreError,\n\ttype ErrorsType,\n};\n", "const GRIDDO_API_URL = process.env.GRIDDO_API_URL;\nconst GRIDDO_PUBLIC_API_URL = process.env.GRIDDO_PUBLIC_API_URL;\n\nconst AI_EMBEDDINGS = `${GRIDDO_API_URL}/ai/embeddings`;\nconst ALERT = `${GRIDDO_PUBLIC_API_URL}/alert`;\nconst DOMAINS = `${GRIDDO_API_URL}/domains`;\nconst GET_ALL = `${GRIDDO_API_URL}/sites/all`;\nconst GET_PAGE = `${GRIDDO_API_URL}/page`;\nconst LOGIN = `${GRIDDO_API_URL}/login_check`;\nconst RESET_RENDER = `${GRIDDO_API_URL}/debug/reset-render`;\nconst ROBOTS = `${GRIDDO_API_URL}/domains/robots`;\nconst SEARCH = `${GRIDDO_API_URL}/search`;\nconst SETTINGS = `${GRIDDO_API_URL}/settings`;\n\n// Site\nconst SITE_URI = `${GRIDDO_API_URL}/site/`;\nconst BUILD_END = [SITE_URI, \"/build/end\"];\nconst BUILD_START = [SITE_URI, \"/build/start\"];\nconst GET_PAGES = [SITE_URI, \"/pages?pagination=false\"];\nconst GET_REFERENCE_FIELD_DATA = [SITE_URI, \"/distributor\"];\nconst GET_SITEMAP = [SITE_URI, \"/sitemap\"];\nconst INFO = [SITE_URI, \"/all\"];\nconst LANGUAGES = [SITE_URI, \"/languages\"];\nconst SOCIALS = [SITE_URI, \"/socials\"];\n\nexport {\n\tAI_EMBEDDINGS,\n\tALERT,\n\tBUILD_END,\n\tBUILD_START,\n\tDOMAINS,\n\tGET_ALL,\n\tGET_PAGE,\n\tGET_PAGES,\n\tGET_REFERENCE_FIELD_DATA,\n\tGET_SITEMAP,\n\tINFO,\n\tLANGUAGES,\n\tLOGIN,\n\tRESET_RENDER,\n\tROBOTS,\n\tSEARCH,\n\tSETTINGS,\n\tSOCIALS,\n};\n", "import type { AuthHeaders } from \"../shared/types/api\";\n\nimport { throwError } from \"../core/errors\";\nimport { LOGIN } from \"../shared/endpoints\";\nimport { GRIDDO_BOT_PASSWORD, GRIDDO_BOT_USER } from \"../shared/envs\";\nimport { LoginError } from \"../shared/errors\";\n\nclass AuthService {\n\theaders: AuthHeaders | undefined;\n\n\tasync login() {\n\t\ttry {\n\t\t\tconst response = await fetch(LOGIN, {\n\t\t\t\tmethod: \"POST\",\n\t\t\t\theaders: { \"Content-Type\": \"application/json\", Connection: \"close\" },\n\t\t\t\tbody: JSON.stringify({\n\t\t\t\t\tusername: GRIDDO_BOT_USER,\n\t\t\t\t\tpassword: GRIDDO_BOT_PASSWORD,\n\t\t\t\t}),\n\t\t\t});\n\n\t\t\tif (!response.ok) {\n\t\t\t\tthrow new Error(\"Error while login in the API\");\n\t\t\t}\n\n\t\t\tconst { token } = await response.json();\n\t\t\tthis.headers = {\n\t\t\t\tAuthorization: `bearer ${token}`,\n\t\t\t\t\"Cache-Control\": \"no-store\",\n\t\t\t};\n\n\t\t\treturn this.headers;\n\t\t} catch (e) {\n\t\t\tthrowError(LoginError, e);\n\t\t}\n\t}\n}\n\nconst authService = new AuthService();\n\nexport { authService as AuthService };\n", "import type { ErrorsType } from \"../shared/errors\";\n\nimport path from \"node:path\";\n\nimport { getRenderPathsHydratedWithDomainFromDB } from \"../services/render\";\nimport { brush } from \"../shared/npm-modules/brush\";\nimport { readDB, writeDB } from \"./db\";\nimport { distRollback } from \"./dist-rollback\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nexport type ErrorData = {\n\terror: ErrorsType;\n\tmessage: string;\n\texpected?: string;\n\thint?: string;\n};\n\nexport class RenderError extends Error {\n\tconstructor(originalError?: unknown) {\n\t\tsuper(originalError instanceof Error ? originalError.message : String(originalError));\n\n\t\tthis.name = \"InternalCXError\";\n\t\tthis.stack = originalError instanceof Error ? originalError.stack : \"\";\n\t}\n}\n\n/**\n * Throws an error with the provided error message, expected value, and hint.\n */\nfunction throwError(options: ErrorData, stack?: unknown): never {\n\tconst { error, message, expected, hint } = options;\n\n\tconst errorColor = GriddoLog.log(brush.red(`[ ${error} ]`));\n\tconst extraText = [expected, hint].filter(Boolean).join(\"\\n\");\n\n\tGriddoLog.log(`\n${errorColor}\n${message}\n${extraText}\n\n${brush.red(\"stack\")}\n${JSON.stringify(stack, null, 2)}`);\n\n\tthrow new RenderError(stack);\n}\n\nasync function withErrorHandler<T>(fn: () => Promise<T>) {\n\ttry {\n\t\tawait fn();\n\t} catch (error) {\n\t\tif (error instanceof RenderError) {\n\t\t\tGriddoLog.error(\"Internal Griddo RenderError\");\n\t\t} else if (error instanceof Error) {\n\t\t\tGriddoLog.error(error.message);\n\t\t} else {\n\t\t\tGriddoLog.error(`An unexpected error occurred ${error}`);\n\t\t}\n\n\t\ttry {\n\t\t\tconst { __root } = await getRenderPathsHydratedWithDomainFromDB();\n\t\t\tconst data = await readDB();\n\t\t\tif (data.needsRollbackOnError) {\n\t\t\t\tGriddoLog.info(\"Cleaning exports dir...\");\n\t\t\t\tGriddoLog.verbose(`Deleting ${path.join(__root, \"exports\")}...`);\n\n\t\t\t\tawait distRollback(data.currentRenderingDomain!);\n\t\t\t} else {\n\t\t\t\tGriddoLog.info(\"No rollback needed, skipping...\");\n\t\t\t}\n\t\t} catch (_e) {\n\t\t\tGriddoLog.info(\"Early render stage, no db.json created yet...\");\n\t\t}\n\n\t\tconst data = await readDB();\n\t\tdata.domains[data.currentRenderingDomain!].isRendering = false;\n\t\tdata.domains[data.currentRenderingDomain!].renderMode = \"ERROR\";\n\t\tawait writeDB(data);\n\t\tthrow error;\n\t}\n}\n\nexport { throwError, withErrorHandler };\n", "import type { RenderDB } from \"../shared/types/render\";\n\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport {\n\tGRIDDO_BUILD_LOGS,\n\tGRIDDO_BUILD_LOGS_BUFFER_SIZE,\n\t// GRIDDO_BUILD_LOGS_TO_FILE,\n} from \"../shared/envs\";\nimport { readDB } from \"./db\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nconst logBuffer: string[] = [];\nlet logFilePath: string | null = null;\nlet flushPromise: Promise<void> | null = null;\nconst LOG_BUFFER_FLUSH_THRESHOLD = GRIDDO_BUILD_LOGS_BUFFER_SIZE;\nconst LOG_FILENAME = \"render-detail-log.txt\";\nconst LOG_TO_FILE_IS_DISABLED = !GRIDDO_BUILD_LOGS; // || !GRIDDO_BUILD_LOGS_TO_FILE;\n\nlet dbData: RenderDB | null = null;\n\n// Universal data cache for this module...\nasync function getDBData() {\n\tif (!dbData) {\n\t\tdbData = await readDB();\n\t}\n\treturn dbData;\n}\n\nasync function initializeLogFile() {\n\tif (LOG_TO_FILE_IS_DISABLED) {\n\t\treturn;\n\t}\n\n\tconst data = await getDBData();\n\tlogFilePath = path.join(data.paths.root, LOG_FILENAME);\n\n\tawait fsp.rm(logFilePath, { force: true });\n}\n\nfunction addLogToBuffer(lineContent: string) {\n\tif (LOG_TO_FILE_IS_DISABLED) {\n\t\treturn;\n\t}\n\n\tlogBuffer.push(lineContent.toString());\n\n\tif (logBuffer.length >= LOG_BUFFER_FLUSH_THRESHOLD) {\n\t\t// Fire-and-forget flush to avoid blocking the main thread.\n\t\tflushLogsToFile().catch((error) => {\n\t\t\tGriddoLog.error(\"Background log flush failed:\", error);\n\t\t});\n\t}\n}\n\n/**\n * Vuelca el contenido del buffer de logs al archivo de logs en disco.\n * Si el buffer est\u00E1 vac\u00EDo, ya se est\u00E1 volcando, o no hay ruta de archivo, no hace nada.\n * Si ocurre un error al escribir, los logs se reinsertan al buffer para reintentar en el siguiente flush.\n */\nasync function flushLogsToFile() {\n\t// Si ya hay un volcado en curso, espera a que termine.\n\tif (flushPromise) {\n\t\tawait flushPromise;\n\t}\n\n\tif (logBuffer.length === 0 || !logFilePath) {\n\t\treturn;\n\t}\n\n\tconst performFlush = async () => {\n\t\tconst logsToFlush = [...logBuffer];\n\t\tlogBuffer.length = 0;\n\n\t\ttry {\n\t\t\tawait fsp.appendFile(logFilePath!, `${logsToFlush.join(\"\\n\")}\\n`);\n\t\t} catch (error) {\n\t\t\tlogBuffer.unshift(...logsToFlush);\n\t\t\tGriddoLog.error(\"Error flushing logs:\", error);\n\t\t\t// No relanzamos el error para no detener el \"fire-and-forget\"\n\t\t}\n\t};\n\n\tflushPromise = performFlush();\n\n\ttry {\n\t\tawait flushPromise;\n\t} finally {\n\t\tflushPromise = null;\n\t}\n}\n\n/**\n * Copia el archivo de log detallado de renderizado (\"render-detail-log.txt\")\n * desde el directorio ra\u00EDz del proyecto al directorio de logs del dominio en exports,\n * usando una marca de tiempo en el nombre de destino.\n * Si el archivo no existe, ignora el error salvo que sea distinto de ENOENT.\n *\n * @param domain - Nombre del dominio para el que se guarda el log.\n */\nasync function saveDetailRenderLog(domain: string) {\n\tawait flushLogsToFile();\n\n\tconst data = await getDBData();\n\n\tconst dateString = getFormattedDateTime();\n\tconst debugDir = path.join(data.paths.exportsDir, domain, \"logs\");\n\n\tawait fsp.mkdir(debugDir, { recursive: true });\n\n\tconst src = path.join(data.paths.root, LOG_FILENAME);\n\tconst dst = path.join(debugDir, `${dateString}-${LOG_FILENAME}`);\n\n\t// Move log to exports\n\ttry {\n\t\tawait fsp.cp(src, dst);\n\t} catch (error) {\n\t\t// It's possible the file doesn't exist if GRIDDO_BUILD_LOGS is false\n\t\t// or if no logs were ever added.\n\t\tif (error instanceof Error && \"code\" in error && error.code !== \"ENOENT\") {\n\t\t\tthrow error;\n\t\t}\n\t}\n}\n\nfunction getFormattedDateTime() {\n\tconst now = new Date();\n\tconst year = now.getFullYear();\n\tconst month = String(now.getMonth() + 1).padStart(2, \"0\");\n\tconst day = String(now.getDate()).padStart(2, \"0\");\n\n\tconst hours = String(now.getHours()).padStart(2, \"0\");\n\tconst minutes = String(now.getMinutes()).padStart(2, \"0\");\n\tconst seconds = String(now.getSeconds()).padStart(2, \"0\");\n\n\t// Formato recomendado para ordenaci\u00F3n de archivos: YYYY-MM-DD_HH-mm-ss\n\treturn `${year}-${month}-${day}_${hours}-${minutes}-${seconds}`;\n}\n\nexport { addLogToBuffer, flushLogsToFile, initializeLogFile, saveDetailRenderLog };\n", "import { GriddoLog } from \"../core/GriddoLog\";\nimport { post } from \"../services/api\";\nimport { AuthService } from \"../services/auth\";\nimport { RESET_RENDER } from \"../shared/endpoints\";\n\nasync function resetRender() {\n\tawait post({\n\t\tendpoint: RESET_RENDER,\n\t\tuseApiCacheDir: false,\n\t});\n}\n\nasync function main() {\n\tawait AuthService.login();\n\tawait resetRender();\n\tGriddoLog.info(\"The render status has been reset.\");\n}\n\nmain().catch(console.error);\n"],
|
|
5
|
+
"mappings": "oeAAA,GAAM,CAAE,IAAAA,CAAI,EAAI,QAKhB,SAASC,EAAYC,EAAyB,CAC7C,GAAI,CAACA,EAAO,MAAO,GAEnB,OAAQA,EAAM,KAAK,EAAE,YAAY,EAAG,CACnC,IAAK,IACL,IAAK,OACL,IAAK,MACL,IAAK,IACL,IAAK,KACJ,MAAO,GACR,QACC,MAAO,EACT,CACD,CAGA,IAAMC,GAAiBH,EAAI,gBAAkBA,EAAI,QAC3CI,GAAwBJ,EAAI,uBAAyBA,EAAI,eACzDK,EAAkBL,EAAI,UAAYA,EAAI,gBACtCM,EAAsBN,EAAI,aAAeA,EAAI,oBAG7CO,GAA+B,OAAO,SAASP,EAAI,8BAAgC,IAAI,EACvFQ,GAA2BP,EAAYD,EAAI,wBAAwB,EACnES,GAA2BR,EAAYD,EAAI,wBAAwB,EACnEU,EAAoBT,EAAYD,EAAI,iBAAiB,EACrDW,GAA4BV,EAAYD,EAAI,yBAAyB,EACrEY,EAAgC,OAAO,SAASZ,EAAI,+BAAiC,KAAK,EAC1Fa,GAA0BZ,EAAYD,EAAI,uBAAuB,EACjEc,GAAwBb,EAAYD,EAAI,qBAAqB,EAC7De,GAAsBf,EAAI,qBAAuBA,EAAI,aACrDgB,GAA4BhB,EAAI,2BAA6BA,EAAI,mBACjEiB,GAAuBhB,EAAYD,EAAI,oBAAoB,EAC3DkB,EAAsBjB,EAAYD,EAAI,mBAAmB,EACzDmB,GAAyBlB,EAAYD,EAAI,sBAAsB,EAC/DoB,GAA6BnB,EAAYD,EAAI,0BAA0B,ECjC7E,IAAMqB,GAAQ,UACRC,EAAQ,CACb,MAAO,WACP,IAAK,WACL,MAAO,WACP,OAAQ,WACR,KAAM,WACN,QAAS,WACT,KAAM,WACN,MAAO,WACP,KAAM,WACN,KAAM,UACN,IAAK,SACN,EAMMC,EAAQ,CAAC,EAEf,QAAWC,KAASF,EAAO,CAC1B,IAAMG,EAAMD,EACZD,EAAME,CAAG,EAAKC,GAA0B,GAAGJ,EAAMG,CAAG,CAAC,GAAGC,CAAI,GAAGL,EAAK,EACrE,CCxBA,IAAMM,EAAN,MAAMC,CAAU,CAEP,aAAc,CAAC,CAEvB,OAAc,WAAWC,EAAsB,CAC1CC,GACH,QAAQ,IAAIC,EAAM,OAAO,SAAS,EAAGA,EAAM,IAAIF,EAAI,KAAK,GAAG,CAAC,CAAC,CAE/D,CAEA,OAAc,SAASA,EAAsB,CACxCG,GACHJ,EAAU,IAAI,GAAGC,CAAG,CAEtB,CAEA,OAAc,QAAQA,EAAsB,CAC3C,QAAQ,IAAI,GAAGE,EAAM,KAAK,MAAM,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACrD,CAEA,OAAc,WAAWA,EAAsB,CAC9C,QAAQ,IAAI,GAAGE,EAAM,MAAM,SAAS,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACzD,CAEA,OAAc,SAASA,EAAsB,CAC5C,QAAQ,MAAM,GAAGE,EAAM,IAAI,OAAO,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACvD,CAEA,OAAc,QAAQA,EAAsB,CAC3C,QAAQ,KAAK,GAAGE,EAAM,OAAO,MAAM,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACxD,CAEA,OAAc,OAAOI,EAA4C,CAChE,QAAQ,IAAI,GAAGA,CAAI,CACpB,CACD,EChCA,IAAAC,GAAmB,0BACnBC,EAAgB,+BAChBC,EAAiB,wBCRjB,IAAAC,EAAiB,wBCFjB,IAAAC,EAAgB,+BAChBC,EAAiB,wBCHjB,IAAAC,EAAiB,wBCAjB,IAAAC,EAAe,sBAEf,IAAAC,EAAiB,wBACjBC,EAAoB,2BACpBC,EAA8B,oBA0BxBC,EAAUC,GACfA,aAAqB,OAAM,iBAAcA,CAAS,EAAIA,EA0ChD,SAASC,EAAWC,EAAcC,EAAmB,CAAC,EAAuB,CACnF,GAAM,CAAE,IAAAC,EAAM,EAAAC,QAAQ,IAAI,EAAG,KAAAC,EAAO,OAAQ,OAAQC,CAAa,EAAIJ,EAEjEK,EAAY,EAAAC,QAAK,QAAQC,EAAON,CAAG,GAAK,EAAE,EACxC,CAAE,KAAAO,CAAK,EAAI,EAAAF,QAAK,MAAMD,CAAS,EAC/BI,EAASL,EAAe,EAAAE,QAAK,QAAQD,EAAWE,EAAOH,CAAY,CAAE,EAAII,EACzEE,EAAiB,EAAAJ,QAAK,WAAWP,CAAI,EAE3C,OAAa,CACZ,IAAMY,EAAWD,EAAiBX,EAAO,EAAAO,QAAK,KAAKD,EAAWN,CAAI,EAClE,GAAI,CACH,IAAMa,EAAQ,EAAAC,QAAG,SAASF,EAAU,CAAE,eAAgB,EAAM,CAAC,EAC7D,GAAKR,IAAS,QAAUS,GAAO,OAAO,GAAOT,IAAS,aAAeS,GAAO,YAAY,EACvF,OAAOD,CAET,MAAQ,CAGR,CAEA,GAAIN,IAAcI,GAAUJ,IAAcG,EACzC,MAGDH,EAAY,EAAAC,QAAK,QAAQD,CAAS,CACnC,CACD,CDzFA,SAASS,EAAWC,EAAqC,CACxD,GAAM,CAAE,IAAAC,CAAI,EAAID,GAAW,CAAC,EACtBE,EAAWC,EAAW,eAAgB,CAAE,IAAAF,CAAI,CAAC,EACnD,OAAOC,GAAY,EAAAE,QAAK,QAAQF,CAAQ,CACzC,CDNA,IAAMG,GAAOC,EAAW,CAAE,IAAK,EAAAC,QAAK,QAAQ,UAAW,UAAU,CAAE,CAAC,GAAK,GACnEC,GAAQ,EAAAD,QAAK,KAAKF,GAAM,eAAe,EACvCI,GAAa,EAAAF,QAAK,KAAKC,GAAO,SAAS,EAE7C,eAAeE,EAAOC,EAAe,GAAI,CACxC,IAAMC,EAAOD,GAAgBF,GAC7B,GAAI,CACH,OAAO,KAAK,MAAM,MAAM,EAAAI,QAAI,SAASD,EAAM,OAAO,CAAC,CACpD,OAASE,EAAO,CACf,MAAAC,EAAU,MAAM,6BAA6BH,CAAI,IAAKE,CAAK,EACrDA,CACP,CACD,CGlBA,IAAAE,EAAgB,+BC4ChB,IAAMC,EAAwB,CAC7B,MAAO,aACP,QAAS,4CACT,SAAU,oFACX,ED0QA,eAAeC,EAAWC,EAAa,CACtC,GAAI,CACH,aAAM,EAAAC,QAAI,OAAOD,CAAG,EACb,EACR,MAAQ,CACP,MAAO,EACR,CACD,CEnUA,IAAME,EAAiB,QAAQ,IAAI,eAC7BC,GAAwB,QAAQ,IAAI,sBAEpCC,GAAgB,GAAGF,CAAc,iBACjCG,GAAQ,GAAGF,EAAqB,SAChCG,GAAU,GAAGJ,CAAc,WAC3BK,GAAU,GAAGL,CAAc,aAC3BM,GAAW,GAAGN,CAAc,QAC5BO,GAAQ,GAAGP,CAAc,eACzBQ,GAAe,GAAGR,CAAc,sBAChCS,GAAS,GAAGT,CAAc,kBAC1BU,GAAS,GAAGV,CAAc,UAC1BW,GAAW,GAAGX,CAAc,YAG5BY,GAAW,GAAGZ,CAAc,SCRlC,IAAMa,EAAN,KAAkB,CACjB,QAEA,MAAM,OAAQ,CACb,GAAI,CACH,IAAMC,EAAW,MAAM,MAAMC,GAAO,CACnC,OAAQ,OACR,QAAS,CAAE,eAAgB,mBAAoB,WAAY,OAAQ,EACnE,KAAM,KAAK,UAAU,CACpB,SAAUC,EACV,SAAUC,CACX,CAAC,CACF,CAAC,EAED,GAAI,CAACH,EAAS,GACb,MAAM,IAAI,MAAM,8BAA8B,EAG/C,GAAM,CAAE,MAAAI,CAAM,EAAI,MAAMJ,EAAS,KAAK,EACtC,YAAK,QAAU,CACd,cAAe,UAAUI,CAAK,GAC9B,gBAAiB,UAClB,EAEO,KAAK,OACb,OAAS,EAAG,CACXC,EAAWC,EAAY,CAAC,CACzB,CACD,CACD,EAEMC,EAAc,IAAIR,EP2HxB,eAAeS,EAAuCC,EAGnD,CACF,GAAM,CAAE,OAAAC,EAAQ,WAAAC,CAAW,EAAIF,GAAW,CAAC,EAGrCG,GADK,MAAMC,EAAOF,CAAU,GACjB,MAEjB,MAAO,CACN,OAAQC,EAAM,KACd,QAAS,EAAAE,QAAK,KAAKF,EAAM,QAASF,GAAU,EAAE,EAC9C,aAAcE,EAAM,WACpB,KAAMA,EAAM,GACZ,QAASA,EAAM,WACf,UAAW,EAAAE,QAAK,KAAKF,EAAM,WAAYF,GAAU,EAAE,EACnD,iBAAkB,EAAAI,QAAK,KAAKF,EAAM,iBAAkBF,GAAU,EAAE,EAChE,MAAOE,EAAM,IACb,eAAgB,EAAAE,QAAK,KAAKF,EAAM,WAAYF,GAAU,GAAI,MAAM,CACjE,CACD,CQpKO,IAAMK,EAAN,cAA0B,KAAM,CACtC,YAAYC,EAAyB,CACpC,MAAMA,aAAyB,MAAQA,EAAc,QAAU,OAAOA,CAAa,CAAC,EAEpF,KAAK,KAAO,kBACZ,KAAK,MAAQA,aAAyB,MAAQA,EAAc,MAAQ,EACrE,CACD,EAKA,SAASC,EAAWC,EAAoBC,EAAwB,CAC/D,GAAM,CAAE,MAAAC,EAAO,QAAAC,EAAS,SAAAC,EAAU,KAAAC,CAAK,EAAIL,EAErCM,EAAaC,EAAU,IAAIC,EAAM,IAAI,KAAKN,CAAK,IAAI,CAAC,EACpDO,EAAY,CAACL,EAAUC,CAAI,EAAE,OAAO,OAAO,EAAE,KAAK;AAAA,CAAI,EAE5D,MAAAE,EAAU,IAAI;AAAA,EACbD,CAAU;AAAA,EACVH,CAAO;AAAA,EACPM,CAAS;AAAA;AAAA,EAETD,EAAM,IAAI,OAAO,CAAC;AAAA,EAClB,KAAK,UAAUP,EAAO,KAAM,CAAC,CAAC,EAAE,EAE3B,IAAIJ,EAAYI,CAAK,CAC5B,CC1CA,IAAAS,GAAgB,+BAWhB,IAAMC,EAAsB,CAAC,EACzBC,GAA6B,KAC7BC,EAAqC,KACnCC,GAA6BC,EAEnC,IAAMC,GAA0B,CAACC,EAuBjC,SAASC,EAAeC,EAAqB,CACxCC,KAIJC,EAAU,KAAKF,EAAY,SAAS,CAAC,EAEjCE,EAAU,QAAUC,IAEvBC,GAAgB,EAAE,MAAOC,GAAU,CAClCC,EAAU,MAAM,+BAAgCD,CAAK,CACtD,CAAC,EAEH,CAOA,eAAeD,IAAkB,CAMhC,GAJIG,GACH,MAAMA,EAGHL,EAAU,SAAW,GAAK,CAACM,GAC9B,OAgBDD,GAbqB,SAAY,CAChC,IAAME,EAAc,CAAC,GAAGP,CAAS,EACjCA,EAAU,OAAS,EAEnB,GAAI,CACH,MAAM,GAAAQ,QAAI,WAAWF,GAAc,GAAGC,EAAY,KAAK;AAAA,CAAI,CAAC;AAAA,CAAI,CACjE,OAASJ,EAAO,CACfH,EAAU,QAAQ,GAAGO,CAAW,EAChCH,EAAU,MAAM,uBAAwBD,CAAK,CAE9C,CACD,GAE4B,EAE5B,GAAI,CACH,MAAME,CACP,QAAE,CACDA,EAAe,IAChB,CACD,CVpEA,GAAM,CAAE,IAAAI,EAAI,EAAI,QACV,CAAE,mBAAAC,GAAqB,IAAK,eAAAC,GAAiB,GAAI,EAAIF,GAe3D,eAAeG,GACdC,EACAC,EACAC,EAAc,GACD,CACb,GAAM,CACL,SAAAC,EACA,KAAAC,EACA,SAAAC,EAAW,GACX,QAAAC,EAAU,EACV,QAAAC,EACA,eAAAC,EAAiB,GACjB,UAAAC,EAAY,EACb,EAAIT,EACEU,EAAe,CAAE,SAAAP,EAAU,KAAAC,EAAM,QAAAG,EAAS,SAAAF,CAAS,EAGzD,GAAIA,GAAYG,EAAgB,CAC/B,IAAMG,EAAQ,IAAI,KACZC,EAAY,MAAMC,GAAmBH,CAAY,EAEvD,GAAIE,EAAW,CACd,GAAIH,EAAW,CACd,IAAMK,EAASC,GAAcH,CAAS,EAChCI,EAAYF,EAAS,SAASA,CAAM,GAAK,GACzCG,EAAWC,GAAQ,KAAK,IAAI,EAAIP,EAAM,QAAQ,CAAC,EACrDQ,EAAe,GAAGlB,CAAM,YAAYe,CAAS,IAAIb,CAAQ,MAAMc,CAAQ,KAAKf,CAAW,EAAE,CAC1F,CACA,OAAOU,CACR,CACD,CAGA,GAAI,CACH,IAAMD,EAAQ,IAAI,KAGZS,EAA4B,CACjC,OAAQnB,EAAO,YAAY,EAC3B,QAAS,OAAO,OAAO,CAAC,EAAGM,EAASc,EAAY,OAAO,CACxD,EAGIpB,EAAO,YAAY,IAAM,OAASG,IACrCgB,EAAa,KAAO,KAAK,UAAUhB,CAAI,EAClCgB,EAAa,UAASA,EAAa,QAAU,CAAC,GAClDA,EAAa,QAAmC,cAAc,EAAI,oBAGpE,IAAME,EAAW,MAAM,MAAMnB,EAAUiB,CAAY,EAGnD,GAAI,CAACE,EAAS,GAAI,CACjB,GAAIA,EAAS,SAAW,IAEvB,OAAO,KAGR,MAAM,IAAI,MAAM,QAAQA,EAAS,MAAM,KAAKA,EAAS,UAAU,EAAE,CAClE,CAEA,IAAMC,EAAU,MAAMD,EAAS,KAAK,EAEpC,GAAIb,EAAW,CACd,IAAMK,EAASC,GAAcQ,CAAI,EAC3BP,GAAYF,EAAS,SAASA,CAAM,GAAK,GACzCG,GAAWC,GAAQ,KAAK,IAAI,EAAIP,EAAM,QAAQ,CAAC,EACrDQ,EAAe,GAAGlB,CAAM,YAAYe,EAAS,IAAIb,CAAQ,MAAMc,EAAQ,KAAKf,CAAW,EAAE,CAC1F,CAEA,OAAIM,GACH,MAAMgB,GAAUd,EAAca,CAAI,EAG5BA,CACR,OAASE,EAAG,CACX,IAAMC,EAAQD,EAEd,GAAInB,EAAU,SAASR,EAAc,EACpC,MAAA6B,EAAU,IAAI;AAAA,eACF7B,EAAc;AAAA;AAAA,IAEzBG,EAAO,YAAY,CAAC,IAAIE,CAAQ;AAAA,UAC1B,KAAK,UAAUC,CAAI,CAAC;AAAA,aACjB,KAAK,UAAUG,CAAO,CAAC;AAAA,WACzBmB,EAAM,OAAO;AAAA;AAAA,CAEvB,EACQ,IAAIE,EAAYF,CAAK,EAG5B,OAAAG,GAAaH,EAAO,CACnB,SAAU,CAAE,SAAAvB,EAAU,KAAAC,CAAK,CAC5B,CAAC,EAEDuB,EAAU,KAAK,sBAAsB1B,CAAM,GAAIE,CAAQ,EAEvD,MAAM2B,GAAM,SAASjC,EAAkB,EAAI,GAAI,EAExCE,GACN,CACC,SAAAI,EACA,KAAAC,EACA,QAAAG,EACA,SAAAF,EACA,QAASC,EAAU,CACpB,EACAL,EACAC,CACD,CACD,CACD,CA4BA,eAAe6B,GAAgCC,EAAgB,CAC9D,GAAM,CAAE,SAAAC,EAAU,KAAAC,EAAM,QAAAC,CAAQ,EAAIH,EAC9BI,EACLH,EAAS,SAAS,cAAc,GAChC,0BAA0B,KAAK,UAAUC,CAAI,CAAC,UAAU,KAAK,UAAUC,GAAS,IAAI,CAAC,GAEtF,OAAOE,GAAcL,EAAO,OAAQI,GAA4B,EAAE,CACnE,CAKA,SAASE,GAAaC,EAAcC,EAA8B,CACjE,GAAM,CAAE,QAAAC,EAAS,MAAAC,CAAM,EAAIH,EACrB,CAAE,SAAAI,CAAS,EAAIH,EACfI,EAAgB,CAAC,EAEvB,QAAWC,KAAQ,OAAO,KAAKF,CAAQ,EACtCC,EAAc,KACb,GAAGC,CAAI,KACN,OAAOF,EAASE,CAAI,GAAM,SAAW,KAAK,UAAUF,EAASE,CAAI,CAAC,EAAIF,EAASE,CAAI,CACpF,EACD,EAID,IAAMC,EAAcF,EAAc,KAAK;AAAA,CAAI,EACrCG,EAAkB,GAAGN,CAAO;AAAA,EAAKC,CAAK,GAG5CM,EAAU,KACTC,EAAM,IAAI;AAAA;AAAA;AAAA;AAAA,EAIVH,CAAW;AAAA;AAAA;AAAA,EAGXC,CAAe;AAAA;AAAA;AAAA,CAGhB,CACA,CACD,CAMA,SAASG,GAAcC,EAAwB,CAC9C,GAAI,SAAOA,GAAa,UAAYA,IAAa,MAAQ,MAAM,QAAQA,CAAQ,GAI/E,MAAO,SAAUA,GAAYA,EAAS,KAAOA,EAAS,KAAO,MAC9D,CAOA,SAASC,GAAMC,EAA2B,CACzC,OAAO,IAAI,QAASC,GAAQ,WAAWA,EAAKD,CAAE,CAAC,CAChD,CASA,SAASE,GAAQF,EAAYG,EAAW,EAAW,CAClD,OAAO,OAAO,YAAYH,EAAK,KAAM,QAAQG,CAAQ,CAAC,CACvD,CAMA,eAAeC,GAAyBC,EAAoB,CAC3D,GAAM,CAAE,OAAAC,CAAO,EAAI,MAAMC,EAAuC,EAC1DC,EAAc,EAAAC,QAAK,KAAKH,EAAQ,UAAU,EAE1CI,EAAU,GAAAC,QAAO,WAAW,QAAQ,EAC1C,OAAAD,EAAQ,OAAO,KAAK,UAAUL,CAAQ,CAAC,EAEhC,GAAGG,CAAW,IAAIE,EAAQ,OAAO,KAAK,CAAC,EAC/C,CAQA,eAAeE,GAAaP,EAAoBQ,EAAY,CAC3D,IAAMC,EAAgB,OAAOD,GAAY,SAAWA,EAAU,KAAK,UAAUA,CAAO,EAC9EE,EAAW,MAAMX,GAAyBC,CAAQ,EAClDW,EAAW,EAAAP,QAAK,QAAQM,CAAQ,EAEhC,MAAME,EAAWD,CAAQ,GAC9B,MAAM,EAAAE,QAAI,MAAMF,EAAU,CAAE,UAAW,EAAK,CAAC,EAG9C,MAAM,EAAAE,QAAI,UAAUH,EAAUD,EAAe,MAAM,CACpD,CAQA,eAAeK,GAAmBd,EAAoB,CACrD,GAAI,CACH,IAAMe,EAAO,MAAMhB,GAAyBC,CAAQ,EAC9CgB,EAAc,MAAM,EAAAH,QAAI,SAASE,EAAM,MAAM,EAGnD,OAFiB,KAAK,MAAMC,CAAW,CAGxC,MAAQ,CACP,OAAO,IACR,CACD,CW1SA,eAAeC,IAAc,CAC5B,MAAMC,GAAK,CACV,SAAUC,GACV,eAAgB,EACjB,CAAC,CACF,CAEA,eAAeC,IAAO,CACrB,MAAMC,EAAY,MAAM,EACxB,MAAMJ,GAAY,EAClBK,EAAU,KAAK,mCAAmC,CACnD,CAEAF,GAAK,EAAE,MAAM,QAAQ,KAAK",
|
|
6
|
+
"names": ["env", "envIsTruthy", "value", "GRIDDO_API_URL", "GRIDDO_PUBLIC_API_URL", "GRIDDO_BOT_USER", "GRIDDO_BOT_PASSWORD", "GRIDDO_API_CONCURRENCY_COUNT", "GRIDDO_SKIP_BUILD_CHECKS", "GRIDDO_RENDER_BY_DOMAINS", "GRIDDO_BUILD_LOGS", "GRIDDO_BUILD_LOGS_TO_FILE", "GRIDDO_BUILD_LOGS_BUFFER_SIZE", "GRIDDO_SSG_VERBOSE_LOGS", "GRIDDO_SEARCH_FEATURE", "GRIDDO_ASSET_PREFIX", "GRIDDO_REACT_APP_INSTANCE", "GRIDDO_AI_EMBEDDINGS", "GRIDDO_VERBOSE_LOGS", "GRIDDO_USE_DIST_BACKUP", "GRIDDO_SSG_BUNDLE_ANALYZER", "RESET", "CODES", "brush", "color", "key", "text", "GriddoLog", "_GriddoLog", "str", "GRIDDO_VERBOSE_LOGS", "brush", "GRIDDO_BUILD_LOGS", "args", "import_node_crypto", "import_promises", "import_node_path", "import_node_path", "import_promises", "import_node_path", "import_node_path", "import_node_fs", "import_node_path", "import_node_process", "import_node_url", "toPath", "urlOrPath", "findUpSync", "name", "options", "cwd", "process", "type", "stopAtOption", "directory", "path", "toPath", "root", "stopAt", "isAbsoluteName", "filePath", "stats", "fs", "pkgDirSync", "options", "cwd", "filePath", "findUpSync", "path", "root", "pkgDirSync", "path", "cache", "dbFilePath", "readDB", "customDBPath", "file", "fsp", "error", "GriddoLog", "import_promises", "LoginError", "pathExists", "dir", "fsp", "GRIDDO_API_URL", "GRIDDO_PUBLIC_API_URL", "AI_EMBEDDINGS", "ALERT", "DOMAINS", "GET_ALL", "GET_PAGE", "LOGIN", "RESET_RENDER", "ROBOTS", "SEARCH", "SETTINGS", "SITE_URI", "AuthService", "response", "LOGIN", "GRIDDO_BOT_USER", "GRIDDO_BOT_PASSWORD", "token", "throwError", "LoginError", "authService", "getRenderPathsHydratedWithDomainFromDB", "options", "domain", "dbFilePath", "paths", "readDB", "path", "RenderError", "originalError", "throwError", "options", "stack", "error", "message", "expected", "hint", "errorColor", "GriddoLog", "brush", "extraText", "import_promises", "logBuffer", "logFilePath", "flushPromise", "LOG_BUFFER_FLUSH_THRESHOLD", "GRIDDO_BUILD_LOGS_BUFFER_SIZE", "LOG_TO_FILE_IS_DISABLED", "GRIDDO_BUILD_LOGS", "addLogToBuffer", "lineContent", "LOG_TO_FILE_IS_DISABLED", "logBuffer", "LOG_BUFFER_FLUSH_THRESHOLD", "flushLogsToFile", "error", "GriddoLog", "flushPromise", "logFilePath", "logsToFlush", "fsp", "env", "RETRY_WAIT_SECONDS", "RETRY_ATTEMPTS", "requestAPI", "props", "method", "appendToLog", "endpoint", "body", "cacheKey", "attempt", "headers", "useApiCacheDir", "logToFile", "cacheOptions", "start", "cacheData", "searchCacheData", "siteId", "getSafeSiteId", "siteIdMsg", "duration", "msToSec", "addLogToBuffer", "fetchOptions", "authService", "response", "data", "saveCache", "e", "error", "GriddoLog", "RenderError", "showApiError", "delay", "postApi", "props", "endpoint", "body", "headers", "referenceFieldBodyParams", "requestAPI", "showApiError", "error", "options", "message", "stack", "callInfo", "callInfoArray", "item", "callInfoStr", "errorDetailsStr", "GriddoLog", "brush", "getSafeSiteId", "response", "delay", "ms", "res", "msToSec", "decimals", "generateFilenameWithHash", "petition", "__root", "getRenderPathsHydratedWithDomainFromDB", "apiCacheDir", "path", "hashSum", "crypto", "saveCache", "content", "stringContent", "filename", "filepath", "pathExists", "fsp", "searchCacheData", "file", "fileContent", "resetRender", "postApi", "RESET_RENDER", "main", "authService", "GriddoLog"]
|
|
7
|
+
}
|
|
@@ -0,0 +1,31 @@
|
|
|
1
|
+
"use strict";var _e=Object.create;var M=Object.defineProperty;var Ie=Object.getOwnPropertyDescriptor;var Se=Object.getOwnPropertyNames;var be=Object.getPrototypeOf,Oe=Object.prototype.hasOwnProperty;var Ae=(r,e,t,n)=>{if(e&&typeof e=="object"||typeof e=="function")for(let o of Se(e))!Oe.call(r,o)&&o!==t&&M(r,o,{get:()=>e[o],enumerable:!(n=Ie(e,o))||n.enumerable});return r};var c=(r,e,t)=>(t=r!=null?_e(be(r)):{},Ae(e||!r||!r.__esModule?M(t,"default",{value:r,enumerable:!0}):t,r));var de=c(require("node:path"));var P=c(require("node:path"));var L=c(require("node:fs/promises")),C=c(require("node:path"));var V=c(require("node:path"));var J=c(require("node:fs"));var w=c(require("node:path")),W=c(require("node:process")),Y=require("node:url"),H=r=>r instanceof URL?(0,Y.fileURLToPath)(r):r;function K(r,e={}){let{cwd:t=W.default.cwd(),type:n="file",stopAt:o}=e,a=w.default.resolve(H(t)??""),{root:u}=w.default.parse(a),p=o?w.default.resolve(a,H(o)):u,g=w.default.isAbsolute(r);for(;;){let I=g?r:w.default.join(a,r);try{let S=J.default.statSync(I,{throwIfNoEntry:!1});if(n==="file"&&S?.isFile()||n==="directory"&&S?.isDirectory())return I}catch{}if(a===p||a===u)break;a=w.default.dirname(a)}}function q(r){let{cwd:e}=r||{},t=K("package.json",{cwd:e});return t&&V.default.dirname(t)}var{env:s}=process;function l(r){if(!r)return!1;switch(r.trim().toLowerCase()){case"1":case"true":case"yes":case"y":case"on":return!0;default:return!1}}var Ve=s.GRIDDO_API_URL||s.API_URL,qe=s.GRIDDO_PUBLIC_API_URL||s.PUBLIC_API_URL,Z=s.botEmail||s.GRIDDO_BOT_USER,z=s.botPassword||s.GRIDDO_BOT_PASSWORD,Ze=Number.parseInt(s.GRIDDO_API_CONCURRENCY_COUNT||"10"),ze=l(s.GRIDDO_SKIP_BUILD_CHECKS),Xe=l(s.GRIDDO_RENDER_BY_DOMAINS),F=l(s.GRIDDO_BUILD_LOGS),Qe=l(s.GRIDDO_BUILD_LOGS_TO_FILE),X=Number.parseInt(s.GRIDDO_BUILD_LOGS_BUFFER_SIZE||"500"),er=l(s.GRIDDO_SSG_VERBOSE_LOGS),Q=l(s.GRIDDO_SEARCH_FEATURE),rr=s.GRIDDO_ASSET_PREFIX||s.ASSET_PREFIX,tr=s.GRIDDO_REACT_APP_INSTANCE||s.REACT_APP_INSTANCE,ee=l(s.GRIDDO_AI_EMBEDDINGS),re=l(s.GRIDDO_VERBOSE_LOGS),or=l(s.GRIDDO_USE_DIST_BACKUP),nr=l(s.GRIDDO_SSG_BUNDLE_ANALYZER);var Te="\x1B[0m",te={black:"\x1B[30m",red:"\x1B[31m",green:"\x1B[32m",yellow:"\x1B[33m",blue:"\x1B[34m",magenta:"\x1B[35m",cyan:"\x1B[36m",white:"\x1B[37m",gray:"\x1B[90m",bold:"\x1B[1m",dim:"\x1B[2m"},d={};for(let r in te){let e=r;d[e]=t=>`${te[e]}${t}${Te}`}var i=class r{constructor(){}static verbose(...e){re&&console.log(d.yellow("verbose"),d.dim(e.join(" ")))}static build(...e){F&&r.log(...e)}static info(...e){console.log(`${d.blue("info")} ${e.join(" ")}`)}static success(...e){console.log(`${d.green("success")} ${e.join(" ")}`)}static error(...e){console.error(`${d.red("error")} ${e.join(" ")}`)}static warn(...e){console.warn(`${d.yellow("warn")} ${e.join(" ")}`)}static log(...e){console.log(...e)}};var $e=q({cwd:C.default.resolve(__dirname,"../../..")})||"",Pe=C.default.join($e,".griddo/cache"),oe=C.default.join(Pe,"db.json");async function h(r=""){let e=r||oe;try{return JSON.parse(await L.default.readFile(e,"utf-8"))}catch(t){throw i.error(`Failed to read DB file at ${e}:`,t),t}}async function G(r,e=""){let t=e||oe;try{await L.default.writeFile(t,JSON.stringify(r,null," "))}catch(n){throw i.error(`Failed to write DB file at ${t}:`,n),n}}var ie=c(require("node:fs/promises"));var ne={error:"LoginError",message:"There was a problem logging in to the API",expected:"This happens if the API is currently not working or the credentials are incorrect."};async function E(r){try{return await ie.default.access(r),!0}catch{return!1}}var m=process.env.GRIDDO_API_URL,Be=process.env.GRIDDO_PUBLIC_API_URL,se=`${m}/ai/embeddings`,Er=`${Be}/alert`,Rr=`${m}/domains`,_r=`${m}/sites/all`,Ir=`${m}/page`,ae=`${m}/login_check`,Sr=`${m}/debug/reset-render`,br=`${m}/domains/robots`,Or=`${m}/search`,Ar=`${m}/settings`,Tr=`${m}/site/`;var v=class{headers;async login(){try{let e=await fetch(ae,{method:"POST",headers:{"Content-Type":"application/json",Connection:"close"},body:JSON.stringify({username:Z,password:z})});if(!e.ok)throw new Error("Error while login in the API");let{token:t}=await e.json();return this.headers={Authorization:`bearer ${t}`,"Cache-Control":"no-store"},this.headers}catch(e){A(ne,e)}}},T=new v;async function $(r){let{domain:e,dbFilePath:t}=r||{},o=(await h(t)).paths;return{__root:o.root,__cache:P.default.join(o.cxCache,e||""),__components:o.components,__cx:o.cx,__sites:o.exportsDir,__exports:P.default.join(o.exportsDir,e||""),__exports_backup:P.default.join(o.exportsDirBackup,e||""),__ssg:o.ssg,__exports_dist:P.default.join(o.exportsDir,e||"","dist")}}var k=c(require("node:fs/promises")),y=c(require("node:path"));async function ce(r){let e=await h(),{exportsDir:t,exportsDirBackup:n}=e.paths;i.info(`Cleaning exports dir for the domain ${r}`),i.verbose(`Deleting ${y.default.join(t,r)}...`),await k.default.rm(y.default.join(t,r),{recursive:!0,force:!0}),await E(y.default.join(n,r))?(await k.default.cp(y.default.join(n,r),y.default.join(t,r),{recursive:!0}),i.info(`export-backup dir for the domain ${r} found. Restoring before exit...`),i.verbose(`Copying ${y.default.join(n,r)} -> ${y.default.join(t,r)}...`)):i.info("No export-backup found, skipping rollback. Next render will create a new exports dir from scratch...")}var R=class extends Error{constructor(e){super(e instanceof Error?e.message:String(e)),this.name="InternalCXError",this.stack=e instanceof Error?e.stack:""}};function A(r,e){let{error:t,message:n,expected:o,hint:a}=r,u=i.log(d.red(`[ ${t} ]`)),p=[o,a].filter(Boolean).join(`
|
|
2
|
+
`);throw i.log(`
|
|
3
|
+
${u}
|
|
4
|
+
${n}
|
|
5
|
+
${p}
|
|
6
|
+
|
|
7
|
+
${d.red("stack")}
|
|
8
|
+
${JSON.stringify(e,null,2)}`),new R(e)}async function fe(r){try{await r()}catch(e){e instanceof R?i.error("Internal Griddo RenderError"):e instanceof Error?i.error(e.message):i.error(`An unexpected error occurred ${e}`);try{let{__root:n}=await $(),o=await h();o.needsRollbackOnError?(i.info("Cleaning exports dir..."),i.verbose(`Deleting ${de.default.join(n,"exports")}...`),await ce(o.currentRenderingDomain)):i.info("No rollback needed, skipping...")}catch{i.info("Early render stage, no db.json created yet...")}let t=await h();throw t.domains[t.currentRenderingDomain].isRendering=!1,t.domains[t.currentRenderingDomain].renderMode="ERROR",await G(t),e}}var he=c(require("node:crypto")),N=c(require("node:fs/promises")),j=c(require("node:path"));var le=c(require("node:fs/promises"));var _=[],pe=null,B=null,xe=X;var Fe=!F;function U(r){Fe||(_.push(r.toString()),_.length>=xe&&Ce().catch(e=>{i.error("Background log flush failed:",e)}))}async function Ce(){if(B&&await B,_.length===0||!pe)return;B=(async()=>{let e=[..._];_.length=0;try{await le.default.appendFile(pe,`${e.join(`
|
|
9
|
+
`)}
|
|
10
|
+
`)}catch(t){_.unshift(...e),i.error("Error flushing logs:",t)}})();try{await B}finally{B=null}}var{env:Ge}=process,{RETRY_WAIT_SECONDS:Ne="4",RETRY_ATTEMPTS:me="4"}=Ge;async function De(r,e,t=""){let{endpoint:n,body:o,cacheKey:a="",attempt:u=1,headers:p,useApiCacheDir:g=!0,logToFile:I=!0}=r,S={endpoint:n,body:o,headers:p,cacheKey:a};if(a&&g){let b=new Date,f=await Ue(S);if(f){if(I){let D=ue(f),O=D?`site: ${D}`:"",x=ge(Date.now()-b.getTime());U(`${e} (cache) ${O} ${n} - ${x}s ${t}`)}return f}}try{let b=new Date,f={method:e.toUpperCase(),headers:Object.assign({},p,T.headers)};e.toLowerCase()!=="get"&&o&&(f.body=JSON.stringify(o),f.headers||(f.headers={}),f.headers["Content-Type"]="application/json");let D=await fetch(n,f);if(!D.ok){if(D.status===404)return null;throw new Error(`HTTP ${D.status}: ${D.statusText}`)}let O=await D.json();if(I){let x=ue(O),Ee=x?`site: ${x}`:"",Re=ge(Date.now()-b.getTime());U(`${e} (fetch) ${Ee} ${n} - ${Re}s ${t}`)}return g&&await ke(S,O),O}catch(b){let f=b;if(u>parseInt(me))throw i.log(`
|
|
11
|
+
Max attempts ${me} reached
|
|
12
|
+
--------------------------------------
|
|
13
|
+
- ${e.toUpperCase()} ${n}
|
|
14
|
+
- BODY: ${JSON.stringify(o)}
|
|
15
|
+
- HEADERS: ${JSON.stringify(p)}
|
|
16
|
+
- ERROR: ${f.message}
|
|
17
|
+
--------------------------------------
|
|
18
|
+
`),new R(f);return Le(f,{callInfo:{endpoint:n,body:o}}),i.warn(`Waiting for retry: ${e}`,n),await ve(parseInt(Ne)*1e3),De({endpoint:n,body:o,headers:p,cacheKey:a,attempt:u+1},e,t)}}async function ye(r){let{endpoint:e,body:t,headers:n}=r,o=e.endsWith("/distributor")&&`# ReferenceField body: ${JSON.stringify(t)} lang: ${JSON.stringify(n?.lang)}`;return De(r,"post",o||"")}function Le(r,e){let{message:t,stack:n}=r,{callInfo:o}=e,a=[];for(let g of Object.keys(o))a.push(`${g}: ${typeof o[g]=="object"?JSON.stringify(o[g]):o[g]}`);let u=a.join(`
|
|
19
|
+
`),p=`${t}
|
|
20
|
+
${n}`;i.warn(d.red(`
|
|
21
|
+
=============
|
|
22
|
+
|
|
23
|
+
{ Call info }
|
|
24
|
+
${u}
|
|
25
|
+
|
|
26
|
+
{ Error details }
|
|
27
|
+
${p}
|
|
28
|
+
|
|
29
|
+
=============
|
|
30
|
+
`))}function ue(r){if(!(typeof r!="object"||r===null||Array.isArray(r)))return"site"in r&&r.site?r.site:void 0}function ve(r){return new Promise(e=>setTimeout(e,r))}function ge(r,e=3){return Number.parseFloat((r/1e3).toFixed(e))}async function we(r){let{__root:e}=await $(),t=j.default.join(e,"apiCache"),n=he.default.createHash("sha256");return n.update(JSON.stringify(r)),`${t}/${n.digest("hex")}`}async function ke(r,e){let t=typeof e=="string"?e:JSON.stringify(e),n=await we(r),o=j.default.dirname(n);await E(o)||await N.default.mkdir(o,{recursive:!0}),await N.default.writeFile(n,t,"utf8")}async function Ue(r){try{let e=await we(r),t=await N.default.readFile(e,"utf8");return JSON.parse(t)}catch{return null}}async function je(){await ye({endpoint:se,useApiCacheDir:!1})}async function Me(){Q&&ee&&(await T.login(),await je())}fe(Me);
|
|
31
|
+
//# sourceMappingURL=start-embeddings.js.map
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
{
|
|
2
|
+
"version": 3,
|
|
3
|
+
"sources": ["../../exporter/core/errors.ts", "../../exporter/services/render.ts", "../../exporter/core/db.ts", "../../exporter/shared/npm-modules/pkg-dir.ts", "../../exporter/shared/npm-modules/find-up-simple.ts", "../../exporter/shared/envs.ts", "../../exporter/shared/npm-modules/brush.ts", "../../exporter/core/GriddoLog.ts", "../../exporter/core/fs.ts", "../../exporter/shared/errors.ts", "../../exporter/shared/endpoints.ts", "../../exporter/services/auth.ts", "../../exporter/core/dist-rollback.ts", "../../exporter/services/api.ts", "../../exporter/core/logger.ts", "../../exporter/commands/start-embeddings.ts"],
|
|
4
|
+
"sourcesContent": ["import type { ErrorsType } from \"../shared/errors\";\n\nimport path from \"node:path\";\n\nimport { getRenderPathsHydratedWithDomainFromDB } from \"../services/render\";\nimport { brush } from \"../shared/npm-modules/brush\";\nimport { readDB, writeDB } from \"./db\";\nimport { distRollback } from \"./dist-rollback\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nexport type ErrorData = {\n\terror: ErrorsType;\n\tmessage: string;\n\texpected?: string;\n\thint?: string;\n};\n\nexport class RenderError extends Error {\n\tconstructor(originalError?: unknown) {\n\t\tsuper(originalError instanceof Error ? originalError.message : String(originalError));\n\n\t\tthis.name = \"InternalCXError\";\n\t\tthis.stack = originalError instanceof Error ? originalError.stack : \"\";\n\t}\n}\n\n/**\n * Throws an error with the provided error message, expected value, and hint.\n */\nfunction throwError(options: ErrorData, stack?: unknown): never {\n\tconst { error, message, expected, hint } = options;\n\n\tconst errorColor = GriddoLog.log(brush.red(`[ ${error} ]`));\n\tconst extraText = [expected, hint].filter(Boolean).join(\"\\n\");\n\n\tGriddoLog.log(`\n${errorColor}\n${message}\n${extraText}\n\n${brush.red(\"stack\")}\n${JSON.stringify(stack, null, 2)}`);\n\n\tthrow new RenderError(stack);\n}\n\nasync function withErrorHandler<T>(fn: () => Promise<T>) {\n\ttry {\n\t\tawait fn();\n\t} catch (error) {\n\t\tif (error instanceof RenderError) {\n\t\t\tGriddoLog.error(\"Internal Griddo RenderError\");\n\t\t} else if (error instanceof Error) {\n\t\t\tGriddoLog.error(error.message);\n\t\t} else {\n\t\t\tGriddoLog.error(`An unexpected error occurred ${error}`);\n\t\t}\n\n\t\ttry {\n\t\t\tconst { __root } = await getRenderPathsHydratedWithDomainFromDB();\n\t\t\tconst data = await readDB();\n\t\t\tif (data.needsRollbackOnError) {\n\t\t\t\tGriddoLog.info(\"Cleaning exports dir...\");\n\t\t\t\tGriddoLog.verbose(`Deleting ${path.join(__root, \"exports\")}...`);\n\n\t\t\t\tawait distRollback(data.currentRenderingDomain!);\n\t\t\t} else {\n\t\t\t\tGriddoLog.info(\"No rollback needed, skipping...\");\n\t\t\t}\n\t\t} catch (_e) {\n\t\t\tGriddoLog.info(\"Early render stage, no db.json created yet...\");\n\t\t}\n\n\t\tconst data = await readDB();\n\t\tdata.domains[data.currentRenderingDomain!].isRendering = false;\n\t\tdata.domains[data.currentRenderingDomain!].renderMode = \"ERROR\";\n\t\tawait writeDB(data);\n\t\tthrow error;\n\t}\n}\n\nexport { throwError, withErrorHandler };\n", "import type { RenderModeTuple } from \"../shared/types/render\";\n\nimport { execSync } from \"node:child_process\";\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { readDB, writeDB } from \"../core/db\";\nimport { throwError } from \"../core/errors\";\nimport { pathExists } from \"../core/fs\";\nimport { GriddoLog } from \"../core/GriddoLog\";\nimport { RenderUUIDError } from \"../shared/errors\";\nimport { brush } from \"../shared/npm-modules/brush\";\nimport { RENDER_MODE } from \"../shared/types/render\";\nimport { AuthService } from \"./auth\";\nimport { getBuildMetadata } from \"./manage-store\";\n\n/**\n * Creates a sentinel file with the current date and time.\n * This file is used to track later if node_modules/@griddo/cx was cleaned by a\n * npm install coming from a deploy.\n */\nasync function markRenderAsStarted(options: { domain: string; basePath: string }) {\n\tconst { domain } = options;\n\n\tconst db = await readDB();\n\tdb.domains[domain].isRendering = true;\n\tawait writeDB(db);\n\n\t// Creamos un archivo centinela, si al terminar el render este archivo no\n\t// existe es que ha habido un deploy por medio y hay que invalidar el render\n\tconst { __ssg } = await getRenderPathsHydratedWithDomainFromDB();\n\n\tconst renderSentinelFile = path.join(__ssg, `.render-sentinel-${domain}`);\n\tawait fsp.writeFile(renderSentinelFile, new Date().toISOString());\n}\n\nasync function markRenderAsCompleted(domain: string) {\n\tconst db = await readDB();\n\tdb.domains[domain].isRendering = false;\n\tdb.currentRenderingDomain = null;\n\tdb.domains[domain].renderMode = \"COMPLETED\";\n\t// db.domains[domain].shouldBeRendered = false;\n\tawait writeDB(db);\n\n\t// Borramos finalmente el archivo centinela\n\tconst { __ssg } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst renderSentinelFile = path.join(__ssg, `.render-sentinel-${domain}`);\n\tawait fsp.unlink(renderSentinelFile);\n}\n\nasync function assertRenderIsValid(domain: string) {\n\t// Comprobamos que .render-sentinel exista, si no es que un deploy lo borro\n\t// y hay que invalidar el render.\n\tconst { __ssg } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst renderSentinelFile = path.join(__ssg, `.render-sentinel-${domain}`);\n\tif (!(await pathExists(renderSentinelFile))) {\n\t\tthrowError(RenderUUIDError);\n\t}\n}\n\n/**\n * Determines the appropriate render mode for a given domain based on its current state,\n * previous render errors, deployment status, and whether rendering is required.\n *\n * @param options - An object containing:\n * - `domain`: The domain name to resolve the render mode for.\n * - `shouldBeRendered`: A boolean indicating if the domain should be rendered.\n * @returns An object with:\n * - `renderMode`: The resolved render mode (`FROM_SCRATCH`, `INCREMENTAL`, or `IDLE`).\n * - `reason`: A string describing the reason for the chosen render mode.\n *\n * @remarks\n * The function checks for missing exports, previous render errors, new deployments,\n * and whether rendering is necessary to decide the render mode.\n *\n * @todo\n * Improve ifs and reason concatenations...\n */\nasync function resolveDomainRenderMode(options: { domain: string; shouldBeRendered: boolean }) {\n\tconst { domain, shouldBeRendered } = options;\n\n\tconst db = await readDB();\n\n\tconst { __cache, __exports } = await getRenderPathsHydratedWithDomainFromDB({ domain });\n\tconst exportsAreMissing = !(await pathExists(path.join(__exports)));\n\tconst previousRenderFailed = db.domains[domain]?.isRendering;\n\tconst newDeployDetected = await hasNewCommit(__cache);\n\n\tif (exportsAreMissing) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.FROM_SCRATCH,\n\t\t\treason: \"missing exports directory\",\n\t\t};\n\t}\n\n\tif (previousRenderFailed) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.FROM_SCRATCH,\n\t\t\treason: \"error in previous render\",\n\t\t};\n\t}\n\n\tif (newDeployDetected) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.FROM_SCRATCH,\n\t\t\treason: \"new commit hash\",\n\t\t};\n\t}\n\n\tif (!shouldBeRendered) {\n\t\treturn {\n\t\t\trenderMode: RENDER_MODE.IDLE,\n\t\t\treason: \"no activity\",\n\t\t};\n\t}\n\n\treturn {\n\t\trenderMode: RENDER_MODE.INCREMENTAL,\n\t\treason: \"has changes\",\n\t};\n}\n\nasync function hasNewCommit(basePath: string): Promise<boolean> {\n\tconst commitFile = path.join(basePath, \"commit\");\n\tconst currentCommit = execSync(\"git rev-parse HEAD\").toString().trim();\n\n\tif (await pathExists(commitFile)) {\n\t\tconst savedCommit = (await fsp.readFile(commitFile, \"utf-8\")).trim();\n\t\tif (savedCommit === currentCommit) {\n\t\t\treturn false; // No hay nuevo commit\n\t\t}\n\n\t\treturn true;\n\t}\n\n\treturn true;\n}\n\nasync function updateCommitFile(options: { basePath: string }) {\n\tconst { basePath } = options;\n\tconst currentCommit = execSync(\"git rev-parse HEAD\").toString().trim();\n\tawait fsp.writeFile(path.join(basePath, \"commit\"), currentCommit);\n}\n\nasync function getRenderModeFromDB(domain: string): Promise<RenderModeTuple> {\n\tconst db = await readDB();\n\n\tif (!db.domains[domain]) {\n\t\tthrow new Error(brush.red(`[!] Error: Domain ${domain} not found in DB`));\n\t}\n\n\tif (!db.domains[domain].renderMode) {\n\t\tthrow new Error(brush.red(`[!] Error: Render mode not found for domain ${domain}`));\n\t}\n\n\treturn {\n\t\trenderMode: db.domains[domain].renderMode,\n\t\treason: db.domains[domain].renderModeReason,\n\t};\n}\n\nasync function getRenderPathsHydratedWithDomainFromDB(options?: {\n\tdomain?: string;\n\tdbFilePath?: string;\n}) {\n\tconst { domain, dbFilePath } = options || {};\n\n\tconst db = await readDB(dbFilePath);\n\tconst paths = db.paths;\n\n\treturn {\n\t\t__root: paths.root,\n\t\t__cache: path.join(paths.cxCache, domain || \"\"),\n\t\t__components: paths.components,\n\t\t__cx: paths.cx,\n\t\t__sites: paths.exportsDir,\n\t\t__exports: path.join(paths.exportsDir, domain || \"\"),\n\t\t__exports_backup: path.join(paths.exportsDirBackup, domain || \"\"),\n\t\t__ssg: paths.ssg,\n\t\t__exports_dist: path.join(paths.exportsDir, domain || \"\", \"dist\"),\n\t};\n}\n\nasync function getRenderMetadataFromDB() {\n\tconst db = await readDB();\n\treturn {\n\t\tgriddoVersion: db.griddoVersion,\n\t\tbuildReportFileName: db.buildReportFileName,\n\t};\n}\n\n/**\n * Save a file with the end of build process to use as `end-render` signal.\n */\nasync function generateBuildReport(domain: string) {\n\tconst authControl = await AuthService.login();\n\n\tconst { __root } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst { buildReportFileName } = await getRenderMetadataFromDB();\n\tconst { buildProcessData } = await getBuildMetadata(domain);\n\n\tconst buildSitesInfo = Object.keys(buildProcessData).map((siteID) => ({\n\t\t...buildProcessData[siteID],\n\t\tsiteId: Number.parseInt(siteID),\n\t}));\n\n\tconst report = {\n\t\tauthControl,\n\t\tsites: buildSitesInfo,\n\t};\n\n\tconst reportFilePath = path.join(__root, \"current-dist\", buildReportFileName);\n\n\tawait fsp.writeFile(reportFilePath, JSON.stringify(report));\n\n\tGriddoLog.verbose(`build report saved in ${reportFilePath}`);\n}\n\nexport {\n\tassertRenderIsValid,\n\tgenerateBuildReport,\n\tgetRenderMetadataFromDB,\n\tgetRenderModeFromDB,\n\tgetRenderPathsHydratedWithDomainFromDB,\n\tmarkRenderAsCompleted,\n\tmarkRenderAsStarted,\n\tresolveDomainRenderMode,\n\tupdateCommitFile,\n};\n", "import type { RenderDB } from \"../shared/types/render\";\n\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { pkgDirSync } from \"../shared/npm-modules/pkg-dir\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nconst root = pkgDirSync({ cwd: path.resolve(__dirname, \"../../..\") }) || \"\";\nconst cache = path.join(root, \".griddo/cache\");\nconst dbFilePath = path.join(cache, \"db.json\");\n\nasync function readDB(customDBPath = \"\") {\n\tconst file = customDBPath || dbFilePath;\n\ttry {\n\t\treturn JSON.parse(await fsp.readFile(file, \"utf-8\")) as RenderDB;\n\t} catch (error) {\n\t\tGriddoLog.error(`Failed to read DB file at ${file}:`, error);\n\t\tthrow error;\n\t}\n}\n\nasync function writeDB(renderDB: RenderDB, customDBPath = \"\") {\n\tconst file = customDBPath || dbFilePath;\n\ttry {\n\t\tawait fsp.writeFile(file, JSON.stringify(renderDB, null, \"\\t\"));\n\t} catch (error) {\n\t\tGriddoLog.error(`Failed to write DB file at ${file}:`, error);\n\t\tthrow error;\n\t}\n}\n\nexport { readDB, writeDB };\n", "import path from \"node:path\";\n\nimport { findUp, findUpSync } from \"./find-up-simple\";\n\nasync function pkgDir(options?: { readonly cwd?: string }) {\n\tconst { cwd } = options || {};\n\tconst filePath = await findUp(\"package.json\", { cwd });\n\treturn filePath && path.dirname(filePath);\n}\n\nfunction pkgDirSync(options?: { readonly cwd?: string }) {\n\tconst { cwd } = options || {};\n\tconst filePath = findUpSync(\"package.json\", { cwd });\n\treturn filePath && path.dirname(filePath);\n}\n\nexport { pkgDir, pkgDirSync };\n", "import fs from \"node:fs\";\nimport fsPromises from \"node:fs/promises\";\nimport path from \"node:path\";\nimport process from \"node:process\";\nimport { fileURLToPath } from \"node:url\";\n\n/**\n * Opciones para controlar el comportamiento de la b\u00FAsqueda.\n */\nexport type Options = {\n\t/**\n\tEl directorio desde donde empezar a buscar.\n\t@default process.cwd()\n\t*/\n\treadonly cwd?: URL | string;\n\n\t/**\n\tEl tipo de ruta a buscar.\n\t@default 'file'\n\t*/\n\treadonly type?: \"file\" | \"directory\";\n\n\t/**\n\tUn directorio en el que la b\u00FAsqueda se detiene si no se encuentran coincidencias.\n\t@default El directorio ra\u00EDz del sistema\n\t*/\n\treadonly stopAt?: URL | string;\n};\n\n// Funci\u00F3n auxiliar para convertir una URL en una ruta de archivo de tipo string.\nconst toPath = (urlOrPath: URL | string | undefined): string | undefined =>\n\turlOrPath instanceof URL ? fileURLToPath(urlOrPath) : urlOrPath;\n\n/**\n * Encuentra un archivo o directorio de forma as\u00EDncrona subiendo por los directorios padre.\n * @param name - El nombre del archivo o directorio a buscar.\n * @param options - Opciones de b\u00FAsqueda.\n * @returns La ruta encontrada o `undefined` si no se pudo encontrar.\n */\nexport async function findUp(name: string, options: Options = {}): Promise<string | undefined> {\n\tconst { cwd = process.cwd(), type = \"file\", stopAt: stopAtOption } = options;\n\n\tlet directory = path.resolve(toPath(cwd) ?? \"\");\n\tconst { root } = path.parse(directory);\n\tconst stopAt = stopAtOption ? path.resolve(directory, toPath(stopAtOption)!) : root;\n\tconst isAbsoluteName = path.isAbsolute(name);\n\n\twhile (true) {\n\t\tconst filePath = isAbsoluteName ? name : path.join(directory, name);\n\t\ttry {\n\t\t\t// eslint-disable-next-line no-await-in-loop\n\t\t\tconst stats = await fsPromises.stat(filePath);\n\t\t\tif ((type === \"file\" && stats.isFile()) || (type === \"directory\" && stats.isDirectory())) {\n\t\t\t\treturn filePath;\n\t\t\t}\n\t\t} catch {\n\t\t\t// Ignora errores (ej. el archivo no existe) y contin\u00FAa la b\u00FAsqueda.\n\t\t}\n\n\t\tif (directory === stopAt || directory === root) {\n\t\t\tbreak;\n\t\t}\n\n\t\tdirectory = path.dirname(directory);\n\t}\n}\n\n/**\n * Encuentra un archivo o directorio de forma s\u00EDncrona subiendo por los directorios padre.\n * @param name - El nombre del archivo o directorio a buscar.\n * @param options - Opciones de b\u00FAsqueda.\n * @returns La ruta encontrada o `undefined` si no se pudo encontrar.\n */\nexport function findUpSync(name: string, options: Options = {}): string | undefined {\n\tconst { cwd = process.cwd(), type = \"file\", stopAt: stopAtOption } = options;\n\n\tlet directory = path.resolve(toPath(cwd) ?? \"\");\n\tconst { root } = path.parse(directory);\n\tconst stopAt = stopAtOption ? path.resolve(directory, toPath(stopAtOption)!) : root;\n\tconst isAbsoluteName = path.isAbsolute(name);\n\n\twhile (true) {\n\t\tconst filePath = isAbsoluteName ? name : path.join(directory, name);\n\t\ttry {\n\t\t\tconst stats = fs.statSync(filePath, { throwIfNoEntry: false });\n\t\t\tif ((type === \"file\" && stats?.isFile()) || (type === \"directory\" && stats?.isDirectory())) {\n\t\t\t\treturn filePath;\n\t\t\t}\n\t\t} catch {\n\t\t\t// En teor\u00EDa, statSync con `throwIfNoEntry: false` no deber\u00EDa lanzar un error,\n\t\t\t// pero se mantiene por seguridad.\n\t\t}\n\n\t\tif (directory === stopAt || directory === root) {\n\t\t\tbreak;\n\t\t}\n\n\t\tdirectory = path.dirname(directory);\n\t}\n}\n", "const { env } = process;\n\n/**\n * Returns true/false from string\n */\nfunction envIsTruthy(value?: string): boolean {\n\tif (!value) return false;\n\n\tswitch (value.trim().toLowerCase()) {\n\t\tcase \"1\":\n\t\tcase \"true\":\n\t\tcase \"yes\":\n\t\tcase \"y\":\n\t\tcase \"on\":\n\t\t\treturn true;\n\t\tdefault:\n\t\t\treturn false;\n\t}\n}\n\n// Credentials\nconst GRIDDO_API_URL = env.GRIDDO_API_URL || env.API_URL;\nconst GRIDDO_PUBLIC_API_URL = env.GRIDDO_PUBLIC_API_URL || env.PUBLIC_API_URL;\nconst GRIDDO_BOT_USER = env.botEmail || env.GRIDDO_BOT_USER;\nconst GRIDDO_BOT_PASSWORD = env.botPassword || env.GRIDDO_BOT_PASSWORD;\n\n// Rendering\nconst GRIDDO_API_CONCURRENCY_COUNT = Number.parseInt(env.GRIDDO_API_CONCURRENCY_COUNT || \"10\");\nconst GRIDDO_SKIP_BUILD_CHECKS = envIsTruthy(env.GRIDDO_SKIP_BUILD_CHECKS);\nconst GRIDDO_RENDER_BY_DOMAINS = envIsTruthy(env.GRIDDO_RENDER_BY_DOMAINS);\nconst GRIDDO_BUILD_LOGS = envIsTruthy(env.GRIDDO_BUILD_LOGS);\nconst GRIDDO_BUILD_LOGS_TO_FILE = envIsTruthy(env.GRIDDO_BUILD_LOGS_TO_FILE);\nconst GRIDDO_BUILD_LOGS_BUFFER_SIZE = Number.parseInt(env.GRIDDO_BUILD_LOGS_BUFFER_SIZE || \"500\");\nconst GRIDDO_SSG_VERBOSE_LOGS = envIsTruthy(env.GRIDDO_SSG_VERBOSE_LOGS);\nconst GRIDDO_SEARCH_FEATURE = envIsTruthy(env.GRIDDO_SEARCH_FEATURE);\nconst GRIDDO_ASSET_PREFIX = env.GRIDDO_ASSET_PREFIX || env.ASSET_PREFIX;\nconst GRIDDO_REACT_APP_INSTANCE = env.GRIDDO_REACT_APP_INSTANCE || env.REACT_APP_INSTANCE;\nconst GRIDDO_AI_EMBEDDINGS = envIsTruthy(env.GRIDDO_AI_EMBEDDINGS);\nconst GRIDDO_VERBOSE_LOGS = envIsTruthy(env.GRIDDO_VERBOSE_LOGS);\nconst GRIDDO_USE_DIST_BACKUP = envIsTruthy(env.GRIDDO_USE_DIST_BACKUP);\nconst GRIDDO_SSG_BUNDLE_ANALYZER = envIsTruthy(env.GRIDDO_SSG_BUNDLE_ANALYZER);\n\nexport {\n\tGRIDDO_AI_EMBEDDINGS,\n\tGRIDDO_API_CONCURRENCY_COUNT,\n\tGRIDDO_API_URL,\n\tGRIDDO_ASSET_PREFIX,\n\tGRIDDO_BOT_PASSWORD,\n\tGRIDDO_BOT_USER,\n\tGRIDDO_BUILD_LOGS,\n\tGRIDDO_BUILD_LOGS_BUFFER_SIZE,\n\tGRIDDO_BUILD_LOGS_TO_FILE,\n\tGRIDDO_PUBLIC_API_URL,\n\tGRIDDO_REACT_APP_INSTANCE,\n\tGRIDDO_RENDER_BY_DOMAINS,\n\tGRIDDO_SEARCH_FEATURE,\n\tGRIDDO_SKIP_BUILD_CHECKS,\n\tGRIDDO_SSG_BUNDLE_ANALYZER,\n\tGRIDDO_SSG_VERBOSE_LOGS,\n\tGRIDDO_USE_DIST_BACKUP,\n\tGRIDDO_VERBOSE_LOGS,\n};\n", "//\n// Brush adds color to a string|number, it does not print it!\n// Its simple, no log, no chains, just color in a string|number\n// usage:\n// console.log(brush.green(\"sucess!\"))\n//\n\nconst RESET = \"\\x1b[0m\";\nconst CODES = {\n\tblack: \"\\x1b[30m\",\n\tred: \"\\x1b[31m\",\n\tgreen: \"\\x1b[32m\",\n\tyellow: \"\\x1b[33m\",\n\tblue: \"\\x1b[34m\",\n\tmagenta: \"\\x1b[35m\",\n\tcyan: \"\\x1b[36m\",\n\twhite: \"\\x1b[37m\",\n\tgray: \"\\x1b[90m\",\n\tbold: \"\\x1b[1m\",\n\tdim: \"\\x1b[2m\",\n} as const;\n\ntype ColorFunction = (text: string | number) => string;\ntype ColorName = keyof typeof CODES;\ntype Brush = Record<ColorName, ColorFunction>;\n\nconst brush = {} as Brush;\n\nfor (const color in CODES) {\n\tconst key = color as ColorName;\n\tbrush[key] = (text: string | number) => `${CODES[key]}${text}${RESET}`;\n}\n\nexport { brush };\n", "import { GRIDDO_BUILD_LOGS, GRIDDO_VERBOSE_LOGS } from \"../shared/envs\";\nimport { brush } from \"../shared/npm-modules/brush\";\n\n/**\n * Clase est\u00E1tica para gestionar los logs de la aplicaci\u00F3n.\n * No se puede instanciar, se usa directamente: GriddoLogs.info(\"mensaje\").\n */\nclass GriddoLog {\n\t/** El constructor es privado para prevenir la instanciaci\u00F3n de la clase. */\n\tprivate constructor() {}\n\n\tpublic static verbose(...str: unknown[]): void {\n\t\tif (GRIDDO_VERBOSE_LOGS) {\n\t\t\tconsole.log(brush.yellow(\"verbose\"), brush.dim(str.join(\" \")));\n\t\t}\n\t}\n\n\tpublic static build(...str: unknown[]): void {\n\t\tif (GRIDDO_BUILD_LOGS) {\n\t\t\tGriddoLog.log(...str);\n\t\t}\n\t}\n\n\tpublic static info(...str: unknown[]): void {\n\t\tconsole.log(`${brush.blue(\"info\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static success(...str: unknown[]): void {\n\t\tconsole.log(`${brush.green(\"success\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static error(...str: unknown[]): void {\n\t\tconsole.error(`${brush.red(\"error\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static warn(...str: unknown[]): void {\n\t\tconsole.warn(`${brush.yellow(\"warn\")} ${str.join(\" \")}`);\n\t}\n\n\tpublic static log(...args: Parameters<typeof console.log>): void {\n\t\tconsole.log(...args);\n\t}\n}\n\nexport { GriddoLog };\n", "import type { MakeDirectoryOptions } from \"node:fs\";\n\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { ArtifactError } from \"../shared/errors\";\nimport { throwError } from \"./errors\";\nimport { GriddoLog } from \"./GriddoLog\";\n\n/**\n * Remove an empty directory from the basePath recursively.\n * If the directory has only .xml files it will handle as empty too (empty site)\n *\n * @param baseDir - The base directory.\n */\nasync function deleteDisposableSiteDirs(baseDir: string) {\n\tif (!(await pathExists(baseDir))) {\n\t\treturn;\n\t}\n\n\tconst sitesDirs = (await fsp.readdir(baseDir, { withFileTypes: true })).filter((file) =>\n\t\tfile.isDirectory(),\n\t);\n\n\tfor (const siteDir of sitesDirs) {\n\t\tconst sitePath = path.join(baseDir, siteDir.name);\n\t\tif (await siteIsEmpty(sitePath)) {\n\t\t\tawait fsp.rm(sitePath, { recursive: true });\n\t\t}\n\t}\n}\n\n/**\n * Creates multiple artifact directories.\n *\n * @param dirs - An array of directory paths.\n * @param options - Same option as `fs.mkdirSync()`\n */\nasync function mkDirs(dirs: string[], options?: MakeDirectoryOptions) {\n\tfor (const dir of dirs) {\n\t\ttry {\n\t\t\tif (!(await pathExists(dir))) {\n\t\t\t\tawait fsp.mkdir(dir, { recursive: true, ...options });\n\t\t\t\tGriddoLog.verbose(`create directory: ${dir}`);\n\t\t\t}\n\t\t} catch (error) {\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\nasync function renamePath(src: string, dst: string) {\n\ttry {\n\t\tif (await pathExists(src)) {\n\t\t\tawait fsp.rename(src, dst);\n\t\t\tGriddoLog.verbose(`rename ${src} to ${dst}`);\n\t\t}\n\t} catch (error) {\n\t\tthrowError(ArtifactError, error);\n\t}\n}\n\n/**\n * Copy multiple directories with backup option.\n *\n * @param src - Source directory.\n * @param dst - Destination directory.\n * @param dirs - Directories to copy.\n * @param options.withBackup - Create a previous backup before copy.\n */\nasync function cpDirs(\n\tsrc: string,\n\tdst: string,\n\tdirs: string[],\n\toptions = {\n\t\twithBackup: false,\n\t},\n) {\n\tconst { withBackup } = options;\n\tfor (const dir of dirs) {\n\t\tconst srcCompose = path.join(src, dir);\n\t\tconst dstCompose = path.join(dst, dir);\n\n\t\t// The dir we want to copy, doesn't exist.\n\t\tif (!(await pathExists(srcCompose))) {\n\t\t\tGriddoLog.info(`(Maybe first render) Source directory does not exist: ${srcCompose}`);\n\t\t\tcontinue;\n\t\t}\n\n\t\t// Create the backup\n\t\tif (withBackup) {\n\t\t\tawait createBackup(dstCompose);\n\t\t\tGriddoLog.verbose(`create backup: ${dstCompose}`);\n\t\t}\n\n\t\t// Copy directory\n\t\ttry {\n\t\t\t// First clean destination\n\t\t\tif (await pathExists(dstCompose)) {\n\t\t\t\tawait fsp.rm(dstCompose, { recursive: true, force: true });\n\t\t\t\tGriddoLog.verbose(`clean destination: ${dstCompose}`);\n\t\t\t}\n\n\t\t\t// Then copy src to dst\n\t\t\tawait fsp.cp(srcCompose, dstCompose, {\n\t\t\t\trecursive: true,\n\t\t\t\tpreserveTimestamps: true,\n\t\t\t});\n\t\t\tGriddoLog.verbose(`copy: ${srcCompose} to ${dstCompose}`);\n\n\t\t\tif (withBackup) {\n\t\t\t\tawait deleteBackup(dstCompose);\n\t\t\t\tGriddoLog.verbose(`delete backup: ${dstCompose}`);\n\t\t\t}\n\t\t} catch (error) {\n\t\t\tif (withBackup) {\n\t\t\t\tawait restoreBackup(dstCompose);\n\t\t\t\tGriddoLog.verbose(\"Backup has been restored.\");\n\t\t\t}\n\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\n/**\n * Move artifacts between cx-paths\n *\n * @param src - Source directory.\n * @param dst - Destination directory.\n * @param dirs - Directories to move.\n * @param options - Options.\n */\nasync function mvDirs(\n\tsrc: string,\n\tdst: string,\n\tdirs: string[],\n\toptions?: { withBackup?: boolean; override?: boolean },\n) {\n\tconst { override, withBackup } = options || {};\n\n\tfor (const dir of dirs) {\n\t\tconst srcCompose = path.join(src, dir);\n\t\tconst dstCompose = path.join(dst, dir);\n\n\t\tif (!(await pathExists(srcCompose))) {\n\t\t\tcontinue;\n\t\t}\n\n\t\tif (withBackup) {\n\t\t\tawait createBackup(dstCompose);\n\t\t}\n\n\t\ttry {\n\t\t\t// Clean destination\n\t\t\tif (override && (await pathExists(dstCompose))) {\n\t\t\t\tawait fsp.rm(dstCompose, { recursive: true, force: true });\n\t\t\t}\n\n\t\t\tawait fsp.rename(srcCompose, dstCompose);\n\t\t\tGriddoLog.verbose(`moved: ${srcCompose} to ${dstCompose}`);\n\n\t\t\tif (withBackup) {\n\t\t\t\tawait deleteBackup(dstCompose);\n\t\t\t}\n\t\t} catch (error) {\n\t\t\tif (withBackup) {\n\t\t\t\tawait restoreBackup(dstCompose);\n\t\t\t\tGriddoLog.info(\"Backup has been restored.\");\n\t\t\t}\n\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\n/**\n * Removes multiple artifact directories.\n *\n * @param dirs - An array of directory paths.\n */\nasync function rmDirs(dirs: string[]) {\n\tfor (const dir of dirs) {\n\t\ttry {\n\t\t\tawait fsp.rm(dir, { recursive: true, force: true });\n\t\t\tGriddoLog.verbose(`artifact removed: ${dir}`);\n\t\t} catch (error) {\n\t\t\tthrowError(ArtifactError, error);\n\t\t}\n\t}\n}\n\nasync function restoreBackup(src: string, suffix = \"-BACKUP\") {\n\tconst dst = src + suffix;\n\ttry {\n\t\tawait fsp.rename(dst, src);\n\t\tGriddoLog.info(`Backup ${dst} has been restored`);\n\t} catch (_error) {\n\t\tthrow new Error(`Error while delete ${dst} backup`);\n\t}\n}\n\nasync function deleteBackup(src: string, suffix = \"-BACKUP\") {\n\tconst dst = src + suffix;\n\n\tif (!(await pathExists(dst))) {\n\t\treturn;\n\t}\n\n\ttry {\n\t\tawait fsp.rm(dst, { recursive: true, force: true });\n\t\tGriddoLog.verbose(`Backup ${dst} has been deleted`);\n\t} catch (_error) {\n\t\tthrow new Error(`Error while delete ${dst} backup`);\n\t}\n}\n\nasync function createBackup(src: string, suffix = \"-BACKUP\") {\n\tconst dst = src + suffix;\n\n\tif (!(await pathExists(src))) {\n\t\treturn;\n\t}\n\n\tif (await pathExists(dst)) {\n\t\tGriddoLog.warn(`Destination ${dst} already exists`);\n\t\treturn;\n\t}\n\n\ttry {\n\t\tawait fsp.rename(src, dst);\n\t\tGriddoLog.verbose(`Backup of ${src} has been created in ${dst}`);\n\t} catch (error) {\n\t\tGriddoLog.error(`Error while coping ${src} to ${dst} backup`);\n\t\tthrowError(ArtifactError, error);\n\t}\n}\n\n/**\n * Return true if the site folder is empty or only has xml files. (Recursively)\n */\nasync function siteIsEmpty(sitePath: string) {\n\tconst siteFiles = (\n\t\tawait fsp.readdir(sitePath, {\n\t\t\twithFileTypes: true,\n\t\t\trecursive: true,\n\t\t})\n\t).filter((file) => file.isFile() && !path.basename(file.name).startsWith(\".\"));\n\n\tconst xmlFiles = siteFiles.filter((file) => file.name.endsWith(\".xml\"));\n\n\tif (siteFiles.length === xmlFiles.length) {\n\t\treturn true;\n\t}\n}\n\n/**\n * Delete empty directories from the given directory in a recursive way.\n */\nasync function deleteEmptyDirectories(dirPath: string) {\n\ttry {\n\t\tconst stats = await fsp.stat(dirPath);\n\n\t\t// Si no es un directorio, no hacemos nada\n\t\tif (!stats.isDirectory()) {\n\t\t\treturn;\n\t\t}\n\n\t\tlet filesInDirectory: string[];\n\t\ttry {\n\t\t\tfilesInDirectory = await fsp.readdir(dirPath);\n\t\t} catch (err: any) {\n\t\t\t// Si el directorio no existe o no se puede leer (ej. permisos), lo saltamos.\n\t\t\tif (err.code === \"ENOENT\") {\n\t\t\t\tGriddoLog.warn(`The directory \"${dirPath}\" does not exist, skipping it.`);\n\t\t\t\treturn;\n\t\t\t}\n\t\t\tGriddoLog.error(`Error al leer el directorio \"${dirPath}\":`, err);\n\t\t\tthrow err; // Re-lanza el error para que sea manejado por el llamador\n\t\t}\n\n\t\t// Recorrer los contenidos del directorio\n\t\tfor (const file of filesInDirectory) {\n\t\t\tconst fullPath = path.join(dirPath, file);\n\t\t\tawait deleteEmptyDirectories(fullPath); // Llamada recursiva s\u00EDncrona\n\t\t}\n\n\t\t// Despu\u00E9s de procesar todos los subdirectorios, verifica si el directorio actual est\u00E1 vac\u00EDo\n\t\tconst remainingFiles = await fsp.readdir(dirPath);\n\n\t\tif (remainingFiles.length === 0) {\n\t\t\ttry {\n\t\t\t\tawait fsp.rmdir(dirPath);\n\t\t\t\tGriddoLog.verbose(`Remove empty directory: ${dirPath}`);\n\t\t\t} catch (err: any) {\n\t\t\t\t// Puede que haya habido un problema de concurrencia o permisos\n\t\t\t\tif (err.code === \"ENOENT\") {\n\t\t\t\t\tGriddoLog.warn(\n\t\t\t\t\t\t`El directorio \"${dirPath}\" ya no existe. Posiblemente fue borrado por otra operaci\u00F3n.`,\n\t\t\t\t\t);\n\t\t\t\t} else {\n\t\t\t\t\tGriddoLog.error(`Error al borrar el directorio \"${dirPath}\":`, err);\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} catch (err: any) {\n\t\tif (err.code === \"ENOENT\") {\n\t\t\t// El directorio ya no existe, no es un error para nosotros en este contexto\n\t\t\tGriddoLog.warn(`The directory \"${dirPath}\" does not exist or has already been processed.`);\n\t\t} else {\n\t\t\tGriddoLog.error(`General error general while processing \"${dirPath}\":`, err);\n\t\t\tthrow err;\n\t\t}\n\t}\n}\n\nasync function pathExists(dir: string) {\n\ttry {\n\t\tawait fsp.access(dir);\n\t\treturn true;\n\t} catch {\n\t\treturn false;\n\t}\n}\n\n/**\n * Busca recursivamente archivos que terminen con un sufijo espec\u00EDfico dentro de un directorio.\n * Esta funci\u00F3n es un generador as\u00EDncrono, lo que la hace muy eficiente en uso de memoria.\n *\n * @param dir El directorio base para comenzar la b\u00FAsqueda.\n * @param suffix El sufijo con el que deben terminar los nombres de archivo (ej: 'page-data.json').\n * @returns Un generador as\u00EDncrono que produce la ruta completa de cada archivo encontrado.\n * @throws Si el directorio inicial `dir` no existe o no se puede leer.\n */\nasync function* findFilesBySuffix(dir: string, suffix: string): AsyncGenerator<string> {\n\tconst dirHandle = await fsp.opendir(dir);\n\tfor await (const item of dirHandle) {\n\t\tconst fullPath = path.join(dir, item.name);\n\t\tif (item.isDirectory()) {\n\t\t\t// yield* para encadenar otro generator.\n\t\t\tyield* findFilesBySuffix(fullPath, suffix);\n\t\t} else if (item.isFile() && item.name.endsWith(suffix)) {\n\t\t\tyield fullPath;\n\t\t}\n\t}\n}\n\n/**\n * Walk a directory and returns the JSON file absolute paths with one level of depth.\n * /abs/.../sotre/331/158.json\n * /abs/.../sotre/114/443.json\n * /abs/.../sotre/131/217.json\n * /abs/.../sotre/191/281.json\n */\nasync function* walkStore(storeDir: string): AsyncGenerator<string> {\n\tconst storeDirHandle = await fsp.opendir(storeDir);\n\n\tfor await (const siteDirent of storeDirHandle) {\n\t\tif (siteDirent.isDirectory()) {\n\t\t\tconst siteDirPath = path.join(storeDir, siteDirent.name);\n\t\t\tconst siteDirHandle = await fsp.opendir(siteDirPath);\n\n\t\t\tfor await (const fileDirent of siteDirHandle) {\n\t\t\t\tconst filePath = path.join(siteDirPath, fileDirent.name);\n\n\t\t\t\tif (fileDirent.isFile() && path.extname(filePath) === \".json\") {\n\t\t\t\t\tyield filePath;\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}\n\nexport {\n\tcpDirs,\n\tdeleteDisposableSiteDirs,\n\tdeleteEmptyDirectories,\n\tfindFilesBySuffix,\n\tmkDirs,\n\tmvDirs,\n\tpathExists,\n\trenamePath,\n\trmDirs,\n\twalkStore,\n};\n", "/**\n * Do you want to add a new error to the list?\n *\n * 1 - Add the new error type name to the `ErrorsType` union type.\n * 2 - Export a new ErrorData object (or a function that returns one) in this\n * file by completing the `error` (ErrosType) and `message` (string) properties\n * obligatorily.\n */\n\nimport type { SpawnSyncReturns } from \"node:child_process\";\nimport type { ErrorData } from \"../core/errors\";\n\ntype ErrorsType =\n\t| \"ArtifactError\"\n\t| \"BundlesInconsistencyError\"\n\t| \"CheckHealthError\"\n\t| \"ErrorInSSGBuildProcess\"\n\t| \"LifecycleExecutionError\"\n\t| \"LoginError\"\n\t| \"NoDomainsFoundError\"\n\t| \"NoJSConfigFileFound\"\n\t| \"ReadFromStoreError\"\n\t| \"ReferenceFieldSourcesNotFoundError\"\n\t| \"RenderUUIDError\"\n\t| \"UploadSearchError\"\n\t| \"WriteToStoreError\";\n\nconst ArtifactError: ErrorData = {\n\terror: \"ArtifactError\",\n\tmessage: \"There was a problem with an artifact\",\n\texpected:\n\t\t\"An external process may have has modified or deleted one of the artifacts (files and directories).\",\n\thint: \"Have there been any recent deployments? These can delete directories from the current render.\",\n};\n\nconst ErrorInSSGBuildProcess = (command: SpawnSyncReturns<string>): ErrorData => ({\n\terror: \"ErrorInSSGBuildProcess\",\n\tmessage: `Error in SSG build process: ${JSON.stringify(command)}`,\n\texpected: \"This can happen if there was a problem with the SSG build process.\",\n});\n\nconst LifecycleExecutionError = (attempts: number, name: string): ErrorData => ({\n\terror: \"LifecycleExecutionError\",\n\tmessage: `Exceeded maximum retry attempts (${attempts}) for ${name} LifeCycle`,\n});\n\nconst LoginError: ErrorData = {\n\terror: \"LoginError\",\n\tmessage: \"There was a problem logging in to the API\",\n\texpected: \"This happens if the API is currently not working or the credentials are incorrect.\",\n};\n\nconst NoDomainsFoundError: ErrorData = {\n\terror: \"NoDomainsFoundError\",\n\tmessage: \"No domains were found in this instance. The process cannot continue.\",\n\texpected:\n\t\t\"This may happen if the API is not functioning, or the site is not properly configured, or the domains are not registered.\",\n\thint: \"You can contact the instance administrator.\",\n};\n\nconst NoJSConfigFileFound: ErrorData = {\n\terror: \"NoJSConfigFileFound\",\n\tmessage: \"Could not find jsconfig.json or tsconfig.json\",\n\texpected:\n\t\t\"This can happen if the instance is not properly configured with a jsconfig.json or tsconfig.json file.\",\n};\n\nconst ReadFromStoreError: ErrorData = {\n\terror: \"ReadFromStoreError\",\n\tmessage: \"There was an error reading a file to the Store directory\",\n\thint: \"There may be an issue such as permissions preventing the file from being read.\",\n};\n\nconst ReferenceFieldSourcesNotFoundError: ErrorData = {\n\terror: \"ReferenceFieldSourcesNotFoundError\",\n\tmessage: \"The distributor has no sources defined.\",\n\texpected:\n\t\t\"It is expected to have at least one data source in the `sources` property, even if it is empty.\",\n};\n\nconst RenderUUIDError: ErrorData = {\n\terror: \"RenderUUIDError\",\n\tmessage: `Render sentinel file does not exist.\nThe rendering UUID cannot be read safely.\nThere was probably an instance deployment during the render, and files were deleted.\n\nThe files generated in this render will not be published.`,\n};\n\nconst WriteToStoreError: ErrorData = {\n\terror: \"WriteToStoreError\",\n\tmessage: \"There was an error writing a file to the Store directory\",\n\thint: \"There may be an issue such as lack of space or permissions preventing the file from being written.\",\n};\n\nconst UploadSearchError: ErrorData = {\n\terror: \"UploadSearchError\",\n\tmessage: \"There was an error uploading content to API for search\",\n\thint: \"This happens if the API is currently not working or the credentials are incorrect.\",\n};\n\nconst CheckHealthError: ErrorData = {\n\terror: \"CheckHealthError\",\n\tmessage: \"There was a problem with environment vars configuration.\",\n\texpected: \"Some of the required environment variables are not set correctly or are missing\",\n\thint: \"Are the environment variables correctly set?\",\n};\n\nexport {\n\tArtifactError,\n\tCheckHealthError,\n\tErrorInSSGBuildProcess,\n\tLifecycleExecutionError,\n\tLoginError,\n\tNoDomainsFoundError,\n\tNoJSConfigFileFound,\n\tReadFromStoreError,\n\tReferenceFieldSourcesNotFoundError,\n\tRenderUUIDError,\n\tUploadSearchError,\n\tWriteToStoreError,\n\ttype ErrorsType,\n};\n", "const GRIDDO_API_URL = process.env.GRIDDO_API_URL;\nconst GRIDDO_PUBLIC_API_URL = process.env.GRIDDO_PUBLIC_API_URL;\n\nconst AI_EMBEDDINGS = `${GRIDDO_API_URL}/ai/embeddings`;\nconst ALERT = `${GRIDDO_PUBLIC_API_URL}/alert`;\nconst DOMAINS = `${GRIDDO_API_URL}/domains`;\nconst GET_ALL = `${GRIDDO_API_URL}/sites/all`;\nconst GET_PAGE = `${GRIDDO_API_URL}/page`;\nconst LOGIN = `${GRIDDO_API_URL}/login_check`;\nconst RESET_RENDER = `${GRIDDO_API_URL}/debug/reset-render`;\nconst ROBOTS = `${GRIDDO_API_URL}/domains/robots`;\nconst SEARCH = `${GRIDDO_API_URL}/search`;\nconst SETTINGS = `${GRIDDO_API_URL}/settings`;\n\n// Site\nconst SITE_URI = `${GRIDDO_API_URL}/site/`;\nconst BUILD_END = [SITE_URI, \"/build/end\"];\nconst BUILD_START = [SITE_URI, \"/build/start\"];\nconst GET_PAGES = [SITE_URI, \"/pages?pagination=false\"];\nconst GET_REFERENCE_FIELD_DATA = [SITE_URI, \"/distributor\"];\nconst GET_SITEMAP = [SITE_URI, \"/sitemap\"];\nconst INFO = [SITE_URI, \"/all\"];\nconst LANGUAGES = [SITE_URI, \"/languages\"];\nconst SOCIALS = [SITE_URI, \"/socials\"];\n\nexport {\n\tAI_EMBEDDINGS,\n\tALERT,\n\tBUILD_END,\n\tBUILD_START,\n\tDOMAINS,\n\tGET_ALL,\n\tGET_PAGE,\n\tGET_PAGES,\n\tGET_REFERENCE_FIELD_DATA,\n\tGET_SITEMAP,\n\tINFO,\n\tLANGUAGES,\n\tLOGIN,\n\tRESET_RENDER,\n\tROBOTS,\n\tSEARCH,\n\tSETTINGS,\n\tSOCIALS,\n};\n", "import type { AuthHeaders } from \"../shared/types/api\";\n\nimport { throwError } from \"../core/errors\";\nimport { LOGIN } from \"../shared/endpoints\";\nimport { GRIDDO_BOT_PASSWORD, GRIDDO_BOT_USER } from \"../shared/envs\";\nimport { LoginError } from \"../shared/errors\";\n\nclass AuthService {\n\theaders: AuthHeaders | undefined;\n\n\tasync login() {\n\t\ttry {\n\t\t\tconst response = await fetch(LOGIN, {\n\t\t\t\tmethod: \"POST\",\n\t\t\t\theaders: { \"Content-Type\": \"application/json\", Connection: \"close\" },\n\t\t\t\tbody: JSON.stringify({\n\t\t\t\t\tusername: GRIDDO_BOT_USER,\n\t\t\t\t\tpassword: GRIDDO_BOT_PASSWORD,\n\t\t\t\t}),\n\t\t\t});\n\n\t\t\tif (!response.ok) {\n\t\t\t\tthrow new Error(\"Error while login in the API\");\n\t\t\t}\n\n\t\t\tconst { token } = await response.json();\n\t\t\tthis.headers = {\n\t\t\t\tAuthorization: `bearer ${token}`,\n\t\t\t\t\"Cache-Control\": \"no-store\",\n\t\t\t};\n\n\t\t\treturn this.headers;\n\t\t} catch (e) {\n\t\t\tthrowError(LoginError, e);\n\t\t}\n\t}\n}\n\nconst authService = new AuthService();\n\nexport { authService as AuthService };\n", "import fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { readDB } from \"./db\";\nimport { pathExists } from \"./fs\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nasync function distRollback(domain: string) {\n\tconst data = await readDB();\n\tconst { exportsDir, exportsDirBackup } = data.paths;\n\n\tGriddoLog.info(`Cleaning exports dir for the domain ${domain}`);\n\tGriddoLog.verbose(`Deleting ${path.join(exportsDir, domain)}...`);\n\n\t// Probar rsync en lugar de borrar y copiar\n\n\t// 1 - Borrar dist corrupto\n\tawait fsp.rm(path.join(exportsDir, domain), {\n\t\trecursive: true,\n\t\tforce: true,\n\t});\n\n\t// 2 - Si hay backup, restaurar\n\tif (await pathExists(path.join(exportsDirBackup, domain))) {\n\t\tawait fsp.cp(path.join(exportsDirBackup, domain), path.join(exportsDir, domain), {\n\t\t\trecursive: true,\n\t\t});\n\n\t\tGriddoLog.info(`export-backup dir for the domain ${domain} found. Restoring before exit...`);\n\t\tGriddoLog.verbose(\n\t\t\t`Copying ${path.join(exportsDirBackup, domain)} -> ${path.join(exportsDir, domain)}...`,\n\t\t);\n\t} else {\n\t\tGriddoLog.info(\n\t\t\t\"No export-backup found, skipping rollback. Next render will create a new exports dir from scratch...\",\n\t\t);\n\t}\n}\n\nexport { distRollback };\n", "import type {\n\tAPIRequest,\n\tAPIResponses,\n\tGetAPI,\n\tPostAPI,\n\tPutAPI,\n\tShowApiErrorOptions,\n} from \"../shared/types/api\";\nimport type { Petition } from \"../shared/types/global\";\n\nimport crypto from \"node:crypto\";\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport { RenderError } from \"../core/errors\";\nimport { pathExists } from \"../core/fs\";\nimport { GriddoLog } from \"../core/GriddoLog\";\nimport { addLogToBuffer } from \"../core/logger\";\nimport { brush } from \"../shared/npm-modules/brush\";\nimport { AuthService } from \"./auth\";\nimport { getRenderPathsHydratedWithDomainFromDB } from \"./render\";\n\n// Envs\nconst { env } = process;\nconst { RETRY_WAIT_SECONDS = \"4\", RETRY_ATTEMPTS = \"4\" } = env;\n\n/**\n * Make a GET/PUT/POST request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns {Promise<T>} A promise that is resolved with the data from the API response.\n *\n * @example\n *\tconst response = await requestAPI<Site>(\n *\t\t{ endpoint: \"...\", cacheKey: \"...\", ... },\n *\t\t\"get\",\n *\t\t\"...\"\n *\t);\n */\nasync function requestAPI<T extends APIResponses>(\n\tprops: APIRequest,\n\tmethod: string,\n\tappendToLog = \"\",\n): Promise<T> {\n\tconst {\n\t\tendpoint,\n\t\tbody,\n\t\tcacheKey = \"\",\n\t\tattempt = 1,\n\t\theaders,\n\t\tuseApiCacheDir = true,\n\t\tlogToFile = true,\n\t} = props;\n\tconst cacheOptions = { endpoint, body, headers, cacheKey };\n\n\t// Cache\n\tif (cacheKey && useApiCacheDir) {\n\t\tconst start = new Date();\n\t\tconst cacheData = await searchCacheData<T>(cacheOptions);\n\n\t\tif (cacheData) {\n\t\t\tif (logToFile) {\n\t\t\t\tconst siteId = getSafeSiteId(cacheData);\n\t\t\t\tconst siteIdMsg = siteId ? `site: ${siteId}` : \"\";\n\t\t\t\tconst duration = msToSec(Date.now() - start.getTime());\n\t\t\t\taddLogToBuffer(`${method} (cache) ${siteIdMsg} ${endpoint} - ${duration}s ${appendToLog}`);\n\t\t\t}\n\t\t\treturn cacheData;\n\t\t}\n\t}\n\n\t// Network\n\ttry {\n\t\tconst start = new Date();\n\n\t\t// Prepare fetch options\n\t\tconst fetchOptions: RequestInit = {\n\t\t\tmethod: method.toUpperCase(),\n\t\t\theaders: Object.assign({}, headers, AuthService.headers) as Record<string, string>,\n\t\t};\n\n\t\t// Add body for non-GET requests\n\t\tif (method.toLowerCase() !== \"get\" && body) {\n\t\t\tfetchOptions.body = JSON.stringify(body);\n\t\t\tif (!fetchOptions.headers) fetchOptions.headers = {};\n\t\t\t(fetchOptions.headers as Record<string, string>)[\"Content-Type\"] = \"application/json\";\n\t\t}\n\n\t\tconst response = await fetch(endpoint, fetchOptions);\n\n\t\t// Handle non-2xx responses\n\t\tif (!response.ok) {\n\t\t\tif (response.status === 404) {\n\t\t\t\t// @ts-expect-error page maybe will be 404\n\t\t\t\treturn null;\n\t\t\t}\n\n\t\t\tthrow new Error(`HTTP ${response.status}: ${response.statusText}`);\n\t\t}\n\n\t\tconst data: T = await response.json();\n\n\t\tif (logToFile) {\n\t\t\tconst siteId = getSafeSiteId(data);\n\t\t\tconst siteIdMsg = siteId ? `site: ${siteId}` : \"\";\n\t\t\tconst duration = msToSec(Date.now() - start.getTime());\n\t\t\taddLogToBuffer(`${method} (fetch) ${siteIdMsg} ${endpoint} - ${duration}s ${appendToLog}`);\n\t\t}\n\n\t\tif (useApiCacheDir) {\n\t\t\tawait saveCache(cacheOptions, data);\n\t\t}\n\n\t\treturn data;\n\t} catch (e) {\n\t\tconst error = e as Error;\n\n\t\tif (attempt > parseInt(RETRY_ATTEMPTS)) {\n\t\t\tGriddoLog.log(`\nMax attempts ${RETRY_ATTEMPTS} reached\n--------------------------------------\n- ${method.toUpperCase()} ${endpoint}\n- BODY: ${JSON.stringify(body)}\n- HEADERS: ${JSON.stringify(headers)}\n- ERROR: ${error.message}\n--------------------------------------\n`);\n\t\t\tthrow new RenderError(error);\n\t\t}\n\n\t\tshowApiError(error, {\n\t\t\tcallInfo: { endpoint, body },\n\t\t});\n\n\t\tGriddoLog.warn(`Waiting for retry: ${method}`, endpoint);\n\n\t\tawait delay(parseInt(RETRY_WAIT_SECONDS) * 1000);\n\n\t\treturn requestAPI<T>(\n\t\t\t{\n\t\t\t\tendpoint,\n\t\t\t\tbody,\n\t\t\t\theaders,\n\t\t\t\tcacheKey,\n\t\t\t\tattempt: attempt + 1,\n\t\t\t},\n\t\t\tmethod,\n\t\t\tappendToLog,\n\t\t);\n\t}\n}\n\n/**\n * Make a GET request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns A promise that is resolved with the data from the API response.\n */\nasync function getApi<T extends APIResponses>(props: GetAPI) {\n\treturn requestAPI<T>(props, \"get\");\n}\n\n/**\n * Make a PUT request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns A promise that is resolved with the data from the API response.\n */\nasync function putApi<T extends APIResponses>(props: PutAPI) {\n\treturn requestAPI<T>(props, \"put\");\n}\n\n/**\n * Make a POST request to the Griddo API.\n *\n * @template T Response Type returned.\n * @returns A promise that is resolved with the data from the API response.\n */\nasync function postApi<T extends APIResponses>(props: PostAPI) {\n\tconst { endpoint, body, headers } = props;\n\tconst referenceFieldBodyParams =\n\t\tendpoint.endsWith(\"/distributor\") &&\n\t\t`# ReferenceField body: ${JSON.stringify(body)} lang: ${JSON.stringify(headers?.lang)}`;\n\n\treturn requestAPI<T>(props, \"post\", referenceFieldBodyParams || \"\");\n}\n\n/**\n * Shows an API error through the terminal.\n */\nfunction showApiError(error: Error, options: ShowApiErrorOptions) {\n\tconst { message, stack } = error;\n\tconst { callInfo } = options;\n\tconst callInfoArray = [];\n\n\tfor (const item of Object.keys(callInfo) as (keyof typeof callInfo)[]) {\n\t\tcallInfoArray.push(\n\t\t\t`${item}: ${\n\t\t\t\ttypeof callInfo[item] === \"object\" ? JSON.stringify(callInfo[item]) : callInfo[item]\n\t\t\t}`,\n\t\t);\n\t}\n\n\t// Compose the errors output\n\tconst callInfoStr = callInfoArray.join(\"\\n\");\n\tconst errorDetailsStr = `${message}\\n${stack}`;\n\n\t// Print the error\n\tGriddoLog.warn(\n\t\tbrush.red(`\n=============\n\n{ Call info }\n${callInfoStr}\n\n{ Error details }\n${errorDetailsStr}\n\n=============\n`),\n\t);\n}\n\n/**\n * Return a siteID from a response object if exist\n * @param response A response object\n */\nfunction getSafeSiteId(response: APIResponses) {\n\tif (typeof response !== \"object\" || response === null || Array.isArray(response)) {\n\t\treturn undefined;\n\t}\n\n\treturn \"site\" in response && response.site ? response.site : undefined;\n}\n\n/**\n * Custom delay using the \"promise hack\",\n *\n * @param ms Amount of miliseconds to be delayed\n */\nfunction delay(ms: number): Promise<void> {\n\treturn new Promise((res) => setTimeout(res, ms));\n}\n\n/**\n * Converts milliseconds to seconds with a fixed number of decimals.\n *\n * @param ms The number in milliseconds.\n * @param fixed The amount of fixed decimals.\n * @returns The converted number in seconds with the fixed number of decimals.\n */\nfunction msToSec(ms: number, decimals = 3): number {\n\treturn Number.parseFloat((ms / 1000).toFixed(decimals));\n}\n\n/**\n * Generate a filename with a hash using a Petition object\n * @param petition An object\n */\nasync function generateFilenameWithHash(petition: Petition) {\n\tconst { __root } = await getRenderPathsHydratedWithDomainFromDB();\n\tconst apiCacheDir = path.join(__root, \"apiCache\");\n\n\tconst hashSum = crypto.createHash(\"sha256\");\n\thashSum.update(JSON.stringify(petition));\n\n\treturn `${apiCacheDir}/${hashSum.digest(\"hex\")}`;\n}\n\n/**\n * Save a file using a hash name.\n *\n * @param petition An object.\n * @param content Content to be saved.\n */\nasync function saveCache<T>(petition: Petition, content: T) {\n\tconst stringContent = typeof content === \"string\" ? content : JSON.stringify(content);\n\tconst filename = await generateFilenameWithHash(petition);\n\tconst filepath = path.dirname(filename);\n\n\tif (!(await pathExists(filepath))) {\n\t\tawait fsp.mkdir(filepath, { recursive: true });\n\t}\n\n\tawait fsp.writeFile(filename, stringContent, \"utf8\");\n}\n\n/**\n * Search in the `apiCache` dir for a file using the petition as hash generator.\n * Return the file content if found or null if not.\n *\n * @param petition An object\n */\nasync function searchCacheData<T>(petition: Petition) {\n\ttry {\n\t\tconst file = await generateFilenameWithHash(petition);\n\t\tconst fileContent = await fsp.readFile(file, \"utf8\");\n\t\tconst jsonData = JSON.parse(fileContent) as T;\n\n\t\treturn jsonData;\n\t} catch {\n\t\treturn null;\n\t}\n}\n\nexport { getApi as get, postApi as post, putApi as put };\n", "import type { RenderDB } from \"../shared/types/render\";\n\nimport fsp from \"node:fs/promises\";\nimport path from \"node:path\";\n\nimport {\n\tGRIDDO_BUILD_LOGS,\n\tGRIDDO_BUILD_LOGS_BUFFER_SIZE,\n\t// GRIDDO_BUILD_LOGS_TO_FILE,\n} from \"../shared/envs\";\nimport { readDB } from \"./db\";\nimport { GriddoLog } from \"./GriddoLog\";\n\nconst logBuffer: string[] = [];\nlet logFilePath: string | null = null;\nlet flushPromise: Promise<void> | null = null;\nconst LOG_BUFFER_FLUSH_THRESHOLD = GRIDDO_BUILD_LOGS_BUFFER_SIZE;\nconst LOG_FILENAME = \"render-detail-log.txt\";\nconst LOG_TO_FILE_IS_DISABLED = !GRIDDO_BUILD_LOGS; // || !GRIDDO_BUILD_LOGS_TO_FILE;\n\nlet dbData: RenderDB | null = null;\n\n// Universal data cache for this module...\nasync function getDBData() {\n\tif (!dbData) {\n\t\tdbData = await readDB();\n\t}\n\treturn dbData;\n}\n\nasync function initializeLogFile() {\n\tif (LOG_TO_FILE_IS_DISABLED) {\n\t\treturn;\n\t}\n\n\tconst data = await getDBData();\n\tlogFilePath = path.join(data.paths.root, LOG_FILENAME);\n\n\tawait fsp.rm(logFilePath, { force: true });\n}\n\nfunction addLogToBuffer(lineContent: string) {\n\tif (LOG_TO_FILE_IS_DISABLED) {\n\t\treturn;\n\t}\n\n\tlogBuffer.push(lineContent.toString());\n\n\tif (logBuffer.length >= LOG_BUFFER_FLUSH_THRESHOLD) {\n\t\t// Fire-and-forget flush to avoid blocking the main thread.\n\t\tflushLogsToFile().catch((error) => {\n\t\t\tGriddoLog.error(\"Background log flush failed:\", error);\n\t\t});\n\t}\n}\n\n/**\n * Vuelca el contenido del buffer de logs al archivo de logs en disco.\n * Si el buffer est\u00E1 vac\u00EDo, ya se est\u00E1 volcando, o no hay ruta de archivo, no hace nada.\n * Si ocurre un error al escribir, los logs se reinsertan al buffer para reintentar en el siguiente flush.\n */\nasync function flushLogsToFile() {\n\t// Si ya hay un volcado en curso, espera a que termine.\n\tif (flushPromise) {\n\t\tawait flushPromise;\n\t}\n\n\tif (logBuffer.length === 0 || !logFilePath) {\n\t\treturn;\n\t}\n\n\tconst performFlush = async () => {\n\t\tconst logsToFlush = [...logBuffer];\n\t\tlogBuffer.length = 0;\n\n\t\ttry {\n\t\t\tawait fsp.appendFile(logFilePath!, `${logsToFlush.join(\"\\n\")}\\n`);\n\t\t} catch (error) {\n\t\t\tlogBuffer.unshift(...logsToFlush);\n\t\t\tGriddoLog.error(\"Error flushing logs:\", error);\n\t\t\t// No relanzamos el error para no detener el \"fire-and-forget\"\n\t\t}\n\t};\n\n\tflushPromise = performFlush();\n\n\ttry {\n\t\tawait flushPromise;\n\t} finally {\n\t\tflushPromise = null;\n\t}\n}\n\n/**\n * Copia el archivo de log detallado de renderizado (\"render-detail-log.txt\")\n * desde el directorio ra\u00EDz del proyecto al directorio de logs del dominio en exports,\n * usando una marca de tiempo en el nombre de destino.\n * Si el archivo no existe, ignora el error salvo que sea distinto de ENOENT.\n *\n * @param domain - Nombre del dominio para el que se guarda el log.\n */\nasync function saveDetailRenderLog(domain: string) {\n\tawait flushLogsToFile();\n\n\tconst data = await getDBData();\n\n\tconst dateString = getFormattedDateTime();\n\tconst debugDir = path.join(data.paths.exportsDir, domain, \"logs\");\n\n\tawait fsp.mkdir(debugDir, { recursive: true });\n\n\tconst src = path.join(data.paths.root, LOG_FILENAME);\n\tconst dst = path.join(debugDir, `${dateString}-${LOG_FILENAME}`);\n\n\t// Move log to exports\n\ttry {\n\t\tawait fsp.cp(src, dst);\n\t} catch (error) {\n\t\t// It's possible the file doesn't exist if GRIDDO_BUILD_LOGS is false\n\t\t// or if no logs were ever added.\n\t\tif (error instanceof Error && \"code\" in error && error.code !== \"ENOENT\") {\n\t\t\tthrow error;\n\t\t}\n\t}\n}\n\nfunction getFormattedDateTime() {\n\tconst now = new Date();\n\tconst year = now.getFullYear();\n\tconst month = String(now.getMonth() + 1).padStart(2, \"0\");\n\tconst day = String(now.getDate()).padStart(2, \"0\");\n\n\tconst hours = String(now.getHours()).padStart(2, \"0\");\n\tconst minutes = String(now.getMinutes()).padStart(2, \"0\");\n\tconst seconds = String(now.getSeconds()).padStart(2, \"0\");\n\n\t// Formato recomendado para ordenaci\u00F3n de archivos: YYYY-MM-DD_HH-mm-ss\n\treturn `${year}-${month}-${day}_${hours}-${minutes}-${seconds}`;\n}\n\nexport { addLogToBuffer, flushLogsToFile, initializeLogFile, saveDetailRenderLog };\n", "//\n//\n// NOTA: Este proceso es para toda la instancia, no discrimina por dominio\n//\n//\n\nimport type { AIEmbeddingsResponse } from \"../shared/types/global\";\n\nimport { withErrorHandler } from \"../core/errors\";\nimport { post } from \"../services/api\";\nimport { AuthService } from \"../services/auth\";\nimport { AI_EMBEDDINGS } from \"../shared/endpoints\";\nimport { GRIDDO_AI_EMBEDDINGS, GRIDDO_SEARCH_FEATURE } from \"../shared/envs\";\n\nasync function triggerAIEmbeddings() {\n\tawait post<AIEmbeddingsResponse>({\n\t\tendpoint: AI_EMBEDDINGS,\n\t\tuseApiCacheDir: false,\n\t});\n}\n\nasync function main() {\n\tif (GRIDDO_SEARCH_FEATURE && GRIDDO_AI_EMBEDDINGS) {\n\t\tawait AuthService.login();\n\t\tawait triggerAIEmbeddings();\n\t}\n}\n\nwithErrorHandler(main);\n"],
|
|
5
|
+
"mappings": "oeAEA,IAAAA,GAAiB,wBCEjB,IAAAC,EAAiB,wBCFjB,IAAAC,EAAgB,+BAChBC,EAAiB,wBCHjB,IAAAC,EAAiB,wBCAjB,IAAAC,EAAe,sBAEf,IAAAC,EAAiB,wBACjBC,EAAoB,2BACpBC,EAA8B,oBA0BxBC,EAAUC,GACfA,aAAqB,OAAM,iBAAcA,CAAS,EAAIA,EA0ChD,SAASC,EAAWC,EAAcC,EAAmB,CAAC,EAAuB,CACnF,GAAM,CAAE,IAAAC,EAAM,EAAAC,QAAQ,IAAI,EAAG,KAAAC,EAAO,OAAQ,OAAQC,CAAa,EAAIJ,EAEjEK,EAAY,EAAAC,QAAK,QAAQC,EAAON,CAAG,GAAK,EAAE,EACxC,CAAE,KAAAO,CAAK,EAAI,EAAAF,QAAK,MAAMD,CAAS,EAC/BI,EAASL,EAAe,EAAAE,QAAK,QAAQD,EAAWE,EAAOH,CAAY,CAAE,EAAII,EACzEE,EAAiB,EAAAJ,QAAK,WAAWP,CAAI,EAE3C,OAAa,CACZ,IAAMY,EAAWD,EAAiBX,EAAO,EAAAO,QAAK,KAAKD,EAAWN,CAAI,EAClE,GAAI,CACH,IAAMa,EAAQ,EAAAC,QAAG,SAASF,EAAU,CAAE,eAAgB,EAAM,CAAC,EAC7D,GAAKR,IAAS,QAAUS,GAAO,OAAO,GAAOT,IAAS,aAAeS,GAAO,YAAY,EACvF,OAAOD,CAET,MAAQ,CAGR,CAEA,GAAIN,IAAcI,GAAUJ,IAAcG,EACzC,MAGDH,EAAY,EAAAC,QAAK,QAAQD,CAAS,CACnC,CACD,CDzFA,SAASS,EAAWC,EAAqC,CACxD,GAAM,CAAE,IAAAC,CAAI,EAAID,GAAW,CAAC,EACtBE,EAAWC,EAAW,eAAgB,CAAE,IAAAF,CAAI,CAAC,EACnD,OAAOC,GAAY,EAAAE,QAAK,QAAQF,CAAQ,CACzC,CEdA,GAAM,CAAE,IAAAG,CAAI,EAAI,QAKhB,SAASC,EAAYC,EAAyB,CAC7C,GAAI,CAACA,EAAO,MAAO,GAEnB,OAAQA,EAAM,KAAK,EAAE,YAAY,EAAG,CACnC,IAAK,IACL,IAAK,OACL,IAAK,MACL,IAAK,IACL,IAAK,KACJ,MAAO,GACR,QACC,MAAO,EACT,CACD,CAGA,IAAMC,GAAiBH,EAAI,gBAAkBA,EAAI,QAC3CI,GAAwBJ,EAAI,uBAAyBA,EAAI,eACzDK,EAAkBL,EAAI,UAAYA,EAAI,gBACtCM,EAAsBN,EAAI,aAAeA,EAAI,oBAG7CO,GAA+B,OAAO,SAASP,EAAI,8BAAgC,IAAI,EACvFQ,GAA2BP,EAAYD,EAAI,wBAAwB,EACnES,GAA2BR,EAAYD,EAAI,wBAAwB,EACnEU,EAAoBT,EAAYD,EAAI,iBAAiB,EACrDW,GAA4BV,EAAYD,EAAI,yBAAyB,EACrEY,EAAgC,OAAO,SAASZ,EAAI,+BAAiC,KAAK,EAC1Fa,GAA0BZ,EAAYD,EAAI,uBAAuB,EACjEc,EAAwBb,EAAYD,EAAI,qBAAqB,EAC7De,GAAsBf,EAAI,qBAAuBA,EAAI,aACrDgB,GAA4BhB,EAAI,2BAA6BA,EAAI,mBACjEiB,GAAuBhB,EAAYD,EAAI,oBAAoB,EAC3DkB,GAAsBjB,EAAYD,EAAI,mBAAmB,EACzDmB,GAAyBlB,EAAYD,EAAI,sBAAsB,EAC/DoB,GAA6BnB,EAAYD,EAAI,0BAA0B,ECjC7E,IAAMqB,GAAQ,UACRC,GAAQ,CACb,MAAO,WACP,IAAK,WACL,MAAO,WACP,OAAQ,WACR,KAAM,WACN,QAAS,WACT,KAAM,WACN,MAAO,WACP,KAAM,WACN,KAAM,UACN,IAAK,SACN,EAMMC,EAAQ,CAAC,EAEf,QAAWC,KAASF,GAAO,CAC1B,IAAMG,EAAMD,EACZD,EAAME,CAAG,EAAKC,GAA0B,GAAGJ,GAAMG,CAAG,CAAC,GAAGC,CAAI,GAAGL,EAAK,EACrE,CCxBA,IAAMM,EAAN,MAAMC,CAAU,CAEP,aAAc,CAAC,CAEvB,OAAc,WAAWC,EAAsB,CAC1CC,IACH,QAAQ,IAAIC,EAAM,OAAO,SAAS,EAAGA,EAAM,IAAIF,EAAI,KAAK,GAAG,CAAC,CAAC,CAE/D,CAEA,OAAc,SAASA,EAAsB,CACxCG,GACHJ,EAAU,IAAI,GAAGC,CAAG,CAEtB,CAEA,OAAc,QAAQA,EAAsB,CAC3C,QAAQ,IAAI,GAAGE,EAAM,KAAK,MAAM,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACrD,CAEA,OAAc,WAAWA,EAAsB,CAC9C,QAAQ,IAAI,GAAGE,EAAM,MAAM,SAAS,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACzD,CAEA,OAAc,SAASA,EAAsB,CAC5C,QAAQ,MAAM,GAAGE,EAAM,IAAI,OAAO,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACvD,CAEA,OAAc,QAAQA,EAAsB,CAC3C,QAAQ,KAAK,GAAGE,EAAM,OAAO,MAAM,CAAC,IAAIF,EAAI,KAAK,GAAG,CAAC,EAAE,CACxD,CAEA,OAAc,OAAOI,EAA4C,CAChE,QAAQ,IAAI,GAAGA,CAAI,CACpB,CACD,ELlCA,IAAMC,GAAOC,EAAW,CAAE,IAAK,EAAAC,QAAK,QAAQ,UAAW,UAAU,CAAE,CAAC,GAAK,GACnEC,GAAQ,EAAAD,QAAK,KAAKF,GAAM,eAAe,EACvCI,GAAa,EAAAF,QAAK,KAAKC,GAAO,SAAS,EAE7C,eAAeE,EAAOC,EAAe,GAAI,CACxC,IAAMC,EAAOD,GAAgBF,GAC7B,GAAI,CACH,OAAO,KAAK,MAAM,MAAM,EAAAI,QAAI,SAASD,EAAM,OAAO,CAAC,CACpD,OAASE,EAAO,CACf,MAAAC,EAAU,MAAM,6BAA6BH,CAAI,IAAKE,CAAK,EACrDA,CACP,CACD,CAEA,eAAeE,EAAQC,EAAoBN,EAAe,GAAI,CAC7D,IAAMC,EAAOD,GAAgBF,GAC7B,GAAI,CACH,MAAM,EAAAI,QAAI,UAAUD,EAAM,KAAK,UAAUK,EAAU,KAAM,GAAI,CAAC,CAC/D,OAASH,EAAO,CACf,MAAAC,EAAU,MAAM,8BAA8BH,CAAI,IAAKE,CAAK,EACtDA,CACP,CACD,CM5BA,IAAAI,GAAgB,+BC4ChB,IAAMC,GAAwB,CAC7B,MAAO,aACP,QAAS,4CACT,SAAU,oFACX,ED0QA,eAAeC,EAAWC,EAAa,CACtC,GAAI,CACH,aAAM,GAAAC,QAAI,OAAOD,CAAG,EACb,EACR,MAAQ,CACP,MAAO,EACR,CACD,CEnUA,IAAME,EAAiB,QAAQ,IAAI,eAC7BC,GAAwB,QAAQ,IAAI,sBAEpCC,GAAgB,GAAGF,CAAc,iBACjCG,GAAQ,GAAGF,EAAqB,SAChCG,GAAU,GAAGJ,CAAc,WAC3BK,GAAU,GAAGL,CAAc,aAC3BM,GAAW,GAAGN,CAAc,QAC5BO,GAAQ,GAAGP,CAAc,eACzBQ,GAAe,GAAGR,CAAc,sBAChCS,GAAS,GAAGT,CAAc,kBAC1BU,GAAS,GAAGV,CAAc,UAC1BW,GAAW,GAAGX,CAAc,YAG5BY,GAAW,GAAGZ,CAAc,SCRlC,IAAMa,EAAN,KAAkB,CACjB,QAEA,MAAM,OAAQ,CACb,GAAI,CACH,IAAMC,EAAW,MAAM,MAAMC,GAAO,CACnC,OAAQ,OACR,QAAS,CAAE,eAAgB,mBAAoB,WAAY,OAAQ,EACnE,KAAM,KAAK,UAAU,CACpB,SAAUC,EACV,SAAUC,CACX,CAAC,CACF,CAAC,EAED,GAAI,CAACH,EAAS,GACb,MAAM,IAAI,MAAM,8BAA8B,EAG/C,GAAM,CAAE,MAAAI,CAAM,EAAI,MAAMJ,EAAS,KAAK,EACtC,YAAK,QAAU,CACd,cAAe,UAAUI,CAAK,GAC9B,gBAAiB,UAClB,EAEO,KAAK,OACb,OAAS,EAAG,CACXC,EAAWC,GAAY,CAAC,CACzB,CACD,CACD,EAEMC,EAAc,IAAIR,EV2HxB,eAAeS,EAAuCC,EAGnD,CACF,GAAM,CAAE,OAAAC,EAAQ,WAAAC,CAAW,EAAIF,GAAW,CAAC,EAGrCG,GADK,MAAMC,EAAOF,CAAU,GACjB,MAEjB,MAAO,CACN,OAAQC,EAAM,KACd,QAAS,EAAAE,QAAK,KAAKF,EAAM,QAASF,GAAU,EAAE,EAC9C,aAAcE,EAAM,WACpB,KAAMA,EAAM,GACZ,QAASA,EAAM,WACf,UAAW,EAAAE,QAAK,KAAKF,EAAM,WAAYF,GAAU,EAAE,EACnD,iBAAkB,EAAAI,QAAK,KAAKF,EAAM,iBAAkBF,GAAU,EAAE,EAChE,MAAOE,EAAM,IACb,eAAgB,EAAAE,QAAK,KAAKF,EAAM,WAAYF,GAAU,GAAI,MAAM,CACjE,CACD,CWrLA,IAAAK,EAAgB,+BAChBC,EAAiB,wBAMjB,eAAeC,GAAaC,EAAgB,CAC3C,IAAMC,EAAO,MAAMC,EAAO,EACpB,CAAE,WAAAC,EAAY,iBAAAC,CAAiB,EAAIH,EAAK,MAE9CI,EAAU,KAAK,uCAAuCL,CAAM,EAAE,EAC9DK,EAAU,QAAQ,YAAY,EAAAC,QAAK,KAAKH,EAAYH,CAAM,CAAC,KAAK,EAKhE,MAAM,EAAAO,QAAI,GAAG,EAAAD,QAAK,KAAKH,EAAYH,CAAM,EAAG,CAC3C,UAAW,GACX,MAAO,EACR,CAAC,EAGG,MAAMQ,EAAW,EAAAF,QAAK,KAAKF,EAAkBJ,CAAM,CAAC,GACvD,MAAM,EAAAO,QAAI,GAAG,EAAAD,QAAK,KAAKF,EAAkBJ,CAAM,EAAG,EAAAM,QAAK,KAAKH,EAAYH,CAAM,EAAG,CAChF,UAAW,EACZ,CAAC,EAEDK,EAAU,KAAK,oCAAoCL,CAAM,kCAAkC,EAC3FK,EAAU,QACT,WAAW,EAAAC,QAAK,KAAKF,EAAkBJ,CAAM,CAAC,OAAO,EAAAM,QAAK,KAAKH,EAAYH,CAAM,CAAC,KACnF,GAEAK,EAAU,KACT,sGACD,CAEF,CZpBO,IAAMI,EAAN,cAA0B,KAAM,CACtC,YAAYC,EAAyB,CACpC,MAAMA,aAAyB,MAAQA,EAAc,QAAU,OAAOA,CAAa,CAAC,EAEpF,KAAK,KAAO,kBACZ,KAAK,MAAQA,aAAyB,MAAQA,EAAc,MAAQ,EACrE,CACD,EAKA,SAASC,EAAWC,EAAoBC,EAAwB,CAC/D,GAAM,CAAE,MAAAC,EAAO,QAAAC,EAAS,SAAAC,EAAU,KAAAC,CAAK,EAAIL,EAErCM,EAAaC,EAAU,IAAIC,EAAM,IAAI,KAAKN,CAAK,IAAI,CAAC,EACpDO,EAAY,CAACL,EAAUC,CAAI,EAAE,OAAO,OAAO,EAAE,KAAK;AAAA,CAAI,EAE5D,MAAAE,EAAU,IAAI;AAAA,EACbD,CAAU;AAAA,EACVH,CAAO;AAAA,EACPM,CAAS;AAAA;AAAA,EAETD,EAAM,IAAI,OAAO,CAAC;AAAA,EAClB,KAAK,UAAUP,EAAO,KAAM,CAAC,CAAC,EAAE,EAE3B,IAAIJ,EAAYI,CAAK,CAC5B,CAEA,eAAeS,GAAoBC,EAAsB,CACxD,GAAI,CACH,MAAMA,EAAG,CACV,OAAST,EAAO,CACXA,aAAiBL,EACpBU,EAAU,MAAM,6BAA6B,EACnCL,aAAiB,MAC3BK,EAAU,MAAML,EAAM,OAAO,EAE7BK,EAAU,MAAM,gCAAgCL,CAAK,EAAE,EAGxD,GAAI,CACH,GAAM,CAAE,OAAAU,CAAO,EAAI,MAAMC,EAAuC,EAC1DC,EAAO,MAAMC,EAAO,EACtBD,EAAK,sBACRP,EAAU,KAAK,yBAAyB,EACxCA,EAAU,QAAQ,YAAY,GAAAS,QAAK,KAAKJ,EAAQ,SAAS,CAAC,KAAK,EAE/D,MAAMK,GAAaH,EAAK,sBAAuB,GAE/CP,EAAU,KAAK,iCAAiC,CAElD,MAAa,CACZA,EAAU,KAAK,+CAA+C,CAC/D,CAEA,IAAMO,EAAO,MAAMC,EAAO,EAC1B,MAAAD,EAAK,QAAQA,EAAK,sBAAuB,EAAE,YAAc,GACzDA,EAAK,QAAQA,EAAK,sBAAuB,EAAE,WAAa,QACxD,MAAMI,EAAQJ,CAAI,EACZZ,CACP,CACD,CarEA,IAAAiB,GAAmB,0BACnBC,EAAgB,+BAChBC,EAAiB,wBCVjB,IAAAC,GAAgB,+BAWhB,IAAMC,EAAsB,CAAC,EACzBC,GAA6B,KAC7BC,EAAqC,KACnCC,GAA6BC,EAEnC,IAAMC,GAA0B,CAACC,EAuBjC,SAASC,EAAeC,EAAqB,CACxCC,KAIJC,EAAU,KAAKF,EAAY,SAAS,CAAC,EAEjCE,EAAU,QAAUC,IAEvBC,GAAgB,EAAE,MAAOC,GAAU,CAClCC,EAAU,MAAM,+BAAgCD,CAAK,CACtD,CAAC,EAEH,CAOA,eAAeD,IAAkB,CAMhC,GAJIG,GACH,MAAMA,EAGHL,EAAU,SAAW,GAAK,CAACM,GAC9B,OAgBDD,GAbqB,SAAY,CAChC,IAAME,EAAc,CAAC,GAAGP,CAAS,EACjCA,EAAU,OAAS,EAEnB,GAAI,CACH,MAAM,GAAAQ,QAAI,WAAWF,GAAc,GAAGC,EAAY,KAAK;AAAA,CAAI,CAAC;AAAA,CAAI,CACjE,OAASJ,EAAO,CACfH,EAAU,QAAQ,GAAGO,CAAW,EAChCH,EAAU,MAAM,uBAAwBD,CAAK,CAE9C,CACD,GAE4B,EAE5B,GAAI,CACH,MAAME,CACP,QAAE,CACDA,EAAe,IAChB,CACD,CDpEA,GAAM,CAAE,IAAAI,EAAI,EAAI,QACV,CAAE,mBAAAC,GAAqB,IAAK,eAAAC,GAAiB,GAAI,EAAIF,GAe3D,eAAeG,GACdC,EACAC,EACAC,EAAc,GACD,CACb,GAAM,CACL,SAAAC,EACA,KAAAC,EACA,SAAAC,EAAW,GACX,QAAAC,EAAU,EACV,QAAAC,EACA,eAAAC,EAAiB,GACjB,UAAAC,EAAY,EACb,EAAIT,EACEU,EAAe,CAAE,SAAAP,EAAU,KAAAC,EAAM,QAAAG,EAAS,SAAAF,CAAS,EAGzD,GAAIA,GAAYG,EAAgB,CAC/B,IAAMG,EAAQ,IAAI,KACZC,EAAY,MAAMC,GAAmBH,CAAY,EAEvD,GAAIE,EAAW,CACd,GAAIH,EAAW,CACd,IAAMK,EAASC,GAAcH,CAAS,EAChCI,EAAYF,EAAS,SAASA,CAAM,GAAK,GACzCG,EAAWC,GAAQ,KAAK,IAAI,EAAIP,EAAM,QAAQ,CAAC,EACrDQ,EAAe,GAAGlB,CAAM,YAAYe,CAAS,IAAIb,CAAQ,MAAMc,CAAQ,KAAKf,CAAW,EAAE,CAC1F,CACA,OAAOU,CACR,CACD,CAGA,GAAI,CACH,IAAMD,EAAQ,IAAI,KAGZS,EAA4B,CACjC,OAAQnB,EAAO,YAAY,EAC3B,QAAS,OAAO,OAAO,CAAC,EAAGM,EAASc,EAAY,OAAO,CACxD,EAGIpB,EAAO,YAAY,IAAM,OAASG,IACrCgB,EAAa,KAAO,KAAK,UAAUhB,CAAI,EAClCgB,EAAa,UAASA,EAAa,QAAU,CAAC,GAClDA,EAAa,QAAmC,cAAc,EAAI,oBAGpE,IAAME,EAAW,MAAM,MAAMnB,EAAUiB,CAAY,EAGnD,GAAI,CAACE,EAAS,GAAI,CACjB,GAAIA,EAAS,SAAW,IAEvB,OAAO,KAGR,MAAM,IAAI,MAAM,QAAQA,EAAS,MAAM,KAAKA,EAAS,UAAU,EAAE,CAClE,CAEA,IAAMC,EAAU,MAAMD,EAAS,KAAK,EAEpC,GAAIb,EAAW,CACd,IAAMK,EAASC,GAAcQ,CAAI,EAC3BP,GAAYF,EAAS,SAASA,CAAM,GAAK,GACzCG,GAAWC,GAAQ,KAAK,IAAI,EAAIP,EAAM,QAAQ,CAAC,EACrDQ,EAAe,GAAGlB,CAAM,YAAYe,EAAS,IAAIb,CAAQ,MAAMc,EAAQ,KAAKf,CAAW,EAAE,CAC1F,CAEA,OAAIM,GACH,MAAMgB,GAAUd,EAAca,CAAI,EAG5BA,CACR,OAASE,EAAG,CACX,IAAMC,EAAQD,EAEd,GAAInB,EAAU,SAASR,EAAc,EACpC,MAAA6B,EAAU,IAAI;AAAA,eACF7B,EAAc;AAAA;AAAA,IAEzBG,EAAO,YAAY,CAAC,IAAIE,CAAQ;AAAA,UAC1B,KAAK,UAAUC,CAAI,CAAC;AAAA,aACjB,KAAK,UAAUG,CAAO,CAAC;AAAA,WACzBmB,EAAM,OAAO;AAAA;AAAA,CAEvB,EACQ,IAAIE,EAAYF,CAAK,EAG5B,OAAAG,GAAaH,EAAO,CACnB,SAAU,CAAE,SAAAvB,EAAU,KAAAC,CAAK,CAC5B,CAAC,EAEDuB,EAAU,KAAK,sBAAsB1B,CAAM,GAAIE,CAAQ,EAEvD,MAAM2B,GAAM,SAASjC,EAAkB,EAAI,GAAI,EAExCE,GACN,CACC,SAAAI,EACA,KAAAC,EACA,QAAAG,EACA,SAAAF,EACA,QAASC,EAAU,CACpB,EACAL,EACAC,CACD,CACD,CACD,CA4BA,eAAe6B,GAAgCC,EAAgB,CAC9D,GAAM,CAAE,SAAAC,EAAU,KAAAC,EAAM,QAAAC,CAAQ,EAAIH,EAC9BI,EACLH,EAAS,SAAS,cAAc,GAChC,0BAA0B,KAAK,UAAUC,CAAI,CAAC,UAAU,KAAK,UAAUC,GAAS,IAAI,CAAC,GAEtF,OAAOE,GAAcL,EAAO,OAAQI,GAA4B,EAAE,CACnE,CAKA,SAASE,GAAaC,EAAcC,EAA8B,CACjE,GAAM,CAAE,QAAAC,EAAS,MAAAC,CAAM,EAAIH,EACrB,CAAE,SAAAI,CAAS,EAAIH,EACfI,EAAgB,CAAC,EAEvB,QAAWC,KAAQ,OAAO,KAAKF,CAAQ,EACtCC,EAAc,KACb,GAAGC,CAAI,KACN,OAAOF,EAASE,CAAI,GAAM,SAAW,KAAK,UAAUF,EAASE,CAAI,CAAC,EAAIF,EAASE,CAAI,CACpF,EACD,EAID,IAAMC,EAAcF,EAAc,KAAK;AAAA,CAAI,EACrCG,EAAkB,GAAGN,CAAO;AAAA,EAAKC,CAAK,GAG5CM,EAAU,KACTC,EAAM,IAAI;AAAA;AAAA;AAAA;AAAA,EAIVH,CAAW;AAAA;AAAA;AAAA,EAGXC,CAAe;AAAA;AAAA;AAAA,CAGhB,CACA,CACD,CAMA,SAASG,GAAcC,EAAwB,CAC9C,GAAI,SAAOA,GAAa,UAAYA,IAAa,MAAQ,MAAM,QAAQA,CAAQ,GAI/E,MAAO,SAAUA,GAAYA,EAAS,KAAOA,EAAS,KAAO,MAC9D,CAOA,SAASC,GAAMC,EAA2B,CACzC,OAAO,IAAI,QAASC,GAAQ,WAAWA,EAAKD,CAAE,CAAC,CAChD,CASA,SAASE,GAAQF,EAAYG,EAAW,EAAW,CAClD,OAAO,OAAO,YAAYH,EAAK,KAAM,QAAQG,CAAQ,CAAC,CACvD,CAMA,eAAeC,GAAyBC,EAAoB,CAC3D,GAAM,CAAE,OAAAC,CAAO,EAAI,MAAMC,EAAuC,EAC1DC,EAAc,EAAAC,QAAK,KAAKH,EAAQ,UAAU,EAE1CI,EAAU,GAAAC,QAAO,WAAW,QAAQ,EAC1C,OAAAD,EAAQ,OAAO,KAAK,UAAUL,CAAQ,CAAC,EAEhC,GAAGG,CAAW,IAAIE,EAAQ,OAAO,KAAK,CAAC,EAC/C,CAQA,eAAeE,GAAaP,EAAoBQ,EAAY,CAC3D,IAAMC,EAAgB,OAAOD,GAAY,SAAWA,EAAU,KAAK,UAAUA,CAAO,EAC9EE,EAAW,MAAMX,GAAyBC,CAAQ,EAClDW,EAAW,EAAAP,QAAK,QAAQM,CAAQ,EAEhC,MAAME,EAAWD,CAAQ,GAC9B,MAAM,EAAAE,QAAI,MAAMF,EAAU,CAAE,UAAW,EAAK,CAAC,EAG9C,MAAM,EAAAE,QAAI,UAAUH,EAAUD,EAAe,MAAM,CACpD,CAQA,eAAeK,GAAmBd,EAAoB,CACrD,GAAI,CACH,IAAMe,EAAO,MAAMhB,GAAyBC,CAAQ,EAC9CgB,EAAc,MAAM,EAAAH,QAAI,SAASE,EAAM,MAAM,EAGnD,OAFiB,KAAK,MAAMC,CAAW,CAGxC,MAAQ,CACP,OAAO,IACR,CACD,CEjSA,eAAeC,IAAsB,CACpC,MAAMC,GAA2B,CAChC,SAAUC,GACV,eAAgB,EACjB,CAAC,CACF,CAEA,eAAeC,IAAO,CACjBC,GAAyBC,KAC5B,MAAMC,EAAY,MAAM,EACxB,MAAMN,GAAoB,EAE5B,CAEAO,GAAiBJ,EAAI",
|
|
6
|
+
"names": ["import_node_path", "import_node_path", "import_promises", "import_node_path", "import_node_path", "import_node_fs", "import_node_path", "import_node_process", "import_node_url", "toPath", "urlOrPath", "findUpSync", "name", "options", "cwd", "process", "type", "stopAtOption", "directory", "path", "toPath", "root", "stopAt", "isAbsoluteName", "filePath", "stats", "fs", "pkgDirSync", "options", "cwd", "filePath", "findUpSync", "path", "env", "envIsTruthy", "value", "GRIDDO_API_URL", "GRIDDO_PUBLIC_API_URL", "GRIDDO_BOT_USER", "GRIDDO_BOT_PASSWORD", "GRIDDO_API_CONCURRENCY_COUNT", "GRIDDO_SKIP_BUILD_CHECKS", "GRIDDO_RENDER_BY_DOMAINS", "GRIDDO_BUILD_LOGS", "GRIDDO_BUILD_LOGS_TO_FILE", "GRIDDO_BUILD_LOGS_BUFFER_SIZE", "GRIDDO_SSG_VERBOSE_LOGS", "GRIDDO_SEARCH_FEATURE", "GRIDDO_ASSET_PREFIX", "GRIDDO_REACT_APP_INSTANCE", "GRIDDO_AI_EMBEDDINGS", "GRIDDO_VERBOSE_LOGS", "GRIDDO_USE_DIST_BACKUP", "GRIDDO_SSG_BUNDLE_ANALYZER", "RESET", "CODES", "brush", "color", "key", "text", "GriddoLog", "_GriddoLog", "str", "GRIDDO_VERBOSE_LOGS", "brush", "GRIDDO_BUILD_LOGS", "args", "root", "pkgDirSync", "path", "cache", "dbFilePath", "readDB", "customDBPath", "file", "fsp", "error", "GriddoLog", "writeDB", "renderDB", "import_promises", "LoginError", "pathExists", "dir", "fsp", "GRIDDO_API_URL", "GRIDDO_PUBLIC_API_URL", "AI_EMBEDDINGS", "ALERT", "DOMAINS", "GET_ALL", "GET_PAGE", "LOGIN", "RESET_RENDER", "ROBOTS", "SEARCH", "SETTINGS", "SITE_URI", "AuthService", "response", "LOGIN", "GRIDDO_BOT_USER", "GRIDDO_BOT_PASSWORD", "token", "throwError", "LoginError", "authService", "getRenderPathsHydratedWithDomainFromDB", "options", "domain", "dbFilePath", "paths", "readDB", "path", "import_promises", "import_node_path", "distRollback", "domain", "data", "readDB", "exportsDir", "exportsDirBackup", "GriddoLog", "path", "fsp", "pathExists", "RenderError", "originalError", "throwError", "options", "stack", "error", "message", "expected", "hint", "errorColor", "GriddoLog", "brush", "extraText", "withErrorHandler", "fn", "__root", "getRenderPathsHydratedWithDomainFromDB", "data", "readDB", "path", "distRollback", "writeDB", "import_node_crypto", "import_promises", "import_node_path", "import_promises", "logBuffer", "logFilePath", "flushPromise", "LOG_BUFFER_FLUSH_THRESHOLD", "GRIDDO_BUILD_LOGS_BUFFER_SIZE", "LOG_TO_FILE_IS_DISABLED", "GRIDDO_BUILD_LOGS", "addLogToBuffer", "lineContent", "LOG_TO_FILE_IS_DISABLED", "logBuffer", "LOG_BUFFER_FLUSH_THRESHOLD", "flushLogsToFile", "error", "GriddoLog", "flushPromise", "logFilePath", "logsToFlush", "fsp", "env", "RETRY_WAIT_SECONDS", "RETRY_ATTEMPTS", "requestAPI", "props", "method", "appendToLog", "endpoint", "body", "cacheKey", "attempt", "headers", "useApiCacheDir", "logToFile", "cacheOptions", "start", "cacheData", "searchCacheData", "siteId", "getSafeSiteId", "siteIdMsg", "duration", "msToSec", "addLogToBuffer", "fetchOptions", "authService", "response", "data", "saveCache", "e", "error", "GriddoLog", "RenderError", "showApiError", "delay", "postApi", "props", "endpoint", "body", "headers", "referenceFieldBodyParams", "requestAPI", "showApiError", "error", "options", "message", "stack", "callInfo", "callInfoArray", "item", "callInfoStr", "errorDetailsStr", "GriddoLog", "brush", "getSafeSiteId", "response", "delay", "ms", "res", "msToSec", "decimals", "generateFilenameWithHash", "petition", "__root", "getRenderPathsHydratedWithDomainFromDB", "apiCacheDir", "path", "hashSum", "crypto", "saveCache", "content", "stringContent", "filename", "filepath", "pathExists", "fsp", "searchCacheData", "file", "fileContent", "triggerAIEmbeddings", "postApi", "AI_EMBEDDINGS", "main", "GRIDDO_SEARCH_FEATURE", "GRIDDO_AI_EMBEDDINGS", "authService", "withErrorHandler"]
|
|
7
|
+
}
|
|
@@ -0,0 +1,66 @@
|
|
|
1
|
+
"use strict";var Xo=Object.create;var yt=Object.defineProperty;var Zo=Object.getOwnPropertyDescriptor;var Qo=Object.getOwnPropertyNames;var en=Object.getPrototypeOf,tn=Object.prototype.hasOwnProperty;var rn=(e,t,r,o)=>{if(t&&typeof t=="object"||typeof t=="function")for(let n of Qo(t))!tn.call(e,n)&&n!==r&&yt(e,n,{get:()=>t[n],enumerable:!(o=Zo(t,n))||o.enumerable});return e};var h=(e,t,r)=>(r=e!=null?Xo(en(e)):{},rn(t||!e||!e.__esModule?yt(r,"default",{value:e,enumerable:!0}):r,e));var wr=h(require("node:path"));var lr=require("node:child_process"),fe=h(require("node:fs/promises")),W=h(require("node:path"));var We=h(require("node:fs/promises")),Ee=h(require("node:path"));var Pt=h(require("node:path"));var Dt=h(require("node:fs"));var Q=h(require("node:path")),Rt=h(require("node:process")),St=require("node:url"),wt=e=>e instanceof URL?(0,St.fileURLToPath)(e):e;function bt(e,t={}){let{cwd:r=Rt.default.cwd(),type:o="file",stopAt:n}=t,i=Q.default.resolve(wt(r)??""),{root:s}=Q.default.parse(i),a=n?Q.default.resolve(i,wt(n)):s,d=Q.default.isAbsolute(e);for(;;){let m=d?e:Q.default.join(i,e);try{let f=Dt.default.statSync(m,{throwIfNoEntry:!1});if(o==="file"&&f?.isFile()||o==="directory"&&f?.isDirectory())return m}catch{}if(i===a||i===s)break;i=Q.default.dirname(i)}}function Et(e){let{cwd:t}=e||{},r=bt("package.json",{cwd:t});return r&&Pt.default.dirname(r)}var{env:R}=process;function k(e){if(!e)return!1;switch(e.trim().toLowerCase()){case"1":case"true":case"yes":case"y":case"on":return!0;default:return!1}}var ke=R.GRIDDO_API_URL||R.API_URL,_t=R.GRIDDO_PUBLIC_API_URL||R.PUBLIC_API_URL,It=R.botEmail||R.GRIDDO_BOT_USER,At=R.botPassword||R.GRIDDO_BOT_PASSWORD,He=Number.parseInt(R.GRIDDO_API_CONCURRENCY_COUNT||"10"),Wn=k(R.GRIDDO_SKIP_BUILD_CHECKS),xt=k(R.GRIDDO_RENDER_BY_DOMAINS),ie=k(R.GRIDDO_BUILD_LOGS),Jn=k(R.GRIDDO_BUILD_LOGS_TO_FILE),Tt=Number.parseInt(R.GRIDDO_BUILD_LOGS_BUFFER_SIZE||"500"),$t=k(R.GRIDDO_SSG_VERBOSE_LOGS),qn=k(R.GRIDDO_SEARCH_FEATURE),Ue=R.GRIDDO_ASSET_PREFIX||R.ASSET_PREFIX,Ot=R.GRIDDO_REACT_APP_INSTANCE||R.REACT_APP_INSTANCE,Vn=k(R.GRIDDO_AI_EMBEDDINGS),vt=k(R.GRIDDO_VERBOSE_LOGS),Kn=k(R.GRIDDO_USE_DIST_BACKUP),Ft=k(R.GRIDDO_SSG_BUNDLE_ANALYZER);var on="\x1B[0m",Ct={black:"\x1B[30m",red:"\x1B[31m",green:"\x1B[32m",yellow:"\x1B[33m",blue:"\x1B[34m",magenta:"\x1B[35m",cyan:"\x1B[36m",white:"\x1B[37m",gray:"\x1B[90m",bold:"\x1B[1m",dim:"\x1B[2m"},D={};for(let e in Ct){let t=e;D[t]=r=>`${Ct[t]}${r}${on}`}var c=class e{constructor(){}static verbose(...t){vt&&console.log(D.yellow("verbose"),D.dim(t.join(" ")))}static build(...t){ie&&e.log(...t)}static info(...t){console.log(`${D.blue("info")} ${t.join(" ")}`)}static success(...t){console.log(`${D.green("success")} ${t.join(" ")}`)}static error(...t){console.error(`${D.red("error")} ${t.join(" ")}`)}static warn(...t){console.warn(`${D.yellow("warn")} ${t.join(" ")}`)}static log(...t){console.log(...t)}};var nn=Et({cwd:Ee.default.resolve(__dirname,"../../..")})||"",sn=Ee.default.join(nn,".griddo/cache"),Lt=Ee.default.join(sn,"db.json");async function b(e=""){let t=e||Lt;try{return JSON.parse(await We.default.readFile(t,"utf-8"))}catch(r){throw c.error(`Failed to read DB file at ${t}:`,r),r}}async function L(e,t=""){let r=t||Lt;try{await We.default.writeFile(r,JSON.stringify(e,null," "))}catch(o){throw c.error(`Failed to write DB file at ${r}:`,o),o}}var P=h(require("node:fs/promises")),q=h(require("node:path"));var se={error:"ArtifactError",message:"There was a problem with an artifact",expected:"An external process may have has modified or deleted one of the artifacts (files and directories).",hint:"Have there been any recent deployments? These can delete directories from the current render."};var jt=(e,t)=>({error:"LifecycleExecutionError",message:`Exceeded maximum retry attempts (${e}) for ${t} LifeCycle`}),Gt={error:"LoginError",message:"There was a problem logging in to the API",expected:"This happens if the API is currently not working or the credentials are incorrect."},Nt={error:"NoDomainsFoundError",message:"No domains were found in this instance. The process cannot continue.",expected:"This may happen if the API is not functioning, or the site is not properly configured, or the domains are not registered.",hint:"You can contact the instance administrator."};var Bt={error:"RenderUUIDError",message:`Render sentinel file does not exist.
|
|
2
|
+
The rendering UUID cannot be read safely.
|
|
3
|
+
There was probably an instance deployment during the render, and files were deleted.
|
|
4
|
+
|
|
5
|
+
The files generated in this render will not be published.`},Mt={error:"WriteToStoreError",message:"There was an error writing a file to the Store directory",hint:"There may be an issue such as lack of space or permissions preventing the file from being written."};async function Je(e){if(!await w(e))return;let t=(await P.default.readdir(e,{withFileTypes:!0})).filter(r=>r.isDirectory());for(let r of t){let o=q.default.join(e,r.name);await an(o)&&await P.default.rm(o,{recursive:!0})}}async function kt(e,t){for(let r of e)try{await w(r)||(await P.default.mkdir(r,{recursive:!0,...t}),c.verbose(`create directory: ${r}`))}catch(o){A(se,o)}}async function Ht(e,t,r,o={withBackup:!1}){let{withBackup:n}=o;for(let i of r){let s=q.default.join(e,i),a=q.default.join(t,i);if(!await w(s)){c.info(`(Maybe first render) Source directory does not exist: ${s}`);continue}n&&(await Jt(a),c.verbose(`create backup: ${a}`));try{await w(a)&&(await P.default.rm(a,{recursive:!0,force:!0}),c.verbose(`clean destination: ${a}`)),await P.default.cp(s,a,{recursive:!0,preserveTimestamps:!0}),c.verbose(`copy: ${s} to ${a}`),n&&(await Wt(a),c.verbose(`delete backup: ${a}`))}catch(d){n&&(await Ut(a),c.verbose("Backup has been restored.")),A(se,d)}}}async function ee(e,t,r,o){let{override:n,withBackup:i}=o||{};for(let s of r){let a=q.default.join(e,s),d=q.default.join(t,s);if(await w(a)){i&&await Jt(d);try{n&&await w(d)&&await P.default.rm(d,{recursive:!0,force:!0}),await P.default.rename(a,d),c.verbose(`moved: ${a} to ${d}`),i&&await Wt(d)}catch(m){i&&(await Ut(d),c.info("Backup has been restored.")),A(se,m)}}}}async function ae(e){for(let t of e)try{await P.default.rm(t,{recursive:!0,force:!0}),c.verbose(`artifact removed: ${t}`)}catch(r){A(se,r)}}async function Ut(e,t="-BACKUP"){let r=e+t;try{await P.default.rename(r,e),c.info(`Backup ${r} has been restored`)}catch{throw new Error(`Error while delete ${r} backup`)}}async function Wt(e,t="-BACKUP"){let r=e+t;if(await w(r))try{await P.default.rm(r,{recursive:!0,force:!0}),c.verbose(`Backup ${r} has been deleted`)}catch{throw new Error(`Error while delete ${r} backup`)}}async function Jt(e,t="-BACKUP"){let r=e+t;if(await w(e)){if(await w(r)){c.warn(`Destination ${r} already exists`);return}try{await P.default.rename(e,r),c.verbose(`Backup of ${e} has been created in ${r}`)}catch(o){c.error(`Error while coping ${e} to ${r} backup`),A(se,o)}}}async function an(e){let t=(await P.default.readdir(e,{withFileTypes:!0,recursive:!0})).filter(o=>o.isFile()&&!q.default.basename(o.name).startsWith(".")),r=t.filter(o=>o.name.endsWith(".xml"));if(t.length===r.length)return!0}async function qe(e){try{if(!(await P.default.stat(e)).isDirectory())return;let r;try{r=await P.default.readdir(e)}catch(n){if(n.code==="ENOENT"){c.warn(`The directory "${e}" does not exist, skipping it.`);return}throw c.error(`Error al leer el directorio "${e}":`,n),n}for(let n of r){let i=q.default.join(e,n);await qe(i)}if((await P.default.readdir(e)).length===0)try{await P.default.rmdir(e),c.verbose(`Remove empty directory: ${e}`)}catch(n){n.code==="ENOENT"?c.warn(`El directorio "${e}" ya no existe. Posiblemente fue borrado por otra operaci\xF3n.`):c.error(`Error al borrar el directorio "${e}":`,n)}}catch(t){if(t.code==="ENOENT")c.warn(`The directory "${e}" does not exist or has already been processed.`);else throw c.error(`General error general while processing "${e}":`,t),t}}async function w(e){try{return await P.default.access(e),!0}catch{return!1}}async function*pe(e,t){let r=await P.default.opendir(e);for await(let o of r){let n=q.default.join(e,o.name);o.isDirectory()?yield*pe(n,t):o.isFile()&&o.name.endsWith(t)&&(yield n)}}var T={FROM_SCRATCH:"FROM_SCRATCH",INCREMENTAL:"INCREMENTAL",IDLE:"IDLE",ERROR:"ERROR",COMPLETED:"COMPLETED"};var H=process.env.GRIDDO_API_URL,cn=process.env.GRIDDO_PUBLIC_API_URL,di=`${H}/ai/embeddings`,li=`${cn}/alert`,qt=`${H}/domains`,Vt=`${H}/sites/all`,Kt=`${H}/page`,zt=`${H}/login_check`,mi=`${H}/debug/reset-render`,Yt=`${H}/domains/robots`,pi=`${H}/search`,Xt=`${H}/settings`,te=`${H}/site/`,Zt=[te,"/build/end"],Qt=[te,"/build/start"];var er=[te,"/distributor"],tr=[te,"/sitemap"],rr=[te,"/all"],or=[te,"/languages"],nr=[te,"/socials"];var Ve=class{headers;async login(){try{let t=await fetch(zt,{method:"POST",headers:{"Content-Type":"application/json",Connection:"close"},body:JSON.stringify({username:It,password:At})});if(!t.ok)throw new Error("Error while login in the API");let{token:r}=await t.json();return this.headers={Authorization:`bearer ${r}`,"Cache-Control":"no-store"},this.headers}catch(t){A(Gt,t)}}},V=new Ve;var ce=h(require("node:fs/promises")),U=h(require("node:path"));async function _e(e){let t=await b(),{sitesToPublish:r,createdPages:o,buildProcessData:n}=t.domains[e].renderInfo||{};if(!r||!o||!n)throw new Error("Build metadata not found");return{buildProcessData:n,createdPages:o,sitesToPublish:r}}async function ir(e,t){let r=await b();r.domains[t]=r.domains[t]||{},r.domains[t].renderInfo=e,await L(r)}async function sr(e,t){let{__root:r}=await y();try{let o=new Set(["editorID","parentEditorID"]);for(let n of t){ln(n,o);let i=`${n.context.page.id}.json`,s=U.default.join(r,"store",e,i);await dn(s,JSON.stringify(n))}}catch(o){A(Mt,o)}}async function ar(e,t){let{__root:r}=await y(),o=async n=>{for(let i of n){let s=U.default.join(r,"store",e,`${i}.json`);try{await w(s)&&(await ce.default.unlink(s),c.verbose(`Removed file: ${s}`))}catch{throw new Error(`Error removing file ${i}`)}}};for(let n of t)await o(n)}async function cr(e,t){let{__root:r}=await y({domain:t}),o=U.default.join(r,"store");if(!await w(o))return;let n=await ce.default.readdir(o),i=e.map(({id:a})=>`${a}`),s=n.filter(a=>!i.includes(a));for(let a of s)await ce.default.rm(U.default.join(o,a),{recursive:!0,force:!0}),c.verbose(`Removed orphan site id: ${a}`)}async function dr(e,t){let{__root:r}=await y(),o=U.default.join(r,"store",e),n=await ce.default.readdir(o),i=[];for(let s of n){let a=s.split(".")[0],d=Number.parseInt(a);!t.includes(d)&&d>0&&i.push(d)}return i}async function dn(e,t){let r=U.default.dirname(e),o=U.default.extname(e),n=U.default.basename(e,o),i=e,s=1;for(;await w(i);){let a=`${n}-${s}${o}`;i=U.default.join(r,a),s++}await ce.default.writeFile(i,t)}function ln(e,t){function r(o){if(!(!o||typeof o!="object"||Array.isArray(o))){for(let n in o)if(Object.hasOwn(o,n))if(t.has(n))delete o[n];else{let i=o[n];typeof i=="object"&&i!==null&&!Array.isArray(i)&&r(i)}}}r(e)}async function mr(e){let{domain:t}=e,r=await b();r.domains[t].isRendering=!0,await L(r);let{__ssg:o}=await y(),n=W.default.join(o,`.render-sentinel-${t}`);await fe.default.writeFile(n,new Date().toISOString())}async function pr(e){let t=await b();t.domains[e].isRendering=!1,t.currentRenderingDomain=null,t.domains[e].renderMode="COMPLETED",await L(t);let{__ssg:r}=await y(),o=W.default.join(r,`.render-sentinel-${e}`);await fe.default.unlink(o)}async function fr(e){let{__ssg:t}=await y(),r=W.default.join(t,`.render-sentinel-${e}`);await w(r)||A(Bt)}async function ur(e){let{basePath:t}=e,r=(0,lr.execSync)("git rev-parse HEAD").toString().trim();await fe.default.writeFile(W.default.join(t,"commit"),r)}async function gr(e){let t=await b();if(!t.domains[e])throw new Error(D.red(`[!] Error: Domain ${e} not found in DB`));if(!t.domains[e].renderMode)throw new Error(D.red(`[!] Error: Render mode not found for domain ${e}`));return{renderMode:t.domains[e].renderMode,reason:t.domains[e].renderModeReason}}async function y(e){let{domain:t,dbFilePath:r}=e||{},n=(await b(r)).paths;return{__root:n.root,__cache:W.default.join(n.cxCache,t||""),__components:n.components,__cx:n.cx,__sites:n.exportsDir,__exports:W.default.join(n.exportsDir,t||""),__exports_backup:W.default.join(n.exportsDirBackup,t||""),__ssg:n.ssg,__exports_dist:W.default.join(n.exportsDir,t||"","dist")}}async function Ke(){let e=await b();return{griddoVersion:e.griddoVersion,buildReportFileName:e.buildReportFileName}}async function hr(e){let t=await V.login(),{__root:r}=await y(),{buildReportFileName:o}=await Ke(),{buildProcessData:n}=await _e(e),i=Object.keys(n).map(d=>({...n[d],siteId:Number.parseInt(d)})),s={authControl:t,sites:i},a=W.default.join(r,"current-dist",o);await fe.default.writeFile(a,JSON.stringify(s)),c.verbose(`build report saved in ${a}`)}var ze=h(require("node:fs/promises")),Y=h(require("node:path"));async function yr(e){let t=await b(),{exportsDir:r,exportsDirBackup:o}=t.paths;c.info(`Cleaning exports dir for the domain ${e}`),c.verbose(`Deleting ${Y.default.join(r,e)}...`),await ze.default.rm(Y.default.join(r,e),{recursive:!0,force:!0}),await w(Y.default.join(o,e))?(await ze.default.cp(Y.default.join(o,e),Y.default.join(r,e),{recursive:!0}),c.info(`export-backup dir for the domain ${e} found. Restoring before exit...`),c.verbose(`Copying ${Y.default.join(o,e)} -> ${Y.default.join(r,e)}...`)):c.info("No export-backup found, skipping rollback. Next render will create a new exports dir from scratch...")}var de=class extends Error{constructor(t){super(t instanceof Error?t.message:String(t)),this.name="InternalCXError",this.stack=t instanceof Error?t.stack:""}};function A(e,t){let{error:r,message:o,expected:n,hint:i}=e,s=c.log(D.red(`[ ${r} ]`)),a=[n,i].filter(Boolean).join(`
|
|
6
|
+
`);throw c.log(`
|
|
7
|
+
${s}
|
|
8
|
+
${o}
|
|
9
|
+
${a}
|
|
10
|
+
|
|
11
|
+
${D.red("stack")}
|
|
12
|
+
${JSON.stringify(t,null,2)}`),new de(t)}async function Dr(e){try{await e()}catch(t){t instanceof de?c.error("Internal Griddo RenderError"):t instanceof Error?c.error(t.message):c.error(`An unexpected error occurred ${t}`);try{let{__root:o}=await y(),n=await b();n.needsRollbackOnError?(c.info("Cleaning exports dir..."),c.verbose(`Deleting ${wr.default.join(o,"exports")}...`),await yr(n.currentRenderingDomain)):c.info("No rollback needed, skipping...")}catch{c.info("Early render stage, no db.json created yet...")}let r=await b();throw r.domains[r.currentRenderingDomain].isRendering=!1,r.domains[r.currentRenderingDomain].renderMode="ERROR",await L(r),t}}var Rr=h(require("node:fs/promises")),Sr=h(require("node:path"));async function br(){let{version:e}=JSON.parse(await Rr.default.readFile(Sr.default.resolve(__dirname,"../..","package.json"),"utf-8")),t=`
|
|
13
|
+
${D.yellow(`Griddo Exporter ${e}`)}
|
|
14
|
+
${D.dim(`Node ${process.version.slice(1)}`)}
|
|
15
|
+
`;c.log(t)}var Or=h(require("node:crypto")),Ae=h(require("node:fs/promises")),Qe=h(require("node:path"));var he=h(require("node:fs/promises")),ge=h(require("node:path"));var le=[],Ie=null,ue=null,mn=Tt,Xe="render-detail-log.txt",Pr=!ie,Ye=null;async function Er(){return Ye||(Ye=await b()),Ye}async function _r(){if(Pr)return;let e=await Er();Ie=ge.default.join(e.paths.root,Xe),await he.default.rm(Ie,{force:!0})}function Ze(e){Pr||(le.push(e.toString()),le.length>=mn&&Ir().catch(t=>{c.error("Background log flush failed:",t)}))}async function Ir(){if(ue&&await ue,le.length===0||!Ie)return;ue=(async()=>{let t=[...le];le.length=0;try{await he.default.appendFile(Ie,`${t.join(`
|
|
16
|
+
`)}
|
|
17
|
+
`)}catch(r){le.unshift(...t),c.error("Error flushing logs:",r)}})();try{await ue}finally{ue=null}}async function Ar(e){await Ir();let t=await Er(),r=pn(),o=ge.default.join(t.paths.exportsDir,e,"logs");await he.default.mkdir(o,{recursive:!0});let n=ge.default.join(t.paths.root,Xe),i=ge.default.join(o,`${r}-${Xe}`);try{await he.default.cp(n,i)}catch(s){if(s instanceof Error&&"code"in s&&s.code!=="ENOENT")throw s}}function pn(){let e=new Date,t=e.getFullYear(),r=String(e.getMonth()+1).padStart(2,"0"),o=String(e.getDate()).padStart(2,"0"),n=String(e.getHours()).padStart(2,"0"),i=String(e.getMinutes()).padStart(2,"0"),s=String(e.getSeconds()).padStart(2,"0");return`${t}-${r}-${o}_${n}-${i}-${s}`}var{env:fn}=process,{RETRY_WAIT_SECONDS:un="4",RETRY_ATTEMPTS:xr="4"}=fn;async function et(e,t,r=""){let{endpoint:o,body:n,cacheKey:i="",attempt:s=1,headers:a,useApiCacheDir:d=!0,logToFile:m=!0}=e,f={endpoint:o,body:n,headers:a,cacheKey:i};if(i&&d){let l=new Date,p=await wn(f);if(p){if(m){let u=Tr(p),g=u?`site: ${u}`:"",I=$r(Date.now()-l.getTime());Ze(`${t} (cache) ${g} ${o} - ${I}s ${r}`)}return p}}try{let l=new Date,p={method:t.toUpperCase(),headers:Object.assign({},a,V.headers)};t.toLowerCase()!=="get"&&n&&(p.body=JSON.stringify(n),p.headers||(p.headers={}),p.headers["Content-Type"]="application/json");let u=await fetch(o,p);if(!u.ok){if(u.status===404)return null;throw new Error(`HTTP ${u.status}: ${u.statusText}`)}let g=await u.json();if(m){let I=Tr(g),j=I?`site: ${I}`:"",S=$r(Date.now()-l.getTime());Ze(`${t} (fetch) ${j} ${o} - ${S}s ${r}`)}return d&&await yn(f,g),g}catch(l){let p=l;if(s>parseInt(xr))throw c.log(`
|
|
18
|
+
Max attempts ${xr} reached
|
|
19
|
+
--------------------------------------
|
|
20
|
+
- ${t.toUpperCase()} ${o}
|
|
21
|
+
- BODY: ${JSON.stringify(n)}
|
|
22
|
+
- HEADERS: ${JSON.stringify(a)}
|
|
23
|
+
- ERROR: ${p.message}
|
|
24
|
+
--------------------------------------
|
|
25
|
+
`),new de(p);return gn(p,{callInfo:{endpoint:o,body:n}}),c.warn(`Waiting for retry: ${t}`,o),await hn(parseInt(un)*1e3),et({endpoint:o,body:n,headers:a,cacheKey:i,attempt:s+1},t,r)}}async function $(e){return et(e,"get")}async function tt(e){let{endpoint:t,body:r,headers:o}=e,n=t.endsWith("/distributor")&&`# ReferenceField body: ${JSON.stringify(r)} lang: ${JSON.stringify(o?.lang)}`;return et(e,"post",n||"")}function gn(e,t){let{message:r,stack:o}=e,{callInfo:n}=t,i=[];for(let d of Object.keys(n))i.push(`${d}: ${typeof n[d]=="object"?JSON.stringify(n[d]):n[d]}`);let s=i.join(`
|
|
26
|
+
`),a=`${r}
|
|
27
|
+
${o}`;c.warn(D.red(`
|
|
28
|
+
=============
|
|
29
|
+
|
|
30
|
+
{ Call info }
|
|
31
|
+
${s}
|
|
32
|
+
|
|
33
|
+
{ Error details }
|
|
34
|
+
${a}
|
|
35
|
+
|
|
36
|
+
=============
|
|
37
|
+
`))}function Tr(e){if(!(typeof e!="object"||e===null||Array.isArray(e)))return"site"in e&&e.site?e.site:void 0}function hn(e){return new Promise(t=>setTimeout(t,e))}function $r(e,t=3){return Number.parseFloat((e/1e3).toFixed(t))}async function vr(e){let{__root:t}=await y(),r=Qe.default.join(t,"apiCache"),o=Or.default.createHash("sha256");return o.update(JSON.stringify(e)),`${r}/${o.digest("hex")}`}async function yn(e,t){let r=typeof t=="string"?t:JSON.stringify(t),o=await vr(e),n=Qe.default.dirname(o);await w(n)||await Ae.default.mkdir(n,{recursive:!0}),await Ae.default.writeFile(o,r,"utf8")}async function wn(e){try{let t=await vr(e),r=await Ae.default.readFile(t,"utf8");return JSON.parse(r)}catch{return null}}async function Fr(){let e=await $({endpoint:qt,useApiCacheDir:!1});e.length||A(Nt),c.verbose(`getting domains names (${e.length})`);let t=e.filter(({slug:r})=>!!r).map(({slug:r})=>r.replace("/",""));return[...new Set(t)]}var Go=h(require("node:path"));async function Dn(e){let t=process.hrtime.bigint();await Promise.resolve(e());let r=process.hrtime.bigint();return Number((Number(r-t)/1e9).toFixed(3))}async function F(e,t,r){let{maxRetrys:o=3,skip:n}=r||{};if(n){c.info(`Skiping life-cycle ${e}. Reason: ${n}
|
|
38
|
+
`);return}let i=0;for(;i<o;)try{c.info(`start ${e} life-cycle`);let s=await Dn(t);c.success(`${e} - ${s}s
|
|
39
|
+
`);break}catch(s){let a=D.red(` Error in ${e} LifeCycle `),d=D.yellow(`Attempt (${i+1})`);c.log(`
|
|
40
|
+
${a} ${d}
|
|
41
|
+
`),c.log(s),c.log(),i++}i===o&&A(jt(o,e))}var C=h(require("node:path"));async function Cr(e){let{__exports:t,__cache:r,__root:o,__ssg:n}=await y({domain:e});return{initials:[t,r,C.default.join(t,"logs"),C.default.join(o,"store"),C.default.join(o,"apiCache")],disposables:[C.default.join(o,"store"),C.default.join(o,"apiCache"),C.default.join(o,"dist"),C.default.join(o,"dist-restored"),C.default.join(o,"assets"),C.default.join(o,"render-detail-log.txt"),C.default.join(o,"current-dist"),C.default.join(n,"domains.json"),C.default.join(n,"render-metadata.json")],cacheables:[],archivables:["assets"],restaurable:["dist"]}}var xe=class{domain;renderMode;pathsHydratedWithDomain;pagesToCreate=[];pagesToDelete=[];ssg;renderMetadata;renderArtifacts;constructor({domain:t,renderMode:r,ssg:o,pathsHydratedWithDomain:n,renderMetadata:i,renderArtifacts:s}){this.domain=t,this.renderMode=r,this.pathsHydratedWithDomain=n,this.renderMetadata=i,this.renderArtifacts=s,this.ssg=o}};var Te=h(require("node:fs/promises")),$e=h(require("node:path"));async function Lr(e){let{renderArtifacts:t,ssg:{ssgArtifacts:r},pathsHydratedWithDomain:{__cache:o,__exports:n},renderMode:i}=e;i===T.FROM_SCRATCH&&(await Te.default.rm($e.default.join(o,".cache"),{recursive:!0,force:!0}),await Te.default.rm($e.default.join(n,"dist"),{recursive:!0,force:!0}),await Te.default.rm($e.default.join(n,"assets"),{recursive:!0,force:!0})),await ae(t.disposables),await ae(r.disposables)}async function jr(e){let{renderArtifacts:t,ssg:{ssgArtifacts:r},pathsHydratedWithDomain:{__cache:o}}=e;await ae(t.disposables),await ae(r.disposables),await ur({basePath:o})}var ve=h(require("node:fs/promises")),De=h(require("node:path"));var rt=class{value;next;constructor(t){this.value=t}},ye=class{#e;#t;#r;constructor(){this.clear()}enqueue(t){let r=new rt(t);this.#e?(this.#t.next=r,this.#t=r):(this.#e=r,this.#t=r),this.#r++}dequeue(){let t=this.#e;if(t)return this.#e=this.#e.next,this.#r--,t.value}peek(){if(this.#e)return this.#e.value}clear(){this.#e=void 0,this.#t=void 0,this.#r=0}get size(){return this.#r}*[Symbol.iterator](){let t=this.#e;for(;t;)yield t.value,t=t.next}*drain(){for(;this.#e;)yield this.dequeue()}};function we(e){Gr(e);let t=new ye,r=0,o=()=>{r<e&&t.size>0&&(r++,t.dequeue()())},n=()=>{r--,o()},i=async(d,m,f)=>{let l=(async()=>d(...f))();m(l);try{await l}catch{}n()},s=(d,m,f)=>{new Promise(l=>{t.enqueue(l)}).then(i.bind(void 0,d,m,f)),r<e&&o()},a=(d,...m)=>new Promise(f=>{s(d,f,m)});return Object.defineProperties(a,{activeCount:{get:()=>r},pendingCount:{get:()=>t.size},clearQueue:{value(){t.clear()}},concurrency:{get:()=>e,set(d){Gr(d),e=d,queueMicrotask(()=>{for(;r<e&&t.size>0;)o()})}},map:{async value(d,m){let f=Array.from(d,(l,p)=>this(m,l,p));return Promise.all(f)}}}),a}function Gr(e){if(!((Number.isInteger(e)||e===Number.POSITIVE_INFINITY)&&e>0))throw new TypeError("Expected `concurrency` to be a number from 1 and up")}var Jr=h(require("node:fs/promises")),qr=h(require("node:path"));async function Nr(e){return $({endpoint:`${Vt}?domainSlug=${e}`})}async function Br(e,t){return $({endpoint:`${Kt}/${e}`,cacheKey:t})}async function Mr(e,t=""){let[r,o]=rr;return $({endpoint:`${r}${e}${o}`,cacheKey:t})}async function ot(e,t=""){let[r,o]=or;return $({endpoint:`${r}${e}${o}`,cacheKey:t})}async function nt(e){let[t,r]=Qt;return $({endpoint:`${t}${e}${r}`})}async function kr(e,t){let[r,o]=Zt;await tt({endpoint:`${r}${e}${o}`,body:t,useApiCacheDir:!1,logToFile:!1})}async function Hr(e,t,r,o,n){let[i,s]=er,a=o||e.site,d=n||e.language;return tt({endpoint:`${i}${a}${s}`,body:t,headers:{lang:d},cacheKey:r})}async function Ur(e){let{siteId:t,headers:r}=e,[o,n]=tr;return $({endpoint:`${o}${t}${n}`,headers:r})}async function Wr(e,t=""){let[r,o]=nr;return $({endpoint:`${r}${e}${o}`,cacheKey:t})}async function Vr(e){let t=await Nr(e);if(t.length)for(let n of t){let{items:i}=await ot(n.id);n.domains=i.filter(s=>s.domain&&(s.domain.slug===e||s.domain.slug===`/${e}`)).map(s=>({[s.id]:`${s.domain.slug}${s.path}`}))}let r=t.filter(n=>!!n.isPublished),o=t.filter(n=>!n.isPublished&&n.shouldBeUpdated);return{sitesToPublish:r,sitesToUnpublish:o}}async function Kr(e){let{__root:t}=await y();for(let r of e){let o=await nt(r.id),{siteHash:n}=o,i={siteHash:n,publishHashes:[],unpublishHashes:[],publishPagesIds:[]};await kr(r.id,i),await Jr.default.rm(qr.default.join(t,"store",r.id.toString()),{force:!0,recursive:!0})}}async function zr(e){let t=await nt(e),r=await Mr(e),o=await ot(e),n=await Wr(e),i=o.items,s=i.find(g=>g.isDefault),{siteHash:a,unpublishHashes:d,publishIds:m}=t,{headers:f,footers:l}=r;return{siteInfo:r,validPagesIds:m,siteHash:a,unpublishHashes:d,siteLangs:i,defaultLang:s,headers:f,footers:l,socials:n}}var Oe=class{_defaultHeaders;_defaultFooters;_navigations;constructor(){this._navigations={footers:[],headers:[]},this._defaultHeaders={},this._defaultFooters={}}set navigations(t){this._navigations=t,this._defaultFooters=this.getDefaultFooters(),this._defaultHeaders=this.getDefaultHeaders()}get navigations(){return this._navigations}getDefaultFooters(){let r=[...this.navigations.footers].filter(n=>n.setAsDefault),o={};for(let n of r)o[n.language]=n;return o}getDefaultHeaders(){let r=[...this.navigations.headers].filter(n=>n.setAsDefault),o={};for(let n of r)o[n.language]=n;return o}getRightLanguage(t,r,o){if(!t||!r)return null;let i=t.find(s=>s.language===o&&s.navigationLanguages?.find(a=>a.navigationId===r))||t.find(s=>s.id===r);return i?{...i}:null}getPageHeader(t,r){return this.getRightLanguage(this.navigations.headers,t,r)}getPageFooter(t,r){return this.getRightLanguage(this.navigations.footers,t,r)}getPageNavigations(t){let{header:r,footer:o,language:n,template:i,templateConfig:s}=t,{templateType:a}=i,{defaultHeader:d,defaultFooter:m,templates:f}=s,l=j=>{let S=j.find(x=>typeof x=="number");return typeof S=="number"?S:null},p=l([r,f?.[a]?.defaultHeader,d]),u=l([o,f?.[a]?.defaultFooter,m]),g=p?this.getPageHeader(p,n):p===0?null:this._defaultHeaders[n],I=u?this.getPageFooter(u,n):u===0?null:this._defaultFooters[n];return{header:g,footer:I}}};function it(e,t,r,o="jpg"){let n=typeof e=="string"?e:e?.url;return n?n.split("/")[2].includes("cloudinary.com")?Sn(n,`c_fill,w_${t},h_${r}`):Rn(n,`f/${o}/w/${t}/h/${r}`):null}function Rn(e,t){let r=e.split("/"),o=r.slice(0,-1).join("/"),n=r.slice(-1)[0];return`${o}/${t}/${n}`}function Sn(e,t){let r=e.replace("https://",""),o=r.split("/").slice(0,4).join("/"),n=r.replace(o,"");return`https://${o}/${t}${n}`}var bn=25;function Pn({socialTitle:e,socialDescription:t,socialImage:r}){return{type:"website",title:e,description:t,image:r?it(r,1280,768):"",twitterImage:r?it(r,1280,768):""}}function En(e){let{title:t,metaTitle:r,metaDescription:o,canonicalURL:n,locale:i,url:s,isIndexed:a,follow:d,metasAdvanced:m,pageLanguages:f,fullUrl:l,metaKeywords:p}=e,u=m?.split(",").filter(Boolean).map(g=>g.trim().toLowerCase())||[];return{title:(r||t||"").trim(),description:o,canonical:n?.trim()&&n!==l?n.trim():a?l:void 0,locale:i,url:s,index:a?"index":"noindex",follow:d?"follow":"nofollow",translate:u.includes("notranslate")?"notranslate":"",metasAdvanced:u.filter(g=>g!=="notranslate").join(),pageLanguages:f,metaKeywords:p?.filter(Boolean).map(g=>g.replace(/"/g,"'")).join(", ")}}async function at(e,t){let{id:r,title:o,fullPath:n,language:i,breadcrumb:s,socialDescription:a,socialImage:d,socialTitle:m}=e,{baseUrl:f,cloudinaryName:l,griddoVersion:p,siteLangs:u,siteMetadata:g,siteOptions:I,siteScript:j,siteSlug:S,socials:x,theme:B,navigations:G}=t,{header:me,footer:Z}=G;e.breadcrumb=s,e.siteSlug=S,e.apiUrl=f,e.publicApiUrl=t.publicBaseUrl,e.instance=t.instance;let J=u.find(({id:Ne})=>Ne===e?.language)?.locale.replace(/_/g,"-"),K=En(e),z=Pn({socialDescription:a,socialImage:d,socialTitle:m}),oe=n.compose,Ge=new Date().toString();return{path:oe,size:void 0,context:{id:r,title:o,fullPath:n,locale:J,languageId:i,theme:B,siteMetadata:g,pageMetadata:K,openGraph:z,socials:x,siteLangs:u,cloudinaryName:l,siteOptions:I,griddoVersion:p,renderDate:Ge,siteScript:j,header:me,footer:Z,page:e}}}async function Zr(e,t){return at(e,t)}async function Qr({page:e,pages:t,isRoot:r=!1,defaultLang:o,template:n,totalQueriedItems:i},s){let a=t.map(async(d,m)=>{let f=m===0,l=m+1,{domainUrl:p,compose:u}=e.fullPath,g={...e,fullPath:{...e.fullPath,compose:st(u,l,{addEndingSlash:!0})},fullUrl:st(e.fullUrl,l,{addEndingSlash:!0}),slug:st(e.slug,l),title:Xr(e.title,l),metaTitle:Xr(e.metaTitle||"",l),disableHrefLangs:l>1,template:{...n,isFirstPage:f,pageNumber:l,totalPages:t.length,baseLink:`${p}${u}`,queriedItems:d,totalQueriedItems:i?.length},isRoot:r,defaultLang:o};return at(g,s)});return Promise.all(a)}function eo(e,t){let{multiPageElements:r,...o}=e,n=JSON.parse(JSON.stringify(r));n.find(({sectionSlug:s})=>s==="/")||n.push({});let i=n.map(async s=>{let a=JSON.parse(JSON.stringify(o)),{sectionSlug:d="/",title:m="",metaTitle:f="",metaDescription:l=""}=s,p=typeof m=="string"?m:m.content,u=a.fullPath.compose||"",g=a.fullUrl.endsWith("/")?a.fullUrl.slice(0,-1):a.fullUrl,I=d?.replace(/\//g,""),S=d==="/"?I:`${I}/`,x=u.endsWith("/")?"":"/",B=`${u}${x}${S}`;return p.trim()&&(a.title=p),l.trim()&&(a.metaDescription=l),a.fullUrl=`${g}/${S}`,a.fullPath.compose=B,a.slug=B,a.template.activeSectionSlug=d,a.template.activeSectionBase=g,a.metaTitle=f.trim()||p.trim()||a.metaTitle,at(a,t)});return Promise.all(i)}function to(e){return new Promise(r=>{let o=(n,i=0)=>{if(!(!n||typeof n!="object")){for(let s in n){let a=n[s],d=a||typeof a=="object",m=JSON.stringify(a).includes('"hasGriddoMultiPage":true');if(!d||!m)continue;let{component:f,hasGriddoMultiPage:l,elements:p}=a;f&&l&&r(p||[]),o(a,i+1)}i||r(null)}};o([e])})}function _n(e,t,r){return t?.slice(e*(r-1),e*r)}function In(e,t){let r=Math.ceil(t.length/e)||1;return Array.from({length:r},(n,i)=>i+1)?.map(n=>_n(e,t,n))}function ro(e){let t=e.queriedItems||[],r=e?.itemsPerPage||bn;return In(r,t)}function Yr(e){return e.replace(/\/+$/,"/")}function st(e,t,r){let o=e.endsWith("/")?"":"/",n=r?.addEndingSlash?"/":"";return t<=1?Yr(`${e}${n}`):Yr(`${e}${o}${t}${n}`)}function Xr(e,t){return!e||t<=1?e:`${e} - ${t}`}function An(e,t){let{order:r,sources:o,quantity:n,mode:i,fixed:s,fullRelations:a=!1,allLanguages:d=!1,preferenceLanguage:m=!1,referenceId:f,fields:l}=e;return i==="auto"?{mode:i,order:r,sources:o,quantity:n,fullRelations:a,allLanguages:d,preferenceLanguage:m,fields:l}:i==="manual"?{mode:i,fixed:s,fullRelations:a,fields:l}:i==="navigation"?{mode:i,order:r,quantity:n,fullRelations:a,referenceId:f||t?.structuredDataContent?.id,fields:l}:(c.error(`ReferenceField mode: ${i} is not recognized on page ${t?.id}.`),e)}async function xn(e){let{page:t,component:r,cacheKey:o}=e,{data:n}=r;if(!n)return c.info(`Error: Page ${t.id} has \`hasDistributorData: true\` or \`getStaticData: true\` but it doesn't have a \`data\` property.`),[];if(Array.isArray(n.sources)&&n.sources.length<1)return c.info(`Warning: Page with id: ${t.id} has a ReferenceField with empty \`data.sources\``),[];let{site:i,lang:s}=n;!n.sources&&n.mode==="auto"&&c.info(`Warning: Page with id: ${t.id} has a ReferenceField with \`undefined\` \`data.sources\``);let a=An(n,t);return await Hr(t,a,o,i,s)}async function oo({page:e,cacheKey:t=""}){try{let{template:r}=e,o=async(s,a=1)=>{if(!(!s||typeof s!="object"||!(JSON.stringify(s).includes('"hasDistributorData":true')||JSON.stringify(s).includes('"getStaticData":true'))))for(let m in s){if(m==="queriedItems")continue;let l=s[m];if(!l||typeof l!="object")continue;(l.hasDistributorData||l.getStaticData)&&(l.queriedItems=await xn({page:e,cacheKey:t,component:l})),await o(l,a+1)}};return await(async s=>(await o([s]),s))(r)}catch(r){throw c.error("Error in getReferenceFieldData:",r),r}}var Tn=Date.now().toString();function $n(){return!1}function no(e,t){let r=Math.max(...t.map(i=>i.name.length+(i.shouldBeUpdated?1:0)+i.id.toString().length)),o=t.map(i=>{let s=i.shouldBeUpdated?"*":"",a=i.name.length+s.length+i.id.toString().length,d=" ".repeat(r-a);return`${D.bold(i.name)} ${D.dim(`(${i.id})`)} ${s} ${d}${D.dim("-")} ${D.dim(i.slug)}`}),n=o.length>0?`
|
|
42
|
+
|
|
43
|
+
${o.join(`
|
|
44
|
+
`)}`:D.dim(`--none--
|
|
45
|
+
`);c.log(`
|
|
46
|
+
${e} ${n}`)}async function io(e){console.info(`API calls with ${He} threads`),console.info(`API URL ${ke}`);let{renderMode:t,domain:r,griddoVersion:o,basePath:n}=e,i=t===T.FROM_SCRATCH,s=De.default.join(n,"store"),a=[],d={},{sitesToPublish:m,sitesToUnpublish:f}=await Vr(r);no("Sites to publish:",m),no("Sites to unpublish:",f),await Kr(f),await cr(m,r);let l=[];for(let p of f){let{pagesStatus:u}=p;l.push(...u.active,...u.offlinePending,...u.uploadPending,...u.deleted,...u.offline)}for(let p of m){let{id:u,slug:g,theme:I,favicon:j,pagesStatus:S}=p,x=u.toString();l.push(...S.offlinePending,...S.deleted);let{siteInfo:B,siteHash:G,unpublishHashes:me,siteLangs:Z,defaultLang:O,headers:J,footers:K,socials:z}=await zr(u),{cloudinaryName:oe,useMetaTitle:Ge,useMetaKeywords:mt,showBasicMetaRobots:Ne,avoidHrefLangsOnCanonicals:No,avoidSelfReferenceCanonicals:Bo,avoidHrefLangXDefault:Mo,avoidDebugMetas:ko}=await $({endpoint:Xt});d[u]={siteHash:G,unpublishHashes:me,publishHashes:[],publishPagesIds:[]};let pt=new Oe;pt.navigations={headers:J,footers:K},p.languages=Z;let Ho=await vn(u,G),Uo=B.siteScript,Wo={siteUrl:B.slug,title:B.name,favicon:j},Jo={baseUrl:ke,publicBaseUrl:_t,instance:Ot,siteSlug:g,theme:I,siteMetadata:Wo,socials:z,siteLangs:Z,cloudinaryName:oe,griddoVersion:o,siteOptions:{useMetaTitle:Ge,useMetaKeywords:mt,showBasicMetaRobots:Ne,avoidHrefLangsOnCanonicals:No,avoidSelfReferenceCanonicals:Bo,avoidHrefLangXDefault:Mo,avoidDebugMetas:ko},siteScript:Uo};c.info(`Getting pages from ${p.name} site`),await ve.default.mkdir(De.default.join(s,x),{recursive:!0});let qo=async(Be,ut)=>{let be=[],v=await Br(ut,Ho);if(!v)return;a.push(ut);let Pe=JSON.parse(JSON.stringify(Jo));Pe.navigations=pt.getPageNavigations(v);let ne=await oo({page:v,cacheKey:Tn}),gt=await to(ne),Me=v?.mode==="list"||v?.mode==="paginated-data",ht=!Me&>,Yo=!ht&&!Me;if(Me){let M={page:v,pages:ro(ne),isRoot:!1,defaultLang:O,template:ne,totalQueriedItems:ne.queriedItems};be=await Qr(M,Pe),l.push(v.id)}if(ht){let M=v;M.template=ne,M.multiPageElements=gt,M.defaultLang=O,be=await eo(M,Pe),l.push(v.id)}if(Yo){let M=v;M.template=ne,M.defaultLang=O,be=[await Zr(M,Pe)]}v.hash!==null&&(d[u].publishHashes.push(v.hash),d[u].publishPagesIds.push(v.id)),await sr(Be,be)},ft=S.uploadPending,Vo=$n();(i||Vo)&&ft.push(...S.active);let Ko=we(He),zo=ft.map(Be=>Ko(()=>qo(x,Be)));await Promise.all(zo),S.zombie=await dr(x,[...S.active,...S.uploadPending]),await ar(x,[S.deleted,S.offline,S.offlinePending,S.zombie])}return await ir({buildProcessData:d,createdPages:a,sitesToPublish:m},r),{pagesToCreate:a,pagesToDelete:l}}async function On(){let{__root:e}=await y(),r=`${De.default.join(e,"apiCache")}/siteHash.json`;try{let o=await ve.default.readFile(r,"utf8");return JSON.parse(o)||{}}catch{return{}}}async function vn(e,t){let r=await On(),o=r[e],n=t||o||Date.now(),{__root:i}=await y(),a=`${De.default.join(i,"apiCache")}/siteHash.json`;return n!==o&&(r[e]=n,await ve.default.writeFile(a,JSON.stringify(r),{encoding:"utf-8"})),n.toString()}async function so(e){let{domain:t,renderMode:r,renderMetadata:{griddoVersion:o},pathsHydratedWithDomain:{__root:n}}=e,{pagesToCreate:i,pagesToDelete:s}=await io({domain:t,renderMode:r,griddoVersion:o,basePath:n});e.pagesToCreate=i,e.pagesToDelete=s}async function ao(e){let{domain:t}=e;await fr(t),await pr(t)}async function co(e){let{domain:t,pathsHydratedWithDomain:{__root:r}}=e;await mr({basePath:r,domain:t})}async function lo(e){let{domain:t}=e;await Ar(t)}var mo=h(require("node:fs/promises")),ct=h(require("node:path"));async function Fn(){return(await $({endpoint:Yt}))?.filter(t=>!!t.path).map(({path:t,content:r})=>({path:t,content:r||`User-agent: *
|
|
47
|
+
\r
|
|
48
|
+
\rAllow: /`}))||[]}async function po(e){let{__root:t}=await y({domain:e}),r=ct.default.join(t,"current-dist"),o=(await Fn()).find(({path:n})=>n===`/${e}`);if(!o){c.info(`Robots not found for ${e}`);return}if(await w(r)){let n=ct.default.join(r,"robots.txt");await mo.default.writeFile(n,o?.content),c.verbose(`wrote robots.txt to ${n}`)}else c.info(`${r} not found`)}var Re=h(require("node:fs/promises")),re=h(require("node:path"));function fo(e){return e.replace(/[<>&'"”]/g,t=>{switch(t){case"<":return"<";case">":return">";case"&":return"&";case"'":return"'";case'"':return""";case"\u201D":return""";default:return t}})}function uo(e){return`<?xml version="1.0" encoding="UTF-8"?>
|
|
49
|
+
<urlset xmlns="http://www.sitemaps.org/schemas/sitemap/0.9" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.sitemaps.org/schemas/sitemap/0.9 http://www.sitemaps.org/schemas/sitemap/0.9/sitemap.xsd">
|
|
50
|
+
${e.map(r=>` <url>
|
|
51
|
+
<loc>${fo(r.loc)}</loc>
|
|
52
|
+
<lastmod>${r.lastmod}</lastmod>
|
|
53
|
+
<priority>${r.priority}</priority>
|
|
54
|
+
</url>`).join(`
|
|
55
|
+
`)}
|
|
56
|
+
</urlset>`}function go(e){return`<?xml version="1.0" encoding="UTF-8"?>
|
|
57
|
+
<sitemapindex xmlns="http://www.sitemaps.org/schemas/sitemap/0.9">
|
|
58
|
+
${e.map(r=>` <sitemap>
|
|
59
|
+
<loc>${fo(r)}</loc>
|
|
60
|
+
</sitemap>`).join(`
|
|
61
|
+
`)}
|
|
62
|
+
</sitemapindex>`}async function yo(e){let{__root:t}=await y(),{sitesToPublish:r}=await _e(e),o=re.default.join(t,"current-dist"),n="sitemap-",i=s=>s==="sitemap.xml"||s.startsWith(n)&&s.endsWith(".xml");for(let s of r){let{id:a,languages:d}=s;for(let m of d){let f=V.headers?{...V.headers}:{};f.lang=m.id.toString();let l=await Ur({siteId:a,headers:f});if(!l)continue;let{items:p,url:u}=l,{home:g,domain:I}=u;if(!g)continue;let j=s.domains.find(O=>Object.keys(O)[0]===m.id.toString());if(!j)continue;let S=Object.values(j)[0],x=[],B=Object.keys(p),G=re.default.join(o,S.replace(I,""));if(await w(G))try{let O=await Re.default.readdir(G);for(let J of O)if(i(J)){let K=re.default.join(G,J);try{await Re.default.rm(K)}catch(z){c.error(`Error deleting file ${K}:`,z)}}}catch(O){c.error(`Error reading ${G}:`,O)}for(let O of B){let J=p[O];if(!J.length)continue;let K=uo(J),z=`/${n}${O.toLowerCase()}.xml`,oe=re.default.join(G,z);await ho(oe,K),c.verbose(`sitemap generated in ${oe}`),x.push(`${g.endsWith("/")?g.slice(0,-1):g}${z}`)}if(!x.length)continue;let me=go(x),Z=re.default.join(G,"sitemap.xml");await ho(Z,me),c.verbose(`sitemap generated in ${Z}`)}}}async function ho(e,t){try{let r=re.default.dirname(e);await w(r)||await Re.default.mkdir(r,{recursive:!0}),await Re.default.writeFile(e,t)}catch{throw new Error("Error saving a file")}}async function wo(e){let{domain:t}=e;await hr(t),await yo(t),await po(t)}async function Do(e){let{renderArtifacts:t}=e;await kt(t.initials)}var Ro=h(require("node:fs/promises")),dt=h(require("node:path"));async function So(e){let{pathsHydratedWithDomain:{__root:t,__ssg:r}}=e,o=dt.default.join(r,"public"),n=dt.default.join(t,"current-dist");await Ro.default.rename(o,n)}async function bo(e){let{renderMode:t,ssg:{ssgArtifacts:r},renderArtifacts:o,pathsHydratedWithDomain:{__components:n,__root:i,__cache:s,__ssg:a}}=e;await Ht(n,a,["static"]),t===T.INCREMENTAL&&(await ee(s,a,r.cacheables),await ee(s,i,o.cacheables))}var Po=require("node:child_process"),Eo=h(require("node:path"));async function _o(e){c.verbose(`read assetPrefixWithDomain, ${e}`),c.verbose(`using this NODE_OPTIONS in gatsby command: ${process.env.NODE_OPTIONS}`);let{__ssg:t}=await y(),r=process.env.NODE_OPTIONS?process.env.NODE_OPTIONS.replace(/--openssl-legacy-provider\s*/g,"").trim():"",o=Eo.default.join(t,"node_modules",".bin","gatsby"),n=["build","--prefix-paths"];return $t&&n.push("--verbose"),new Promise((i,s)=>{let a=(0,Po.spawn)(o,n,{cwd:t,stdio:["ignore","inherit","inherit"],env:{...process.env,SPAWN_ASSET_PREFIX_WITH_DOMAIN:e,NODE_OPTIONS:r}});a.on("close",d=>{d===0?i():s(new Error(`Gatsby build failed with exit code ${d}`))}),a.on("error",d=>{s(new Error(`Failed to start Gatsby build process: ${d.message}`))})})}async function Io(e){let{ssg:{assetPrefix:t}}=e;await _o(t)}var Ce=h(require("node:fs/promises")),Le=h(require("node:path"));var X=h(require("node:fs/promises")),E=h(require("node:path"));async function Ao(e){let{__root:t,__exports:r}=await y({domain:e}),o=(await X.default.readdir(E.default.join(r,"dist"))).filter(i=>E.default.extname(i)===".js"||E.default.extname(i)===".json"||E.default.extname(i)===".css");await X.default.mkdir(E.default.join(t,"assets"),{recursive:!0}),await X.default.cp(E.default.join(r,"dist","page-data"),E.default.join(t,"assets","page-data"),{preserveTimestamps:!0,recursive:!0}),await w(E.default.join(t,"static"))&&await X.default.cp(E.default.join(t,"static"),E.default.join(t,"assets"),{force:!1,preserveTimestamps:!0,recursive:!0}),await w(E.default.join(r,"dist","static"))&&await X.default.cp(E.default.join(r,"dist","static"),E.default.join(t,"assets","static"),{force:!1,preserveTimestamps:!0,recursive:!0}),await w(E.default.join(t,"static"))&&await X.default.cp(E.default.join(t,"static"),E.default.join(r,"dist",e),{preserveTimestamps:!0,recursive:!0,force:!1});let n=o.map(i=>{let s=E.default.join(r,"dist",i),a=E.default.join(t,"assets",i);return X.default.cp(s,a,{preserveTimestamps:!0,recursive:!0})});await Promise.all(n)}var N=h(require("node:fs/promises")),_=h(require("node:path"));var To=require("node:crypto"),Se=h(require("node:fs/promises")),$o=require("node:path");var Cn=[/\.js$/];async function xo(e){let t=new Map;try{let o=(await Se.default.readdir(e)).filter(n=>Cn.some(i=>i.test(n)));for(let n of o){let i=(0,$o.join)(e,n),s=await Se.default.readFile(i),a=(0,To.createHash)("sha256").update(s).digest("hex");t.set(n,a)}return t}catch(r){return c.build(`ERROR: Error processing directory ${e}: ${r}`),t}}async function Oo(e,t){let[r,o]=await Promise.all([xo(e),xo(t)]);if(r.size!==o.size)return!0;for(let[n,i]of r.entries()){let s=o.get(n);if(!s||i!==s)return!0}return!1}async function vo(e,t){try{let r=await Se.default.readFile(e,"utf-8"),o=/(window\.___webpackCompilationHash=")([^"]*)(";<\/script>)/,n=r.replace(o,`$1${t}$3`);if(r===n)return;await Se.default.writeFile(e,n,"utf-8")}catch(r){c.error(`Failed to process file ${e}: ${r}`)}}var Fe=class{bundleDir;currentRenderDir;pagesToCreate;pagesToDelete;assetArtifacts;state={htmlToAdd:[],jsonToAdd:[],htmlToDelete:[],jsonToDelete:[]};constructor(t){this.bundleDir=t.dst,this.currentRenderDir=t.src,this.pagesToCreate=t.pagesToCreate,this.pagesToDelete=t.pagesToDelete,this.assetArtifacts=t.artifactsToCopyToExports}async execute(){await this.setPagesToDelete(),await this.setPagesToAdd(),await this.sync(),Ft&&await this.assertAssetsAreValid()}async assertAssetsAreValid(){await Oo(this.currentRenderDir,this.bundleDir)?c.info("Warning: JavaScript and CSS bundles must have the same files and content. You can ignore this warning if the rendering process only has pages to unpublish."):c.info("Previous and current JavaScript and CSS packages successfully matched")}async setPagesToDelete(){let t=new Set([...this.pagesToDelete,...this.pagesToCreate]),r=await this.scanPages(this.bundleDir);for(let o of r)t.has(o.id)&&(this.state.htmlToDelete.push(o.htmlPath),this.state.jsonToDelete.push(o.jsonPath));c.verbose(`${this.state.htmlToDelete.length} pages HTML to delete`),c.verbose(`${this.state.jsonToDelete.length} pages JSON to delete`)}async setPagesToAdd(){let t=new Set(this.pagesToCreate),r=await this.scanPages(this.currentRenderDir);for(let o of r)if(t.has(o.id)){let n=_.default.join(this.bundleDir,o.composePath,"index.html"),i=o.composePath==="/"?"index":o.composePath,s=_.default.join(this.bundleDir,"page-data",i,"page-data.json");this.state.htmlToAdd.push({from:o.htmlPath,to:n}),this.state.jsonToAdd.push({from:o.jsonPath,to:s})}c.verbose(`${this.state.htmlToAdd.length} pages HTML to create`),c.verbose(`${this.state.jsonToAdd.length} pages JSON to create`)}async sync(){let t=[...this.state.htmlToDelete,...this.state.jsonToDelete];for(let o of t)try{await N.default.rm(o,{force:!0}),c.verbose(`Sync (remove) : ${o}`)}catch(n){c.error(`Failed to remove ${o}:`,n)}await this.restoreWebpackCompilationHash();let r=[...this.state.htmlToAdd,...this.state.jsonToAdd];for(let o of r)try{await N.default.mkdir(_.default.dirname(o.to),{recursive:!0}),await N.default.copyFile(o.from,o.to),c.verbose(`Sync (copy) : ${o.from} -> ${o.to}`)}catch(n){c.error(`Failed to copy ${o.from} to ${o.to}:`,n)}await this.copyAssetArtifacts(),await this.copySitemaps()}async restoreWebpackCompilationHash(){let t=_.default.join(this.bundleDir,"page-data","app-data.json"),r=JSON.parse(await N.default.readFile(t,"utf8")).webpackCompilationHash;c.build(`Reading webpackCompilationHash from ${t} = ${r}`);let o=this.state.htmlToAdd.map(({from:n})=>n);for(let n of o)await vo(n,r)}async copyAssetArtifacts(){for(let t of this.assetArtifacts){let r=_.default.join(this.currentRenderDir,t),o=_.default.join(this.bundleDir,t);try{await N.default.mkdir(_.default.dirname(o),{recursive:!0}),await N.default.copyFile(r,o),c.verbose(`Copied artifact: ${t}`)}catch(n){c.error(`Failed to copy artifact ${r} to ${o}:`,n)}}}async copySitemaps(){let t=_.default.resolve(this.currentRenderDir),r=_.default.resolve(this.bundleDir),o=pe(r,".xml");for await(let i of o){let s=_.default.basename(i);(s.startsWith("sitemap-")||s==="sitemap.xml")&&await N.default.rm(i,{force:!0})}let n=pe(t,".xml");for await(let i of n){let s=_.default.relative(t,i),a=_.default.join(r,s);try{await N.default.mkdir(_.default.dirname(a),{recursive:!0}),await N.default.copyFile(i,a)}catch(d){c.error(`Failed to copy sitemap ${i} to ${a}:`,d)}}c.verbose("Copied sitemap files.")}async scanPages(t){let r=we(128),o=pe(_.default.join(t,"page-data"),"page-data.json"),n=[];for await(let a of o)n.push(a);let i=n.map(a=>r(async()=>{try{let d=await N.default.readFile(a,"utf-8"),m=JSON.parse(d),f=m.result.pageContext.id,l=m.result.pageContext.fullPath.compose,p=l==="/"?"index":l;return{id:f,composePath:l,htmlPath:_.default.join(t,l,"index.html"),jsonPath:_.default.join(t,"page-data",p,"page-data.json")}}catch(d){return c.error(`Error reading or parsing page data from ${a}:`,d),null}}));return(await Promise.all(i)).filter(a=>a!==null)}};async function Fo(e){let{renderMode:t,renderArtifacts:r,domain:o,pagesToCreate:n,pagesToDelete:i,pathsHydratedWithDomain:{__root:s,__ssg:a,__cache:d,__exports:m},ssg:{ssgArtifacts:f,assetPrefix:l}}=e,p=!!l&&l!=="",u=Le.default.join(s,"current-dist"),g=Le.default.join(m,"dist");await ee(s,d,r.cacheables,{override:!0}),await ee(a,d,f.cacheables,{override:!0}),t===T.FROM_SCRATCH&&(await Je(u),await Ce.default.rm(g,{force:!0,recursive:!0}),await Ce.default.rename(u,g)),t===T.INCREMENTAL&&(await Je(g),await new Fe({src:u,dst:g,pagesToCreate:n,pagesToDelete:i,artifactsToCopyToExports:["build-report.json"]}).execute(),await Ce.default.rm(u,{force:!0,recursive:!0}),await qe(Le.default.join(m,"dist"))),p&&await Ao(o),await ee(s,m,r.archivables,{override:!0})}var je=h(require("node:path"));async function Co(e){let{__ssg:t}=await y({domain:e});return{disposables:[je.default.join(t,"public"),je.default.join(t,"static"),je.default.join(t,".cache")],cacheables:[".cache"]}}async function Lo(){let e=await b();e.needsRollbackOnError=!0,await L(e),c.info(`Render lifecycles that follow will restore the exports dir on error if needed
|
|
63
|
+
`)}async function jo(){let e=await b();e.needsRollbackOnError=!1,await L(e),c.info(`Render lifecycles that follow will not restore the exports dir on error
|
|
64
|
+
`)}var Ln=ie;async function lt(e){await _r();let{renderMode:t,reason:r}=await gr(e),o=await Ke(),n=await y({domain:e}),{__ssg:i}=n,s=await Cr(e),a=await Co(e),d=Ue?`${Ue}/${e}`:"",m=await w(Go.default.join(i,`.render-sentinel-${e}`)),f=m?T.FROM_SCRATCH:t,l=m?"previous render error":r,p=await b();if(p.domains[e].renderMode=f,p.domains[e].renderModeReason=l,p.domains[e].isRendering=!1,t===T.IDLE&&f===T.IDLE){c.info(`(From Current Render) [${e}]: Skipping start-render as it is marked as IDLE with the reason ${r}.`);return}p.domains[e].isRendering=!0,p.currentRenderingDomain=e,await L(p);let u=l?` <${l}>`:"";c.info(`Init render (${f})${u} for domain ${e}
|
|
65
|
+
`);let g=new xe({domain:e,pathsHydratedWithDomain:n,renderArtifacts:s,renderMetadata:o,renderMode:f,ssg:{assetPrefix:d,ssgArtifacts:a}});await F("Init",async()=>co(g)),await F("Clean",async()=>Lr(g)),await F("Prepare",async()=>Do(g)),await F("Restore",async()=>bo(g)),await Lo(),await F("Data",async()=>so(g)),await F("SSG",async()=>Io(g)),await F("Relocation",async()=>So(g)),await F("Meta",async()=>wo(g)),await F("Sync",async()=>Fo(g)),await F("HealthCheck",async()=>ao(g)),await F("Logs",async()=>lo(g),{skip:Ln?!1:"Build logs to file are disabled"}),await F("Close",async()=>jr(g)),await jo()}async function jn(){c.warn("Legacy Render Mode");let e=await Fr();for(let t of e)await lt(t)}async function Gn(){let[e]=process.argv.slice(2);await lt(e)}async function Nn(){await br(),await V.login(),xt?await Gn():await jn()}Dr(Nn);
|
|
66
|
+
//# sourceMappingURL=start-render.js.map
|