@payloadcms/plugin-import-export 3.77.0-internal.8cf758f → 3.77.0-internal.9837b1e
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/export/createExport.d.ts.map +1 -1
- package/dist/export/createExport.js +2 -9
- package/dist/export/createExport.js.map +1 -1
- package/dist/export/handlePreview.d.ts.map +1 -1
- package/dist/export/handlePreview.js +34 -17
- package/dist/export/handlePreview.js.map +1 -1
- package/dist/utilities/filterToSelectedFields.spec.js +204 -0
- package/dist/utilities/filterToSelectedFields.spec.js.map +1 -0
- package/dist/utilities/flattenObject.d.ts +1 -7
- package/dist/utilities/flattenObject.d.ts.map +1 -1
- package/dist/utilities/flattenObject.js +14 -25
- package/dist/utilities/flattenObject.js.map +1 -1
- package/dist/utilities/getSchemaColumns.d.ts +11 -7
- package/dist/utilities/getSchemaColumns.d.ts.map +1 -1
- package/dist/utilities/getSchemaColumns.js +26 -21
- package/dist/utilities/getSchemaColumns.js.map +1 -1
- package/package.json +7 -7
- package/dist/utilities/collectTimezoneCompanionFields.d.ts +0 -24
- package/dist/utilities/collectTimezoneCompanionFields.d.ts.map +0 -1
- package/dist/utilities/collectTimezoneCompanionFields.js +0 -89
- package/dist/utilities/collectTimezoneCompanionFields.js.map +0 -1
- package/dist/utilities/collectTimezoneCompanionFields.spec.js +0 -319
- package/dist/utilities/collectTimezoneCompanionFields.spec.js.map +0 -1
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"createExport.d.ts","sourceRoot":"","sources":["../../src/export/createExport.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,cAAc,EAAE,IAAI,EAAa,KAAK,EAAE,MAAM,SAAS,CAAA;
|
|
1
|
+
{"version":3,"file":"createExport.d.ts","sourceRoot":"","sources":["../../src/export/createExport.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,cAAc,EAAE,IAAI,EAAa,KAAK,EAAE,MAAM,SAAS,CAAA;AAerE,MAAM,MAAM,MAAM,GAAG;IACnB;;;OAGG;IACH,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,cAAc,EAAE,MAAM,CAAA;IACtB;;OAEG;IACH,KAAK,CAAC,EAAE,OAAO,CAAA;IACf,MAAM,CAAC,EAAE,IAAI,GAAG,KAAK,CAAA;IACrB,gBAAgB,EAAE,MAAM,CAAA;IACxB,MAAM,CAAC,EAAE,MAAM,EAAE,CAAA;IACjB,MAAM,EAAE,KAAK,GAAG,MAAM,CAAA;IACtB,OAAO,CAAC,EAAE,MAAM,EAAE,CAAA;IAClB,EAAE,EAAE,MAAM,GAAG,MAAM,CAAA;IACnB,KAAK,CAAC,EAAE,MAAM,CAAA;IACd,MAAM,CAAC,EAAE,MAAM,CAAA;IACf;;;OAGG;IACH,QAAQ,CAAC,EAAE,MAAM,CAAA;IACjB,IAAI,EAAE,MAAM,CAAA;IACZ,IAAI,CAAC,EAAE,MAAM,CAAA;IACb,IAAI,EAAE,MAAM,CAAA;IACZ,IAAI,EAAE,IAAI,CAAA;IACV,cAAc,EAAE,MAAM,CAAA;IACtB,MAAM,EAAE,MAAM,GAAG,MAAM,CAAA;IACvB,KAAK,CAAC,EAAE,KAAK,CAAA;CACd,CAAA;AAED,MAAM,MAAM,gBAAgB,GAAG;IAC7B;;OAEG;IACH,QAAQ,CAAC,EAAE,OAAO,CAAA;IAClB,GAAG,EAAE,cAAc,CAAA;CACpB,GAAG,MAAM,CAAA;AAEV,eAAO,MAAM,YAAY,SAAgB,gBAAgB,kCA8dxD,CAAA"}
|
|
@@ -2,7 +2,6 @@
|
|
|
2
2
|
import { APIError } from 'payload';
|
|
3
3
|
import { Readable } from 'stream';
|
|
4
4
|
import { buildDisabledFieldRegex } from '../utilities/buildDisabledFieldRegex.js';
|
|
5
|
-
import { collectTimezoneCompanionFields } from '../utilities/collectTimezoneCompanionFields.js';
|
|
6
5
|
import { flattenObject } from '../utilities/flattenObject.js';
|
|
7
6
|
import { getExportFieldFunctions } from '../utilities/getExportFieldFunctions.js';
|
|
8
7
|
import { getFilename } from '../utilities/getFilename.js';
|
|
@@ -129,8 +128,6 @@ export const createExport = async (args)=>{
|
|
|
129
128
|
const toCSVFunctions = getExportFieldFunctions({
|
|
130
129
|
fields: collectionConfig.flattenedFields
|
|
131
130
|
});
|
|
132
|
-
// Collect auto-generated timezone companion fields from schema
|
|
133
|
-
const timezoneCompanionFields = collectTimezoneCompanionFields(collectionConfig.flattenedFields);
|
|
134
131
|
const disabledFields = collectionConfig.admin?.custom?.['plugin-import-export']?.disabledFields ?? [];
|
|
135
132
|
const disabledMatchers = disabledFields.map(buildDisabledFieldRegex);
|
|
136
133
|
const filterDisabledCSV = (row)=>{
|
|
@@ -175,8 +172,7 @@ export const createExport = async (args)=>{
|
|
|
175
172
|
disabledFields,
|
|
176
173
|
fields,
|
|
177
174
|
locale,
|
|
178
|
-
localeCodes
|
|
179
|
-
timezoneCompanionFields
|
|
175
|
+
localeCodes
|
|
180
176
|
});
|
|
181
177
|
if (debug) {
|
|
182
178
|
req.payload.logger.debug({
|
|
@@ -226,7 +222,6 @@ export const createExport = async (args)=>{
|
|
|
226
222
|
const batchRows = result.docs.map((doc)=>filterDisabledCSV(flattenObject({
|
|
227
223
|
doc,
|
|
228
224
|
fields,
|
|
229
|
-
timezoneCompanionFields,
|
|
230
225
|
toCSVFunctions
|
|
231
226
|
})));
|
|
232
227
|
// On first batch, discover additional columns from data and merge with schema
|
|
@@ -310,7 +305,6 @@ export const createExport = async (args)=>{
|
|
|
310
305
|
const transformDoc = (doc)=>isCSV ? filterDisabledCSV(flattenObject({
|
|
311
306
|
doc,
|
|
312
307
|
fields,
|
|
313
|
-
timezoneCompanionFields,
|
|
314
308
|
toCSVFunctions
|
|
315
309
|
})) : filterDisabledJSON(doc);
|
|
316
310
|
// Skip fetching if access was denied - we'll create an empty export
|
|
@@ -341,8 +335,7 @@ export const createExport = async (args)=>{
|
|
|
341
335
|
disabledFields,
|
|
342
336
|
fields,
|
|
343
337
|
locale,
|
|
344
|
-
localeCodes
|
|
345
|
-
timezoneCompanionFields
|
|
338
|
+
localeCodes
|
|
346
339
|
});
|
|
347
340
|
// Merge schema columns with data-discovered columns
|
|
348
341
|
// Schema provides ordering, data provides additional columns (e.g., array indices > 0)
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/export/createExport.ts"],"sourcesContent":["/* eslint-disable perfectionist/sort-objects */\nimport type { PayloadRequest, Sort, TypedUser, Where } from 'payload'\n\nimport { stringify } from 'csv-stringify/sync'\nimport { APIError } from 'payload'\nimport { Readable } from 'stream'\n\nimport { buildDisabledFieldRegex } from '../utilities/buildDisabledFieldRegex.js'\nimport { collectTimezoneCompanionFields } from '../utilities/collectTimezoneCompanionFields.js'\nimport { flattenObject } from '../utilities/flattenObject.js'\nimport { getExportFieldFunctions } from '../utilities/getExportFieldFunctions.js'\nimport { getFilename } from '../utilities/getFilename.js'\nimport { getSchemaColumns, mergeColumns } from '../utilities/getSchemaColumns.js'\nimport { getSelect } from '../utilities/getSelect.js'\nimport { validateLimitValue } from '../utilities/validateLimitValue.js'\nimport { createExportBatchProcessor, type ExportFindArgs } from './batchProcessor.js'\n\nexport type Export = {\n /**\n * Number of documents to process in each batch during export\n * @default 100\n */\n batchSize?: number\n collectionSlug: string\n /**\n * If true, enables debug logging\n */\n debug?: boolean\n drafts?: 'no' | 'yes'\n exportCollection: string\n fields?: string[]\n format: 'csv' | 'json'\n globals?: string[]\n id: number | string\n limit?: number\n locale?: string\n /**\n * Maximum number of documents that can be exported in a single operation.\n * This value has already been resolved from the plugin config.\n */\n maxLimit?: number\n name: string\n page?: number\n slug: string\n sort: Sort\n userCollection: string\n userID: number | string\n where?: Where\n}\n\nexport type CreateExportArgs = {\n /**\n * If true, stream the file instead of saving it\n */\n download?: boolean\n req: PayloadRequest\n} & Export\n\nexport const createExport = async (args: CreateExportArgs) => {\n const {\n id,\n name: nameArg,\n batchSize = 100,\n collectionSlug,\n debug = false,\n download,\n drafts: draftsFromInput,\n exportCollection,\n fields,\n format,\n limit: incomingLimit,\n maxLimit,\n locale: localeFromInput,\n page,\n req,\n sort,\n userCollection,\n userID,\n where: whereFromInput = {},\n } = args\n const { locale: localeFromReq, payload } = req\n\n if (debug) {\n req.payload.logger.debug({\n message: 'Starting export process with args:',\n collectionSlug,\n draft: draftsFromInput,\n fields,\n format,\n })\n }\n\n const locale = localeFromInput ?? localeFromReq\n const collectionConfig = payload.config.collections.find(({ slug }) => slug === collectionSlug)\n\n if (!collectionConfig) {\n throw new APIError(`Collection with slug ${collectionSlug} not found.`)\n }\n\n let user: TypedUser | undefined\n\n if (userCollection && userID) {\n user = (await req.payload.findByID({\n id: userID,\n collection: userCollection,\n overrideAccess: true,\n })) as TypedUser\n }\n\n if (!user && req.user) {\n user = req?.user?.id ? req.user : req?.user?.user\n }\n\n if (!user) {\n throw new APIError('User authentication is required to create exports.')\n }\n\n const draft = draftsFromInput === 'yes'\n const hasVersions = Boolean(collectionConfig.versions)\n\n // Only filter by _status for versioned collections\n const publishedWhere: Where = hasVersions ? { _status: { equals: 'published' } } : {}\n\n const where: Where = {\n and: [whereFromInput, draft ? {} : publishedWhere],\n }\n\n const name = `${nameArg ?? `${getFilename()}-${collectionSlug}`}.${format}`\n const isCSV = format === 'csv'\n const select = Array.isArray(fields) && fields.length > 0 ? getSelect(fields) : undefined\n\n if (debug) {\n req.payload.logger.debug({ message: 'Export configuration:', name, isCSV, locale })\n }\n\n // Determine maximum export documents:\n // 1. If maxLimit is defined, it sets the absolute ceiling\n // 2. User's limit is applied but clamped to maxLimit if it exceeds it\n let maxExportDocuments: number | undefined\n\n if (typeof maxLimit === 'number' && maxLimit > 0) {\n if (typeof incomingLimit === 'number' && incomingLimit > 0) {\n // User provided a limit - clamp it to maxLimit\n maxExportDocuments = Math.min(incomingLimit, maxLimit)\n } else {\n // No user limit - use maxLimit as the ceiling\n maxExportDocuments = maxLimit\n }\n } else {\n // No maxLimit - use user's limit if provided\n maxExportDocuments =\n typeof incomingLimit === 'number' && incomingLimit > 0 ? incomingLimit : undefined\n }\n\n // Try to count documents - if access is denied, treat as 0 documents\n let totalDocs = 0\n let accessDenied = false\n try {\n const countResult = await payload.count({\n collection: collectionSlug,\n user,\n locale,\n overrideAccess: false,\n })\n totalDocs = countResult.totalDocs\n } catch (error) {\n // Access denied - user can't read from this collection\n // We'll create an empty export file\n accessDenied = true\n if (debug) {\n req.payload.logger.debug({\n message: 'Access denied for collection, creating empty export',\n collectionSlug,\n })\n }\n }\n\n const totalPages = Math.max(1, Math.ceil(totalDocs / batchSize))\n const requestedPage = page || 1\n const adjustedPage = requestedPage > totalPages ? 1 : requestedPage\n\n const findArgs = {\n collection: collectionSlug,\n depth: 1,\n draft,\n limit: batchSize,\n locale,\n overrideAccess: false,\n page: 0,\n select,\n sort,\n user,\n where,\n }\n\n if (debug) {\n req.payload.logger.debug({ message: 'Find arguments:', findArgs })\n }\n\n const toCSVFunctions = getExportFieldFunctions({\n fields: collectionConfig.flattenedFields,\n })\n\n // Collect auto-generated timezone companion fields from schema\n const timezoneCompanionFields = collectTimezoneCompanionFields(collectionConfig.flattenedFields)\n\n const disabledFields =\n collectionConfig.admin?.custom?.['plugin-import-export']?.disabledFields ?? []\n\n const disabledMatchers = disabledFields.map(buildDisabledFieldRegex)\n\n const filterDisabledCSV = (row: Record<string, unknown>): Record<string, unknown> => {\n const filtered: Record<string, unknown> = {}\n\n for (const [key, value] of Object.entries(row)) {\n const isDisabled = disabledMatchers.some((matcher) => matcher.test(key))\n if (!isDisabled) {\n filtered[key] = value\n }\n }\n\n return filtered\n }\n\n const filterDisabledJSON = (doc: any, parentPath = ''): any => {\n if (Array.isArray(doc)) {\n return doc.map((item) => filterDisabledJSON(item, parentPath))\n }\n\n if (typeof doc !== 'object' || doc === null) {\n return doc\n }\n\n const filtered: Record<string, any> = {}\n for (const [key, value] of Object.entries(doc)) {\n const currentPath = parentPath ? `${parentPath}.${key}` : key\n\n // Only remove if this exact path is disabled\n const isDisabled = disabledFields.includes(currentPath)\n\n if (!isDisabled) {\n filtered[key] = filterDisabledJSON(value, currentPath)\n }\n }\n\n return filtered\n }\n\n if (download) {\n const limitErrorMsg = validateLimitValue(incomingLimit, req.t)\n if (limitErrorMsg) {\n throw new APIError(limitErrorMsg)\n }\n\n // Get schema-based columns first (provides base ordering and handles empty exports)\n let schemaColumns: string[] = []\n if (isCSV) {\n const localeCodes =\n locale === 'all' && payload.config.localization\n ? payload.config.localization.localeCodes\n : undefined\n\n schemaColumns = getSchemaColumns({\n collectionConfig,\n disabledFields,\n fields,\n locale,\n localeCodes,\n timezoneCompanionFields,\n })\n\n if (debug) {\n req.payload.logger.debug({\n columnCount: schemaColumns.length,\n msg: 'Schema-based column inference complete',\n })\n }\n }\n\n // allColumns will be finalized after first batch (schema + data columns merged)\n let allColumns: string[] = []\n let columnsFinalized = false\n\n const encoder = new TextEncoder()\n let isFirstBatch = true\n let currentBatchPage = adjustedPage\n let fetched = 0\n const maxDocs =\n typeof maxExportDocuments === 'number' ? maxExportDocuments : Number.POSITIVE_INFINITY\n\n const stream = new Readable({\n async read() {\n const remaining = Math.max(0, maxDocs - fetched)\n\n if (remaining === 0) {\n if (!isCSV) {\n // If first batch with no remaining, output empty array; otherwise just close\n this.push(encoder.encode(isFirstBatch ? '[]' : ']'))\n }\n this.push(null)\n return\n }\n\n const result = await payload.find({\n ...findArgs,\n page: currentBatchPage,\n limit: Math.min(batchSize, remaining),\n })\n\n if (debug) {\n req.payload.logger.debug(\n `Streaming batch ${currentBatchPage} with ${result.docs.length} docs`,\n )\n }\n\n if (result.docs.length === 0) {\n // Close JSON array properly if JSON\n if (!isCSV) {\n // If first batch with no docs, output empty array; otherwise just close\n this.push(encoder.encode(isFirstBatch ? '[]' : ']'))\n }\n this.push(null)\n return\n }\n\n if (isCSV) {\n // --- CSV Streaming ---\n const batchRows = result.docs.map((doc) =>\n filterDisabledCSV(\n flattenObject({ doc, fields, timezoneCompanionFields, toCSVFunctions }),\n ),\n )\n\n // On first batch, discover additional columns from data and merge with schema\n if (!columnsFinalized) {\n const dataColumns: string[] = []\n const seenCols = new Set<string>()\n for (const row of batchRows) {\n for (const key of Object.keys(row)) {\n if (!seenCols.has(key)) {\n seenCols.add(key)\n dataColumns.push(key)\n }\n }\n }\n // Merge schema columns with data-discovered columns\n allColumns = mergeColumns(schemaColumns, dataColumns)\n columnsFinalized = true\n\n if (debug) {\n req.payload.logger.debug({\n dataColumnsCount: dataColumns.length,\n finalColumnsCount: allColumns.length,\n msg: 'Merged schema and data columns',\n })\n }\n }\n\n const paddedRows = batchRows.map((row) => {\n const fullRow: Record<string, unknown> = {}\n for (const col of allColumns) {\n fullRow[col] = row[col] ?? ''\n }\n return fullRow\n })\n\n const csvString = stringify(paddedRows, {\n bom: isFirstBatch,\n header: isFirstBatch,\n columns: allColumns,\n })\n\n this.push(encoder.encode(csvString))\n } else {\n // --- JSON Streaming ---\n const batchRows = result.docs.map((doc) => filterDisabledJSON(doc))\n\n // Convert each filtered/flattened row into JSON string\n const batchJSON = batchRows.map((row) => JSON.stringify(row)).join(',')\n\n if (isFirstBatch) {\n this.push(encoder.encode('[' + batchJSON))\n } else {\n this.push(encoder.encode(',' + batchJSON))\n }\n }\n\n fetched += result.docs.length\n isFirstBatch = false\n currentBatchPage += 1\n\n if (!result.hasNextPage || fetched >= maxDocs) {\n if (debug) {\n req.payload.logger.debug('Stream complete - no more pages')\n }\n if (!isCSV) {\n this.push(encoder.encode(']'))\n }\n this.push(null) // End the stream\n }\n },\n })\n\n return new Response(Readable.toWeb(stream) as ReadableStream, {\n headers: {\n 'Content-Disposition': `attachment; filename=\"${name}\"`,\n 'Content-Type': isCSV ? 'text/csv' : 'application/json',\n },\n })\n }\n\n // Non-download path (buffered export)\n if (debug) {\n req.payload.logger.debug('Starting file generation')\n }\n\n // Create export batch processor\n const processor = createExportBatchProcessor({ batchSize, debug })\n\n // Transform function based on format\n const transformDoc = (doc: unknown) =>\n isCSV\n ? filterDisabledCSV(flattenObject({ doc, fields, timezoneCompanionFields, toCSVFunctions }))\n : filterDisabledJSON(doc)\n\n // Skip fetching if access was denied - we'll create an empty export\n let exportResult = {\n columns: [] as string[],\n docs: [] as Record<string, unknown>[],\n fetchedCount: 0,\n }\n\n if (!accessDenied) {\n exportResult = await processor.processExport({\n collectionSlug,\n findArgs: findArgs as ExportFindArgs,\n format,\n maxDocs:\n typeof maxExportDocuments === 'number' ? maxExportDocuments : Number.POSITIVE_INFINITY,\n req,\n startPage: adjustedPage,\n transformDoc,\n })\n }\n\n const { columns: dataColumns, docs: rows } = exportResult\n const outputData: string[] = []\n\n // Prepare final output\n if (isCSV) {\n // Get schema-based columns for consistent ordering\n const localeCodes =\n locale === 'all' && payload.config.localization\n ? payload.config.localization.localeCodes\n : undefined\n\n const schemaColumns = getSchemaColumns({\n collectionConfig,\n disabledFields,\n fields,\n locale,\n localeCodes,\n timezoneCompanionFields,\n })\n\n // Merge schema columns with data-discovered columns\n // Schema provides ordering, data provides additional columns (e.g., array indices > 0)\n const finalColumns = mergeColumns(schemaColumns, dataColumns)\n\n const paddedRows = rows.map((row) => {\n const fullRow: Record<string, unknown> = {}\n for (const col of finalColumns) {\n fullRow[col] = row[col] ?? ''\n }\n return fullRow\n })\n\n // Always output CSV with header, even if empty\n outputData.push(\n stringify(paddedRows, {\n bom: true,\n header: true,\n columns: finalColumns,\n }),\n )\n } else {\n // JSON format\n outputData.push(rows.map((doc) => JSON.stringify(doc)).join(',\\n'))\n }\n\n // Ensure we always have valid content for the file\n // For JSON, empty exports produce \"[]\"\n // For CSV, if completely empty (no columns, no rows), produce at least a newline to ensure file creation\n const content = format === 'json' ? `[${outputData.join(',')}]` : outputData.join('')\n const buffer = Buffer.from(content.length > 0 ? content : '\\n')\n if (debug) {\n req.payload.logger.debug(`${format} file generation complete`)\n }\n\n if (!id) {\n if (debug) {\n req.payload.logger.debug('Creating new export file')\n }\n req.file = {\n name,\n data: buffer,\n mimetype: isCSV ? 'text/csv' : 'application/json',\n size: buffer.length,\n }\n } else {\n if (debug) {\n req.payload.logger.debug(`Updating existing export with id: ${id}`)\n }\n await req.payload.update({\n id,\n collection: exportCollection,\n data: {},\n file: {\n name,\n data: buffer,\n mimetype: isCSV ? 'text/csv' : 'application/json',\n size: buffer.length,\n },\n // Override access only here so that we can be sure the export collection itself is updated as expected\n overrideAccess: true,\n req,\n })\n }\n if (debug) {\n req.payload.logger.debug('Export process completed successfully')\n }\n}\n"],"names":["stringify","APIError","Readable","buildDisabledFieldRegex","collectTimezoneCompanionFields","flattenObject","getExportFieldFunctions","getFilename","getSchemaColumns","mergeColumns","getSelect","validateLimitValue","createExportBatchProcessor","createExport","args","id","name","nameArg","batchSize","collectionSlug","debug","download","drafts","draftsFromInput","exportCollection","fields","format","limit","incomingLimit","maxLimit","locale","localeFromInput","page","req","sort","userCollection","userID","where","whereFromInput","localeFromReq","payload","logger","message","draft","collectionConfig","config","collections","find","slug","user","findByID","collection","overrideAccess","hasVersions","Boolean","versions","publishedWhere","_status","equals","and","isCSV","select","Array","isArray","length","undefined","maxExportDocuments","Math","min","totalDocs","accessDenied","countResult","count","error","totalPages","max","ceil","requestedPage","adjustedPage","findArgs","depth","toCSVFunctions","flattenedFields","timezoneCompanionFields","disabledFields","admin","custom","disabledMatchers","map","filterDisabledCSV","row","filtered","key","value","Object","entries","isDisabled","some","matcher","test","filterDisabledJSON","doc","parentPath","item","currentPath","includes","limitErrorMsg","t","schemaColumns","localeCodes","localization","columnCount","msg","allColumns","columnsFinalized","encoder","TextEncoder","isFirstBatch","currentBatchPage","fetched","maxDocs","Number","POSITIVE_INFINITY","stream","read","remaining","push","encode","result","docs","batchRows","dataColumns","seenCols","Set","keys","has","add","dataColumnsCount","finalColumnsCount","paddedRows","fullRow","col","csvString","bom","header","columns","batchJSON","JSON","join","hasNextPage","Response","toWeb","headers","processor","transformDoc","exportResult","fetchedCount","processExport","startPage","rows","outputData","finalColumns","content","buffer","Buffer","from","file","data","mimetype","size","update"],"mappings":"AAAA,6CAA6C,GAG7C,SAASA,SAAS,QAAQ,qBAAoB;AAC9C,SAASC,QAAQ,QAAQ,UAAS;AAClC,SAASC,QAAQ,QAAQ,SAAQ;AAEjC,SAASC,uBAAuB,QAAQ,0CAAyC;AACjF,SAASC,8BAA8B,QAAQ,iDAAgD;AAC/F,SAASC,aAAa,QAAQ,gCAA+B;AAC7D,SAASC,uBAAuB,QAAQ,0CAAyC;AACjF,SAASC,WAAW,QAAQ,8BAA6B;AACzD,SAASC,gBAAgB,EAAEC,YAAY,QAAQ,mCAAkC;AACjF,SAASC,SAAS,QAAQ,4BAA2B;AACrD,SAASC,kBAAkB,QAAQ,qCAAoC;AACvE,SAASC,0BAA0B,QAA6B,sBAAqB;AA2CrF,OAAO,MAAMC,eAAe,OAAOC;IACjC,MAAM,EACJC,EAAE,EACFC,MAAMC,OAAO,EACbC,YAAY,GAAG,EACfC,cAAc,EACdC,QAAQ,KAAK,EACbC,QAAQ,EACRC,QAAQC,eAAe,EACvBC,gBAAgB,EAChBC,MAAM,EACNC,MAAM,EACNC,OAAOC,aAAa,EACpBC,QAAQ,EACRC,QAAQC,eAAe,EACvBC,IAAI,EACJC,GAAG,EACHC,IAAI,EACJC,cAAc,EACdC,MAAM,EACNC,OAAOC,iBAAiB,CAAC,CAAC,EAC3B,GAAGxB;IACJ,MAAM,EAAEgB,QAAQS,aAAa,EAAEC,OAAO,EAAE,GAAGP;IAE3C,IAAIb,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;YACvBsB,SAAS;YACTvB;YACAwB,OAAOpB;YACPE;YACAC;QACF;IACF;IAEA,MAAMI,SAASC,mBAAmBQ;IAClC,MAAMK,mBAAmBJ,QAAQK,MAAM,CAACC,WAAW,CAACC,IAAI,CAAC,CAAC,EAAEC,IAAI,EAAE,GAAKA,SAAS7B;IAEhF,IAAI,CAACyB,kBAAkB;QACrB,MAAM,IAAI3C,SAAS,CAAC,qBAAqB,EAAEkB,eAAe,WAAW,CAAC;IACxE;IAEA,IAAI8B;IAEJ,IAAId,kBAAkBC,QAAQ;QAC5Ba,OAAQ,MAAMhB,IAAIO,OAAO,CAACU,QAAQ,CAAC;YACjCnC,IAAIqB;YACJe,YAAYhB;YACZiB,gBAAgB;QAClB;IACF;IAEA,IAAI,CAACH,QAAQhB,IAAIgB,IAAI,EAAE;QACrBA,OAAOhB,KAAKgB,MAAMlC,KAAKkB,IAAIgB,IAAI,GAAGhB,KAAKgB,MAAMA;IAC/C;IAEA,IAAI,CAACA,MAAM;QACT,MAAM,IAAIhD,SAAS;IACrB;IAEA,MAAM0C,QAAQpB,oBAAoB;IAClC,MAAM8B,cAAcC,QAAQV,iBAAiBW,QAAQ;IAErD,mDAAmD;IACnD,MAAMC,iBAAwBH,cAAc;QAAEI,SAAS;YAAEC,QAAQ;QAAY;IAAE,IAAI,CAAC;IAEpF,MAAMrB,QAAe;QACnBsB,KAAK;YAACrB;YAAgBK,QAAQ,CAAC,IAAIa;SAAe;IACpD;IAEA,MAAMxC,OAAO,GAAGC,WAAW,GAAGV,cAAc,CAAC,EAAEY,gBAAgB,CAAC,CAAC,EAAEO,QAAQ;IAC3E,MAAMkC,QAAQlC,WAAW;IACzB,MAAMmC,SAASC,MAAMC,OAAO,CAACtC,WAAWA,OAAOuC,MAAM,GAAG,IAAItD,UAAUe,UAAUwC;IAEhF,IAAI7C,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;YAAEsB,SAAS;YAAyB1B;YAAM4C;YAAO9B;QAAO;IACnF;IAEA,sCAAsC;IACtC,0DAA0D;IAC1D,sEAAsE;IACtE,IAAIoC;IAEJ,IAAI,OAAOrC,aAAa,YAAYA,WAAW,GAAG;QAChD,IAAI,OAAOD,kBAAkB,YAAYA,gBAAgB,GAAG;YAC1D,+CAA+C;YAC/CsC,qBAAqBC,KAAKC,GAAG,CAACxC,eAAeC;QAC/C,OAAO;YACL,8CAA8C;YAC9CqC,qBAAqBrC;QACvB;IACF,OAAO;QACL,6CAA6C;QAC7CqC,qBACE,OAAOtC,kBAAkB,YAAYA,gBAAgB,IAAIA,gBAAgBqC;IAC7E;IAEA,qEAAqE;IACrE,IAAII,YAAY;IAChB,IAAIC,eAAe;IACnB,IAAI;QACF,MAAMC,cAAc,MAAM/B,QAAQgC,KAAK,CAAC;YACtCrB,YAAYhC;YACZ8B;YACAnB;YACAsB,gBAAgB;QAClB;QACAiB,YAAYE,YAAYF,SAAS;IACnC,EAAE,OAAOI,OAAO;QACd,uDAAuD;QACvD,oCAAoC;QACpCH,eAAe;QACf,IAAIlD,OAAO;YACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;gBACvBsB,SAAS;gBACTvB;YACF;QACF;IACF;IAEA,MAAMuD,aAAaP,KAAKQ,GAAG,CAAC,GAAGR,KAAKS,IAAI,CAACP,YAAYnD;IACrD,MAAM2D,gBAAgB7C,QAAQ;IAC9B,MAAM8C,eAAeD,gBAAgBH,aAAa,IAAIG;IAEtD,MAAME,WAAW;QACf5B,YAAYhC;QACZ6D,OAAO;QACPrC;QACAhB,OAAOT;QACPY;QACAsB,gBAAgB;QAChBpB,MAAM;QACN6B;QACA3B;QACAe;QACAZ;IACF;IAEA,IAAIjB,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;YAAEsB,SAAS;YAAmBqC;QAAS;IAClE;IAEA,MAAME,iBAAiB3E,wBAAwB;QAC7CmB,QAAQmB,iBAAiBsC,eAAe;IAC1C;IAEA,+DAA+D;IAC/D,MAAMC,0BAA0B/E,+BAA+BwC,iBAAiBsC,eAAe;IAE/F,MAAME,iBACJxC,iBAAiByC,KAAK,EAAEC,QAAQ,CAAC,uBAAuB,EAAEF,kBAAkB,EAAE;IAEhF,MAAMG,mBAAmBH,eAAeI,GAAG,CAACrF;IAE5C,MAAMsF,oBAAoB,CAACC;QACzB,MAAMC,WAAoC,CAAC;QAE3C,KAAK,MAAM,CAACC,KAAKC,MAAM,IAAIC,OAAOC,OAAO,CAACL,KAAM;YAC9C,MAAMM,aAAaT,iBAAiBU,IAAI,CAAC,CAACC,UAAYA,QAAQC,IAAI,CAACP;YACnE,IAAI,CAACI,YAAY;gBACfL,QAAQ,CAACC,IAAI,GAAGC;YAClB;QACF;QAEA,OAAOF;IACT;IAEA,MAAMS,qBAAqB,CAACC,KAAUC,aAAa,EAAE;QACnD,IAAIxC,MAAMC,OAAO,CAACsC,MAAM;YACtB,OAAOA,IAAIb,GAAG,CAAC,CAACe,OAASH,mBAAmBG,MAAMD;QACpD;QAEA,IAAI,OAAOD,QAAQ,YAAYA,QAAQ,MAAM;YAC3C,OAAOA;QACT;QAEA,MAAMV,WAAgC,CAAC;QACvC,KAAK,MAAM,CAACC,KAAKC,MAAM,IAAIC,OAAOC,OAAO,CAACM,KAAM;YAC9C,MAAMG,cAAcF,aAAa,GAAGA,WAAW,CAAC,EAAEV,KAAK,GAAGA;YAE1D,6CAA6C;YAC7C,MAAMI,aAAaZ,eAAeqB,QAAQ,CAACD;YAE3C,IAAI,CAACR,YAAY;gBACfL,QAAQ,CAACC,IAAI,GAAGQ,mBAAmBP,OAAOW;YAC5C;QACF;QAEA,OAAOb;IACT;IAEA,IAAItE,UAAU;QACZ,MAAMqF,gBAAgB/F,mBAAmBiB,eAAeK,IAAI0E,CAAC;QAC7D,IAAID,eAAe;YACjB,MAAM,IAAIzG,SAASyG;QACrB;QAEA,oFAAoF;QACpF,IAAIE,gBAA0B,EAAE;QAChC,IAAIhD,OAAO;YACT,MAAMiD,cACJ/E,WAAW,SAASU,QAAQK,MAAM,CAACiE,YAAY,GAC3CtE,QAAQK,MAAM,CAACiE,YAAY,CAACD,WAAW,GACvC5C;YAEN2C,gBAAgBpG,iBAAiB;gBAC/BoC;gBACAwC;gBACA3D;gBACAK;gBACA+E;gBACA1B;YACF;YAEA,IAAI/D,OAAO;gBACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;oBACvB2F,aAAaH,cAAc5C,MAAM;oBACjCgD,KAAK;gBACP;YACF;QACF;QAEA,gFAAgF;QAChF,IAAIC,aAAuB,EAAE;QAC7B,IAAIC,mBAAmB;QAEvB,MAAMC,UAAU,IAAIC;QACpB,IAAIC,eAAe;QACnB,IAAIC,mBAAmBxC;QACvB,IAAIyC,UAAU;QACd,MAAMC,UACJ,OAAOtD,uBAAuB,WAAWA,qBAAqBuD,OAAOC,iBAAiB;QAExF,MAAMC,SAAS,IAAIzH,SAAS;YAC1B,MAAM0H;gBACJ,MAAMC,YAAY1D,KAAKQ,GAAG,CAAC,GAAG6C,UAAUD;gBAExC,IAAIM,cAAc,GAAG;oBACnB,IAAI,CAACjE,OAAO;wBACV,6EAA6E;wBAC7E,IAAI,CAACkE,IAAI,CAACX,QAAQY,MAAM,CAACV,eAAe,OAAO;oBACjD;oBACA,IAAI,CAACS,IAAI,CAAC;oBACV;gBACF;gBAEA,MAAME,SAAS,MAAMxF,QAAQO,IAAI,CAAC;oBAChC,GAAGgC,QAAQ;oBACX/C,MAAMsF;oBACN3F,OAAOwC,KAAKC,GAAG,CAAClD,WAAW2G;gBAC7B;gBAEA,IAAIzG,OAAO;oBACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CACtB,CAAC,gBAAgB,EAAEkG,iBAAiB,MAAM,EAAEU,OAAOC,IAAI,CAACjE,MAAM,CAAC,KAAK,CAAC;gBAEzE;gBAEA,IAAIgE,OAAOC,IAAI,CAACjE,MAAM,KAAK,GAAG;oBAC5B,oCAAoC;oBACpC,IAAI,CAACJ,OAAO;wBACV,wEAAwE;wBACxE,IAAI,CAACkE,IAAI,CAACX,QAAQY,MAAM,CAACV,eAAe,OAAO;oBACjD;oBACA,IAAI,CAACS,IAAI,CAAC;oBACV;gBACF;gBAEA,IAAIlE,OAAO;oBACT,wBAAwB;oBACxB,MAAMsE,YAAYF,OAAOC,IAAI,CAACzC,GAAG,CAAC,CAACa,MACjCZ,kBACEpF,cAAc;4BAAEgG;4BAAK5E;4BAAQ0D;4BAAyBF;wBAAe;oBAIzE,8EAA8E;oBAC9E,IAAI,CAACiC,kBAAkB;wBACrB,MAAMiB,cAAwB,EAAE;wBAChC,MAAMC,WAAW,IAAIC;wBACrB,KAAK,MAAM3C,OAAOwC,UAAW;4BAC3B,KAAK,MAAMtC,OAAOE,OAAOwC,IAAI,CAAC5C,KAAM;gCAClC,IAAI,CAAC0C,SAASG,GAAG,CAAC3C,MAAM;oCACtBwC,SAASI,GAAG,CAAC5C;oCACbuC,YAAYL,IAAI,CAAClC;gCACnB;4BACF;wBACF;wBACA,oDAAoD;wBACpDqB,aAAaxG,aAAamG,eAAeuB;wBACzCjB,mBAAmB;wBAEnB,IAAI9F,OAAO;4BACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;gCACvBqH,kBAAkBN,YAAYnE,MAAM;gCACpC0E,mBAAmBzB,WAAWjD,MAAM;gCACpCgD,KAAK;4BACP;wBACF;oBACF;oBAEA,MAAM2B,aAAaT,UAAU1C,GAAG,CAAC,CAACE;wBAChC,MAAMkD,UAAmC,CAAC;wBAC1C,KAAK,MAAMC,OAAO5B,WAAY;4BAC5B2B,OAAO,CAACC,IAAI,GAAGnD,GAAG,CAACmD,IAAI,IAAI;wBAC7B;wBACA,OAAOD;oBACT;oBAEA,MAAME,YAAY9I,UAAU2I,YAAY;wBACtCI,KAAK1B;wBACL2B,QAAQ3B;wBACR4B,SAAShC;oBACX;oBAEA,IAAI,CAACa,IAAI,CAACX,QAAQY,MAAM,CAACe;gBAC3B,OAAO;oBACL,yBAAyB;oBACzB,MAAMZ,YAAYF,OAAOC,IAAI,CAACzC,GAAG,CAAC,CAACa,MAAQD,mBAAmBC;oBAE9D,uDAAuD;oBACvD,MAAM6C,YAAYhB,UAAU1C,GAAG,CAAC,CAACE,MAAQyD,KAAKnJ,SAAS,CAAC0F,MAAM0D,IAAI,CAAC;oBAEnE,IAAI/B,cAAc;wBAChB,IAAI,CAACS,IAAI,CAACX,QAAQY,MAAM,CAAC,MAAMmB;oBACjC,OAAO;wBACL,IAAI,CAACpB,IAAI,CAACX,QAAQY,MAAM,CAAC,MAAMmB;oBACjC;gBACF;gBAEA3B,WAAWS,OAAOC,IAAI,CAACjE,MAAM;gBAC7BqD,eAAe;gBACfC,oBAAoB;gBAEpB,IAAI,CAACU,OAAOqB,WAAW,IAAI9B,WAAWC,SAAS;oBAC7C,IAAIpG,OAAO;wBACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;oBAC3B;oBACA,IAAI,CAACwC,OAAO;wBACV,IAAI,CAACkE,IAAI,CAACX,QAAQY,MAAM,CAAC;oBAC3B;oBACA,IAAI,CAACD,IAAI,CAAC,OAAM,iBAAiB;gBACnC;YACF;QACF;QAEA,OAAO,IAAIwB,SAASpJ,SAASqJ,KAAK,CAAC5B,SAA2B;YAC5D6B,SAAS;gBACP,uBAAuB,CAAC,sBAAsB,EAAExI,KAAK,CAAC,CAAC;gBACvD,gBAAgB4C,QAAQ,aAAa;YACvC;QACF;IACF;IAEA,sCAAsC;IACtC,IAAIxC,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;IAC3B;IAEA,gCAAgC;IAChC,MAAMqI,YAAY7I,2BAA2B;QAAEM;QAAWE;IAAM;IAEhE,qCAAqC;IACrC,MAAMsI,eAAe,CAACrD,MACpBzC,QACI6B,kBAAkBpF,cAAc;YAAEgG;YAAK5E;YAAQ0D;YAAyBF;QAAe,MACvFmB,mBAAmBC;IAEzB,oEAAoE;IACpE,IAAIsD,eAAe;QACjBV,SAAS,EAAE;QACXhB,MAAM,EAAE;QACR2B,cAAc;IAChB;IAEA,IAAI,CAACtF,cAAc;QACjBqF,eAAe,MAAMF,UAAUI,aAAa,CAAC;YAC3C1I;YACA4D,UAAUA;YACVrD;YACA8F,SACE,OAAOtD,uBAAuB,WAAWA,qBAAqBuD,OAAOC,iBAAiB;YACxFzF;YACA6H,WAAWhF;YACX4E;QACF;IACF;IAEA,MAAM,EAAET,SAASd,WAAW,EAAEF,MAAM8B,IAAI,EAAE,GAAGJ;IAC7C,MAAMK,aAAuB,EAAE;IAE/B,uBAAuB;IACvB,IAAIpG,OAAO;QACT,mDAAmD;QACnD,MAAMiD,cACJ/E,WAAW,SAASU,QAAQK,MAAM,CAACiE,YAAY,GAC3CtE,QAAQK,MAAM,CAACiE,YAAY,CAACD,WAAW,GACvC5C;QAEN,MAAM2C,gBAAgBpG,iBAAiB;YACrCoC;YACAwC;YACA3D;YACAK;YACA+E;YACA1B;QACF;QAEA,oDAAoD;QACpD,uFAAuF;QACvF,MAAM8E,eAAexJ,aAAamG,eAAeuB;QAEjD,MAAMQ,aAAaoB,KAAKvE,GAAG,CAAC,CAACE;YAC3B,MAAMkD,UAAmC,CAAC;YAC1C,KAAK,MAAMC,OAAOoB,aAAc;gBAC9BrB,OAAO,CAACC,IAAI,GAAGnD,GAAG,CAACmD,IAAI,IAAI;YAC7B;YACA,OAAOD;QACT;QAEA,+CAA+C;QAC/CoB,WAAWlC,IAAI,CACb9H,UAAU2I,YAAY;YACpBI,KAAK;YACLC,QAAQ;YACRC,SAASgB;QACX;IAEJ,OAAO;QACL,cAAc;QACdD,WAAWlC,IAAI,CAACiC,KAAKvE,GAAG,CAAC,CAACa,MAAQ8C,KAAKnJ,SAAS,CAACqG,MAAM+C,IAAI,CAAC;IAC9D;IAEA,mDAAmD;IACnD,uCAAuC;IACvC,yGAAyG;IACzG,MAAMc,UAAUxI,WAAW,SAAS,CAAC,CAAC,EAAEsI,WAAWZ,IAAI,CAAC,KAAK,CAAC,CAAC,GAAGY,WAAWZ,IAAI,CAAC;IAClF,MAAMe,SAASC,OAAOC,IAAI,CAACH,QAAQlG,MAAM,GAAG,IAAIkG,UAAU;IAC1D,IAAI9I,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC,GAAGM,OAAO,yBAAyB,CAAC;IAC/D;IAEA,IAAI,CAACX,IAAI;QACP,IAAIK,OAAO;YACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;QAC3B;QACAa,IAAIqI,IAAI,GAAG;YACTtJ;YACAuJ,MAAMJ;YACNK,UAAU5G,QAAQ,aAAa;YAC/B6G,MAAMN,OAAOnG,MAAM;QACrB;IACF,OAAO;QACL,IAAI5C,OAAO;YACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC,CAAC,kCAAkC,EAAEL,IAAI;QACpE;QACA,MAAMkB,IAAIO,OAAO,CAACkI,MAAM,CAAC;YACvB3J;YACAoC,YAAY3B;YACZ+I,MAAM,CAAC;YACPD,MAAM;gBACJtJ;gBACAuJ,MAAMJ;gBACNK,UAAU5G,QAAQ,aAAa;gBAC/B6G,MAAMN,OAAOnG,MAAM;YACrB;YACA,uGAAuG;YACvGZ,gBAAgB;YAChBnB;QACF;IACF;IACA,IAAIb,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;IAC3B;AACF,EAAC"}
|
|
1
|
+
{"version":3,"sources":["../../src/export/createExport.ts"],"sourcesContent":["/* eslint-disable perfectionist/sort-objects */\nimport type { PayloadRequest, Sort, TypedUser, Where } from 'payload'\n\nimport { stringify } from 'csv-stringify/sync'\nimport { APIError } from 'payload'\nimport { Readable } from 'stream'\n\nimport { buildDisabledFieldRegex } from '../utilities/buildDisabledFieldRegex.js'\nimport { flattenObject } from '../utilities/flattenObject.js'\nimport { getExportFieldFunctions } from '../utilities/getExportFieldFunctions.js'\nimport { getFilename } from '../utilities/getFilename.js'\nimport { getSchemaColumns, mergeColumns } from '../utilities/getSchemaColumns.js'\nimport { getSelect } from '../utilities/getSelect.js'\nimport { validateLimitValue } from '../utilities/validateLimitValue.js'\nimport { createExportBatchProcessor, type ExportFindArgs } from './batchProcessor.js'\n\nexport type Export = {\n /**\n * Number of documents to process in each batch during export\n * @default 100\n */\n batchSize?: number\n collectionSlug: string\n /**\n * If true, enables debug logging\n */\n debug?: boolean\n drafts?: 'no' | 'yes'\n exportCollection: string\n fields?: string[]\n format: 'csv' | 'json'\n globals?: string[]\n id: number | string\n limit?: number\n locale?: string\n /**\n * Maximum number of documents that can be exported in a single operation.\n * This value has already been resolved from the plugin config.\n */\n maxLimit?: number\n name: string\n page?: number\n slug: string\n sort: Sort\n userCollection: string\n userID: number | string\n where?: Where\n}\n\nexport type CreateExportArgs = {\n /**\n * If true, stream the file instead of saving it\n */\n download?: boolean\n req: PayloadRequest\n} & Export\n\nexport const createExport = async (args: CreateExportArgs) => {\n const {\n id,\n name: nameArg,\n batchSize = 100,\n collectionSlug,\n debug = false,\n download,\n drafts: draftsFromInput,\n exportCollection,\n fields,\n format,\n limit: incomingLimit,\n maxLimit,\n locale: localeFromInput,\n page,\n req,\n sort,\n userCollection,\n userID,\n where: whereFromInput = {},\n } = args\n const { locale: localeFromReq, payload } = req\n\n if (debug) {\n req.payload.logger.debug({\n message: 'Starting export process with args:',\n collectionSlug,\n draft: draftsFromInput,\n fields,\n format,\n })\n }\n\n const locale = localeFromInput ?? localeFromReq\n const collectionConfig = payload.config.collections.find(({ slug }) => slug === collectionSlug)\n\n if (!collectionConfig) {\n throw new APIError(`Collection with slug ${collectionSlug} not found.`)\n }\n\n let user: TypedUser | undefined\n\n if (userCollection && userID) {\n user = (await req.payload.findByID({\n id: userID,\n collection: userCollection,\n overrideAccess: true,\n })) as TypedUser\n }\n\n if (!user && req.user) {\n user = req?.user?.id ? req.user : req?.user?.user\n }\n\n if (!user) {\n throw new APIError('User authentication is required to create exports.')\n }\n\n const draft = draftsFromInput === 'yes'\n const hasVersions = Boolean(collectionConfig.versions)\n\n // Only filter by _status for versioned collections\n const publishedWhere: Where = hasVersions ? { _status: { equals: 'published' } } : {}\n\n const where: Where = {\n and: [whereFromInput, draft ? {} : publishedWhere],\n }\n\n const name = `${nameArg ?? `${getFilename()}-${collectionSlug}`}.${format}`\n const isCSV = format === 'csv'\n const select = Array.isArray(fields) && fields.length > 0 ? getSelect(fields) : undefined\n\n if (debug) {\n req.payload.logger.debug({ message: 'Export configuration:', name, isCSV, locale })\n }\n\n // Determine maximum export documents:\n // 1. If maxLimit is defined, it sets the absolute ceiling\n // 2. User's limit is applied but clamped to maxLimit if it exceeds it\n let maxExportDocuments: number | undefined\n\n if (typeof maxLimit === 'number' && maxLimit > 0) {\n if (typeof incomingLimit === 'number' && incomingLimit > 0) {\n // User provided a limit - clamp it to maxLimit\n maxExportDocuments = Math.min(incomingLimit, maxLimit)\n } else {\n // No user limit - use maxLimit as the ceiling\n maxExportDocuments = maxLimit\n }\n } else {\n // No maxLimit - use user's limit if provided\n maxExportDocuments =\n typeof incomingLimit === 'number' && incomingLimit > 0 ? incomingLimit : undefined\n }\n\n // Try to count documents - if access is denied, treat as 0 documents\n let totalDocs = 0\n let accessDenied = false\n try {\n const countResult = await payload.count({\n collection: collectionSlug,\n user,\n locale,\n overrideAccess: false,\n })\n totalDocs = countResult.totalDocs\n } catch (error) {\n // Access denied - user can't read from this collection\n // We'll create an empty export file\n accessDenied = true\n if (debug) {\n req.payload.logger.debug({\n message: 'Access denied for collection, creating empty export',\n collectionSlug,\n })\n }\n }\n\n const totalPages = Math.max(1, Math.ceil(totalDocs / batchSize))\n const requestedPage = page || 1\n const adjustedPage = requestedPage > totalPages ? 1 : requestedPage\n\n const findArgs = {\n collection: collectionSlug,\n depth: 1,\n draft,\n limit: batchSize,\n locale,\n overrideAccess: false,\n page: 0,\n select,\n sort,\n user,\n where,\n }\n\n if (debug) {\n req.payload.logger.debug({ message: 'Find arguments:', findArgs })\n }\n\n const toCSVFunctions = getExportFieldFunctions({\n fields: collectionConfig.flattenedFields,\n })\n\n const disabledFields =\n collectionConfig.admin?.custom?.['plugin-import-export']?.disabledFields ?? []\n\n const disabledMatchers = disabledFields.map(buildDisabledFieldRegex)\n\n const filterDisabledCSV = (row: Record<string, unknown>): Record<string, unknown> => {\n const filtered: Record<string, unknown> = {}\n\n for (const [key, value] of Object.entries(row)) {\n const isDisabled = disabledMatchers.some((matcher) => matcher.test(key))\n if (!isDisabled) {\n filtered[key] = value\n }\n }\n\n return filtered\n }\n\n const filterDisabledJSON = (doc: any, parentPath = ''): any => {\n if (Array.isArray(doc)) {\n return doc.map((item) => filterDisabledJSON(item, parentPath))\n }\n\n if (typeof doc !== 'object' || doc === null) {\n return doc\n }\n\n const filtered: Record<string, any> = {}\n for (const [key, value] of Object.entries(doc)) {\n const currentPath = parentPath ? `${parentPath}.${key}` : key\n\n // Only remove if this exact path is disabled\n const isDisabled = disabledFields.includes(currentPath)\n\n if (!isDisabled) {\n filtered[key] = filterDisabledJSON(value, currentPath)\n }\n }\n\n return filtered\n }\n\n if (download) {\n const limitErrorMsg = validateLimitValue(incomingLimit, req.t)\n if (limitErrorMsg) {\n throw new APIError(limitErrorMsg)\n }\n\n // Get schema-based columns first (provides base ordering and handles empty exports)\n let schemaColumns: string[] = []\n if (isCSV) {\n const localeCodes =\n locale === 'all' && payload.config.localization\n ? payload.config.localization.localeCodes\n : undefined\n\n schemaColumns = getSchemaColumns({\n collectionConfig,\n disabledFields,\n fields,\n locale,\n localeCodes,\n })\n\n if (debug) {\n req.payload.logger.debug({\n columnCount: schemaColumns.length,\n msg: 'Schema-based column inference complete',\n })\n }\n }\n\n // allColumns will be finalized after first batch (schema + data columns merged)\n let allColumns: string[] = []\n let columnsFinalized = false\n\n const encoder = new TextEncoder()\n let isFirstBatch = true\n let currentBatchPage = adjustedPage\n let fetched = 0\n const maxDocs =\n typeof maxExportDocuments === 'number' ? maxExportDocuments : Number.POSITIVE_INFINITY\n\n const stream = new Readable({\n async read() {\n const remaining = Math.max(0, maxDocs - fetched)\n\n if (remaining === 0) {\n if (!isCSV) {\n // If first batch with no remaining, output empty array; otherwise just close\n this.push(encoder.encode(isFirstBatch ? '[]' : ']'))\n }\n this.push(null)\n return\n }\n\n const result = await payload.find({\n ...findArgs,\n page: currentBatchPage,\n limit: Math.min(batchSize, remaining),\n })\n\n if (debug) {\n req.payload.logger.debug(\n `Streaming batch ${currentBatchPage} with ${result.docs.length} docs`,\n )\n }\n\n if (result.docs.length === 0) {\n // Close JSON array properly if JSON\n if (!isCSV) {\n // If first batch with no docs, output empty array; otherwise just close\n this.push(encoder.encode(isFirstBatch ? '[]' : ']'))\n }\n this.push(null)\n return\n }\n\n if (isCSV) {\n // --- CSV Streaming ---\n const batchRows = result.docs.map((doc) =>\n filterDisabledCSV(\n flattenObject({\n doc,\n fields,\n toCSVFunctions,\n }),\n ),\n )\n\n // On first batch, discover additional columns from data and merge with schema\n if (!columnsFinalized) {\n const dataColumns: string[] = []\n const seenCols = new Set<string>()\n for (const row of batchRows) {\n for (const key of Object.keys(row)) {\n if (!seenCols.has(key)) {\n seenCols.add(key)\n dataColumns.push(key)\n }\n }\n }\n // Merge schema columns with data-discovered columns\n allColumns = mergeColumns(schemaColumns, dataColumns)\n columnsFinalized = true\n\n if (debug) {\n req.payload.logger.debug({\n dataColumnsCount: dataColumns.length,\n finalColumnsCount: allColumns.length,\n msg: 'Merged schema and data columns',\n })\n }\n }\n\n const paddedRows = batchRows.map((row) => {\n const fullRow: Record<string, unknown> = {}\n for (const col of allColumns) {\n fullRow[col] = row[col] ?? ''\n }\n return fullRow\n })\n\n const csvString = stringify(paddedRows, {\n bom: isFirstBatch,\n header: isFirstBatch,\n columns: allColumns,\n })\n\n this.push(encoder.encode(csvString))\n } else {\n // --- JSON Streaming ---\n const batchRows = result.docs.map((doc) => filterDisabledJSON(doc))\n\n // Convert each filtered/flattened row into JSON string\n const batchJSON = batchRows.map((row) => JSON.stringify(row)).join(',')\n\n if (isFirstBatch) {\n this.push(encoder.encode('[' + batchJSON))\n } else {\n this.push(encoder.encode(',' + batchJSON))\n }\n }\n\n fetched += result.docs.length\n isFirstBatch = false\n currentBatchPage += 1\n\n if (!result.hasNextPage || fetched >= maxDocs) {\n if (debug) {\n req.payload.logger.debug('Stream complete - no more pages')\n }\n if (!isCSV) {\n this.push(encoder.encode(']'))\n }\n this.push(null) // End the stream\n }\n },\n })\n\n return new Response(Readable.toWeb(stream) as ReadableStream, {\n headers: {\n 'Content-Disposition': `attachment; filename=\"${name}\"`,\n 'Content-Type': isCSV ? 'text/csv' : 'application/json',\n },\n })\n }\n\n // Non-download path (buffered export)\n if (debug) {\n req.payload.logger.debug('Starting file generation')\n }\n\n // Create export batch processor\n const processor = createExportBatchProcessor({ batchSize, debug })\n\n // Transform function based on format\n const transformDoc = (doc: unknown) =>\n isCSV\n ? filterDisabledCSV(\n flattenObject({\n doc,\n fields,\n toCSVFunctions,\n }),\n )\n : filterDisabledJSON(doc)\n\n // Skip fetching if access was denied - we'll create an empty export\n let exportResult = {\n columns: [] as string[],\n docs: [] as Record<string, unknown>[],\n fetchedCount: 0,\n }\n\n if (!accessDenied) {\n exportResult = await processor.processExport({\n collectionSlug,\n findArgs: findArgs as ExportFindArgs,\n format,\n maxDocs:\n typeof maxExportDocuments === 'number' ? maxExportDocuments : Number.POSITIVE_INFINITY,\n req,\n startPage: adjustedPage,\n transformDoc,\n })\n }\n\n const { columns: dataColumns, docs: rows } = exportResult\n const outputData: string[] = []\n\n // Prepare final output\n if (isCSV) {\n // Get schema-based columns for consistent ordering\n const localeCodes =\n locale === 'all' && payload.config.localization\n ? payload.config.localization.localeCodes\n : undefined\n\n const schemaColumns = getSchemaColumns({\n collectionConfig,\n disabledFields,\n fields,\n locale,\n localeCodes,\n })\n\n // Merge schema columns with data-discovered columns\n // Schema provides ordering, data provides additional columns (e.g., array indices > 0)\n const finalColumns = mergeColumns(schemaColumns, dataColumns)\n\n const paddedRows = rows.map((row) => {\n const fullRow: Record<string, unknown> = {}\n for (const col of finalColumns) {\n fullRow[col] = row[col] ?? ''\n }\n return fullRow\n })\n\n // Always output CSV with header, even if empty\n outputData.push(\n stringify(paddedRows, {\n bom: true,\n header: true,\n columns: finalColumns,\n }),\n )\n } else {\n // JSON format\n outputData.push(rows.map((doc) => JSON.stringify(doc)).join(',\\n'))\n }\n\n // Ensure we always have valid content for the file\n // For JSON, empty exports produce \"[]\"\n // For CSV, if completely empty (no columns, no rows), produce at least a newline to ensure file creation\n const content = format === 'json' ? `[${outputData.join(',')}]` : outputData.join('')\n const buffer = Buffer.from(content.length > 0 ? content : '\\n')\n if (debug) {\n req.payload.logger.debug(`${format} file generation complete`)\n }\n\n if (!id) {\n if (debug) {\n req.payload.logger.debug('Creating new export file')\n }\n req.file = {\n name,\n data: buffer,\n mimetype: isCSV ? 'text/csv' : 'application/json',\n size: buffer.length,\n }\n } else {\n if (debug) {\n req.payload.logger.debug(`Updating existing export with id: ${id}`)\n }\n await req.payload.update({\n id,\n collection: exportCollection,\n data: {},\n file: {\n name,\n data: buffer,\n mimetype: isCSV ? 'text/csv' : 'application/json',\n size: buffer.length,\n },\n // Override access only here so that we can be sure the export collection itself is updated as expected\n overrideAccess: true,\n req,\n })\n }\n if (debug) {\n req.payload.logger.debug('Export process completed successfully')\n }\n}\n"],"names":["stringify","APIError","Readable","buildDisabledFieldRegex","flattenObject","getExportFieldFunctions","getFilename","getSchemaColumns","mergeColumns","getSelect","validateLimitValue","createExportBatchProcessor","createExport","args","id","name","nameArg","batchSize","collectionSlug","debug","download","drafts","draftsFromInput","exportCollection","fields","format","limit","incomingLimit","maxLimit","locale","localeFromInput","page","req","sort","userCollection","userID","where","whereFromInput","localeFromReq","payload","logger","message","draft","collectionConfig","config","collections","find","slug","user","findByID","collection","overrideAccess","hasVersions","Boolean","versions","publishedWhere","_status","equals","and","isCSV","select","Array","isArray","length","undefined","maxExportDocuments","Math","min","totalDocs","accessDenied","countResult","count","error","totalPages","max","ceil","requestedPage","adjustedPage","findArgs","depth","toCSVFunctions","flattenedFields","disabledFields","admin","custom","disabledMatchers","map","filterDisabledCSV","row","filtered","key","value","Object","entries","isDisabled","some","matcher","test","filterDisabledJSON","doc","parentPath","item","currentPath","includes","limitErrorMsg","t","schemaColumns","localeCodes","localization","columnCount","msg","allColumns","columnsFinalized","encoder","TextEncoder","isFirstBatch","currentBatchPage","fetched","maxDocs","Number","POSITIVE_INFINITY","stream","read","remaining","push","encode","result","docs","batchRows","dataColumns","seenCols","Set","keys","has","add","dataColumnsCount","finalColumnsCount","paddedRows","fullRow","col","csvString","bom","header","columns","batchJSON","JSON","join","hasNextPage","Response","toWeb","headers","processor","transformDoc","exportResult","fetchedCount","processExport","startPage","rows","outputData","finalColumns","content","buffer","Buffer","from","file","data","mimetype","size","update"],"mappings":"AAAA,6CAA6C,GAG7C,SAASA,SAAS,QAAQ,qBAAoB;AAC9C,SAASC,QAAQ,QAAQ,UAAS;AAClC,SAASC,QAAQ,QAAQ,SAAQ;AAEjC,SAASC,uBAAuB,QAAQ,0CAAyC;AACjF,SAASC,aAAa,QAAQ,gCAA+B;AAC7D,SAASC,uBAAuB,QAAQ,0CAAyC;AACjF,SAASC,WAAW,QAAQ,8BAA6B;AACzD,SAASC,gBAAgB,EAAEC,YAAY,QAAQ,mCAAkC;AACjF,SAASC,SAAS,QAAQ,4BAA2B;AACrD,SAASC,kBAAkB,QAAQ,qCAAoC;AACvE,SAASC,0BAA0B,QAA6B,sBAAqB;AA2CrF,OAAO,MAAMC,eAAe,OAAOC;IACjC,MAAM,EACJC,EAAE,EACFC,MAAMC,OAAO,EACbC,YAAY,GAAG,EACfC,cAAc,EACdC,QAAQ,KAAK,EACbC,QAAQ,EACRC,QAAQC,eAAe,EACvBC,gBAAgB,EAChBC,MAAM,EACNC,MAAM,EACNC,OAAOC,aAAa,EACpBC,QAAQ,EACRC,QAAQC,eAAe,EACvBC,IAAI,EACJC,GAAG,EACHC,IAAI,EACJC,cAAc,EACdC,MAAM,EACNC,OAAOC,iBAAiB,CAAC,CAAC,EAC3B,GAAGxB;IACJ,MAAM,EAAEgB,QAAQS,aAAa,EAAEC,OAAO,EAAE,GAAGP;IAE3C,IAAIb,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;YACvBsB,SAAS;YACTvB;YACAwB,OAAOpB;YACPE;YACAC;QACF;IACF;IAEA,MAAMI,SAASC,mBAAmBQ;IAClC,MAAMK,mBAAmBJ,QAAQK,MAAM,CAACC,WAAW,CAACC,IAAI,CAAC,CAAC,EAAEC,IAAI,EAAE,GAAKA,SAAS7B;IAEhF,IAAI,CAACyB,kBAAkB;QACrB,MAAM,IAAI1C,SAAS,CAAC,qBAAqB,EAAEiB,eAAe,WAAW,CAAC;IACxE;IAEA,IAAI8B;IAEJ,IAAId,kBAAkBC,QAAQ;QAC5Ba,OAAQ,MAAMhB,IAAIO,OAAO,CAACU,QAAQ,CAAC;YACjCnC,IAAIqB;YACJe,YAAYhB;YACZiB,gBAAgB;QAClB;IACF;IAEA,IAAI,CAACH,QAAQhB,IAAIgB,IAAI,EAAE;QACrBA,OAAOhB,KAAKgB,MAAMlC,KAAKkB,IAAIgB,IAAI,GAAGhB,KAAKgB,MAAMA;IAC/C;IAEA,IAAI,CAACA,MAAM;QACT,MAAM,IAAI/C,SAAS;IACrB;IAEA,MAAMyC,QAAQpB,oBAAoB;IAClC,MAAM8B,cAAcC,QAAQV,iBAAiBW,QAAQ;IAErD,mDAAmD;IACnD,MAAMC,iBAAwBH,cAAc;QAAEI,SAAS;YAAEC,QAAQ;QAAY;IAAE,IAAI,CAAC;IAEpF,MAAMrB,QAAe;QACnBsB,KAAK;YAACrB;YAAgBK,QAAQ,CAAC,IAAIa;SAAe;IACpD;IAEA,MAAMxC,OAAO,GAAGC,WAAW,GAAGV,cAAc,CAAC,EAAEY,gBAAgB,CAAC,CAAC,EAAEO,QAAQ;IAC3E,MAAMkC,QAAQlC,WAAW;IACzB,MAAMmC,SAASC,MAAMC,OAAO,CAACtC,WAAWA,OAAOuC,MAAM,GAAG,IAAItD,UAAUe,UAAUwC;IAEhF,IAAI7C,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;YAAEsB,SAAS;YAAyB1B;YAAM4C;YAAO9B;QAAO;IACnF;IAEA,sCAAsC;IACtC,0DAA0D;IAC1D,sEAAsE;IACtE,IAAIoC;IAEJ,IAAI,OAAOrC,aAAa,YAAYA,WAAW,GAAG;QAChD,IAAI,OAAOD,kBAAkB,YAAYA,gBAAgB,GAAG;YAC1D,+CAA+C;YAC/CsC,qBAAqBC,KAAKC,GAAG,CAACxC,eAAeC;QAC/C,OAAO;YACL,8CAA8C;YAC9CqC,qBAAqBrC;QACvB;IACF,OAAO;QACL,6CAA6C;QAC7CqC,qBACE,OAAOtC,kBAAkB,YAAYA,gBAAgB,IAAIA,gBAAgBqC;IAC7E;IAEA,qEAAqE;IACrE,IAAII,YAAY;IAChB,IAAIC,eAAe;IACnB,IAAI;QACF,MAAMC,cAAc,MAAM/B,QAAQgC,KAAK,CAAC;YACtCrB,YAAYhC;YACZ8B;YACAnB;YACAsB,gBAAgB;QAClB;QACAiB,YAAYE,YAAYF,SAAS;IACnC,EAAE,OAAOI,OAAO;QACd,uDAAuD;QACvD,oCAAoC;QACpCH,eAAe;QACf,IAAIlD,OAAO;YACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;gBACvBsB,SAAS;gBACTvB;YACF;QACF;IACF;IAEA,MAAMuD,aAAaP,KAAKQ,GAAG,CAAC,GAAGR,KAAKS,IAAI,CAACP,YAAYnD;IACrD,MAAM2D,gBAAgB7C,QAAQ;IAC9B,MAAM8C,eAAeD,gBAAgBH,aAAa,IAAIG;IAEtD,MAAME,WAAW;QACf5B,YAAYhC;QACZ6D,OAAO;QACPrC;QACAhB,OAAOT;QACPY;QACAsB,gBAAgB;QAChBpB,MAAM;QACN6B;QACA3B;QACAe;QACAZ;IACF;IAEA,IAAIjB,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;YAAEsB,SAAS;YAAmBqC;QAAS;IAClE;IAEA,MAAME,iBAAiB3E,wBAAwB;QAC7CmB,QAAQmB,iBAAiBsC,eAAe;IAC1C;IAEA,MAAMC,iBACJvC,iBAAiBwC,KAAK,EAAEC,QAAQ,CAAC,uBAAuB,EAAEF,kBAAkB,EAAE;IAEhF,MAAMG,mBAAmBH,eAAeI,GAAG,CAACnF;IAE5C,MAAMoF,oBAAoB,CAACC;QACzB,MAAMC,WAAoC,CAAC;QAE3C,KAAK,MAAM,CAACC,KAAKC,MAAM,IAAIC,OAAOC,OAAO,CAACL,KAAM;YAC9C,MAAMM,aAAaT,iBAAiBU,IAAI,CAAC,CAACC,UAAYA,QAAQC,IAAI,CAACP;YACnE,IAAI,CAACI,YAAY;gBACfL,QAAQ,CAACC,IAAI,GAAGC;YAClB;QACF;QAEA,OAAOF;IACT;IAEA,MAAMS,qBAAqB,CAACC,KAAUC,aAAa,EAAE;QACnD,IAAIvC,MAAMC,OAAO,CAACqC,MAAM;YACtB,OAAOA,IAAIb,GAAG,CAAC,CAACe,OAASH,mBAAmBG,MAAMD;QACpD;QAEA,IAAI,OAAOD,QAAQ,YAAYA,QAAQ,MAAM;YAC3C,OAAOA;QACT;QAEA,MAAMV,WAAgC,CAAC;QACvC,KAAK,MAAM,CAACC,KAAKC,MAAM,IAAIC,OAAOC,OAAO,CAACM,KAAM;YAC9C,MAAMG,cAAcF,aAAa,GAAGA,WAAW,CAAC,EAAEV,KAAK,GAAGA;YAE1D,6CAA6C;YAC7C,MAAMI,aAAaZ,eAAeqB,QAAQ,CAACD;YAE3C,IAAI,CAACR,YAAY;gBACfL,QAAQ,CAACC,IAAI,GAAGQ,mBAAmBP,OAAOW;YAC5C;QACF;QAEA,OAAOb;IACT;IAEA,IAAIrE,UAAU;QACZ,MAAMoF,gBAAgB9F,mBAAmBiB,eAAeK,IAAIyE,CAAC;QAC7D,IAAID,eAAe;YACjB,MAAM,IAAIvG,SAASuG;QACrB;QAEA,oFAAoF;QACpF,IAAIE,gBAA0B,EAAE;QAChC,IAAI/C,OAAO;YACT,MAAMgD,cACJ9E,WAAW,SAASU,QAAQK,MAAM,CAACgE,YAAY,GAC3CrE,QAAQK,MAAM,CAACgE,YAAY,CAACD,WAAW,GACvC3C;YAEN0C,gBAAgBnG,iBAAiB;gBAC/BoC;gBACAuC;gBACA1D;gBACAK;gBACA8E;YACF;YAEA,IAAIxF,OAAO;gBACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;oBACvB0F,aAAaH,cAAc3C,MAAM;oBACjC+C,KAAK;gBACP;YACF;QACF;QAEA,gFAAgF;QAChF,IAAIC,aAAuB,EAAE;QAC7B,IAAIC,mBAAmB;QAEvB,MAAMC,UAAU,IAAIC;QACpB,IAAIC,eAAe;QACnB,IAAIC,mBAAmBvC;QACvB,IAAIwC,UAAU;QACd,MAAMC,UACJ,OAAOrD,uBAAuB,WAAWA,qBAAqBsD,OAAOC,iBAAiB;QAExF,MAAMC,SAAS,IAAIvH,SAAS;YAC1B,MAAMwH;gBACJ,MAAMC,YAAYzD,KAAKQ,GAAG,CAAC,GAAG4C,UAAUD;gBAExC,IAAIM,cAAc,GAAG;oBACnB,IAAI,CAAChE,OAAO;wBACV,6EAA6E;wBAC7E,IAAI,CAACiE,IAAI,CAACX,QAAQY,MAAM,CAACV,eAAe,OAAO;oBACjD;oBACA,IAAI,CAACS,IAAI,CAAC;oBACV;gBACF;gBAEA,MAAME,SAAS,MAAMvF,QAAQO,IAAI,CAAC;oBAChC,GAAGgC,QAAQ;oBACX/C,MAAMqF;oBACN1F,OAAOwC,KAAKC,GAAG,CAAClD,WAAW0G;gBAC7B;gBAEA,IAAIxG,OAAO;oBACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CACtB,CAAC,gBAAgB,EAAEiG,iBAAiB,MAAM,EAAEU,OAAOC,IAAI,CAAChE,MAAM,CAAC,KAAK,CAAC;gBAEzE;gBAEA,IAAI+D,OAAOC,IAAI,CAAChE,MAAM,KAAK,GAAG;oBAC5B,oCAAoC;oBACpC,IAAI,CAACJ,OAAO;wBACV,wEAAwE;wBACxE,IAAI,CAACiE,IAAI,CAACX,QAAQY,MAAM,CAACV,eAAe,OAAO;oBACjD;oBACA,IAAI,CAACS,IAAI,CAAC;oBACV;gBACF;gBAEA,IAAIjE,OAAO;oBACT,wBAAwB;oBACxB,MAAMqE,YAAYF,OAAOC,IAAI,CAACzC,GAAG,CAAC,CAACa,MACjCZ,kBACEnF,cAAc;4BACZ+F;4BACA3E;4BACAwD;wBACF;oBAIJ,8EAA8E;oBAC9E,IAAI,CAACgC,kBAAkB;wBACrB,MAAMiB,cAAwB,EAAE;wBAChC,MAAMC,WAAW,IAAIC;wBACrB,KAAK,MAAM3C,OAAOwC,UAAW;4BAC3B,KAAK,MAAMtC,OAAOE,OAAOwC,IAAI,CAAC5C,KAAM;gCAClC,IAAI,CAAC0C,SAASG,GAAG,CAAC3C,MAAM;oCACtBwC,SAASI,GAAG,CAAC5C;oCACbuC,YAAYL,IAAI,CAAClC;gCACnB;4BACF;wBACF;wBACA,oDAAoD;wBACpDqB,aAAavG,aAAakG,eAAeuB;wBACzCjB,mBAAmB;wBAEnB,IAAI7F,OAAO;4BACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;gCACvBoH,kBAAkBN,YAAYlE,MAAM;gCACpCyE,mBAAmBzB,WAAWhD,MAAM;gCACpC+C,KAAK;4BACP;wBACF;oBACF;oBAEA,MAAM2B,aAAaT,UAAU1C,GAAG,CAAC,CAACE;wBAChC,MAAMkD,UAAmC,CAAC;wBAC1C,KAAK,MAAMC,OAAO5B,WAAY;4BAC5B2B,OAAO,CAACC,IAAI,GAAGnD,GAAG,CAACmD,IAAI,IAAI;wBAC7B;wBACA,OAAOD;oBACT;oBAEA,MAAME,YAAY5I,UAAUyI,YAAY;wBACtCI,KAAK1B;wBACL2B,QAAQ3B;wBACR4B,SAAShC;oBACX;oBAEA,IAAI,CAACa,IAAI,CAACX,QAAQY,MAAM,CAACe;gBAC3B,OAAO;oBACL,yBAAyB;oBACzB,MAAMZ,YAAYF,OAAOC,IAAI,CAACzC,GAAG,CAAC,CAACa,MAAQD,mBAAmBC;oBAE9D,uDAAuD;oBACvD,MAAM6C,YAAYhB,UAAU1C,GAAG,CAAC,CAACE,MAAQyD,KAAKjJ,SAAS,CAACwF,MAAM0D,IAAI,CAAC;oBAEnE,IAAI/B,cAAc;wBAChB,IAAI,CAACS,IAAI,CAACX,QAAQY,MAAM,CAAC,MAAMmB;oBACjC,OAAO;wBACL,IAAI,CAACpB,IAAI,CAACX,QAAQY,MAAM,CAAC,MAAMmB;oBACjC;gBACF;gBAEA3B,WAAWS,OAAOC,IAAI,CAAChE,MAAM;gBAC7BoD,eAAe;gBACfC,oBAAoB;gBAEpB,IAAI,CAACU,OAAOqB,WAAW,IAAI9B,WAAWC,SAAS;oBAC7C,IAAInG,OAAO;wBACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;oBAC3B;oBACA,IAAI,CAACwC,OAAO;wBACV,IAAI,CAACiE,IAAI,CAACX,QAAQY,MAAM,CAAC;oBAC3B;oBACA,IAAI,CAACD,IAAI,CAAC,OAAM,iBAAiB;gBACnC;YACF;QACF;QAEA,OAAO,IAAIwB,SAASlJ,SAASmJ,KAAK,CAAC5B,SAA2B;YAC5D6B,SAAS;gBACP,uBAAuB,CAAC,sBAAsB,EAAEvI,KAAK,CAAC,CAAC;gBACvD,gBAAgB4C,QAAQ,aAAa;YACvC;QACF;IACF;IAEA,sCAAsC;IACtC,IAAIxC,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;IAC3B;IAEA,gCAAgC;IAChC,MAAMoI,YAAY5I,2BAA2B;QAAEM;QAAWE;IAAM;IAEhE,qCAAqC;IACrC,MAAMqI,eAAe,CAACrD,MACpBxC,QACI4B,kBACEnF,cAAc;YACZ+F;YACA3E;YACAwD;QACF,MAEFkB,mBAAmBC;IAEzB,oEAAoE;IACpE,IAAIsD,eAAe;QACjBV,SAAS,EAAE;QACXhB,MAAM,EAAE;QACR2B,cAAc;IAChB;IAEA,IAAI,CAACrF,cAAc;QACjBoF,eAAe,MAAMF,UAAUI,aAAa,CAAC;YAC3CzI;YACA4D,UAAUA;YACVrD;YACA6F,SACE,OAAOrD,uBAAuB,WAAWA,qBAAqBsD,OAAOC,iBAAiB;YACxFxF;YACA4H,WAAW/E;YACX2E;QACF;IACF;IAEA,MAAM,EAAET,SAASd,WAAW,EAAEF,MAAM8B,IAAI,EAAE,GAAGJ;IAC7C,MAAMK,aAAuB,EAAE;IAE/B,uBAAuB;IACvB,IAAInG,OAAO;QACT,mDAAmD;QACnD,MAAMgD,cACJ9E,WAAW,SAASU,QAAQK,MAAM,CAACgE,YAAY,GAC3CrE,QAAQK,MAAM,CAACgE,YAAY,CAACD,WAAW,GACvC3C;QAEN,MAAM0C,gBAAgBnG,iBAAiB;YACrCoC;YACAuC;YACA1D;YACAK;YACA8E;QACF;QAEA,oDAAoD;QACpD,uFAAuF;QACvF,MAAMoD,eAAevJ,aAAakG,eAAeuB;QAEjD,MAAMQ,aAAaoB,KAAKvE,GAAG,CAAC,CAACE;YAC3B,MAAMkD,UAAmC,CAAC;YAC1C,KAAK,MAAMC,OAAOoB,aAAc;gBAC9BrB,OAAO,CAACC,IAAI,GAAGnD,GAAG,CAACmD,IAAI,IAAI;YAC7B;YACA,OAAOD;QACT;QAEA,+CAA+C;QAC/CoB,WAAWlC,IAAI,CACb5H,UAAUyI,YAAY;YACpBI,KAAK;YACLC,QAAQ;YACRC,SAASgB;QACX;IAEJ,OAAO;QACL,cAAc;QACdD,WAAWlC,IAAI,CAACiC,KAAKvE,GAAG,CAAC,CAACa,MAAQ8C,KAAKjJ,SAAS,CAACmG,MAAM+C,IAAI,CAAC;IAC9D;IAEA,mDAAmD;IACnD,uCAAuC;IACvC,yGAAyG;IACzG,MAAMc,UAAUvI,WAAW,SAAS,CAAC,CAAC,EAAEqI,WAAWZ,IAAI,CAAC,KAAK,CAAC,CAAC,GAAGY,WAAWZ,IAAI,CAAC;IAClF,MAAMe,SAASC,OAAOC,IAAI,CAACH,QAAQjG,MAAM,GAAG,IAAIiG,UAAU;IAC1D,IAAI7I,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC,GAAGM,OAAO,yBAAyB,CAAC;IAC/D;IAEA,IAAI,CAACX,IAAI;QACP,IAAIK,OAAO;YACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;QAC3B;QACAa,IAAIoI,IAAI,GAAG;YACTrJ;YACAsJ,MAAMJ;YACNK,UAAU3G,QAAQ,aAAa;YAC/B4G,MAAMN,OAAOlG,MAAM;QACrB;IACF,OAAO;QACL,IAAI5C,OAAO;YACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC,CAAC,kCAAkC,EAAEL,IAAI;QACpE;QACA,MAAMkB,IAAIO,OAAO,CAACiI,MAAM,CAAC;YACvB1J;YACAoC,YAAY3B;YACZ8I,MAAM,CAAC;YACPD,MAAM;gBACJrJ;gBACAsJ,MAAMJ;gBACNK,UAAU3G,QAAQ,aAAa;gBAC/B4G,MAAMN,OAAOlG,MAAM;YACrB;YACA,uGAAuG;YACvGZ,gBAAgB;YAChBnB;QACF;IACF;IACA,IAAIb,OAAO;QACTa,IAAIO,OAAO,CAACC,MAAM,CAACrB,KAAK,CAAC;IAC3B;AACF,EAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"handlePreview.d.ts","sourceRoot":"","sources":["../../src/export/handlePreview.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAkB,cAAc,EAAS,MAAM,SAAS,CAAA;
|
|
1
|
+
{"version":3,"file":"handlePreview.d.ts","sourceRoot":"","sources":["../../src/export/handlePreview.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAkB,cAAc,EAAS,MAAM,SAAS,CAAA;AAsBpE,eAAO,MAAM,aAAa,QAAe,cAAc,KAAG,OAAO,CAAC,QAAQ,CAsPzE,CAAA"}
|
|
@@ -1,11 +1,10 @@
|
|
|
1
1
|
import { addDataAndFileToRequest } from 'payload';
|
|
2
2
|
import { getObjectDotNotation } from 'payload/shared';
|
|
3
3
|
import { DEFAULT_PREVIEW_LIMIT, MAX_PREVIEW_LIMIT, MIN_PREVIEW_LIMIT, MIN_PREVIEW_PAGE } from '../constants.js';
|
|
4
|
-
import { collectTimezoneCompanionFields } from '../utilities/collectTimezoneCompanionFields.js';
|
|
5
4
|
import { flattenObject } from '../utilities/flattenObject.js';
|
|
6
5
|
import { getExportFieldFunctions } from '../utilities/getExportFieldFunctions.js';
|
|
7
6
|
import { getFlattenedFieldKeys } from '../utilities/getFlattenedFieldKeys.js';
|
|
8
|
-
import { getSchemaColumns } from '../utilities/getSchemaColumns.js';
|
|
7
|
+
import { getSchemaColumns, mergeColumns } from '../utilities/getSchemaColumns.js';
|
|
9
8
|
import { getSelect } from '../utilities/getSelect.js';
|
|
10
9
|
import { removeDisabledFields } from '../utilities/removeDisabledFields.js';
|
|
11
10
|
import { resolveLimit } from '../utilities/resolveLimit.js';
|
|
@@ -73,17 +72,16 @@ export const handlePreview = async (req)=>{
|
|
|
73
72
|
const localeCodes = locale === 'all' && req.payload.config.localization ? req.payload.config.localization.localeCodes : undefined;
|
|
74
73
|
// Get disabled fields configuration
|
|
75
74
|
const disabledFields = targetCollection.config.admin?.custom?.['plugin-import-export']?.disabledFields ?? [];
|
|
76
|
-
//
|
|
77
|
-
const
|
|
78
|
-
// Always compute columns for CSV (even if no docs) for consistent schema
|
|
79
|
-
const columns = isCSV ? getSchemaColumns({
|
|
75
|
+
// Compute schema-based columns for CSV (provides base ordering and handles empty exports)
|
|
76
|
+
const schemaColumns = isCSV ? getSchemaColumns({
|
|
80
77
|
collectionConfig: targetCollection.config,
|
|
81
78
|
disabledFields,
|
|
82
79
|
fields,
|
|
83
80
|
locale: locale ?? undefined,
|
|
84
|
-
localeCodes
|
|
85
|
-
timezoneCompanionFields
|
|
81
|
+
localeCodes
|
|
86
82
|
}) : undefined;
|
|
83
|
+
// columns will be finalized after data is available (merged with data-discovered columns)
|
|
84
|
+
let columns = schemaColumns;
|
|
87
85
|
// If we're beyond the effective limit (considering both user limit and maxLimit), return empty docs
|
|
88
86
|
if (exportTotalDocs > 0 && previewStartIndex >= exportTotalDocs) {
|
|
89
87
|
const response = {
|
|
@@ -132,20 +130,39 @@ export const handlePreview = async (req)=>{
|
|
|
132
130
|
const possibleKeys = getFlattenedFieldKeys(targetCollection.config.fields, '', {
|
|
133
131
|
localeCodes
|
|
134
132
|
});
|
|
135
|
-
|
|
136
|
-
|
|
133
|
+
// Flatten docs without padding yet. This preserves the exact keys produced by toCSV hooks,
|
|
134
|
+
// allowing mergeColumns to detect which schema columns were replaced with derived ones.
|
|
135
|
+
transformed = docs.map((doc)=>flattenObject({
|
|
137
136
|
doc,
|
|
138
137
|
fields,
|
|
139
|
-
timezoneCompanionFields,
|
|
140
138
|
toCSVFunctions
|
|
141
|
-
});
|
|
142
|
-
|
|
143
|
-
|
|
144
|
-
|
|
139
|
+
}));
|
|
140
|
+
// Merge schema columns with data-discovered columns (e.g., derived columns from toCSV)
|
|
141
|
+
// This ensures the preview headers match what the actual export will produce
|
|
142
|
+
if (schemaColumns && transformed.length > 0) {
|
|
143
|
+
const dataColumns = [];
|
|
144
|
+
const seenCols = new Set();
|
|
145
|
+
for (const row of transformed){
|
|
146
|
+
for (const key of Object.keys(row)){
|
|
147
|
+
if (!seenCols.has(key)) {
|
|
148
|
+
seenCols.add(key);
|
|
149
|
+
dataColumns.push(key);
|
|
150
|
+
}
|
|
145
151
|
}
|
|
146
152
|
}
|
|
147
|
-
|
|
148
|
-
}
|
|
153
|
+
columns = mergeColumns(schemaColumns, dataColumns);
|
|
154
|
+
}
|
|
155
|
+
// Pad rows with null for missing columns (uses merged columns, not raw schema)
|
|
156
|
+
if (!fields || fields.length === 0) {
|
|
157
|
+
const paddingKeys = columns ?? possibleKeys;
|
|
158
|
+
for (const row of transformed){
|
|
159
|
+
for (const key of paddingKeys){
|
|
160
|
+
if (!(key in row)) {
|
|
161
|
+
row[key] = null;
|
|
162
|
+
}
|
|
163
|
+
}
|
|
164
|
+
}
|
|
165
|
+
}
|
|
149
166
|
} else {
|
|
150
167
|
transformed = docs.map((doc)=>{
|
|
151
168
|
let output = {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/export/handlePreview.ts"],"sourcesContent":["import type { FlattenedField, PayloadRequest, Where } from 'payload'\n\nimport { addDataAndFileToRequest } from 'payload'\nimport { getObjectDotNotation } from 'payload/shared'\n\nimport type { ExportPreviewResponse } from '../types.js'\n\nimport {\n DEFAULT_PREVIEW_LIMIT,\n MAX_PREVIEW_LIMIT,\n MIN_PREVIEW_LIMIT,\n MIN_PREVIEW_PAGE,\n} from '../constants.js'\nimport { collectTimezoneCompanionFields } from '../utilities/collectTimezoneCompanionFields.js'\nimport { flattenObject } from '../utilities/flattenObject.js'\nimport { getExportFieldFunctions } from '../utilities/getExportFieldFunctions.js'\nimport { getFlattenedFieldKeys } from '../utilities/getFlattenedFieldKeys.js'\nimport { getSchemaColumns } from '../utilities/getSchemaColumns.js'\nimport { getSelect } from '../utilities/getSelect.js'\nimport { removeDisabledFields } from '../utilities/removeDisabledFields.js'\nimport { resolveLimit } from '../utilities/resolveLimit.js'\nimport { setNestedValue } from '../utilities/setNestedValue.js'\n\nexport const handlePreview = async (req: PayloadRequest): Promise<Response> => {\n await addDataAndFileToRequest(req)\n\n const {\n collectionSlug,\n draft: draftFromReq,\n fields,\n limit: exportLimit,\n locale,\n previewLimit: rawPreviewLimit = DEFAULT_PREVIEW_LIMIT,\n previewPage: rawPreviewPage = 1,\n sort,\n where: whereFromReq = {},\n } = req.data as {\n collectionSlug: string\n draft?: 'no' | 'yes'\n fields?: string[]\n format?: 'csv' | 'json'\n limit?: number\n locale?: string\n previewLimit?: number\n previewPage?: number\n sort?: any\n where?: any\n }\n\n // Validate and clamp pagination values to safe bounds\n const previewLimit = Math.max(MIN_PREVIEW_LIMIT, Math.min(rawPreviewLimit, MAX_PREVIEW_LIMIT))\n const previewPage = Math.max(MIN_PREVIEW_PAGE, rawPreviewPage)\n\n const targetCollection = req.payload.collections[collectionSlug]\n if (!targetCollection) {\n return Response.json(\n { error: `Collection with slug ${collectionSlug} not found` },\n { status: 400 },\n )\n }\n\n const pluginConfig = targetCollection.config.custom?.['plugin-import-export']\n const maxLimit = await resolveLimit({\n limit: pluginConfig?.exportLimit,\n req,\n })\n\n const select = Array.isArray(fields) && fields.length > 0 ? getSelect(fields) : undefined\n const draft = draftFromReq === 'yes'\n const collectionHasVersions = Boolean(targetCollection.config.versions)\n\n // Only filter by _status for versioned collections\n const publishedWhere: Where = collectionHasVersions ? { _status: { equals: 'published' } } : {}\n\n const where: Where = {\n and: [whereFromReq, draft ? {} : publishedWhere],\n }\n\n // Count total docs matching export criteria\n const countResult = await req.payload.count({\n collection: collectionSlug,\n overrideAccess: false,\n req,\n where,\n })\n\n const totalMatchingDocs = countResult.totalDocs\n\n // Calculate actual export count (respecting both export limit and max limit)\n let effectiveLimit = totalMatchingDocs\n\n // Apply user's export limit if provided\n if (exportLimit && exportLimit > 0) {\n effectiveLimit = Math.min(effectiveLimit, exportLimit)\n }\n\n // Apply max limit if configured\n if (typeof maxLimit === 'number' && maxLimit > 0) {\n effectiveLimit = Math.min(effectiveLimit, maxLimit)\n }\n\n const exportTotalDocs = effectiveLimit\n\n // Calculate preview pagination that respects export limit\n // Preview should only show docs that will actually be exported\n const previewStartIndex = (previewPage - 1) * previewLimit\n\n // Calculate pagination info based on export limit (not raw DB results)\n const previewTotalPages = exportTotalDocs === 0 ? 0 : Math.ceil(exportTotalDocs / previewLimit)\n\n const isCSV = req?.data?.format === 'csv'\n\n // Get locale codes for locale expansion when locale='all'\n const localeCodes =\n locale === 'all' && req.payload.config.localization\n ? req.payload.config.localization.localeCodes\n : undefined\n\n // Get disabled fields configuration\n const disabledFields =\n targetCollection.config.admin?.custom?.['plugin-import-export']?.disabledFields ?? []\n\n // Collect auto-generated timezone companion fields from schema\n const timezoneCompanionFields = collectTimezoneCompanionFields(\n targetCollection.config.flattenedFields,\n )\n\n // Always compute columns for CSV (even if no docs) for consistent schema\n const columns = isCSV\n ? getSchemaColumns({\n collectionConfig: targetCollection.config,\n disabledFields,\n fields,\n locale: locale ?? undefined,\n localeCodes,\n timezoneCompanionFields,\n })\n : undefined\n\n // If we're beyond the effective limit (considering both user limit and maxLimit), return empty docs\n if (exportTotalDocs > 0 && previewStartIndex >= exportTotalDocs) {\n const response: ExportPreviewResponse = {\n columns,\n docs: [],\n exportTotalDocs,\n hasNextPage: false,\n hasPrevPage: previewPage > 1,\n limit: previewLimit,\n maxLimit,\n page: previewPage,\n totalDocs: exportTotalDocs,\n totalPages: previewTotalPages,\n }\n return Response.json(response)\n }\n\n // Fetch preview page with full previewLimit to maintain consistent pagination offsets\n // We'll trim the results afterwards if needed to respect export limit\n const result = await req.payload.find({\n collection: collectionSlug,\n depth: 1,\n draft,\n limit: previewLimit,\n locale,\n overrideAccess: false,\n page: previewPage,\n req,\n select,\n sort,\n where,\n })\n\n // Trim docs to respect effective limit boundary (user limit clamped by maxLimit)\n let docs = result.docs\n if (exportTotalDocs > 0) {\n const remainingInExport = exportTotalDocs - previewStartIndex\n if (remainingInExport < docs.length) {\n docs = docs.slice(0, remainingInExport)\n }\n }\n\n // Transform docs based on format\n let transformed: Record<string, unknown>[]\n\n if (isCSV) {\n const toCSVFunctions = getExportFieldFunctions({\n fields: targetCollection.config.fields as FlattenedField[],\n })\n\n const possibleKeys = getFlattenedFieldKeys(\n targetCollection.config.fields as FlattenedField[],\n '',\n { localeCodes },\n )\n\n transformed = docs.map((doc) => {\n const row = flattenObject({\n doc,\n fields,\n timezoneCompanionFields,\n toCSVFunctions,\n })\n\n for (const key of possibleKeys) {\n if (!(key in row)) {\n row[key] = null\n }\n }\n\n return row\n })\n } else {\n transformed = docs.map((doc) => {\n let output: Record<string, unknown> = { ...doc }\n\n // Remove disabled fields first\n output = removeDisabledFields(output, disabledFields)\n\n // Then trim to selected fields only (if fields are provided)\n if (Array.isArray(fields) && fields.length > 0) {\n const trimmed: Record<string, unknown> = {}\n\n for (const key of fields) {\n const value = getObjectDotNotation(output, key)\n setNestedValue(trimmed, key, value ?? null)\n }\n\n output = trimmed\n }\n\n return output\n })\n }\n\n const hasNextPage = previewPage < previewTotalPages\n const hasPrevPage = previewPage > 1\n\n const response: ExportPreviewResponse = {\n columns,\n docs: transformed,\n exportTotalDocs,\n hasNextPage,\n hasPrevPage,\n limit: previewLimit,\n maxLimit,\n page: previewPage,\n totalDocs: exportTotalDocs,\n totalPages: previewTotalPages,\n }\n\n return Response.json(response)\n}\n"],"names":["addDataAndFileToRequest","getObjectDotNotation","DEFAULT_PREVIEW_LIMIT","MAX_PREVIEW_LIMIT","MIN_PREVIEW_LIMIT","MIN_PREVIEW_PAGE","collectTimezoneCompanionFields","flattenObject","getExportFieldFunctions","getFlattenedFieldKeys","getSchemaColumns","getSelect","removeDisabledFields","resolveLimit","setNestedValue","handlePreview","req","collectionSlug","draft","draftFromReq","fields","limit","exportLimit","locale","previewLimit","rawPreviewLimit","previewPage","rawPreviewPage","sort","where","whereFromReq","data","Math","max","min","targetCollection","payload","collections","Response","json","error","status","pluginConfig","config","custom","maxLimit","select","Array","isArray","length","undefined","collectionHasVersions","Boolean","versions","publishedWhere","_status","equals","and","countResult","count","collection","overrideAccess","totalMatchingDocs","totalDocs","effectiveLimit","exportTotalDocs","previewStartIndex","previewTotalPages","ceil","isCSV","format","localeCodes","localization","disabledFields","admin","timezoneCompanionFields","flattenedFields","columns","collectionConfig","response","docs","hasNextPage","hasPrevPage","page","totalPages","result","find","depth","remainingInExport","slice","transformed","toCSVFunctions","possibleKeys","map","doc","row","key","output","trimmed","value"],"mappings":"AAEA,SAASA,uBAAuB,QAAQ,UAAS;AACjD,SAASC,oBAAoB,QAAQ,iBAAgB;AAIrD,SACEC,qBAAqB,EACrBC,iBAAiB,EACjBC,iBAAiB,EACjBC,gBAAgB,QACX,kBAAiB;AACxB,SAASC,8BAA8B,QAAQ,iDAAgD;AAC/F,SAASC,aAAa,QAAQ,gCAA+B;AAC7D,SAASC,uBAAuB,QAAQ,0CAAyC;AACjF,SAASC,qBAAqB,QAAQ,wCAAuC;AAC7E,SAASC,gBAAgB,QAAQ,mCAAkC;AACnE,SAASC,SAAS,QAAQ,4BAA2B;AACrD,SAASC,oBAAoB,QAAQ,uCAAsC;AAC3E,SAASC,YAAY,QAAQ,+BAA8B;AAC3D,SAASC,cAAc,QAAQ,iCAAgC;AAE/D,OAAO,MAAMC,gBAAgB,OAAOC;IAClC,MAAMhB,wBAAwBgB;IAE9B,MAAM,EACJC,cAAc,EACdC,OAAOC,YAAY,EACnBC,MAAM,EACNC,OAAOC,WAAW,EAClBC,MAAM,EACNC,cAAcC,kBAAkBvB,qBAAqB,EACrDwB,aAAaC,iBAAiB,CAAC,EAC/BC,IAAI,EACJC,OAAOC,eAAe,CAAC,CAAC,EACzB,GAAGd,IAAIe,IAAI;IAaZ,sDAAsD;IACtD,MAAMP,eAAeQ,KAAKC,GAAG,CAAC7B,mBAAmB4B,KAAKE,GAAG,CAACT,iBAAiBtB;IAC3E,MAAMuB,cAAcM,KAAKC,GAAG,CAAC5B,kBAAkBsB;IAE/C,MAAMQ,mBAAmBnB,IAAIoB,OAAO,CAACC,WAAW,CAACpB,eAAe;IAChE,IAAI,CAACkB,kBAAkB;QACrB,OAAOG,SAASC,IAAI,CAClB;YAAEC,OAAO,CAAC,qBAAqB,EAAEvB,eAAe,UAAU,CAAC;QAAC,GAC5D;YAAEwB,QAAQ;QAAI;IAElB;IAEA,MAAMC,eAAeP,iBAAiBQ,MAAM,CAACC,MAAM,EAAE,CAAC,uBAAuB;IAC7E,MAAMC,WAAW,MAAMhC,aAAa;QAClCQ,OAAOqB,cAAcpB;QACrBN;IACF;IAEA,MAAM8B,SAASC,MAAMC,OAAO,CAAC5B,WAAWA,OAAO6B,MAAM,GAAG,IAAItC,UAAUS,UAAU8B;IAChF,MAAMhC,QAAQC,iBAAiB;IAC/B,MAAMgC,wBAAwBC,QAAQjB,iBAAiBQ,MAAM,CAACU,QAAQ;IAEtE,mDAAmD;IACnD,MAAMC,iBAAwBH,wBAAwB;QAAEI,SAAS;YAAEC,QAAQ;QAAY;IAAE,IAAI,CAAC;IAE9F,MAAM3B,QAAe;QACnB4B,KAAK;YAAC3B;YAAcZ,QAAQ,CAAC,IAAIoC;SAAe;IAClD;IAEA,4CAA4C;IAC5C,MAAMI,cAAc,MAAM1C,IAAIoB,OAAO,CAACuB,KAAK,CAAC;QAC1CC,YAAY3C;QACZ4C,gBAAgB;QAChB7C;QACAa;IACF;IAEA,MAAMiC,oBAAoBJ,YAAYK,SAAS;IAE/C,6EAA6E;IAC7E,IAAIC,iBAAiBF;IAErB,wCAAwC;IACxC,IAAIxC,eAAeA,cAAc,GAAG;QAClC0C,iBAAiBhC,KAAKE,GAAG,CAAC8B,gBAAgB1C;IAC5C;IAEA,gCAAgC;IAChC,IAAI,OAAOuB,aAAa,YAAYA,WAAW,GAAG;QAChDmB,iBAAiBhC,KAAKE,GAAG,CAAC8B,gBAAgBnB;IAC5C;IAEA,MAAMoB,kBAAkBD;IAExB,0DAA0D;IAC1D,+DAA+D;IAC/D,MAAME,oBAAoB,AAACxC,CAAAA,cAAc,CAAA,IAAKF;IAE9C,uEAAuE;IACvE,MAAM2C,oBAAoBF,oBAAoB,IAAI,IAAIjC,KAAKoC,IAAI,CAACH,kBAAkBzC;IAElF,MAAM6C,QAAQrD,KAAKe,MAAMuC,WAAW;IAEpC,0DAA0D;IAC1D,MAAMC,cACJhD,WAAW,SAASP,IAAIoB,OAAO,CAACO,MAAM,CAAC6B,YAAY,GAC/CxD,IAAIoB,OAAO,CAACO,MAAM,CAAC6B,YAAY,CAACD,WAAW,GAC3CrB;IAEN,oCAAoC;IACpC,MAAMuB,iBACJtC,iBAAiBQ,MAAM,CAAC+B,KAAK,EAAE9B,QAAQ,CAAC,uBAAuB,EAAE6B,kBAAkB,EAAE;IAEvF,+DAA+D;IAC/D,MAAME,0BAA0BrE,+BAC9B6B,iBAAiBQ,MAAM,CAACiC,eAAe;IAGzC,yEAAyE;IACzE,MAAMC,UAAUR,QACZ3D,iBAAiB;QACfoE,kBAAkB3C,iBAAiBQ,MAAM;QACzC8B;QACArD;QACAG,QAAQA,UAAU2B;QAClBqB;QACAI;IACF,KACAzB;IAEJ,oGAAoG;IACpG,IAAIe,kBAAkB,KAAKC,qBAAqBD,iBAAiB;QAC/D,MAAMc,WAAkC;YACtCF;YACAG,MAAM,EAAE;YACRf;YACAgB,aAAa;YACbC,aAAaxD,cAAc;YAC3BL,OAAOG;YACPqB;YACAsC,MAAMzD;YACNqC,WAAWE;YACXmB,YAAYjB;QACd;QACA,OAAO7B,SAASC,IAAI,CAACwC;IACvB;IAEA,sFAAsF;IACtF,sEAAsE;IACtE,MAAMM,SAAS,MAAMrE,IAAIoB,OAAO,CAACkD,IAAI,CAAC;QACpC1B,YAAY3C;QACZsE,OAAO;QACPrE;QACAG,OAAOG;QACPD;QACAsC,gBAAgB;QAChBsB,MAAMzD;QACNV;QACA8B;QACAlB;QACAC;IACF;IAEA,iFAAiF;IACjF,IAAImD,OAAOK,OAAOL,IAAI;IACtB,IAAIf,kBAAkB,GAAG;QACvB,MAAMuB,oBAAoBvB,kBAAkBC;QAC5C,IAAIsB,oBAAoBR,KAAK/B,MAAM,EAAE;YACnC+B,OAAOA,KAAKS,KAAK,CAAC,GAAGD;QACvB;IACF;IAEA,iCAAiC;IACjC,IAAIE;IAEJ,IAAIrB,OAAO;QACT,MAAMsB,iBAAiBnF,wBAAwB;YAC7CY,QAAQe,iBAAiBQ,MAAM,CAACvB,MAAM;QACxC;QAEA,MAAMwE,eAAenF,sBACnB0B,iBAAiBQ,MAAM,CAACvB,MAAM,EAC9B,IACA;YAAEmD;QAAY;QAGhBmB,cAAcV,KAAKa,GAAG,CAAC,CAACC;YACtB,MAAMC,MAAMxF,cAAc;gBACxBuF;gBACA1E;gBACAuD;gBACAgB;YACF;YAEA,KAAK,MAAMK,OAAOJ,aAAc;gBAC9B,IAAI,CAAEI,CAAAA,OAAOD,GAAE,GAAI;oBACjBA,GAAG,CAACC,IAAI,GAAG;gBACb;YACF;YAEA,OAAOD;QACT;IACF,OAAO;QACLL,cAAcV,KAAKa,GAAG,CAAC,CAACC;YACtB,IAAIG,SAAkC;gBAAE,GAAGH,GAAG;YAAC;YAE/C,+BAA+B;YAC/BG,SAASrF,qBAAqBqF,QAAQxB;YAEtC,6DAA6D;YAC7D,IAAI1B,MAAMC,OAAO,CAAC5B,WAAWA,OAAO6B,MAAM,GAAG,GAAG;gBAC9C,MAAMiD,UAAmC,CAAC;gBAE1C,KAAK,MAAMF,OAAO5E,OAAQ;oBACxB,MAAM+E,QAAQlG,qBAAqBgG,QAAQD;oBAC3ClF,eAAeoF,SAASF,KAAKG,SAAS;gBACxC;gBAEAF,SAASC;YACX;YAEA,OAAOD;QACT;IACF;IAEA,MAAMhB,cAAcvD,cAAcyC;IAClC,MAAMe,cAAcxD,cAAc;IAElC,MAAMqD,WAAkC;QACtCF;QACAG,MAAMU;QACNzB;QACAgB;QACAC;QACA7D,OAAOG;QACPqB;QACAsC,MAAMzD;QACNqC,WAAWE;QACXmB,YAAYjB;IACd;IAEA,OAAO7B,SAASC,IAAI,CAACwC;AACvB,EAAC"}
|
|
1
|
+
{"version":3,"sources":["../../src/export/handlePreview.ts"],"sourcesContent":["import type { FlattenedField, PayloadRequest, Where } from 'payload'\n\nimport { addDataAndFileToRequest } from 'payload'\nimport { getObjectDotNotation } from 'payload/shared'\n\nimport type { ExportPreviewResponse } from '../types.js'\n\nimport {\n DEFAULT_PREVIEW_LIMIT,\n MAX_PREVIEW_LIMIT,\n MIN_PREVIEW_LIMIT,\n MIN_PREVIEW_PAGE,\n} from '../constants.js'\nimport { flattenObject } from '../utilities/flattenObject.js'\nimport { getExportFieldFunctions } from '../utilities/getExportFieldFunctions.js'\nimport { getFlattenedFieldKeys } from '../utilities/getFlattenedFieldKeys.js'\nimport { getSchemaColumns, mergeColumns } from '../utilities/getSchemaColumns.js'\nimport { getSelect } from '../utilities/getSelect.js'\nimport { removeDisabledFields } from '../utilities/removeDisabledFields.js'\nimport { resolveLimit } from '../utilities/resolveLimit.js'\nimport { setNestedValue } from '../utilities/setNestedValue.js'\n\nexport const handlePreview = async (req: PayloadRequest): Promise<Response> => {\n await addDataAndFileToRequest(req)\n\n const {\n collectionSlug,\n draft: draftFromReq,\n fields,\n limit: exportLimit,\n locale,\n previewLimit: rawPreviewLimit = DEFAULT_PREVIEW_LIMIT,\n previewPage: rawPreviewPage = 1,\n sort,\n where: whereFromReq = {},\n } = req.data as {\n collectionSlug: string\n draft?: 'no' | 'yes'\n fields?: string[]\n format?: 'csv' | 'json'\n limit?: number\n locale?: string\n previewLimit?: number\n previewPage?: number\n sort?: any\n where?: any\n }\n\n // Validate and clamp pagination values to safe bounds\n const previewLimit = Math.max(MIN_PREVIEW_LIMIT, Math.min(rawPreviewLimit, MAX_PREVIEW_LIMIT))\n const previewPage = Math.max(MIN_PREVIEW_PAGE, rawPreviewPage)\n\n const targetCollection = req.payload.collections[collectionSlug]\n if (!targetCollection) {\n return Response.json(\n { error: `Collection with slug ${collectionSlug} not found` },\n { status: 400 },\n )\n }\n\n const pluginConfig = targetCollection.config.custom?.['plugin-import-export']\n const maxLimit = await resolveLimit({\n limit: pluginConfig?.exportLimit,\n req,\n })\n\n const select = Array.isArray(fields) && fields.length > 0 ? getSelect(fields) : undefined\n const draft = draftFromReq === 'yes'\n const collectionHasVersions = Boolean(targetCollection.config.versions)\n\n // Only filter by _status for versioned collections\n const publishedWhere: Where = collectionHasVersions ? { _status: { equals: 'published' } } : {}\n\n const where: Where = {\n and: [whereFromReq, draft ? {} : publishedWhere],\n }\n\n // Count total docs matching export criteria\n const countResult = await req.payload.count({\n collection: collectionSlug,\n overrideAccess: false,\n req,\n where,\n })\n\n const totalMatchingDocs = countResult.totalDocs\n\n // Calculate actual export count (respecting both export limit and max limit)\n let effectiveLimit = totalMatchingDocs\n\n // Apply user's export limit if provided\n if (exportLimit && exportLimit > 0) {\n effectiveLimit = Math.min(effectiveLimit, exportLimit)\n }\n\n // Apply max limit if configured\n if (typeof maxLimit === 'number' && maxLimit > 0) {\n effectiveLimit = Math.min(effectiveLimit, maxLimit)\n }\n\n const exportTotalDocs = effectiveLimit\n\n // Calculate preview pagination that respects export limit\n // Preview should only show docs that will actually be exported\n const previewStartIndex = (previewPage - 1) * previewLimit\n\n // Calculate pagination info based on export limit (not raw DB results)\n const previewTotalPages = exportTotalDocs === 0 ? 0 : Math.ceil(exportTotalDocs / previewLimit)\n\n const isCSV = req?.data?.format === 'csv'\n\n // Get locale codes for locale expansion when locale='all'\n const localeCodes =\n locale === 'all' && req.payload.config.localization\n ? req.payload.config.localization.localeCodes\n : undefined\n\n // Get disabled fields configuration\n const disabledFields =\n targetCollection.config.admin?.custom?.['plugin-import-export']?.disabledFields ?? []\n\n // Compute schema-based columns for CSV (provides base ordering and handles empty exports)\n const schemaColumns = isCSV\n ? getSchemaColumns({\n collectionConfig: targetCollection.config,\n disabledFields,\n fields,\n locale: locale ?? undefined,\n localeCodes,\n })\n : undefined\n\n // columns will be finalized after data is available (merged with data-discovered columns)\n let columns = schemaColumns\n\n // If we're beyond the effective limit (considering both user limit and maxLimit), return empty docs\n if (exportTotalDocs > 0 && previewStartIndex >= exportTotalDocs) {\n const response: ExportPreviewResponse = {\n columns,\n docs: [],\n exportTotalDocs,\n hasNextPage: false,\n hasPrevPage: previewPage > 1,\n limit: previewLimit,\n maxLimit,\n page: previewPage,\n totalDocs: exportTotalDocs,\n totalPages: previewTotalPages,\n }\n return Response.json(response)\n }\n\n // Fetch preview page with full previewLimit to maintain consistent pagination offsets\n // We'll trim the results afterwards if needed to respect export limit\n const result = await req.payload.find({\n collection: collectionSlug,\n depth: 1,\n draft,\n limit: previewLimit,\n locale,\n overrideAccess: false,\n page: previewPage,\n req,\n select,\n sort,\n where,\n })\n\n // Trim docs to respect effective limit boundary (user limit clamped by maxLimit)\n let docs = result.docs\n if (exportTotalDocs > 0) {\n const remainingInExport = exportTotalDocs - previewStartIndex\n if (remainingInExport < docs.length) {\n docs = docs.slice(0, remainingInExport)\n }\n }\n\n // Transform docs based on format\n let transformed: Record<string, unknown>[]\n\n if (isCSV) {\n const toCSVFunctions = getExportFieldFunctions({\n fields: targetCollection.config.fields as FlattenedField[],\n })\n\n const possibleKeys = getFlattenedFieldKeys(\n targetCollection.config.fields as FlattenedField[],\n '',\n { localeCodes },\n )\n\n // Flatten docs without padding yet. This preserves the exact keys produced by toCSV hooks,\n // allowing mergeColumns to detect which schema columns were replaced with derived ones.\n transformed = docs.map((doc) =>\n flattenObject({\n doc,\n fields,\n toCSVFunctions,\n }),\n )\n\n // Merge schema columns with data-discovered columns (e.g., derived columns from toCSV)\n // This ensures the preview headers match what the actual export will produce\n if (schemaColumns && transformed.length > 0) {\n const dataColumns: string[] = []\n const seenCols = new Set<string>()\n for (const row of transformed) {\n for (const key of Object.keys(row)) {\n if (!seenCols.has(key)) {\n seenCols.add(key)\n dataColumns.push(key)\n }\n }\n }\n columns = mergeColumns(schemaColumns, dataColumns)\n }\n\n // Pad rows with null for missing columns (uses merged columns, not raw schema)\n if (!fields || fields.length === 0) {\n const paddingKeys = columns ?? possibleKeys\n for (const row of transformed) {\n for (const key of paddingKeys) {\n if (!(key in row)) {\n row[key] = null\n }\n }\n }\n }\n } else {\n transformed = docs.map((doc) => {\n let output: Record<string, unknown> = { ...doc }\n\n // Remove disabled fields first\n output = removeDisabledFields(output, disabledFields)\n\n // Then trim to selected fields only (if fields are provided)\n if (Array.isArray(fields) && fields.length > 0) {\n const trimmed: Record<string, unknown> = {}\n\n for (const key of fields) {\n const value = getObjectDotNotation(output, key)\n setNestedValue(trimmed, key, value ?? null)\n }\n\n output = trimmed\n }\n\n return output\n })\n }\n\n const hasNextPage = previewPage < previewTotalPages\n const hasPrevPage = previewPage > 1\n\n const response: ExportPreviewResponse = {\n columns,\n docs: transformed,\n exportTotalDocs,\n hasNextPage,\n hasPrevPage,\n limit: previewLimit,\n maxLimit,\n page: previewPage,\n totalDocs: exportTotalDocs,\n totalPages: previewTotalPages,\n }\n\n return Response.json(response)\n}\n"],"names":["addDataAndFileToRequest","getObjectDotNotation","DEFAULT_PREVIEW_LIMIT","MAX_PREVIEW_LIMIT","MIN_PREVIEW_LIMIT","MIN_PREVIEW_PAGE","flattenObject","getExportFieldFunctions","getFlattenedFieldKeys","getSchemaColumns","mergeColumns","getSelect","removeDisabledFields","resolveLimit","setNestedValue","handlePreview","req","collectionSlug","draft","draftFromReq","fields","limit","exportLimit","locale","previewLimit","rawPreviewLimit","previewPage","rawPreviewPage","sort","where","whereFromReq","data","Math","max","min","targetCollection","payload","collections","Response","json","error","status","pluginConfig","config","custom","maxLimit","select","Array","isArray","length","undefined","collectionHasVersions","Boolean","versions","publishedWhere","_status","equals","and","countResult","count","collection","overrideAccess","totalMatchingDocs","totalDocs","effectiveLimit","exportTotalDocs","previewStartIndex","previewTotalPages","ceil","isCSV","format","localeCodes","localization","disabledFields","admin","schemaColumns","collectionConfig","columns","response","docs","hasNextPage","hasPrevPage","page","totalPages","result","find","depth","remainingInExport","slice","transformed","toCSVFunctions","possibleKeys","map","doc","dataColumns","seenCols","Set","row","key","Object","keys","has","add","push","paddingKeys","output","trimmed","value"],"mappings":"AAEA,SAASA,uBAAuB,QAAQ,UAAS;AACjD,SAASC,oBAAoB,QAAQ,iBAAgB;AAIrD,SACEC,qBAAqB,EACrBC,iBAAiB,EACjBC,iBAAiB,EACjBC,gBAAgB,QACX,kBAAiB;AACxB,SAASC,aAAa,QAAQ,gCAA+B;AAC7D,SAASC,uBAAuB,QAAQ,0CAAyC;AACjF,SAASC,qBAAqB,QAAQ,wCAAuC;AAC7E,SAASC,gBAAgB,EAAEC,YAAY,QAAQ,mCAAkC;AACjF,SAASC,SAAS,QAAQ,4BAA2B;AACrD,SAASC,oBAAoB,QAAQ,uCAAsC;AAC3E,SAASC,YAAY,QAAQ,+BAA8B;AAC3D,SAASC,cAAc,QAAQ,iCAAgC;AAE/D,OAAO,MAAMC,gBAAgB,OAAOC;IAClC,MAAMhB,wBAAwBgB;IAE9B,MAAM,EACJC,cAAc,EACdC,OAAOC,YAAY,EACnBC,MAAM,EACNC,OAAOC,WAAW,EAClBC,MAAM,EACNC,cAAcC,kBAAkBvB,qBAAqB,EACrDwB,aAAaC,iBAAiB,CAAC,EAC/BC,IAAI,EACJC,OAAOC,eAAe,CAAC,CAAC,EACzB,GAAGd,IAAIe,IAAI;IAaZ,sDAAsD;IACtD,MAAMP,eAAeQ,KAAKC,GAAG,CAAC7B,mBAAmB4B,KAAKE,GAAG,CAACT,iBAAiBtB;IAC3E,MAAMuB,cAAcM,KAAKC,GAAG,CAAC5B,kBAAkBsB;IAE/C,MAAMQ,mBAAmBnB,IAAIoB,OAAO,CAACC,WAAW,CAACpB,eAAe;IAChE,IAAI,CAACkB,kBAAkB;QACrB,OAAOG,SAASC,IAAI,CAClB;YAAEC,OAAO,CAAC,qBAAqB,EAAEvB,eAAe,UAAU,CAAC;QAAC,GAC5D;YAAEwB,QAAQ;QAAI;IAElB;IAEA,MAAMC,eAAeP,iBAAiBQ,MAAM,CAACC,MAAM,EAAE,CAAC,uBAAuB;IAC7E,MAAMC,WAAW,MAAMhC,aAAa;QAClCQ,OAAOqB,cAAcpB;QACrBN;IACF;IAEA,MAAM8B,SAASC,MAAMC,OAAO,CAAC5B,WAAWA,OAAO6B,MAAM,GAAG,IAAItC,UAAUS,UAAU8B;IAChF,MAAMhC,QAAQC,iBAAiB;IAC/B,MAAMgC,wBAAwBC,QAAQjB,iBAAiBQ,MAAM,CAACU,QAAQ;IAEtE,mDAAmD;IACnD,MAAMC,iBAAwBH,wBAAwB;QAAEI,SAAS;YAAEC,QAAQ;QAAY;IAAE,IAAI,CAAC;IAE9F,MAAM3B,QAAe;QACnB4B,KAAK;YAAC3B;YAAcZ,QAAQ,CAAC,IAAIoC;SAAe;IAClD;IAEA,4CAA4C;IAC5C,MAAMI,cAAc,MAAM1C,IAAIoB,OAAO,CAACuB,KAAK,CAAC;QAC1CC,YAAY3C;QACZ4C,gBAAgB;QAChB7C;QACAa;IACF;IAEA,MAAMiC,oBAAoBJ,YAAYK,SAAS;IAE/C,6EAA6E;IAC7E,IAAIC,iBAAiBF;IAErB,wCAAwC;IACxC,IAAIxC,eAAeA,cAAc,GAAG;QAClC0C,iBAAiBhC,KAAKE,GAAG,CAAC8B,gBAAgB1C;IAC5C;IAEA,gCAAgC;IAChC,IAAI,OAAOuB,aAAa,YAAYA,WAAW,GAAG;QAChDmB,iBAAiBhC,KAAKE,GAAG,CAAC8B,gBAAgBnB;IAC5C;IAEA,MAAMoB,kBAAkBD;IAExB,0DAA0D;IAC1D,+DAA+D;IAC/D,MAAME,oBAAoB,AAACxC,CAAAA,cAAc,CAAA,IAAKF;IAE9C,uEAAuE;IACvE,MAAM2C,oBAAoBF,oBAAoB,IAAI,IAAIjC,KAAKoC,IAAI,CAACH,kBAAkBzC;IAElF,MAAM6C,QAAQrD,KAAKe,MAAMuC,WAAW;IAEpC,0DAA0D;IAC1D,MAAMC,cACJhD,WAAW,SAASP,IAAIoB,OAAO,CAACO,MAAM,CAAC6B,YAAY,GAC/CxD,IAAIoB,OAAO,CAACO,MAAM,CAAC6B,YAAY,CAACD,WAAW,GAC3CrB;IAEN,oCAAoC;IACpC,MAAMuB,iBACJtC,iBAAiBQ,MAAM,CAAC+B,KAAK,EAAE9B,QAAQ,CAAC,uBAAuB,EAAE6B,kBAAkB,EAAE;IAEvF,0FAA0F;IAC1F,MAAME,gBAAgBN,QAClB5D,iBAAiB;QACfmE,kBAAkBzC,iBAAiBQ,MAAM;QACzC8B;QACArD;QACAG,QAAQA,UAAU2B;QAClBqB;IACF,KACArB;IAEJ,0FAA0F;IAC1F,IAAI2B,UAAUF;IAEd,oGAAoG;IACpG,IAAIV,kBAAkB,KAAKC,qBAAqBD,iBAAiB;QAC/D,MAAMa,WAAkC;YACtCD;YACAE,MAAM,EAAE;YACRd;YACAe,aAAa;YACbC,aAAavD,cAAc;YAC3BL,OAAOG;YACPqB;YACAqC,MAAMxD;YACNqC,WAAWE;YACXkB,YAAYhB;QACd;QACA,OAAO7B,SAASC,IAAI,CAACuC;IACvB;IAEA,sFAAsF;IACtF,sEAAsE;IACtE,MAAMM,SAAS,MAAMpE,IAAIoB,OAAO,CAACiD,IAAI,CAAC;QACpCzB,YAAY3C;QACZqE,OAAO;QACPpE;QACAG,OAAOG;QACPD;QACAsC,gBAAgB;QAChBqB,MAAMxD;QACNV;QACA8B;QACAlB;QACAC;IACF;IAEA,iFAAiF;IACjF,IAAIkD,OAAOK,OAAOL,IAAI;IACtB,IAAId,kBAAkB,GAAG;QACvB,MAAMsB,oBAAoBtB,kBAAkBC;QAC5C,IAAIqB,oBAAoBR,KAAK9B,MAAM,EAAE;YACnC8B,OAAOA,KAAKS,KAAK,CAAC,GAAGD;QACvB;IACF;IAEA,iCAAiC;IACjC,IAAIE;IAEJ,IAAIpB,OAAO;QACT,MAAMqB,iBAAiBnF,wBAAwB;YAC7Ca,QAAQe,iBAAiBQ,MAAM,CAACvB,MAAM;QACxC;QAEA,MAAMuE,eAAenF,sBACnB2B,iBAAiBQ,MAAM,CAACvB,MAAM,EAC9B,IACA;YAAEmD;QAAY;QAGhB,2FAA2F;QAC3F,wFAAwF;QACxFkB,cAAcV,KAAKa,GAAG,CAAC,CAACC,MACtBvF,cAAc;gBACZuF;gBACAzE;gBACAsE;YACF;QAGF,uFAAuF;QACvF,6EAA6E;QAC7E,IAAIf,iBAAiBc,YAAYxC,MAAM,GAAG,GAAG;YAC3C,MAAM6C,cAAwB,EAAE;YAChC,MAAMC,WAAW,IAAIC;YACrB,KAAK,MAAMC,OAAOR,YAAa;gBAC7B,KAAK,MAAMS,OAAOC,OAAOC,IAAI,CAACH,KAAM;oBAClC,IAAI,CAACF,SAASM,GAAG,CAACH,MAAM;wBACtBH,SAASO,GAAG,CAACJ;wBACbJ,YAAYS,IAAI,CAACL;oBACnB;gBACF;YACF;YACArB,UAAUnE,aAAaiE,eAAemB;QACxC;QAEA,+EAA+E;QAC/E,IAAI,CAAC1E,UAAUA,OAAO6B,MAAM,KAAK,GAAG;YAClC,MAAMuD,cAAc3B,WAAWc;YAC/B,KAAK,MAAMM,OAAOR,YAAa;gBAC7B,KAAK,MAAMS,OAAOM,YAAa;oBAC7B,IAAI,CAAEN,CAAAA,OAAOD,GAAE,GAAI;wBACjBA,GAAG,CAACC,IAAI,GAAG;oBACb;gBACF;YACF;QACF;IACF,OAAO;QACLT,cAAcV,KAAKa,GAAG,CAAC,CAACC;YACtB,IAAIY,SAAkC;gBAAE,GAAGZ,GAAG;YAAC;YAE/C,+BAA+B;YAC/BY,SAAS7F,qBAAqB6F,QAAQhC;YAEtC,6DAA6D;YAC7D,IAAI1B,MAAMC,OAAO,CAAC5B,WAAWA,OAAO6B,MAAM,GAAG,GAAG;gBAC9C,MAAMyD,UAAmC,CAAC;gBAE1C,KAAK,MAAMR,OAAO9E,OAAQ;oBACxB,MAAMuF,QAAQ1G,qBAAqBwG,QAAQP;oBAC3CpF,eAAe4F,SAASR,KAAKS,SAAS;gBACxC;gBAEAF,SAASC;YACX;YAEA,OAAOD;QACT;IACF;IAEA,MAAMzB,cAActD,cAAcyC;IAClC,MAAMc,cAAcvD,cAAc;IAElC,MAAMoD,WAAkC;QACtCD;QACAE,MAAMU;QACNxB;QACAe;QACAC;QACA5D,OAAOG;QACPqB;QACAqC,MAAMxD;QACNqC,WAAWE;QACXkB,YAAYhB;IACd;IAEA,OAAO7B,SAASC,IAAI,CAACuC;AACvB,EAAC"}
|
|
@@ -0,0 +1,204 @@
|
|
|
1
|
+
import { describe, expect, it } from 'vitest';
|
|
2
|
+
import { filterToSelectedFields } from './getSchemaColumns.js';
|
|
3
|
+
describe('filterToSelectedFields', ()=>{
|
|
4
|
+
describe('fields with their own column', ()=>{
|
|
5
|
+
it('should include only the exact column for a date with timezones field', ()=>{
|
|
6
|
+
const columns = [
|
|
7
|
+
'id',
|
|
8
|
+
'title',
|
|
9
|
+
'dateWithTimezone',
|
|
10
|
+
'dateWithTimezone_tz',
|
|
11
|
+
'createdAt'
|
|
12
|
+
];
|
|
13
|
+
const result = filterToSelectedFields(columns, [
|
|
14
|
+
'title'
|
|
15
|
+
]);
|
|
16
|
+
expect(result).toEqual([
|
|
17
|
+
'title'
|
|
18
|
+
]);
|
|
19
|
+
});
|
|
20
|
+
it('should not include timezone companion when only date field is selected', ()=>{
|
|
21
|
+
const columns = [
|
|
22
|
+
'id',
|
|
23
|
+
'dateWithTimezone',
|
|
24
|
+
'dateWithTimezone_tz',
|
|
25
|
+
'createdAt'
|
|
26
|
+
];
|
|
27
|
+
const result = filterToSelectedFields(columns, [
|
|
28
|
+
'dateWithTimezone'
|
|
29
|
+
]);
|
|
30
|
+
expect(result).toEqual([
|
|
31
|
+
'dateWithTimezone'
|
|
32
|
+
]);
|
|
33
|
+
expect(result).not.toContain('dateWithTimezone_tz');
|
|
34
|
+
});
|
|
35
|
+
it('should include both date and timezone companion when both are explicitly selected', ()=>{
|
|
36
|
+
const columns = [
|
|
37
|
+
'id',
|
|
38
|
+
'dateWithTimezone',
|
|
39
|
+
'dateWithTimezone_tz',
|
|
40
|
+
'createdAt'
|
|
41
|
+
];
|
|
42
|
+
const result = filterToSelectedFields(columns, [
|
|
43
|
+
'dateWithTimezone',
|
|
44
|
+
'dateWithTimezone_tz'
|
|
45
|
+
]);
|
|
46
|
+
expect(result).toEqual([
|
|
47
|
+
'dateWithTimezone',
|
|
48
|
+
'dateWithTimezone_tz'
|
|
49
|
+
]);
|
|
50
|
+
});
|
|
51
|
+
it('should preserve user-specified ordering', ()=>{
|
|
52
|
+
const columns = [
|
|
53
|
+
'id',
|
|
54
|
+
'title',
|
|
55
|
+
'email',
|
|
56
|
+
'createdAt'
|
|
57
|
+
];
|
|
58
|
+
const result = filterToSelectedFields(columns, [
|
|
59
|
+
'email',
|
|
60
|
+
'title',
|
|
61
|
+
'id'
|
|
62
|
+
]);
|
|
63
|
+
expect(result).toEqual([
|
|
64
|
+
'email',
|
|
65
|
+
'title',
|
|
66
|
+
'id'
|
|
67
|
+
]);
|
|
68
|
+
});
|
|
69
|
+
});
|
|
70
|
+
describe('fields without their own column (groups, arrays, blocks)', ()=>{
|
|
71
|
+
it('should expand a group to include its nested fields', ()=>{
|
|
72
|
+
const columns = [
|
|
73
|
+
'id',
|
|
74
|
+
'group_name',
|
|
75
|
+
'group_age',
|
|
76
|
+
'createdAt'
|
|
77
|
+
];
|
|
78
|
+
const result = filterToSelectedFields(columns, [
|
|
79
|
+
'group'
|
|
80
|
+
]);
|
|
81
|
+
expect(result).toEqual([
|
|
82
|
+
'group_name',
|
|
83
|
+
'group_age'
|
|
84
|
+
]);
|
|
85
|
+
});
|
|
86
|
+
it('should expand an array to include its nested fields', ()=>{
|
|
87
|
+
const columns = [
|
|
88
|
+
'id',
|
|
89
|
+
'items_0_title',
|
|
90
|
+
'items_0_value',
|
|
91
|
+
'createdAt'
|
|
92
|
+
];
|
|
93
|
+
const result = filterToSelectedFields(columns, [
|
|
94
|
+
'items'
|
|
95
|
+
]);
|
|
96
|
+
expect(result).toEqual([
|
|
97
|
+
'items_0_title',
|
|
98
|
+
'items_0_value'
|
|
99
|
+
]);
|
|
100
|
+
});
|
|
101
|
+
it('should expand blocks to include their nested fields', ()=>{
|
|
102
|
+
const columns = [
|
|
103
|
+
'id',
|
|
104
|
+
'blocks_0_hero_blockType',
|
|
105
|
+
'blocks_0_hero_title',
|
|
106
|
+
'createdAt'
|
|
107
|
+
];
|
|
108
|
+
const result = filterToSelectedFields(columns, [
|
|
109
|
+
'blocks'
|
|
110
|
+
]);
|
|
111
|
+
expect(result).toEqual([
|
|
112
|
+
'blocks_0_hero_blockType',
|
|
113
|
+
'blocks_0_hero_title'
|
|
114
|
+
]);
|
|
115
|
+
});
|
|
116
|
+
it('should expand a hasMany select to include indexed columns', ()=>{
|
|
117
|
+
const columns = [
|
|
118
|
+
'id',
|
|
119
|
+
'colors_0',
|
|
120
|
+
'createdAt'
|
|
121
|
+
];
|
|
122
|
+
const result = filterToSelectedFields(columns, [
|
|
123
|
+
'colors'
|
|
124
|
+
]);
|
|
125
|
+
expect(result).toEqual([
|
|
126
|
+
'colors_0'
|
|
127
|
+
]);
|
|
128
|
+
});
|
|
129
|
+
it('should expand a polymorphic relationship to include relationTo and id columns', ()=>{
|
|
130
|
+
const columns = [
|
|
131
|
+
'id',
|
|
132
|
+
'relField_relationTo',
|
|
133
|
+
'relField_id',
|
|
134
|
+
'createdAt'
|
|
135
|
+
];
|
|
136
|
+
const result = filterToSelectedFields(columns, [
|
|
137
|
+
'relField'
|
|
138
|
+
]);
|
|
139
|
+
expect(result).toEqual([
|
|
140
|
+
'relField_relationTo',
|
|
141
|
+
'relField_id'
|
|
142
|
+
]);
|
|
143
|
+
});
|
|
144
|
+
});
|
|
145
|
+
describe('nested field selections with dot notation', ()=>{
|
|
146
|
+
it('should resolve a specific nested field within a group', ()=>{
|
|
147
|
+
const columns = [
|
|
148
|
+
'id',
|
|
149
|
+
'group_name',
|
|
150
|
+
'group_age',
|
|
151
|
+
'createdAt'
|
|
152
|
+
];
|
|
153
|
+
const result = filterToSelectedFields(columns, [
|
|
154
|
+
'group.name'
|
|
155
|
+
]);
|
|
156
|
+
expect(result).toEqual([
|
|
157
|
+
'group_name'
|
|
158
|
+
]);
|
|
159
|
+
expect(result).not.toContain('group_age');
|
|
160
|
+
});
|
|
161
|
+
});
|
|
162
|
+
describe('mixed selections', ()=>{
|
|
163
|
+
it('should handle groups and date fields together without sibling leakage', ()=>{
|
|
164
|
+
const columns = [
|
|
165
|
+
'id',
|
|
166
|
+
'title',
|
|
167
|
+
'group_name',
|
|
168
|
+
'group_age',
|
|
169
|
+
'dateWithTimezone',
|
|
170
|
+
'dateWithTimezone_tz',
|
|
171
|
+
'createdAt'
|
|
172
|
+
];
|
|
173
|
+
const result = filterToSelectedFields(columns, [
|
|
174
|
+
'id',
|
|
175
|
+
'group',
|
|
176
|
+
'dateWithTimezone'
|
|
177
|
+
]);
|
|
178
|
+
expect(result).toEqual([
|
|
179
|
+
'id',
|
|
180
|
+
'group_name',
|
|
181
|
+
'group_age',
|
|
182
|
+
'dateWithTimezone'
|
|
183
|
+
]);
|
|
184
|
+
expect(result).not.toContain('dateWithTimezone_tz');
|
|
185
|
+
});
|
|
186
|
+
it('should not produce duplicates when a group and its nested field are both selected', ()=>{
|
|
187
|
+
const columns = [
|
|
188
|
+
'id',
|
|
189
|
+
'group_name',
|
|
190
|
+
'group_age'
|
|
191
|
+
];
|
|
192
|
+
const result = filterToSelectedFields(columns, [
|
|
193
|
+
'group',
|
|
194
|
+
'group.name'
|
|
195
|
+
]);
|
|
196
|
+
const uniqueResult = [
|
|
197
|
+
...new Set(result)
|
|
198
|
+
];
|
|
199
|
+
expect(result).toEqual(uniqueResult);
|
|
200
|
+
});
|
|
201
|
+
});
|
|
202
|
+
});
|
|
203
|
+
|
|
204
|
+
//# sourceMappingURL=filterToSelectedFields.spec.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../../src/utilities/filterToSelectedFields.spec.ts"],"sourcesContent":["import { describe, expect, it } from 'vitest'\n\nimport { filterToSelectedFields } from './getSchemaColumns.js'\n\ndescribe('filterToSelectedFields', () => {\n describe('fields with their own column', () => {\n it('should include only the exact column for a date with timezones field', () => {\n const columns = ['id', 'title', 'dateWithTimezone', 'dateWithTimezone_tz', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['title'])\n\n expect(result).toEqual(['title'])\n })\n\n it('should not include timezone companion when only date field is selected', () => {\n const columns = ['id', 'dateWithTimezone', 'dateWithTimezone_tz', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['dateWithTimezone'])\n\n expect(result).toEqual(['dateWithTimezone'])\n expect(result).not.toContain('dateWithTimezone_tz')\n })\n\n it('should include both date and timezone companion when both are explicitly selected', () => {\n const columns = ['id', 'dateWithTimezone', 'dateWithTimezone_tz', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['dateWithTimezone', 'dateWithTimezone_tz'])\n\n expect(result).toEqual(['dateWithTimezone', 'dateWithTimezone_tz'])\n })\n\n it('should preserve user-specified ordering', () => {\n const columns = ['id', 'title', 'email', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['email', 'title', 'id'])\n\n expect(result).toEqual(['email', 'title', 'id'])\n })\n })\n\n describe('fields without their own column (groups, arrays, blocks)', () => {\n it('should expand a group to include its nested fields', () => {\n const columns = ['id', 'group_name', 'group_age', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['group'])\n\n expect(result).toEqual(['group_name', 'group_age'])\n })\n\n it('should expand an array to include its nested fields', () => {\n const columns = ['id', 'items_0_title', 'items_0_value', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['items'])\n\n expect(result).toEqual(['items_0_title', 'items_0_value'])\n })\n\n it('should expand blocks to include their nested fields', () => {\n const columns = ['id', 'blocks_0_hero_blockType', 'blocks_0_hero_title', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['blocks'])\n\n expect(result).toEqual(['blocks_0_hero_blockType', 'blocks_0_hero_title'])\n })\n\n it('should expand a hasMany select to include indexed columns', () => {\n const columns = ['id', 'colors_0', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['colors'])\n\n expect(result).toEqual(['colors_0'])\n })\n\n it('should expand a polymorphic relationship to include relationTo and id columns', () => {\n const columns = ['id', 'relField_relationTo', 'relField_id', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['relField'])\n\n expect(result).toEqual(['relField_relationTo', 'relField_id'])\n })\n })\n\n describe('nested field selections with dot notation', () => {\n it('should resolve a specific nested field within a group', () => {\n const columns = ['id', 'group_name', 'group_age', 'createdAt']\n\n const result = filterToSelectedFields(columns, ['group.name'])\n\n expect(result).toEqual(['group_name'])\n expect(result).not.toContain('group_age')\n })\n })\n\n describe('mixed selections', () => {\n it('should handle groups and date fields together without sibling leakage', () => {\n const columns = [\n 'id',\n 'title',\n 'group_name',\n 'group_age',\n 'dateWithTimezone',\n 'dateWithTimezone_tz',\n 'createdAt',\n ]\n\n const result = filterToSelectedFields(columns, ['id', 'group', 'dateWithTimezone'])\n\n expect(result).toEqual(['id', 'group_name', 'group_age', 'dateWithTimezone'])\n expect(result).not.toContain('dateWithTimezone_tz')\n })\n\n it('should not produce duplicates when a group and its nested field are both selected', () => {\n const columns = ['id', 'group_name', 'group_age']\n\n const result = filterToSelectedFields(columns, ['group', 'group.name'])\n\n const uniqueResult = [...new Set(result)]\n\n expect(result).toEqual(uniqueResult)\n })\n })\n})\n"],"names":["describe","expect","it","filterToSelectedFields","columns","result","toEqual","not","toContain","uniqueResult","Set"],"mappings":"AAAA,SAASA,QAAQ,EAAEC,MAAM,EAAEC,EAAE,QAAQ,SAAQ;AAE7C,SAASC,sBAAsB,QAAQ,wBAAuB;AAE9DH,SAAS,0BAA0B;IACjCA,SAAS,gCAAgC;QACvCE,GAAG,wEAAwE;YACzE,MAAME,UAAU;gBAAC;gBAAM;gBAAS;gBAAoB;gBAAuB;aAAY;YAEvF,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAQ;YAExDH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;aAAQ;QAClC;QAEAJ,GAAG,0EAA0E;YAC3E,MAAME,UAAU;gBAAC;gBAAM;gBAAoB;gBAAuB;aAAY;YAE9E,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAmB;YAEnEH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;aAAmB;YAC3CL,OAAOI,QAAQE,GAAG,CAACC,SAAS,CAAC;QAC/B;QAEAN,GAAG,qFAAqF;YACtF,MAAME,UAAU;gBAAC;gBAAM;gBAAoB;gBAAuB;aAAY;YAE9E,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;gBAAoB;aAAsB;YAE1FH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAAoB;aAAsB;QACpE;QAEAJ,GAAG,2CAA2C;YAC5C,MAAME,UAAU;gBAAC;gBAAM;gBAAS;gBAAS;aAAY;YAErD,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;gBAAS;gBAAS;aAAK;YAEvEH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAAS;gBAAS;aAAK;QACjD;IACF;IAEAN,SAAS,4DAA4D;QACnEE,GAAG,sDAAsD;YACvD,MAAME,UAAU;gBAAC;gBAAM;gBAAc;gBAAa;aAAY;YAE9D,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAQ;YAExDH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAAc;aAAY;QACpD;QAEAJ,GAAG,uDAAuD;YACxD,MAAME,UAAU;gBAAC;gBAAM;gBAAiB;gBAAiB;aAAY;YAErE,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAQ;YAExDH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAAiB;aAAgB;QAC3D;QAEAJ,GAAG,uDAAuD;YACxD,MAAME,UAAU;gBAAC;gBAAM;gBAA2B;gBAAuB;aAAY;YAErF,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAS;YAEzDH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAA2B;aAAsB;QAC3E;QAEAJ,GAAG,6DAA6D;YAC9D,MAAME,UAAU;gBAAC;gBAAM;gBAAY;aAAY;YAE/C,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAS;YAEzDH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;aAAW;QACrC;QAEAJ,GAAG,iFAAiF;YAClF,MAAME,UAAU;gBAAC;gBAAM;gBAAuB;gBAAe;aAAY;YAEzE,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAW;YAE3DH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAAuB;aAAc;QAC/D;IACF;IAEAN,SAAS,6CAA6C;QACpDE,GAAG,yDAAyD;YAC1D,MAAME,UAAU;gBAAC;gBAAM;gBAAc;gBAAa;aAAY;YAE9D,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;aAAa;YAE7DH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;aAAa;YACrCL,OAAOI,QAAQE,GAAG,CAACC,SAAS,CAAC;QAC/B;IACF;IAEAR,SAAS,oBAAoB;QAC3BE,GAAG,yEAAyE;YAC1E,MAAME,UAAU;gBACd;gBACA;gBACA;gBACA;gBACA;gBACA;gBACA;aACD;YAED,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;gBAAM;gBAAS;aAAmB;YAElFH,OAAOI,QAAQC,OAAO,CAAC;gBAAC;gBAAM;gBAAc;gBAAa;aAAmB;YAC5EL,OAAOI,QAAQE,GAAG,CAACC,SAAS,CAAC;QAC/B;QAEAN,GAAG,qFAAqF;YACtF,MAAME,UAAU;gBAAC;gBAAM;gBAAc;aAAY;YAEjD,MAAMC,SAASF,uBAAuBC,SAAS;gBAAC;gBAAS;aAAa;YAEtE,MAAMK,eAAe;mBAAI,IAAIC,IAAIL;aAAQ;YAEzCJ,OAAOI,QAAQC,OAAO,CAACG;QACzB;IACF;AACF"}
|
|
@@ -4,14 +4,8 @@ type Args = {
|
|
|
4
4
|
doc: Document;
|
|
5
5
|
fields?: string[];
|
|
6
6
|
prefix?: string;
|
|
7
|
-
/**
|
|
8
|
-
* Set of auto-generated timezone companion field names (from collectTimezoneCompanionFields).
|
|
9
|
-
* These fields are excluded unless explicitly selected.
|
|
10
|
-
* If not provided, no timezone filtering is applied.
|
|
11
|
-
*/
|
|
12
|
-
timezoneCompanionFields?: Set<string>;
|
|
13
7
|
toCSVFunctions: Record<string, ToCSVFunction>;
|
|
14
8
|
};
|
|
15
|
-
export declare const flattenObject: ({ doc, fields, prefix,
|
|
9
|
+
export declare const flattenObject: ({ doc, fields, prefix, toCSVFunctions, }: Args) => Record<string, unknown>;
|
|
16
10
|
export {};
|
|
17
11
|
//# sourceMappingURL=flattenObject.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"flattenObject.d.ts","sourceRoot":"","sources":["../../src/utilities/flattenObject.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,QAAQ,EAAE,MAAM,SAAS,CAAA;AAEvC,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,aAAa,CAAA;AAIhD,KAAK,IAAI,GAAG;IACV,GAAG,EAAE,QAAQ,CAAA;IACb,MAAM,CAAC,EAAE,MAAM,EAAE,CAAA;IACjB,MAAM,CAAC,EAAE,MAAM,CAAA;IACf
|
|
1
|
+
{"version":3,"file":"flattenObject.d.ts","sourceRoot":"","sources":["../../src/utilities/flattenObject.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,QAAQ,EAAE,MAAM,SAAS,CAAA;AAEvC,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,aAAa,CAAA;AAIhD,KAAK,IAAI,GAAG;IACV,GAAG,EAAE,QAAQ,CAAA;IACb,MAAM,CAAC,EAAE,MAAM,EAAE,CAAA;IACjB,MAAM,CAAC,EAAE,MAAM,CAAA;IACf,cAAc,EAAE,MAAM,CAAC,MAAM,EAAE,aAAa,CAAC,CAAA;CAC9C,CAAA;AAED,eAAO,MAAM,aAAa,6CAKvB,IAAI,KAAG,MAAM,CAAC,MAAM,EAAE,OAAO,CAoK/B,CAAA"}
|