@rpcbase/server 0.538.0 → 0.540.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/email-DK8uUU4X.js +8045 -0
- package/dist/email-DK8uUU4X.js.map +1 -0
- package/dist/handler--FFBJMl6.js +153 -0
- package/dist/handler--FFBJMl6.js.map +1 -0
- package/dist/handler-0rPClEv4.js +663 -0
- package/dist/handler-0rPClEv4.js.map +1 -0
- package/dist/handler-COnCnprN.js +203 -0
- package/dist/handler-COnCnprN.js.map +1 -0
- package/dist/handler-ClQF4MOn.js +931 -0
- package/dist/handler-ClQF4MOn.js.map +1 -0
- package/dist/index.js +4988 -4830
- package/dist/index.js.map +1 -1
- package/dist/notifications.js +199 -134
- package/dist/notifications.js.map +1 -1
- package/dist/queryExecutor-Bol_iR8f.js +453 -0
- package/dist/queryExecutor-Bol_iR8f.js.map +1 -0
- package/dist/render_resend_false-MiC__Smr.js +6 -0
- package/dist/render_resend_false-MiC__Smr.js.map +1 -0
- package/dist/rts/index.d.ts +0 -1
- package/dist/rts/index.d.ts.map +1 -1
- package/dist/rts/index.js +1003 -842
- package/dist/rts/index.js.map +1 -1
- package/dist/schemas-Cjdjgehl.js +4225 -0
- package/dist/schemas-Cjdjgehl.js.map +1 -0
- package/dist/shared-nE84Or5W.js +111 -0
- package/dist/shared-nE84Or5W.js.map +1 -0
- package/dist/ssrMiddleware.d.ts +1 -1
- package/dist/uploads.js +99 -84
- package/dist/uploads.js.map +1 -1
- package/package.json +9 -9
- package/dist/email-H8nTAGxe.js +0 -12449
- package/dist/email-H8nTAGxe.js.map +0 -1
- package/dist/handler-BBzEodA0.js +0 -182
- package/dist/handler-BBzEodA0.js.map +0 -1
- package/dist/handler-BLwgdQv-.js +0 -544
- package/dist/handler-BLwgdQv-.js.map +0 -1
- package/dist/handler-CZD5p1Jv.js +0 -28
- package/dist/handler-CZD5p1Jv.js.map +0 -1
- package/dist/handler-Cq6MsoD4.js +0 -124
- package/dist/handler-Cq6MsoD4.js.map +0 -1
- package/dist/handler-DBtnVvP2.js +0 -756
- package/dist/handler-DBtnVvP2.js.map +0 -1
- package/dist/queryExecutor-JadZcQSQ.js +0 -318
- package/dist/queryExecutor-JadZcQSQ.js.map +0 -1
- package/dist/render_resend-DQANggpW.js +0 -7
- package/dist/render_resend-DQANggpW.js.map +0 -1
- package/dist/rts/api/cleanup/handler.d.ts +0 -9
- package/dist/rts/api/cleanup/handler.d.ts.map +0 -1
- package/dist/rts/api/cleanup/index.d.ts +0 -11
- package/dist/rts/api/cleanup/index.d.ts.map +0 -1
- package/dist/schemas-BR3K5Luo.js +0 -3824
- package/dist/schemas-BR3K5Luo.js.map +0 -1
- package/dist/shared-DhZ_rDdo.js +0 -87
- package/dist/shared-DhZ_rDdo.js.map +0 -1
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"handler-DBtnVvP2.js","names":["JSDOM","createDOMPurify","UploadFileProcessor","MAX_SVG_BYTES","window","DOMPurify","normalizeForSniff","raw","replace","trimStart","looksLikeSvgText","text","normalized","startsWith","test","looksLikeSvg","sniff","Buffer","toString","sanitizeSvg","svg","sanitize","USE_PROFILES","svgFilters","sanitizeSvgProcessor","id","maxBytes","match","process","data","mimeType","length","Error","svgText","sanitized","trim","sanitizedBuffer","from","sanitizeSvgProcessor","UploadFileProcessorContext","filename","clientMimeType","totalSize","sniff","Buffer","UploadFileProcessorResult","data","mimeType","UploadFileProcessor","id","maxBytes","match","ctx","process","Promise","uploadProcessors","Object","freeze","getMaxUploadProcessorBytes","reduce","max","processor","Math","selectUploadProcessors","filter","applyUploadProcessors","Omit","applied","currentData","currentMimeType","processorCtx","length","Error","result","trim","push","ApiHandler","getTenantFilesystemDb","models","GridFSBucket","Model","Uploads","enqueueUploadPostProcessors","applyUploadProcessors","getMaxUploadProcessorBytes","selectUploadProcessors","SessionUser","UploadChunkDoc","UploadSessionDoc","buildUploadsAbility","ensureUploadIndexes","getBucketName","getModelCtx","getUploadSessionAccessQuery","getTenantId","waitForStreamFinished","stream","NodeJS","WritableStream","Promise","resolve","reject","once","writeToStream","chunk","Buffer","ok","write","onDrain","cleanup","onError","error","off","on","abortUploadStream","abort","destroy","completeUpload","Record","CompleteResponsePayload","_payload","ctx","tenantId","res","status","uploadId","String","req","params","trim","ability","modelCtx","UploadSession","UploadChunk","all","get","can","existing","findOne","$and","_id","lean","fileId","locked","findOneAndUpdate","$set","$unset","returnDocument","fsDb","nativeDb","db","updateOne","bucketName","bucket","lockedUserId","userId","undefined","maxProcessorBytes","shouldBufferForProcessing","totalSize","declaredMimeType","mimeType","toLowerCase","declaredSvg","filename","endsWith","uploadStream","finalMimeType","inlineProcessors","finalMetadata","isPublic","ownerKeyHash","Error","cursor","find","sort","index","AsyncIterable","close","expectedIndex","chunks","bufferedBytes","pendingChunks","sniffParts","sniffBytes","chunkDoc","data","push","length","slice","subarray","Math","min","sniff","concat","processors","clientMimeType","openUploadStream","metadata","pending","chunksTotal","assembled","processed","processedMimeType","applied","processedSize","finished","end","id","catch","console","deleteMany","message","ApiHandler","models","Model","Uploads","SessionUser","UploadChunkDoc","UploadSessionDoc","buildUploadsAbility","getModelCtx","getUploadSessionAccessQuery","getTenantId","getStatus","Record","StatusResponsePayload","_payload","ctx","Promise","tenantId","res","status","ok","error","uploadId","String","req","params","trim","ability","modelCtx","UploadSession","UploadChunk","all","get","can","session","findOne","$and","_id","lean","receivedDocs","find","index","sort","received","Array","map","doc","filter","n","Number","isInteger","chunkSize","chunksTotal","fileId","z","InitRoute","ChunkRoute","StatusRoute","CompleteRoute","initRequestSchema","object","filename","string","min","mimeType","isPublic","boolean","optional","totalSize","number","int","InitRequestPayload","infer","initResponseSchema","ok","error","uploadId","uploadKey","chunkSize","chunksTotal","InitResponsePayload","statusResponseSchema","status","enum","received","array","fileId","StatusResponsePayload","completeResponseSchema","CompleteResponsePayload","randomBytes","ApiHandler","models","ObjectId","Model","Uploads","SessionUser","UploadChunkDoc","UploadSessionDoc","buildUploadsAbility","computeSha256Hex","ensureUploadIndexes","getChunkSizeBytes","getModelCtx","getSessionTtlMs","getTenantId","getUserId","initUpload","InitRequestPayload","InitResponsePayload","payload","ctx","Promise","tenantId","res","status","ok","error","userId","parsed","initRequestSchema","safeParse","success","chunkSize","filename","mimeType","totalSize","isPublic","data","chunksTotal","Math","ceil","ability","modelCtx","UploadSession","UploadChunk","all","get","uploadId","toString","now","Date","expiresAt","uploadKey","ownerKeyHash","Buffer","from","undefined","create","_id","createdAt","ApiHandler","models","Model","SessionUser","UploadChunkDoc","UploadSessionDoc","buildUploadsAbility","computeSha256Hex","ensureUploadIndexes","getModelCtx","getUploadSessionAccessQuery","getTenantId","normalizeSha256Hex","toBufferPayload","ChunkResponsePayload","ok","error","uploadChunk","Buffer","payload","ctx","Promise","tenantId","res","status","uploadId","String","req","params","trim","indexRaw","index","Number","isInteger","ability","modelCtx","UploadSession","UploadChunk","all","get","can","session","findOne","$and","_id","lean","chunksTotal","data","expectedSize","totalSize","chunkSize","length","checksumHeader","sha256","undefined","expectedSha256","updateOne","$set","size","expiresAt","$setOnInsert","createdAt","Date","upsert","rawBodyParser","limitBytes","maxClientBytesPerSecond","req","res","next","contentType","headers","String","includes","total","chunks","Buffer","done","paused","throttleTimeout","ReturnType","setTimeout","rateBytesPerSecond","cleanup","off","onData","onEnd","onError","onAborted","clearTimeout","finish","error","body","concat","chunk","buffer","isBuffer","from","length","destroy","status","json","ok","push","now","Date","clientKey","getClientKey","state","getClientRateState","waitMs","consumeRateBudget","pause","resume","err","Error","on","ClientRateState","tokens","lastRefillMs","lastSeenMs","MAX_BURST_SECONDS","STALE_CLIENT_MS","clientRateStates","Map","lastCleanupMs","rawClientIp","clientIp","trim","rawIp","ip","maybeCleanupStates","size","key","delete","capacity","existing","get","Math","min","set","bytes","elapsedMs","max","ceil","Api","completeUpload","getStatus","initUpload","uploadChunk","rawBodyParser","getChunkSizeBytes","getMaxClientUploadBytesPerSecond","getRawBodyLimitBytes","SessionUser","Uploads","api","chunkSizeBytes","use","InitRoute","limitBytes","maxClientBytesPerSecond","post","put","ChunkRoute","get","StatusRoute","CompleteRoute"],"sources":["../src/uploads/api/file-uploads/processors/sanitizeSvg.ts","../src/uploads/api/file-uploads/processors/index.ts","../src/uploads/api/file-uploads/handlers/completeUpload.ts","../src/uploads/api/file-uploads/handlers/getStatus.ts","../src/uploads/api/file-uploads/index.ts","../src/uploads/api/file-uploads/handlers/initUpload.ts","../src/uploads/api/file-uploads/handlers/uploadChunk.ts","../src/uploads/api/file-uploads/middleware/rawBodyParser.ts","../src/uploads/api/file-uploads/handler.ts"],"sourcesContent":["import { JSDOM } from \"jsdom\"\nimport createDOMPurify from \"dompurify\"\n\nimport type { UploadFileProcessor } from \"./index\"\n\n\nconst MAX_SVG_BYTES = 128 * 1024\n\nconst window = new JSDOM(\"\").window\nconst DOMPurify = createDOMPurify(window)\n\nconst normalizeForSniff = (raw: string): string => raw.replace(/^\\uFEFF/, \"\").trimStart()\n\nconst looksLikeSvgText = (text: string): boolean => {\n const normalized = normalizeForSniff(text)\n if (!normalized.startsWith(\"<\")) return false\n return /<svg(?:\\s|>)/i.test(normalized)\n}\n\nexport const looksLikeSvg = (sniff: Buffer): boolean => looksLikeSvgText(sniff.toString(\"utf8\"))\n\nexport const sanitizeSvg = (svg: string): string =>\n DOMPurify.sanitize(svg, {\n USE_PROFILES: { svg: true, svgFilters: true },\n })\n\nexport const sanitizeSvgProcessor: UploadFileProcessor = {\n id: \"sanitize-svg\",\n maxBytes: MAX_SVG_BYTES,\n match: ({ sniff }) => looksLikeSvg(sniff),\n process: (data): { data: Buffer; mimeType: string } => {\n if (data.length > MAX_SVG_BYTES) {\n throw new Error(\"svg_too_large\")\n }\n\n const svgText = data.toString(\"utf8\")\n if (!looksLikeSvgText(svgText)) {\n throw new Error(\"svg_invalid\")\n }\n\n const sanitized = sanitizeSvg(svgText)\n if (!sanitized.trim() || !looksLikeSvgText(sanitized)) {\n throw new Error(\"svg_sanitize_failed\")\n }\n\n const sanitizedBuffer = Buffer.from(sanitized, \"utf8\")\n if (sanitizedBuffer.length > MAX_SVG_BYTES) {\n throw new Error(\"svg_too_large\")\n }\n\n return { data: sanitizedBuffer, mimeType: \"image/svg+xml\" }\n },\n}\n\n","import { sanitizeSvgProcessor } from \"./sanitizeSvg\"\n\n\nexport type UploadFileProcessorContext = {\n filename: string\n clientMimeType: string\n totalSize: number\n sniff: Buffer\n}\n\nexport type UploadFileProcessorResult = {\n data: Buffer\n mimeType?: string\n}\n\nexport type UploadFileProcessor = {\n id: string\n maxBytes: number\n match: (ctx: UploadFileProcessorContext) => boolean\n process: (data: Buffer, ctx: UploadFileProcessorContext) => Promise<UploadFileProcessorResult> | UploadFileProcessorResult\n}\n\nexport const uploadProcessors = Object.freeze([sanitizeSvgProcessor] satisfies UploadFileProcessor[])\n\nexport const getMaxUploadProcessorBytes = (): number =>\n uploadProcessors.reduce((max, processor) => Math.max(max, processor.maxBytes), 0)\n\nexport const selectUploadProcessors = (ctx: UploadFileProcessorContext): UploadFileProcessor[] =>\n uploadProcessors.filter((processor) => processor.match(ctx))\n\nexport const applyUploadProcessors = async (\n data: Buffer,\n ctx: Omit<UploadFileProcessorContext, \"sniff\" | \"totalSize\">,\n): Promise<{ data: Buffer; mimeType: string; applied: string[] }> => {\n let currentData = data\n let currentMimeType = ctx.clientMimeType\n const applied: string[] = []\n\n for (const processor of uploadProcessors) {\n const processorCtx: UploadFileProcessorContext = {\n filename: ctx.filename,\n clientMimeType: currentMimeType,\n totalSize: currentData.length,\n sniff: currentData,\n }\n\n if (!processor.match(processorCtx)) continue\n\n if (currentData.length > processor.maxBytes) {\n throw new Error(\"processor_input_too_large\")\n }\n\n const result = await processor.process(currentData, processorCtx)\n currentData = result.data\n if (typeof result.mimeType === \"string\" && result.mimeType.trim()) {\n currentMimeType = result.mimeType.trim()\n }\n applied.push(processor.id)\n }\n\n return {\n data: currentData,\n mimeType: currentMimeType,\n applied,\n }\n}\n","import { ApiHandler } from \"@rpcbase/api\"\nimport { getTenantFilesystemDb, models } from \"@rpcbase/db\"\nimport { GridFSBucket } from \"mongodb\"\nimport type { Model } from \"mongoose\"\n\nimport * as Uploads from \"../index\"\nimport { enqueueUploadPostProcessors } from \"../postProcessors\"\nimport { applyUploadProcessors, getMaxUploadProcessorBytes, selectUploadProcessors } from \"../processors\"\nimport {\n type SessionUser,\n type UploadChunkDoc,\n type UploadSessionDoc,\n buildUploadsAbility,\n ensureUploadIndexes,\n getBucketName,\n getModelCtx,\n getUploadSessionAccessQuery,\n getTenantId,\n} from \"../shared\"\n\n\nconst waitForStreamFinished = async (stream: NodeJS.WritableStream): Promise<void> => new Promise((resolve, reject) => {\n stream.once(\"finish\", resolve)\n stream.once(\"error\", reject)\n})\n\nconst writeToStream = async (stream: NodeJS.WritableStream, chunk: Buffer): Promise<void> => {\n const ok = stream.write(chunk)\n if (ok) return\n await new Promise<void>((resolve, reject) => {\n const onDrain = () => {\n cleanup()\n resolve()\n }\n\n const onError = (error: unknown) => {\n cleanup()\n reject(error)\n }\n\n const cleanup = () => {\n stream.off(\"drain\", onDrain)\n stream.off(\"error\", onError)\n }\n\n stream.on(\"drain\", onDrain)\n stream.on(\"error\", onError)\n })\n}\n\nconst abortUploadStream = async (stream: unknown): Promise<void> => {\n if (!stream) return\n if (typeof (stream as { abort?: unknown }).abort === \"function\") {\n try {\n await (stream as { abort: () => Promise<void> | void }).abort()\n return\n } catch {\n //\n }\n }\n try {\n ;(stream as { destroy?: () => void }).destroy?.()\n } catch {\n //\n }\n}\n\nexport const completeUpload: ApiHandler<Record<string, never>, Uploads.CompleteResponsePayload, SessionUser> = async (\n _payload,\n ctx,\n): Promise<Uploads.CompleteResponsePayload> => {\n const tenantId = getTenantId(ctx)\n if (!tenantId) {\n ctx.res.status(400)\n return { ok: false, error: \"tenant_missing\" }\n }\n\n const uploadId = String(ctx.req.params?.uploadId ?? \"\").trim()\n if (!uploadId) {\n ctx.res.status(400)\n return { ok: false, error: \"invalid_upload_id\" }\n }\n\n const ability = buildUploadsAbility(ctx, tenantId)\n const modelCtx = getModelCtx(ctx, tenantId, ability)\n\n const [UploadSession, UploadChunk] = await Promise.all([\n models.get(\"RBUploadSession\", modelCtx) as Promise<Model<UploadSessionDoc>>,\n models.get(\"RBUploadChunk\", modelCtx) as Promise<Model<UploadChunkDoc>>,\n ])\n\n if (!ability.can(\"update\", \"RBUploadSession\")) {\n ctx.res.status(401)\n return { ok: false, error: \"unauthorized\" }\n }\n\n const existing = await UploadSession.findOne({ $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"read\")] }).lean()\n if (!existing) {\n ctx.res.status(404)\n return { ok: false, error: \"not_found\" }\n }\n\n if (existing.status === \"done\" && existing.fileId) {\n return { ok: true, fileId: existing.fileId }\n }\n\n const locked = await UploadSession.findOneAndUpdate(\n { $and: [{ _id: uploadId }, { status: \"uploading\" }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"assembling\" }, $unset: { error: \"\" } },\n { returnDocument: \"after\" },\n ).lean()\n\n if (!locked) {\n ctx.res.status(409)\n return { ok: false, error: \"not_uploading\" }\n }\n\n await ensureUploadIndexes(UploadSession, UploadChunk)\n\n const fsDb = await getTenantFilesystemDb(tenantId)\n const nativeDb = fsDb.db\n if (!nativeDb) {\n await UploadSession.updateOne(\n { $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"error\", error: \"filesystem_db_unavailable\" } },\n )\n ctx.res.status(500)\n return { ok: false, error: \"assembly_failed\" }\n }\n const bucketName = getBucketName()\n const bucket = new GridFSBucket(nativeDb, { bucketName })\n\n const lockedUserId = typeof locked.userId === \"string\" ? locked.userId : undefined\n const maxProcessorBytes = getMaxUploadProcessorBytes()\n const shouldBufferForProcessing = locked.totalSize <= maxProcessorBytes\n const declaredMimeType = locked.mimeType.trim().toLowerCase()\n const declaredSvg = declaredMimeType === \"image/svg+xml\" || locked.filename.trim().toLowerCase().endsWith(\".svg\")\n\n let uploadStream: NodeJS.WritableStream | null = null\n let finalMimeType = locked.mimeType\n let inlineProcessors: string[] = []\n let finalMetadata: Record<string, unknown> = {\n uploadId,\n tenantId,\n mimeType: locked.mimeType,\n totalSize: locked.totalSize,\n ...(typeof locked.isPublic === \"boolean\" ? { isPublic: locked.isPublic } : {}),\n ...(typeof locked.ownerKeyHash === \"string\" ? { ownerKeyHash: locked.ownerKeyHash } : {}),\n ...(lockedUserId ? { userId: lockedUserId } : {}),\n }\n\n try {\n if (!shouldBufferForProcessing && declaredSvg) {\n throw new Error(\"svg_too_large\")\n }\n\n const cursor = UploadChunk.find({ uploadId }).sort({ index: 1 }).cursor() as unknown as AsyncIterable<UploadChunkDoc> & {\n close: () => Promise<void>\n }\n\n let expectedIndex = 0\n const chunks: Buffer[] = []\n let bufferedBytes = 0\n\n const pendingChunks: Buffer[] = []\n const sniffParts: Buffer[] = []\n let sniffBytes = 0\n\n try {\n for await (const chunkDoc of cursor) {\n if (chunkDoc.index !== expectedIndex) {\n throw new Error(\"missing_chunks\")\n }\n\n const chunk = chunkDoc.data\n\n if (shouldBufferForProcessing) {\n chunks.push(chunk)\n bufferedBytes += chunk.length\n } else if (!uploadStream) {\n pendingChunks.push(chunk)\n\n if (sniffBytes < maxProcessorBytes) {\n const slice = chunk.subarray(0, Math.min(chunk.length, maxProcessorBytes - sniffBytes))\n if (slice.length) {\n sniffParts.push(slice)\n sniffBytes += slice.length\n }\n }\n\n if (sniffBytes >= maxProcessorBytes) {\n const sniff = Buffer.concat(sniffParts, sniffBytes)\n const processors = selectUploadProcessors({\n filename: locked.filename,\n clientMimeType: locked.mimeType,\n totalSize: locked.totalSize,\n sniff,\n })\n\n if (processors.length) {\n throw new Error(\"svg_too_large\")\n }\n\n finalMetadata = {\n uploadId,\n tenantId,\n mimeType: locked.mimeType,\n totalSize: locked.totalSize,\n ...(typeof locked.isPublic === \"boolean\" ? { isPublic: locked.isPublic } : {}),\n ...(typeof locked.ownerKeyHash === \"string\" ? { ownerKeyHash: locked.ownerKeyHash } : {}),\n ...(lockedUserId ? { userId: lockedUserId } : {}),\n }\n uploadStream = bucket.openUploadStream(locked.filename, {\n metadata: finalMetadata,\n })\n\n for (const pending of pendingChunks) {\n await writeToStream(uploadStream, pending)\n }\n pendingChunks.length = 0\n }\n } else {\n await writeToStream(uploadStream, chunk)\n }\n\n expectedIndex += 1\n }\n } finally {\n try {\n await cursor.close()\n } catch {\n //\n }\n }\n\n if (expectedIndex !== locked.chunksTotal) {\n throw new Error(\"missing_chunks\")\n }\n\n if (shouldBufferForProcessing) {\n const assembled = Buffer.concat(chunks, bufferedBytes)\n const { data: processed, mimeType: processedMimeType, applied } = await applyUploadProcessors(assembled, {\n filename: locked.filename,\n clientMimeType: locked.mimeType,\n })\n\n finalMimeType = processedMimeType\n inlineProcessors = applied\n finalMetadata = {\n uploadId,\n tenantId,\n mimeType: processedMimeType,\n totalSize: locked.totalSize,\n ...(applied.length ? { processors: applied, processedSize: processed.length } : {}),\n ...(typeof locked.isPublic === \"boolean\" ? { isPublic: locked.isPublic } : {}),\n ...(typeof locked.ownerKeyHash === \"string\" ? { ownerKeyHash: locked.ownerKeyHash } : {}),\n ...(lockedUserId ? { userId: lockedUserId } : {}),\n }\n uploadStream = bucket.openUploadStream(locked.filename, {\n metadata: finalMetadata,\n })\n\n const finished = waitForStreamFinished(uploadStream)\n uploadStream.end(processed)\n await finished\n } else {\n if (!uploadStream) {\n const sniff = Buffer.concat(sniffParts, sniffBytes)\n const processors = selectUploadProcessors({\n filename: locked.filename,\n clientMimeType: locked.mimeType,\n totalSize: locked.totalSize,\n sniff,\n })\n\n if (processors.length) {\n throw new Error(\"svg_too_large\")\n }\n\n finalMetadata = {\n uploadId,\n tenantId,\n mimeType: locked.mimeType,\n totalSize: locked.totalSize,\n ...(typeof locked.isPublic === \"boolean\" ? { isPublic: locked.isPublic } : {}),\n ...(typeof locked.ownerKeyHash === \"string\" ? { ownerKeyHash: locked.ownerKeyHash } : {}),\n ...(lockedUserId ? { userId: lockedUserId } : {}),\n }\n uploadStream = bucket.openUploadStream(locked.filename, {\n metadata: finalMetadata,\n })\n\n for (const pending of pendingChunks) {\n await writeToStream(uploadStream, pending)\n }\n pendingChunks.length = 0\n }\n\n const finished = waitForStreamFinished(uploadStream)\n uploadStream.end()\n await finished\n }\n\n const fileId = String((uploadStream as unknown as { id?: unknown }).id ?? \"\")\n if (!fileId) {\n throw new Error(\"missing_file_id\")\n }\n\n await UploadSession.updateOne(\n { $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"done\", fileId }, $unset: { error: \"\" } },\n )\n\n await enqueueUploadPostProcessors({\n tenantId,\n uploadId,\n fileId,\n filename: locked.filename,\n mimeType: finalMimeType,\n clientMimeType: locked.mimeType,\n totalSize: locked.totalSize,\n ...(typeof locked.isPublic === \"boolean\" ? { isPublic: locked.isPublic } : {}),\n ...(typeof locked.ownerKeyHash === \"string\" ? { ownerKeyHash: locked.ownerKeyHash } : {}),\n ...(lockedUserId ? { userId: lockedUserId } : {}),\n inlineProcessors,\n metadata: finalMetadata,\n }).catch((error) => {\n console.error(\"Upload post processor enqueue failed\", {\n tenantId,\n uploadId,\n fileId,\n error,\n })\n })\n\n try {\n await UploadChunk.deleteMany({ uploadId })\n } catch {\n //\n }\n\n return { ok: true, fileId }\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error)\n\n await abortUploadStream(uploadStream)\n\n if (message === \"missing_chunks\") {\n await UploadSession.updateOne(\n { $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"uploading\" } },\n )\n ctx.res.status(409)\n return { ok: false, error: \"missing_chunks\" }\n }\n\n if (message === \"svg_too_large\") {\n await UploadSession.updateOne(\n { $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"error\", error: message } },\n )\n ctx.res.status(413)\n return { ok: false, error: message }\n }\n\n if (message === \"svg_invalid\" || message === \"svg_sanitize_failed\") {\n await UploadSession.updateOne(\n { $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"error\", error: message } },\n )\n ctx.res.status(400)\n return { ok: false, error: message }\n }\n\n await UploadSession.updateOne(\n { $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] },\n { $set: { status: \"error\", error: message } },\n )\n\n ctx.res.status(500)\n return { ok: false, error: \"assembly_failed\" }\n }\n}\n","import { ApiHandler } from \"@rpcbase/api\"\nimport { models } from \"@rpcbase/db\"\nimport type { Model } from \"mongoose\"\n\nimport * as Uploads from \"../index\"\nimport {\n type SessionUser,\n type UploadChunkDoc,\n type UploadSessionDoc,\n buildUploadsAbility,\n getModelCtx,\n getUploadSessionAccessQuery,\n getTenantId,\n} from \"../shared\"\n\n\nexport const getStatus: ApiHandler<Record<string, never>, Uploads.StatusResponsePayload, SessionUser> = async (\n _payload,\n ctx,\n): Promise<Uploads.StatusResponsePayload> => {\n const tenantId = getTenantId(ctx)\n if (!tenantId) {\n ctx.res.status(400)\n return { ok: false, error: \"tenant_missing\" }\n }\n\n const uploadId = String(ctx.req.params?.uploadId ?? \"\").trim()\n if (!uploadId) {\n ctx.res.status(400)\n return { ok: false, error: \"invalid_upload_id\" }\n }\n\n const ability = buildUploadsAbility(ctx, tenantId)\n const modelCtx = getModelCtx(ctx, tenantId, ability)\n\n const [UploadSession, UploadChunk] = await Promise.all([\n models.get(\"RBUploadSession\", modelCtx) as Promise<Model<UploadSessionDoc>>,\n models.get(\"RBUploadChunk\", modelCtx) as Promise<Model<UploadChunkDoc>>,\n ])\n\n if (!ability.can(\"read\", \"RBUploadSession\")) {\n ctx.res.status(401)\n return { ok: false, error: \"unauthorized\" }\n }\n\n const session = await UploadSession.findOne({ $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"read\")] }).lean()\n if (!session) {\n ctx.res.status(404)\n return { ok: false, error: \"not_found\" }\n }\n\n const receivedDocs = await UploadChunk.find(\n { uploadId },\n { index: 1, _id: 0 },\n ).sort({ index: 1 }).lean()\n\n const received = (receivedDocs as unknown as Array<{ index?: unknown }>)\n .map((doc) => (typeof doc.index === \"number\" ? doc.index : -1))\n .filter((n) => Number.isInteger(n) && n >= 0)\n\n return {\n ok: true,\n status: session.status,\n chunkSize: session.chunkSize,\n chunksTotal: session.chunksTotal,\n received,\n ...(session.fileId ? { fileId: session.fileId } : {}),\n }\n}\n","import { z } from \"zod\"\n\n\nexport const InitRoute = \"/api/rb/file-uploads\"\nexport const ChunkRoute = \"/api/rb/file-uploads/:uploadId/chunks/:index\"\nexport const StatusRoute = \"/api/rb/file-uploads/:uploadId/status\"\nexport const CompleteRoute = \"/api/rb/file-uploads/:uploadId/complete\"\n\nexport const initRequestSchema = z.object({\n filename: z.string().min(1),\n mimeType: z.string().min(1),\n isPublic: z.boolean().optional(),\n totalSize: z.number().int().min(1),\n})\n\nexport type InitRequestPayload = z.infer<typeof initRequestSchema>\n\nexport const initResponseSchema = z.object({\n ok: z.boolean(),\n error: z.string().optional(),\n uploadId: z.string().optional(),\n uploadKey: z.string().optional(),\n chunkSize: z.number().int().optional(),\n chunksTotal: z.number().int().optional(),\n})\n\nexport type InitResponsePayload = z.infer<typeof initResponseSchema>\n\nexport const statusResponseSchema = z.object({\n ok: z.boolean(),\n error: z.string().optional(),\n status: z.enum([\"uploading\", \"assembling\", \"done\", \"error\"]).optional(),\n chunkSize: z.number().int().optional(),\n chunksTotal: z.number().int().optional(),\n received: z.array(z.number().int().min(0)).optional(),\n fileId: z.string().optional(),\n})\n\nexport type StatusResponsePayload = z.infer<typeof statusResponseSchema>\n\nexport const completeResponseSchema = z.object({\n ok: z.boolean(),\n error: z.string().optional(),\n fileId: z.string().optional(),\n})\n\nexport type CompleteResponsePayload = z.infer<typeof completeResponseSchema>\n","import { randomBytes } from \"node:crypto\"\n\nimport { ApiHandler } from \"@rpcbase/api\"\nimport { models } from \"@rpcbase/db\"\nimport { ObjectId } from \"mongodb\"\nimport type { Model } from \"mongoose\"\n\nimport * as Uploads from \"../index\"\nimport {\n type SessionUser,\n type UploadChunkDoc,\n type UploadSessionDoc,\n buildUploadsAbility,\n computeSha256Hex,\n ensureUploadIndexes,\n getChunkSizeBytes,\n getModelCtx,\n getSessionTtlMs,\n getTenantId,\n getUserId,\n} from \"../shared\"\n\n\nexport const initUpload: ApiHandler<Uploads.InitRequestPayload, Uploads.InitResponsePayload, SessionUser> = async (\n payload,\n ctx,\n): Promise<Uploads.InitResponsePayload> => {\n const tenantId = getTenantId(ctx)\n if (!tenantId) {\n ctx.res.status(400)\n return { ok: false, error: \"tenant_missing\" }\n }\n\n const userId = getUserId(ctx)\n\n const parsed = Uploads.initRequestSchema.safeParse(payload ?? {})\n if (!parsed.success) {\n ctx.res.status(400)\n return { ok: false, error: \"invalid_payload\" }\n }\n\n const chunkSize = getChunkSizeBytes()\n const { filename, mimeType, totalSize, isPublic } = parsed.data\n const chunksTotal = Math.ceil(totalSize / chunkSize)\n\n const ability = buildUploadsAbility(ctx, tenantId)\n const modelCtx = getModelCtx(ctx, tenantId, ability)\n\n const [UploadSession, UploadChunk] = await Promise.all([\n models.get(\"RBUploadSession\", modelCtx) as Promise<Model<UploadSessionDoc>>,\n models.get(\"RBUploadChunk\", modelCtx) as Promise<Model<UploadChunkDoc>>,\n ])\n\n await ensureUploadIndexes(UploadSession, UploadChunk)\n\n const uploadId = new ObjectId().toString()\n const now = Date.now()\n const expiresAt = new Date(now + getSessionTtlMs())\n\n const uploadKey = userId ? null : randomBytes(32).toString(\"base64url\")\n const ownerKeyHash = uploadKey ? computeSha256Hex(Buffer.from(uploadKey)) : undefined\n\n await UploadSession.create({\n _id: uploadId,\n ...(userId ? { userId } : {}),\n ...(ownerKeyHash ? { ownerKeyHash } : {}),\n filename,\n mimeType,\n ...(typeof isPublic === \"boolean\" ? { isPublic } : {}),\n totalSize,\n chunkSize,\n chunksTotal,\n status: \"uploading\",\n createdAt: new Date(now),\n expiresAt,\n })\n\n return {\n ok: true,\n uploadId,\n chunkSize,\n chunksTotal,\n ...(uploadKey ? { uploadKey } : {}),\n }\n}\n","import { ApiHandler } from \"@rpcbase/api\"\nimport { models } from \"@rpcbase/db\"\nimport type { Model } from \"mongoose\"\n\nimport {\n type SessionUser,\n type UploadChunkDoc,\n type UploadSessionDoc,\n buildUploadsAbility,\n computeSha256Hex,\n ensureUploadIndexes,\n getModelCtx,\n getUploadSessionAccessQuery,\n getTenantId,\n normalizeSha256Hex,\n toBufferPayload,\n} from \"../shared\"\n\n\ntype ChunkResponsePayload = {\n ok: boolean\n error?: string\n}\n\nexport const uploadChunk: ApiHandler<Buffer, ChunkResponsePayload, SessionUser> = async (\n payload,\n ctx,\n): Promise<ChunkResponsePayload> => {\n const tenantId = getTenantId(ctx)\n if (!tenantId) {\n ctx.res.status(400)\n return { ok: false, error: \"tenant_missing\" }\n }\n\n const uploadId = String(ctx.req.params?.uploadId ?? \"\").trim()\n const indexRaw = String(ctx.req.params?.index ?? \"\").trim()\n const index = Number(indexRaw)\n\n if (!uploadId || !Number.isInteger(index) || index < 0) {\n ctx.res.status(400)\n return { ok: false, error: \"invalid_chunk_ref\" }\n }\n\n const ability = buildUploadsAbility(ctx, tenantId)\n const modelCtx = getModelCtx(ctx, tenantId, ability)\n\n const [UploadSession, UploadChunk] = await Promise.all([\n models.get(\"RBUploadSession\", modelCtx) as Promise<Model<UploadSessionDoc>>,\n models.get(\"RBUploadChunk\", modelCtx) as Promise<Model<UploadChunkDoc>>,\n ])\n\n if (!ability.can(\"update\", \"RBUploadSession\")) {\n ctx.res.status(401)\n return { ok: false, error: \"unauthorized\" }\n }\n\n const session = await UploadSession.findOne({ $and: [{ _id: uploadId }, getUploadSessionAccessQuery(ability, \"update\")] }).lean()\n if (!session) {\n ctx.res.status(404)\n return { ok: false, error: \"not_found\" }\n }\n\n if (session.status !== \"uploading\") {\n ctx.res.status(409)\n return { ok: false, error: \"not_uploading\" }\n }\n\n if (index >= session.chunksTotal) {\n ctx.res.status(400)\n return { ok: false, error: \"index_out_of_range\" }\n }\n\n const data = toBufferPayload(payload)\n if (!data) {\n ctx.res.status(400)\n return { ok: false, error: \"invalid_body\" }\n }\n\n const expectedSize = index === session.chunksTotal - 1\n ? session.totalSize - session.chunkSize * (session.chunksTotal - 1)\n : session.chunkSize\n\n if (data.length > expectedSize) {\n ctx.res.status(413)\n return { ok: false, error: \"chunk_too_large\" }\n }\n\n if (data.length !== expectedSize) {\n ctx.res.status(400)\n return { ok: false, error: \"invalid_chunk_size\" }\n }\n\n const checksumHeader = ctx.req.get(\"X-Chunk-SHA256\")\n const sha256 = checksumHeader ? computeSha256Hex(data) : undefined\n\n if (checksumHeader) {\n const expectedSha256 = normalizeSha256Hex(checksumHeader)\n if (sha256 !== expectedSha256) {\n ctx.res.status(400)\n return { ok: false, error: \"checksum_mismatch\" }\n }\n }\n\n await ensureUploadIndexes(UploadSession, UploadChunk)\n\n await UploadChunk.updateOne(\n { uploadId, index },\n {\n $set: {\n uploadId,\n index,\n data,\n size: data.length,\n sha256,\n expiresAt: session.expiresAt,\n },\n $setOnInsert: {\n createdAt: new Date(),\n },\n },\n { upsert: true },\n )\n\n ctx.res.status(204)\n return { ok: true }\n}\n","export const rawBodyParser = ({\n limitBytes,\n maxClientBytesPerSecond,\n}: {\n limitBytes: number\n maxClientBytesPerSecond?: number | null\n}) => {\n return (req: any, res: any, next: any) => {\n const contentType = typeof req?.headers?.[\"content-type\"] === \"string\"\n ? String(req.headers[\"content-type\"])\n : \"\"\n\n if (!contentType.includes(\"application/octet-stream\")) {\n next()\n return\n }\n\n let total = 0\n const chunks: Buffer[] = []\n let done = false\n let paused = false\n let throttleTimeout: ReturnType<typeof setTimeout> | null = null\n\n const rateBytesPerSecond = typeof maxClientBytesPerSecond === \"number\" && maxClientBytesPerSecond > 0\n ? maxClientBytesPerSecond\n : null\n\n const cleanup = () => {\n req.off(\"data\", onData)\n req.off(\"end\", onEnd)\n req.off(\"error\", onError)\n req.off(\"aborted\", onAborted)\n if (throttleTimeout) {\n clearTimeout(throttleTimeout)\n throttleTimeout = null\n }\n }\n\n const finish = (error?: unknown) => {\n if (done) return\n done = true\n\n cleanup()\n\n if (error) {\n next(error)\n return\n }\n\n req.body = Buffer.concat(chunks, total)\n next()\n }\n\n const onData = (chunk: any) => {\n if (done) return\n const buffer = Buffer.isBuffer(chunk) ? chunk : Buffer.from(chunk)\n total += buffer.length\n\n if (total > limitBytes) {\n done = true\n cleanup()\n req.destroy()\n res.status(413).json({ ok: false, error: \"chunk_too_large\" })\n return\n }\n\n chunks.push(buffer)\n\n if (!rateBytesPerSecond) return\n\n const now = Date.now()\n const clientKey = getClientKey(req)\n const state = getClientRateState(clientKey, rateBytesPerSecond, now)\n const waitMs = consumeRateBudget(state, buffer.length, rateBytesPerSecond, now)\n\n if (waitMs > 0 && !paused) {\n paused = true\n req.pause()\n throttleTimeout = setTimeout(() => {\n throttleTimeout = null\n paused = false\n if (done) return\n try {\n req.resume()\n } catch {\n //\n }\n }, waitMs)\n }\n }\n\n const onEnd = () => finish()\n const onError = (err: unknown) => finish(err)\n const onAborted = () => finish(new Error(\"request_aborted\"))\n\n req.on(\"data\", onData)\n req.on(\"end\", onEnd)\n req.on(\"error\", onError)\n req.on(\"aborted\", onAborted)\n }\n}\n\ntype ClientRateState = {\n tokens: number\n lastRefillMs: number\n lastSeenMs: number\n}\n\nconst MAX_BURST_SECONDS = 1\nconst STALE_CLIENT_MS = 15 * 60 * 1000\n\nconst clientRateStates = new Map<string, ClientRateState>()\nlet lastCleanupMs = 0\n\nconst getClientKey = (req: any): string => {\n const rawClientIp = typeof req?.clientIp === \"string\" ? req.clientIp : \"\"\n if (rawClientIp.trim()) return rawClientIp.trim()\n const rawIp = typeof req?.ip === \"string\" ? req.ip : \"\"\n return rawIp.trim() || \"unknown\"\n}\n\nconst maybeCleanupStates = (now: number) => {\n if (now - lastCleanupMs < 60_000) return\n lastCleanupMs = now\n\n if (clientRateStates.size < 2000) return\n\n for (const [key, state] of clientRateStates) {\n if (now - state.lastSeenMs > STALE_CLIENT_MS) {\n clientRateStates.delete(key)\n }\n }\n}\n\nconst getClientRateState = (key: string, rateBytesPerSecond: number, now: number): ClientRateState => {\n maybeCleanupStates(now)\n\n const capacity = rateBytesPerSecond * MAX_BURST_SECONDS\n const existing = clientRateStates.get(key)\n if (existing) {\n existing.lastSeenMs = now\n existing.tokens = Math.min(capacity, existing.tokens)\n return existing\n }\n\n const next: ClientRateState = {\n tokens: capacity,\n lastRefillMs: now,\n lastSeenMs: now,\n }\n clientRateStates.set(key, next)\n return next\n}\n\nconst consumeRateBudget = (\n state: ClientRateState,\n bytes: number,\n rateBytesPerSecond: number,\n now: number,\n): number => {\n const capacity = rateBytesPerSecond * MAX_BURST_SECONDS\n const elapsedMs = Math.max(0, now - state.lastRefillMs)\n\n if (elapsedMs > 0) {\n state.tokens = Math.min(capacity, state.tokens + (elapsedMs * rateBytesPerSecond) / 1000)\n state.lastRefillMs = now\n }\n\n state.tokens -= bytes\n\n if (state.tokens >= 0) return 0\n return Math.ceil((-state.tokens / rateBytesPerSecond) * 1000)\n}\n","import { Api } from \"@rpcbase/api\"\n\nimport { completeUpload } from \"./handlers/completeUpload\"\nimport { getStatus } from \"./handlers/getStatus\"\nimport { initUpload } from \"./handlers/initUpload\"\nimport { uploadChunk } from \"./handlers/uploadChunk\"\nimport { rawBodyParser } from \"./middleware/rawBodyParser\"\nimport { getChunkSizeBytes, getMaxClientUploadBytesPerSecond, getRawBodyLimitBytes, type SessionUser } from \"./shared\"\n\nimport * as Uploads from \"./index\"\n\n\nexport default (api: Api<SessionUser>) => {\n const chunkSizeBytes = getChunkSizeBytes()\n api.use(\n Uploads.InitRoute,\n rawBodyParser({\n limitBytes: getRawBodyLimitBytes(chunkSizeBytes),\n maxClientBytesPerSecond: getMaxClientUploadBytesPerSecond(),\n }),\n )\n\n api.post(Uploads.InitRoute, initUpload)\n api.put(Uploads.ChunkRoute, uploadChunk)\n api.get(Uploads.StatusRoute, getStatus)\n api.post(Uploads.CompleteRoute, completeUpload)\n}\n"],"mappings":";;;;;;;;;AAMA,IAAMG,gBAAgB,MAAM;AAE5B,IAAMC,SAAS,IAAIJ,MAAM,GAAG,CAACI;AAC7B,IAAMC,YAAYJ,gBAAgBG,OAAO;AAEzC,IAAME,qBAAqBC,QAAwBA,IAAIC,QAAQ,WAAW,GAAG,CAACC,WAAW;AAEzF,IAAMC,oBAAoBC,SAA0B;CAClD,MAAMC,aAAaN,kBAAkBK,KAAK;AAC1C,KAAI,CAACC,WAAWC,WAAW,IAAI,CAAE,QAAO;AACxC,QAAO,gBAAgBC,KAAKF,WAAW;;AAGzC,IAAaG,gBAAgBC,UAA2BN,iBAAiBM,MAAME,SAAS,OAAO,CAAC;AAEhG,IAAaC,eAAeC,QAC1Bf,UAAUgB,SAASD,KAAK,EACtBE,cAAc;CAAEF,KAAK;CAAMG,YAAY;CAAK,EAC7C,CAAC;;;ACFJ,IAAagC,mBAAmBC,OAAOC,OAAO,CDIW;CACvDhC,IAAI;CACJC,UAAUvB;CACVwB,QAAQ,EAAEX,YAAYD,aAAaC,MAAM;CACzCY,UAAUC,SAA6C;AACrD,MAAIA,KAAKE,SAAS5B,cAChB,OAAM,IAAI6B,MAAM,gBAAgB;EAGlC,MAAMC,UAAUJ,KAAKX,SAAS,OAAO;AACrC,MAAI,CAACR,iBAAiBuB,QAAQ,CAC5B,OAAM,IAAID,MAAM,cAAc;EAGhC,MAAME,YAAYf,YAAYc,QAAQ;AACtC,MAAI,CAACC,UAAUC,MAAM,IAAI,CAACzB,iBAAiBwB,UAAU,CACnD,OAAM,IAAIF,MAAM,sBAAsB;EAGxC,MAAMI,kBAAkBnB,OAAOoB,KAAKH,WAAW,OAAO;AACtD,MAAIE,gBAAgBL,SAAS5B,cAC3B,OAAM,IAAI6B,MAAM,gBAAgB;AAGlC,SAAO;GAAEH,MAAMO;GAAiBN,UAAU;GAAiB;;CAE9D,CC9BmE,CAAiC;AAErG,IAAa4B,mCACXH,iBAAiBI,QAAQC,KAAKC,cAAcC,KAAKF,IAAIA,KAAKC,UAAUX,SAAS,EAAE,EAAE;AAEnF,IAAaa,0BAA0BX,QACrCG,iBAAiBS,QAAQH,cAAcA,UAAUV,MAAMC,IAAI,CAAC;AAE9D,IAAaa,wBAAwB,OACnCnB,MACAM,QACmE;CACnE,IAAIgB,cAActB;CAClB,IAAIuB,kBAAkBjB,IAAIX;CAC1B,MAAM0B,UAAoB,EAAE;AAE5B,MAAK,MAAMN,aAAaN,kBAAkB;EACxC,MAAMe,eAA2C;GAC/C9B,UAAUY,IAAIZ;GACdC,gBAAgB4B;GAChB3B,WAAW0B,YAAYG;GACvB5B,OAAOyB;GACR;AAED,MAAI,CAACP,UAAUV,MAAMmB,aAAa,CAAE;AAEpC,MAAIF,YAAYG,SAASV,UAAUX,SACjC,OAAM,IAAIsB,MAAM,4BAA4B;EAG9C,MAAMC,SAAS,MAAMZ,UAAUR,QAAQe,aAAaE,aAAa;AACjEF,gBAAcK,OAAO3B;AACrB,MAAI,OAAO2B,OAAO1B,aAAa,YAAY0B,OAAO1B,SAAS2B,MAAM,CAC/DL,mBAAkBI,OAAO1B,SAAS2B,MAAM;AAE1CP,UAAQQ,KAAKd,UAAUZ,GAAG;;AAG5B,QAAO;EACLH,MAAMsB;EACNrB,UAAUsB;EACVF;EACD;;;;AC3CH,IAAM4B,wBAAwB,OAAOC,WAAiD,IAAIG,SAASC,SAASC,WAAW;AACrHL,QAAOM,KAAK,UAAUF,QAAQ;AAC9BJ,QAAOM,KAAK,SAASD,OAAO;EAC5B;AAEF,IAAME,gBAAgB,OAAOP,QAA+BQ,UAAiC;AAE3F,KADWR,OAAOW,MAAMH,MAAM,CACtB;AACR,OAAM,IAAIL,SAAeC,SAASC,WAAW;EAC3C,MAAMO,gBAAgB;AACpBC,YAAS;AACTT,YAAS;;EAGX,MAAMU,WAAWC,UAAmB;AAClCF,YAAS;AACTR,UAAOU,MAAM;;EAGf,MAAMF,gBAAgB;AACpBb,UAAOgB,IAAI,SAASJ,QAAQ;AAC5BZ,UAAOgB,IAAI,SAASF,QAAQ;;AAG9Bd,SAAOiB,GAAG,SAASL,QAAQ;AAC3BZ,SAAOiB,GAAG,SAASH,QAAQ;GAC3B;;AAGJ,IAAMI,oBAAoB,OAAOlB,WAAmC;AAClE,KAAI,CAACA,OAAQ;AACb,KAAI,OAAQA,OAA+BmB,UAAU,WACnD,KAAI;AACF,QAAOnB,OAAiDmB,OAAO;AAC/D;SACM;AAIV,KAAI;AACAnB,SAAoCoB,WAAW;SAC3C;;AAKV,IAAaC,iBAAkG,OAC7GG,UACAC,QAC6C;CAC7C,MAAMC,WAAW5B,YAAY2B,IAAI;AACjC,KAAI,CAACC,UAAU;AACbD,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAkB;;CAG/C,MAAMc,WAAWC,OAAOL,IAAIM,IAAIC,QAAQH,YAAY,GAAG,CAACI,MAAM;AAC9D,KAAI,CAACJ,UAAU;AACbJ,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAqB;;CAGlD,MAAMmB,UAAUzC,oBAAoBgC,KAAKC,SAAS;CAClD,MAAMS,WAAWvC,YAAY6B,KAAKC,UAAUQ,QAAQ;CAEpD,MAAM,CAACE,eAAeC,eAAe,MAAMlC,QAAQmC,IAAI,CACrDxD,OAAOyD,IAAI,mBAAmBJ,SAAS,EACvCrD,OAAOyD,IAAI,iBAAiBJ,SAAS,CACtC,CAAC;AAEF,KAAI,CAACD,QAAQM,IAAI,UAAU,kBAAkB,EAAE;AAC7Cf,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAgB;;CAG7C,MAAM0B,WAAW,MAAML,cAAcM,QAAQ,EAAEC,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,OAAO,CAAA,EAAG,CAAC,CAACW,MAAM;AAChI,KAAI,CAACJ,UAAU;AACbhB,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAa;;AAG1C,KAAI0B,SAASb,WAAW,UAAUa,SAASK,OACzC,QAAO;EAAEpC,IAAI;EAAMoC,QAAQL,SAASK;EAAQ;CAG9C,MAAMC,SAAS,MAAMX,cAAcY,iBACjC,EAAEL,MAAM;EAAC,EAAEC,KAAKf,UAAU;EAAE,EAAED,QAAQ,aAAa;EAAE/B,4BAA4BqC,SAAS,SAAS;EAAA,EAAG,EACtG;EAAEe,MAAM,EAAErB,QAAQ,cAAc;EAAEsB,QAAQ,EAAEnC,OAAO,IAAG;EAAG,EACzD,EAAEoC,gBAAgB,SACpB,CAAC,CAACN,MAAM;AAER,KAAI,CAACE,QAAQ;AACXtB,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAiB;;AAG9C,OAAMrB,oBAAoB0C,eAAeC,YAAY;CAGrD,MAAMgB,YADO,MAAMxE,sBAAsB6C,SAAS,EAC5B4B;AACtB,KAAI,CAACD,UAAU;AACb,QAAMjB,cAAcmB,UAClB,EAAEZ,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,SAAS,CAAA,EAAG,EAC7E,EAAEe,MAAM;GAAErB,QAAQ;GAASb,OAAO;GAA4B,EAChE,CAAC;AACDU,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAmB;;CAGhD,MAAM0C,SAAS,IAAI1E,aAAasE,UAAU,EAAEG,YADzB7D,eAAe,EACsB,CAAC;CAEzD,MAAM+D,eAAe,OAAOX,OAAOY,WAAW,WAAWZ,OAAOY,SAASC,KAAAA;CACzE,MAAMC,oBAAoBzE,4BAA4B;CACtD,MAAM0E,4BAA4Bf,OAAOgB,aAAaF;CAEtD,MAAMM,cADmBpB,OAAOkB,SAAShC,MAAM,CAACiC,aAAa,KACpB,mBAAmBnB,OAAOqB,SAASnC,MAAM,CAACiC,aAAa,CAACG,SAAS,OAAO;CAEjH,IAAIC,eAA6C;CACjD,IAAIC,gBAAgBxB,OAAOkB;CAC3B,IAAIO,mBAA6B,EAAE;CACnC,IAAIC,gBAAyC;EAC3C5C;EACAH;EACAuC,UAAUlB,OAAOkB;EACjBF,WAAWhB,OAAOgB;EAClB,GAAI,OAAOhB,OAAO2B,aAAa,YAAY,EAAEA,UAAU3B,OAAO2B,UAAU,GAAG,EAAE;EAC7E,GAAI,OAAO3B,OAAO4B,iBAAiB,WAAW,EAAEA,cAAc5B,OAAO4B,cAAc,GAAG,EAAE;EACxF,GAAIjB,eAAe,EAAEC,QAAQD,cAAc,GAAG,EAAE;EACjD;AAED,KAAI;AACF,MAAI,CAACI,6BAA6BK,YAChC,OAAM,IAAIS,MAAM,gBAAgB;EAGlC,MAAMC,SAASxC,YAAYyC,KAAK,EAAEjD,UAAU,CAAC,CAACkD,KAAK,EAAEC,OAAO,GAAG,CAAC,CAACH,QAAQ;EAIzE,IAAIM,gBAAgB;EACpB,MAAMC,SAAmB,EAAE;EAC3B,IAAIC,gBAAgB;EAEpB,MAAMC,gBAA0B,EAAE;EAClC,MAAMC,aAAuB,EAAE;EAC/B,IAAIC,aAAa;AAEjB,MAAI;AACF,cAAW,MAAMC,YAAYZ,QAAQ;AACnC,QAAIY,SAAST,UAAUG,cACrB,OAAM,IAAIP,MAAM,iBAAiB;IAGnC,MAAMpE,QAAQiF,SAASC;AAEvB,QAAI5B,2BAA2B;AAC7BsB,YAAOO,KAAKnF,MAAM;AAClB6E,sBAAiB7E,MAAMoF;eACd,CAACtB,cAAc;AACxBgB,mBAAcK,KAAKnF,MAAM;AAEzB,SAAIgF,aAAa3B,mBAAmB;MAClC,MAAMgC,QAAQrF,MAAMsF,SAAS,GAAGC,KAAKC,IAAIxF,MAAMoF,QAAQ/B,oBAAoB2B,WAAW,CAAC;AACvF,UAAIK,MAAMD,QAAQ;AAChBL,kBAAWI,KAAKE,MAAM;AACtBL,qBAAcK,MAAMD;;;AAIxB,SAAIJ,cAAc3B,mBAAmB;MACnC,MAAMoC,QAAQxF,OAAOyF,OAAOX,YAAYC,WAAW;AAQnD,UAPmBnG,uBAAuB;OACxC+E,UAAUrB,OAAOqB;OACjBgC,gBAAgBrD,OAAOkB;OACvBF,WAAWhB,OAAOgB;OAClBkC;OACD,CAAC,CAEaL,OACb,OAAM,IAAIhB,MAAM,gBAAgB;AAGlCH,sBAAgB;OACd5C;OACAH;OACAuC,UAAUlB,OAAOkB;OACjBF,WAAWhB,OAAOgB;OAClB,GAAI,OAAOhB,OAAO2B,aAAa,YAAY,EAAEA,UAAU3B,OAAO2B,UAAU,GAAG,EAAE;OAC7E,GAAI,OAAO3B,OAAO4B,iBAAiB,WAAW,EAAEA,cAAc5B,OAAO4B,cAAc,GAAG,EAAE;OACxF,GAAIjB,eAAe,EAAEC,QAAQD,cAAc,GAAG,EAAE;OACjD;AACDY,qBAAeb,OAAO4C,iBAAiBtD,OAAOqB,UAAU,EACtDkC,UAAU7B,eACX,CAAC;AAEF,WAAK,MAAM8B,WAAWjB,cACpB,OAAM/E,cAAc+D,cAAciC,QAAQ;AAE5CjB,oBAAcM,SAAS;;UAGzB,OAAMrF,cAAc+D,cAAc9D,MAAM;AAG1C2E,qBAAiB;;YAEX;AACR,OAAI;AACF,UAAMN,OAAOK,OAAO;WACd;;AAKV,MAAIC,kBAAkBpC,OAAOyD,YAC3B,OAAM,IAAI5B,MAAM,iBAAiB;AAGnC,MAAId,2BAA2B;GAE7B,MAAM,EAAE4B,MAAMgB,WAAWzC,UAAU0C,mBAAmBC,YAAY,MAAMzH,sBADtDsB,OAAOyF,OAAOd,QAAQC,cAAc,EACmD;IACvGjB,UAAUrB,OAAOqB;IACjBgC,gBAAgBrD,OAAOkB;IACxB,CAAC;AAEFM,mBAAgBoC;AAChBnC,sBAAmBoC;AACnBnC,mBAAgB;IACd5C;IACAH;IACAuC,UAAU0C;IACV5C,WAAWhB,OAAOgB;IAClB,GAAI6C,QAAQhB,SAAS;KAAEO,YAAYS;KAASC,eAAeH,UAAUd;KAAQ,GAAG,EAAE;IAClF,GAAI,OAAO7C,OAAO2B,aAAa,YAAY,EAAEA,UAAU3B,OAAO2B,UAAU,GAAG,EAAE;IAC7E,GAAI,OAAO3B,OAAO4B,iBAAiB,WAAW,EAAEA,cAAc5B,OAAO4B,cAAc,GAAG,EAAE;IACxF,GAAIjB,eAAe,EAAEC,QAAQD,cAAc,GAAG,EAAE;IACjD;AACDY,kBAAeb,OAAO4C,iBAAiBtD,OAAOqB,UAAU,EACtDkC,UAAU7B,eACX,CAAC;GAEF,MAAMqC,WAAW/G,sBAAsBuE,aAAa;AACpDA,gBAAayC,IAAIL,UAAU;AAC3B,SAAMI;SACD;AACL,OAAI,CAACxC,cAAc;IACjB,MAAM2B,QAAQxF,OAAOyF,OAAOX,YAAYC,WAAW;AAQnD,QAPmBnG,uBAAuB;KACxC+E,UAAUrB,OAAOqB;KACjBgC,gBAAgBrD,OAAOkB;KACvBF,WAAWhB,OAAOgB;KAClBkC;KACD,CAAC,CAEaL,OACb,OAAM,IAAIhB,MAAM,gBAAgB;AAGlCH,oBAAgB;KACd5C;KACAH;KACAuC,UAAUlB,OAAOkB;KACjBF,WAAWhB,OAAOgB;KAClB,GAAI,OAAOhB,OAAO2B,aAAa,YAAY,EAAEA,UAAU3B,OAAO2B,UAAU,GAAG,EAAE;KAC7E,GAAI,OAAO3B,OAAO4B,iBAAiB,WAAW,EAAEA,cAAc5B,OAAO4B,cAAc,GAAG,EAAE;KACxF,GAAIjB,eAAe,EAAEC,QAAQD,cAAc,GAAG,EAAE;KACjD;AACDY,mBAAeb,OAAO4C,iBAAiBtD,OAAOqB,UAAU,EACtDkC,UAAU7B,eACX,CAAC;AAEF,SAAK,MAAM8B,WAAWjB,cACpB,OAAM/E,cAAc+D,cAAciC,QAAQ;AAE5CjB,kBAAcM,SAAS;;GAGzB,MAAMkB,WAAW/G,sBAAsBuE,aAAa;AACpDA,gBAAayC,KAAK;AAClB,SAAMD;;EAGR,MAAMhE,SAAShB,OAAQwC,aAA6C0C,MAAM,GAAG;AAC7E,MAAI,CAAClE,OACH,OAAM,IAAI8B,MAAM,kBAAkB;AAGpC,QAAMxC,cAAcmB,UAClB,EAAEZ,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,SAAS,CAAA,EAAG,EAC7E;GAAEe,MAAM;IAAErB,QAAQ;IAAQkB;IAAQ;GAAEI,QAAQ,EAAEnC,OAAO,IAAG;GAC1D,CAAC;AAED,QAAM7B,4BAA4B;GAChCwC;GACAG;GACAiB;GACAsB,UAAUrB,OAAOqB;GACjBH,UAAUM;GACV6B,gBAAgBrD,OAAOkB;GACvBF,WAAWhB,OAAOgB;GAClB,GAAI,OAAOhB,OAAO2B,aAAa,YAAY,EAAEA,UAAU3B,OAAO2B,UAAU,GAAG,EAAE;GAC7E,GAAI,OAAO3B,OAAO4B,iBAAiB,WAAW,EAAEA,cAAc5B,OAAO4B,cAAc,GAAG,EAAE;GACxF,GAAIjB,eAAe,EAAEC,QAAQD,cAAc,GAAG,EAAE;GAChDc;GACA8B,UAAU7B;GACX,CAAC,CAACwC,OAAOlG,UAAU;AAClBmG,WAAQnG,MAAM,wCAAwC;IACpDW;IACAG;IACAiB;IACA/B;IACD,CAAC;IACF;AAEF,MAAI;AACF,SAAMsB,YAAY8E,WAAW,EAAEtF,UAAU,CAAC;UACpC;AAIR,SAAO;GAAEnB,IAAI;GAAMoC;GAAQ;UACpB/B,OAAO;EACd,MAAMqG,UAAUrG,iBAAiB6D,QAAQ7D,MAAMqG,UAAUtF,OAAOf,MAAM;AAEtE,QAAMG,kBAAkBoD,aAAa;AAErC,MAAI8C,YAAY,kBAAkB;AAChC,SAAMhF,cAAcmB,UAClB,EAAEZ,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,SAAS,CAAA,EAAG,EAC7E,EAAEe,MAAM,EAAErB,QAAQ,aAAY,EAChC,CAAC;AACDH,OAAIE,IAAIC,OAAO,IAAI;AACnB,UAAO;IAAElB,IAAI;IAAOK,OAAO;IAAkB;;AAG/C,MAAIqG,YAAY,iBAAiB;AAC/B,SAAMhF,cAAcmB,UAClB,EAAEZ,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,SAAS,CAAA,EAAG,EAC7E,EAAEe,MAAM;IAAErB,QAAQ;IAASb,OAAOqG;IAAQ,EAC5C,CAAC;AACD3F,OAAIE,IAAIC,OAAO,IAAI;AACnB,UAAO;IAAElB,IAAI;IAAOK,OAAOqG;IAAS;;AAGtC,MAAIA,YAAY,iBAAiBA,YAAY,uBAAuB;AAClE,SAAMhF,cAAcmB,UAClB,EAAEZ,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,SAAS,CAAA,EAAG,EAC7E,EAAEe,MAAM;IAAErB,QAAQ;IAASb,OAAOqG;IAAQ,EAC5C,CAAC;AACD3F,OAAIE,IAAIC,OAAO,IAAI;AACnB,UAAO;IAAElB,IAAI;IAAOK,OAAOqG;IAAS;;AAGtC,QAAMhF,cAAcmB,UAClB,EAAEZ,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEhC,4BAA4BqC,SAAS,SAAS,CAAA,EAAG,EAC7E,EAAEe,MAAM;GAAErB,QAAQ;GAASb,OAAOqG;GAAQ,EAC5C,CAAC;AAED3F,MAAIE,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAElB,IAAI;GAAOK,OAAO;GAAmB;;;;;AC5WlD,IAAaiH,YAA2F,OACtGG,UACAC,QAC2C;CAC3C,MAAME,WAAWP,YAAYK,IAAI;AACjC,KAAI,CAACE,UAAU;AACbF,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAEC,IAAI;GAAOC,OAAO;GAAkB;;CAG/C,MAAMC,WAAWC,OAAOR,IAAIS,IAAIC,QAAQH,YAAY,GAAG,CAACI,MAAM;AAC9D,KAAI,CAACJ,UAAU;AACbP,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAEC,IAAI;GAAOC,OAAO;GAAqB;;CAGlD,MAAMM,UAAUpB,oBAAoBQ,KAAKE,SAAS;CAClD,MAAMW,WAAWpB,YAAYO,KAAKE,UAAUU,QAAQ;CAEpD,MAAM,CAACE,eAAeC,eAAe,MAAMd,QAAQe,IAAI,CACrD9B,OAAO+B,IAAI,mBAAmBJ,SAAS,EACvC3B,OAAO+B,IAAI,iBAAiBJ,SAAS,CACtC,CAAC;AAEF,KAAI,CAACD,QAAQM,IAAI,QAAQ,kBAAkB,EAAE;AAC3ClB,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAEC,IAAI;GAAOC,OAAO;GAAgB;;CAG7C,MAAMa,UAAU,MAAML,cAAcM,QAAQ,EAAEC,MAAM,CAAC,EAAEC,KAAKf,UAAU,EAAEb,4BAA4BkB,SAAS,OAAO,CAAA,EAAG,CAAC,CAACW,MAAM;AAC/H,KAAI,CAACJ,SAAS;AACZnB,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAEC,IAAI;GAAOC,OAAO;GAAa;;CAQ1C,MAAMsB,YALe,MAAMb,YAAYU,KACrC,EAAElB,UAAU,EACZ;EAAEmB,OAAO;EAAGJ,KAAK;EACnB,CAAC,CAACK,KAAK,EAAED,OAAO,GAAG,CAAC,CAACH,MAAM,EAGxBO,KAAKC,QAAS,OAAOA,IAAIL,UAAU,WAAWK,IAAIL,QAAQ,GAAI,CAC9DM,QAAQC,MAAMC,OAAOC,UAAUF,EAAE,IAAIA,KAAK,EAAE;AAE/C,QAAO;EACL5B,IAAI;EACJD,QAAQe,QAAQf;EAChBgC,WAAWjB,QAAQiB;EACnBC,aAAalB,QAAQkB;EACrBT;EACA,GAAIT,QAAQmB,SAAS,EAAEA,QAAQnB,QAAQmB,QAAQ,GAAG,EAAE;EACrD;;;;AChEH,IAAaE,YAAY;AACzB,IAAaC,aAAa;AAC1B,IAAaC,cAAc;AAC3B,IAAaC,gBAAgB;AAE7B,IAAaC,oBAAoBL,OAAS;CACxCO,UAAUP,QAAU,CAACS,IAAI,EAAE;CAC3BC,UAAUV,QAAU,CAACS,IAAI,EAAE;CAC3BE,UAAUX,SAAW,CAACa,UAAU;CAChCC,WAAWd,QAAU,CAACgB,KAAK,CAACP,IAAI,EAAC;CAClC,CAAC;AAIgCT,OAAS;CACzCoB,IAAIpB,SAAW;CACfqB,OAAOrB,QAAU,CAACa,UAAU;CAC5BS,UAAUtB,QAAU,CAACa,UAAU;CAC/BU,WAAWvB,QAAU,CAACa,UAAU;CAChCW,WAAWxB,QAAU,CAACgB,KAAK,CAACH,UAAU;CACtCY,aAAazB,QAAU,CAACgB,KAAK,CAACH,UAAS;CACxC,CAAC;AAIkCb,OAAS;CAC3CoB,IAAIpB,SAAW;CACfqB,OAAOrB,QAAU,CAACa,UAAU;CAC5Be,QAAQ5B,MAAO;EAAC;EAAa;EAAc;EAAQ;EAAQ,CAAC,CAACa,UAAU;CACvEW,WAAWxB,QAAU,CAACgB,KAAK,CAACH,UAAU;CACtCY,aAAazB,QAAU,CAACgB,KAAK,CAACH,UAAU;CACxCiB,UAAU9B,MAAQA,QAAU,CAACgB,KAAK,CAACP,IAAI,EAAE,CAAC,CAACI,UAAU;CACrDmB,QAAQhC,QAAU,CAACa,UAAS;CAC7B,CAAC;AAIoCb,OAAS;CAC7CoB,IAAIpB,SAAW;CACfqB,OAAOrB,QAAU,CAACa,UAAU;CAC5BmB,QAAQhC,QAAU,CAACa,UAAS;CAC7B,CAAC;;;ACrBF,IAAawC,aAA+F,OAC1GG,SACAC,QACyC;CACzC,MAAME,WAAWR,YAAYM,IAAI;AACjC,KAAI,CAACE,UAAU;AACbF,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAEC,IAAI;GAAOC,OAAO;GAAkB;;CAG/C,MAAMC,SAASZ,UAAUK,IAAI;CAE7B,MAAMQ,SAAAA,kBAAmCE,UAAUX,WAAW,EAAE,CAAC;AACjE,KAAI,CAACS,OAAOG,SAAS;AACnBX,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAEC,IAAI;GAAOC,OAAO;GAAmB;;CAGhD,MAAMM,YAAYrB,mBAAmB;CACrC,MAAM,EAAEsB,UAAUC,UAAUC,WAAWC,aAAaR,OAAOS;CAC3D,MAAMC,cAAcC,KAAKC,KAAKL,YAAYH,UAAU;CAGpD,MAAMU,WAAW9B,YAAYQ,KAAKE,UADlBd,oBAAoBY,KAAKE,SAAS,CACE;CAEpD,MAAM,CAACqB,eAAeC,eAAe,MAAMvB,QAAQwB,IAAI,CACrD5C,OAAO6C,IAAI,mBAAmBJ,SAAS,EACvCzC,OAAO6C,IAAI,iBAAiBJ,SAAS,CACtC,CAAC;AAEF,OAAMhC,oBAAoBiC,eAAeC,YAAY;CAErD,MAAMG,WAAW,IAAI7C,UAAU,CAAC8C,UAAU;CAC1C,MAAMC,MAAMC,KAAKD,KAAK;CACtB,MAAME,YAAY,IAAID,KAAKD,MAAMpC,iBAAiB,CAAC;CAEnD,MAAMuC,YAAYzB,SAAS,OAAO5B,YAAY,GAAG,CAACiD,SAAS,YAAY;CACvE,MAAMK,eAAeD,YAAY3C,iBAAiB6C,OAAOC,KAAKH,UAAU,CAAC,GAAGI,KAAAA;AAE5E,OAAMb,cAAcc,OAAO;EACzBC,KAAKX;EACL,GAAIpB,SAAS,EAAEA,QAAQ,GAAG,EAAE;EAC5B,GAAI0B,eAAe,EAAEA,cAAc,GAAG,EAAE;EACxCpB;EACAC;EACA,GAAI,OAAOE,aAAa,YAAY,EAAEA,UAAU,GAAG,EAAE;EACrDD;EACAH;EACAM;EACAd,QAAQ;EACRmC,WAAW,IAAIT,KAAKD,IAAI;EACxBE;EACD,CAAC;AAEF,QAAO;EACL1B,IAAI;EACJsB;EACAf;EACAM;EACA,GAAIc,YAAY,EAAEA,WAAW,GAAG,EAAE;EACnC;;;;AC3DH,IAAayB,cAAqE,OAChFE,SACAC,QACkC;CAClC,MAAME,WAAWX,YAAYS,IAAI;AACjC,KAAI,CAACE,UAAU;AACbF,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAkB;;CAG/C,MAAMS,WAAWC,OAAON,IAAIO,IAAIC,QAAQH,YAAY,GAAG,CAACI,MAAM;CAC9D,MAAMC,WAAWJ,OAAON,IAAIO,IAAIC,QAAQG,SAAS,GAAG,CAACF,MAAM;CAC3D,MAAME,QAAQC,OAAOF,SAAS;AAE9B,KAAI,CAACL,YAAY,CAACO,OAAOC,UAAUF,MAAM,IAAIA,QAAQ,GAAG;AACtDX,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAqB;;CAGlD,MAAMkB,UAAU5B,oBAAoBc,KAAKE,SAAS;CAClD,MAAMa,WAAW1B,YAAYW,KAAKE,UAAUY,QAAQ;CAEpD,MAAM,CAACE,eAAeC,eAAe,MAAMhB,QAAQiB,IAAI,CACrDrC,OAAOsC,IAAI,mBAAmBJ,SAAS,EACvClC,OAAOsC,IAAI,iBAAiBJ,SAAS,CACtC,CAAC;AAEF,KAAI,CAACD,QAAQM,IAAI,UAAU,kBAAkB,EAAE;AAC7CpB,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAgB;;CAG7C,MAAMyB,UAAU,MAAML,cAAcM,QAAQ,EAAEC,MAAM,CAAC,EAAEC,KAAKnB,UAAU,EAAEf,4BAA4BwB,SAAS,SAAS,CAAA,EAAG,CAAC,CAACW,MAAM;AACjI,KAAI,CAACJ,SAAS;AACZrB,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAa;;AAG1C,KAAIyB,QAAQjB,WAAW,aAAa;AAClCJ,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAiB;;AAG9C,KAAIe,SAASU,QAAQK,aAAa;AAChC1B,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAsB;;CAGnD,MAAM+B,OAAOlC,gBAAgBM,QAAQ;AACrC,KAAI,CAAC4B,MAAM;AACT3B,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAgB;;CAG7C,MAAMgC,eAAejB,UAAUU,QAAQK,cAAc,IACjDL,QAAQQ,YAAYR,QAAQS,aAAaT,QAAQK,cAAc,KAC/DL,QAAQS;AAEZ,KAAIH,KAAKI,SAASH,cAAc;AAC9B5B,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAmB;;AAGhD,KAAI+B,KAAKI,WAAWH,cAAc;AAChC5B,MAAIG,IAAIC,OAAO,IAAI;AACnB,SAAO;GAAET,IAAI;GAAOC,OAAO;GAAsB;;CAGnD,MAAMoC,iBAAiBhC,IAAIO,IAAIY,IAAI,iBAAiB;CACpD,MAAMc,SAASD,iBAAiB7C,iBAAiBwC,KAAK,GAAGO,KAAAA;AAEzD,KAAIF;MAEEC,WADmBzC,mBAAmBwC,eAAe,EAC1B;AAC7BhC,OAAIG,IAAIC,OAAO,IAAI;AACnB,UAAO;IAAET,IAAI;IAAOC,OAAO;IAAqB;;;AAIpD,OAAMR,oBAAoB4B,eAAeC,YAAY;AAErD,OAAMA,YAAYmB,UAChB;EAAE/B;EAAUM;EAAO,EACnB;EACE0B,MAAM;GACJhC;GACAM;GACAgB;GACAW,MAAMX,KAAKI;GACXE;GACAM,WAAWlB,QAAQkB;GACpB;EACDC,cAAc,EACZC,2BAAW,IAAIC,MAAK,EACtB;EACD,EACD,EAAEC,QAAQ,MACZ,CAAC;AAED3C,KAAIG,IAAIC,OAAO,IAAI;AACnB,QAAO,EAAET,IAAI,MAAM;;;;AC5HrB,IAAaiD,iBAAiB,EAC5BC,YACAC,8BAII;AACJ,SAAQC,KAAUC,KAAUC,SAAc;AAKxC,MAAI,EAJgB,OAAOF,KAAKI,UAAU,oBAAoB,WAC1DC,OAAOL,IAAII,QAAQ,gBAAgB,GACnC,IAEaE,SAAS,2BAA2B,EAAE;AACrDJ,SAAM;AACN;;EAGF,IAAIK,QAAQ;EACZ,MAAMC,SAAmB,EAAE;EAC3B,IAAIE,OAAO;EACX,IAAIC,SAAS;EACb,IAAIC,kBAAwD;EAE5D,MAAMG,qBAAqB,OAAOhB,4BAA4B,YAAYA,0BAA0B,IAChGA,0BACA;EAEJ,MAAMiB,gBAAgB;AACpBhB,OAAIiB,IAAI,QAAQC,OAAO;AACvBlB,OAAIiB,IAAI,OAAOE,MAAM;AACrBnB,OAAIiB,IAAI,SAASG,QAAQ;AACzBpB,OAAIiB,IAAI,WAAWI,UAAU;AAC7B,OAAIT,iBAAiB;AACnBU,iBAAaV,gBAAgB;AAC7BA,sBAAkB;;;EAItB,MAAMW,UAAUC,UAAoB;AAClC,OAAId,KAAM;AACVA,UAAO;AAEPM,YAAS;AAET,OAAIQ,OAAO;AACTtB,SAAKsB,MAAM;AACX;;AAGFxB,OAAIyB,OAAOhB,OAAOiB,OAAOlB,QAAQD,MAAM;AACvCL,SAAM;;EAGR,MAAMgB,UAAUS,UAAe;AAC7B,OAAIjB,KAAM;GACV,MAAMkB,SAASnB,OAAOoB,SAASF,MAAM,GAAGA,QAAQlB,OAAOqB,KAAKH,MAAM;AAClEpB,YAASqB,OAAOG;AAEhB,OAAIxB,QAAQT,YAAY;AACtBY,WAAO;AACPM,aAAS;AACThB,QAAIgC,SAAS;AACb/B,QAAIgC,OAAO,IAAI,CAACC,KAAK;KAAEC,IAAI;KAAOX,OAAO;KAAmB,CAAC;AAC7D;;AAGFhB,UAAO4B,KAAKR,OAAO;AAEnB,OAAI,CAACb,mBAAoB;GAEzB,MAAMsB,MAAMC,KAAKD,KAAK;GAGtB,MAAMM,SAASC,kBADDF,mBADIF,aAAaxC,IAAI,EACSe,oBAAoBsB,IAAI,EAC5BT,OAAOG,QAAQhB,oBAAoBsB,IAAI;AAE/E,OAAIM,SAAS,KAAK,CAAChC,QAAQ;AACzBA,aAAS;AACTX,QAAI6C,OAAO;AACXjC,sBAAkBE,iBAAiB;AACjCF,uBAAkB;AAClBD,cAAS;AACT,SAAID,KAAM;AACV,SAAI;AACFV,UAAI8C,QAAQ;aACN;OAGPH,OAAO;;;EAId,MAAMxB,cAAcI,QAAQ;EAC5B,MAAMH,WAAW2B,QAAiBxB,OAAOwB,IAAI;EAC7C,MAAM1B,kBAAkBE,uBAAO,IAAIyB,MAAM,kBAAkB,CAAC;AAE5DhD,MAAIiD,GAAG,QAAQ/B,OAAO;AACtBlB,MAAIiD,GAAG,OAAO9B,MAAM;AACpBnB,MAAIiD,GAAG,SAAS7B,QAAQ;AACxBpB,MAAIiD,GAAG,WAAW5B,UAAU;;;AAUhC,IAAMiC,oBAAoB;AAC1B,IAAMC,kBAAkB,MAAU;AAElC,IAAMC,mCAAmB,IAAIC,KAA8B;AAC3D,IAAIC,gBAAgB;AAEpB,IAAMlB,gBAAgBxC,QAAqB;CACzC,MAAM2D,cAAc,OAAO3D,KAAK4D,aAAa,WAAW5D,IAAI4D,WAAW;AACvE,KAAID,YAAYE,MAAM,CAAE,QAAOF,YAAYE,MAAM;AAEjD,SADc,OAAO7D,KAAK+D,OAAO,WAAW/D,IAAI+D,KAAK,IACxCF,MAAM,IAAI;;AAGzB,IAAMG,sBAAsB3B,QAAgB;AAC1C,KAAIA,MAAMqB,gBAAgB,IAAQ;AAClCA,iBAAgBrB;AAEhB,KAAImB,iBAAiBS,OAAO,IAAM;AAElC,MAAK,MAAM,CAACC,KAAKzB,UAAUe,iBACzB,KAAInB,MAAMI,MAAMY,aAAaE,gBAC3BC,kBAAiBW,OAAOD,IAAI;;AAKlC,IAAMxB,sBAAsBwB,KAAanD,oBAA4BsB,QAAiC;AACpG2B,oBAAmB3B,IAAI;CAEvB,MAAM+B,WAAWrD,qBAAqBuC;CACtC,MAAMe,WAAWb,iBAAiBc,IAAIJ,IAAI;AAC1C,KAAIG,UAAU;AACZA,WAAShB,aAAahB;AACtBgC,WAASlB,SAASoB,KAAKC,IAAIJ,UAAUC,SAASlB,OAAO;AACrD,SAAOkB;;CAGT,MAAMnE,OAAwB;EAC5BiD,QAAQiB;EACRhB,cAAcf;EACdgB,YAAYhB;EACb;AACDmB,kBAAiBiB,IAAIP,KAAKhE,KAAK;AAC/B,QAAOA;;AAGT,IAAM0C,qBACJH,OACAiC,OACA3D,oBACAsB,QACW;CACX,MAAM+B,WAAWrD,qBAAqBuC;CACtC,MAAMqB,YAAYJ,KAAKK,IAAI,GAAGvC,MAAMI,MAAMW,aAAa;AAEvD,KAAIuB,YAAY,GAAG;AACjBlC,QAAMU,SAASoB,KAAKC,IAAIJ,UAAU3B,MAAMU,SAAUwB,YAAY5D,qBAAsB,IAAK;AACzF0B,QAAMW,eAAef;;AAGvBI,OAAMU,UAAUuB;AAEhB,KAAIjC,MAAMU,UAAU,EAAG,QAAO;AAC9B,QAAOoB,KAAKM,KAAM,CAACpC,MAAMU,SAASpC,qBAAsB,IAAK;;;;AC/J/D,IAAA,mBAAgB0E,QAA0B;CACxC,MAAMC,iBAAiBN,mBAAmB;AAC1CK,KAAIE,IACFH,WACAL,cAAc;EACZU,YAAYP,qBAAqBI,eAAe;EAChDI,yBAAyBT,kCAAiC;EAC3D,CACH,CAAC;AAEDI,KAAIM,KAAKP,WAAmBP,WAAW;AACvCQ,KAAIO,IAAIR,YAAoBN,YAAY;AACxCO,KAAIS,IAAIV,aAAqBR,UAAU;AACvCS,KAAIM,KAAKP,eAAuBT,eAAe"}
|
|
@@ -1,318 +0,0 @@
|
|
|
1
|
-
import assert from "assert";
|
|
2
|
-
import { hkdfSync } from "crypto";
|
|
3
|
-
import { models } from "@rpcbase/db";
|
|
4
|
-
import { buildAbility, buildAbilityFromSession, getAccessibleByQuery, getTenantRolesFromSessionUser } from "@rpcbase/db/acl";
|
|
5
|
-
//#region src/getDerivedKey.ts
|
|
6
|
-
var getDerivedKey = (masterKey, info, length = 32, salt = "") => {
|
|
7
|
-
assert(masterKey?.length >= 32, "MASTER_KEY must be 32 chars or longer.");
|
|
8
|
-
return Buffer.from(hkdfSync("sha256", masterKey, Buffer.from(salt), Buffer.from(info), length)).toString("hex");
|
|
9
|
-
};
|
|
10
|
-
//#endregion
|
|
11
|
-
//#region src/rts/queryExecutor.ts
|
|
12
|
-
var RTS_TENANT_ID_QUERY_PARAM = "rb-tenant-id";
|
|
13
|
-
var RTS_USER_ID_HEADER = "rb-user-id";
|
|
14
|
-
var QUERY_MAX_LIMIT = 4096;
|
|
15
|
-
var INTERNAL_MODEL_NAMES = new Set(["RBRtsChange", "RBRtsCounter"]);
|
|
16
|
-
var paginationCursorSigningSecret = null;
|
|
17
|
-
var getPaginationCursorSigningSecret = () => {
|
|
18
|
-
if (paginationCursorSigningSecret) return paginationCursorSigningSecret;
|
|
19
|
-
const masterKey = process.env.MASTER_KEY?.trim();
|
|
20
|
-
if (!masterKey) throw new Error("MASTER_KEY must be defined to derive pagination cursor signing secret");
|
|
21
|
-
paginationCursorSigningSecret = getDerivedKey(masterKey, "pagination_cursor_signing");
|
|
22
|
-
return paginationCursorSigningSecret;
|
|
23
|
-
};
|
|
24
|
-
var normalizeTenantId = (value) => {
|
|
25
|
-
if (typeof value !== "string") return null;
|
|
26
|
-
const normalized = value.trim();
|
|
27
|
-
return normalized ? normalized : null;
|
|
28
|
-
};
|
|
29
|
-
var normalizeSignedInTenants = (value) => {
|
|
30
|
-
if (!Array.isArray(value)) return [];
|
|
31
|
-
return value.map((tenantId) => normalizeTenantId(String(tenantId))).filter((tenantId) => Boolean(tenantId));
|
|
32
|
-
};
|
|
33
|
-
var getTenantIdFromRequest = (req) => {
|
|
34
|
-
const rawQuery = req.query?.[RTS_TENANT_ID_QUERY_PARAM];
|
|
35
|
-
const normalizedFromQuery = normalizeTenantId(Array.isArray(rawQuery) ? rawQuery[0] : rawQuery);
|
|
36
|
-
if (normalizedFromQuery) return normalizedFromQuery;
|
|
37
|
-
return normalizeTenantId((req.session?.user)?.currentTenantId);
|
|
38
|
-
};
|
|
39
|
-
var resolveRtsRequestTenantId = (req) => {
|
|
40
|
-
return getTenantIdFromRequest(req);
|
|
41
|
-
};
|
|
42
|
-
var isRtsRequestAuthorized = (req, tenantId) => {
|
|
43
|
-
const sessionUser = req.session?.user;
|
|
44
|
-
if (!sessionUser) return false;
|
|
45
|
-
const signedInTenants = normalizeSignedInTenants(sessionUser.signedInTenants);
|
|
46
|
-
if (signedInTenants.length > 0) return signedInTenants.includes(tenantId);
|
|
47
|
-
const currentTenantId = normalizeTenantId(sessionUser.currentTenantId);
|
|
48
|
-
if (!currentTenantId) return false;
|
|
49
|
-
return currentTenantId === tenantId;
|
|
50
|
-
};
|
|
51
|
-
var buildRtsAbilityFromRequest = async (req, tenantId) => {
|
|
52
|
-
const sessionUserId = normalizeTenantId((req.session?.user)?.id);
|
|
53
|
-
if (sessionUserId) return {
|
|
54
|
-
ability: buildAbilityFromSession({
|
|
55
|
-
tenantId,
|
|
56
|
-
session: req.session
|
|
57
|
-
}),
|
|
58
|
-
userId: sessionUserId
|
|
59
|
-
};
|
|
60
|
-
const headerValue = req.headers[RTS_USER_ID_HEADER];
|
|
61
|
-
const headerUserId = normalizeTenantId(Array.isArray(headerValue) ? headerValue[0] : headerValue);
|
|
62
|
-
if (!headerUserId) return {
|
|
63
|
-
ability: buildAbilityFromSession({
|
|
64
|
-
tenantId,
|
|
65
|
-
session: req.session
|
|
66
|
-
}),
|
|
67
|
-
userId: null
|
|
68
|
-
};
|
|
69
|
-
const user = await (await models.getGlobal("RBUser", { req: { session: null } })).findById(headerUserId, {
|
|
70
|
-
tenants: 1,
|
|
71
|
-
tenantRoles: 1
|
|
72
|
-
}).lean();
|
|
73
|
-
const tenantsRaw = user?.tenants;
|
|
74
|
-
if (!(Array.isArray(tenantsRaw) ? tenantsRaw.map((tenant) => String(tenant)) : []).includes(tenantId)) throw new Error("Tenant not authorized for this session");
|
|
75
|
-
const roles = getTenantRolesFromSessionUser(user, tenantId);
|
|
76
|
-
return {
|
|
77
|
-
ability: buildAbility({
|
|
78
|
-
tenantId,
|
|
79
|
-
userId: headerUserId,
|
|
80
|
-
roles: roles.length ? roles : ["owner"]
|
|
81
|
-
}),
|
|
82
|
-
userId: headerUserId
|
|
83
|
-
};
|
|
84
|
-
};
|
|
85
|
-
var getTenantModel = async (tenantId, modelName, ability) => {
|
|
86
|
-
const ctx = {
|
|
87
|
-
req: { session: { user: { currentTenantId: tenantId } } },
|
|
88
|
-
ability
|
|
89
|
-
};
|
|
90
|
-
return models.get(modelName, ctx);
|
|
91
|
-
};
|
|
92
|
-
var normalizeLimit = (limit) => {
|
|
93
|
-
if (typeof limit !== "number") return QUERY_MAX_LIMIT;
|
|
94
|
-
if (!Number.isFinite(limit)) return QUERY_MAX_LIMIT;
|
|
95
|
-
return Math.min(QUERY_MAX_LIMIT, Math.abs(limit));
|
|
96
|
-
};
|
|
97
|
-
var normalizeString = (value) => {
|
|
98
|
-
return typeof value === "string" ? value.trim() : "";
|
|
99
|
-
};
|
|
100
|
-
var normalizeObject = (value) => {
|
|
101
|
-
if (!value || typeof value !== "object" || Array.isArray(value)) return void 0;
|
|
102
|
-
return value;
|
|
103
|
-
};
|
|
104
|
-
var normalizePagination = (value) => {
|
|
105
|
-
if (!value || typeof value !== "object" || Array.isArray(value)) return void 0;
|
|
106
|
-
return value;
|
|
107
|
-
};
|
|
108
|
-
var normalizePopulateSelect = (value) => {
|
|
109
|
-
if (typeof value === "string") return value.trim() || void 0;
|
|
110
|
-
return normalizeObject(value);
|
|
111
|
-
};
|
|
112
|
-
var normalizePopulateOptions = (value) => {
|
|
113
|
-
if (!value || typeof value !== "object" || Array.isArray(value)) return void 0;
|
|
114
|
-
const raw = value;
|
|
115
|
-
const normalized = {};
|
|
116
|
-
if (raw.sort && typeof raw.sort === "object" && !Array.isArray(raw.sort)) normalized.sort = raw.sort;
|
|
117
|
-
if (typeof raw.limit === "number" && Number.isFinite(raw.limit)) normalized.limit = Math.max(0, Math.floor(Math.abs(raw.limit)));
|
|
118
|
-
if (!normalized.sort && normalized.limit === void 0) return void 0;
|
|
119
|
-
return normalized;
|
|
120
|
-
};
|
|
121
|
-
var normalizePopulateObject = (value) => {
|
|
122
|
-
if (!value || typeof value !== "object" || Array.isArray(value)) return void 0;
|
|
123
|
-
const raw = value;
|
|
124
|
-
const path = normalizeString(raw.path);
|
|
125
|
-
if (!path) return void 0;
|
|
126
|
-
const normalized = { path };
|
|
127
|
-
const model = normalizeString(raw.model);
|
|
128
|
-
if (model) normalized.model = model;
|
|
129
|
-
const select = normalizePopulateSelect(raw.select);
|
|
130
|
-
if (select !== void 0) normalized.select = select;
|
|
131
|
-
const match = normalizeObject(raw.match);
|
|
132
|
-
if (match) normalized.match = match;
|
|
133
|
-
const nestedPopulate = normalizeRtsPopulateOption(raw.populate);
|
|
134
|
-
if (nestedPopulate !== void 0) normalized.populate = nestedPopulate;
|
|
135
|
-
const options = normalizePopulateOptions(raw.options);
|
|
136
|
-
if (options) normalized.options = options;
|
|
137
|
-
return normalized;
|
|
138
|
-
};
|
|
139
|
-
var normalizeRtsPopulateOption = (value) => {
|
|
140
|
-
if (typeof value === "string") return value.trim() || void 0;
|
|
141
|
-
if (Array.isArray(value)) {
|
|
142
|
-
const normalized = value.map((entry) => {
|
|
143
|
-
if (typeof entry === "string") return entry.trim() || null;
|
|
144
|
-
return normalizePopulateObject(entry) ?? null;
|
|
145
|
-
}).filter((entry) => entry !== null);
|
|
146
|
-
return normalized.length > 0 ? normalized : void 0;
|
|
147
|
-
}
|
|
148
|
-
return normalizePopulateObject(value);
|
|
149
|
-
};
|
|
150
|
-
var normalizeModelName = (value) => {
|
|
151
|
-
if (typeof value !== "string") return null;
|
|
152
|
-
return value.trim() || null;
|
|
153
|
-
};
|
|
154
|
-
var resolvePopulateRefModelName = (model, path, explicitModelName) => {
|
|
155
|
-
if (explicitModelName) return explicitModelName;
|
|
156
|
-
const schema = model.schema;
|
|
157
|
-
const schemaPath = typeof schema.path === "function" ? schema.path(path) : null;
|
|
158
|
-
const directRef = normalizeModelName(schemaPath?.options?.ref);
|
|
159
|
-
if (directRef) return directRef;
|
|
160
|
-
const arrayRef = normalizeModelName(schemaPath?.caster?.options?.ref);
|
|
161
|
-
if (arrayRef) return arrayRef;
|
|
162
|
-
const virtualRef = normalizeModelName((typeof schema.virtualpath === "function" ? schema.virtualpath(path) : null)?.options?.ref);
|
|
163
|
-
if (virtualRef) return virtualRef;
|
|
164
|
-
return null;
|
|
165
|
-
};
|
|
166
|
-
var mergePopulateMatchWithAcl = (populateMatch, aclMatch) => {
|
|
167
|
-
if (!populateMatch || Object.keys(populateMatch).length === 0) return aclMatch;
|
|
168
|
-
return { $and: [populateMatch, aclMatch] };
|
|
169
|
-
};
|
|
170
|
-
var resolvePopulateSpecForModel = async ({ tenantId, model, ability, populate, allowInternalModels, modelCache, dependencyModelNames }) => {
|
|
171
|
-
if (!populate) return void 0;
|
|
172
|
-
const getModelCached = async (targetModelName) => {
|
|
173
|
-
const cached = modelCache.get(targetModelName);
|
|
174
|
-
if (cached) return cached;
|
|
175
|
-
const loaded = await getTenantModel(tenantId, targetModelName, ability);
|
|
176
|
-
modelCache.set(targetModelName, loaded);
|
|
177
|
-
return loaded;
|
|
178
|
-
};
|
|
179
|
-
const resolveOne = async (entry, parentModel) => {
|
|
180
|
-
if (typeof entry === "string") {
|
|
181
|
-
const path = entry.trim();
|
|
182
|
-
if (!path) return null;
|
|
183
|
-
const refModelName = resolvePopulateRefModelName(parentModel, path, null);
|
|
184
|
-
if (!refModelName) return path;
|
|
185
|
-
if (!allowInternalModels && INTERNAL_MODEL_NAMES.has(refModelName)) throw new Error("Model not allowed");
|
|
186
|
-
if (!ability.can("read", refModelName)) throw new Error("forbidden");
|
|
187
|
-
dependencyModelNames.add(refModelName);
|
|
188
|
-
return {
|
|
189
|
-
path,
|
|
190
|
-
match: getAccessibleByQuery(ability, "read", refModelName)
|
|
191
|
-
};
|
|
192
|
-
}
|
|
193
|
-
const path = entry.path.trim();
|
|
194
|
-
if (!path) return null;
|
|
195
|
-
const explicitModelName = normalizeModelName(entry.model);
|
|
196
|
-
const refModelName = resolvePopulateRefModelName(parentModel, path, explicitModelName);
|
|
197
|
-
let nestedModel = parentModel;
|
|
198
|
-
const normalizedEntry = { path };
|
|
199
|
-
if (entry.select !== void 0) normalizedEntry.select = entry.select;
|
|
200
|
-
if (entry.options !== void 0) normalizedEntry.options = entry.options;
|
|
201
|
-
if (explicitModelName) normalizedEntry.model = explicitModelName;
|
|
202
|
-
if (entry.match !== void 0) normalizedEntry.match = entry.match;
|
|
203
|
-
if (refModelName) {
|
|
204
|
-
if (!allowInternalModels && INTERNAL_MODEL_NAMES.has(refModelName)) throw new Error("Model not allowed");
|
|
205
|
-
if (!ability.can("read", refModelName)) throw new Error("forbidden");
|
|
206
|
-
dependencyModelNames.add(refModelName);
|
|
207
|
-
nestedModel = await getModelCached(refModelName);
|
|
208
|
-
const aclMatch = getAccessibleByQuery(ability, "read", refModelName);
|
|
209
|
-
normalizedEntry.match = mergePopulateMatchWithAcl(normalizedEntry.match, aclMatch);
|
|
210
|
-
} else if (entry.populate !== void 0) throw new Error("Populate path must reference a model when nested populate is used");
|
|
211
|
-
const nestedPopulate = await resolvePopulateSpecForModel({
|
|
212
|
-
tenantId,
|
|
213
|
-
model: nestedModel,
|
|
214
|
-
ability,
|
|
215
|
-
populate: entry.populate,
|
|
216
|
-
allowInternalModels,
|
|
217
|
-
modelCache,
|
|
218
|
-
dependencyModelNames
|
|
219
|
-
});
|
|
220
|
-
if (nestedPopulate !== void 0) normalizedEntry.populate = nestedPopulate;
|
|
221
|
-
return normalizedEntry;
|
|
222
|
-
};
|
|
223
|
-
if (Array.isArray(populate)) {
|
|
224
|
-
const filtered = (await Promise.all(populate.map((entry) => resolveOne(entry, model)))).filter((entry) => entry !== null);
|
|
225
|
-
return filtered.length > 0 ? filtered : void 0;
|
|
226
|
-
}
|
|
227
|
-
return await resolveOne(populate, model) ?? void 0;
|
|
228
|
-
};
|
|
229
|
-
var normalizeRtsQueryOptions = (options) => {
|
|
230
|
-
if (!options || typeof options !== "object") return {};
|
|
231
|
-
const normalized = {};
|
|
232
|
-
if (options.projection && typeof options.projection === "object" && !Array.isArray(options.projection)) normalized.projection = options.projection;
|
|
233
|
-
if (options.sort && typeof options.sort === "object" && !Array.isArray(options.sort)) normalized.sort = options.sort;
|
|
234
|
-
normalized.limit = normalizeLimit(options.limit);
|
|
235
|
-
normalized.populate = normalizeRtsPopulateOption(options.populate);
|
|
236
|
-
normalized.pagination = normalizePagination(options.pagination);
|
|
237
|
-
return normalized;
|
|
238
|
-
};
|
|
239
|
-
var resolveRtsQueryDependencyModelNames = async ({ tenantId, ability, modelName, options, allowInternalModels = false }) => {
|
|
240
|
-
const model = await getTenantModel(tenantId, modelName, ability);
|
|
241
|
-
const modelCache = /* @__PURE__ */ new Map();
|
|
242
|
-
modelCache.set(modelName, model);
|
|
243
|
-
const dependencyModelNames = /* @__PURE__ */ new Set();
|
|
244
|
-
await resolvePopulateSpecForModel({
|
|
245
|
-
tenantId,
|
|
246
|
-
model,
|
|
247
|
-
ability,
|
|
248
|
-
populate: options.populate,
|
|
249
|
-
allowInternalModels,
|
|
250
|
-
modelCache,
|
|
251
|
-
dependencyModelNames
|
|
252
|
-
});
|
|
253
|
-
return Array.from(dependencyModelNames);
|
|
254
|
-
};
|
|
255
|
-
var runRtsQuery = async ({ tenantId, ability, modelName, query, options, allowInternalModels = false }) => {
|
|
256
|
-
const { model, finalQuery } = await prepareRtsExecution({
|
|
257
|
-
tenantId,
|
|
258
|
-
ability,
|
|
259
|
-
modelName,
|
|
260
|
-
query,
|
|
261
|
-
allowInternalModels
|
|
262
|
-
});
|
|
263
|
-
const projection = options.projection ?? void 0;
|
|
264
|
-
const sort = options.sort;
|
|
265
|
-
const limit = normalizeLimit(options.limit);
|
|
266
|
-
const modelCache = /* @__PURE__ */ new Map();
|
|
267
|
-
modelCache.set(modelName, model);
|
|
268
|
-
const populate = await resolvePopulateSpecForModel({
|
|
269
|
-
tenantId,
|
|
270
|
-
model,
|
|
271
|
-
ability,
|
|
272
|
-
populate: options.populate,
|
|
273
|
-
allowInternalModels,
|
|
274
|
-
modelCache,
|
|
275
|
-
dependencyModelNames: /* @__PURE__ */ new Set()
|
|
276
|
-
});
|
|
277
|
-
if (options.pagination) {
|
|
278
|
-
const paginatedQuery = model.find(finalQuery, projection);
|
|
279
|
-
if (populate !== void 0) paginatedQuery.populate(populate);
|
|
280
|
-
const paginatedResult = await paginatedQuery.paginate(options.pagination, { cursor: { signingSecret: getPaginationCursorSigningSecret() } });
|
|
281
|
-
const totalCount = typeof paginatedResult.totalCount === "number" && Number.isFinite(paginatedResult.totalCount) && paginatedResult.totalCount >= 0 ? Math.floor(paginatedResult.totalCount) : void 0;
|
|
282
|
-
return {
|
|
283
|
-
data: Array.isArray(paginatedResult.nodes) ? paginatedResult.nodes : [],
|
|
284
|
-
pageInfo: paginatedResult.pageInfo,
|
|
285
|
-
...totalCount !== void 0 ? { totalCount } : {}
|
|
286
|
-
};
|
|
287
|
-
}
|
|
288
|
-
const queryPromise = model.find(finalQuery, projection);
|
|
289
|
-
if (populate !== void 0) queryPromise.populate(populate);
|
|
290
|
-
if (sort && Object.keys(sort).length) queryPromise.sort(sort);
|
|
291
|
-
queryPromise.limit(limit);
|
|
292
|
-
const data = await queryPromise;
|
|
293
|
-
return { data: Array.isArray(data) ? data : [] };
|
|
294
|
-
};
|
|
295
|
-
var prepareRtsExecution = async ({ tenantId, ability, modelName, query, allowInternalModels = false }) => {
|
|
296
|
-
if (!allowInternalModels && INTERNAL_MODEL_NAMES.has(modelName)) throw new Error("Model not allowed");
|
|
297
|
-
if (!ability.can("read", modelName)) throw new Error("forbidden");
|
|
298
|
-
return {
|
|
299
|
-
model: await getTenantModel(tenantId, modelName, ability),
|
|
300
|
-
finalQuery: { $and: [query, getAccessibleByQuery(ability, "read", modelName)] }
|
|
301
|
-
};
|
|
302
|
-
};
|
|
303
|
-
var runRtsCount = async ({ tenantId, ability, modelName, query, allowInternalModels = false }) => {
|
|
304
|
-
const { model, finalQuery } = await prepareRtsExecution({
|
|
305
|
-
tenantId,
|
|
306
|
-
ability,
|
|
307
|
-
modelName,
|
|
308
|
-
query,
|
|
309
|
-
allowInternalModels
|
|
310
|
-
});
|
|
311
|
-
const count = await model.countDocuments(finalQuery);
|
|
312
|
-
if (typeof count !== "number" || !Number.isFinite(count) || count < 0) return 0;
|
|
313
|
-
return Math.floor(count);
|
|
314
|
-
};
|
|
315
|
-
//#endregion
|
|
316
|
-
export { normalizeRtsQueryOptions as a, runRtsCount as c, isRtsRequestAuthorized as i, runRtsQuery as l, RTS_USER_ID_HEADER as n, resolveRtsQueryDependencyModelNames as o, buildRtsAbilityFromRequest as r, resolveRtsRequestTenantId as s, RTS_TENANT_ID_QUERY_PARAM as t, getDerivedKey as u };
|
|
317
|
-
|
|
318
|
-
//# sourceMappingURL=queryExecutor-JadZcQSQ.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"queryExecutor-JadZcQSQ.js","names":["assert","hkdfSync","getDerivedKey","masterKey","info","length","salt","Buffer","from","toString","Request","PaginationPageInfo","PaginationSpec","models","LoadModelCtx","buildAbility","buildAbilityFromSession","getAccessibleByQuery","getTenantRolesFromSessionUser","AclSubjectType","AppAbility","Model","getDerivedKey","JsonObject","Record","SessionUser","id","currentTenantId","signedInTenants","HeaderUserDoc","tenants","tenantRoles","RtsPopulateObject","path","model","select","match","options","sort","limit","populate","RtsPopulateOption","Array","RtsQueryOptions","projection","pagination","RtsQueryResult","data","pageInfo","totalCount","PreparedRtsExecution","finalQuery","RTS_TENANT_ID_QUERY_PARAM","RTS_USER_ID_HEADER","QUERY_MAX_LIMIT","INTERNAL_MODEL_NAMES","Set","paginationCursorSigningSecret","getPaginationCursorSigningSecret","masterKey","process","env","MASTER_KEY","trim","Error","normalizeTenantId","value","normalized","normalizeSignedInTenants","isArray","map","tenantId","String","filter","Boolean","getTenantIdFromRequest","req","rawQuery","query","queryTenantId","normalizedFromQuery","session","user","resolveRtsRequestTenantId","resolveRtsRequestUserId","sessionUserId","headerValue","headers","headerUserId","isRtsRequestAuthorized","sessionUser","length","includes","buildRtsAbilityFromRequest","Promise","ability","userId","headerUserIdRaw","rbCtx","User","getGlobal","findById","lean","tenantsRaw","tenant","roles","getTenantModel","modelName","ctx","get","normalizeLimit","Number","isFinite","Math","min","abs","normalizeString","normalizeObject","undefined","normalizePagination","normalizePopulateSelect","normalizePopulateOptions","raw","max","floor","normalizePopulateObject","nestedPopulate","normalizeRtsPopulateOption","entry","normalizeModelName","resolvePopulateRefModelName","explicitModelName","schema","schemaPath","directRef","ref","arrayRef","caster","virtualPath","virtualpath","virtualRef","mergePopulateMatchWithAcl","populateMatch","aclMatch","Object","keys","$and","PreparedPopulateObject","PreparedPopulateOption","resolvePopulateSpecForModel","allowInternalModels","modelCache","dependencyModelNames","Map","getModelCached","targetModelName","cached","loaded","set","resolveOne","parentModel","refModelName","has","can","add","Exclude","nestedModel","normalizedEntry","resolved","all","filtered","normalizeRtsQueryOptions","resolveRtsQueryDependencyModelNames","from","runRtsQuery","prepareRtsExecution","paginatedQuery","find","paginatedResult","paginate","cursor","signingSecret","nodes","queryPromise","accessQuery","runRtsCount","count","countDocuments"],"sources":["../src/getDerivedKey.ts","../src/rts/queryExecutor.ts"],"sourcesContent":["import assert from \"assert\"\nimport { hkdfSync } from \"crypto\"\n\n\nexport const getDerivedKey = (\n masterKey: string,\n info: string,\n length: number = 32, // Default to 256-bit keys\n salt: string = \"\",\n): string => {\n assert(masterKey?.length >= 32, \"MASTER_KEY must be 32 chars or longer.\")\n\n return Buffer.from(hkdfSync(\n \"sha256\",\n masterKey,\n Buffer.from(salt),\n Buffer.from(info),\n length,\n )).toString(\"hex\")\n}\n","import type { Request } from \"express\"\nimport type { PaginationPageInfo, PaginationSpec } from \"@rpcbase/api\"\nimport { models, type LoadModelCtx } from \"@rpcbase/db\"\nimport { buildAbility, buildAbilityFromSession, getAccessibleByQuery, getTenantRolesFromSessionUser, type AclSubjectType, type AppAbility } from \"@rpcbase/db/acl\"\nimport type { Model } from \"mongoose\"\n\nimport { getDerivedKey } from \"../getDerivedKey\"\n\n\ntype JsonObject = Record<string, unknown>\n\ntype SessionUser = {\n id?: unknown\n currentTenantId?: unknown\n signedInTenants?: unknown\n}\n\ntype HeaderUserDoc = {\n tenants?: unknown\n tenantRoles?: unknown\n}\n\nexport type RtsPopulateObject = {\n path: string\n model?: string\n select?: string | JsonObject\n match?: JsonObject\n options?: {\n sort?: Record<string, 1 | -1>\n limit?: number\n }\n populate?: RtsPopulateOption\n}\n\nexport type RtsPopulateOption =\n | string\n | RtsPopulateObject\n | Array<string | RtsPopulateObject>\n\nexport type RtsQueryOptions = {\n projection?: JsonObject\n sort?: Record<string, 1 | -1>\n limit?: number\n populate?: RtsPopulateOption\n pagination?: PaginationSpec\n}\n\nexport type RtsQueryResult = {\n data: unknown[]\n pageInfo?: PaginationPageInfo\n totalCount?: number\n}\n\ntype PreparedRtsExecution = {\n model: Model<any>\n finalQuery: JsonObject\n}\n\nexport const RTS_TENANT_ID_QUERY_PARAM = \"rb-tenant-id\"\nexport const RTS_USER_ID_HEADER = \"rb-user-id\"\n\nconst QUERY_MAX_LIMIT = 4096\nconst INTERNAL_MODEL_NAMES = new Set([\"RBRtsChange\", \"RBRtsCounter\"])\nlet paginationCursorSigningSecret: string | null = null\n\nconst getPaginationCursorSigningSecret = (): string => {\n if (paginationCursorSigningSecret) return paginationCursorSigningSecret\n const masterKey = process.env.MASTER_KEY?.trim()\n if (!masterKey) {\n throw new Error(\"MASTER_KEY must be defined to derive pagination cursor signing secret\")\n }\n paginationCursorSigningSecret = getDerivedKey(masterKey, \"pagination_cursor_signing\")\n return paginationCursorSigningSecret\n}\n\nconst normalizeTenantId = (value: unknown): string | null => {\n if (typeof value !== \"string\") return null\n const normalized = value.trim()\n return normalized ? normalized : null\n}\n\nconst normalizeSignedInTenants = (value: unknown): string[] => {\n if (!Array.isArray(value)) return []\n return value\n .map((tenantId) => normalizeTenantId(String(tenantId)))\n .filter((tenantId): tenantId is string => Boolean(tenantId))\n}\n\nconst getTenantIdFromRequest = (req: Request): string | null => {\n const rawQuery = req.query?.[RTS_TENANT_ID_QUERY_PARAM]\n const queryTenantId = Array.isArray(rawQuery) ? rawQuery[0] : rawQuery\n const normalizedFromQuery = normalizeTenantId(queryTenantId)\n if (normalizedFromQuery) return normalizedFromQuery\n\n return normalizeTenantId((req.session?.user as SessionUser | undefined)?.currentTenantId)\n}\n\nexport const resolveRtsRequestTenantId = (req: Request): string | null => {\n return getTenantIdFromRequest(req)\n}\n\nexport const resolveRtsRequestUserId = (req: Request): string | null => {\n const sessionUserId = normalizeTenantId((req.session?.user as SessionUser | undefined)?.id)\n if (sessionUserId) return sessionUserId\n\n const headerValue = req.headers[RTS_USER_ID_HEADER]\n const headerUserId = Array.isArray(headerValue) ? headerValue[0] : headerValue\n return normalizeTenantId(headerUserId)\n}\n\nexport const isRtsRequestAuthorized = (req: Request, tenantId: string): boolean => {\n const sessionUser = req.session?.user as SessionUser | undefined\n if (!sessionUser) return false\n\n const signedInTenants = normalizeSignedInTenants(sessionUser.signedInTenants)\n if (signedInTenants.length > 0) {\n return signedInTenants.includes(tenantId)\n }\n\n const currentTenantId = normalizeTenantId(sessionUser.currentTenantId)\n if (!currentTenantId) return false\n return currentTenantId === tenantId\n}\n\nexport const buildRtsAbilityFromRequest = async (\n req: Request,\n tenantId: string,\n): Promise<{ ability: AppAbility; userId: string | null }> => {\n const sessionUserId = normalizeTenantId((req.session?.user as SessionUser | undefined)?.id)\n if (sessionUserId) {\n const ability = buildAbilityFromSession({ tenantId, session: req.session })\n return { ability, userId: sessionUserId }\n }\n\n const headerValue = req.headers[RTS_USER_ID_HEADER]\n const headerUserIdRaw = Array.isArray(headerValue) ? headerValue[0] : headerValue\n const headerUserId = normalizeTenantId(headerUserIdRaw)\n if (!headerUserId) {\n const ability = buildAbilityFromSession({ tenantId, session: req.session })\n return { ability, userId: null }\n }\n\n const rbCtx: LoadModelCtx = { req: { session: null } }\n const User = await models.getGlobal(\"RBUser\", rbCtx)\n const user = await User.findById(headerUserId, { tenants: 1, tenantRoles: 1 }).lean() as HeaderUserDoc | null\n\n const tenantsRaw = user?.tenants\n const tenants = Array.isArray(tenantsRaw) ? tenantsRaw.map((tenant) => String(tenant)) : []\n if (!tenants.includes(tenantId)) {\n throw new Error(\"Tenant not authorized for this session\")\n }\n\n const roles = getTenantRolesFromSessionUser(user, tenantId)\n return {\n ability: buildAbility({ tenantId, userId: headerUserId, roles: roles.length ? roles : [\"owner\"] }),\n userId: headerUserId,\n }\n}\n\nconst getTenantModel = async (tenantId: string, modelName: string, ability: AppAbility): Promise<Model<any>> => {\n const ctx: LoadModelCtx = {\n req: {\n session: {\n user: {\n currentTenantId: tenantId,\n },\n },\n },\n ability,\n }\n\n return models.get(modelName, ctx)\n}\n\nconst normalizeLimit = (limit?: number): number => {\n if (typeof limit !== \"number\") return QUERY_MAX_LIMIT\n if (!Number.isFinite(limit)) return QUERY_MAX_LIMIT\n return Math.min(QUERY_MAX_LIMIT, Math.abs(limit))\n}\n\nconst normalizeString = (value: unknown): string => {\n return typeof value === \"string\" ? value.trim() : \"\"\n}\n\nconst normalizeObject = (value: unknown): JsonObject | undefined => {\n if (!value || typeof value !== \"object\" || Array.isArray(value)) return undefined\n return value as JsonObject\n}\n\nconst normalizePagination = (value: unknown): PaginationSpec | undefined => {\n if (!value || typeof value !== \"object\" || Array.isArray(value)) return undefined\n return value as PaginationSpec\n}\n\nconst normalizePopulateSelect = (value: unknown): string | JsonObject | undefined => {\n if (typeof value === \"string\") {\n const normalized = value.trim()\n return normalized || undefined\n }\n return normalizeObject(value)\n}\n\nconst normalizePopulateOptions = (value: unknown): RtsPopulateObject[\"options\"] | undefined => {\n if (!value || typeof value !== \"object\" || Array.isArray(value)) return undefined\n const raw = value as { sort?: unknown; limit?: unknown }\n const normalized: RtsPopulateObject[\"options\"] = {}\n\n if (raw.sort && typeof raw.sort === \"object\" && !Array.isArray(raw.sort)) {\n normalized.sort = raw.sort as Record<string, 1 | -1>\n }\n\n if (typeof raw.limit === \"number\" && Number.isFinite(raw.limit)) {\n normalized.limit = Math.max(0, Math.floor(Math.abs(raw.limit)))\n }\n\n if (!normalized.sort && normalized.limit === undefined) return undefined\n return normalized\n}\n\nconst normalizePopulateObject = (value: unknown): RtsPopulateObject | undefined => {\n if (!value || typeof value !== \"object\" || Array.isArray(value)) return undefined\n const raw = value as Record<string, unknown>\n const path = normalizeString(raw.path)\n if (!path) return undefined\n\n const normalized: RtsPopulateObject = { path }\n\n const model = normalizeString(raw.model)\n if (model) normalized.model = model\n\n const select = normalizePopulateSelect(raw.select)\n if (select !== undefined) normalized.select = select\n\n const match = normalizeObject(raw.match)\n if (match) normalized.match = match\n\n const nestedPopulate = normalizeRtsPopulateOption(raw.populate)\n if (nestedPopulate !== undefined) normalized.populate = nestedPopulate\n\n const options = normalizePopulateOptions(raw.options)\n if (options) normalized.options = options\n\n return normalized\n}\n\nconst normalizeRtsPopulateOption = (value: unknown): RtsPopulateOption | undefined => {\n if (typeof value === \"string\") {\n const normalized = value.trim()\n return normalized || undefined\n }\n\n if (Array.isArray(value)) {\n const normalized = value\n .map((entry) => {\n if (typeof entry === \"string\") {\n const path = entry.trim()\n return path || null\n }\n return normalizePopulateObject(entry) ?? null\n })\n .filter((entry): entry is string | RtsPopulateObject => entry !== null)\n\n return normalized.length > 0 ? normalized : undefined\n }\n\n return normalizePopulateObject(value)\n}\n\nconst normalizeModelName = (value: unknown): string | null => {\n if (typeof value !== \"string\") return null\n const normalized = value.trim()\n return normalized || null\n}\n\nconst resolvePopulateRefModelName = (\n model: Model<any>,\n path: string,\n explicitModelName: string | null,\n): string | null => {\n if (explicitModelName) return explicitModelName\n\n const schema = model.schema as any\n const schemaPath = typeof schema.path === \"function\" ? schema.path(path) : null\n const directRef = normalizeModelName(schemaPath?.options?.ref)\n if (directRef) return directRef\n\n const arrayRef = normalizeModelName(schemaPath?.caster?.options?.ref)\n if (arrayRef) return arrayRef\n\n const virtualPath = typeof schema.virtualpath === \"function\" ? schema.virtualpath(path) : null\n const virtualRef = normalizeModelName(virtualPath?.options?.ref)\n if (virtualRef) return virtualRef\n\n return null\n}\n\nconst mergePopulateMatchWithAcl = (\n populateMatch: JsonObject | undefined,\n aclMatch: JsonObject,\n): JsonObject => {\n if (!populateMatch || Object.keys(populateMatch).length === 0) return aclMatch\n return { $and: [populateMatch, aclMatch] }\n}\n\ntype PreparedPopulateObject = {\n path: string\n model?: string\n select?: string | JsonObject\n match?: JsonObject\n options?: {\n sort?: Record<string, 1 | -1>\n limit?: number\n }\n populate?: PreparedPopulateOption\n}\n\ntype PreparedPopulateOption =\n | string\n | PreparedPopulateObject\n | Array<string | PreparedPopulateObject>\n\nconst resolvePopulateSpecForModel = async ({\n tenantId,\n model,\n ability,\n populate,\n allowInternalModels,\n modelCache,\n dependencyModelNames,\n}: {\n tenantId: string\n model: Model<any>\n ability: AppAbility\n populate: RtsPopulateOption | undefined\n allowInternalModels: boolean\n modelCache: Map<string, Model<any>>\n dependencyModelNames: Set<string>\n}): Promise<PreparedPopulateOption | undefined> => {\n if (!populate) return undefined\n\n const getModelCached = async (targetModelName: string): Promise<Model<any>> => {\n const cached = modelCache.get(targetModelName)\n if (cached) return cached\n const loaded = await getTenantModel(tenantId, targetModelName, ability)\n modelCache.set(targetModelName, loaded)\n return loaded\n }\n\n const resolveOne = async (\n entry: string | RtsPopulateObject,\n parentModel: Model<any>,\n ): Promise<string | PreparedPopulateObject | null> => {\n if (typeof entry === \"string\") {\n const path = entry.trim()\n if (!path) return null\n\n const refModelName = resolvePopulateRefModelName(parentModel, path, null)\n if (!refModelName) return path\n if (!allowInternalModels && INTERNAL_MODEL_NAMES.has(refModelName)) {\n throw new Error(\"Model not allowed\")\n }\n if (!ability.can(\"read\", refModelName as AclSubjectType)) {\n throw new Error(\"forbidden\")\n }\n\n dependencyModelNames.add(refModelName)\n\n const aclMatch = getAccessibleByQuery(\n ability,\n \"read\",\n refModelName as Exclude<AclSubjectType, \"all\">,\n )\n return {\n path,\n match: aclMatch as JsonObject,\n }\n }\n\n const path = entry.path.trim()\n if (!path) return null\n\n const explicitModelName = normalizeModelName(entry.model)\n const refModelName = resolvePopulateRefModelName(parentModel, path, explicitModelName)\n let nestedModel = parentModel\n\n const normalizedEntry: PreparedPopulateObject = {\n path,\n }\n\n if (entry.select !== undefined) normalizedEntry.select = entry.select\n if (entry.options !== undefined) normalizedEntry.options = entry.options\n if (explicitModelName) normalizedEntry.model = explicitModelName\n if (entry.match !== undefined) normalizedEntry.match = entry.match\n\n if (refModelName) {\n if (!allowInternalModels && INTERNAL_MODEL_NAMES.has(refModelName)) {\n throw new Error(\"Model not allowed\")\n }\n if (!ability.can(\"read\", refModelName as AclSubjectType)) {\n throw new Error(\"forbidden\")\n }\n\n dependencyModelNames.add(refModelName)\n nestedModel = await getModelCached(refModelName)\n\n const aclMatch = getAccessibleByQuery(\n ability,\n \"read\",\n refModelName as Exclude<AclSubjectType, \"all\">,\n ) as JsonObject\n normalizedEntry.match = mergePopulateMatchWithAcl(\n normalizedEntry.match,\n aclMatch,\n )\n } else if (entry.populate !== undefined) {\n throw new Error(\"Populate path must reference a model when nested populate is used\")\n }\n\n const nestedPopulate = await resolvePopulateSpecForModel({\n tenantId,\n model: nestedModel,\n ability,\n populate: entry.populate,\n allowInternalModels,\n modelCache,\n dependencyModelNames,\n })\n if (nestedPopulate !== undefined) normalizedEntry.populate = nestedPopulate\n\n return normalizedEntry\n }\n\n if (Array.isArray(populate)) {\n const resolved = await Promise.all(populate.map((entry) => resolveOne(entry, model)))\n const filtered = resolved.filter((entry): entry is string | PreparedPopulateObject => entry !== null)\n return filtered.length > 0 ? filtered : undefined\n }\n\n const resolved = await resolveOne(populate, model)\n return resolved ?? undefined\n}\n\nexport const normalizeRtsQueryOptions = (options: RtsQueryOptions | undefined): RtsQueryOptions => {\n if (!options || typeof options !== \"object\") return {}\n const normalized: RtsQueryOptions = {}\n\n if (options.projection && typeof options.projection === \"object\" && !Array.isArray(options.projection)) {\n normalized.projection = options.projection\n }\n\n if (options.sort && typeof options.sort === \"object\" && !Array.isArray(options.sort)) {\n normalized.sort = options.sort\n }\n\n normalized.limit = normalizeLimit(options.limit)\n normalized.populate = normalizeRtsPopulateOption(options.populate)\n normalized.pagination = normalizePagination(options.pagination)\n\n return normalized\n}\n\nexport const resolveRtsQueryDependencyModelNames = async ({\n tenantId,\n ability,\n modelName,\n options,\n allowInternalModels = false,\n}: {\n tenantId: string\n ability: AppAbility\n modelName: string\n options: RtsQueryOptions\n allowInternalModels?: boolean\n}): Promise<string[]> => {\n const model = await getTenantModel(tenantId, modelName, ability)\n const modelCache = new Map<string, Model<any>>()\n modelCache.set(modelName, model)\n\n const dependencyModelNames = new Set<string>()\n await resolvePopulateSpecForModel({\n tenantId,\n model,\n ability,\n populate: options.populate,\n allowInternalModels,\n modelCache,\n dependencyModelNames,\n })\n\n return Array.from(dependencyModelNames)\n}\n\nexport const runRtsQuery = async ({\n tenantId,\n ability,\n modelName,\n query,\n options,\n allowInternalModels = false,\n}: {\n tenantId: string\n ability: AppAbility\n modelName: string\n query: JsonObject\n options: RtsQueryOptions\n allowInternalModels?: boolean\n}): Promise<RtsQueryResult> => {\n const { model, finalQuery } = await prepareRtsExecution({\n tenantId,\n ability,\n modelName,\n query,\n allowInternalModels,\n })\n const projection = options.projection ?? undefined\n const sort = options.sort\n const limit = normalizeLimit(options.limit)\n const modelCache = new Map<string, Model<any>>()\n modelCache.set(modelName, model)\n\n const populate = await resolvePopulateSpecForModel({\n tenantId,\n model,\n ability,\n populate: options.populate,\n allowInternalModels,\n modelCache,\n dependencyModelNames: new Set<string>(),\n })\n\n if (options.pagination) {\n const paginatedQuery = model.find(finalQuery, projection)\n if (populate !== undefined) {\n paginatedQuery.populate(populate as any)\n }\n\n const paginatedResult = await paginatedQuery.paginate(options.pagination, {\n cursor: {\n signingSecret: getPaginationCursorSigningSecret(),\n },\n })\n const totalCount = typeof paginatedResult.totalCount === \"number\"\n && Number.isFinite(paginatedResult.totalCount)\n && paginatedResult.totalCount >= 0\n ? Math.floor(paginatedResult.totalCount)\n : undefined\n\n return {\n data: Array.isArray(paginatedResult.nodes) ? paginatedResult.nodes : [],\n pageInfo: paginatedResult.pageInfo,\n ...(totalCount !== undefined ? { totalCount } : {}),\n }\n }\n\n const queryPromise = model.find(finalQuery, projection)\n if (populate !== undefined) {\n queryPromise.populate(populate as any)\n }\n if (sort && Object.keys(sort).length) {\n queryPromise.sort(sort)\n }\n queryPromise.limit(limit)\n\n const data = await queryPromise\n return { data: Array.isArray(data) ? data : [] }\n}\n\nconst prepareRtsExecution = async ({\n tenantId,\n ability,\n modelName,\n query,\n allowInternalModels = false,\n}: {\n tenantId: string\n ability: AppAbility\n modelName: string\n query: JsonObject\n allowInternalModels?: boolean\n}): Promise<PreparedRtsExecution> => {\n if (!allowInternalModels && INTERNAL_MODEL_NAMES.has(modelName)) {\n throw new Error(\"Model not allowed\")\n }\n\n if (!ability.can(\"read\", modelName as AclSubjectType)) {\n throw new Error(\"forbidden\")\n }\n\n const model = await getTenantModel(tenantId, modelName, ability)\n const accessQuery = getAccessibleByQuery(ability, \"read\", modelName as Exclude<AclSubjectType, \"all\">)\n const finalQuery: JsonObject = { $and: [query, accessQuery] }\n\n return { model, finalQuery }\n}\n\nexport const runRtsCount = async ({\n tenantId,\n ability,\n modelName,\n query,\n allowInternalModels = false,\n}: {\n tenantId: string\n ability: AppAbility\n modelName: string\n query: JsonObject\n allowInternalModels?: boolean\n}): Promise<number> => {\n const { model, finalQuery } = await prepareRtsExecution({\n tenantId,\n ability,\n modelName,\n query,\n allowInternalModels,\n })\n\n const count = await model.countDocuments(finalQuery)\n if (typeof count !== \"number\" || !Number.isFinite(count) || count < 0) {\n return 0\n }\n\n return Math.floor(count)\n}\n"],"mappings":";;;;;AAIA,IAAaE,iBACXC,WACAC,MACAC,SAAiB,IACjBC,OAAe,OACJ;AACXN,QAAOG,WAAWE,UAAU,IAAI,yCAAyC;AAEzE,QAAOE,OAAOC,KAAKP,SACjB,UACAE,WACAI,OAAOC,KAAKF,KAAK,EACjBC,OAAOC,KAAKJ,KAAK,EACjBC,OACD,CAAC,CAACI,SAAS,MAAM;;;;ACwCpB,IAAa2C,4BAA4B;AACzC,IAAaC,qBAAqB;AAElC,IAAMC,kBAAkB;AACxB,IAAMC,uBAAuB,IAAIC,IAAI,CAAC,eAAe,eAAe,CAAC;AACrE,IAAIC,gCAA+C;AAEnD,IAAMC,yCAAiD;AACrD,KAAID,8BAA+B,QAAOA;CAC1C,MAAME,YAAYC,QAAQC,IAAIC,YAAYC,MAAM;AAChD,KAAI,CAACJ,UACH,OAAM,IAAIK,MAAM,wEAAwE;AAE1FP,iCAAgCnC,cAAcqC,WAAW,4BAA4B;AACrF,QAAOF;;AAGT,IAAMQ,qBAAqBC,UAAkC;AAC3D,KAAI,OAAOA,UAAU,SAAU,QAAO;CACtC,MAAMC,aAAaD,MAAMH,MAAM;AAC/B,QAAOI,aAAaA,aAAa;;AAGnC,IAAMC,4BAA4BF,UAA6B;AAC7D,KAAI,CAACxB,MAAM2B,QAAQH,MAAM,CAAE,QAAO,EAAE;AACpC,QAAOA,MACJI,KAAKC,aAAaN,kBAAkBO,OAAOD,SAAS,CAAC,CAAC,CACtDE,QAAQF,aAAiCG,QAAQH,SAAS,CAAC;;AAGhE,IAAMI,0BAA0BC,QAAgC;CAC9D,MAAMC,WAAWD,IAAIE,QAAQ1B;CAE7B,MAAM4B,sBAAsBf,kBADNvB,MAAM2B,QAAQQ,SAAS,GAAGA,SAAS,KAAKA,SACF;AAC5D,KAAIG,oBAAqB,QAAOA;AAEhC,QAAOf,mBAAmBW,IAAIK,SAASC,OAAkCvD,gBAAgB;;AAG3F,IAAawD,6BAA6BP,QAAgC;AACxE,QAAOD,uBAAuBC,IAAI;;AAYpC,IAAaa,0BAA0Bb,KAAcL,aAA8B;CACjF,MAAMmB,cAAcd,IAAIK,SAASC;AACjC,KAAI,CAACQ,YAAa,QAAO;CAEzB,MAAM9D,kBAAkBwC,yBAAyBsB,YAAY9D,gBAAgB;AAC7E,KAAIA,gBAAgB+D,SAAS,EAC3B,QAAO/D,gBAAgBgE,SAASrB,SAAS;CAG3C,MAAM5C,kBAAkBsC,kBAAkByB,YAAY/D,gBAAgB;AACtE,KAAI,CAACA,gBAAiB,QAAO;AAC7B,QAAOA,oBAAoB4C;;AAG7B,IAAasB,6BAA6B,OACxCjB,KACAL,aAC4D;CAC5D,MAAMc,gBAAgBpB,mBAAmBW,IAAIK,SAASC,OAAkCxD,GAAG;AAC3F,KAAI2D,cAEF,QAAO;EAAEU,SADO/E,wBAAwB;GAAEuD;GAAUU,SAASL,IAAIK;GAAS,CAAC;EACzDe,QAAQX;EAAe;CAG3C,MAAMC,cAAcV,IAAIW,QAAQlC;CAEhC,MAAMmC,eAAevB,kBADGvB,MAAM2B,QAAQiB,YAAY,GAAGA,YAAY,KAAKA,YACf;AACvD,KAAI,CAACE,aAEH,QAAO;EAAEO,SADO/E,wBAAwB;GAAEuD;GAAUU,SAASL,IAAIK;GAAS,CAAC;EACzDe,QAAQ;EAAM;CAKlC,MAAMd,OAAO,OADA,MAAMrE,OAAOuF,UAAU,UADR,EAAExB,KAAK,EAAEK,SAAS,MAAK,EAAG,CACF,EAC5BoB,SAASb,cAAc;EAAE1D,SAAS;EAAGC,aAAa;EAAG,CAAC,CAACuE,MAAM;CAErF,MAAMC,aAAarB,MAAMpD;AAEzB,KAAI,EADYY,MAAM2B,QAAQkC,WAAW,GAAGA,WAAWjC,KAAKkC,WAAWhC,OAAOgC,OAAO,CAAC,GAAG,EAAE,EAC9EZ,SAASrB,SAAS,CAC7B,OAAM,IAAIP,MAAM,yCAAyC;CAG3D,MAAMyC,QAAQvF,8BAA8BgE,MAAMX,SAAS;AAC3D,QAAO;EACLwB,SAAShF,aAAa;GAAEwD;GAAUyB,QAAQR;GAAciB,OAAOA,MAAMd,SAASc,QAAQ,CAAC,QAAO;GAAG,CAAC;EAClGT,QAAQR;EACT;;AAGH,IAAMkB,iBAAiB,OAAOnC,UAAkBoC,WAAmBZ,YAA6C;CAC9G,MAAMa,MAAoB;EACxBhC,KAAK,EACHK,SAAS,EACPC,MAAM,EACJvD,iBAAiB4C,UACnB,EACF,EACD;EACDwB;EACD;AAED,QAAOlF,OAAOgG,IAAIF,WAAWC,IAAI;;AAGnC,IAAME,kBAAkBvE,UAA2B;AACjD,KAAI,OAAOA,UAAU,SAAU,QAAOe;AACtC,KAAI,CAACyD,OAAOC,SAASzE,MAAM,CAAE,QAAOe;AACpC,QAAO2D,KAAKC,IAAI5D,iBAAiB2D,KAAKE,IAAI5E,MAAM,CAAC;;AAGnD,IAAM6E,mBAAmBlD,UAA2B;AAClD,QAAO,OAAOA,UAAU,WAAWA,MAAMH,MAAM,GAAG;;AAGpD,IAAMsD,mBAAmBnD,UAA2C;AAClE,KAAI,CAACA,SAAS,OAAOA,UAAU,YAAYxB,MAAM2B,QAAQH,MAAM,CAAE,QAAOoD,KAAAA;AACxE,QAAOpD;;AAGT,IAAMqD,uBAAuBrD,UAA+C;AAC1E,KAAI,CAACA,SAAS,OAAOA,UAAU,YAAYxB,MAAM2B,QAAQH,MAAM,CAAE,QAAOoD,KAAAA;AACxE,QAAOpD;;AAGT,IAAMsD,2BAA2BtD,UAAoD;AACnF,KAAI,OAAOA,UAAU,SAEnB,QADmBA,MAAMH,MAAM,IACVuD,KAAAA;AAEvB,QAAOD,gBAAgBnD,MAAM;;AAG/B,IAAMuD,4BAA4BvD,UAA6D;AAC7F,KAAI,CAACA,SAAS,OAAOA,UAAU,YAAYxB,MAAM2B,QAAQH,MAAM,CAAE,QAAOoD,KAAAA;CACxE,MAAMI,MAAMxD;CACZ,MAAMC,aAA2C,EAAE;AAEnD,KAAIuD,IAAIpF,QAAQ,OAAOoF,IAAIpF,SAAS,YAAY,CAACI,MAAM2B,QAAQqD,IAAIpF,KAAK,CACtE6B,YAAW7B,OAAOoF,IAAIpF;AAGxB,KAAI,OAAOoF,IAAInF,UAAU,YAAYwE,OAAOC,SAASU,IAAInF,MAAM,CAC7D4B,YAAW5B,QAAQ0E,KAAKU,IAAI,GAAGV,KAAKW,MAAMX,KAAKE,IAAIO,IAAInF,MAAM,CAAC,CAAC;AAGjE,KAAI,CAAC4B,WAAW7B,QAAQ6B,WAAW5B,UAAU+E,KAAAA,EAAW,QAAOA,KAAAA;AAC/D,QAAOnD;;AAGT,IAAM0D,2BAA2B3D,UAAkD;AACjF,KAAI,CAACA,SAAS,OAAOA,UAAU,YAAYxB,MAAM2B,QAAQH,MAAM,CAAE,QAAOoD,KAAAA;CACxE,MAAMI,MAAMxD;CACZ,MAAMjC,OAAOmF,gBAAgBM,IAAIzF,KAAK;AACtC,KAAI,CAACA,KAAM,QAAOqF,KAAAA;CAElB,MAAMnD,aAAgC,EAAElC,MAAM;CAE9C,MAAMC,QAAQkF,gBAAgBM,IAAIxF,MAAM;AACxC,KAAIA,MAAOiC,YAAWjC,QAAQA;CAE9B,MAAMC,SAASqF,wBAAwBE,IAAIvF,OAAO;AAClD,KAAIA,WAAWmF,KAAAA,EAAWnD,YAAWhC,SAASA;CAE9C,MAAMC,QAAQiF,gBAAgBK,IAAItF,MAAM;AACxC,KAAIA,MAAO+B,YAAW/B,QAAQA;CAE9B,MAAM0F,iBAAiBC,2BAA2BL,IAAIlF,SAAS;AAC/D,KAAIsF,mBAAmBR,KAAAA,EAAWnD,YAAW3B,WAAWsF;CAExD,MAAMzF,UAAUoF,yBAAyBC,IAAIrF,QAAQ;AACrD,KAAIA,QAAS8B,YAAW9B,UAAUA;AAElC,QAAO8B;;AAGT,IAAM4D,8BAA8B7D,UAAkD;AACpF,KAAI,OAAOA,UAAU,SAEnB,QADmBA,MAAMH,MAAM,IACVuD,KAAAA;AAGvB,KAAI5E,MAAM2B,QAAQH,MAAM,EAAE;EACxB,MAAMC,aAAaD,MAChBI,KAAK0D,UAAU;AACd,OAAI,OAAOA,UAAU,SAEnB,QADaA,MAAMjE,MAAM,IACV;AAEjB,UAAO8D,wBAAwBG,MAAM,IAAI;IACzC,CACDvD,QAAQuD,UAA+CA,UAAU,KAAK;AAEzE,SAAO7D,WAAWwB,SAAS,IAAIxB,aAAamD,KAAAA;;AAG9C,QAAOO,wBAAwB3D,MAAM;;AAGvC,IAAM+D,sBAAsB/D,UAAkC;AAC5D,KAAI,OAAOA,UAAU,SAAU,QAAO;AAEtC,QADmBA,MAAMH,MAAM,IACV;;AAGvB,IAAMmE,+BACJhG,OACAD,MACAkG,sBACkB;AAClB,KAAIA,kBAAmB,QAAOA;CAE9B,MAAMC,SAASlG,MAAMkG;CACrB,MAAMC,aAAa,OAAOD,OAAOnG,SAAS,aAAamG,OAAOnG,KAAKA,KAAK,GAAG;CAC3E,MAAMqG,YAAYL,mBAAmBI,YAAYhG,SAASkG,IAAI;AAC9D,KAAID,UAAW,QAAOA;CAEtB,MAAME,WAAWP,mBAAmBI,YAAYI,QAAQpG,SAASkG,IAAI;AACrE,KAAIC,SAAU,QAAOA;CAGrB,MAAMI,aAAaX,oBADC,OAAOG,OAAOO,gBAAgB,aAAaP,OAAOO,YAAY1G,KAAK,GAAG,OACvCI,SAASkG,IAAI;AAChE,KAAIK,WAAY,QAAOA;AAEvB,QAAO;;AAGT,IAAMC,6BACJC,eACAC,aACe;AACf,KAAI,CAACD,iBAAiBE,OAAOC,KAAKH,cAAc,CAACnD,WAAW,EAAG,QAAOoD;AACtE,QAAO,EAAEG,MAAM,CAACJ,eAAeC,SAAQ,EAAG;;AAoB5C,IAAMM,8BAA8B,OAAO,EACzC9E,UACArC,OACA6D,SACAvD,UACA8G,qBACAC,YACAC,2BASiD;AACjD,KAAI,CAAChH,SAAU,QAAO8E,KAAAA;CAEtB,MAAMoC,iBAAiB,OAAOC,oBAAiD;EAC7E,MAAMC,SAASL,WAAW1C,IAAI8C,gBAAgB;AAC9C,MAAIC,OAAQ,QAAOA;EACnB,MAAMC,SAAS,MAAMnD,eAAenC,UAAUoF,iBAAiB5D,QAAQ;AACvEwD,aAAWO,IAAIH,iBAAiBE,OAAO;AACvC,SAAOA;;CAGT,MAAME,aAAa,OACjB/B,OACAgC,gBACoD;AACpD,MAAI,OAAOhC,UAAU,UAAU;GAC7B,MAAM/F,OAAO+F,MAAMjE,MAAM;AACzB,OAAI,CAAC9B,KAAM,QAAO;GAElB,MAAMgI,eAAe/B,4BAA4B8B,aAAa/H,MAAM,KAAK;AACzE,OAAI,CAACgI,aAAc,QAAOhI;AAC1B,OAAI,CAACqH,uBAAuB/F,qBAAqB2G,IAAID,aAAa,CAChE,OAAM,IAAIjG,MAAM,oBAAoB;AAEtC,OAAI,CAAC+B,QAAQoE,IAAI,QAAQF,aAA+B,CACtD,OAAM,IAAIjG,MAAM,YAAY;AAG9BwF,wBAAqBY,IAAIH,aAAa;AAOtC,UAAO;IACLhI;IACAG,OAPenB,qBACf8E,SACA,QACAkE,aACD;IAIA;;EAGH,MAAMhI,OAAO+F,MAAM/F,KAAK8B,MAAM;AAC9B,MAAI,CAAC9B,KAAM,QAAO;EAElB,MAAMkG,oBAAoBF,mBAAmBD,MAAM9F,MAAM;EACzD,MAAM+H,eAAe/B,4BAA4B8B,aAAa/H,MAAMkG,kBAAkB;EACtF,IAAImC,cAAcN;EAElB,MAAMO,kBAA0C,EAC9CtI,MACD;AAED,MAAI+F,MAAM7F,WAAWmF,KAAAA,EAAWiD,iBAAgBpI,SAAS6F,MAAM7F;AAC/D,MAAI6F,MAAM3F,YAAYiF,KAAAA,EAAWiD,iBAAgBlI,UAAU2F,MAAM3F;AACjE,MAAI8F,kBAAmBoC,iBAAgBrI,QAAQiG;AAC/C,MAAIH,MAAM5F,UAAUkF,KAAAA,EAAWiD,iBAAgBnI,QAAQ4F,MAAM5F;AAE7D,MAAI6H,cAAc;AAChB,OAAI,CAACX,uBAAuB/F,qBAAqB2G,IAAID,aAAa,CAChE,OAAM,IAAIjG,MAAM,oBAAoB;AAEtC,OAAI,CAAC+B,QAAQoE,IAAI,QAAQF,aAA+B,CACtD,OAAM,IAAIjG,MAAM,YAAY;AAG9BwF,wBAAqBY,IAAIH,aAAa;AACtCK,iBAAc,MAAMZ,eAAeO,aAAa;GAEhD,MAAMlB,WAAW9H,qBACf8E,SACA,QACAkE,aACD;AACDM,mBAAgBnI,QAAQyG,0BACtB0B,gBAAgBnI,OAChB2G,SACD;aACQf,MAAMxF,aAAa8E,KAAAA,EAC5B,OAAM,IAAItD,MAAM,oEAAoE;EAGtF,MAAM8D,iBAAiB,MAAMuB,4BAA4B;GACvD9E;GACArC,OAAOoI;GACPvE;GACAvD,UAAUwF,MAAMxF;GAChB8G;GACAC;GACAC;GACD,CAAC;AACF,MAAI1B,mBAAmBR,KAAAA,EAAWiD,iBAAgB/H,WAAWsF;AAE7D,SAAOyC;;AAGT,KAAI7H,MAAM2B,QAAQ7B,SAAS,EAAE;EAE3B,MAAMkI,YADW,MAAM5E,QAAQ2E,IAAIjI,SAAS8B,KAAK0D,UAAU+B,WAAW/B,OAAO9F,MAAM,CAAC,CAAC,EAC3DuC,QAAQuD,UAAoDA,UAAU,KAAK;AACrG,SAAO0C,SAAS/E,SAAS,IAAI+E,WAAWpD,KAAAA;;AAI1C,QADiB,MAAMyC,WAAWvH,UAAUN,MAAM,IAC/BoF,KAAAA;;AAGrB,IAAaqD,4BAA4BtI,YAA0D;AACjG,KAAI,CAACA,WAAW,OAAOA,YAAY,SAAU,QAAO,EAAE;CACtD,MAAM8B,aAA8B,EAAE;AAEtC,KAAI9B,QAAQO,cAAc,OAAOP,QAAQO,eAAe,YAAY,CAACF,MAAM2B,QAAQhC,QAAQO,WAAW,CACpGuB,YAAWvB,aAAaP,QAAQO;AAGlC,KAAIP,QAAQC,QAAQ,OAAOD,QAAQC,SAAS,YAAY,CAACI,MAAM2B,QAAQhC,QAAQC,KAAK,CAClF6B,YAAW7B,OAAOD,QAAQC;AAG5B6B,YAAW5B,QAAQuE,eAAezE,QAAQE,MAAM;AAChD4B,YAAW3B,WAAWuF,2BAA2B1F,QAAQG,SAAS;AAClE2B,YAAWtB,aAAa0E,oBAAoBlF,QAAQQ,WAAW;AAE/D,QAAOsB;;AAGT,IAAayG,sCAAsC,OAAO,EACxDrG,UACAwB,SACAY,WACAtE,SACAiH,sBAAsB,YAOC;CACvB,MAAMpH,QAAQ,MAAMwE,eAAenC,UAAUoC,WAAWZ,QAAQ;CAChE,MAAMwD,6BAAa,IAAIE,KAAyB;AAChDF,YAAWO,IAAInD,WAAWzE,MAAM;CAEhC,MAAMsH,uCAAuB,IAAIhG,KAAa;AAC9C,OAAM6F,4BAA4B;EAChC9E;EACArC;EACA6D;EACAvD,UAAUH,QAAQG;EAClB8G;EACAC;EACAC;EACD,CAAC;AAEF,QAAO9G,MAAMmI,KAAKrB,qBAAqB;;AAGzC,IAAasB,cAAc,OAAO,EAChCvG,UACAwB,SACAY,WACA7B,OACAzC,SACAiH,sBAAsB,YAQO;CAC7B,MAAM,EAAEpH,OAAOiB,eAAe,MAAM4H,oBAAoB;EACtDxG;EACAwB;EACAY;EACA7B;EACAwE;EACD,CAAC;CACF,MAAM1G,aAAaP,QAAQO,cAAc0E,KAAAA;CACzC,MAAMhF,OAAOD,QAAQC;CACrB,MAAMC,QAAQuE,eAAezE,QAAQE,MAAM;CAC3C,MAAMgH,6BAAa,IAAIE,KAAyB;AAChDF,YAAWO,IAAInD,WAAWzE,MAAM;CAEhC,MAAMM,WAAW,MAAM6G,4BAA4B;EACjD9E;EACArC;EACA6D;EACAvD,UAAUH,QAAQG;EAClB8G;EACAC;EACAC,sCAAsB,IAAIhG,KAAY;EACvC,CAAC;AAEF,KAAInB,QAAQQ,YAAY;EACtB,MAAMmI,iBAAiB9I,MAAM+I,KAAK9H,YAAYP,WAAW;AACzD,MAAIJ,aAAa8E,KAAAA,EACf0D,gBAAexI,SAASA,SAAgB;EAG1C,MAAM0I,kBAAkB,MAAMF,eAAeG,SAAS9I,QAAQQ,YAAY,EACxEuI,QAAQ,EACNC,eAAe3H,kCAAiC,EAClD,EACD,CAAC;EACF,MAAMT,aAAa,OAAOiI,gBAAgBjI,eAAe,YACpD8D,OAAOC,SAASkE,gBAAgBjI,WAAW,IAC3CiI,gBAAgBjI,cAAc,IAC/BgE,KAAKW,MAAMsD,gBAAgBjI,WAAW,GACtCqE,KAAAA;AAEJ,SAAO;GACLvE,MAAML,MAAM2B,QAAQ6G,gBAAgBI,MAAM,GAAGJ,gBAAgBI,QAAQ,EAAE;GACvEtI,UAAUkI,gBAAgBlI;GAC1B,GAAIC,eAAeqE,KAAAA,IAAY,EAAErE,YAAY,GAAG,EAAE;GACnD;;CAGH,MAAMsI,eAAerJ,MAAM+I,KAAK9H,YAAYP,WAAW;AACvD,KAAIJ,aAAa8E,KAAAA,EACfiE,cAAa/I,SAASA,SAAgB;AAExC,KAAIF,QAAQ0G,OAAOC,KAAK3G,KAAK,CAACqD,OAC5B4F,cAAajJ,KAAKA,KAAK;AAEzBiJ,cAAahJ,MAAMA,MAAM;CAEzB,MAAMQ,OAAO,MAAMwI;AACnB,QAAO,EAAExI,MAAML,MAAM2B,QAAQtB,KAAK,GAAGA,OAAO,EAAA,EAAI;;AAGlD,IAAMgI,sBAAsB,OAAO,EACjCxG,UACAwB,SACAY,WACA7B,OACAwE,sBAAsB,YAOa;AACnC,KAAI,CAACA,uBAAuB/F,qBAAqB2G,IAAIvD,UAAU,CAC7D,OAAM,IAAI3C,MAAM,oBAAoB;AAGtC,KAAI,CAAC+B,QAAQoE,IAAI,QAAQxD,UAA4B,CACnD,OAAM,IAAI3C,MAAM,YAAY;AAO9B,QAAO;EAAE9B,OAJK,MAAMwE,eAAenC,UAAUoC,WAAWZ,QAAQ;EAIhD5C,YAFe,EAAE+F,MAAM,CAACpE,OADpB7D,qBAAqB8E,SAAS,QAAQY,UAA4C,CAC5C,EAAG;EAEjC;;AAG9B,IAAa8E,cAAc,OAAO,EAChClH,UACAwB,SACAY,WACA7B,OACAwE,sBAAsB,YAOD;CACrB,MAAM,EAAEpH,OAAOiB,eAAe,MAAM4H,oBAAoB;EACtDxG;EACAwB;EACAY;EACA7B;EACAwE;EACD,CAAC;CAEF,MAAMoC,QAAQ,MAAMxJ,MAAMyJ,eAAexI,WAAW;AACpD,KAAI,OAAOuI,UAAU,YAAY,CAAC3E,OAAOC,SAAS0E,MAAM,IAAIA,QAAQ,EAClE,QAAO;AAGT,QAAOzE,KAAKW,MAAM8D,MAAM"}
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
//#region __vite-optional-peer-dep:@react-email/render:resend
|
|
2
|
-
var render_resend_default = {};
|
|
3
|
-
throw new Error(`Could not resolve "@react-email/render" imported by "resend". Is it installed?`);
|
|
4
|
-
//#endregion
|
|
5
|
-
export { render_resend_default as default };
|
|
6
|
-
|
|
7
|
-
//# sourceMappingURL=render_resend-DQANggpW.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"render_resend-DQANggpW.js","names":[],"sources":["../__vite-optional-peer-dep:@react-email/render:resend"],"sourcesContent":["export default {};\nthrow new Error(`Could not resolve \"@react-email/render\" imported by \"resend\". Is it installed?`)"],"mappings":";AAAA,IAAA,wBAAe,EAAE;AACjB,MAAM,IAAI,MAAM,iFAAA"}
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"handler.d.ts","sourceRoot":"","sources":["../../../../src/rts/api/cleanup/handler.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,GAAG,EAAc,MAAM,cAAc,CAAA;AAO9C,KAAK,WAAW,GAAG;IACjB,EAAE,CAAC,EAAE,MAAM,CAAA;IACX,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,eAAe,CAAC,EAAE,MAAM,EAAE,CAAA;CAC3B,CAAA;yBAqBe,KAAK,GAAG,CAAC,WAAW,CAAC;AAArC,wBAEC"}
|
|
@@ -1,11 +0,0 @@
|
|
|
1
|
-
import { z } from 'zod';
|
|
2
|
-
export declare const Route = "/api/rb/rts/cleanup";
|
|
3
|
-
export declare const requestSchema: z.ZodObject<{
|
|
4
|
-
tenantId: z.ZodString;
|
|
5
|
-
}, z.core.$strip>;
|
|
6
|
-
export type RequestPayload = z.infer<typeof requestSchema>;
|
|
7
|
-
export declare const responseSchema: z.ZodObject<{
|
|
8
|
-
ok: z.ZodBoolean;
|
|
9
|
-
}, z.core.$strip>;
|
|
10
|
-
export type ResponsePayload = z.infer<typeof responseSchema>;
|
|
11
|
-
//# sourceMappingURL=index.d.ts.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../../src/rts/api/cleanup/index.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,CAAC,EAAE,MAAM,KAAK,CAAA;AAGvB,eAAO,MAAM,KAAK,wBAAwB,CAAA;AAE1C,eAAO,MAAM,aAAa;;iBAExB,CAAA;AAEF,MAAM,MAAM,cAAc,GAAG,CAAC,CAAC,KAAK,CAAC,OAAO,aAAa,CAAC,CAAA;AAE1D,eAAO,MAAM,cAAc;;iBAEzB,CAAA;AAEF,MAAM,MAAM,eAAe,GAAG,CAAC,CAAC,KAAK,CAAC,OAAO,cAAc,CAAC,CAAA"}
|