@maas/payload-plugin-media-cloud 0.0.33 → 0.0.34
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/adapter/handleDelete.mjs.map +1 -1
- package/dist/adapter/handleUpload.mjs.map +1 -1
- package/dist/adapter/staticHandler.mjs.map +1 -1
- package/dist/adapter/storageAdapter.mjs.map +1 -1
- package/dist/collectionHooks/afterChange.mjs.map +1 -1
- package/dist/collectionHooks/beforeChange.mjs.map +1 -1
- package/dist/collections/mediaCollection.mjs.map +1 -1
- package/dist/endpoints/fileExistsHandler.mjs.map +1 -1
- package/dist/endpoints/muxAssetHandler.mjs.map +1 -1
- package/dist/endpoints/muxCreateUploadHandler.mjs.map +1 -1
- package/dist/endpoints/muxWebhookHandler.mjs.map +1 -1
- package/dist/endpoints/tusCleanupHandler.mjs.map +1 -1
- package/dist/endpoints/tusFolderHandler.mjs.map +1 -1
- package/dist/endpoints/tusPostProcessorHandler.mjs.map +1 -1
- package/dist/fields/alt.mjs.map +1 -1
- package/dist/fields/filename.mjs.map +1 -1
- package/dist/fields/height.mjs.map +1 -1
- package/dist/fields/mux.mjs.map +1 -1
- package/dist/fields/path.mjs.map +1 -1
- package/dist/fields/storage.mjs.map +1 -1
- package/dist/fields/thumbnail.mjs.map +1 -1
- package/dist/fields/width.mjs.map +1 -1
- package/dist/hooks/useErrorHandler.mjs +8 -25
- package/dist/hooks/useErrorHandler.mjs.map +1 -1
- package/dist/hooks/useMediaCloudEmitter.mjs.map +1 -1
- package/dist/plugin.mjs.map +1 -1
- package/dist/tus/stores/s3/expirationManager.mjs.map +1 -1
- package/dist/tus/stores/s3/fileOperations.mjs.map +1 -1
- package/dist/tus/stores/s3/index.mjs.map +1 -1
- package/dist/tus/stores/s3/metadataManager.mjs.map +1 -1
- package/dist/tus/stores/s3/partsManager.mjs.map +1 -1
- package/dist/tus/stores/s3/s3Store.mjs +2 -2
- package/dist/tus/stores/s3/s3Store.mjs.map +1 -1
- package/dist/tus/stores/s3/semaphore.mjs.map +1 -1
- package/dist/types/errors.mjs.map +1 -1
- package/dist/utils/buildS3Path.mjs.map +1 -1
- package/dist/utils/buildThumbnailURL.mjs.map +1 -1
- package/dist/utils/defaultOptions.mjs.map +1 -1
- package/dist/utils/file.mjs.map +1 -1
- package/dist/utils/mux.mjs.map +1 -1
- package/dist/utils/tus.mjs.map +1 -1
- package/package.json +1 -1
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"expirationManager.mjs","names":["
|
|
1
|
+
{"version":3,"file":"expirationManager.mjs","names":["client: S3","bucket: string","expirationPeriodInMilliseconds: number","generateInfoKey: S3MetadataManager['generateInfoKey']","generatePartKey: S3MetadataManager['generatePartKey']","keyMarker: string | undefined","uploadIdMarker: string | undefined","listResponse: AWS.ListMultipartUploadsCommandOutput","deletions: Promise<AWS.DeleteObjectsCommandOutput>[]"],"sources":["../../../../src/tus/stores/s3/expirationManager.ts"],"sourcesContent":["import type AWS from '@aws-sdk/client-s3'\nimport type { S3 } from '@aws-sdk/client-s3'\nimport type { S3MetadataManager } from './metadataManager'\n\ntype GetExpirationDateArgs = {\n createdAt: string\n}\n\nexport class S3ExpirationManager {\n constructor(\n private client: S3,\n private bucket: string,\n private expirationPeriodInMilliseconds: number,\n private generateInfoKey: S3MetadataManager['generateInfoKey'],\n private generatePartKey: S3MetadataManager['generatePartKey']\n ) {}\n\n /**\n * Calculates the expiration date for a file based on a creation date\n * @param args - The function arguments\n * @param args.createdAt - The creation date as a string\n * @returns The expiration date\n */\n getExpirationDate(args: GetExpirationDateArgs): Date {\n const { createdAt } = args\n const date = new Date(createdAt)\n return new Date(date.getTime() + this.expirationPeriodInMilliseconds)\n }\n\n /**\n * Gets the expiration period in milliseconds\n * @returns The expiration period in milliseconds\n */\n getExpiration(): number {\n return this.expirationPeriodInMilliseconds\n }\n\n /**\n * Deletes expired incomplete uploads based on their initialization date.\n * Returns the number of deleted uploads.\n * @param args - The function arguments (empty object)\n * @returns Promise that resolves to the number of deleted uploads\n */\n async deleteExpired(): Promise<number> {\n // No arguments to destructure\n if (this.getExpiration() === 0) {\n return 0\n }\n\n let keyMarker: string | undefined = undefined\n let uploadIdMarker: string | undefined = undefined\n let isTruncated = true\n let deleted = 0\n\n while (isTruncated) {\n const listResponse: AWS.ListMultipartUploadsCommandOutput =\n await this.client.listMultipartUploads({\n Bucket: this.bucket,\n KeyMarker: keyMarker,\n UploadIdMarker: uploadIdMarker,\n })\n\n const expiredUploads =\n listResponse.Uploads?.filter((multiPartUpload: AWS.MultipartUpload) => {\n const initiatedDate = multiPartUpload.Initiated\n return (\n initiatedDate &&\n new Date().getTime() >\n this.getExpirationDate({\n createdAt: initiatedDate.toISOString(),\n }).getTime()\n )\n }) || []\n\n const objectsToDelete = expiredUploads.reduce(\n (all: { Key: string }[], expiredUpload: AWS.MultipartUpload) => {\n all.push(\n {\n Key: this.generateInfoKey({ id: expiredUpload.Key as string }),\n },\n {\n Key: this.generatePartKey({\n id: expiredUpload.Key as string,\n isIncomplete: true,\n }),\n }\n )\n return all\n },\n [] as { Key: string }[]\n )\n\n const deletions: Promise<AWS.DeleteObjectsCommandOutput>[] = []\n\n if (objectsToDelete.length > 0) {\n deletions.push(\n this.client.deleteObjects({\n Bucket: this.bucket,\n Delete: {\n Objects: objectsToDelete,\n },\n })\n )\n }\n\n const abortions = expiredUploads.map(\n (expiredUpload: AWS.MultipartUpload) =>\n this.client.abortMultipartUpload({\n Bucket: this.bucket,\n Key: expiredUpload.Key,\n UploadId: expiredUpload.UploadId,\n })\n )\n\n await Promise.all([...deletions, ...abortions])\n\n deleted += expiredUploads.length\n\n isTruncated = listResponse.IsTruncated || false\n keyMarker = listResponse.NextKeyMarker\n uploadIdMarker = listResponse.NextUploadIdMarker\n }\n\n return deleted\n }\n}\n"],"mappings":";AAQA,IAAa,sBAAb,MAAiC;CAC/B,YACE,AAAQA,QACR,AAAQC,QACR,AAAQC,gCACR,AAAQC,iBACR,AAAQC,iBACR;EALQ;EACA;EACA;EACA;EACA;;;;;;;;CASV,kBAAkB,MAAmC;EACnD,MAAM,EAAE,cAAc;EACtB,MAAM,OAAO,IAAI,KAAK,UAAU;AAChC,SAAO,IAAI,KAAK,KAAK,SAAS,GAAG,KAAK,+BAA+B;;;;;;CAOvE,gBAAwB;AACtB,SAAO,KAAK;;;;;;;;CASd,MAAM,gBAAiC;AAErC,MAAI,KAAK,eAAe,KAAK,EAC3B,QAAO;EAGT,IAAIC,YAAgC;EACpC,IAAIC,iBAAqC;EACzC,IAAI,cAAc;EAClB,IAAI,UAAU;AAEd,SAAO,aAAa;GAClB,MAAMC,eACJ,MAAM,KAAK,OAAO,qBAAqB;IACrC,QAAQ,KAAK;IACb,WAAW;IACX,gBAAgB;IACjB,CAAC;GAEJ,MAAM,iBACJ,aAAa,SAAS,QAAQ,oBAAyC;IACrE,MAAM,gBAAgB,gBAAgB;AACtC,WACE,kCACA,IAAI,MAAM,EAAC,SAAS,GAClB,KAAK,kBAAkB,EACrB,WAAW,cAAc,aAAa,EACvC,CAAC,CAAC,SAAS;KAEhB,IAAI,EAAE;GAEV,MAAM,kBAAkB,eAAe,QACpC,KAAwB,kBAAuC;AAC9D,QAAI,KACF,EACE,KAAK,KAAK,gBAAgB,EAAE,IAAI,cAAc,KAAe,CAAC,EAC/D,EACD,EACE,KAAK,KAAK,gBAAgB;KACxB,IAAI,cAAc;KAClB,cAAc;KACf,CAAC,EACH,CACF;AACD,WAAO;MAET,EAAE,CACH;GAED,MAAMC,YAAuD,EAAE;AAE/D,OAAI,gBAAgB,SAAS,EAC3B,WAAU,KACR,KAAK,OAAO,cAAc;IACxB,QAAQ,KAAK;IACb,QAAQ,EACN,SAAS,iBACV;IACF,CAAC,CACH;GAGH,MAAM,YAAY,eAAe,KAC9B,kBACC,KAAK,OAAO,qBAAqB;IAC/B,QAAQ,KAAK;IACb,KAAK,cAAc;IACnB,UAAU,cAAc;IACzB,CAAC,CACL;AAED,SAAM,QAAQ,IAAI,CAAC,GAAG,WAAW,GAAG,UAAU,CAAC;AAE/C,cAAW,eAAe;AAE1B,iBAAc,aAAa,eAAe;AAC1C,eAAY,aAAa;AACzB,oBAAiB,aAAa;;AAGhC,SAAO"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"fileOperations.mjs","names":["
|
|
1
|
+
{"version":3,"file":"fileOperations.mjs","names":["maxMultipartParts: number","maxUploadSize: number","minPartSize: number","preferredPartSize: number","optimalPartSize: number"],"sources":["../../../../src/tus/stores/s3/fileOperations.ts"],"sourcesContent":["import crypto from 'node:crypto'\nimport fs from 'node:fs'\nimport os from 'node:os'\nimport path from 'node:path'\n\nimport { useErrorHandler } from '../../../hooks/useErrorHandler'\nimport { MediaCloudErrors } from '../../../types/errors'\n\nimport type { NodeFSError } from '../../../types'\nimport type AWS from '@aws-sdk/client-s3'\n\ntype CalculateOptimalPartSizeArgs = {\n size?: number\n}\n\ntype CalculateOffsetFromPartsArgs = {\n parts?: Array<AWS.Part>\n}\n\ntype CalculatePartNumberArgs = {\n parts: Array<AWS.Part>\n}\n\ntype GenerateUniqueTmpFileNameArgs = {\n template: string\n}\n\ntype CalculateOffsetFromPartsExportArgs = {\n parts?: Array<{ Size?: number }>\n}\n\nconst { throwError } = useErrorHandler()\n\nexport class S3FileOperations {\n constructor(\n private maxMultipartParts: number,\n private maxUploadSize: number,\n private minPartSize: number,\n private preferredPartSize: number\n ) {}\n\n /**\n * Calculates the optimal part size for S3 multipart upload\n * @param args - The function arguments\n * @param args.size - The upload size in bytes (optional)\n * @returns The optimal part size in bytes\n */\n calculateOptimalPartSize(args: CalculateOptimalPartSizeArgs): number {\n const { size } = args\n // When upload size is not known, default to preferredPartSize\n // Assuming maxUploadSize can lead to extremely large chunks (hundreds of GB),\n // which breaks streaming/resumable uploads in practice\n let uploadSize = size\n if (uploadSize === undefined) {\n uploadSize = this.preferredPartSize\n }\n\n let optimalPartSize: number\n\n // When upload is smaller or equal to preferredPartSize, upload in just one part\n if (uploadSize <= this.preferredPartSize) {\n optimalPartSize = uploadSize\n }\n // Does the upload fit in maxMultipartParts parts or less with preferredPartSize\n else if (uploadSize <= this.preferredPartSize * this.maxMultipartParts) {\n optimalPartSize = this.preferredPartSize\n // The upload is too big for the preferred size\n // We divide the size with the max amount of parts and round it up\n } else {\n optimalPartSize = Math.ceil(uploadSize / this.maxMultipartParts)\n }\n\n // Always ensure the part size is at least minPartSize\n return Math.max(optimalPartSize, this.minPartSize)\n }\n\n /**\n * Calculates the offset based on uploaded parts\n * @param args - The function arguments\n * @param args.parts - Array of uploaded parts (optional)\n * @returns The total offset in bytes\n */\n calculateOffsetFromParts(args: CalculateOffsetFromPartsArgs): number {\n const { parts } = args\n return parts && parts.length > 0\n ? parts.reduce((a, b) => a + (b.Size ?? 0), 0)\n : 0\n }\n\n /**\n * Calculates the next part number based on existing parts\n * @param args - The function arguments\n * @param args.parts - Array of uploaded parts\n * @returns The next part number to use\n */\n calculatePartNumber(args: CalculatePartNumberArgs): number {\n const { parts } = args\n return parts.length > 0 ? parts[parts.length - 1].PartNumber! + 1 : 1\n }\n\n /**\n * Generates a unique temporary file name\n * @param args - The function arguments\n * @param args.template - The template string for the file name\n * @returns Promise that resolves to the unique file path\n * @throws Error if unable to find unique name after max tries\n */\n async generateUniqueTmpFileName(\n args: GenerateUniqueTmpFileNameArgs\n ): Promise<string> {\n const { template } = args\n const tries = 5\n for (let i = 0; i < tries; i++) {\n const randomId = crypto.randomBytes(16).toString('hex')\n const filePath = path.join(os.tmpdir(), `${template}${randomId}`)\n try {\n await fs.promises.access(filePath, fs.constants.F_OK)\n } catch (error) {\n const nodeError = error as NodeFSError\n if (nodeError.code === 'ENOENT') {\n return filePath\n }\n }\n }\n throwError(MediaCloudErrors.S3_UNIQUE_NAME_ERROR)\n throw new Error() // This will never execute but satisfies TypeScript\n }\n}\n\n/**\n * Calculates the total offset from uploaded parts\n * @param args - The function arguments\n * @param args.parts - Array of parts with size information (optional)\n * @returns The total offset in bytes\n */\nexport function calculateOffsetFromParts(\n args: CalculateOffsetFromPartsExportArgs\n) {\n const { parts } = args\n return parts && parts.length > 0\n ? parts.reduce((a, b) => a + (b.Size ?? 0), 0)\n : 0\n}\n"],"mappings":";;;;;;;;AA+BA,MAAM,EAAE,eAAe,iBAAiB;AAExC,IAAa,mBAAb,MAA8B;CAC5B,YACE,AAAQA,mBACR,AAAQC,eACR,AAAQC,aACR,AAAQC,mBACR;EAJQ;EACA;EACA;EACA;;;;;;;;CASV,yBAAyB,MAA4C;EACnE,MAAM,EAAE,SAAS;EAIjB,IAAI,aAAa;AACjB,MAAI,eAAe,OACjB,cAAa,KAAK;EAGpB,IAAIC;AAGJ,MAAI,cAAc,KAAK,kBACrB,mBAAkB;WAGX,cAAc,KAAK,oBAAoB,KAAK,kBACnD,mBAAkB,KAAK;MAIvB,mBAAkB,KAAK,KAAK,aAAa,KAAK,kBAAkB;AAIlE,SAAO,KAAK,IAAI,iBAAiB,KAAK,YAAY;;;;;;;;CASpD,yBAAyB,MAA4C;EACnE,MAAM,EAAE,UAAU;AAClB,SAAO,SAAS,MAAM,SAAS,IAC3B,MAAM,QAAQ,GAAG,MAAM,KAAK,EAAE,QAAQ,IAAI,EAAE,GAC5C;;;;;;;;CASN,oBAAoB,MAAuC;EACzD,MAAM,EAAE,UAAU;AAClB,SAAO,MAAM,SAAS,IAAI,MAAM,MAAM,SAAS,GAAG,aAAc,IAAI;;;;;;;;;CAUtE,MAAM,0BACJ,MACiB;EACjB,MAAM,EAAE,aAAa;EACrB,MAAM,QAAQ;AACd,OAAK,IAAI,IAAI,GAAG,IAAI,OAAO,KAAK;GAC9B,MAAM,WAAW,OAAO,YAAY,GAAG,CAAC,SAAS,MAAM;GACvD,MAAM,WAAW,KAAK,KAAK,GAAG,QAAQ,EAAE,GAAG,WAAW,WAAW;AACjE,OAAI;AACF,UAAM,GAAG,SAAS,OAAO,UAAU,GAAG,UAAU,KAAK;YAC9C,OAAO;AAEd,QADkB,MACJ,SAAS,SACrB,QAAO;;;AAIb,aAAW,iBAAiB,qBAAqB;AACjD,QAAM,IAAI,OAAO;;;;;;;;;AAUrB,SAAgB,yBACd,MACA;CACA,MAAM,EAAE,UAAU;AAClB,QAAO,SAAS,MAAM,SAAS,IAC3B,MAAM,QAAQ,GAAG,MAAM,KAAK,EAAE,QAAQ,IAAI,EAAE,GAC5C"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.mjs","names":["
|
|
1
|
+
{"version":3,"file":"index.mjs","names":["magicError: UseMagicErrorReturn"],"sources":["../../../../src/tus/stores/s3/index.ts"],"sourcesContent":["import { MediaCloudErrors } from '../../../types/errors'\nimport { S3Store } from './s3Store'\nimport { useMagicError } from '@maas/error-handler'\n\nimport type { MediaCloudPluginOptions } from '../../../types'\nimport type { UseMagicErrorReturn } from '@maas/error-handler'\n\nconst magicError: UseMagicErrorReturn = useMagicError({\n prefix: 'PLUGIN-MEDIA-CLOUD',\n})\n\n/**\n * Validates S3 configuration options\n * @param s3Config - The S3 configuration to validate\n * @throws {Error} When required S3 configuration is missing\n */\nfunction validateS3Config(s3Config: MediaCloudPluginOptions['s3']): void {\n magicError.assert(s3Config?.bucket, MediaCloudErrors.S3_CONFIG_MISSING)\n magicError.assert(s3Config?.region, MediaCloudErrors.S3_CONFIG_MISSING)\n magicError.assert(s3Config?.accessKeyId, MediaCloudErrors.S3_CONFIG_MISSING)\n magicError.assert(\n s3Config?.secretAccessKey,\n MediaCloudErrors.S3_CONFIG_MISSING\n )\n}\n\n/**\n * Creates and configures an S3Store instance\n * @param s3Config - The S3 configuration options\n * @returns A configured S3Store instance\n * @throws {Error} When required S3 configuration is missing\n */\nexport function createS3Store(\n s3Config: MediaCloudPluginOptions['s3']\n): S3Store {\n validateS3Config(s3Config)\n\n return new S3Store({\n s3ClientConfig: {\n acl: 'public-read',\n bucket: s3Config?.bucket ?? '',\n credentials: {\n accessKeyId: s3Config?.accessKeyId ?? '',\n secretAccessKey: s3Config?.secretAccessKey ?? '',\n },\n endpoint: s3Config?.endpoint ?? '',\n forcePathStyle: true,\n region: s3Config?.region ?? '',\n },\n })\n}\n"],"mappings":";;;;;AAOA,MAAMA,aAAkC,cAAc,EACpD,QAAQ,sBACT,CAAC;;;;;;AAOF,SAAS,iBAAiB,UAA+C;AACvE,YAAW,OAAO,UAAU,QAAQ,iBAAiB,kBAAkB;AACvE,YAAW,OAAO,UAAU,QAAQ,iBAAiB,kBAAkB;AACvE,YAAW,OAAO,UAAU,aAAa,iBAAiB,kBAAkB;AAC5E,YAAW,OACT,UAAU,iBACV,iBAAiB,kBAClB;;;;;;;;AASH,SAAgB,cACd,UACS;AACT,kBAAiB,SAAS;AAE1B,QAAO,IAAI,QAAQ,EACjB,gBAAgB;EACd,KAAK;EACL,QAAQ,UAAU,UAAU;EAC5B,aAAa;GACX,aAAa,UAAU,eAAe;GACtC,iBAAiB,UAAU,mBAAmB;GAC/C;EACD,UAAU,UAAU,YAAY;EAChC,gBAAgB;EAChB,QAAQ,UAAU,UAAU;EAC7B,EACF,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"metadataManager.mjs","names":["
|
|
1
|
+
{"version":3,"file":"metadataManager.mjs","names":["client: S3","bucket: string","shouldUseExpirationTags: () => boolean","generateCompleteTag: (value: 'false' | 'true') => string | undefined"],"sources":["../../../../src/tus/stores/s3/metadataManager.ts"],"sourcesContent":["import { TUS_RESUMABLE, Upload } from '@tus/utils'\n\nimport { useErrorHandler } from '../../../hooks/useErrorHandler'\nimport { MediaCloudLogs } from '../../../types/errors'\n\nimport type { S3 } from '@aws-sdk/client-s3'\nimport type { TusUploadMetadata } from '../../../types'\n\ntype GenerateInfoKeyArgs = {\n id: string\n}\n\ntype GeneratePartKeyArgs = {\n id: string\n prefix?: string\n isIncomplete?: boolean\n}\n\ntype GetMetadataArgs = {\n id: string\n}\n\ntype SaveMetadataArgs = {\n upload: Upload\n uploadId: string\n}\n\ntype CompleteMetadataArgs = {\n upload: Upload\n}\n\nconst { log } = useErrorHandler()\n\nexport class S3MetadataManager {\n constructor(\n private client: S3,\n private bucket: string,\n private shouldUseExpirationTags: () => boolean,\n private generateCompleteTag: (value: 'false' | 'true') => string | undefined\n ) {}\n /**\n * Generates the S3 key for metadata info files\n * @param args - The function arguments\n * @param args.id - The file ID\n * @returns The S3 key for the metadata file\n */\n generateInfoKey(args: GenerateInfoKeyArgs): string {\n return `${args.id}.info`\n }\n\n /**\n * Generates the S3 key for part files\n * @param args - The function arguments\n * @param args.id - The file ID\n * @param args.prefix - Path prefix for nested media items\n * @param args.isIncomplete - Whether this is an incomplete part\n * @returns The S3 key for the part file\n */\n generatePartKey(args: GeneratePartKeyArgs): string {\n const { id, prefix, isIncomplete = false } = args\n let key = id\n\n if (prefix) {\n key = `${prefix}/${key}`\n }\n\n if (isIncomplete) {\n key += '.part'\n }\n\n // TODO: introduce ObjectPrefixing for parts and incomplete parts\n // ObjectPrefix is prepended to the name of each S3 object that is created\n // to store uploaded files. It can be used to create a pseudo-directory\n // structure in the bucket, e.g. \"path/to/my/uploads\".\n return key\n }\n\n /**\n * Retrieves upload metadata previously saved in `${file_id}.info`.\n * Always fetches fresh data from S3 to ensure correctness.\n * @param args - The function arguments\n * @param args.id - The file ID to retrieve metadata for\n * @returns Promise that resolves to the upload metadata\n */\n async getMetadata(args: GetMetadataArgs): Promise<TusUploadMetadata> {\n const { id } = args\n\n const { Body, Metadata } = await this.client.getObject({\n Bucket: this.bucket,\n Key: this.generateInfoKey({ id }),\n })\n\n const file = JSON.parse((await Body?.transformToString()) as string)\n\n const parsedSize =\n file.size === undefined || file.size === null\n ? undefined\n : Number(file.size)\n\n const metadata: TusUploadMetadata = {\n file: new Upload({\n id,\n creation_date: file.creation_date,\n metadata: file.metadata,\n offset: Number.parseInt(file.offset, 10),\n size: Number.isFinite(parsedSize) ? parsedSize : undefined,\n storage: file.storage,\n }),\n 'tus-version': Metadata?.['tus-version'] as string,\n 'upload-id': (Metadata?.['upload-id'] || file['upload-id']) as string,\n }\n\n return metadata\n }\n\n /**\n * Saves upload metadata to a `${file_id}.info` file on S3.\n * The upload-id is stored in the S3 object metadata.\n * @param args - The function arguments\n * @param args.upload - The upload object containing metadata\n * @param args.uploadId - The upload ID for the multipart upload\n * @returns Promise that resolves when metadata is saved\n */\n async saveMetadata(args: SaveMetadataArgs): Promise<void> {\n const { upload, uploadId } = args\n log(MediaCloudLogs.S3_STORE_METADATA_SAVING)\n\n await this.client.putObject({\n Body: JSON.stringify(upload),\n Bucket: this.bucket,\n Key: this.generateInfoKey({ id: upload.id }),\n Metadata: {\n 'tus-version': TUS_RESUMABLE,\n 'upload-id': uploadId,\n },\n Tagging: this.generateCompleteTag('false'),\n })\n log(MediaCloudLogs.S3_STORE_METADATA_SAVED)\n }\n\n /**\n * Completes metadata by updating it with completion tags\n * @param args - The function arguments\n * @param args.upload - The completed upload object\n * @returns Promise that resolves when metadata is updated\n */\n async completeMetadata(args: CompleteMetadataArgs): Promise<void> {\n const { upload } = args\n\n if (!this.shouldUseExpirationTags()) {\n return\n }\n\n const metadata = await this.getMetadata({ id: upload.id })\n const { 'upload-id': uploadId } = metadata\n\n await this.client.putObject({\n Body: JSON.stringify(upload),\n Bucket: this.bucket,\n Key: this.generateInfoKey({ id: upload.id }),\n Metadata: {\n 'tus-version': TUS_RESUMABLE,\n 'upload-id': uploadId,\n },\n Tagging: this.generateCompleteTag('true'),\n })\n }\n}\n"],"mappings":";;;;;AA+BA,MAAM,EAAE,QAAQ,iBAAiB;AAEjC,IAAa,oBAAb,MAA+B;CAC7B,YACE,AAAQA,QACR,AAAQC,QACR,AAAQC,yBACR,AAAQC,qBACR;EAJQ;EACA;EACA;EACA;;;;;;;;CAQV,gBAAgB,MAAmC;AACjD,SAAO,GAAG,KAAK,GAAG;;;;;;;;;;CAWpB,gBAAgB,MAAmC;EACjD,MAAM,EAAE,IAAI,QAAQ,eAAe,UAAU;EAC7C,IAAI,MAAM;AAEV,MAAI,OACF,OAAM,GAAG,OAAO,GAAG;AAGrB,MAAI,aACF,QAAO;AAOT,SAAO;;;;;;;;;CAUT,MAAM,YAAY,MAAmD;EACnE,MAAM,EAAE,OAAO;EAEf,MAAM,EAAE,MAAM,aAAa,MAAM,KAAK,OAAO,UAAU;GACrD,QAAQ,KAAK;GACb,KAAK,KAAK,gBAAgB,EAAE,IAAI,CAAC;GAClC,CAAC;EAEF,MAAM,OAAO,KAAK,MAAO,MAAM,MAAM,mBAAmB,CAAY;EAEpE,MAAM,aACJ,KAAK,SAAS,UAAa,KAAK,SAAS,OACrC,SACA,OAAO,KAAK,KAAK;AAevB,SAboC;GAClC,MAAM,IAAI,OAAO;IACf;IACA,eAAe,KAAK;IACpB,UAAU,KAAK;IACf,QAAQ,OAAO,SAAS,KAAK,QAAQ,GAAG;IACxC,MAAM,OAAO,SAAS,WAAW,GAAG,aAAa;IACjD,SAAS,KAAK;IACf,CAAC;GACF,eAAe,WAAW;GAC1B,aAAc,WAAW,gBAAgB,KAAK;GAC/C;;;;;;;;;;CAaH,MAAM,aAAa,MAAuC;EACxD,MAAM,EAAE,QAAQ,aAAa;AAC7B,MAAI,eAAe,yBAAyB;AAE5C,QAAM,KAAK,OAAO,UAAU;GAC1B,MAAM,KAAK,UAAU,OAAO;GAC5B,QAAQ,KAAK;GACb,KAAK,KAAK,gBAAgB,EAAE,IAAI,OAAO,IAAI,CAAC;GAC5C,UAAU;IACR,eAAe;IACf,aAAa;IACd;GACD,SAAS,KAAK,oBAAoB,QAAQ;GAC3C,CAAC;AACF,MAAI,eAAe,wBAAwB;;;;;;;;CAS7C,MAAM,iBAAiB,MAA2C;EAChE,MAAM,EAAE,WAAW;AAEnB,MAAI,CAAC,KAAK,yBAAyB,CACjC;EAIF,MAAM,EAAE,aAAa,aADJ,MAAM,KAAK,YAAY,EAAE,IAAI,OAAO,IAAI,CAAC;AAG1D,QAAM,KAAK,OAAO,UAAU;GAC1B,MAAM,KAAK,UAAU,OAAO;GAC5B,QAAQ,KAAK;GACb,KAAK,KAAK,gBAAgB,EAAE,IAAI,OAAO,IAAI,CAAC;GAC5C,UAAU;IACR,eAAe;IACf,aAAa;IACd;GACD,SAAS,KAAK,oBAAoB,OAAO;GAC1C,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"partsManager.mjs","names":["fs","os","stream","NoSuchKey","NotFound","S3","StreamSplitter","useErrorHandler","MediaCloudErrors","MediaCloudLogs","AWS","Readable","IncompletePartInfo","TusUploadMetadata","S3FileOperations","S3MetadataManager","Semaphore","SemaphorePermit","RetrievePartsArgs","id","partNumberMarker","FinishMultipartUploadArgs","metadata","parts","Array","Part","GetIncompletePartArgs","GetIncompletePartSizeArgs","DeleteIncompletePartArgs","DownloadIncompletePartArgs","UploadIncompletePartArgs","readStream","ReadStream","UploadPartArgs","partNumber","UploadPartsArgs","currentPartNumber","offset","log","throwError","S3PartsManager","constructor","client","bucket","minPartSize","partUploadSemaphore","metadataManager","fileOperations","generateCompleteTag","value","retrieveParts","args","Promise","getMetadata","MUX_UPLOAD_ID_MISSING","Error","params","ListPartsCommandInput","Bucket","Key","generatePartKey","prefix","file","undefined","PartNumberMarker","UploadId","data","listParts","Parts","IsTruncated","rest","NextPartNumberMarker","sort","a","b","PartNumber","finishMultipartUpload","MultipartUpload","map","part","ETag","result","completeMultipartUpload","Location","error","TUS_UPLOAD_ERROR","cause","getIncompletePart","getObject","isIncomplete","Body","getIncompletePartSize","headObject","ContentLength","deleteIncompletePart","deleteObject","S3_DELETE_ERROR","downloadIncompletePart","incompletePart","filePath","generateUniqueTmpFileName","template","incompletePartSize","byteCounterTransform","Transform","transform","chunk","_","callback","length","promises","pipeline","createWriteStream","createReadStream","options","cleanUpOnEnd","fileReader","on","unlink","err","destroy","createReader","path","size","rm","catch","uploadIncompletePart","putObject","Tagging","S3_STORE_INCOMPLETE_PART_UPLOADED","uploadPart","permit","acquire","UploadPartCommandInput","uploadParts","initialOffset","pendingChunkFilepath","bytesUploaded","splitterStream","chunkSize","calculateOptimalPartSize","directory","tmpdir","filepath","partSize","acquiredPermit","isFinalPart","uploadChunk","readable","mappedError","String","deferred","push","S3_STORE_CHUNK_REMOVAL_FAILED","reject","allSettled","all"],"sources":["../../../../src/tus/stores/s3/partsManager.ts"],"sourcesContent":["import fs from 'node:fs'\nimport os from 'node:os'\nimport stream from 'node:stream'\n\nimport { NoSuchKey, NotFound, type S3 } from '@aws-sdk/client-s3'\nimport { StreamSplitter } from '@tus/utils'\n\nimport { useErrorHandler } from '../../../hooks/useErrorHandler'\nimport { MediaCloudErrors, MediaCloudLogs } from '../../../types/errors'\n\nimport type AWS from '@aws-sdk/client-s3'\nimport type { Readable } from 'node:stream'\nimport type { IncompletePartInfo, TusUploadMetadata } from '../../../types'\nimport type { S3FileOperations } from './fileOperations'\nimport type { S3MetadataManager } from './metadataManager'\nimport type { Semaphore, SemaphorePermit } from './semaphore'\n\ntype RetrievePartsArgs = {\n id: string\n partNumberMarker?: string\n}\n\ntype FinishMultipartUploadArgs = {\n metadata: TusUploadMetadata\n parts: Array<AWS.Part>\n}\n\ntype GetIncompletePartArgs = {\n id: string\n}\n\ntype GetIncompletePartSizeArgs = {\n id: string\n}\n\ntype DeleteIncompletePartArgs = {\n id: string\n}\n\ntype DownloadIncompletePartArgs = {\n id: string\n}\n\ntype UploadIncompletePartArgs = {\n id: string\n readStream: fs.ReadStream | Readable\n}\n\ntype UploadPartArgs = {\n metadata: TusUploadMetadata\n readStream: fs.ReadStream | Readable\n partNumber: number\n}\n\ntype UploadPartsArgs = {\n metadata: TusUploadMetadata\n readStream: stream.Readable\n currentPartNumber: number\n offset: number\n}\n\nconst { log, throwError } = useErrorHandler()\n\nexport class S3PartsManager {\n constructor(\n private client: S3,\n private bucket: string,\n private minPartSize: number,\n private partUploadSemaphore: Semaphore,\n private metadataManager: S3MetadataManager,\n private fileOperations: S3FileOperations,\n private generateCompleteTag: (value: 'false' | 'true') => string | undefined\n ) {}\n\n /**\n * Gets the number of complete parts/chunks already uploaded to S3.\n * Retrieves only consecutive parts.\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @param args.partNumberMarker - Marker for pagination (optional)\n * @returns Promise that resolves to array of uploaded parts\n */\n async retrieveParts(args: RetrievePartsArgs): Promise<Array<AWS.Part>> {\n const { id, partNumberMarker } = args\n const metadata = await this.metadataManager.getMetadata({ id })\n\n if (!metadata['upload-id']) {\n throwError(MediaCloudErrors.MUX_UPLOAD_ID_MISSING)\n throw new Error() // This will never execute but satisfies TypeScript\n }\n\n const params: AWS.ListPartsCommandInput = {\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: metadata.file.metadata?.prefix ?? undefined,\n }),\n PartNumberMarker: partNumberMarker,\n UploadId: metadata['upload-id'],\n }\n\n const data = await this.client.listParts(params)\n\n let parts = data.Parts ?? []\n\n if (data.IsTruncated) {\n const rest = await this.retrieveParts({\n id,\n partNumberMarker: data.NextPartNumberMarker,\n })\n parts = [...parts, ...rest]\n }\n\n if (!partNumberMarker) {\n parts.sort((a, b) => (a.PartNumber || 0) - (b.PartNumber || 0))\n }\n\n return parts\n }\n\n /**\n * Completes a multipart upload on S3.\n * This is where S3 concatenates all the uploaded parts.\n * @param args - The function arguments\n * @param args.metadata - The upload metadata\n * @param args.parts - Array of uploaded parts to complete\n * @returns Promise that resolves to the location URL (optional)\n */\n async finishMultipartUpload(\n args: FinishMultipartUploadArgs\n ): Promise<string | undefined> {\n const { metadata, parts } = args\n\n const params = {\n Key: this.metadataManager.generatePartKey({\n id: metadata.file.id,\n prefix: metadata.file.metadata?.prefix ?? undefined,\n }),\n Bucket: this.bucket,\n MultipartUpload: {\n Parts: parts.map((part) => {\n return {\n ETag: part.ETag,\n PartNumber: part.PartNumber,\n }\n }),\n },\n UploadId: metadata['upload-id'],\n }\n\n try {\n const result = await this.client.completeMultipartUpload(params)\n return result.Location\n } catch (error) {\n throwError({ ...MediaCloudErrors.TUS_UPLOAD_ERROR, cause: error })\n throw new Error() // This will never execute but satisfies TypeScript\n }\n }\n\n /**\n * Gets incomplete part from S3\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves to readable stream or undefined if not found\n */\n async getIncompletePart(\n args: GetIncompletePartArgs\n ): Promise<Readable | undefined> {\n const { id } = args\n\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n const data = await this.client.getObject({\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n isIncomplete: true,\n }),\n })\n return data.Body as Readable\n } catch (error) {\n if (error instanceof NoSuchKey) {\n return undefined\n }\n throw error\n }\n }\n\n /**\n * Gets the size of an incomplete part\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves to part size or undefined if not found\n */\n async getIncompletePartSize(\n args: GetIncompletePartSizeArgs\n ): Promise<number | undefined> {\n const { id } = args\n\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n const data = await this.client.headObject({\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n isIncomplete: true,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n })\n return data.ContentLength\n } catch (error) {\n if (error instanceof NotFound) {\n return undefined\n }\n throw error\n }\n }\n\n /**\n * Deletes an incomplete part\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves when deletion is complete\n */\n async deleteIncompletePart(args: DeleteIncompletePartArgs): Promise<void> {\n const { id } = args\n\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n await this.client.deleteObject({\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n isIncomplete: true,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n })\n } catch (error) {\n throwError({\n ...MediaCloudErrors.S3_DELETE_ERROR,\n cause: error,\n })\n }\n }\n\n /**\n * Downloads incomplete part to temporary file\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves to incomplete part info or undefined if not found\n */\n async downloadIncompletePart(\n args: DownloadIncompletePartArgs\n ): Promise<IncompletePartInfo | undefined> {\n const { id } = args\n const incompletePart = await this.getIncompletePart({ id })\n\n if (!incompletePart) {\n return\n }\n const filePath = await this.fileOperations.generateUniqueTmpFileName({\n template: 'tus-s3-incomplete-part-',\n })\n\n try {\n let incompletePartSize = 0\n\n const byteCounterTransform = new stream.Transform({\n transform(chunk, _, callback) {\n incompletePartSize += chunk.length\n callback(null, chunk)\n },\n })\n\n // Write to temporary file\n await stream.promises.pipeline(\n incompletePart,\n byteCounterTransform,\n fs.createWriteStream(filePath)\n )\n\n const createReadStream = (options: { cleanUpOnEnd: boolean }) => {\n const fileReader = fs.createReadStream(filePath)\n\n if (options.cleanUpOnEnd) {\n fileReader.on('end', () => {\n fs.unlink(filePath, () => {})\n })\n\n fileReader.on('error', (err) => {\n fileReader.destroy(err)\n fs.unlink(filePath, () => {})\n })\n }\n\n return fileReader\n }\n\n return {\n createReader: createReadStream,\n path: filePath,\n size: incompletePartSize,\n }\n } catch (err) {\n fs.promises.rm(filePath).catch(() => {})\n throw err\n }\n }\n\n /**\n * Uploads an incomplete part\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @param args.readStream - The stream to read data from\n * @returns Promise that resolves to the ETag of the uploaded part\n */\n async uploadIncompletePart(args: UploadIncompletePartArgs): Promise<string> {\n const { id, readStream } = args\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n const data = await this.client.putObject({\n Body: readStream,\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n isIncomplete: true,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n Tagging: this.generateCompleteTag('false'),\n })\n log(MediaCloudLogs.S3_STORE_INCOMPLETE_PART_UPLOADED)\n return data.ETag as string\n } catch (error) {\n throwError({ ...MediaCloudErrors.TUS_UPLOAD_ERROR, cause: error })\n throw error\n }\n }\n\n /**\n * Uploads a single part\n * @param args - The function arguments\n * @param args.metadata - The upload metadata\n * @param args.readStream - The stream to read data from\n * @param args.partNumber - The part number to upload\n * @returns Promise that resolves to the ETag of the uploaded part\n */\n async uploadPart(args: UploadPartArgs): Promise<AWS.Part> {\n const { metadata, readStream, partNumber } = args\n const permit = await this.partUploadSemaphore.acquire()\n\n if (!metadata['upload-id']) {\n throwError(MediaCloudErrors.MUX_UPLOAD_ID_MISSING)\n throw new Error() // This will never execute but satisfies TypeScript\n }\n\n const params: AWS.UploadPartCommandInput = {\n Body: readStream,\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id: metadata.file.id,\n prefix: metadata.file.metadata?.prefix ?? undefined,\n }),\n PartNumber: partNumber,\n UploadId: metadata['upload-id'],\n }\n\n try {\n const data = await this.client.uploadPart(params)\n return { ETag: data.ETag, PartNumber: partNumber }\n } catch (error) {\n throwError({ ...MediaCloudErrors.TUS_UPLOAD_ERROR, cause: error })\n throw error\n } finally {\n permit()\n }\n }\n\n /**\n * Uploads a stream to s3 using multiple parts\n * @param args - The function arguments\n * @param args.metadata - The upload metadata\n * @param args.readStream - The stream to read data from\n * @param args.currentPartNumber - The current part number to start from\n * @param args.offset - The byte offset to start from\n * @returns Promise that resolves to the number of bytes uploaded\n */\n async uploadParts(args: UploadPartsArgs): Promise<number> {\n const { metadata, readStream, offset: initialOffset } = args\n let { currentPartNumber } = args\n let offset = initialOffset\n const size = metadata.file.size\n const promises: Promise<void>[] = []\n let pendingChunkFilepath: null | string = null\n let bytesUploaded = 0\n let permit: SemaphorePermit | undefined = undefined\n\n const splitterStream = new StreamSplitter({\n chunkSize: this.fileOperations.calculateOptimalPartSize({ size }),\n directory: os.tmpdir(),\n })\n .on('beforeChunkStarted', async () => {\n permit = await this.partUploadSemaphore.acquire()\n })\n .on('chunkStarted', (filepath) => {\n pendingChunkFilepath = filepath\n })\n .on('chunkFinished', ({ path, size: partSize }) => {\n pendingChunkFilepath = null\n\n const acquiredPermit = permit\n const partNumber = currentPartNumber++\n\n offset += partSize\n\n const isFinalPart = size === offset\n\n const uploadChunk = async () => {\n try {\n // Only the first chunk of each PATCH request can prepend\n // an incomplete part (last chunk) from the previous request.\n const readable = fs.createReadStream(path)\n readable.on('error', function (error) {\n throw error\n })\n\n switch (true) {\n case partSize >= this.minPartSize || isFinalPart:\n await this.uploadPart({\n metadata,\n readStream: readable,\n partNumber,\n })\n break\n default:\n await this.uploadIncompletePart({\n id: metadata.file.id,\n readStream: readable,\n })\n break\n }\n\n bytesUploaded += partSize\n } catch (error) {\n // Destroy the splitter to stop processing more chunks\n const mappedError =\n error instanceof Error ? error : new Error(String(error))\n splitterStream.destroy(mappedError)\n throw mappedError\n } finally {\n fs.promises.rm(path).catch(function () {})\n acquiredPermit?.()\n }\n }\n\n const deferred = uploadChunk()\n\n promises.push(deferred)\n })\n .on('chunkError', () => {\n permit?.()\n })\n\n try {\n await stream.promises.pipeline(readStream, splitterStream)\n } catch (error) {\n if (pendingChunkFilepath !== null) {\n try {\n await fs.promises.rm(pendingChunkFilepath)\n } catch {\n log(MediaCloudLogs.S3_STORE_CHUNK_REMOVAL_FAILED)\n }\n }\n const mappedError =\n error instanceof Error ? error : new Error(String(error))\n promises.push(Promise.reject(mappedError))\n } finally {\n // Wait for all promises\n await Promise.allSettled(promises)\n // Reject the promise if any of the promises reject\n await Promise.all(promises)\n }\n\n return bytesUploaded\n }\n}\n"],"mappings":";;;;;;;;;AA6DA,MAAM,EAAEsC,KAAKC,eAAehC,iBAAiB;AAE7C,IAAaiC,iBAAb,MAA4B;CAC1BC,YACE,AAAQC,QACR,AAAQC,QACR,AAAQC,aACR,AAAQC,qBACR,AAAQC,iBACR,AAAQC,gBACR,AAAQC,qBACR;EAPQN;EACAC;EACAC;EACAC;EACAC;EACAC;EACAC;;;;;;;;;;CAWV,MAAME,cAAcC,MAAmD;EACrE,MAAM,EAAEhC,IAAIC,qBAAqB+B;EACjC,MAAM7B,WAAW,MAAM,KAAKwB,gBAAgBO,YAAY,EAAElC,IAAI,CAAC;AAE/D,MAAI,CAACG,SAAS,cAAc;AAC1BiB,cAAW/B,iBAAiB8C,sBAAsB;AAClD,SAAM,IAAIC,OAAO;;EAGnB,MAAMC,SAAoC;GACxCE,QAAQ,KAAKf;GACbgB,KAAK,KAAKb,gBAAgBc,gBAAgB;IACxCzC;IACA0C,QAAQvC,SAASwC,KAAKxC,UAAUuC,UAAUE;IAC3C,CAAC;GACFC,kBAAkB5C;GAClB6C,UAAU3C,SAAS;GACpB;EAED,MAAM4C,OAAO,MAAM,KAAKxB,OAAOyB,UAAUX,OAAO;EAEhD,IAAIjC,QAAQ2C,KAAKE,SAAS,EAAE;AAE5B,MAAIF,KAAKG,aAAa;GACpB,MAAMC,OAAO,MAAM,KAAKpB,cAAc;IACpC/B;IACAC,kBAAkB8C,KAAKK;IACxB,CAAC;AACFhD,WAAQ,CAAC,GAAGA,OAAO,GAAG+C,KAAK;;AAG7B,MAAI,CAAClD,iBACHG,OAAMiD,MAAMC,GAAGC,OAAOD,EAAEE,cAAc,MAAMD,EAAEC,cAAc,GAAG;AAGjE,SAAOpD;;;;;;;;;;CAWT,MAAMqD,sBACJzB,MAC6B;EAC7B,MAAM,EAAE7B,UAAUC,UAAU4B;EAE5B,MAAMK,SAAS;GACbG,KAAK,KAAKb,gBAAgBc,gBAAgB;IACxCzC,IAAIG,SAASwC,KAAK3C;IAClB0C,QAAQvC,SAASwC,KAAKxC,UAAUuC,UAAUE;IAC3C,CAAC;GACFL,QAAQ,KAAKf;GACbkC,iBAAiB,EACfT,OAAO7C,MAAMuD,KAAKC,SAAS;AACzB,WAAO;KACLC,MAAMD,KAAKC;KACXL,YAAYI,KAAKJ;KAClB;KACF,EACF;GACDV,UAAU3C,SAAS;GACpB;AAED,MAAI;AAEF,WADe,MAAM,KAAKoB,OAAOwC,wBAAwB1B,OAAO,EAClD2B;WACPC,OAAO;AACd7C,cAAW;IAAE,GAAG/B,iBAAiB6E;IAAkBC,OAAOF;IAAO,CAAC;AAClE,SAAM,IAAI7B,OAAO;;;;;;;;;CAUrB,MAAMgC,kBACJpC,MAC+B;EAC/B,MAAM,EAAEhC,OAAOgC;AAEf,MAAI;GACF,MAAM,EAAEW,SAAS,MAAM,KAAKhB,gBAAgBO,YAAY,EAAElC,IAAI,CAAC;AAU/D,WARa,MAAM,KAAKuB,OAAO8C,UAAU;IACvC9B,QAAQ,KAAKf;IACbgB,KAAK,KAAKb,gBAAgBc,gBAAgB;KACxCzC;KACA0C,QAAQC,MAAMxC,UAAUuC,UAAUE;KAClC0B,cAAc;KACf,CAAA;IACF,CAAC,EACUC;WACLN,OAAO;AACd,OAAIA,iBAAiBjF,UACnB;AAEF,SAAMiF;;;;;;;;;CAUV,MAAMO,sBACJxC,MAC6B;EAC7B,MAAM,EAAEhC,OAAOgC;AAEf,MAAI;GACF,MAAM,EAAEW,SAAS,MAAM,KAAKhB,gBAAgBO,YAAY,EAAElC,IAAI,CAAC;AAU/D,WARa,MAAM,KAAKuB,OAAOkD,WAAW;IACxClC,QAAQ,KAAKf;IACbgB,KAAK,KAAKb,gBAAgBc,gBAAgB;KACxCzC;KACAsE,cAAc;KACd5B,QAAQC,MAAMxC,UAAUuC,UAAUE;KACnC,CAAA;IACF,CAAC,EACU8B;WACLT,OAAO;AACd,OAAIA,iBAAiBhF,SACnB;AAEF,SAAMgF;;;;;;;;;CAUV,MAAMU,qBAAqB3C,MAA+C;EACxE,MAAM,EAAEhC,OAAOgC;AAEf,MAAI;GACF,MAAM,EAAEW,SAAS,MAAM,KAAKhB,gBAAgBO,YAAY,EAAElC,IAAI,CAAC;AAE/D,SAAM,KAAKuB,OAAOqD,aAAa;IAC7BrC,QAAQ,KAAKf;IACbgB,KAAK,KAAKb,gBAAgBc,gBAAgB;KACxCzC;KACAsE,cAAc;KACd5B,QAAQC,MAAMxC,UAAUuC,UAAUE;KACnC,CAAA;IACF,CAAC;WACKqB,OAAO;AACd7C,cAAW;IACT,GAAG/B,iBAAiBwF;IACpBV,OAAOF;IACR,CAAC;;;;;;;;;CAUN,MAAMa,uBACJ9C,MACyC;EACzC,MAAM,EAAEhC,OAAOgC;EACf,MAAM+C,iBAAiB,MAAM,KAAKX,kBAAkB,EAAEpE,IAAI,CAAC;AAE3D,MAAI,CAAC+E,eACH;EAEF,MAAMC,WAAW,MAAM,KAAKpD,eAAeqD,0BAA0B,EACnEC,UAAU,2BACX,CAAC;AAEF,MAAI;GACF,IAAIC,qBAAqB;GAEzB,MAAMC,uBAAuB,IAAIrG,OAAOsG,UAAU,EAChDC,UAAUC,OAAOC,GAAGC,UAAU;AAC5BN,0BAAsBI,MAAMG;AAC5BD,aAAS,MAAMF,MAAM;MAExB,CAAC;AAGF,SAAMxG,OAAO4G,SAASC,SACpBb,gBACAK,sBACAvG,GAAGgH,kBAAkBb,SACvB,CAAC;GAED,MAAMc,oBAAoBC,YAAuC;IAC/D,MAAME,aAAapH,GAAGiH,iBAAiBd,SAAS;AAEhD,QAAIe,QAAQC,cAAc;AACxBC,gBAAWC,GAAG,aAAa;AACzBrH,SAAGsH,OAAOnB,gBAAgB,GAAG;OAC7B;AAEFiB,gBAAWC,GAAG,UAAUE,QAAQ;AAC9BH,iBAAWI,QAAQD,IAAI;AACvBvH,SAAGsH,OAAOnB,gBAAgB,GAAG;OAC7B;;AAGJ,WAAOiB;;AAGT,UAAO;IACLK,cAAcR;IACdS,MAAMvB;IACNwB,MAAMrB;IACP;WACMiB,KAAK;AACZvH,MAAG8G,SAASc,GAAGzB,SAAS,CAAC0B,YAAY,GAAG;AACxC,SAAMN;;;;;;;;;;CAWV,MAAMO,qBAAqB3E,MAAiD;EAC1E,MAAM,EAAEhC,IAAIY,eAAeoB;AAC3B,MAAI;GACF,MAAM,EAAEW,SAAS,MAAM,KAAKhB,gBAAgBO,YAAY,EAAElC,IAAI,CAAC;GAE/D,MAAM+C,OAAO,MAAM,KAAKxB,OAAOqF,UAAU;IACvCrC,MAAM3D;IACN2B,QAAQ,KAAKf;IACbgB,KAAK,KAAKb,gBAAgBc,gBAAgB;KACxCzC;KACAsE,cAAc;KACd5B,QAAQC,MAAMxC,UAAUuC,UAAUE;KACnC,CAAC;IACFiE,SAAS,KAAKhF,oBAAoB,QAAO;IAC1C,CAAC;AACFV,OAAI7B,eAAewH,kCAAkC;AACrD,UAAO/D,KAAKc;WACLI,OAAO;AACd7C,cAAW;IAAE,GAAG/B,iBAAiB6E;IAAkBC,OAAOF;IAAO,CAAC;AAClE,SAAMA;;;;;;;;;;;CAYV,MAAM8C,WAAW/E,MAAyC;EACxD,MAAM,EAAE7B,UAAUS,YAAYG,eAAeiB;EAC7C,MAAMgF,SAAS,MAAM,KAAKtF,oBAAoBuF,SAAS;AAEvD,MAAI,CAAC9G,SAAS,cAAc;AAC1BiB,cAAW/B,iBAAiB8C,sBAAsB;AAClD,SAAM,IAAIC,OAAO;;EAGnB,MAAMC,SAAqC;GACzCkC,MAAM3D;GACN2B,QAAQ,KAAKf;GACbgB,KAAK,KAAKb,gBAAgBc,gBAAgB;IACxCzC,IAAIG,SAASwC,KAAK3C;IAClB0C,QAAQvC,SAASwC,KAAKxC,UAAUuC,UAAUE;IAC3C,CAAC;GACFY,YAAYzC;GACZ+B,UAAU3C,SAAS;GACpB;AAED,MAAI;AAEF,UAAO;IAAE0D,OADI,MAAM,KAAKtC,OAAOwF,WAAW1E,OAAO,EAC7BwB;IAAML,YAAYzC;IAAY;WAC3CkD,OAAO;AACd7C,cAAW;IAAE,GAAG/B,iBAAiB6E;IAAkBC,OAAOF;IAAO,CAAC;AAClE,SAAMA;YACE;AACR+C,WAAQ;;;;;;;;;;;;CAaZ,MAAMG,YAAYnF,MAAwC;EACxD,MAAM,EAAE7B,UAAUS,YAAYM,QAAQkG,kBAAkBpF;EACxD,IAAI,EAAEf,sBAAsBe;EAC5B,IAAId,SAASkG;EACb,MAAMZ,OAAOrG,SAASwC,KAAK6D;EAC3B,MAAMb,WAA4B,EAAE;EACpC,IAAI0B,uBAAsC;EAC1C,IAAIC,gBAAgB;EACpB,IAAIN,SAAsCpE;EAE1C,MAAM2E,iBAAiB,IAAIpI,eAAe;GACxCqI,WAAW,KAAK5F,eAAe6F,yBAAyB,EAAEjB,MAAM,CAAC;GACjEkB,WAAW5I,GAAG6I,QAAO;GACtB,CAAC,CACCzB,GAAG,sBAAsB,YAAY;AACpCc,YAAS,MAAM,KAAKtF,oBAAoBuF,SAAS;IACjD,CACDf,GAAG,iBAAiB0B,aAAa;AAChCP,0BAAuBO;IACvB,CACD1B,GAAG,kBAAkB,EAAEK,MAAMC,MAAMqB,eAAe;AACjDR,0BAAuB;GAEvB,MAAMS,iBAAiBd;GACvB,MAAMjG,aAAaE;AAEnBC,aAAU2G;GAEV,MAAME,cAAcvB,SAAStF;GAE7B,MAAM8G,cAAc,YAAY;AAC9B,QAAI;KAGF,MAAMC,WAAWpJ,GAAGiH,iBAAiBS,KAAK;AAC1C0B,cAAS/B,GAAG,SAAS,SAAUjC,OAAO;AACpC,YAAMA;OACN;AAEF,aAAQ,MAAR;MACE,KAAK4D,YAAY,KAAKpG,eAAesG;AACnC,aAAM,KAAKhB,WAAW;QACpB5G;QACAS,YAAYqH;QACZlH;QACD,CAAC;AACF;MACF;AACE,aAAM,KAAK4F,qBAAqB;QAC9B3G,IAAIG,SAASwC,KAAK3C;QAClBY,YAAYqH;QACb,CAAC;AACF;;AAGJX,sBAAiBO;aACV5D,OAAO;KAEd,MAAMiE,cACJjE,iBAAiB7B,QAAQ6B,QAAQ,IAAI7B,MAAM+F,OAAOlE,MAAM,CAAC;AAC3DsD,oBAAelB,QAAQ6B,YAAY;AACnC,WAAMA;cACE;AACRrJ,QAAG8G,SAASc,GAAGF,KAAK,CAACG,MAAM,WAAY,GAAG;AAC1CoB,uBAAkB;;;GAItB,MAAMM,WAAWJ,aAAa;AAE9BrC,YAAS0C,KAAKD,SAAS;IACvB,CACDlC,GAAG,oBAAoB;AACtBc,aAAU;IACV;AAEJ,MAAI;AACF,SAAMjI,OAAO4G,SAASC,SAAShF,YAAY2G,eAAe;WACnDtD,OAAO;AACd,OAAIoD,yBAAyB,KAC3B,KAAI;AACF,UAAMxI,GAAG8G,SAASc,GAAGY,qBAAqB;WACpC;AACNlG,QAAI7B,eAAegJ,8BAA8B;;GAGrD,MAAMJ,cACJjE,iBAAiB7B,QAAQ6B,QAAQ,IAAI7B,MAAM+F,OAAOlE,MAAM,CAAC;AAC3D0B,YAAS0C,KAAKpG,QAAQsG,OAAOL,YAAY,CAAC;YAClC;AAER,SAAMjG,QAAQuG,WAAW7C,SAAS;AAElC,SAAM1D,QAAQwG,IAAI9C,SAAS;;AAG7B,SAAO2B"}
|
|
1
|
+
{"version":3,"file":"partsManager.mjs","names":["client: S3","bucket: string","minPartSize: number","partUploadSemaphore: Semaphore","metadataManager: S3MetadataManager","fileOperations: S3FileOperations","generateCompleteTag: (value: 'false' | 'true') => string | undefined","params: AWS.ListPartsCommandInput","params: AWS.UploadPartCommandInput","promises: Promise<void>[]","pendingChunkFilepath: null | string","permit: SemaphorePermit | undefined"],"sources":["../../../../src/tus/stores/s3/partsManager.ts"],"sourcesContent":["import fs from 'node:fs'\nimport os from 'node:os'\nimport stream from 'node:stream'\n\nimport { NoSuchKey, NotFound, type S3 } from '@aws-sdk/client-s3'\nimport { StreamSplitter } from '@tus/utils'\n\nimport { useErrorHandler } from '../../../hooks/useErrorHandler'\nimport { MediaCloudErrors, MediaCloudLogs } from '../../../types/errors'\n\nimport type AWS from '@aws-sdk/client-s3'\nimport type { Readable } from 'node:stream'\nimport type { IncompletePartInfo, TusUploadMetadata } from '../../../types'\nimport type { S3FileOperations } from './fileOperations'\nimport type { S3MetadataManager } from './metadataManager'\nimport type { Semaphore, SemaphorePermit } from './semaphore'\n\ntype RetrievePartsArgs = {\n id: string\n partNumberMarker?: string\n}\n\ntype FinishMultipartUploadArgs = {\n metadata: TusUploadMetadata\n parts: Array<AWS.Part>\n}\n\ntype GetIncompletePartArgs = {\n id: string\n}\n\ntype GetIncompletePartSizeArgs = {\n id: string\n}\n\ntype DeleteIncompletePartArgs = {\n id: string\n}\n\ntype DownloadIncompletePartArgs = {\n id: string\n}\n\ntype UploadIncompletePartArgs = {\n id: string\n readStream: fs.ReadStream | Readable\n}\n\ntype UploadPartArgs = {\n metadata: TusUploadMetadata\n readStream: fs.ReadStream | Readable\n partNumber: number\n}\n\ntype UploadPartsArgs = {\n metadata: TusUploadMetadata\n readStream: stream.Readable\n currentPartNumber: number\n offset: number\n}\n\nconst { log, throwError } = useErrorHandler()\n\nexport class S3PartsManager {\n constructor(\n private client: S3,\n private bucket: string,\n private minPartSize: number,\n private partUploadSemaphore: Semaphore,\n private metadataManager: S3MetadataManager,\n private fileOperations: S3FileOperations,\n private generateCompleteTag: (value: 'false' | 'true') => string | undefined\n ) {}\n\n /**\n * Gets the number of complete parts/chunks already uploaded to S3.\n * Retrieves only consecutive parts.\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @param args.partNumberMarker - Marker for pagination (optional)\n * @returns Promise that resolves to array of uploaded parts\n */\n async retrieveParts(args: RetrievePartsArgs): Promise<Array<AWS.Part>> {\n const { id, partNumberMarker } = args\n const metadata = await this.metadataManager.getMetadata({ id })\n\n if (!metadata['upload-id']) {\n throwError(MediaCloudErrors.MUX_UPLOAD_ID_MISSING)\n throw new Error() // This will never execute but satisfies TypeScript\n }\n\n const params: AWS.ListPartsCommandInput = {\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: metadata.file.metadata?.prefix ?? undefined,\n }),\n PartNumberMarker: partNumberMarker,\n UploadId: metadata['upload-id'],\n }\n\n const data = await this.client.listParts(params)\n\n let parts = data.Parts ?? []\n\n if (data.IsTruncated) {\n const rest = await this.retrieveParts({\n id,\n partNumberMarker: data.NextPartNumberMarker,\n })\n parts = [...parts, ...rest]\n }\n\n if (!partNumberMarker) {\n parts.sort((a, b) => (a.PartNumber || 0) - (b.PartNumber || 0))\n }\n\n return parts\n }\n\n /**\n * Completes a multipart upload on S3.\n * This is where S3 concatenates all the uploaded parts.\n * @param args - The function arguments\n * @param args.metadata - The upload metadata\n * @param args.parts - Array of uploaded parts to complete\n * @returns Promise that resolves to the location URL (optional)\n */\n async finishMultipartUpload(\n args: FinishMultipartUploadArgs\n ): Promise<string | undefined> {\n const { metadata, parts } = args\n\n const params = {\n Key: this.metadataManager.generatePartKey({\n id: metadata.file.id,\n prefix: metadata.file.metadata?.prefix ?? undefined,\n }),\n Bucket: this.bucket,\n MultipartUpload: {\n Parts: parts.map((part) => {\n return {\n ETag: part.ETag,\n PartNumber: part.PartNumber,\n }\n }),\n },\n UploadId: metadata['upload-id'],\n }\n\n try {\n const result = await this.client.completeMultipartUpload(params)\n return result.Location\n } catch (error) {\n throwError({ ...MediaCloudErrors.TUS_UPLOAD_ERROR, cause: error })\n throw new Error() // This will never execute but satisfies TypeScript\n }\n }\n\n /**\n * Gets incomplete part from S3\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves to readable stream or undefined if not found\n */\n async getIncompletePart(\n args: GetIncompletePartArgs\n ): Promise<Readable | undefined> {\n const { id } = args\n\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n const data = await this.client.getObject({\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n isIncomplete: true,\n }),\n })\n return data.Body as Readable\n } catch (error) {\n if (error instanceof NoSuchKey) {\n return undefined\n }\n throw error\n }\n }\n\n /**\n * Gets the size of an incomplete part\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves to part size or undefined if not found\n */\n async getIncompletePartSize(\n args: GetIncompletePartSizeArgs\n ): Promise<number | undefined> {\n const { id } = args\n\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n const data = await this.client.headObject({\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n isIncomplete: true,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n })\n return data.ContentLength\n } catch (error) {\n if (error instanceof NotFound) {\n return undefined\n }\n throw error\n }\n }\n\n /**\n * Deletes an incomplete part\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves when deletion is complete\n */\n async deleteIncompletePart(args: DeleteIncompletePartArgs): Promise<void> {\n const { id } = args\n\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n await this.client.deleteObject({\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n isIncomplete: true,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n })\n } catch (error) {\n throwError({\n ...MediaCloudErrors.S3_DELETE_ERROR,\n cause: error,\n })\n }\n }\n\n /**\n * Downloads incomplete part to temporary file\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @returns Promise that resolves to incomplete part info or undefined if not found\n */\n async downloadIncompletePart(\n args: DownloadIncompletePartArgs\n ): Promise<IncompletePartInfo | undefined> {\n const { id } = args\n const incompletePart = await this.getIncompletePart({ id })\n\n if (!incompletePart) {\n return\n }\n const filePath = await this.fileOperations.generateUniqueTmpFileName({\n template: 'tus-s3-incomplete-part-',\n })\n\n try {\n let incompletePartSize = 0\n\n const byteCounterTransform = new stream.Transform({\n transform(chunk, _, callback) {\n incompletePartSize += chunk.length\n callback(null, chunk)\n },\n })\n\n // Write to temporary file\n await stream.promises.pipeline(\n incompletePart,\n byteCounterTransform,\n fs.createWriteStream(filePath)\n )\n\n const createReadStream = (options: { cleanUpOnEnd: boolean }) => {\n const fileReader = fs.createReadStream(filePath)\n\n if (options.cleanUpOnEnd) {\n fileReader.on('end', () => {\n fs.unlink(filePath, () => {})\n })\n\n fileReader.on('error', (err) => {\n fileReader.destroy(err)\n fs.unlink(filePath, () => {})\n })\n }\n\n return fileReader\n }\n\n return {\n createReader: createReadStream,\n path: filePath,\n size: incompletePartSize,\n }\n } catch (err) {\n fs.promises.rm(filePath).catch(() => {})\n throw err\n }\n }\n\n /**\n * Uploads an incomplete part\n * @param args - The function arguments\n * @param args.id - The upload ID\n * @param args.readStream - The stream to read data from\n * @returns Promise that resolves to the ETag of the uploaded part\n */\n async uploadIncompletePart(args: UploadIncompletePartArgs): Promise<string> {\n const { id, readStream } = args\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n const data = await this.client.putObject({\n Body: readStream,\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id,\n isIncomplete: true,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n Tagging: this.generateCompleteTag('false'),\n })\n log(MediaCloudLogs.S3_STORE_INCOMPLETE_PART_UPLOADED)\n return data.ETag as string\n } catch (error) {\n throwError({ ...MediaCloudErrors.TUS_UPLOAD_ERROR, cause: error })\n throw error\n }\n }\n\n /**\n * Uploads a single part\n * @param args - The function arguments\n * @param args.metadata - The upload metadata\n * @param args.readStream - The stream to read data from\n * @param args.partNumber - The part number to upload\n * @returns Promise that resolves to the ETag of the uploaded part\n */\n async uploadPart(args: UploadPartArgs): Promise<AWS.Part> {\n const { metadata, readStream, partNumber } = args\n const permit = await this.partUploadSemaphore.acquire()\n\n if (!metadata['upload-id']) {\n throwError(MediaCloudErrors.MUX_UPLOAD_ID_MISSING)\n throw new Error() // This will never execute but satisfies TypeScript\n }\n\n const params: AWS.UploadPartCommandInput = {\n Body: readStream,\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id: metadata.file.id,\n prefix: metadata.file.metadata?.prefix ?? undefined,\n }),\n PartNumber: partNumber,\n UploadId: metadata['upload-id'],\n }\n\n try {\n const data = await this.client.uploadPart(params)\n return { ETag: data.ETag, PartNumber: partNumber }\n } catch (error) {\n throwError({ ...MediaCloudErrors.TUS_UPLOAD_ERROR, cause: error })\n throw error\n } finally {\n permit()\n }\n }\n\n /**\n * Uploads a stream to s3 using multiple parts\n * @param args - The function arguments\n * @param args.metadata - The upload metadata\n * @param args.readStream - The stream to read data from\n * @param args.currentPartNumber - The current part number to start from\n * @param args.offset - The byte offset to start from\n * @returns Promise that resolves to the number of bytes uploaded\n */\n async uploadParts(args: UploadPartsArgs): Promise<number> {\n const { metadata, readStream, offset: initialOffset } = args\n let { currentPartNumber } = args\n let offset = initialOffset\n const size = metadata.file.size\n const promises: Promise<void>[] = []\n let pendingChunkFilepath: null | string = null\n let bytesUploaded = 0\n let permit: SemaphorePermit | undefined = undefined\n\n const splitterStream = new StreamSplitter({\n chunkSize: this.fileOperations.calculateOptimalPartSize({ size }),\n directory: os.tmpdir(),\n })\n .on('beforeChunkStarted', async () => {\n permit = await this.partUploadSemaphore.acquire()\n })\n .on('chunkStarted', (filepath) => {\n pendingChunkFilepath = filepath\n })\n .on('chunkFinished', ({ path, size: partSize }) => {\n pendingChunkFilepath = null\n\n const acquiredPermit = permit\n const partNumber = currentPartNumber++\n\n offset += partSize\n\n const isFinalPart = size === offset\n\n const uploadChunk = async () => {\n try {\n // Only the first chunk of each PATCH request can prepend\n // an incomplete part (last chunk) from the previous request.\n const readable = fs.createReadStream(path)\n readable.on('error', function (error) {\n throw error\n })\n\n switch (true) {\n case partSize >= this.minPartSize || isFinalPart:\n await this.uploadPart({\n metadata,\n readStream: readable,\n partNumber,\n })\n break\n default:\n await this.uploadIncompletePart({\n id: metadata.file.id,\n readStream: readable,\n })\n break\n }\n\n bytesUploaded += partSize\n } catch (error) {\n // Destroy the splitter to stop processing more chunks\n const mappedError =\n error instanceof Error ? error : new Error(String(error))\n splitterStream.destroy(mappedError)\n throw mappedError\n } finally {\n fs.promises.rm(path).catch(function () {})\n acquiredPermit?.()\n }\n }\n\n const deferred = uploadChunk()\n\n promises.push(deferred)\n })\n .on('chunkError', () => {\n permit?.()\n })\n\n try {\n await stream.promises.pipeline(readStream, splitterStream)\n } catch (error) {\n if (pendingChunkFilepath !== null) {\n try {\n await fs.promises.rm(pendingChunkFilepath)\n } catch {\n log(MediaCloudLogs.S3_STORE_CHUNK_REMOVAL_FAILED)\n }\n }\n const mappedError =\n error instanceof Error ? error : new Error(String(error))\n promises.push(Promise.reject(mappedError))\n } finally {\n // Wait for all promises\n await Promise.allSettled(promises)\n // Reject the promise if any of the promises reject\n await Promise.all(promises)\n }\n\n return bytesUploaded\n }\n}\n"],"mappings":";;;;;;;;;AA6DA,MAAM,EAAE,KAAK,eAAe,iBAAiB;AAE7C,IAAa,iBAAb,MAA4B;CAC1B,YACE,AAAQA,QACR,AAAQC,QACR,AAAQC,aACR,AAAQC,qBACR,AAAQC,iBACR,AAAQC,gBACR,AAAQC,qBACR;EAPQ;EACA;EACA;EACA;EACA;EACA;EACA;;;;;;;;;;CAWV,MAAM,cAAc,MAAmD;EACrE,MAAM,EAAE,IAAI,qBAAqB;EACjC,MAAM,WAAW,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;AAE/D,MAAI,CAAC,SAAS,cAAc;AAC1B,cAAW,iBAAiB,sBAAsB;AAClD,SAAM,IAAI,OAAO;;EAGnB,MAAMC,SAAoC;GACxC,QAAQ,KAAK;GACb,KAAK,KAAK,gBAAgB,gBAAgB;IACxC;IACA,QAAQ,SAAS,KAAK,UAAU,UAAU;IAC3C,CAAC;GACF,kBAAkB;GAClB,UAAU,SAAS;GACpB;EAED,MAAM,OAAO,MAAM,KAAK,OAAO,UAAU,OAAO;EAEhD,IAAI,QAAQ,KAAK,SAAS,EAAE;AAE5B,MAAI,KAAK,aAAa;GACpB,MAAM,OAAO,MAAM,KAAK,cAAc;IACpC;IACA,kBAAkB,KAAK;IACxB,CAAC;AACF,WAAQ,CAAC,GAAG,OAAO,GAAG,KAAK;;AAG7B,MAAI,CAAC,iBACH,OAAM,MAAM,GAAG,OAAO,EAAE,cAAc,MAAM,EAAE,cAAc,GAAG;AAGjE,SAAO;;;;;;;;;;CAWT,MAAM,sBACJ,MAC6B;EAC7B,MAAM,EAAE,UAAU,UAAU;EAE5B,MAAM,SAAS;GACb,KAAK,KAAK,gBAAgB,gBAAgB;IACxC,IAAI,SAAS,KAAK;IAClB,QAAQ,SAAS,KAAK,UAAU,UAAU;IAC3C,CAAC;GACF,QAAQ,KAAK;GACb,iBAAiB,EACf,OAAO,MAAM,KAAK,SAAS;AACzB,WAAO;KACL,MAAM,KAAK;KACX,YAAY,KAAK;KAClB;KACD,EACH;GACD,UAAU,SAAS;GACpB;AAED,MAAI;AAEF,WADe,MAAM,KAAK,OAAO,wBAAwB,OAAO,EAClD;WACP,OAAO;AACd,cAAW;IAAE,GAAG,iBAAiB;IAAkB,OAAO;IAAO,CAAC;AAClE,SAAM,IAAI,OAAO;;;;;;;;;CAUrB,MAAM,kBACJ,MAC+B;EAC/B,MAAM,EAAE,OAAO;AAEf,MAAI;GACF,MAAM,EAAE,SAAS,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;AAU/D,WARa,MAAM,KAAK,OAAO,UAAU;IACvC,QAAQ,KAAK;IACb,KAAK,KAAK,gBAAgB,gBAAgB;KACxC;KACA,QAAQ,MAAM,UAAU,UAAU;KAClC,cAAc;KACf,CAAC;IACH,CAAC,EACU;WACL,OAAO;AACd,OAAI,iBAAiB,UACnB;AAEF,SAAM;;;;;;;;;CAUV,MAAM,sBACJ,MAC6B;EAC7B,MAAM,EAAE,OAAO;AAEf,MAAI;GACF,MAAM,EAAE,SAAS,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;AAU/D,WARa,MAAM,KAAK,OAAO,WAAW;IACxC,QAAQ,KAAK;IACb,KAAK,KAAK,gBAAgB,gBAAgB;KACxC;KACA,cAAc;KACd,QAAQ,MAAM,UAAU,UAAU;KACnC,CAAC;IACH,CAAC,EACU;WACL,OAAO;AACd,OAAI,iBAAiB,SACnB;AAEF,SAAM;;;;;;;;;CAUV,MAAM,qBAAqB,MAA+C;EACxE,MAAM,EAAE,OAAO;AAEf,MAAI;GACF,MAAM,EAAE,SAAS,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;AAE/D,SAAM,KAAK,OAAO,aAAa;IAC7B,QAAQ,KAAK;IACb,KAAK,KAAK,gBAAgB,gBAAgB;KACxC;KACA,cAAc;KACd,QAAQ,MAAM,UAAU,UAAU;KACnC,CAAC;IACH,CAAC;WACK,OAAO;AACd,cAAW;IACT,GAAG,iBAAiB;IACpB,OAAO;IACR,CAAC;;;;;;;;;CAUN,MAAM,uBACJ,MACyC;EACzC,MAAM,EAAE,OAAO;EACf,MAAM,iBAAiB,MAAM,KAAK,kBAAkB,EAAE,IAAI,CAAC;AAE3D,MAAI,CAAC,eACH;EAEF,MAAM,WAAW,MAAM,KAAK,eAAe,0BAA0B,EACnE,UAAU,2BACX,CAAC;AAEF,MAAI;GACF,IAAI,qBAAqB;GAEzB,MAAM,uBAAuB,IAAI,OAAO,UAAU,EAChD,UAAU,OAAO,GAAG,UAAU;AAC5B,0BAAsB,MAAM;AAC5B,aAAS,MAAM,MAAM;MAExB,CAAC;AAGF,SAAM,OAAO,SAAS,SACpB,gBACA,sBACA,GAAG,kBAAkB,SAAS,CAC/B;GAED,MAAM,oBAAoB,YAAuC;IAC/D,MAAM,aAAa,GAAG,iBAAiB,SAAS;AAEhD,QAAI,QAAQ,cAAc;AACxB,gBAAW,GAAG,aAAa;AACzB,SAAG,OAAO,gBAAgB,GAAG;OAC7B;AAEF,gBAAW,GAAG,UAAU,QAAQ;AAC9B,iBAAW,QAAQ,IAAI;AACvB,SAAG,OAAO,gBAAgB,GAAG;OAC7B;;AAGJ,WAAO;;AAGT,UAAO;IACL,cAAc;IACd,MAAM;IACN,MAAM;IACP;WACM,KAAK;AACZ,MAAG,SAAS,GAAG,SAAS,CAAC,YAAY,GAAG;AACxC,SAAM;;;;;;;;;;CAWV,MAAM,qBAAqB,MAAiD;EAC1E,MAAM,EAAE,IAAI,eAAe;AAC3B,MAAI;GACF,MAAM,EAAE,SAAS,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;GAE/D,MAAM,OAAO,MAAM,KAAK,OAAO,UAAU;IACvC,MAAM;IACN,QAAQ,KAAK;IACb,KAAK,KAAK,gBAAgB,gBAAgB;KACxC;KACA,cAAc;KACd,QAAQ,MAAM,UAAU,UAAU;KACnC,CAAC;IACF,SAAS,KAAK,oBAAoB,QAAQ;IAC3C,CAAC;AACF,OAAI,eAAe,kCAAkC;AACrD,UAAO,KAAK;WACL,OAAO;AACd,cAAW;IAAE,GAAG,iBAAiB;IAAkB,OAAO;IAAO,CAAC;AAClE,SAAM;;;;;;;;;;;CAYV,MAAM,WAAW,MAAyC;EACxD,MAAM,EAAE,UAAU,YAAY,eAAe;EAC7C,MAAM,SAAS,MAAM,KAAK,oBAAoB,SAAS;AAEvD,MAAI,CAAC,SAAS,cAAc;AAC1B,cAAW,iBAAiB,sBAAsB;AAClD,SAAM,IAAI,OAAO;;EAGnB,MAAMC,SAAqC;GACzC,MAAM;GACN,QAAQ,KAAK;GACb,KAAK,KAAK,gBAAgB,gBAAgB;IACxC,IAAI,SAAS,KAAK;IAClB,QAAQ,SAAS,KAAK,UAAU,UAAU;IAC3C,CAAC;GACF,YAAY;GACZ,UAAU,SAAS;GACpB;AAED,MAAI;AAEF,UAAO;IAAE,OADI,MAAM,KAAK,OAAO,WAAW,OAAO,EAC7B;IAAM,YAAY;IAAY;WAC3C,OAAO;AACd,cAAW;IAAE,GAAG,iBAAiB;IAAkB,OAAO;IAAO,CAAC;AAClE,SAAM;YACE;AACR,WAAQ;;;;;;;;;;;;CAaZ,MAAM,YAAY,MAAwC;EACxD,MAAM,EAAE,UAAU,YAAY,QAAQ,kBAAkB;EACxD,IAAI,EAAE,sBAAsB;EAC5B,IAAI,SAAS;EACb,MAAM,OAAO,SAAS,KAAK;EAC3B,MAAMC,WAA4B,EAAE;EACpC,IAAIC,uBAAsC;EAC1C,IAAI,gBAAgB;EACpB,IAAIC,SAAsC;EAE1C,MAAM,iBAAiB,IAAI,eAAe;GACxC,WAAW,KAAK,eAAe,yBAAyB,EAAE,MAAM,CAAC;GACjE,WAAW,GAAG,QAAQ;GACvB,CAAC,CACC,GAAG,sBAAsB,YAAY;AACpC,YAAS,MAAM,KAAK,oBAAoB,SAAS;IACjD,CACD,GAAG,iBAAiB,aAAa;AAChC,0BAAuB;IACvB,CACD,GAAG,kBAAkB,EAAE,MAAM,MAAM,eAAe;AACjD,0BAAuB;GAEvB,MAAM,iBAAiB;GACvB,MAAM,aAAa;AAEnB,aAAU;GAEV,MAAM,cAAc,SAAS;GAE7B,MAAM,cAAc,YAAY;AAC9B,QAAI;KAGF,MAAM,WAAW,GAAG,iBAAiB,KAAK;AAC1C,cAAS,GAAG,SAAS,SAAU,OAAO;AACpC,YAAM;OACN;AAEF,aAAQ,MAAR;MACE,KAAK,YAAY,KAAK,eAAe;AACnC,aAAM,KAAK,WAAW;QACpB;QACA,YAAY;QACZ;QACD,CAAC;AACF;MACF;AACE,aAAM,KAAK,qBAAqB;QAC9B,IAAI,SAAS,KAAK;QAClB,YAAY;QACb,CAAC;AACF;;AAGJ,sBAAiB;aACV,OAAO;KAEd,MAAM,cACJ,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,MAAM,CAAC;AAC3D,oBAAe,QAAQ,YAAY;AACnC,WAAM;cACE;AACR,QAAG,SAAS,GAAG,KAAK,CAAC,MAAM,WAAY,GAAG;AAC1C,uBAAkB;;;GAItB,MAAM,WAAW,aAAa;AAE9B,YAAS,KAAK,SAAS;IACvB,CACD,GAAG,oBAAoB;AACtB,aAAU;IACV;AAEJ,MAAI;AACF,SAAM,OAAO,SAAS,SAAS,YAAY,eAAe;WACnD,OAAO;AACd,OAAI,yBAAyB,KAC3B,KAAI;AACF,UAAM,GAAG,SAAS,GAAG,qBAAqB;WACpC;AACN,QAAI,eAAe,8BAA8B;;GAGrD,MAAM,cACJ,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO,MAAM,CAAC;AAC3D,YAAS,KAAK,QAAQ,OAAO,YAAY,CAAC;YAClC;AAER,SAAM,QAAQ,WAAW,SAAS;AAElC,SAAM,QAAQ,IAAI,SAAS;;AAG7B,SAAO"}
|
|
@@ -115,10 +115,10 @@ var S3Store = class extends DataStore {
|
|
|
115
115
|
if (incompletePart.size !== offsetDiff) throw ERRORS.FILE_WRITE_ERROR;
|
|
116
116
|
await this.partsManager.deleteIncompletePart({ id });
|
|
117
117
|
offset = requestedOffset - incompletePart.size;
|
|
118
|
-
finalReadable = stream.Readable.from(async function* () {
|
|
118
|
+
finalReadable = stream.Readable.from((async function* () {
|
|
119
119
|
yield* incompletePart.createReader({ cleanUpOnEnd: true });
|
|
120
120
|
yield* readable;
|
|
121
|
-
}());
|
|
121
|
+
})());
|
|
122
122
|
}
|
|
123
123
|
const partNumber = this.fileOperations.calculatePartNumber({ parts: await this.partsManager.retrieveParts({ id }) });
|
|
124
124
|
const newOffset = requestedOffset + await this.partsManager.uploadParts({
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"s3Store.mjs","names":["NoSuchKey","NotFound","S3","DataStore","ERRORS","TUS_RESUMABLE","Upload","stream","Readable","useErrorHandler","Semaphore","S3ExpirationManager","S3FileOperations","S3MetadataManager","S3PartsManager","MediaCloudLogs","AWS","AWSError","S3StoreConfig","TusUploadMetadata","log","S3Store","client","bucket","partSize","minPartSize","maxMultipartParts","maxUploadSize","const","useTags","expirationPeriodInMilliseconds","acl","partUploadSemaphore","metadataManager","fileOperations","partsManager","expirationManager","customEndpoint","constructor","options","s3ClientConfig","maxConcurrentPartUploads","restS3ClientConfig","extensions","String","endpoint","shouldUseExpirationTags","bind","generateCompleteTag","generateInfoKey","generatePartKey","deleteExpired","value","undefined","create","upload","Promise","S3_STORE_MULTIPART_INIT","request","CreateMultipartUploadCommandInput","Bucket","Key","id","prefix","metadata","Metadata","contentType","ContentType","cacheControl","CacheControl","ACL","ObjectCannedACL","creation_date","Date","toISOString","response","createMultipartUpload","storage","type","path","saveMetadata","uploadId","UploadId","S3_STORE_MULTIPART_CREATED","write","readable","offset","getMetadata","calculatedOffset","calculateOffsetFromParts","parts","retrieveParts","offsetDiff","requestedOffset","finalReadable","FILE_WRITE_ERROR","incompletePart","downloadIncompletePart","size","deleteIncompletePart","from","createReader","cleanUpOnEnd","partNumber","calculatePartNumber","bytesUploaded","uploadParts","readStream","currentPartNumber","newOffset","file","finishMultipartUpload","completedUpload","completeMetadata","error","S3_STORE_UPLOAD_FAILED","getUpload","Code","FILE_NOT_FOUND","S3_STORE_RETRIEVE_PARTS_ERROR","incompletePartSize","getIncompletePartSize","read","S3_STORE_READ_ATTEMPT","retries","lastError","Error","data","getObject","S3_STORE_READ_SUCCESS","Body","S3_STORE_READ_RETRY","resolve","setTimeout","S3_STORE_READ_FAILED","copy","oldKey","newKey","copyObject","CopySource","encodeURI","deleteObject","code","includes","S3_STORE_FILE_NOT_FOUND","remove","abortMultipartUpload","deleteObjects","Delete","Objects","isIncomplete","cleanup","completeMultipartUpload","incompletePartInfo","uploadPart","length","updatedParts","getUrl","regionConfig","config","region","getExpiration"],"sources":["../../../../src/tus/stores/s3/s3Store.ts"],"sourcesContent":["import { NoSuchKey, NotFound, S3 } from '@aws-sdk/client-s3'\nimport { DataStore, ERRORS, TUS_RESUMABLE, Upload } from '@tus/utils'\nimport stream, { type Readable } from 'node:stream'\n\nimport { useErrorHandler } from '../../../hooks/useErrorHandler'\nimport { Semaphore } from './semaphore'\nimport { S3ExpirationManager } from './expirationManager'\nimport { S3FileOperations } from './fileOperations'\nimport { S3MetadataManager } from './metadataManager'\nimport { S3PartsManager } from './partsManager'\n\nimport { MediaCloudLogs } from '../../../types/errors'\n\nimport type AWS from '@aws-sdk/client-s3'\nimport type { AWSError, S3StoreConfig, TusUploadMetadata } from '../../../types'\n\nconst { log } = useErrorHandler()\n\nexport class S3Store extends DataStore {\n public client: S3\n public bucket: string\n public partSize = 8 * 1024 * 1024 // 8MB preferred part size\n public minPartSize = 5 * 1024 * 1024 // 5MB minimum part size\n public maxMultipartParts = 10_000\n public maxUploadSize = 5_497_558_138_880 as const // 5TiB\n public useTags = false\n public expirationPeriodInMilliseconds = 0\n protected acl?: string\n\n protected partUploadSemaphore: Semaphore\n protected metadataManager: S3MetadataManager\n protected fileOperations: S3FileOperations\n protected partsManager: S3PartsManager\n protected expirationManager: S3ExpirationManager\n protected customEndpoint: string\n\n constructor(options: S3StoreConfig) {\n super()\n const {\n maxMultipartParts,\n minPartSize,\n partSize,\n s3ClientConfig,\n maxConcurrentPartUploads,\n useTags,\n expirationPeriodInMilliseconds,\n } = options\n const { acl, bucket, ...restS3ClientConfig } = s3ClientConfig\n\n this.extensions = [\n 'creation',\n 'creation-with-upload',\n 'creation-defer-length',\n 'termination',\n 'expiration',\n ]\n\n this.bucket = bucket\n this.acl = acl\n this.client = new S3(restS3ClientConfig)\n this.customEndpoint = String(restS3ClientConfig.endpoint)\n\n this.partSize = partSize ?? this.partSize\n this.minPartSize = minPartSize ?? this.minPartSize\n this.maxMultipartParts = maxMultipartParts ?? this.maxMultipartParts\n\n this.useTags = useTags ?? this.useTags\n this.expirationPeriodInMilliseconds =\n expirationPeriodInMilliseconds ?? this.expirationPeriodInMilliseconds\n this.partUploadSemaphore = new Semaphore(maxConcurrentPartUploads ?? 60)\n\n // Initialize component managers\n this.metadataManager = new S3MetadataManager(\n this.client,\n this.bucket,\n this.shouldUseExpirationTags.bind(this),\n this.generateCompleteTag.bind(this)\n )\n\n this.fileOperations = new S3FileOperations(\n this.maxMultipartParts,\n this.maxUploadSize,\n this.minPartSize,\n this.partSize\n )\n\n this.partsManager = new S3PartsManager(\n this.client,\n this.bucket,\n this.minPartSize,\n this.partUploadSemaphore,\n this.metadataManager,\n this.fileOperations,\n this.generateCompleteTag.bind(this)\n )\n\n this.expirationManager = new S3ExpirationManager(\n this.client,\n this.bucket,\n this.expirationPeriodInMilliseconds,\n this.metadataManager.generateInfoKey.bind(this.metadataManager),\n this.metadataManager.generatePartKey.bind(this.metadataManager)\n )\n\n // Cleanup expired uploads when the store is initialized\n this.deleteExpired()\n }\n\n /**\n * Helper method to check if expiration tags should be used\n * @returns True if expiration tags should be used\n */\n protected shouldUseExpirationTags(): boolean {\n return this.expirationPeriodInMilliseconds !== 0 && this.useTags\n }\n\n /**\n * Generates a tag for marking complete/incomplete uploads\n * @param value - Either 'false' or 'true' to mark completion status\n * @returns The tag string or undefined if tags shouldn’t be used\n */\n protected generateCompleteTag(value: 'false' | 'true'): string | undefined {\n if (!this.shouldUseExpirationTags()) {\n return undefined\n }\n return `Tus-Completed=${value}`\n }\n\n /**\n * Creates a multipart upload on S3 attaching any metadata to it.\n * Also, a `${file_id}.info` file is created which holds some information\n * about the upload itself like: `upload-id`, `upload-length`, etc.\n * @param upload - The upload object to create\n * @return Promise that resolves to the created upload object with storage information\n */\n public async create(upload: Upload): Promise<Upload> {\n log(MediaCloudLogs.S3_STORE_MULTIPART_INIT)\n\n const request: AWS.CreateMultipartUploadCommandInput = {\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id: upload.id,\n prefix: upload.metadata?.prefix ?? undefined,\n }),\n Metadata: { 'tus-version': TUS_RESUMABLE },\n }\n\n if (upload.metadata?.contentType) {\n request.ContentType = upload.metadata.contentType as string\n }\n\n if (upload.metadata?.cacheControl) {\n request.CacheControl = upload.metadata.cacheControl as string\n }\n\n if (this.acl) {\n request.ACL = this.acl as AWS.ObjectCannedACL\n }\n\n upload.creation_date = new Date().toISOString()\n\n const response = await this.client.createMultipartUpload(request)\n\n upload.storage = {\n type: 's3',\n bucket: this.bucket,\n path: response.Key as string,\n }\n await this.metadataManager.saveMetadata({\n upload,\n uploadId: response.UploadId as string,\n })\n log(MediaCloudLogs.S3_STORE_MULTIPART_CREATED)\n\n return upload\n }\n\n /**\n * Writes `buffer` to the file specified by the upload’s `id` at `offset`\n * @param readable - The readable stream to write\n * @param id - The upload ID\n * @param offset - The byte offset to write at\n * @return Promise that resolves to the new offset after writing the data\n */\n public async write(\n readable: stream.Readable,\n id: string,\n offset: number\n ): Promise<number> {\n const metadata = await this.metadataManager.getMetadata({ id })\n\n // TUS sends PATCH requests with an `upload-offset` header.\n // Offset the write by the offset in the PATCH request.\n const calculatedOffset = this.fileOperations.calculateOffsetFromParts({\n parts: await this.partsManager.retrieveParts({ id }),\n })\n const offsetDiff = offset - calculatedOffset\n const requestedOffset = offset\n\n let finalReadable = readable\n\n if (offsetDiff < 0) {\n throw ERRORS.FILE_WRITE_ERROR\n }\n\n // If the offset given in the PATCH request is higher than\n // the expected offset, we need to prepend an incomplete\n // part to the readable stream, if one exists.\n if (offsetDiff > 0) {\n const incompletePart = await this.partsManager.downloadIncompletePart({\n id,\n })\n\n if (!incompletePart) {\n throw ERRORS.FILE_WRITE_ERROR\n }\n\n if (incompletePart.size !== offsetDiff) {\n throw ERRORS.FILE_WRITE_ERROR\n }\n\n // Clear the incomplete part from S3 since it's going to be combined with the current request\n await this.partsManager.deleteIncompletePart({ id })\n\n // Adjust offset to account for the incomplete part\n offset = requestedOffset - incompletePart.size\n\n finalReadable = stream.Readable.from(\n (async function* () {\n yield* incompletePart.createReader({ cleanUpOnEnd: true })\n yield* readable\n })()\n )\n }\n\n const partNumber = this.fileOperations.calculatePartNumber({\n parts: await this.partsManager.retrieveParts({ id }),\n })\n\n const bytesUploaded = await this.partsManager.uploadParts({\n metadata,\n readStream: finalReadable,\n currentPartNumber: partNumber,\n offset,\n })\n\n // The size of the incomplete part should not be counted, because the\n // process of the incomplete part should be fully transparent to the user.\n const newOffset =\n requestedOffset + bytesUploaded - (offsetDiff > 0 ? offsetDiff : 0)\n\n // Check if the upload is complete\n if (metadata.file.size === newOffset) {\n try {\n const parts = await this.partsManager.retrieveParts({ id })\n await this.partsManager.finishMultipartUpload({ metadata, parts })\n\n // Update the metadata with completed state\n const completedUpload = new Upload({\n ...metadata.file,\n offset: newOffset,\n size: metadata.file.size,\n storage: metadata.file.storage,\n })\n\n await this.metadataManager.completeMetadata({ upload: completedUpload })\n } catch (error) {\n log(MediaCloudLogs.S3_STORE_UPLOAD_FAILED)\n throw error\n }\n }\n\n return newOffset\n }\n\n /**\n * Returns the current state of the upload, i.e how much data has been\n * uploaded and if the upload is complete.\n * @param id - The upload ID to retrieve\n * @returns Promise that resolves to the upload object with current offset and storage information\n */\n public async getUpload(id: string): Promise<Upload> {\n let metadata: TusUploadMetadata\n\n try {\n metadata = await this.metadataManager.getMetadata({ id })\n } catch (error) {\n if (\n error instanceof NoSuchKey ||\n error instanceof NotFound ||\n (error as AWSError)?.Code === 'NotFound' ||\n (error as AWSError)?.Code === 'NoSuchKey'\n ) {\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error\n }\n\n let offset: number\n\n try {\n const parts = await this.partsManager.retrieveParts({ id })\n offset = this.fileOperations.calculateOffsetFromParts({ parts })\n } catch (error) {\n // Check if the error is caused by the upload not being found. This happens\n // when the multipart upload has already been completed or aborted.\n if (\n (error as AWSError)?.Code === 'NoSuchUpload' ||\n (error as AWSError)?.Code === 'NoSuchKey'\n ) {\n return new Upload({\n ...metadata.file,\n metadata: metadata.file.metadata,\n offset: metadata.file.size as number,\n size: metadata.file.size,\n storage: metadata.file.storage,\n })\n }\n\n log(MediaCloudLogs.S3_STORE_RETRIEVE_PARTS_ERROR)\n throw error\n }\n\n const incompletePartSize = await this.partsManager.getIncompletePartSize({\n id,\n })\n\n return new Upload({\n ...metadata.file,\n offset: offset + (incompletePartSize ?? 0),\n size: metadata.file.size,\n storage: metadata.file.storage,\n })\n }\n\n /**\n * Reads the file specified by the upload’s `id` and returns a readable stream\n * @param id - The upload ID to read\n * @returns Promise that resolves to a readable stream of the file's contents\n */\n async read(id: string): Promise<Readable> {\n log(MediaCloudLogs.S3_STORE_READ_ATTEMPT)\n let retries = 3\n let lastError: Error | null = null\n\n while (retries > 0) {\n try {\n const data = await this.client.getObject({\n Bucket: this.bucket,\n Key: id,\n })\n log(MediaCloudLogs.S3_STORE_READ_SUCCESS)\n return data.Body as Readable\n } catch (error) {\n log(MediaCloudLogs.S3_STORE_READ_RETRY)\n lastError = error as Error\n retries--\n\n if (retries > 0) {\n // Wait a bit before retrying, in case S3 needs time for consistency\n await new Promise((resolve) => setTimeout(resolve, 100))\n }\n }\n }\n\n log(MediaCloudLogs.S3_STORE_READ_FAILED)\n throw lastError ?? new Error(`Failed to read file ${id}`)\n }\n\n /**\n *\n * Moves the file specified by its `oldKey` to `newKey`.\n * @param oldKey - The current S3 key of the file to be moved\n * @param newKey - The new S3 key to move the file to\n * @return Promise that resolves when the file has been successfully moved, or rejects with an error if the move operation fails\n */\n public async copy(oldKey: string, newKey: string): Promise<void> {\n try {\n await this.client.copyObject({\n Bucket: this.bucket,\n CopySource: encodeURI(`${this.bucket}/${oldKey}`),\n Key: newKey,\n })\n\n await this.client.deleteObject({\n Bucket: this.bucket,\n Key: oldKey,\n })\n } catch (error) {\n if (\n (error as AWSError)?.code &&\n ['NoSuchKey', 'NoSuchUpload', 'NotFound'].includes(\n (error as AWSError).Code || ''\n )\n ) {\n log(MediaCloudLogs.S3_STORE_FILE_NOT_FOUND)\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error as Error\n }\n }\n\n /**\n * Removes files specified by the upload’s `id`\n * @param id - The upload ID to remove\n * @returns Promise that resolves when the file and its metadata have been removed\n */\n public async remove(id: string): Promise<void> {\n try {\n const { 'upload-id': uploadId, file } =\n await this.metadataManager.getMetadata({ id })\n if (uploadId) {\n await this.client.abortMultipartUpload({\n Bucket: this.bucket,\n Key: id,\n UploadId: uploadId,\n })\n }\n\n await this.client.deleteObjects({\n Bucket: this.bucket,\n Delete: {\n Objects: [\n {\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n },\n {\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n isIncomplete: true,\n }),\n },\n {\n Key: this.metadataManager.generateInfoKey({ id }),\n },\n ],\n },\n })\n } catch (error) {\n if (\n (error as AWSError)?.code &&\n ['NoSuchKey', 'NoSuchUpload', 'NotFound'].includes(\n (error as AWSError).Code || ''\n )\n ) {\n log(MediaCloudLogs.S3_STORE_FILE_NOT_FOUND)\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error\n }\n }\n\n /**\n * Removes the .info file specified by the upload’s `id`\n * @param id - The upload ID to clean up\n * @returns Promise that resolves when the metadata file has been removed\n */\n public async cleanup(id: string): Promise<void> {\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n await this.client.deleteObjects({\n Bucket: this.bucket,\n Delete: {\n Objects: [\n {\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n isIncomplete: true,\n }),\n },\n {\n Key: this.metadataManager.generateInfoKey({ id }),\n },\n ],\n },\n })\n } catch (error) {\n if (\n (error as AWSError)?.code &&\n ['NoSuchKey', 'NoSuchUpload', 'NotFound'].includes(\n (error as AWSError).Code || ''\n )\n ) {\n log(MediaCloudLogs.S3_STORE_FILE_NOT_FOUND)\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error\n }\n }\n\n /**\n * Combine all multipart uploads into a single object\n * @param id - The upload ID to complete\n * @returns Promise that resolves to the completed upload object with storage information\n */\n public async completeMultipartUpload(id: string): Promise<Upload> {\n const metadata = await this.metadataManager.getMetadata({ id })\n const parts = await this.partsManager.retrieveParts({ id })\n\n const incompletePartInfo = await this.partsManager.downloadIncompletePart({\n id,\n })\n\n if (incompletePartInfo) {\n // Upload the incomplete part as a regular part\n await this.partsManager.uploadPart({\n metadata,\n readStream: incompletePartInfo.createReader({ cleanUpOnEnd: true }),\n partNumber: parts.length + 1,\n })\n\n // Remove the incomplete part\n await this.partsManager.deleteIncompletePart({ id })\n\n // Re-fetch parts to include the newly uploaded part\n const updatedParts = await this.partsManager.retrieveParts({ id })\n await this.partsManager.finishMultipartUpload({\n metadata,\n parts: updatedParts,\n })\n } else {\n await this.partsManager.finishMultipartUpload({ metadata, parts })\n }\n\n const completedUpload = new Upload({\n ...metadata.file,\n offset: metadata.file.size ?? 0,\n size: metadata.file.size ?? 0,\n storage: metadata.file.storage,\n })\n\n await this.metadataManager.completeMetadata({ upload: completedUpload })\n\n return completedUpload\n }\n\n /**\n * Get the full S3 URL for an uploaded file\n * @param id - The upload ID to get the URL for\n * @returns The full URL to access the uploaded file on S3\n */\n public getUrl(id: string): string {\n // Use the custom endpoint if available\n if (this.customEndpoint) {\n return `${this.customEndpoint}/${this.bucket}/${id}`\n }\n\n // Fallback to standard AWS S3 URL format\n const regionConfig = this.client.config.region\n let region: string\n\n // If region is a function, we can't resolve it synchronously, so use a fallback\n if (typeof regionConfig === 'function') {\n region = 'us-east-1' // fallback for sync calls\n } else {\n region = regionConfig || 'us-east-1'\n }\n\n // Standard AWS S3 URL format\n if (region === 'us-east-1') {\n return `https://${this.bucket}.s3.amazonaws.com/${id}`\n } else {\n return `https://${this.bucket}.s3.${region}.amazonaws.com/${id}`\n }\n }\n\n /**\n * Deletes expired incomplete uploads.\n * Returns the number of deleted uploads.\n * @returns Promise that resolves to the number of deleted uploads\n */\n async deleteExpired(): Promise<number> {\n return this.expirationManager.deleteExpired()\n }\n\n /**\n * Returns the expiration period in milliseconds\n * @return The expiration period in milliseconds\n */\n getExpiration(): number {\n return this.expirationManager.getExpiration()\n }\n}\n"],"mappings":";;;;;;;;;;;;AAgBA,MAAM,EAAEoB,QAAQX,iBAAiB;AAEjC,IAAaY,UAAb,cAA6BlB,UAAU;CAkBrCmC,YAAYC,SAAwB;AAClC,SAAO;kBAhBS,IAAI,OAAO;qBACR,IAAI,OAAO;2BACL;uBACJ;iBACN;wCACuB;EAYtC,MAAM,EACJb,mBACAD,aACAD,UACAgB,gBACAC,0BACAZ,SACAC,mCACES;EACJ,MAAM,EAAER,KAAKR,QAAQ,GAAGmB,uBAAuBF;AAE/C,OAAKG,aAAa;GAChB;GACA;GACA;GACA;GACA;GACD;AAED,OAAKpB,SAASA;AACd,OAAKQ,MAAMA;AACX,OAAKT,SAAS,IAAIpB,GAAGwC,mBAAmB;AACxC,OAAKL,iBAAiBO,OAAOF,mBAAmBG,SAAS;AAEzD,OAAKrB,WAAWA,YAAY,KAAKA;AACjC,OAAKC,cAAcA,eAAe,KAAKA;AACvC,OAAKC,oBAAoBA,qBAAqB,KAAKA;AAEnD,OAAKG,UAAUA,WAAW,KAAKA;AAC/B,OAAKC,iCACHA,kCAAkC,KAAKA;AACzC,OAAKE,sBAAsB,IAAItB,UAAU+B,4BAA4B,GAAG;AAGxE,OAAKR,kBAAkB,IAAIpB,kBACzB,KAAKS,QACL,KAAKC,QACL,KAAKuB,wBAAwBC,KAAK,KAAK,EACvC,KAAKC,oBAAoBD,KAAK,KAChC,CAAC;AAED,OAAKb,iBAAiB,IAAItB,iBACxB,KAAKc,mBACL,KAAKC,eACL,KAAKF,aACL,KAAKD,SACN;AAED,OAAKW,eAAe,IAAIrB,eACtB,KAAKQ,QACL,KAAKC,QACL,KAAKE,aACL,KAAKO,qBACL,KAAKC,iBACL,KAAKC,gBACL,KAAKc,oBAAoBD,KAAK,KAChC,CAAC;AAED,OAAKX,oBAAoB,IAAIzB,oBAC3B,KAAKW,QACL,KAAKC,QACL,KAAKO,gCACL,KAAKG,gBAAgBgB,gBAAgBF,KAAK,KAAKd,gBAAgB,EAC/D,KAAKA,gBAAgBiB,gBAAgBH,KAAK,KAAKd,gBACjD,CAAC;AAGD,OAAKkB,eAAe;;;;;;CAOtB,AAAUL,0BAAmC;AAC3C,SAAO,KAAKhB,mCAAmC,KAAK,KAAKD;;;;;;;CAQ3D,AAAUmB,oBAAoBI,OAA6C;AACzE,MAAI,CAAC,KAAKN,yBAAyB,CACjC;AAEF,SAAO,iBAAiBM;;;;;;;;;CAU1B,MAAaE,OAAOC,QAAiC;AACnDnC,MAAIL,eAAe0C,wBAAwB;EAE3C,MAAMC,UAAiD;GACrDE,QAAQ,KAAKrC;GACbsC,KAAK,KAAK5B,gBAAgBiB,gBAAgB;IACxCY,IAAIP,OAAOO;IACXC,QAAQR,OAAOS,UAAUD,UAAUV;IACpC,CAAC;GACFY,UAAU,EAAE,eAAe5D,eAAc;GAC1C;AAED,MAAIkD,OAAOS,UAAUE,YACnBR,SAAQS,cAAcZ,OAAOS,SAASE;AAGxC,MAAIX,OAAOS,UAAUI,aACnBV,SAAQW,eAAed,OAAOS,SAASI;AAGzC,MAAI,KAAKrC,IACP2B,SAAQY,MAAM,KAAKvC;AAGrBwB,SAAOiB,iCAAgB,IAAIC,MAAM,EAACC,aAAa;EAE/C,MAAMC,WAAW,MAAM,KAAKrD,OAAOsD,sBAAsBlB,QAAQ;AAEjEH,SAAOsB,UAAU;GACfC,MAAM;GACNvD,QAAQ,KAAKA;GACbwD,MAAMJ,SAASd;GAChB;AACD,QAAM,KAAK5B,gBAAgB+C,aAAa;GACtCzB;GACA0B,UAAUN,SAASO;GACpB,CAAC;AACF9D,MAAIL,eAAeoE,2BAA2B;AAE9C,SAAO5B;;;;;;;;;CAUT,MAAa6B,MACXC,UACAvB,IACAwB,QACiB;EACjB,MAAMtB,WAAW,MAAM,KAAK/B,gBAAgBsD,YAAY,EAAEzB,IAAI,CAAC;EAI/D,MAAM0B,mBAAmB,KAAKtD,eAAeuD,yBAAyB,EACpEC,OAAO,MAAM,KAAKvD,aAAawD,cAAc,EAAE7B,IAAI,CAAA,EACpD,CAAC;EACF,MAAM8B,aAAaN,SAASE;EAC5B,MAAMK,kBAAkBP;EAExB,IAAIQ,gBAAgBT;AAEpB,MAAIO,aAAa,EACf,OAAMxF,OAAO2F;AAMf,MAAIH,aAAa,GAAG;GAClB,MAAMI,iBAAiB,MAAM,KAAK7D,aAAa8D,uBAAuB,EACpEnC,IACD,CAAC;AAEF,OAAI,CAACkC,eACH,OAAM5F,OAAO2F;AAGf,OAAIC,eAAeE,SAASN,WAC1B,OAAMxF,OAAO2F;AAIf,SAAM,KAAK5D,aAAagE,qBAAqB,EAAErC,IAAI,CAAC;AAGpDwB,YAASO,kBAAkBG,eAAeE;AAE1CJ,mBAAgBvF,OAAOC,SAAS4F,KAC7B,mBAAmB;AAClB,WAAOJ,eAAeK,aAAa,EAAEC,cAAc,MAAM,CAAC;AAC1D,WAAOjB;MAEX,CAAC;;EAGH,MAAMkB,aAAa,KAAKrE,eAAesE,oBAAoB,EACzDd,OAAO,MAAM,KAAKvD,aAAawD,cAAc,EAAE7B,IAAI,CAAA,EACpD,CAAC;EAWF,MAAM+C,YACJhB,kBAVoB,MAAM,KAAK1D,aAAauE,YAAY;GACxD1C;GACA2C,YAAYb;GACZc,mBAAmBL;GACnBjB;GACD,CAAC,IAKmCM,aAAa,IAAIA,aAAa;AAGnE,MAAI5B,SAAS8C,KAAKZ,SAASW,UACzB,KAAI;GACF,MAAMnB,QAAQ,MAAM,KAAKvD,aAAawD,cAAc,EAAE7B,IAAI,CAAC;AAC3D,SAAM,KAAK3B,aAAa4E,sBAAsB;IAAE/C;IAAU0B;IAAO,CAAC;GAGlE,MAAMsB,kBAAkB,IAAI1G,OAAO;IACjC,GAAG0D,SAAS8C;IACZxB,QAAQuB;IACRX,MAAMlC,SAAS8C,KAAKZ;IACpBrB,SAASb,SAAS8C,KAAKjC;IACxB,CAAC;AAEF,SAAM,KAAK5C,gBAAgBgF,iBAAiB,EAAE1D,QAAQyD,iBAAiB,CAAC;WACjEE,OAAO;AACd9F,OAAIL,eAAeoG,uBAAuB;AAC1C,SAAMD;;AAIV,SAAOL;;;;;;;;CAST,MAAaO,UAAUtD,IAA6B;EAClD,IAAIE;AAEJ,MAAI;AACFA,cAAW,MAAM,KAAK/B,gBAAgBsD,YAAY,EAAEzB,IAAI,CAAC;WAClDoD,OAAO;AACd,OACEA,iBAAiBlH,aACjBkH,iBAAiBjH,YAChBiH,OAAoBG,SAAS,cAC7BH,OAAoBG,SAAS,YAE9B,OAAMjH,OAAOkH;AAEf,SAAMJ;;EAGR,IAAI5B;AAEJ,MAAI;GACF,MAAMI,QAAQ,MAAM,KAAKvD,aAAawD,cAAc,EAAE7B,IAAI,CAAC;AAC3DwB,YAAS,KAAKpD,eAAeuD,yBAAyB,EAAEC,OAAO,CAAC;WACzDwB,OAAO;AAGd,OACGA,OAAoBG,SAAS,kBAC7BH,OAAoBG,SAAS,YAE9B,QAAO,IAAI/G,OAAO;IAChB,GAAG0D,SAAS8C;IACZ9C,UAAUA,SAAS8C,KAAK9C;IACxBsB,QAAQtB,SAAS8C,KAAKZ;IACtBA,MAAMlC,SAAS8C,KAAKZ;IACpBrB,SAASb,SAAS8C,KAAKjC;IACxB,CAAC;AAGJzD,OAAIL,eAAewG,8BAA8B;AACjD,SAAML;;EAGR,MAAMM,qBAAqB,MAAM,KAAKrF,aAAasF,sBAAsB,EACvE3D,IACD,CAAC;AAEF,SAAO,IAAIxD,OAAO;GAChB,GAAG0D,SAAS8C;GACZxB,QAAQA,UAAUkC,sBAAsB;GACxCtB,MAAMlC,SAAS8C,KAAKZ;GACpBrB,SAASb,SAAS8C,KAAKjC;GACxB,CAAC;;;;;;;CAQJ,MAAM6C,KAAK5D,IAA+B;AACxC1C,MAAIL,eAAe4G,sBAAsB;EACzC,IAAIC,UAAU;EACd,IAAIC,YAA0B;AAE9B,SAAOD,UAAU,EACf,KAAI;GACF,MAAMG,OAAO,MAAM,KAAKzG,OAAO0G,UAAU;IACvCpE,QAAQ,KAAKrC;IACbsC,KAAKC;IACN,CAAC;AACF1C,OAAIL,eAAekH,sBAAsB;AACzC,UAAOF,KAAKG;WACLhB,OAAO;AACd9F,OAAIL,eAAeoH,oBAAoB;AACvCN,eAAYX;AACZU;AAEA,OAAIA,UAAU,EAEZ,OAAM,IAAIpE,SAAS4E,YAAYC,WAAWD,SAAS,IAAI,CAAC;;AAK9DhH,MAAIL,eAAeuH,qBAAqB;AACxC,QAAMT,6BAAa,IAAIC,MAAM,uBAAuBhE,KAAK;;;;;;;;;CAU3D,MAAayE,KAAKC,QAAgBC,QAA+B;AAC/D,MAAI;AACF,SAAM,KAAKnH,OAAOoH,WAAW;IAC3B9E,QAAQ,KAAKrC;IACboH,YAAYC,UAAU,GAAG,KAAKrH,OAAM,GAAIiH,SAAS;IACjD3E,KAAK4E;IACN,CAAC;AAEF,SAAM,KAAKnH,OAAOuH,aAAa;IAC7BjF,QAAQ,KAAKrC;IACbsC,KAAK2E;IACN,CAAC;WACKtB,OAAO;AACd,OACGA,OAAoB4B,QACrB;IAAC;IAAa;IAAgB;IAAW,CAACC,SACvC7B,MAAmBG,QAAQ,GAC7B,EACD;AACAjG,QAAIL,eAAeiI,wBAAwB;AAC3C,UAAM5I,OAAOkH;;AAEf,SAAMJ;;;;;;;;CASV,MAAa+B,OAAOnF,IAA2B;AAC7C,MAAI;GACF,MAAM,EAAE,aAAamB,UAAU6B,SAC7B,MAAM,KAAK7E,gBAAgBsD,YAAY,EAAEzB,IAAI,CAAC;AAChD,OAAImB,SACF,OAAM,KAAK3D,OAAO4H,qBAAqB;IACrCtF,QAAQ,KAAKrC;IACbsC,KAAKC;IACLoB,UAAUD;IACX,CAAC;AAGJ,SAAM,KAAK3D,OAAO6H,cAAc;IAC9BvF,QAAQ,KAAKrC;IACb6H,QAAQ,EACNC,SAAS;KACP,EACExF,KAAK,KAAK5B,gBAAgBiB,gBAAgB;MACxCY;MACAC,QAAQ+C,MAAM9C,UAAUD,UAAUV;MACnC,CAAA,EACF;KACD,EACEQ,KAAK,KAAK5B,gBAAgBiB,gBAAgB;MACxCY;MACAC,QAAQ+C,MAAM9C,UAAUD,UAAUV;MAClCiG,cAAc;MACf,CAAA,EACF;KACD,EACEzF,KAAK,KAAK5B,gBAAgBgB,gBAAgB,EAAEa,IAAI,CAAA,EACjD;KAAA,EAEL;IACD,CAAC;WACKoD,OAAO;AACd,OACGA,OAAoB4B,QACrB;IAAC;IAAa;IAAgB;IAAW,CAACC,SACvC7B,MAAmBG,QAAQ,GAC7B,EACD;AACAjG,QAAIL,eAAeiI,wBAAwB;AAC3C,UAAM5I,OAAOkH;;AAEf,SAAMJ;;;;;;;;CASV,MAAaqC,QAAQzF,IAA2B;AAC9C,MAAI;GACF,MAAM,EAAEgD,SAAS,MAAM,KAAK7E,gBAAgBsD,YAAY,EAAEzB,IAAI,CAAC;AAE/D,SAAM,KAAKxC,OAAO6H,cAAc;IAC9BvF,QAAQ,KAAKrC;IACb6H,QAAQ,EACNC,SAAS,CACP,EACExF,KAAK,KAAK5B,gBAAgBiB,gBAAgB;KACxCY;KACAC,QAAQ+C,MAAM9C,UAAUD,UAAUV;KAClCiG,cAAc;KACf,CAAA,EACF,EACD,EACEzF,KAAK,KAAK5B,gBAAgBgB,gBAAgB,EAAEa,IAAI,CAAA,EACjD,CAAA,EAEL;IACD,CAAC;WACKoD,OAAO;AACd,OACGA,OAAoB4B,QACrB;IAAC;IAAa;IAAgB;IAAW,CAACC,SACvC7B,MAAmBG,QAAQ,GAC7B,EACD;AACAjG,QAAIL,eAAeiI,wBAAwB;AAC3C,UAAM5I,OAAOkH;;AAEf,SAAMJ;;;;;;;;CASV,MAAasC,wBAAwB1F,IAA6B;EAChE,MAAME,WAAW,MAAM,KAAK/B,gBAAgBsD,YAAY,EAAEzB,IAAI,CAAC;EAC/D,MAAM4B,QAAQ,MAAM,KAAKvD,aAAawD,cAAc,EAAE7B,IAAI,CAAC;EAE3D,MAAM2F,qBAAqB,MAAM,KAAKtH,aAAa8D,uBAAuB,EACxEnC,IACD,CAAC;AAEF,MAAI2F,oBAAoB;AAEtB,SAAM,KAAKtH,aAAauH,WAAW;IACjC1F;IACA2C,YAAY8C,mBAAmBpD,aAAa,EAAEC,cAAc,MAAM,CAAC;IACnEC,YAAYb,MAAMiE,SAAS;IAC5B,CAAC;AAGF,SAAM,KAAKxH,aAAagE,qBAAqB,EAAErC,IAAI,CAAC;GAGpD,MAAM8F,eAAe,MAAM,KAAKzH,aAAawD,cAAc,EAAE7B,IAAI,CAAC;AAClE,SAAM,KAAK3B,aAAa4E,sBAAsB;IAC5C/C;IACA0B,OAAOkE;IACR,CAAC;QAEF,OAAM,KAAKzH,aAAa4E,sBAAsB;GAAE/C;GAAU0B;GAAO,CAAC;EAGpE,MAAMsB,kBAAkB,IAAI1G,OAAO;GACjC,GAAG0D,SAAS8C;GACZxB,QAAQtB,SAAS8C,KAAKZ,QAAQ;GAC9BA,MAAMlC,SAAS8C,KAAKZ,QAAQ;GAC5BrB,SAASb,SAAS8C,KAAKjC;GACxB,CAAC;AAEF,QAAM,KAAK5C,gBAAgBgF,iBAAiB,EAAE1D,QAAQyD,iBAAiB,CAAC;AAExE,SAAOA;;;;;;;CAQT,AAAO6C,OAAO/F,IAAoB;AAEhC,MAAI,KAAKzB,eACP,QAAO,GAAG,KAAKA,eAAc,GAAI,KAAKd,OAAM,GAAIuC;EAIlD,MAAMgG,eAAe,KAAKxI,OAAOyI,OAAOC;EACxC,IAAIA;AAGJ,MAAI,OAAOF,iBAAiB,WAC1BE,UAAS;MAETA,UAASF,gBAAgB;AAI3B,MAAIE,WAAW,YACb,QAAO,WAAW,KAAKzI,OAAM,oBAAqBuC;MAElD,QAAO,WAAW,KAAKvC,OAAM,MAAOyI,OAAM,iBAAkBlG;;;;;;;CAShE,MAAMX,gBAAiC;AACrC,SAAO,KAAKf,kBAAkBe,eAAe;;;;;;CAO/C8G,gBAAwB;AACtB,SAAO,KAAK7H,kBAAkB6H,eAAe"}
|
|
1
|
+
{"version":3,"file":"s3Store.mjs","names":["request: AWS.CreateMultipartUploadCommandInput","metadata: TusUploadMetadata","offset: number","lastError: Error | null","region: string"],"sources":["../../../../src/tus/stores/s3/s3Store.ts"],"sourcesContent":["import { NoSuchKey, NotFound, S3 } from '@aws-sdk/client-s3'\nimport { DataStore, ERRORS, TUS_RESUMABLE, Upload } from '@tus/utils'\nimport stream, { type Readable } from 'node:stream'\n\nimport { useErrorHandler } from '../../../hooks/useErrorHandler'\nimport { Semaphore } from './semaphore'\nimport { S3ExpirationManager } from './expirationManager'\nimport { S3FileOperations } from './fileOperations'\nimport { S3MetadataManager } from './metadataManager'\nimport { S3PartsManager } from './partsManager'\n\nimport { MediaCloudLogs } from '../../../types/errors'\n\nimport type AWS from '@aws-sdk/client-s3'\nimport type { AWSError, S3StoreConfig, TusUploadMetadata } from '../../../types'\n\nconst { log } = useErrorHandler()\n\nexport class S3Store extends DataStore {\n public client: S3\n public bucket: string\n public partSize = 8 * 1024 * 1024 // 8MB preferred part size\n public minPartSize = 5 * 1024 * 1024 // 5MB minimum part size\n public maxMultipartParts = 10_000\n public maxUploadSize = 5_497_558_138_880 as const // 5TiB\n public useTags = false\n public expirationPeriodInMilliseconds = 0\n protected acl?: string\n\n protected partUploadSemaphore: Semaphore\n protected metadataManager: S3MetadataManager\n protected fileOperations: S3FileOperations\n protected partsManager: S3PartsManager\n protected expirationManager: S3ExpirationManager\n protected customEndpoint: string\n\n constructor(options: S3StoreConfig) {\n super()\n const {\n maxMultipartParts,\n minPartSize,\n partSize,\n s3ClientConfig,\n maxConcurrentPartUploads,\n useTags,\n expirationPeriodInMilliseconds,\n } = options\n const { acl, bucket, ...restS3ClientConfig } = s3ClientConfig\n\n this.extensions = [\n 'creation',\n 'creation-with-upload',\n 'creation-defer-length',\n 'termination',\n 'expiration',\n ]\n\n this.bucket = bucket\n this.acl = acl\n this.client = new S3(restS3ClientConfig)\n this.customEndpoint = String(restS3ClientConfig.endpoint)\n\n this.partSize = partSize ?? this.partSize\n this.minPartSize = minPartSize ?? this.minPartSize\n this.maxMultipartParts = maxMultipartParts ?? this.maxMultipartParts\n\n this.useTags = useTags ?? this.useTags\n this.expirationPeriodInMilliseconds =\n expirationPeriodInMilliseconds ?? this.expirationPeriodInMilliseconds\n this.partUploadSemaphore = new Semaphore(maxConcurrentPartUploads ?? 60)\n\n // Initialize component managers\n this.metadataManager = new S3MetadataManager(\n this.client,\n this.bucket,\n this.shouldUseExpirationTags.bind(this),\n this.generateCompleteTag.bind(this)\n )\n\n this.fileOperations = new S3FileOperations(\n this.maxMultipartParts,\n this.maxUploadSize,\n this.minPartSize,\n this.partSize\n )\n\n this.partsManager = new S3PartsManager(\n this.client,\n this.bucket,\n this.minPartSize,\n this.partUploadSemaphore,\n this.metadataManager,\n this.fileOperations,\n this.generateCompleteTag.bind(this)\n )\n\n this.expirationManager = new S3ExpirationManager(\n this.client,\n this.bucket,\n this.expirationPeriodInMilliseconds,\n this.metadataManager.generateInfoKey.bind(this.metadataManager),\n this.metadataManager.generatePartKey.bind(this.metadataManager)\n )\n\n // Cleanup expired uploads when the store is initialized\n this.deleteExpired()\n }\n\n /**\n * Helper method to check if expiration tags should be used\n * @returns True if expiration tags should be used\n */\n protected shouldUseExpirationTags(): boolean {\n return this.expirationPeriodInMilliseconds !== 0 && this.useTags\n }\n\n /**\n * Generates a tag for marking complete/incomplete uploads\n * @param value - Either 'false' or 'true' to mark completion status\n * @returns The tag string or undefined if tags shouldn’t be used\n */\n protected generateCompleteTag(value: 'false' | 'true'): string | undefined {\n if (!this.shouldUseExpirationTags()) {\n return undefined\n }\n return `Tus-Completed=${value}`\n }\n\n /**\n * Creates a multipart upload on S3 attaching any metadata to it.\n * Also, a `${file_id}.info` file is created which holds some information\n * about the upload itself like: `upload-id`, `upload-length`, etc.\n * @param upload - The upload object to create\n * @return Promise that resolves to the created upload object with storage information\n */\n public async create(upload: Upload): Promise<Upload> {\n log(MediaCloudLogs.S3_STORE_MULTIPART_INIT)\n\n const request: AWS.CreateMultipartUploadCommandInput = {\n Bucket: this.bucket,\n Key: this.metadataManager.generatePartKey({\n id: upload.id,\n prefix: upload.metadata?.prefix ?? undefined,\n }),\n Metadata: { 'tus-version': TUS_RESUMABLE },\n }\n\n if (upload.metadata?.contentType) {\n request.ContentType = upload.metadata.contentType as string\n }\n\n if (upload.metadata?.cacheControl) {\n request.CacheControl = upload.metadata.cacheControl as string\n }\n\n if (this.acl) {\n request.ACL = this.acl as AWS.ObjectCannedACL\n }\n\n upload.creation_date = new Date().toISOString()\n\n const response = await this.client.createMultipartUpload(request)\n\n upload.storage = {\n type: 's3',\n bucket: this.bucket,\n path: response.Key as string,\n }\n await this.metadataManager.saveMetadata({\n upload,\n uploadId: response.UploadId as string,\n })\n log(MediaCloudLogs.S3_STORE_MULTIPART_CREATED)\n\n return upload\n }\n\n /**\n * Writes `buffer` to the file specified by the upload’s `id` at `offset`\n * @param readable - The readable stream to write\n * @param id - The upload ID\n * @param offset - The byte offset to write at\n * @return Promise that resolves to the new offset after writing the data\n */\n public async write(\n readable: stream.Readable,\n id: string,\n offset: number\n ): Promise<number> {\n const metadata = await this.metadataManager.getMetadata({ id })\n\n // TUS sends PATCH requests with an `upload-offset` header.\n // Offset the write by the offset in the PATCH request.\n const calculatedOffset = this.fileOperations.calculateOffsetFromParts({\n parts: await this.partsManager.retrieveParts({ id }),\n })\n const offsetDiff = offset - calculatedOffset\n const requestedOffset = offset\n\n let finalReadable = readable\n\n if (offsetDiff < 0) {\n throw ERRORS.FILE_WRITE_ERROR\n }\n\n // If the offset given in the PATCH request is higher than\n // the expected offset, we need to prepend an incomplete\n // part to the readable stream, if one exists.\n if (offsetDiff > 0) {\n const incompletePart = await this.partsManager.downloadIncompletePart({\n id,\n })\n\n if (!incompletePart) {\n throw ERRORS.FILE_WRITE_ERROR\n }\n\n if (incompletePart.size !== offsetDiff) {\n throw ERRORS.FILE_WRITE_ERROR\n }\n\n // Clear the incomplete part from S3 since it's going to be combined with the current request\n await this.partsManager.deleteIncompletePart({ id })\n\n // Adjust offset to account for the incomplete part\n offset = requestedOffset - incompletePart.size\n\n finalReadable = stream.Readable.from(\n (async function* () {\n yield* incompletePart.createReader({ cleanUpOnEnd: true })\n yield* readable\n })()\n )\n }\n\n const partNumber = this.fileOperations.calculatePartNumber({\n parts: await this.partsManager.retrieveParts({ id }),\n })\n\n const bytesUploaded = await this.partsManager.uploadParts({\n metadata,\n readStream: finalReadable,\n currentPartNumber: partNumber,\n offset,\n })\n\n // The size of the incomplete part should not be counted, because the\n // process of the incomplete part should be fully transparent to the user.\n const newOffset =\n requestedOffset + bytesUploaded - (offsetDiff > 0 ? offsetDiff : 0)\n\n // Check if the upload is complete\n if (metadata.file.size === newOffset) {\n try {\n const parts = await this.partsManager.retrieveParts({ id })\n await this.partsManager.finishMultipartUpload({ metadata, parts })\n\n // Update the metadata with completed state\n const completedUpload = new Upload({\n ...metadata.file,\n offset: newOffset,\n size: metadata.file.size,\n storage: metadata.file.storage,\n })\n\n await this.metadataManager.completeMetadata({ upload: completedUpload })\n } catch (error) {\n log(MediaCloudLogs.S3_STORE_UPLOAD_FAILED)\n throw error\n }\n }\n\n return newOffset\n }\n\n /**\n * Returns the current state of the upload, i.e how much data has been\n * uploaded and if the upload is complete.\n * @param id - The upload ID to retrieve\n * @returns Promise that resolves to the upload object with current offset and storage information\n */\n public async getUpload(id: string): Promise<Upload> {\n let metadata: TusUploadMetadata\n\n try {\n metadata = await this.metadataManager.getMetadata({ id })\n } catch (error) {\n if (\n error instanceof NoSuchKey ||\n error instanceof NotFound ||\n (error as AWSError)?.Code === 'NotFound' ||\n (error as AWSError)?.Code === 'NoSuchKey'\n ) {\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error\n }\n\n let offset: number\n\n try {\n const parts = await this.partsManager.retrieveParts({ id })\n offset = this.fileOperations.calculateOffsetFromParts({ parts })\n } catch (error) {\n // Check if the error is caused by the upload not being found. This happens\n // when the multipart upload has already been completed or aborted.\n if (\n (error as AWSError)?.Code === 'NoSuchUpload' ||\n (error as AWSError)?.Code === 'NoSuchKey'\n ) {\n return new Upload({\n ...metadata.file,\n metadata: metadata.file.metadata,\n offset: metadata.file.size as number,\n size: metadata.file.size,\n storage: metadata.file.storage,\n })\n }\n\n log(MediaCloudLogs.S3_STORE_RETRIEVE_PARTS_ERROR)\n throw error\n }\n\n const incompletePartSize = await this.partsManager.getIncompletePartSize({\n id,\n })\n\n return new Upload({\n ...metadata.file,\n offset: offset + (incompletePartSize ?? 0),\n size: metadata.file.size,\n storage: metadata.file.storage,\n })\n }\n\n /**\n * Reads the file specified by the upload’s `id` and returns a readable stream\n * @param id - The upload ID to read\n * @returns Promise that resolves to a readable stream of the file's contents\n */\n async read(id: string): Promise<Readable> {\n log(MediaCloudLogs.S3_STORE_READ_ATTEMPT)\n let retries = 3\n let lastError: Error | null = null\n\n while (retries > 0) {\n try {\n const data = await this.client.getObject({\n Bucket: this.bucket,\n Key: id,\n })\n log(MediaCloudLogs.S3_STORE_READ_SUCCESS)\n return data.Body as Readable\n } catch (error) {\n log(MediaCloudLogs.S3_STORE_READ_RETRY)\n lastError = error as Error\n retries--\n\n if (retries > 0) {\n // Wait a bit before retrying, in case S3 needs time for consistency\n await new Promise((resolve) => setTimeout(resolve, 100))\n }\n }\n }\n\n log(MediaCloudLogs.S3_STORE_READ_FAILED)\n throw lastError ?? new Error(`Failed to read file ${id}`)\n }\n\n /**\n *\n * Moves the file specified by its `oldKey` to `newKey`.\n * @param oldKey - The current S3 key of the file to be moved\n * @param newKey - The new S3 key to move the file to\n * @return Promise that resolves when the file has been successfully moved, or rejects with an error if the move operation fails\n */\n public async copy(oldKey: string, newKey: string): Promise<void> {\n try {\n await this.client.copyObject({\n Bucket: this.bucket,\n CopySource: encodeURI(`${this.bucket}/${oldKey}`),\n Key: newKey,\n })\n\n await this.client.deleteObject({\n Bucket: this.bucket,\n Key: oldKey,\n })\n } catch (error) {\n if (\n (error as AWSError)?.code &&\n ['NoSuchKey', 'NoSuchUpload', 'NotFound'].includes(\n (error as AWSError).Code || ''\n )\n ) {\n log(MediaCloudLogs.S3_STORE_FILE_NOT_FOUND)\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error as Error\n }\n }\n\n /**\n * Removes files specified by the upload’s `id`\n * @param id - The upload ID to remove\n * @returns Promise that resolves when the file and its metadata have been removed\n */\n public async remove(id: string): Promise<void> {\n try {\n const { 'upload-id': uploadId, file } =\n await this.metadataManager.getMetadata({ id })\n if (uploadId) {\n await this.client.abortMultipartUpload({\n Bucket: this.bucket,\n Key: id,\n UploadId: uploadId,\n })\n }\n\n await this.client.deleteObjects({\n Bucket: this.bucket,\n Delete: {\n Objects: [\n {\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n }),\n },\n {\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n isIncomplete: true,\n }),\n },\n {\n Key: this.metadataManager.generateInfoKey({ id }),\n },\n ],\n },\n })\n } catch (error) {\n if (\n (error as AWSError)?.code &&\n ['NoSuchKey', 'NoSuchUpload', 'NotFound'].includes(\n (error as AWSError).Code || ''\n )\n ) {\n log(MediaCloudLogs.S3_STORE_FILE_NOT_FOUND)\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error\n }\n }\n\n /**\n * Removes the .info file specified by the upload’s `id`\n * @param id - The upload ID to clean up\n * @returns Promise that resolves when the metadata file has been removed\n */\n public async cleanup(id: string): Promise<void> {\n try {\n const { file } = await this.metadataManager.getMetadata({ id })\n\n await this.client.deleteObjects({\n Bucket: this.bucket,\n Delete: {\n Objects: [\n {\n Key: this.metadataManager.generatePartKey({\n id,\n prefix: file?.metadata?.prefix ?? undefined,\n isIncomplete: true,\n }),\n },\n {\n Key: this.metadataManager.generateInfoKey({ id }),\n },\n ],\n },\n })\n } catch (error) {\n if (\n (error as AWSError)?.code &&\n ['NoSuchKey', 'NoSuchUpload', 'NotFound'].includes(\n (error as AWSError).Code || ''\n )\n ) {\n log(MediaCloudLogs.S3_STORE_FILE_NOT_FOUND)\n throw ERRORS.FILE_NOT_FOUND\n }\n throw error\n }\n }\n\n /**\n * Combine all multipart uploads into a single object\n * @param id - The upload ID to complete\n * @returns Promise that resolves to the completed upload object with storage information\n */\n public async completeMultipartUpload(id: string): Promise<Upload> {\n const metadata = await this.metadataManager.getMetadata({ id })\n const parts = await this.partsManager.retrieveParts({ id })\n\n const incompletePartInfo = await this.partsManager.downloadIncompletePart({\n id,\n })\n\n if (incompletePartInfo) {\n // Upload the incomplete part as a regular part\n await this.partsManager.uploadPart({\n metadata,\n readStream: incompletePartInfo.createReader({ cleanUpOnEnd: true }),\n partNumber: parts.length + 1,\n })\n\n // Remove the incomplete part\n await this.partsManager.deleteIncompletePart({ id })\n\n // Re-fetch parts to include the newly uploaded part\n const updatedParts = await this.partsManager.retrieveParts({ id })\n await this.partsManager.finishMultipartUpload({\n metadata,\n parts: updatedParts,\n })\n } else {\n await this.partsManager.finishMultipartUpload({ metadata, parts })\n }\n\n const completedUpload = new Upload({\n ...metadata.file,\n offset: metadata.file.size ?? 0,\n size: metadata.file.size ?? 0,\n storage: metadata.file.storage,\n })\n\n await this.metadataManager.completeMetadata({ upload: completedUpload })\n\n return completedUpload\n }\n\n /**\n * Get the full S3 URL for an uploaded file\n * @param id - The upload ID to get the URL for\n * @returns The full URL to access the uploaded file on S3\n */\n public getUrl(id: string): string {\n // Use the custom endpoint if available\n if (this.customEndpoint) {\n return `${this.customEndpoint}/${this.bucket}/${id}`\n }\n\n // Fallback to standard AWS S3 URL format\n const regionConfig = this.client.config.region\n let region: string\n\n // If region is a function, we can't resolve it synchronously, so use a fallback\n if (typeof regionConfig === 'function') {\n region = 'us-east-1' // fallback for sync calls\n } else {\n region = regionConfig || 'us-east-1'\n }\n\n // Standard AWS S3 URL format\n if (region === 'us-east-1') {\n return `https://${this.bucket}.s3.amazonaws.com/${id}`\n } else {\n return `https://${this.bucket}.s3.${region}.amazonaws.com/${id}`\n }\n }\n\n /**\n * Deletes expired incomplete uploads.\n * Returns the number of deleted uploads.\n * @returns Promise that resolves to the number of deleted uploads\n */\n async deleteExpired(): Promise<number> {\n return this.expirationManager.deleteExpired()\n }\n\n /**\n * Returns the expiration period in milliseconds\n * @return The expiration period in milliseconds\n */\n getExpiration(): number {\n return this.expirationManager.getExpiration()\n }\n}\n"],"mappings":";;;;;;;;;;;;AAgBA,MAAM,EAAE,QAAQ,iBAAiB;AAEjC,IAAa,UAAb,cAA6B,UAAU;CAkBrC,YAAY,SAAwB;AAClC,SAAO;kBAhBS,IAAI,OAAO;qBACR,IAAI,OAAO;2BACL;uBACJ;iBACN;wCACuB;EAYtC,MAAM,EACJ,mBACA,aACA,UACA,gBACA,0BACA,SACA,mCACE;EACJ,MAAM,EAAE,KAAK,QAAQ,GAAG,uBAAuB;AAE/C,OAAK,aAAa;GAChB;GACA;GACA;GACA;GACA;GACD;AAED,OAAK,SAAS;AACd,OAAK,MAAM;AACX,OAAK,SAAS,IAAI,GAAG,mBAAmB;AACxC,OAAK,iBAAiB,OAAO,mBAAmB,SAAS;AAEzD,OAAK,WAAW,YAAY,KAAK;AACjC,OAAK,cAAc,eAAe,KAAK;AACvC,OAAK,oBAAoB,qBAAqB,KAAK;AAEnD,OAAK,UAAU,WAAW,KAAK;AAC/B,OAAK,iCACH,kCAAkC,KAAK;AACzC,OAAK,sBAAsB,IAAI,UAAU,4BAA4B,GAAG;AAGxE,OAAK,kBAAkB,IAAI,kBACzB,KAAK,QACL,KAAK,QACL,KAAK,wBAAwB,KAAK,KAAK,EACvC,KAAK,oBAAoB,KAAK,KAAK,CACpC;AAED,OAAK,iBAAiB,IAAI,iBACxB,KAAK,mBACL,KAAK,eACL,KAAK,aACL,KAAK,SACN;AAED,OAAK,eAAe,IAAI,eACtB,KAAK,QACL,KAAK,QACL,KAAK,aACL,KAAK,qBACL,KAAK,iBACL,KAAK,gBACL,KAAK,oBAAoB,KAAK,KAAK,CACpC;AAED,OAAK,oBAAoB,IAAI,oBAC3B,KAAK,QACL,KAAK,QACL,KAAK,gCACL,KAAK,gBAAgB,gBAAgB,KAAK,KAAK,gBAAgB,EAC/D,KAAK,gBAAgB,gBAAgB,KAAK,KAAK,gBAAgB,CAChE;AAGD,OAAK,eAAe;;;;;;CAOtB,AAAU,0BAAmC;AAC3C,SAAO,KAAK,mCAAmC,KAAK,KAAK;;;;;;;CAQ3D,AAAU,oBAAoB,OAA6C;AACzE,MAAI,CAAC,KAAK,yBAAyB,CACjC;AAEF,SAAO,iBAAiB;;;;;;;;;CAU1B,MAAa,OAAO,QAAiC;AACnD,MAAI,eAAe,wBAAwB;EAE3C,MAAMA,UAAiD;GACrD,QAAQ,KAAK;GACb,KAAK,KAAK,gBAAgB,gBAAgB;IACxC,IAAI,OAAO;IACX,QAAQ,OAAO,UAAU,UAAU;IACpC,CAAC;GACF,UAAU,EAAE,eAAe,eAAe;GAC3C;AAED,MAAI,OAAO,UAAU,YACnB,SAAQ,cAAc,OAAO,SAAS;AAGxC,MAAI,OAAO,UAAU,aACnB,SAAQ,eAAe,OAAO,SAAS;AAGzC,MAAI,KAAK,IACP,SAAQ,MAAM,KAAK;AAGrB,SAAO,iCAAgB,IAAI,MAAM,EAAC,aAAa;EAE/C,MAAM,WAAW,MAAM,KAAK,OAAO,sBAAsB,QAAQ;AAEjE,SAAO,UAAU;GACf,MAAM;GACN,QAAQ,KAAK;GACb,MAAM,SAAS;GAChB;AACD,QAAM,KAAK,gBAAgB,aAAa;GACtC;GACA,UAAU,SAAS;GACpB,CAAC;AACF,MAAI,eAAe,2BAA2B;AAE9C,SAAO;;;;;;;;;CAUT,MAAa,MACX,UACA,IACA,QACiB;EACjB,MAAM,WAAW,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;EAI/D,MAAM,mBAAmB,KAAK,eAAe,yBAAyB,EACpE,OAAO,MAAM,KAAK,aAAa,cAAc,EAAE,IAAI,CAAC,EACrD,CAAC;EACF,MAAM,aAAa,SAAS;EAC5B,MAAM,kBAAkB;EAExB,IAAI,gBAAgB;AAEpB,MAAI,aAAa,EACf,OAAM,OAAO;AAMf,MAAI,aAAa,GAAG;GAClB,MAAM,iBAAiB,MAAM,KAAK,aAAa,uBAAuB,EACpE,IACD,CAAC;AAEF,OAAI,CAAC,eACH,OAAM,OAAO;AAGf,OAAI,eAAe,SAAS,WAC1B,OAAM,OAAO;AAIf,SAAM,KAAK,aAAa,qBAAqB,EAAE,IAAI,CAAC;AAGpD,YAAS,kBAAkB,eAAe;AAE1C,mBAAgB,OAAO,SAAS,MAC7B,mBAAmB;AAClB,WAAO,eAAe,aAAa,EAAE,cAAc,MAAM,CAAC;AAC1D,WAAO;OACL,CACL;;EAGH,MAAM,aAAa,KAAK,eAAe,oBAAoB,EACzD,OAAO,MAAM,KAAK,aAAa,cAAc,EAAE,IAAI,CAAC,EACrD,CAAC;EAWF,MAAM,YACJ,kBAVoB,MAAM,KAAK,aAAa,YAAY;GACxD;GACA,YAAY;GACZ,mBAAmB;GACnB;GACD,CAAC,IAKmC,aAAa,IAAI,aAAa;AAGnE,MAAI,SAAS,KAAK,SAAS,UACzB,KAAI;GACF,MAAM,QAAQ,MAAM,KAAK,aAAa,cAAc,EAAE,IAAI,CAAC;AAC3D,SAAM,KAAK,aAAa,sBAAsB;IAAE;IAAU;IAAO,CAAC;GAGlE,MAAM,kBAAkB,IAAI,OAAO;IACjC,GAAG,SAAS;IACZ,QAAQ;IACR,MAAM,SAAS,KAAK;IACpB,SAAS,SAAS,KAAK;IACxB,CAAC;AAEF,SAAM,KAAK,gBAAgB,iBAAiB,EAAE,QAAQ,iBAAiB,CAAC;WACjE,OAAO;AACd,OAAI,eAAe,uBAAuB;AAC1C,SAAM;;AAIV,SAAO;;;;;;;;CAST,MAAa,UAAU,IAA6B;EAClD,IAAIC;AAEJ,MAAI;AACF,cAAW,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;WAClD,OAAO;AACd,OACE,iBAAiB,aACjB,iBAAiB,YAChB,OAAoB,SAAS,cAC7B,OAAoB,SAAS,YAE9B,OAAM,OAAO;AAEf,SAAM;;EAGR,IAAIC;AAEJ,MAAI;GACF,MAAM,QAAQ,MAAM,KAAK,aAAa,cAAc,EAAE,IAAI,CAAC;AAC3D,YAAS,KAAK,eAAe,yBAAyB,EAAE,OAAO,CAAC;WACzD,OAAO;AAGd,OACG,OAAoB,SAAS,kBAC7B,OAAoB,SAAS,YAE9B,QAAO,IAAI,OAAO;IAChB,GAAG,SAAS;IACZ,UAAU,SAAS,KAAK;IACxB,QAAQ,SAAS,KAAK;IACtB,MAAM,SAAS,KAAK;IACpB,SAAS,SAAS,KAAK;IACxB,CAAC;AAGJ,OAAI,eAAe,8BAA8B;AACjD,SAAM;;EAGR,MAAM,qBAAqB,MAAM,KAAK,aAAa,sBAAsB,EACvE,IACD,CAAC;AAEF,SAAO,IAAI,OAAO;GAChB,GAAG,SAAS;GACZ,QAAQ,UAAU,sBAAsB;GACxC,MAAM,SAAS,KAAK;GACpB,SAAS,SAAS,KAAK;GACxB,CAAC;;;;;;;CAQJ,MAAM,KAAK,IAA+B;AACxC,MAAI,eAAe,sBAAsB;EACzC,IAAI,UAAU;EACd,IAAIC,YAA0B;AAE9B,SAAO,UAAU,EACf,KAAI;GACF,MAAM,OAAO,MAAM,KAAK,OAAO,UAAU;IACvC,QAAQ,KAAK;IACb,KAAK;IACN,CAAC;AACF,OAAI,eAAe,sBAAsB;AACzC,UAAO,KAAK;WACL,OAAO;AACd,OAAI,eAAe,oBAAoB;AACvC,eAAY;AACZ;AAEA,OAAI,UAAU,EAEZ,OAAM,IAAI,SAAS,YAAY,WAAW,SAAS,IAAI,CAAC;;AAK9D,MAAI,eAAe,qBAAqB;AACxC,QAAM,6BAAa,IAAI,MAAM,uBAAuB,KAAK;;;;;;;;;CAU3D,MAAa,KAAK,QAAgB,QAA+B;AAC/D,MAAI;AACF,SAAM,KAAK,OAAO,WAAW;IAC3B,QAAQ,KAAK;IACb,YAAY,UAAU,GAAG,KAAK,OAAO,GAAG,SAAS;IACjD,KAAK;IACN,CAAC;AAEF,SAAM,KAAK,OAAO,aAAa;IAC7B,QAAQ,KAAK;IACb,KAAK;IACN,CAAC;WACK,OAAO;AACd,OACG,OAAoB,QACrB;IAAC;IAAa;IAAgB;IAAW,CAAC,SACvC,MAAmB,QAAQ,GAC7B,EACD;AACA,QAAI,eAAe,wBAAwB;AAC3C,UAAM,OAAO;;AAEf,SAAM;;;;;;;;CASV,MAAa,OAAO,IAA2B;AAC7C,MAAI;GACF,MAAM,EAAE,aAAa,UAAU,SAC7B,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;AAChD,OAAI,SACF,OAAM,KAAK,OAAO,qBAAqB;IACrC,QAAQ,KAAK;IACb,KAAK;IACL,UAAU;IACX,CAAC;AAGJ,SAAM,KAAK,OAAO,cAAc;IAC9B,QAAQ,KAAK;IACb,QAAQ,EACN,SAAS;KACP,EACE,KAAK,KAAK,gBAAgB,gBAAgB;MACxC;MACA,QAAQ,MAAM,UAAU,UAAU;MACnC,CAAC,EACH;KACD,EACE,KAAK,KAAK,gBAAgB,gBAAgB;MACxC;MACA,QAAQ,MAAM,UAAU,UAAU;MAClC,cAAc;MACf,CAAC,EACH;KACD,EACE,KAAK,KAAK,gBAAgB,gBAAgB,EAAE,IAAI,CAAC,EAClD;KACF,EACF;IACF,CAAC;WACK,OAAO;AACd,OACG,OAAoB,QACrB;IAAC;IAAa;IAAgB;IAAW,CAAC,SACvC,MAAmB,QAAQ,GAC7B,EACD;AACA,QAAI,eAAe,wBAAwB;AAC3C,UAAM,OAAO;;AAEf,SAAM;;;;;;;;CASV,MAAa,QAAQ,IAA2B;AAC9C,MAAI;GACF,MAAM,EAAE,SAAS,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;AAE/D,SAAM,KAAK,OAAO,cAAc;IAC9B,QAAQ,KAAK;IACb,QAAQ,EACN,SAAS,CACP,EACE,KAAK,KAAK,gBAAgB,gBAAgB;KACxC;KACA,QAAQ,MAAM,UAAU,UAAU;KAClC,cAAc;KACf,CAAC,EACH,EACD,EACE,KAAK,KAAK,gBAAgB,gBAAgB,EAAE,IAAI,CAAC,EAClD,CACF,EACF;IACF,CAAC;WACK,OAAO;AACd,OACG,OAAoB,QACrB;IAAC;IAAa;IAAgB;IAAW,CAAC,SACvC,MAAmB,QAAQ,GAC7B,EACD;AACA,QAAI,eAAe,wBAAwB;AAC3C,UAAM,OAAO;;AAEf,SAAM;;;;;;;;CASV,MAAa,wBAAwB,IAA6B;EAChE,MAAM,WAAW,MAAM,KAAK,gBAAgB,YAAY,EAAE,IAAI,CAAC;EAC/D,MAAM,QAAQ,MAAM,KAAK,aAAa,cAAc,EAAE,IAAI,CAAC;EAE3D,MAAM,qBAAqB,MAAM,KAAK,aAAa,uBAAuB,EACxE,IACD,CAAC;AAEF,MAAI,oBAAoB;AAEtB,SAAM,KAAK,aAAa,WAAW;IACjC;IACA,YAAY,mBAAmB,aAAa,EAAE,cAAc,MAAM,CAAC;IACnE,YAAY,MAAM,SAAS;IAC5B,CAAC;AAGF,SAAM,KAAK,aAAa,qBAAqB,EAAE,IAAI,CAAC;GAGpD,MAAM,eAAe,MAAM,KAAK,aAAa,cAAc,EAAE,IAAI,CAAC;AAClE,SAAM,KAAK,aAAa,sBAAsB;IAC5C;IACA,OAAO;IACR,CAAC;QAEF,OAAM,KAAK,aAAa,sBAAsB;GAAE;GAAU;GAAO,CAAC;EAGpE,MAAM,kBAAkB,IAAI,OAAO;GACjC,GAAG,SAAS;GACZ,QAAQ,SAAS,KAAK,QAAQ;GAC9B,MAAM,SAAS,KAAK,QAAQ;GAC5B,SAAS,SAAS,KAAK;GACxB,CAAC;AAEF,QAAM,KAAK,gBAAgB,iBAAiB,EAAE,QAAQ,iBAAiB,CAAC;AAExE,SAAO;;;;;;;CAQT,AAAO,OAAO,IAAoB;AAEhC,MAAI,KAAK,eACP,QAAO,GAAG,KAAK,eAAe,GAAG,KAAK,OAAO,GAAG;EAIlD,MAAM,eAAe,KAAK,OAAO,OAAO;EACxC,IAAIC;AAGJ,MAAI,OAAO,iBAAiB,WAC1B,UAAS;MAET,UAAS,gBAAgB;AAI3B,MAAI,WAAW,YACb,QAAO,WAAW,KAAK,OAAO,oBAAoB;MAElD,QAAO,WAAW,KAAK,OAAO,MAAM,OAAO,iBAAiB;;;;;;;CAShE,MAAM,gBAAiC;AACrC,SAAO,KAAK,kBAAkB,eAAe;;;;;;CAO/C,gBAAwB;AACtB,SAAO,KAAK,kBAAkB,eAAe"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"semaphore.mjs","names":[
|
|
1
|
+
{"version":3,"file":"semaphore.mjs","names":[],"sources":["../../../../src/tus/stores/s3/semaphore.ts"],"sourcesContent":["/**\n * A semaphore implementation for controlling concurrent operations.\n * Used to limit the number of simultaneous part uploads to S3.\n */\nexport class Semaphore {\n private permits: number\n private queue: (() => void)[] = []\n\n constructor(permits: number) {\n this.permits = permits\n }\n\n private release(): void {\n this.permits++\n const next = this.queue.shift()\n if (next) {\n next()\n }\n }\n\n async acquire(): Promise<() => void> {\n return new Promise((resolve) => {\n if (this.permits > 0) {\n this.permits--\n resolve(() => this.release())\n } else {\n this.queue.push(() => {\n this.permits--\n resolve(() => this.release())\n })\n }\n })\n }\n}\n\nexport type SemaphorePermit = () => void\n"],"mappings":";;;;;AAIA,IAAa,YAAb,MAAuB;CAIrB,YAAY,SAAiB;eAFG,EAAE;AAGhC,OAAK,UAAU;;CAGjB,AAAQ,UAAgB;AACtB,OAAK;EACL,MAAM,OAAO,KAAK,MAAM,OAAO;AAC/B,MAAI,KACF,OAAM;;CAIV,MAAM,UAA+B;AACnC,SAAO,IAAI,SAAS,YAAY;AAC9B,OAAI,KAAK,UAAU,GAAG;AACpB,SAAK;AACL,kBAAc,KAAK,SAAS,CAAC;SAE7B,MAAK,MAAM,WAAW;AACpB,SAAK;AACL,kBAAc,KAAK,SAAS,CAAC;KAC7B;IAEJ"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"errors.mjs","names":[
|
|
1
|
+
{"version":3,"file":"errors.mjs","names":[],"sources":["../../src/types/errors.ts"],"sourcesContent":["export const MediaCloudErrors = {\n // Mux Errors\n MUX_CONFIG_MISSING: {\n message:\n 'Mux configuration (tokenId and tokenSecret) must be provided in pluginOptions to use Mux',\n errorCode: 400,\n },\n MUX_CONFIG_INCOMPLETE: {\n message: 'Mux configuration is missing. Mux features will not be available',\n errorCode: 400,\n },\n MUX_UPLOAD_ID_MISSING: {\n message: 'No upload-id found for upload',\n errorCode: 400,\n },\n MUX_ASSET_DELETE_ERROR: {\n message: 'Error deleting Mux asset',\n errorCode: 500,\n },\n MUX_UPLOAD_ERROR: {\n message: 'Mux video upload failed',\n errorCode: 500,\n },\n MUX_DIRECT_UPLOAD_ERROR: {\n message: 'Mux direct upload failed',\n errorCode: 500,\n },\n MUX_CREATE_UPLOAD_ERROR: {\n message: 'Error in Mux create upload handler',\n errorCode: 500,\n },\n MUX_REQUEST_NO_JSON: {\n message: 'Request does not support json() method',\n errorCode: 400,\n },\n MUX_WEBHOOK_BODY_INVALID: {\n message: 'Invalid Mux webhook body',\n errorCode: 400,\n },\n\n // S3 Errors\n S3_CONFIG_MISSING: {\n message:\n 'S3 configuration (bucket, region, accessKeyId, secretAccessKey) must be provided in pluginOptions',\n errorCode: 400,\n },\n S3_DELETE_ERROR: {\n message: 'Error deleting file from S3',\n errorCode: 500,\n },\n S3_RENAME_ERROR: {\n message: 'Error renaming file in S3',\n errorCode: 500,\n },\n S3_MOVE_ERROR: {\n message: 'Error moving file in S3',\n errorCode: 500,\n },\n S3_PREFIX_ERROR: {\n message: 'Error updating S3 prefix',\n errorCode: 500,\n },\n S3_GET_OBJECT_ERROR: {\n message: 'Error getting object from S3',\n errorCode: 500,\n },\n S3_UNIQUE_NAME_ERROR: {\n message: 'Could not find a unique file name after maximum tries',\n errorCode: 500,\n },\n\n // TUS Errors\n TUS_UPLOAD_ERROR: {\n message: 'TUS file upload error occurred',\n errorCode: 500,\n },\n TUS_CLEANUP_ERROR: {\n message: 'Error during cleanup',\n errorCode: 500,\n },\n\n // General Errors\n UNKNOWN_STORAGE_TYPE: {\n message: 'Unknown storage type for media',\n errorCode: 500,\n },\n COLLECTION_REQUIRED: {\n message: 'A collection must be specified in the plugin options',\n errorCode: 400,\n },\n\n // File Errors\n FILE_TYPE_UNKNOWN: {\n message: 'Unable to determine file type',\n errorCode: 400,\n },\n FILE_TYPE_ERROR: {\n message: 'Error determining file type',\n errorCode: 500,\n },\n FILE_TYPE_UNSUPPORTED_ERROR: {\n message: 'Unsupported file type',\n errorCode: 400,\n },\n FILE_TYPE_NOT_ALLOWED: {\n message: 'File type not allowed',\n errorCode: 400,\n },\n FILE_SIZE_EXCEEDED: {\n message: 'File size exceeds the maximum allowed limit',\n errorCode: 400,\n },\n FILENAME_SANITIZE_ERROR: {\n message: 'Error sanitizing filename',\n errorCode: 500,\n },\n FILE_MISSING_NAME: {\n message: 'Filename is missing, cannot parse file',\n errorCode: 500,\n },\n FILE_NOT_FOUND: {\n message: 'File not found',\n errorCode: 404,\n },\n FILE_DIMENSIONS_ERROR: {\n message: 'Error setting media dimensions',\n errorCode: 500,\n },\n\n // Upload Errors\n UPLOAD_NO_URL: {\n message: 'No upload URL provided, cannot parse upload ID',\n errorCode: 400,\n },\n UPLOAD_HANDLER_ERROR: {\n message: 'Upload handler error occurred',\n errorCode: 500,\n },\n UPLOAD_POLLING_ERROR: {\n message: 'Polling error for upload',\n errorCode: 500,\n },\n\n // Plugin Errors\n PLUGIN_NOT_CONFIGURED: {\n message: 'Payload Media Cloud plugin is not configured',\n errorCode: 500,\n },\n NAMING_FUNCTION_ERROR: {\n message: 'Error in namingFunction',\n errorCode: 500,\n },\n\n // HOOK Errors\n FOLDER_FETCH_ERROR: {\n message: 'Error fetching folder data',\n errorCode: 500,\n },\n} as const\n\nexport enum MediaCloudLogs {\n S3_STORE_MULTIPART_INIT = 'Initializing multipart upload',\n S3_STORE_MULTIPART_CREATED = 'Multipart upload created',\n S3_STORE_UPLOAD_FAILED = 'Failed to finish upload',\n S3_STORE_READ_ATTEMPT = 'Attempting to read file from S3',\n S3_STORE_READ_SUCCESS = 'Successfully read file from S3',\n S3_STORE_READ_RETRY = 'Failed to read file, trying again',\n S3_STORE_READ_FAILED = 'Failed to read file',\n S3_STORE_FILE_NOT_FOUND = 'No file found',\n S3_STORE_FILE_FOUND = 'File found',\n S3_STORE_RETRIEVE_PARTS_ERROR = 'Error retrieving parts',\n S3_STORE_METADATA_SAVING = 'Saving metadata',\n S3_STORE_METADATA_SAVED = 'Metadata file saved',\n S3_STORE_INCOMPLETE_PART_UPLOADED = 'Finished uploading incomplete part',\n S3_STORE_CHUNK_REMOVAL_FAILED = 'Failed to remove chunk',\n}\n"],"mappings":";AAAA,MAAa,mBAAmB;CAE9B,oBAAoB;EAClB,SACE;EACF,WAAW;EACZ;CACD,uBAAuB;EACrB,SAAS;EACT,WAAW;EACZ;CACD,uBAAuB;EACrB,SAAS;EACT,WAAW;EACZ;CACD,wBAAwB;EACtB,SAAS;EACT,WAAW;EACZ;CACD,kBAAkB;EAChB,SAAS;EACT,WAAW;EACZ;CACD,yBAAyB;EACvB,SAAS;EACT,WAAW;EACZ;CACD,yBAAyB;EACvB,SAAS;EACT,WAAW;EACZ;CACD,qBAAqB;EACnB,SAAS;EACT,WAAW;EACZ;CACD,0BAA0B;EACxB,SAAS;EACT,WAAW;EACZ;CAGD,mBAAmB;EACjB,SACE;EACF,WAAW;EACZ;CACD,iBAAiB;EACf,SAAS;EACT,WAAW;EACZ;CACD,iBAAiB;EACf,SAAS;EACT,WAAW;EACZ;CACD,eAAe;EACb,SAAS;EACT,WAAW;EACZ;CACD,iBAAiB;EACf,SAAS;EACT,WAAW;EACZ;CACD,qBAAqB;EACnB,SAAS;EACT,WAAW;EACZ;CACD,sBAAsB;EACpB,SAAS;EACT,WAAW;EACZ;CAGD,kBAAkB;EAChB,SAAS;EACT,WAAW;EACZ;CACD,mBAAmB;EACjB,SAAS;EACT,WAAW;EACZ;CAGD,sBAAsB;EACpB,SAAS;EACT,WAAW;EACZ;CACD,qBAAqB;EACnB,SAAS;EACT,WAAW;EACZ;CAGD,mBAAmB;EACjB,SAAS;EACT,WAAW;EACZ;CACD,iBAAiB;EACf,SAAS;EACT,WAAW;EACZ;CACD,6BAA6B;EAC3B,SAAS;EACT,WAAW;EACZ;CACD,uBAAuB;EACrB,SAAS;EACT,WAAW;EACZ;CACD,oBAAoB;EAClB,SAAS;EACT,WAAW;EACZ;CACD,yBAAyB;EACvB,SAAS;EACT,WAAW;EACZ;CACD,mBAAmB;EACjB,SAAS;EACT,WAAW;EACZ;CACD,gBAAgB;EACd,SAAS;EACT,WAAW;EACZ;CACD,uBAAuB;EACrB,SAAS;EACT,WAAW;EACZ;CAGD,eAAe;EACb,SAAS;EACT,WAAW;EACZ;CACD,sBAAsB;EACpB,SAAS;EACT,WAAW;EACZ;CACD,sBAAsB;EACpB,SAAS;EACT,WAAW;EACZ;CAGD,uBAAuB;EACrB,SAAS;EACT,WAAW;EACZ;CACD,uBAAuB;EACrB,SAAS;EACT,WAAW;EACZ;CAGD,oBAAoB;EAClB,SAAS;EACT,WAAW;EACZ;CACF;AAED,IAAY,4DAAL;AACL;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"buildS3Path.mjs","names":["
|
|
1
|
+
{"version":3,"file":"buildS3Path.mjs","names":["parts: Array<string>"],"sources":["../../src/utils/buildS3Path.ts"],"sourcesContent":["interface Folder {\n id: string | number\n name: unknown\n folder?: unknown\n}\n\ninterface BuildS3PathData {\n id: string | number\n name: unknown\n folder?: unknown\n}\n\nexport function buildS3Path(data: BuildS3PathData): string {\n const parts: Array<string> = []\n\n function traverseFolder(current: Folder): void {\n if (!current || typeof current !== 'object') {\n return\n }\n\n if (current.folder) {\n traverseFolder(current.folder as Folder)\n }\n\n if (current.name && typeof current.name === 'string') {\n parts.push(current.name)\n }\n }\n\n traverseFolder(data.folder as Folder)\n\n if (data.name && typeof data.name === 'string') {\n parts.push(data.name)\n }\n\n return parts.join('/')\n}\n"],"mappings":";AAYA,SAAgB,YAAY,MAA+B;CACzD,MAAMA,QAAuB,EAAE;CAE/B,SAAS,eAAe,SAAuB;AAC7C,MAAI,CAAC,WAAW,OAAO,YAAY,SACjC;AAGF,MAAI,QAAQ,OACV,gBAAe,QAAQ,OAAiB;AAG1C,MAAI,QAAQ,QAAQ,OAAO,QAAQ,SAAS,SAC1C,OAAM,KAAK,QAAQ,KAAK;;AAI5B,gBAAe,KAAK,OAAiB;AAErC,KAAI,KAAK,QAAQ,OAAO,KAAK,SAAS,SACpC,OAAM,KAAK,KAAK,KAAK;AAGvB,QAAO,MAAM,KAAK,IAAI"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"buildThumbnailURL.mjs","names":[
|
|
1
|
+
{"version":3,"file":"buildThumbnailURL.mjs","names":[],"sources":["../../src/utils/buildThumbnailURL.ts"],"sourcesContent":["import type { S3Store } from '../tus/stores/s3/s3Store'\nimport type { Storage } from '../types'\n\ninterface BuildThumbnailURLArgs {\n storage: Storage\n playbackId?: string\n s3Key?: string\n s3Store?: S3Store | null\n}\n\nexport function buildThumbnailURL(args: BuildThumbnailURLArgs) {\n const { storage, s3Store, playbackId, s3Key } = args\n\n if (storage === 'mux' && playbackId) {\n return `https://image.mux.com/${playbackId}/thumbnail.jpg?width=200&height=200&fit_mode=pad`\n }\n\n if (storage === 's3' && s3Key) {\n const url = s3Store?.getUrl(s3Key)\n return `https://wsrv.nl/?url=${url}?width=512&height=512&default=1&q=80`\n }\n}\n"],"mappings":";AAUA,SAAgB,kBAAkB,MAA6B;CAC7D,MAAM,EAAE,SAAS,SAAS,YAAY,UAAU;AAEhD,KAAI,YAAY,SAAS,WACvB,QAAO,yBAAyB,WAAW;AAG7C,KAAI,YAAY,QAAQ,MAEtB,QAAO,wBADK,SAAS,OAAO,MAAM,CACC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"defaultOptions.mjs","names":["MediaCloudDefaultPluginOptions"
|
|
1
|
+
{"version":3,"file":"defaultOptions.mjs","names":["defaultOptions: MediaCloudDefaultPluginOptions"],"sources":["../../src/utils/defaultOptions.ts"],"sourcesContent":["import { MediaCloudDefaultPluginOptions } from '../types'\n\nexport const defaultOptions: MediaCloudDefaultPluginOptions = {\n enabled: true,\n collection: 'media',\n view: 'grid',\n folders: true,\n storage: {\n 'video/*': 'mux',\n },\n}\n"],"mappings":";AAEA,MAAaA,iBAAiD;CAC5D,SAAS;CACT,YAAY;CACZ,MAAM;CACN,SAAS;CACT,SAAS,EACP,WAAW,OACZ;CACF"}
|
package/dist/utils/file.mjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"file.mjs","names":["
|
|
1
|
+
{"version":3,"file":"file.mjs","names":["magicError: UseMagicErrorReturn"],"sources":["../../src/utils/file.ts"],"sourcesContent":["import { fileTypeFromBuffer } from 'file-type'\nimport { parse } from 'pathe'\n\nimport { getFileExistsHandler } from '../endpoints/fileExistsHandler'\nimport { MediaCloudErrors } from '../types/errors'\n\nimport { MediaCloudPluginOptions, MimeType } from '../types'\nimport { useMagicError, type UseMagicErrorReturn } from '@maas/error-handler'\n\nimport type { S3Store } from '../tus/stores/s3/s3Store'\n\nconst magicError: UseMagicErrorReturn = useMagicError({\n prefix: 'PLUGIN-MEDIA-CLOUD',\n})\n\nconst { logError, throwError } = magicError\n\nconst MUX_SUPPORTED_VIDEO_FORMATS = new Set([\n 'application/mxf',\n 'video/3gpp',\n 'video/3gpp2',\n 'video/mp2t',\n 'video/mp4',\n 'video/mpeg',\n 'video/quicktime',\n 'video/webm',\n 'video/x-f4v',\n 'video/x-flv',\n 'video/x-matroska',\n 'video/x-ms-asf',\n 'video/x-ms-wmv',\n 'video/x-msvideo',\n 'video/x-mxf',\n])\n\n/**\n * Checks if a file is a video file supported by Mux\n * @param file - The file to check\n * @returns Promise that resolves to true if the file is a supported video format, false otherwise\n */\nexport async function isVideo(file: File): Promise<boolean> {\n try {\n const buffer = new Uint8Array(await file.arrayBuffer())\n const fileType = await fileTypeFromBuffer(buffer)\n\n return fileType?.mime\n ? MUX_SUPPORTED_VIDEO_FORMATS.has(fileType.mime)\n : false\n } catch {\n return false\n }\n}\n\nconst allowedMimeTypes = new Set<MimeType>([\n 'application/mxf',\n 'video/3gpp',\n 'video/3gpp2',\n 'video/mp2t',\n 'video/mp4',\n 'video/mpeg',\n 'video/quicktime',\n 'video/webm',\n 'video/x-f4v',\n 'video/x-flv',\n 'video/x-matroska',\n 'video/x-ms-asf',\n 'video/x-ms-wmv',\n 'video/x-msvideo',\n 'video/x-mxf',\n 'image/apng',\n 'image/avif',\n 'image/gif',\n 'image/jpeg',\n 'image/png',\n 'image/svg+xml',\n 'image/webp',\n])\n\n/**\n * Check the mime type against allowed mime types\n * @param mimeType - The file to check\n */\nfunction validateMimeType(mimeType: string) {\n if (!allowedMimeTypes.has(mimeType as MimeType)) {\n throwError(MediaCloudErrors.FILE_TYPE_UNSUPPORTED_ERROR)\n }\n}\n\n/**\n * Gets the MIME type of a file using file-type library\n * @param file - The file to check\n * @returns Promise that resolves to the MIME type of the file or undefined if it cannot be determined\n */\nexport async function getMimeType(file: File): Promise<MimeType | undefined> {\n try {\n const buffer = new Uint8Array(await file.arrayBuffer())\n const fileType = await fileTypeFromBuffer(buffer)\n\n if (!fileType?.mime) {\n logError(MediaCloudErrors.FILE_TYPE_ERROR.message)\n return undefined\n }\n\n validateMimeType(fileType.mime)\n return fileType.mime as MimeType\n } catch (_error) {\n logError(MediaCloudErrors.FILE_TYPE_ERROR.message)\n return undefined\n }\n}\n\n/**\n * Generates a unique filename by appending a random suffix if needed\n * @param originalFilename - The original filename\n * @returns A unique filename with a random suffix appended before the extension\n */\nexport function generateUniqueFilename(originalFilename: string): string {\n const timestamp = Date.now().toString(36)\n // Parse name and extension\n const { name, ext } = parse(originalFilename)\n\n // Sanitize name\n const sanitizedName = sanitizeFilename(name)\n\n return `${sanitizedName}-${timestamp}${ext}`\n}\n\n/**\n * Sanitizes a filename by removing/replacing invalid characters and handling edge cases\n * Converts all non-alphanumeric characters (except dots for extensions) to underscores\n * @param filename - The filename to sanitize\n * @returns A sanitized filename safe for all operating systems\n * @throws {TypeError} When filename is not a string\n */\nexport function sanitizeFilename(filename: string): string {\n try {\n const parsed = parse(filename)\n const sanitized = parsed.name\n .replace(/[^a-zA-Z0-9_.-]/g, '_')\n .replace(/_{2,}/g, '_')\n .replace(/^_|_$/g, '')\n return `${sanitized}${parsed.ext}`\n } catch (_error) {\n logError(MediaCloudErrors.FILENAME_SANITIZE_ERROR.message)\n return filename\n }\n}\n\ninterface CreateFileEndpointsArgs {\n getS3Store: () => S3Store\n pluginOptions: MediaCloudPluginOptions\n}\n\nexport function createFileEndpoints(args: CreateFileEndpointsArgs) {\n const { getS3Store, pluginOptions } = args\n const collection = pluginOptions.collection\n\n magicError.assert(collection, MediaCloudErrors.COLLECTION_REQUIRED)\n\n return [\n {\n handler: getFileExistsHandler({ getS3Store, collection }),\n method: 'get' as const,\n path: '/uploads/:filename/exists',\n },\n ]\n}\n"],"mappings":";;;;;;;AAWA,MAAMA,aAAkC,cAAc,EACpD,QAAQ,sBACT,CAAC;AAEF,MAAM,EAAE,UAAU,eAAe;AAEjC,MAAM,8BAA8B,IAAI,IAAI;CAC1C;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACD,CAAC;;;;;;AAOF,eAAsB,QAAQ,MAA8B;AAC1D,KAAI;EAEF,MAAM,WAAW,MAAM,mBADR,IAAI,WAAW,MAAM,KAAK,aAAa,CAAC,CACN;AAEjD,SAAO,UAAU,OACb,4BAA4B,IAAI,SAAS,KAAK,GAC9C;SACE;AACN,SAAO;;;AAIX,MAAM,mBAAmB,IAAI,IAAc;CACzC;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACA;CACD,CAAC;;;;;AAMF,SAAS,iBAAiB,UAAkB;AAC1C,KAAI,CAAC,iBAAiB,IAAI,SAAqB,CAC7C,YAAW,iBAAiB,4BAA4B;;;;;;;AAS5D,eAAsB,YAAY,MAA2C;AAC3E,KAAI;EAEF,MAAM,WAAW,MAAM,mBADR,IAAI,WAAW,MAAM,KAAK,aAAa,CAAC,CACN;AAEjD,MAAI,CAAC,UAAU,MAAM;AACnB,YAAS,iBAAiB,gBAAgB,QAAQ;AAClD;;AAGF,mBAAiB,SAAS,KAAK;AAC/B,SAAO,SAAS;UACT,QAAQ;AACf,WAAS,iBAAiB,gBAAgB,QAAQ;AAClD;;;;;;;;AASJ,SAAgB,uBAAuB,kBAAkC;CACvE,MAAM,YAAY,KAAK,KAAK,CAAC,SAAS,GAAG;CAEzC,MAAM,EAAE,MAAM,QAAQ,MAAM,iBAAiB;AAK7C,QAAO,GAFe,iBAAiB,KAAK,CAEpB,GAAG,YAAY;;;;;;;;;AAUzC,SAAgB,iBAAiB,UAA0B;AACzD,KAAI;EACF,MAAM,SAAS,MAAM,SAAS;AAK9B,SAAO,GAJW,OAAO,KACtB,QAAQ,oBAAoB,IAAI,CAChC,QAAQ,UAAU,IAAI,CACtB,QAAQ,UAAU,GAAG,GACF,OAAO;UACtB,QAAQ;AACf,WAAS,iBAAiB,wBAAwB,QAAQ;AAC1D,SAAO;;;AASX,SAAgB,oBAAoB,MAA+B;CACjE,MAAM,EAAE,YAAY,kBAAkB;CACtC,MAAM,aAAa,cAAc;AAEjC,YAAW,OAAO,YAAY,iBAAiB,oBAAoB;AAEnE,QAAO,CACL;EACE,SAAS,qBAAqB;GAAE;GAAY;GAAY,CAAC;EACzD,QAAQ;EACR,MAAM;EACP,CACF"}
|
package/dist/utils/mux.mjs.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"mux.mjs","names":["
|
|
1
|
+
{"version":3,"file":"mux.mjs","names":["magicError: UseMagicErrorReturn"],"sources":["../../src/utils/mux.ts"],"sourcesContent":["import Mux from '@mux/mux-node'\n\nimport { MediaCloudErrors } from '../types/errors'\nimport { getMuxAssetHandler } from '../endpoints/muxAssetHandler'\nimport { getMuxCreateUploadHandler } from '../endpoints/muxCreateUploadHandler'\nimport { getMuxWebhookHandler } from '../endpoints/muxWebhookHandler'\nimport { useMagicError } from '@maas/error-handler'\n\nimport type { MediaCloudPluginOptions } from '../types'\nimport type { UseMagicErrorReturn } from '@maas/error-handler'\n\nconst magicError: UseMagicErrorReturn = useMagicError({\n prefix: 'PLUGIN-MEDIA-CLOUD',\n})\n\n/**\n * Creates Mux-related endpoints for asset handling\n * @param args - The arguments for creating Mux endpoints\n * @returns An array of Mux endpoint configurations\n */\ninterface CreateMuxEndpointsArgs {\n getMuxClient: () => Mux\n pluginOptions: MediaCloudPluginOptions\n}\n\nexport function createMuxEndpoints(args: CreateMuxEndpointsArgs) {\n const { getMuxClient, pluginOptions } = args\n\n const collection = pluginOptions.collection\n\n magicError.assert(collection, MediaCloudErrors.COLLECTION_REQUIRED)\n\n return [\n {\n handler: getMuxWebhookHandler({ getMuxClient, collection }),\n method: 'post' as const,\n path: '/mux/webhook',\n },\n {\n handler: getMuxCreateUploadHandler({ getMuxClient, pluginOptions }),\n method: 'post' as const,\n path: '/mux/upload',\n },\n {\n handler: getMuxAssetHandler({ getMuxClient, collection }),\n method: 'get' as const,\n path: '/mux/asset',\n },\n ]\n}\n\n/**\n * Validates Mux configuration options\n * @param muxConfig - The Mux configuration to validate\n * @throws {Error} When required Mux configuration is missing\n */\nexport function validateMuxConfig(\n muxConfig: MediaCloudPluginOptions['mux']\n): void {\n magicError.assert(muxConfig, MediaCloudErrors.MUX_CONFIG_INCOMPLETE)\n magicError.assert(muxConfig?.tokenId, MediaCloudErrors.MUX_CONFIG_MISSING)\n magicError.assert(muxConfig?.tokenSecret, MediaCloudErrors.MUX_CONFIG_MISSING)\n}\n\n/**\n * Creates a new Mux client instance\n * @param muxConfig - The Mux configuration options\n * @returns A configured Mux client\n */\nexport function createMuxClient(\n muxConfig: MediaCloudPluginOptions['mux']\n): Mux {\n validateMuxConfig(muxConfig)\n\n return new Mux({\n tokenId: muxConfig?.tokenId,\n tokenSecret: muxConfig?.tokenSecret,\n webhookSecret: muxConfig?.webhookSecret,\n })\n}\n"],"mappings":";;;;;;;;AAWA,MAAMA,aAAkC,cAAc,EACpD,QAAQ,sBACT,CAAC;AAYF,SAAgB,mBAAmB,MAA8B;CAC/D,MAAM,EAAE,cAAc,kBAAkB;CAExC,MAAM,aAAa,cAAc;AAEjC,YAAW,OAAO,YAAY,iBAAiB,oBAAoB;AAEnE,QAAO;EACL;GACE,SAAS,qBAAqB;IAAE;IAAc;IAAY,CAAC;GAC3D,QAAQ;GACR,MAAM;GACP;EACD;GACE,SAAS,0BAA0B;IAAE;IAAc;IAAe,CAAC;GACnE,QAAQ;GACR,MAAM;GACP;EACD;GACE,SAAS,mBAAmB;IAAE;IAAc;IAAY,CAAC;GACzD,QAAQ;GACR,MAAM;GACP;EACF;;;;;;;AAQH,SAAgB,kBACd,WACM;AACN,YAAW,OAAO,WAAW,iBAAiB,sBAAsB;AACpE,YAAW,OAAO,WAAW,SAAS,iBAAiB,mBAAmB;AAC1E,YAAW,OAAO,WAAW,aAAa,iBAAiB,mBAAmB;;;;;;;AAQhF,SAAgB,gBACd,WACK;AACL,mBAAkB,UAAU;AAE5B,QAAO,IAAI,IAAI;EACb,SAAS,WAAW;EACpB,aAAa,WAAW;EACxB,eAAe,WAAW;EAC3B,CAAC"}
|