@payloadcms/storage-s3 3.83.0 → 3.84.0-canary.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/client/S3ClientUploadHandler.d.ts.map +1 -1
- package/dist/client/S3ClientUploadHandler.js +5 -2
- package/dist/client/S3ClientUploadHandler.js.map +1 -1
- package/dist/deleteFile.js +2 -2
- package/dist/deleteFile.js.map +1 -1
- package/dist/generateSignedURL.d.ts.map +1 -1
- package/dist/generateSignedURL.js +4 -6
- package/dist/generateSignedURL.js.map +1 -1
- package/dist/generateURL.d.ts.map +1 -1
- package/dist/generateURL.js +6 -2
- package/dist/generateURL.js.map +1 -1
- package/dist/getFile.d.ts.map +1 -1
- package/dist/getFile.js +2 -3
- package/dist/getFile.js.map +1 -1
- package/dist/uploadFile.js +2 -2
- package/dist/uploadFile.js.map +1 -1
- package/package.json +4 -4
- package/dist/handleDelete.d.ts +0 -9
- package/dist/handleDelete.d.ts.map +0 -1
- package/dist/handleDelete.js +0 -11
- package/dist/handleDelete.js.map +0 -1
- package/dist/handleUpload.d.ts +0 -13
- package/dist/handleUpload.d.ts.map +0 -1
- package/dist/handleUpload.js +0 -37
- package/dist/handleUpload.js.map +0 -1
- package/dist/staticHandler.d.ts +0 -21
- package/dist/staticHandler.d.ts.map +0 -1
- package/dist/staticHandler.js +0 -169
- package/dist/staticHandler.js.map +0 -1
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"S3ClientUploadHandler.d.ts","sourceRoot":"","sources":["../../src/client/S3ClientUploadHandler.ts"],"names":[],"mappings":"AAIA,eAAO,MAAM,qBAAqB;;;;;;;
|
|
1
|
+
{"version":3,"file":"S3ClientUploadHandler.d.ts","sourceRoot":"","sources":["../../src/client/S3ClientUploadHandler.ts"],"names":[],"mappings":"AAIA,eAAO,MAAM,qBAAqB;;;;;;;aA0B5B,OAAM,aAkBV,CAAA"}
|
|
@@ -8,6 +8,7 @@ export const S3ClientUploadHandler = createClientUploadHandler({
|
|
|
8
8
|
path: serverHandlerPath,
|
|
9
9
|
serverURL
|
|
10
10
|
});
|
|
11
|
+
// get the signed URL from the server
|
|
11
12
|
const response = await fetch(endpointRoute, {
|
|
12
13
|
body: JSON.stringify({
|
|
13
14
|
collectionSlug,
|
|
@@ -23,7 +24,8 @@ export const S3ClientUploadHandler = createClientUploadHandler({
|
|
|
23
24
|
const { errors } = await response.json();
|
|
24
25
|
throw new Error(errors.reduce((acc, err)=>`${acc ? `${acc}, ` : ''}${err.message}`, ''));
|
|
25
26
|
}
|
|
26
|
-
const { docPrefix:
|
|
27
|
+
const { docPrefix: sanitizedDocPrefix, url } = await response.json();
|
|
28
|
+
// upload the file directly to S3 using the signed URL
|
|
27
29
|
await fetch(url, {
|
|
28
30
|
body: file,
|
|
29
31
|
headers: {
|
|
@@ -32,8 +34,9 @@ export const S3ClientUploadHandler = createClientUploadHandler({
|
|
|
32
34
|
},
|
|
33
35
|
method: 'PUT'
|
|
34
36
|
});
|
|
37
|
+
// return the docPrefix so the client can update the field value accordingly
|
|
35
38
|
return {
|
|
36
|
-
prefix:
|
|
39
|
+
prefix: sanitizedDocPrefix
|
|
37
40
|
};
|
|
38
41
|
}
|
|
39
42
|
});
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../../src/client/S3ClientUploadHandler.ts"],"sourcesContent":["'use client'\nimport { createClientUploadHandler } from '@payloadcms/plugin-cloud-storage/client'\nimport { formatAdminURL } from 'payload/shared'\n\nexport const S3ClientUploadHandler = createClientUploadHandler({\n handler: async ({ apiRoute, collectionSlug, docPrefix, file, serverHandlerPath, serverURL }) => {\n const endpointRoute = formatAdminURL({\n apiRoute,\n path: serverHandlerPath,\n serverURL,\n })\n\n const response = await fetch(endpointRoute, {\n body: JSON.stringify({\n collectionSlug,\n docPrefix,\n filename: file.name,\n filesize: file.size,\n mimeType: file.type,\n }),\n credentials: 'include',\n method: 'POST',\n })\n\n if (!response.ok) {\n const { errors } = (await response.json()) as {\n errors: { message: string }[]\n }\n\n throw new Error(errors.reduce((acc, err) => `${acc ? `${acc}, ` : ''}${err.message}`, ''))\n }\n\n const { docPrefix:
|
|
1
|
+
{"version":3,"sources":["../../src/client/S3ClientUploadHandler.ts"],"sourcesContent":["'use client'\nimport { createClientUploadHandler } from '@payloadcms/plugin-cloud-storage/client'\nimport { formatAdminURL } from 'payload/shared'\n\nexport const S3ClientUploadHandler = createClientUploadHandler({\n handler: async ({ apiRoute, collectionSlug, docPrefix, file, serverHandlerPath, serverURL }) => {\n const endpointRoute = formatAdminURL({\n apiRoute,\n path: serverHandlerPath,\n serverURL,\n })\n\n // get the signed URL from the server\n const response = await fetch(endpointRoute, {\n body: JSON.stringify({\n collectionSlug,\n docPrefix,\n filename: file.name,\n filesize: file.size,\n mimeType: file.type,\n }),\n credentials: 'include',\n method: 'POST',\n })\n\n if (!response.ok) {\n const { errors } = (await response.json()) as {\n errors: { message: string }[]\n }\n\n throw new Error(errors.reduce((acc, err) => `${acc ? `${acc}, ` : ''}${err.message}`, ''))\n }\n\n const { docPrefix: sanitizedDocPrefix, url } = (await response.json()) as {\n docPrefix: string\n url: string\n }\n\n // upload the file directly to S3 using the signed URL\n await fetch(url, {\n body: file,\n headers: { 'Content-Length': file.size.toString(), 'Content-Type': file.type },\n method: 'PUT',\n })\n\n // return the docPrefix so the client can update the field value accordingly\n return { prefix: sanitizedDocPrefix }\n },\n})\n"],"names":["createClientUploadHandler","formatAdminURL","S3ClientUploadHandler","handler","apiRoute","collectionSlug","docPrefix","file","serverHandlerPath","serverURL","endpointRoute","path","response","fetch","body","JSON","stringify","filename","name","filesize","size","mimeType","type","credentials","method","ok","errors","json","Error","reduce","acc","err","message","sanitizedDocPrefix","url","headers","toString","prefix"],"mappings":"AAAA;AACA,SAASA,yBAAyB,QAAQ,0CAAyC;AACnF,SAASC,cAAc,QAAQ,iBAAgB;AAE/C,OAAO,MAAMC,wBAAwBF,0BAA0B;IAC7DG,SAAS,OAAO,EAAEC,QAAQ,EAAEC,cAAc,EAAEC,SAAS,EAAEC,IAAI,EAAEC,iBAAiB,EAAEC,SAAS,EAAE;QACzF,MAAMC,gBAAgBT,eAAe;YACnCG;YACAO,MAAMH;YACNC;QACF;QAEA,qCAAqC;QACrC,MAAMG,WAAW,MAAMC,MAAMH,eAAe;YAC1CI,MAAMC,KAAKC,SAAS,CAAC;gBACnBX;gBACAC;gBACAW,UAAUV,KAAKW,IAAI;gBACnBC,UAAUZ,KAAKa,IAAI;gBACnBC,UAAUd,KAAKe,IAAI;YACrB;YACAC,aAAa;YACbC,QAAQ;QACV;QAEA,IAAI,CAACZ,SAASa,EAAE,EAAE;YAChB,MAAM,EAAEC,MAAM,EAAE,GAAI,MAAMd,SAASe,IAAI;YAIvC,MAAM,IAAIC,MAAMF,OAAOG,MAAM,CAAC,CAACC,KAAKC,MAAQ,GAAGD,MAAM,GAAGA,IAAI,EAAE,CAAC,GAAG,KAAKC,IAAIC,OAAO,EAAE,EAAE;QACxF;QAEA,MAAM,EAAE1B,WAAW2B,kBAAkB,EAAEC,GAAG,EAAE,GAAI,MAAMtB,SAASe,IAAI;QAKnE,sDAAsD;QACtD,MAAMd,MAAMqB,KAAK;YACfpB,MAAMP;YACN4B,SAAS;gBAAE,kBAAkB5B,KAAKa,IAAI,CAACgB,QAAQ;gBAAI,gBAAgB7B,KAAKe,IAAI;YAAC;YAC7EE,QAAQ;QACV;QAEA,4EAA4E;QAC5E,OAAO;YAAEa,QAAQJ;QAAmB;IACtC;AACF,GAAE"}
|
package/dist/deleteFile.js
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities';
|
|
2
2
|
export async function deleteFile({ bucket, client, collectionPrefix = '', docPrefix, filename, useCompositePrefixes = false }) {
|
|
3
|
-
const
|
|
3
|
+
const { fileKey } = getFileKey({
|
|
4
4
|
collectionPrefix,
|
|
5
5
|
docPrefix,
|
|
6
6
|
filename,
|
|
@@ -8,7 +8,7 @@ export async function deleteFile({ bucket, client, collectionPrefix = '', docPre
|
|
|
8
8
|
});
|
|
9
9
|
await client.deleteObject({
|
|
10
10
|
Bucket: bucket,
|
|
11
|
-
Key:
|
|
11
|
+
Key: fileKey
|
|
12
12
|
});
|
|
13
13
|
}
|
|
14
14
|
|
package/dist/deleteFile.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/deleteFile.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\n\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\n\ninterface DeleteArgs {\n bucket: string\n client: AWS.S3\n collectionPrefix?: string\n docPrefix: string\n filename: string\n useCompositePrefixes?: boolean\n}\n\nexport async function deleteFile({\n bucket,\n client,\n collectionPrefix = '',\n docPrefix,\n filename,\n useCompositePrefixes = false,\n}: DeleteArgs): Promise<void> {\n const
|
|
1
|
+
{"version":3,"sources":["../src/deleteFile.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\n\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\n\ninterface DeleteArgs {\n bucket: string\n client: AWS.S3\n collectionPrefix?: string\n docPrefix: string\n filename: string\n useCompositePrefixes?: boolean\n}\n\nexport async function deleteFile({\n bucket,\n client,\n collectionPrefix = '',\n docPrefix,\n filename,\n useCompositePrefixes = false,\n}: DeleteArgs): Promise<void> {\n const { fileKey } = getFileKey({\n collectionPrefix,\n docPrefix,\n filename,\n useCompositePrefixes,\n })\n\n await client.deleteObject({\n Bucket: bucket,\n Key: fileKey,\n })\n}\n"],"names":["getFileKey","deleteFile","bucket","client","collectionPrefix","docPrefix","filename","useCompositePrefixes","fileKey","deleteObject","Bucket","Key"],"mappings":"AAEA,SAASA,UAAU,QAAQ,6CAA4C;AAWvE,OAAO,eAAeC,WAAW,EAC/BC,MAAM,EACNC,MAAM,EACNC,mBAAmB,EAAE,EACrBC,SAAS,EACTC,QAAQ,EACRC,uBAAuB,KAAK,EACjB;IACX,MAAM,EAAEC,OAAO,EAAE,GAAGR,WAAW;QAC7BI;QACAC;QACAC;QACAC;IACF;IAEA,MAAMJ,OAAOM,YAAY,CAAC;QACxBC,QAAQR;QACRS,KAAKH;IACP;AACF"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"generateSignedURL.d.ts","sourceRoot":"","sources":["../src/generateSignedURL.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AACjF,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,SAAS,CAAA;AAE7C,OAAO,KAAK,GAAG,MAAM,oBAAoB,CAAA;
|
|
1
|
+
{"version":3,"file":"generateSignedURL.d.ts","sourceRoot":"","sources":["../src/generateSignedURL.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AACjF,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,SAAS,CAAA;AAE7C,OAAO,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAKzC,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAA;AAMlD,UAAU,IAAI;IACZ,MAAM,CAAC,EAAE,mBAAmB,CAAA;IAC5B,GAAG,CAAC,EAAE,SAAS,GAAG,aAAa,CAAA;IAC/B,MAAM,EAAE,MAAM,CAAA;IACd,WAAW,EAAE,gBAAgB,CAAC,aAAa,CAAC,CAAA;IAC5C,gBAAgB,EAAE,MAAM,GAAG,CAAC,EAAE,CAAA;IAC9B,oBAAoB,CAAC,EAAE,OAAO,CAAA;CAC/B;AAID,eAAO,MAAM,2BAA2B,kFAOrC,IAAI,KAAG,cAyET,CAAA"}
|
|
@@ -2,7 +2,6 @@ import * as AWS from '@aws-sdk/client-s3';
|
|
|
2
2
|
import { getSignedUrl } from '@aws-sdk/s3-request-presigner';
|
|
3
3
|
import { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities';
|
|
4
4
|
import { APIError, Forbidden } from 'payload';
|
|
5
|
-
import { sanitizeFilename } from 'payload/shared';
|
|
6
5
|
const bytesToMB = (bytes)=>{
|
|
7
6
|
return bytes / 1024 / 1024;
|
|
8
7
|
};
|
|
@@ -28,11 +27,10 @@ export const getGenerateSignedURLHandler = ({ access = defaultAccess, acl, bucke
|
|
|
28
27
|
})) {
|
|
29
28
|
throw new Forbidden();
|
|
30
29
|
}
|
|
31
|
-
const
|
|
32
|
-
const fileKey = getFileKey({
|
|
30
|
+
const { fileKey, sanitizedDocPrefix } = getFileKey({
|
|
33
31
|
collectionPrefix,
|
|
34
|
-
docPrefix
|
|
35
|
-
filename
|
|
32
|
+
docPrefix,
|
|
33
|
+
filename,
|
|
36
34
|
useCompositePrefixes
|
|
37
35
|
});
|
|
38
36
|
const signableHeaders = new Set();
|
|
@@ -54,7 +52,7 @@ export const getGenerateSignedURLHandler = ({ access = defaultAccess, acl, bucke
|
|
|
54
52
|
signableHeaders
|
|
55
53
|
});
|
|
56
54
|
return Response.json({
|
|
57
|
-
docPrefix:
|
|
55
|
+
docPrefix: sanitizedDocPrefix,
|
|
58
56
|
url
|
|
59
57
|
});
|
|
60
58
|
};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/generateSignedURL.ts"],"sourcesContent":["import type { ClientUploadsAccess } from '@payloadcms/plugin-cloud-storage/types'\nimport type { PayloadHandler } from 'payload'\n\nimport * as AWS from '@aws-sdk/client-s3'\nimport { getSignedUrl } from '@aws-sdk/s3-request-presigner'\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\nimport { APIError, Forbidden } from 'payload'\
|
|
1
|
+
{"version":3,"sources":["../src/generateSignedURL.ts"],"sourcesContent":["import type { ClientUploadsAccess } from '@payloadcms/plugin-cloud-storage/types'\nimport type { PayloadHandler } from 'payload'\n\nimport * as AWS from '@aws-sdk/client-s3'\nimport { getSignedUrl } from '@aws-sdk/s3-request-presigner'\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\nimport { APIError, Forbidden } from 'payload'\n\nimport type { S3StorageOptions } from './index.js'\n\nconst bytesToMB = (bytes: number) => {\n return bytes / 1024 / 1024\n}\n\ninterface Args {\n access?: ClientUploadsAccess\n acl?: 'private' | 'public-read'\n bucket: string\n collections: S3StorageOptions['collections']\n getStorageClient: () => AWS.S3\n useCompositePrefixes?: boolean\n}\n\nconst defaultAccess: Args['access'] = ({ req }) => !!req.user\n\nexport const getGenerateSignedURLHandler = ({\n access = defaultAccess,\n acl,\n bucket,\n collections,\n getStorageClient,\n useCompositePrefixes = false,\n}: Args): PayloadHandler => {\n return async (req) => {\n if (!req.json) {\n throw new APIError('Content-Type expected to be application/json', 400)\n }\n\n let filesizeLimit = req.payload.config.upload.limits?.fileSize\n\n if (filesizeLimit === Infinity) {\n filesizeLimit = undefined\n }\n\n const { collectionSlug, docPrefix, filename, filesize, mimeType } = (await req.json()) as {\n collectionSlug: string\n docPrefix?: string\n filename: string\n filesize: number\n mimeType: string\n }\n\n const collectionS3Config = collections[collectionSlug]\n if (!collectionS3Config) {\n throw new APIError(`Collection ${collectionSlug} was not found in S3 options`)\n }\n\n const collectionPrefix =\n (typeof collectionS3Config === 'object' && collectionS3Config.prefix) || ''\n\n if (!(await access({ collectionSlug, req }))) {\n throw new Forbidden()\n }\n\n const { fileKey, sanitizedDocPrefix } = getFileKey({\n collectionPrefix,\n docPrefix,\n filename,\n useCompositePrefixes,\n })\n\n const signableHeaders = new Set<string>()\n\n if (filesizeLimit) {\n if (filesize > filesizeLimit) {\n throw new APIError(\n `Exceeded file size limit. Limit: ${bytesToMB(filesizeLimit).toFixed(2)}MB, got: ${bytesToMB(filesize).toFixed(2)}MB`,\n 400,\n )\n }\n\n // Still force S3 to validate\n signableHeaders.add('content-length')\n }\n\n const url = await getSignedUrl(\n getStorageClient(),\n new AWS.PutObjectCommand({\n ACL: acl,\n Bucket: bucket,\n ContentLength: filesizeLimit ? Math.min(filesize, filesizeLimit) : undefined,\n ContentType: mimeType,\n Key: fileKey,\n }),\n {\n expiresIn: 600,\n signableHeaders,\n },\n )\n\n return Response.json({\n docPrefix: sanitizedDocPrefix,\n url,\n })\n }\n}\n"],"names":["AWS","getSignedUrl","getFileKey","APIError","Forbidden","bytesToMB","bytes","defaultAccess","req","user","getGenerateSignedURLHandler","access","acl","bucket","collections","getStorageClient","useCompositePrefixes","json","filesizeLimit","payload","config","upload","limits","fileSize","Infinity","undefined","collectionSlug","docPrefix","filename","filesize","mimeType","collectionS3Config","collectionPrefix","prefix","fileKey","sanitizedDocPrefix","signableHeaders","Set","toFixed","add","url","PutObjectCommand","ACL","Bucket","ContentLength","Math","min","ContentType","Key","expiresIn","Response"],"mappings":"AAGA,YAAYA,SAAS,qBAAoB;AACzC,SAASC,YAAY,QAAQ,gCAA+B;AAC5D,SAASC,UAAU,QAAQ,6CAA4C;AACvE,SAASC,QAAQ,EAAEC,SAAS,QAAQ,UAAS;AAI7C,MAAMC,YAAY,CAACC;IACjB,OAAOA,QAAQ,OAAO;AACxB;AAWA,MAAMC,gBAAgC,CAAC,EAAEC,GAAG,EAAE,GAAK,CAAC,CAACA,IAAIC,IAAI;AAE7D,OAAO,MAAMC,8BAA8B,CAAC,EAC1CC,SAASJ,aAAa,EACtBK,GAAG,EACHC,MAAM,EACNC,WAAW,EACXC,gBAAgB,EAChBC,uBAAuB,KAAK,EACvB;IACL,OAAO,OAAOR;QACZ,IAAI,CAACA,IAAIS,IAAI,EAAE;YACb,MAAM,IAAId,SAAS,gDAAgD;QACrE;QAEA,IAAIe,gBAAgBV,IAAIW,OAAO,CAACC,MAAM,CAACC,MAAM,CAACC,MAAM,EAAEC;QAEtD,IAAIL,kBAAkBM,UAAU;YAC9BN,gBAAgBO;QAClB;QAEA,MAAM,EAAEC,cAAc,EAAEC,SAAS,EAAEC,QAAQ,EAAEC,QAAQ,EAAEC,QAAQ,EAAE,GAAI,MAAMtB,IAAIS,IAAI;QAQnF,MAAMc,qBAAqBjB,WAAW,CAACY,eAAe;QACtD,IAAI,CAACK,oBAAoB;YACvB,MAAM,IAAI5B,SAAS,CAAC,WAAW,EAAEuB,eAAe,4BAA4B,CAAC;QAC/E;QAEA,MAAMM,mBACJ,AAAC,OAAOD,uBAAuB,YAAYA,mBAAmBE,MAAM,IAAK;QAE3E,IAAI,CAAE,MAAMtB,OAAO;YAAEe;YAAgBlB;QAAI,IAAK;YAC5C,MAAM,IAAIJ;QACZ;QAEA,MAAM,EAAE8B,OAAO,EAAEC,kBAAkB,EAAE,GAAGjC,WAAW;YACjD8B;YACAL;YACAC;YACAZ;QACF;QAEA,MAAMoB,kBAAkB,IAAIC;QAE5B,IAAInB,eAAe;YACjB,IAAIW,WAAWX,eAAe;gBAC5B,MAAM,IAAIf,SACR,CAAC,iCAAiC,EAAEE,UAAUa,eAAeoB,OAAO,CAAC,GAAG,SAAS,EAAEjC,UAAUwB,UAAUS,OAAO,CAAC,GAAG,EAAE,CAAC,EACrH;YAEJ;YAEA,6BAA6B;YAC7BF,gBAAgBG,GAAG,CAAC;QACtB;QAEA,MAAMC,MAAM,MAAMvC,aAChBc,oBACA,IAAIf,IAAIyC,gBAAgB,CAAC;YACvBC,KAAK9B;YACL+B,QAAQ9B;YACR+B,eAAe1B,gBAAgB2B,KAAKC,GAAG,CAACjB,UAAUX,iBAAiBO;YACnEsB,aAAajB;YACbkB,KAAKd;QACP,IACA;YACEe,WAAW;YACXb;QACF;QAGF,OAAOc,SAASjC,IAAI,CAAC;YACnBU,WAAWQ;YACXK;QACF;IACF;AACF,EAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"generateURL.d.ts","sourceRoot":"","sources":["../src/generateURL.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;
|
|
1
|
+
{"version":3,"file":"generateURL.d.ts","sourceRoot":"","sources":["../src/generateURL.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAK9C,UAAU,eAAe;IACvB,MAAM,EAAE,MAAM,CAAA;IACd,gBAAgB,CAAC,EAAE,MAAM,CAAA;IACzB,QAAQ,CAAC,EAAE,GAAG,CAAC,cAAc,CAAC,UAAU,CAAC,CAAA;IACzC,QAAQ,EAAE,MAAM,CAAA;IAChB,MAAM,EAAE,MAAM,CAAA;IACd,oBAAoB,CAAC,EAAE,OAAO,CAAA;CAC/B;AAED,wBAAgB,WAAW,CAAC,EAC1B,MAAM,EACN,gBAAqB,EACrB,QAAQ,EACR,QAAQ,EACR,MAAM,EACN,oBAA4B,GAC7B,EAAE,eAAe,GAAG,MAAM,CAa1B"}
|
package/dist/generateURL.js
CHANGED
|
@@ -1,11 +1,15 @@
|
|
|
1
1
|
import { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities';
|
|
2
|
+
import path from 'path';
|
|
2
3
|
export function generateURL({ bucket, collectionPrefix = '', endpoint, filename, prefix, useCompositePrefixes = false }) {
|
|
3
|
-
const fileKey = getFileKey({
|
|
4
|
+
const { fileKey: rawFileKey } = getFileKey({
|
|
4
5
|
collectionPrefix,
|
|
5
6
|
docPrefix: prefix,
|
|
6
|
-
filename
|
|
7
|
+
filename,
|
|
7
8
|
useCompositePrefixes
|
|
8
9
|
});
|
|
10
|
+
const dir = path.posix.dirname(rawFileKey);
|
|
11
|
+
const encodedFilename = encodeURIComponent(path.posix.basename(rawFileKey));
|
|
12
|
+
const fileKey = dir === '.' ? encodedFilename : path.posix.join(dir, encodedFilename);
|
|
9
13
|
const stringifiedEndpoint = typeof endpoint === 'string' ? endpoint : endpoint?.toString();
|
|
10
14
|
return `${stringifiedEndpoint}/${bucket}/${fileKey}`;
|
|
11
15
|
}
|
package/dist/generateURL.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/generateURL.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\n\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\n\ninterface GenerateURLArgs {\n bucket: string\n collectionPrefix?: string\n endpoint?: AWS.S3ClientConfig['endpoint']\n filename: string\n prefix: string\n useCompositePrefixes?: boolean\n}\n\nexport function generateURL({\n bucket,\n collectionPrefix = '',\n endpoint,\n filename,\n prefix,\n useCompositePrefixes = false,\n}: GenerateURLArgs): string {\n const fileKey = getFileKey({\n collectionPrefix,\n docPrefix: prefix,\n filename
|
|
1
|
+
{"version":3,"sources":["../src/generateURL.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\n\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\nimport path from 'path'\n\ninterface GenerateURLArgs {\n bucket: string\n collectionPrefix?: string\n endpoint?: AWS.S3ClientConfig['endpoint']\n filename: string\n prefix: string\n useCompositePrefixes?: boolean\n}\n\nexport function generateURL({\n bucket,\n collectionPrefix = '',\n endpoint,\n filename,\n prefix,\n useCompositePrefixes = false,\n}: GenerateURLArgs): string {\n const { fileKey: rawFileKey } = getFileKey({\n collectionPrefix,\n docPrefix: prefix,\n filename,\n useCompositePrefixes,\n })\n const dir = path.posix.dirname(rawFileKey)\n const encodedFilename = encodeURIComponent(path.posix.basename(rawFileKey))\n const fileKey = dir === '.' ? encodedFilename : path.posix.join(dir, encodedFilename)\n\n const stringifiedEndpoint = typeof endpoint === 'string' ? endpoint : endpoint?.toString()\n return `${stringifiedEndpoint}/${bucket}/${fileKey}`\n}\n"],"names":["getFileKey","path","generateURL","bucket","collectionPrefix","endpoint","filename","prefix","useCompositePrefixes","fileKey","rawFileKey","docPrefix","dir","posix","dirname","encodedFilename","encodeURIComponent","basename","join","stringifiedEndpoint","toString"],"mappings":"AAEA,SAASA,UAAU,QAAQ,6CAA4C;AACvE,OAAOC,UAAU,OAAM;AAWvB,OAAO,SAASC,YAAY,EAC1BC,MAAM,EACNC,mBAAmB,EAAE,EACrBC,QAAQ,EACRC,QAAQ,EACRC,MAAM,EACNC,uBAAuB,KAAK,EACZ;IAChB,MAAM,EAAEC,SAASC,UAAU,EAAE,GAAGV,WAAW;QACzCI;QACAO,WAAWJ;QACXD;QACAE;IACF;IACA,MAAMI,MAAMX,KAAKY,KAAK,CAACC,OAAO,CAACJ;IAC/B,MAAMK,kBAAkBC,mBAAmBf,KAAKY,KAAK,CAACI,QAAQ,CAACP;IAC/D,MAAMD,UAAUG,QAAQ,MAAMG,kBAAkBd,KAAKY,KAAK,CAACK,IAAI,CAACN,KAAKG;IAErE,MAAMI,sBAAsB,OAAOd,aAAa,WAAWA,WAAWA,UAAUe;IAChF,OAAO,GAAGD,oBAAoB,CAAC,EAAEhB,OAAO,CAAC,EAAEM,SAAS;AACtD"}
|
package/dist/getFile.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"getFile.d.ts","sourceRoot":"","sources":["../src/getFile.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAC9C,OAAO,KAAK,EAAE,gBAAgB,EAAE,cAAc,EAAE,MAAM,SAAS,CAAA;
|
|
1
|
+
{"version":3,"file":"getFile.d.ts","sourceRoot":"","sources":["../src/getFile.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAC9C,OAAO,KAAK,EAAE,gBAAgB,EAAE,cAAc,EAAE,MAAM,SAAS,CAAA;AAW/D,MAAM,MAAM,qBAAqB,GAC7B;IACE,oBAAoB;IACpB,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,kBAAkB,CAAC,CAAC,IAAI,EAAE;QACxB,UAAU,EAAE,gBAAgB,CAAA;QAC5B,QAAQ,EAAE,MAAM,CAAA;QAChB,GAAG,EAAE,cAAc,CAAA;KACpB,GAAG,OAAO,GAAG,OAAO,CAAC,OAAO,CAAC,CAAA;CAC/B,GACD,OAAO,CAAA;AAEX,UAAU,WAAW;IACnB,MAAM,EAAE,MAAM,CAAA;IACd,MAAM,EAAE,GAAG,CAAC,EAAE,CAAA;IACd,mBAAmB,CAAC,EAAE,OAAO,CAAA;IAC7B,UAAU,EAAE,gBAAgB,CAAA;IAC5B,gBAAgB,CAAC,EAAE,MAAM,CAAA;IACzB,QAAQ,EAAE,MAAM,CAAA;IAChB,eAAe,CAAC,EAAE,OAAO,CAAA;IACzB,gBAAgB,CAAC,EAAE,MAAM,CAAA;IACzB,GAAG,EAAE,cAAc,CAAA;IACnB,eAAe,EAAE,qBAAqB,CAAA;IACtC,oBAAoB,CAAC,EAAE,OAAO,CAAA;CAC/B;AA8BD,wBAAsB,OAAO,CAAC,EAC5B,MAAM,EACN,MAAM,EACN,mBAAmB,EACnB,UAAU,EACV,gBAAqB,EACrB,QAAQ,EACR,eAAe,EACf,gBAAgB,EAChB,GAAG,EACH,eAAe,EACf,oBAA4B,GAC7B,EAAE,WAAW,GAAG,OAAO,CAAC,QAAQ,CAAC,CA8JjC"}
|
package/dist/getFile.js
CHANGED
|
@@ -2,7 +2,6 @@ import { GetObjectCommand } from '@aws-sdk/client-s3';
|
|
|
2
2
|
import { getSignedUrl } from '@aws-sdk/s3-request-presigner';
|
|
3
3
|
import { getFilePrefix as getDocPrefix, getFileKey } from '@payloadcms/plugin-cloud-storage/utilities';
|
|
4
4
|
import { getRangeRequestInfo } from 'payload/internal';
|
|
5
|
-
import { sanitizeFilename } from 'payload/shared';
|
|
6
5
|
const isNodeReadableStream = (body)=>{
|
|
7
6
|
return typeof body === 'object' && body !== null && 'pipe' in body && typeof body.pipe === 'function' && 'destroy' in body && typeof body.destroy === 'function';
|
|
8
7
|
};
|
|
@@ -35,10 +34,10 @@ export async function getFile({ bucket, client, clientUploadContext, collection,
|
|
|
35
34
|
prefixQueryParam,
|
|
36
35
|
req
|
|
37
36
|
});
|
|
38
|
-
const key = getFileKey({
|
|
37
|
+
const { fileKey: key } = getFileKey({
|
|
39
38
|
collectionPrefix,
|
|
40
39
|
docPrefix,
|
|
41
|
-
filename
|
|
40
|
+
filename,
|
|
42
41
|
useCompositePrefixes
|
|
43
42
|
});
|
|
44
43
|
if (signedDownloads && !clientUploadContext) {
|
package/dist/getFile.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/getFile.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\nimport type { CollectionConfig, PayloadRequest } from 'payload'\nimport type { Readable } from 'stream'\n\nimport { GetObjectCommand } from '@aws-sdk/client-s3'\nimport { getSignedUrl } from '@aws-sdk/s3-request-presigner'\nimport {\n getFilePrefix as getDocPrefix,\n getFileKey,\n} from '@payloadcms/plugin-cloud-storage/utilities'\nimport { getRangeRequestInfo } from 'payload/internal'\nimport { sanitizeFilename } from 'payload/shared'\n\nexport type SignedDownloadsConfig =\n | {\n /** @default 7200 */\n expiresIn?: number\n shouldUseSignedURL?(args: {\n collection: CollectionConfig\n filename: string\n req: PayloadRequest\n }): boolean | Promise<boolean>\n }\n | boolean\n\ninterface GetFileArgs {\n bucket: string\n client: AWS.S3\n clientUploadContext?: unknown\n collection: CollectionConfig\n collectionPrefix?: string\n filename: string\n incomingHeaders?: Headers\n prefixQueryParam?: string\n req: PayloadRequest\n signedDownloads: SignedDownloadsConfig\n useCompositePrefixes?: boolean\n}\n\nconst isNodeReadableStream = (body: AWS.GetObjectOutput['Body']): body is Readable => {\n return (\n typeof body === 'object' &&\n body !== null &&\n 'pipe' in body &&\n typeof body.pipe === 'function' &&\n 'destroy' in body &&\n typeof body.destroy === 'function'\n )\n}\n\nconst abortRequestAndDestroyStream = ({\n abortController,\n object,\n}: {\n abortController: AbortController\n object?: AWS.GetObjectOutput\n}) => {\n try {\n abortController.abort()\n } catch {\n /* noop */\n }\n if (object?.Body && isNodeReadableStream(object.Body)) {\n object.Body.destroy()\n }\n}\n\nexport async function getFile({\n bucket,\n client,\n clientUploadContext,\n collection,\n collectionPrefix = '',\n filename,\n incomingHeaders,\n prefixQueryParam,\n req,\n signedDownloads,\n useCompositePrefixes = false,\n}: GetFileArgs): Promise<Response> {\n let object: AWS.GetObjectOutput | undefined = undefined\n let streamed = false\n\n const abortController = new AbortController()\n if (req.signal) {\n req.signal.addEventListener('abort', () => {\n abortRequestAndDestroyStream({ abortController, object })\n })\n }\n\n try {\n const docPrefix = await getDocPrefix({\n clientUploadContext,\n collection,\n filename,\n prefixQueryParam,\n req,\n })\n\n const key = getFileKey({\n collectionPrefix,\n docPrefix,\n filename: sanitizeFilename(filename),\n useCompositePrefixes,\n })\n\n if (signedDownloads && !clientUploadContext) {\n let useSignedURL = true\n if (\n typeof signedDownloads === 'object' &&\n typeof signedDownloads.shouldUseSignedURL === 'function'\n ) {\n useSignedURL = await signedDownloads.shouldUseSignedURL({ collection, filename, req })\n }\n\n if (useSignedURL) {\n const command = new GetObjectCommand({ Bucket: bucket, Key: key })\n const signedUrl = await getSignedUrl(\n client,\n command,\n typeof signedDownloads === 'object' ? signedDownloads : { expiresIn: 7200 },\n )\n return Response.redirect(signedUrl, 302)\n }\n }\n\n // Get file size first for range validation and to set Content-Length header before streaming\n const headObject = await client.headObject({\n Bucket: bucket,\n Key: key,\n })\n const fileSize = headObject.ContentLength\n\n if (!fileSize) {\n return new Response('Internal Server Error', { status: 500 })\n }\n\n // Handle range request\n const rangeHeader = req.headers.get('range')\n const rangeResult = getRangeRequestInfo({ fileSize, rangeHeader })\n\n if (rangeResult.type === 'invalid') {\n return new Response(null, {\n headers: new Headers(rangeResult.headers),\n status: rangeResult.status,\n })\n }\n\n const rangeForS3 =\n rangeResult.type === 'partial'\n ? `bytes=${rangeResult.rangeStart}-${rangeResult.rangeEnd}`\n : undefined\n\n let headers = new Headers(incomingHeaders)\n\n // Add range-related headers from the result\n for (const [headerKey, value] of Object.entries(rangeResult.headers)) {\n headers.append(headerKey, value)\n }\n\n headers.append('Content-Type', String(headObject.ContentType))\n if (headObject.ETag) {\n headers.append('ETag', headObject.ETag)\n }\n\n // Add Content-Security-Policy header for SVG files to prevent executable code\n if (headObject.ContentType === 'image/svg+xml') {\n headers.append('Content-Security-Policy', \"script-src 'none'\")\n }\n\n const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match')\n const objectEtag = headObject.ETag\n\n if (\n collection.upload &&\n typeof collection.upload === 'object' &&\n typeof collection.upload.modifyResponseHeaders === 'function'\n ) {\n headers = collection.upload.modifyResponseHeaders({ headers }) || headers\n }\n\n if (etagFromHeaders && etagFromHeaders === objectEtag) {\n return new Response(null, {\n headers,\n status: 304,\n })\n }\n\n object = await client.getObject(\n {\n Bucket: bucket,\n Key: key,\n Range: rangeForS3,\n },\n { abortSignal: abortController.signal },\n )\n\n if (!object.Body) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n if (!isNodeReadableStream(object.Body)) {\n req.payload.logger.error({\n key,\n msg: 'S3 object body is not a readable stream',\n })\n return new Response('Internal Server Error', { status: 500 })\n }\n\n const stream = object.Body\n stream.on('error', (err: Error) => {\n req.payload.logger.error({\n err,\n key,\n msg: 'Error while streaming S3 object (aborting)',\n })\n abortRequestAndDestroyStream({ abortController, object })\n })\n\n streamed = true\n return new Response(stream, { headers, status: rangeResult.status })\n } catch (err) {\n if (\n err &&\n typeof err === 'object' &&\n (('name' in err && (err.name === 'NoSuchKey' || err.name === 'NotFound')) ||\n ('httpStatusCode' in err && err.httpStatusCode === 404))\n ) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n req.payload.logger.error(err)\n return new Response('Internal Server Error', { status: 500 })\n } finally {\n if (!streamed) {\n abortRequestAndDestroyStream({ abortController, object })\n }\n }\n}\n"],"names":["GetObjectCommand","getSignedUrl","getFilePrefix","getDocPrefix","getFileKey","getRangeRequestInfo","sanitizeFilename","isNodeReadableStream","body","pipe","destroy","abortRequestAndDestroyStream","abortController","object","abort","Body","getFile","bucket","client","clientUploadContext","collection","collectionPrefix","filename","incomingHeaders","prefixQueryParam","req","signedDownloads","useCompositePrefixes","undefined","streamed","AbortController","signal","addEventListener","docPrefix","key","useSignedURL","shouldUseSignedURL","command","Bucket","Key","signedUrl","expiresIn","Response","redirect","headObject","fileSize","ContentLength","status","rangeHeader","headers","get","rangeResult","type","Headers","rangeForS3","rangeStart","rangeEnd","headerKey","value","Object","entries","append","String","ContentType","ETag","etagFromHeaders","objectEtag","upload","modifyResponseHeaders","getObject","Range","abortSignal","statusText","payload","logger","error","msg","stream","on","err","name","httpStatusCode"],"mappings":"AAIA,SAASA,gBAAgB,QAAQ,qBAAoB;AACrD,SAASC,YAAY,QAAQ,gCAA+B;AAC5D,SACEC,iBAAiBC,YAAY,EAC7BC,UAAU,QACL,6CAA4C;AACnD,SAASC,mBAAmB,QAAQ,mBAAkB;AACtD,SAASC,gBAAgB,QAAQ,iBAAgB;AA4BjD,MAAMC,uBAAuB,CAACC;IAC5B,OACE,OAAOA,SAAS,YAChBA,SAAS,QACT,UAAUA,QACV,OAAOA,KAAKC,IAAI,KAAK,cACrB,aAAaD,QACb,OAAOA,KAAKE,OAAO,KAAK;AAE5B;AAEA,MAAMC,+BAA+B,CAAC,EACpCC,eAAe,EACfC,MAAM,EAIP;IACC,IAAI;QACFD,gBAAgBE,KAAK;IACvB,EAAE,OAAM;IACN,QAAQ,GACV;IACA,IAAID,QAAQE,QAAQR,qBAAqBM,OAAOE,IAAI,GAAG;QACrDF,OAAOE,IAAI,CAACL,OAAO;IACrB;AACF;AAEA,OAAO,eAAeM,QAAQ,EAC5BC,MAAM,EACNC,MAAM,EACNC,mBAAmB,EACnBC,UAAU,EACVC,mBAAmB,EAAE,EACrBC,QAAQ,EACRC,eAAe,EACfC,gBAAgB,EAChBC,GAAG,EACHC,eAAe,EACfC,uBAAuB,KAAK,EAChB;IACZ,IAAId,SAA0Ce;IAC9C,IAAIC,WAAW;IAEf,MAAMjB,kBAAkB,IAAIkB;IAC5B,IAAIL,IAAIM,MAAM,EAAE;QACdN,IAAIM,MAAM,CAACC,gBAAgB,CAAC,SAAS;YACnCrB,6BAA6B;gBAAEC;gBAAiBC;YAAO;QACzD;IACF;IAEA,IAAI;QACF,MAAMoB,YAAY,MAAM9B,aAAa;YACnCgB;YACAC;YACAE;YACAE;YACAC;QACF;QAEA,MAAMS,MAAM9B,WAAW;YACrBiB;YACAY;YACAX,UAAUhB,iBAAiBgB;YAC3BK;QACF;QAEA,IAAID,mBAAmB,CAACP,qBAAqB;YAC3C,IAAIgB,eAAe;YACnB,IACE,OAAOT,oBAAoB,YAC3B,OAAOA,gBAAgBU,kBAAkB,KAAK,YAC9C;gBACAD,eAAe,MAAMT,gBAAgBU,kBAAkB,CAAC;oBAAEhB;oBAAYE;oBAAUG;gBAAI;YACtF;YAEA,IAAIU,cAAc;gBAChB,MAAME,UAAU,IAAIrC,iBAAiB;oBAAEsC,QAAQrB;oBAAQsB,KAAKL;gBAAI;gBAChE,MAAMM,YAAY,MAAMvC,aACtBiB,QACAmB,SACA,OAAOX,oBAAoB,WAAWA,kBAAkB;oBAAEe,WAAW;gBAAK;gBAE5E,OAAOC,SAASC,QAAQ,CAACH,WAAW;YACtC;QACF;QAEA,6FAA6F;QAC7F,MAAMI,aAAa,MAAM1B,OAAO0B,UAAU,CAAC;YACzCN,QAAQrB;YACRsB,KAAKL;QACP;QACA,MAAMW,WAAWD,WAAWE,aAAa;QAEzC,IAAI,CAACD,UAAU;YACb,OAAO,IAAIH,SAAS,yBAAyB;gBAAEK,QAAQ;YAAI;QAC7D;QAEA,uBAAuB;QACvB,MAAMC,cAAcvB,IAAIwB,OAAO,CAACC,GAAG,CAAC;QACpC,MAAMC,cAAc9C,oBAAoB;YAAEwC;YAAUG;QAAY;QAEhE,IAAIG,YAAYC,IAAI,KAAK,WAAW;YAClC,OAAO,IAAIV,SAAS,MAAM;gBACxBO,SAAS,IAAII,QAAQF,YAAYF,OAAO;gBACxCF,QAAQI,YAAYJ,MAAM;YAC5B;QACF;QAEA,MAAMO,aACJH,YAAYC,IAAI,KAAK,YACjB,CAAC,MAAM,EAAED,YAAYI,UAAU,CAAC,CAAC,EAAEJ,YAAYK,QAAQ,EAAE,GACzD5B;QAEN,IAAIqB,UAAU,IAAII,QAAQ9B;QAE1B,4CAA4C;QAC5C,KAAK,MAAM,CAACkC,WAAWC,MAAM,IAAIC,OAAOC,OAAO,CAACT,YAAYF,OAAO,EAAG;YACpEA,QAAQY,MAAM,CAACJ,WAAWC;QAC5B;QAEAT,QAAQY,MAAM,CAAC,gBAAgBC,OAAOlB,WAAWmB,WAAW;QAC5D,IAAInB,WAAWoB,IAAI,EAAE;YACnBf,QAAQY,MAAM,CAAC,QAAQjB,WAAWoB,IAAI;QACxC;QAEA,8EAA8E;QAC9E,IAAIpB,WAAWmB,WAAW,KAAK,iBAAiB;YAC9Cd,QAAQY,MAAM,CAAC,2BAA2B;QAC5C;QAEA,MAAMI,kBAAkBxC,IAAIwB,OAAO,CAACC,GAAG,CAAC,WAAWzB,IAAIwB,OAAO,CAACC,GAAG,CAAC;QACnE,MAAMgB,aAAatB,WAAWoB,IAAI;QAElC,IACE5C,WAAW+C,MAAM,IACjB,OAAO/C,WAAW+C,MAAM,KAAK,YAC7B,OAAO/C,WAAW+C,MAAM,CAACC,qBAAqB,KAAK,YACnD;YACAnB,UAAU7B,WAAW+C,MAAM,CAACC,qBAAqB,CAAC;gBAAEnB;YAAQ,MAAMA;QACpE;QAEA,IAAIgB,mBAAmBA,oBAAoBC,YAAY;YACrD,OAAO,IAAIxB,SAAS,MAAM;gBACxBO;gBACAF,QAAQ;YACV;QACF;QAEAlC,SAAS,MAAMK,OAAOmD,SAAS,CAC7B;YACE/B,QAAQrB;YACRsB,KAAKL;YACLoC,OAAOhB;QACT,GACA;YAAEiB,aAAa3D,gBAAgBmB,MAAM;QAAC;QAGxC,IAAI,CAAClB,OAAOE,IAAI,EAAE;YAChB,OAAO,IAAI2B,SAAS,MAAM;gBAAEK,QAAQ;gBAAKyB,YAAY;YAAY;QACnE;QAEA,IAAI,CAACjE,qBAAqBM,OAAOE,IAAI,GAAG;YACtCU,IAAIgD,OAAO,CAACC,MAAM,CAACC,KAAK,CAAC;gBACvBzC;gBACA0C,KAAK;YACP;YACA,OAAO,IAAIlC,SAAS,yBAAyB;gBAAEK,QAAQ;YAAI;QAC7D;QAEA,MAAM8B,SAAShE,OAAOE,IAAI;QAC1B8D,OAAOC,EAAE,CAAC,SAAS,CAACC;YAClBtD,IAAIgD,OAAO,CAACC,MAAM,CAACC,KAAK,CAAC;gBACvBI;gBACA7C;gBACA0C,KAAK;YACP;YACAjE,6BAA6B;gBAAEC;gBAAiBC;YAAO;QACzD;QAEAgB,WAAW;QACX,OAAO,IAAIa,SAASmC,QAAQ;YAAE5B;YAASF,QAAQI,YAAYJ,MAAM;QAAC;IACpE,EAAE,OAAOgC,KAAK;QACZ,IACEA,OACA,OAAOA,QAAQ,YACd,CAAA,AAAC,UAAUA,OAAQA,CAAAA,IAAIC,IAAI,KAAK,eAAeD,IAAIC,IAAI,KAAK,UAAS,KACnE,oBAAoBD,OAAOA,IAAIE,cAAc,KAAK,GAAG,GACxD;YACA,OAAO,IAAIvC,SAAS,MAAM;gBAAEK,QAAQ;gBAAKyB,YAAY;YAAY;QACnE;QACA/C,IAAIgD,OAAO,CAACC,MAAM,CAACC,KAAK,CAACI;QACzB,OAAO,IAAIrC,SAAS,yBAAyB;YAAEK,QAAQ;QAAI;IAC7D,SAAU;QACR,IAAI,CAAClB,UAAU;YACblB,6BAA6B;gBAAEC;gBAAiBC;YAAO;QACzD;IACF;AACF"}
|
|
1
|
+
{"version":3,"sources":["../src/getFile.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\nimport type { CollectionConfig, PayloadRequest } from 'payload'\nimport type { Readable } from 'stream'\n\nimport { GetObjectCommand } from '@aws-sdk/client-s3'\nimport { getSignedUrl } from '@aws-sdk/s3-request-presigner'\nimport {\n getFilePrefix as getDocPrefix,\n getFileKey,\n} from '@payloadcms/plugin-cloud-storage/utilities'\nimport { getRangeRequestInfo } from 'payload/internal'\n\nexport type SignedDownloadsConfig =\n | {\n /** @default 7200 */\n expiresIn?: number\n shouldUseSignedURL?(args: {\n collection: CollectionConfig\n filename: string\n req: PayloadRequest\n }): boolean | Promise<boolean>\n }\n | boolean\n\ninterface GetFileArgs {\n bucket: string\n client: AWS.S3\n clientUploadContext?: unknown\n collection: CollectionConfig\n collectionPrefix?: string\n filename: string\n incomingHeaders?: Headers\n prefixQueryParam?: string\n req: PayloadRequest\n signedDownloads: SignedDownloadsConfig\n useCompositePrefixes?: boolean\n}\n\nconst isNodeReadableStream = (body: AWS.GetObjectOutput['Body']): body is Readable => {\n return (\n typeof body === 'object' &&\n body !== null &&\n 'pipe' in body &&\n typeof body.pipe === 'function' &&\n 'destroy' in body &&\n typeof body.destroy === 'function'\n )\n}\n\nconst abortRequestAndDestroyStream = ({\n abortController,\n object,\n}: {\n abortController: AbortController\n object?: AWS.GetObjectOutput\n}) => {\n try {\n abortController.abort()\n } catch {\n /* noop */\n }\n if (object?.Body && isNodeReadableStream(object.Body)) {\n object.Body.destroy()\n }\n}\n\nexport async function getFile({\n bucket,\n client,\n clientUploadContext,\n collection,\n collectionPrefix = '',\n filename,\n incomingHeaders,\n prefixQueryParam,\n req,\n signedDownloads,\n useCompositePrefixes = false,\n}: GetFileArgs): Promise<Response> {\n let object: AWS.GetObjectOutput | undefined = undefined\n let streamed = false\n\n const abortController = new AbortController()\n if (req.signal) {\n req.signal.addEventListener('abort', () => {\n abortRequestAndDestroyStream({ abortController, object })\n })\n }\n\n try {\n const docPrefix = await getDocPrefix({\n clientUploadContext,\n collection,\n filename,\n prefixQueryParam,\n req,\n })\n\n const { fileKey: key } = getFileKey({\n collectionPrefix,\n docPrefix,\n filename,\n useCompositePrefixes,\n })\n\n if (signedDownloads && !clientUploadContext) {\n let useSignedURL = true\n if (\n typeof signedDownloads === 'object' &&\n typeof signedDownloads.shouldUseSignedURL === 'function'\n ) {\n useSignedURL = await signedDownloads.shouldUseSignedURL({ collection, filename, req })\n }\n\n if (useSignedURL) {\n const command = new GetObjectCommand({ Bucket: bucket, Key: key })\n const signedUrl = await getSignedUrl(\n client,\n command,\n typeof signedDownloads === 'object' ? signedDownloads : { expiresIn: 7200 },\n )\n return Response.redirect(signedUrl, 302)\n }\n }\n\n // Get file size first for range validation and to set Content-Length header before streaming\n const headObject = await client.headObject({\n Bucket: bucket,\n Key: key,\n })\n const fileSize = headObject.ContentLength\n\n if (!fileSize) {\n return new Response('Internal Server Error', { status: 500 })\n }\n\n // Handle range request\n const rangeHeader = req.headers.get('range')\n const rangeResult = getRangeRequestInfo({ fileSize, rangeHeader })\n\n if (rangeResult.type === 'invalid') {\n return new Response(null, {\n headers: new Headers(rangeResult.headers),\n status: rangeResult.status,\n })\n }\n\n const rangeForS3 =\n rangeResult.type === 'partial'\n ? `bytes=${rangeResult.rangeStart}-${rangeResult.rangeEnd}`\n : undefined\n\n let headers = new Headers(incomingHeaders)\n\n // Add range-related headers from the result\n for (const [headerKey, value] of Object.entries(rangeResult.headers)) {\n headers.append(headerKey, value)\n }\n\n headers.append('Content-Type', String(headObject.ContentType))\n if (headObject.ETag) {\n headers.append('ETag', headObject.ETag)\n }\n\n // Add Content-Security-Policy header for SVG files to prevent executable code\n if (headObject.ContentType === 'image/svg+xml') {\n headers.append('Content-Security-Policy', \"script-src 'none'\")\n }\n\n const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match')\n const objectEtag = headObject.ETag\n\n if (\n collection.upload &&\n typeof collection.upload === 'object' &&\n typeof collection.upload.modifyResponseHeaders === 'function'\n ) {\n headers = collection.upload.modifyResponseHeaders({ headers }) || headers\n }\n\n if (etagFromHeaders && etagFromHeaders === objectEtag) {\n return new Response(null, {\n headers,\n status: 304,\n })\n }\n\n object = await client.getObject(\n {\n Bucket: bucket,\n Key: key,\n Range: rangeForS3,\n },\n { abortSignal: abortController.signal },\n )\n\n if (!object.Body) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n if (!isNodeReadableStream(object.Body)) {\n req.payload.logger.error({\n key,\n msg: 'S3 object body is not a readable stream',\n })\n return new Response('Internal Server Error', { status: 500 })\n }\n\n const stream = object.Body\n stream.on('error', (err: Error) => {\n req.payload.logger.error({\n err,\n key,\n msg: 'Error while streaming S3 object (aborting)',\n })\n abortRequestAndDestroyStream({ abortController, object })\n })\n\n streamed = true\n return new Response(stream, { headers, status: rangeResult.status })\n } catch (err) {\n if (\n err &&\n typeof err === 'object' &&\n (('name' in err && (err.name === 'NoSuchKey' || err.name === 'NotFound')) ||\n ('httpStatusCode' in err && err.httpStatusCode === 404))\n ) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n req.payload.logger.error(err)\n return new Response('Internal Server Error', { status: 500 })\n } finally {\n if (!streamed) {\n abortRequestAndDestroyStream({ abortController, object })\n }\n }\n}\n"],"names":["GetObjectCommand","getSignedUrl","getFilePrefix","getDocPrefix","getFileKey","getRangeRequestInfo","isNodeReadableStream","body","pipe","destroy","abortRequestAndDestroyStream","abortController","object","abort","Body","getFile","bucket","client","clientUploadContext","collection","collectionPrefix","filename","incomingHeaders","prefixQueryParam","req","signedDownloads","useCompositePrefixes","undefined","streamed","AbortController","signal","addEventListener","docPrefix","fileKey","key","useSignedURL","shouldUseSignedURL","command","Bucket","Key","signedUrl","expiresIn","Response","redirect","headObject","fileSize","ContentLength","status","rangeHeader","headers","get","rangeResult","type","Headers","rangeForS3","rangeStart","rangeEnd","headerKey","value","Object","entries","append","String","ContentType","ETag","etagFromHeaders","objectEtag","upload","modifyResponseHeaders","getObject","Range","abortSignal","statusText","payload","logger","error","msg","stream","on","err","name","httpStatusCode"],"mappings":"AAIA,SAASA,gBAAgB,QAAQ,qBAAoB;AACrD,SAASC,YAAY,QAAQ,gCAA+B;AAC5D,SACEC,iBAAiBC,YAAY,EAC7BC,UAAU,QACL,6CAA4C;AACnD,SAASC,mBAAmB,QAAQ,mBAAkB;AA4BtD,MAAMC,uBAAuB,CAACC;IAC5B,OACE,OAAOA,SAAS,YAChBA,SAAS,QACT,UAAUA,QACV,OAAOA,KAAKC,IAAI,KAAK,cACrB,aAAaD,QACb,OAAOA,KAAKE,OAAO,KAAK;AAE5B;AAEA,MAAMC,+BAA+B,CAAC,EACpCC,eAAe,EACfC,MAAM,EAIP;IACC,IAAI;QACFD,gBAAgBE,KAAK;IACvB,EAAE,OAAM;IACN,QAAQ,GACV;IACA,IAAID,QAAQE,QAAQR,qBAAqBM,OAAOE,IAAI,GAAG;QACrDF,OAAOE,IAAI,CAACL,OAAO;IACrB;AACF;AAEA,OAAO,eAAeM,QAAQ,EAC5BC,MAAM,EACNC,MAAM,EACNC,mBAAmB,EACnBC,UAAU,EACVC,mBAAmB,EAAE,EACrBC,QAAQ,EACRC,eAAe,EACfC,gBAAgB,EAChBC,GAAG,EACHC,eAAe,EACfC,uBAAuB,KAAK,EAChB;IACZ,IAAId,SAA0Ce;IAC9C,IAAIC,WAAW;IAEf,MAAMjB,kBAAkB,IAAIkB;IAC5B,IAAIL,IAAIM,MAAM,EAAE;QACdN,IAAIM,MAAM,CAACC,gBAAgB,CAAC,SAAS;YACnCrB,6BAA6B;gBAAEC;gBAAiBC;YAAO;QACzD;IACF;IAEA,IAAI;QACF,MAAMoB,YAAY,MAAM7B,aAAa;YACnCe;YACAC;YACAE;YACAE;YACAC;QACF;QAEA,MAAM,EAAES,SAASC,GAAG,EAAE,GAAG9B,WAAW;YAClCgB;YACAY;YACAX;YACAK;QACF;QAEA,IAAID,mBAAmB,CAACP,qBAAqB;YAC3C,IAAIiB,eAAe;YACnB,IACE,OAAOV,oBAAoB,YAC3B,OAAOA,gBAAgBW,kBAAkB,KAAK,YAC9C;gBACAD,eAAe,MAAMV,gBAAgBW,kBAAkB,CAAC;oBAAEjB;oBAAYE;oBAAUG;gBAAI;YACtF;YAEA,IAAIW,cAAc;gBAChB,MAAME,UAAU,IAAIrC,iBAAiB;oBAAEsC,QAAQtB;oBAAQuB,KAAKL;gBAAI;gBAChE,MAAMM,YAAY,MAAMvC,aACtBgB,QACAoB,SACA,OAAOZ,oBAAoB,WAAWA,kBAAkB;oBAAEgB,WAAW;gBAAK;gBAE5E,OAAOC,SAASC,QAAQ,CAACH,WAAW;YACtC;QACF;QAEA,6FAA6F;QAC7F,MAAMI,aAAa,MAAM3B,OAAO2B,UAAU,CAAC;YACzCN,QAAQtB;YACRuB,KAAKL;QACP;QACA,MAAMW,WAAWD,WAAWE,aAAa;QAEzC,IAAI,CAACD,UAAU;YACb,OAAO,IAAIH,SAAS,yBAAyB;gBAAEK,QAAQ;YAAI;QAC7D;QAEA,uBAAuB;QACvB,MAAMC,cAAcxB,IAAIyB,OAAO,CAACC,GAAG,CAAC;QACpC,MAAMC,cAAc9C,oBAAoB;YAAEwC;YAAUG;QAAY;QAEhE,IAAIG,YAAYC,IAAI,KAAK,WAAW;YAClC,OAAO,IAAIV,SAAS,MAAM;gBACxBO,SAAS,IAAII,QAAQF,YAAYF,OAAO;gBACxCF,QAAQI,YAAYJ,MAAM;YAC5B;QACF;QAEA,MAAMO,aACJH,YAAYC,IAAI,KAAK,YACjB,CAAC,MAAM,EAAED,YAAYI,UAAU,CAAC,CAAC,EAAEJ,YAAYK,QAAQ,EAAE,GACzD7B;QAEN,IAAIsB,UAAU,IAAII,QAAQ/B;QAE1B,4CAA4C;QAC5C,KAAK,MAAM,CAACmC,WAAWC,MAAM,IAAIC,OAAOC,OAAO,CAACT,YAAYF,OAAO,EAAG;YACpEA,QAAQY,MAAM,CAACJ,WAAWC;QAC5B;QAEAT,QAAQY,MAAM,CAAC,gBAAgBC,OAAOlB,WAAWmB,WAAW;QAC5D,IAAInB,WAAWoB,IAAI,EAAE;YACnBf,QAAQY,MAAM,CAAC,QAAQjB,WAAWoB,IAAI;QACxC;QAEA,8EAA8E;QAC9E,IAAIpB,WAAWmB,WAAW,KAAK,iBAAiB;YAC9Cd,QAAQY,MAAM,CAAC,2BAA2B;QAC5C;QAEA,MAAMI,kBAAkBzC,IAAIyB,OAAO,CAACC,GAAG,CAAC,WAAW1B,IAAIyB,OAAO,CAACC,GAAG,CAAC;QACnE,MAAMgB,aAAatB,WAAWoB,IAAI;QAElC,IACE7C,WAAWgD,MAAM,IACjB,OAAOhD,WAAWgD,MAAM,KAAK,YAC7B,OAAOhD,WAAWgD,MAAM,CAACC,qBAAqB,KAAK,YACnD;YACAnB,UAAU9B,WAAWgD,MAAM,CAACC,qBAAqB,CAAC;gBAAEnB;YAAQ,MAAMA;QACpE;QAEA,IAAIgB,mBAAmBA,oBAAoBC,YAAY;YACrD,OAAO,IAAIxB,SAAS,MAAM;gBACxBO;gBACAF,QAAQ;YACV;QACF;QAEAnC,SAAS,MAAMK,OAAOoD,SAAS,CAC7B;YACE/B,QAAQtB;YACRuB,KAAKL;YACLoC,OAAOhB;QACT,GACA;YAAEiB,aAAa5D,gBAAgBmB,MAAM;QAAC;QAGxC,IAAI,CAAClB,OAAOE,IAAI,EAAE;YAChB,OAAO,IAAI4B,SAAS,MAAM;gBAAEK,QAAQ;gBAAKyB,YAAY;YAAY;QACnE;QAEA,IAAI,CAAClE,qBAAqBM,OAAOE,IAAI,GAAG;YACtCU,IAAIiD,OAAO,CAACC,MAAM,CAACC,KAAK,CAAC;gBACvBzC;gBACA0C,KAAK;YACP;YACA,OAAO,IAAIlC,SAAS,yBAAyB;gBAAEK,QAAQ;YAAI;QAC7D;QAEA,MAAM8B,SAASjE,OAAOE,IAAI;QAC1B+D,OAAOC,EAAE,CAAC,SAAS,CAACC;YAClBvD,IAAIiD,OAAO,CAACC,MAAM,CAACC,KAAK,CAAC;gBACvBI;gBACA7C;gBACA0C,KAAK;YACP;YACAlE,6BAA6B;gBAAEC;gBAAiBC;YAAO;QACzD;QAEAgB,WAAW;QACX,OAAO,IAAIc,SAASmC,QAAQ;YAAE5B;YAASF,QAAQI,YAAYJ,MAAM;QAAC;IACpE,EAAE,OAAOgC,KAAK;QACZ,IACEA,OACA,OAAOA,QAAQ,YACd,CAAA,AAAC,UAAUA,OAAQA,CAAAA,IAAIC,IAAI,KAAK,eAAeD,IAAIC,IAAI,KAAK,UAAS,KACnE,oBAAoBD,OAAOA,IAAIE,cAAc,KAAK,GAAG,GACxD;YACA,OAAO,IAAIvC,SAAS,MAAM;gBAAEK,QAAQ;gBAAKyB,YAAY;YAAY;QACnE;QACAhD,IAAIiD,OAAO,CAACC,MAAM,CAACC,KAAK,CAACI;QACzB,OAAO,IAAIrC,SAAS,yBAAyB;YAAEK,QAAQ;QAAI;IAC7D,SAAU;QACR,IAAI,CAACnB,UAAU;YACblB,6BAA6B;gBAAEC;gBAAiBC;YAAO;QACzD;IACF;AACF"}
|
package/dist/uploadFile.js
CHANGED
|
@@ -4,9 +4,9 @@ import fs from 'fs';
|
|
|
4
4
|
const multipartThreshold = 1024 * 1024 * 50 // 50MB
|
|
5
5
|
;
|
|
6
6
|
export async function uploadFile({ acl, bucket, buffer, client, collectionPrefix = '', docPrefix, filename, mimeType, tempFilePath, useCompositePrefixes = false }) {
|
|
7
|
-
const fileKey = getFileKey({
|
|
7
|
+
const { fileKey } = getFileKey({
|
|
8
8
|
collectionPrefix,
|
|
9
|
-
docPrefix
|
|
9
|
+
docPrefix,
|
|
10
10
|
filename,
|
|
11
11
|
useCompositePrefixes
|
|
12
12
|
});
|
package/dist/uploadFile.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/uploadFile.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\n\nimport { Upload } from '@aws-sdk/lib-storage'\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\nimport fs from 'fs'\n\ninterface UploadArgs {\n acl?: 'private' | 'public-read'\n bucket: string\n buffer: Buffer\n client: AWS.S3\n collectionPrefix?: string\n docPrefix?: string\n filename: string\n mimeType: string\n tempFilePath?: string\n useCompositePrefixes?: boolean\n}\n\nconst multipartThreshold = 1024 * 1024 * 50 // 50MB\n\nexport async function uploadFile({\n acl,\n bucket,\n buffer,\n client,\n collectionPrefix = '',\n docPrefix,\n filename,\n mimeType,\n tempFilePath,\n useCompositePrefixes = false,\n}: UploadArgs): Promise<void> {\n const fileKey = getFileKey({\n collectionPrefix,\n docPrefix
|
|
1
|
+
{"version":3,"sources":["../src/uploadFile.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\n\nimport { Upload } from '@aws-sdk/lib-storage'\nimport { getFileKey } from '@payloadcms/plugin-cloud-storage/utilities'\nimport fs from 'fs'\n\ninterface UploadArgs {\n acl?: 'private' | 'public-read'\n bucket: string\n buffer: Buffer\n client: AWS.S3\n collectionPrefix?: string\n docPrefix?: string\n filename: string\n mimeType: string\n tempFilePath?: string\n useCompositePrefixes?: boolean\n}\n\nconst multipartThreshold = 1024 * 1024 * 50 // 50MB\n\nexport async function uploadFile({\n acl,\n bucket,\n buffer,\n client,\n collectionPrefix = '',\n docPrefix,\n filename,\n mimeType,\n tempFilePath,\n useCompositePrefixes = false,\n}: UploadArgs): Promise<void> {\n const { fileKey } = getFileKey({\n collectionPrefix,\n docPrefix,\n filename,\n useCompositePrefixes,\n })\n\n const fileBufferOrStream = tempFilePath ? fs.createReadStream(tempFilePath) : buffer\n\n if (buffer.length > 0 && buffer.length < multipartThreshold) {\n await client.putObject({\n ACL: acl,\n Body: fileBufferOrStream,\n Bucket: bucket,\n ContentType: mimeType,\n Key: fileKey,\n })\n\n return\n }\n\n const parallelUploadS3 = new Upload({\n client,\n params: {\n ACL: acl,\n Body: fileBufferOrStream,\n Bucket: bucket,\n ContentType: mimeType,\n Key: fileKey,\n },\n partSize: multipartThreshold,\n queueSize: 4,\n })\n\n await parallelUploadS3.done()\n}\n"],"names":["Upload","getFileKey","fs","multipartThreshold","uploadFile","acl","bucket","buffer","client","collectionPrefix","docPrefix","filename","mimeType","tempFilePath","useCompositePrefixes","fileKey","fileBufferOrStream","createReadStream","length","putObject","ACL","Body","Bucket","ContentType","Key","parallelUploadS3","params","partSize","queueSize","done"],"mappings":"AAEA,SAASA,MAAM,QAAQ,uBAAsB;AAC7C,SAASC,UAAU,QAAQ,6CAA4C;AACvE,OAAOC,QAAQ,KAAI;AAenB,MAAMC,qBAAqB,OAAO,OAAO,GAAG,OAAO;;AAEnD,OAAO,eAAeC,WAAW,EAC/BC,GAAG,EACHC,MAAM,EACNC,MAAM,EACNC,MAAM,EACNC,mBAAmB,EAAE,EACrBC,SAAS,EACTC,QAAQ,EACRC,QAAQ,EACRC,YAAY,EACZC,uBAAuB,KAAK,EACjB;IACX,MAAM,EAAEC,OAAO,EAAE,GAAGd,WAAW;QAC7BQ;QACAC;QACAC;QACAG;IACF;IAEA,MAAME,qBAAqBH,eAAeX,GAAGe,gBAAgB,CAACJ,gBAAgBN;IAE9E,IAAIA,OAAOW,MAAM,GAAG,KAAKX,OAAOW,MAAM,GAAGf,oBAAoB;QAC3D,MAAMK,OAAOW,SAAS,CAAC;YACrBC,KAAKf;YACLgB,MAAML;YACNM,QAAQhB;YACRiB,aAAaX;YACbY,KAAKT;QACP;QAEA;IACF;IAEA,MAAMU,mBAAmB,IAAIzB,OAAO;QAClCQ;QACAkB,QAAQ;YACNN,KAAKf;YACLgB,MAAML;YACNM,QAAQhB;YACRiB,aAAaX;YACbY,KAAKT;QACP;QACAY,UAAUxB;QACVyB,WAAW;IACb;IAEA,MAAMH,iBAAiBI,IAAI;AAC7B"}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@payloadcms/storage-s3",
|
|
3
|
-
"version": "3.
|
|
3
|
+
"version": "3.84.0-canary.1",
|
|
4
4
|
"description": "Payload storage adapter for Amazon S3",
|
|
5
5
|
"homepage": "https://payloadcms.com",
|
|
6
6
|
"repository": {
|
|
@@ -40,14 +40,14 @@
|
|
|
40
40
|
"@aws-sdk/client-s3": "^3.614.0",
|
|
41
41
|
"@aws-sdk/lib-storage": "^3.614.0",
|
|
42
42
|
"@aws-sdk/s3-request-presigner": "^3.614.0",
|
|
43
|
-
"@payloadcms/plugin-cloud-storage": "3.
|
|
43
|
+
"@payloadcms/plugin-cloud-storage": "3.84.0-canary.1"
|
|
44
44
|
},
|
|
45
45
|
"devDependencies": {
|
|
46
46
|
"@smithy/node-http-handler": "4.0.3",
|
|
47
|
-
"payload": "3.
|
|
47
|
+
"payload": "3.84.0-canary.1"
|
|
48
48
|
},
|
|
49
49
|
"peerDependencies": {
|
|
50
|
-
"payload": "3.
|
|
50
|
+
"payload": "3.84.0-canary.1"
|
|
51
51
|
},
|
|
52
52
|
"engines": {
|
|
53
53
|
"node": "^18.20.2 || >=20.9.0"
|
package/dist/handleDelete.d.ts
DELETED
|
@@ -1,9 +0,0 @@
|
|
|
1
|
-
import type * as AWS from '@aws-sdk/client-s3';
|
|
2
|
-
import type { HandleDelete } from '@payloadcms/plugin-cloud-storage/types';
|
|
3
|
-
interface Args {
|
|
4
|
-
bucket: string;
|
|
5
|
-
getStorageClient: () => AWS.S3;
|
|
6
|
-
}
|
|
7
|
-
export declare const getHandleDelete: ({ bucket, getStorageClient }: Args) => HandleDelete;
|
|
8
|
-
export {};
|
|
9
|
-
//# sourceMappingURL=handleDelete.d.ts.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"handleDelete.d.ts","sourceRoot":"","sources":["../src/handleDelete.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,wCAAwC,CAAA;AAI1E,UAAU,IAAI;IACZ,MAAM,EAAE,MAAM,CAAA;IACd,gBAAgB,EAAE,MAAM,GAAG,CAAC,EAAE,CAAA;CAC/B;AAED,eAAO,MAAM,eAAe,iCAAkC,IAAI,KAAG,YAOpE,CAAA"}
|
package/dist/handleDelete.js
DELETED
|
@@ -1,11 +0,0 @@
|
|
|
1
|
-
import path from 'path';
|
|
2
|
-
export const getHandleDelete = ({ bucket, getStorageClient })=>{
|
|
3
|
-
return async ({ doc: { prefix = '' }, filename })=>{
|
|
4
|
-
await getStorageClient().deleteObject({
|
|
5
|
-
Bucket: bucket,
|
|
6
|
-
Key: path.posix.join(prefix, filename)
|
|
7
|
-
});
|
|
8
|
-
};
|
|
9
|
-
};
|
|
10
|
-
|
|
11
|
-
//# sourceMappingURL=handleDelete.js.map
|
package/dist/handleDelete.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/handleDelete.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\nimport type { HandleDelete } from '@payloadcms/plugin-cloud-storage/types'\n\nimport path from 'path'\n\ninterface Args {\n bucket: string\n getStorageClient: () => AWS.S3\n}\n\nexport const getHandleDelete = ({ bucket, getStorageClient }: Args): HandleDelete => {\n return async ({ doc: { prefix = '' }, filename }) => {\n await getStorageClient().deleteObject({\n Bucket: bucket,\n Key: path.posix.join(prefix, filename),\n })\n }\n}\n"],"names":["path","getHandleDelete","bucket","getStorageClient","doc","prefix","filename","deleteObject","Bucket","Key","posix","join"],"mappings":"AAGA,OAAOA,UAAU,OAAM;AAOvB,OAAO,MAAMC,kBAAkB,CAAC,EAAEC,MAAM,EAAEC,gBAAgB,EAAQ;IAChE,OAAO,OAAO,EAAEC,KAAK,EAAEC,SAAS,EAAE,EAAE,EAAEC,QAAQ,EAAE;QAC9C,MAAMH,mBAAmBI,YAAY,CAAC;YACpCC,QAAQN;YACRO,KAAKT,KAAKU,KAAK,CAACC,IAAI,CAACN,QAAQC;QAC/B;IACF;AACF,EAAC"}
|
package/dist/handleUpload.d.ts
DELETED
|
@@ -1,13 +0,0 @@
|
|
|
1
|
-
import type * as AWS from '@aws-sdk/client-s3';
|
|
2
|
-
import type { HandleUpload } from '@payloadcms/plugin-cloud-storage/types';
|
|
3
|
-
import type { CollectionConfig } from 'payload';
|
|
4
|
-
interface Args {
|
|
5
|
-
acl?: 'private' | 'public-read';
|
|
6
|
-
bucket: string;
|
|
7
|
-
collection: CollectionConfig;
|
|
8
|
-
getStorageClient: () => AWS.S3;
|
|
9
|
-
prefix?: string;
|
|
10
|
-
}
|
|
11
|
-
export declare const getHandleUpload: ({ acl, bucket, getStorageClient, prefix, }: Args) => HandleUpload;
|
|
12
|
-
export {};
|
|
13
|
-
//# sourceMappingURL=handleUpload.d.ts.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"handleUpload.d.ts","sourceRoot":"","sources":["../src/handleUpload.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAC9C,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,wCAAwC,CAAA;AAC1E,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,SAAS,CAAA;AAM/C,UAAU,IAAI;IACZ,GAAG,CAAC,EAAE,SAAS,GAAG,aAAa,CAAA;IAC/B,MAAM,EAAE,MAAM,CAAA;IACd,UAAU,EAAE,gBAAgB,CAAA;IAC5B,gBAAgB,EAAE,MAAM,GAAG,CAAC,EAAE,CAAA;IAC9B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAID,eAAO,MAAM,eAAe,+CAKzB,IAAI,KAAG,YAqCT,CAAA"}
|
package/dist/handleUpload.js
DELETED
|
@@ -1,37 +0,0 @@
|
|
|
1
|
-
import { Upload } from '@aws-sdk/lib-storage';
|
|
2
|
-
import fs from 'fs';
|
|
3
|
-
import path from 'path';
|
|
4
|
-
const multipartThreshold = 1024 * 1024 * 50 // 50MB
|
|
5
|
-
;
|
|
6
|
-
export const getHandleUpload = ({ acl, bucket, getStorageClient, prefix = '' })=>{
|
|
7
|
-
return async ({ data, file })=>{
|
|
8
|
-
const fileKey = path.posix.join(data.prefix || prefix, file.filename);
|
|
9
|
-
const fileBufferOrStream = file.tempFilePath ? fs.createReadStream(file.tempFilePath) : file.buffer;
|
|
10
|
-
if (file.buffer.length > 0 && file.buffer.length < multipartThreshold) {
|
|
11
|
-
await getStorageClient().putObject({
|
|
12
|
-
ACL: acl,
|
|
13
|
-
Body: fileBufferOrStream,
|
|
14
|
-
Bucket: bucket,
|
|
15
|
-
ContentType: file.mimeType,
|
|
16
|
-
Key: fileKey
|
|
17
|
-
});
|
|
18
|
-
return data;
|
|
19
|
-
}
|
|
20
|
-
const parallelUploadS3 = new Upload({
|
|
21
|
-
client: getStorageClient(),
|
|
22
|
-
params: {
|
|
23
|
-
ACL: acl,
|
|
24
|
-
Body: fileBufferOrStream,
|
|
25
|
-
Bucket: bucket,
|
|
26
|
-
ContentType: file.mimeType,
|
|
27
|
-
Key: fileKey
|
|
28
|
-
},
|
|
29
|
-
partSize: multipartThreshold,
|
|
30
|
-
queueSize: 4
|
|
31
|
-
});
|
|
32
|
-
await parallelUploadS3.done();
|
|
33
|
-
return data;
|
|
34
|
-
};
|
|
35
|
-
};
|
|
36
|
-
|
|
37
|
-
//# sourceMappingURL=handleUpload.js.map
|
package/dist/handleUpload.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/handleUpload.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\nimport type { HandleUpload } from '@payloadcms/plugin-cloud-storage/types'\nimport type { CollectionConfig } from 'payload'\n\nimport { Upload } from '@aws-sdk/lib-storage'\nimport fs from 'fs'\nimport path from 'path'\n\ninterface Args {\n acl?: 'private' | 'public-read'\n bucket: string\n collection: CollectionConfig\n getStorageClient: () => AWS.S3\n prefix?: string\n}\n\nconst multipartThreshold = 1024 * 1024 * 50 // 50MB\n\nexport const getHandleUpload = ({\n acl,\n bucket,\n getStorageClient,\n prefix = '',\n}: Args): HandleUpload => {\n return async ({ data, file }) => {\n const fileKey = path.posix.join(data.prefix || prefix, file.filename)\n\n const fileBufferOrStream = file.tempFilePath\n ? fs.createReadStream(file.tempFilePath)\n : file.buffer\n\n if (file.buffer.length > 0 && file.buffer.length < multipartThreshold) {\n await getStorageClient().putObject({\n ACL: acl,\n Body: fileBufferOrStream,\n Bucket: bucket,\n ContentType: file.mimeType,\n Key: fileKey,\n })\n\n return data\n }\n\n const parallelUploadS3 = new Upload({\n client: getStorageClient(),\n params: {\n ACL: acl,\n Body: fileBufferOrStream,\n Bucket: bucket,\n ContentType: file.mimeType,\n Key: fileKey,\n },\n partSize: multipartThreshold,\n queueSize: 4,\n })\n\n await parallelUploadS3.done()\n\n return data\n }\n}\n"],"names":["Upload","fs","path","multipartThreshold","getHandleUpload","acl","bucket","getStorageClient","prefix","data","file","fileKey","posix","join","filename","fileBufferOrStream","tempFilePath","createReadStream","buffer","length","putObject","ACL","Body","Bucket","ContentType","mimeType","Key","parallelUploadS3","client","params","partSize","queueSize","done"],"mappings":"AAIA,SAASA,MAAM,QAAQ,uBAAsB;AAC7C,OAAOC,QAAQ,KAAI;AACnB,OAAOC,UAAU,OAAM;AAUvB,MAAMC,qBAAqB,OAAO,OAAO,GAAG,OAAO;;AAEnD,OAAO,MAAMC,kBAAkB,CAAC,EAC9BC,GAAG,EACHC,MAAM,EACNC,gBAAgB,EAChBC,SAAS,EAAE,EACN;IACL,OAAO,OAAO,EAAEC,IAAI,EAAEC,IAAI,EAAE;QAC1B,MAAMC,UAAUT,KAAKU,KAAK,CAACC,IAAI,CAACJ,KAAKD,MAAM,IAAIA,QAAQE,KAAKI,QAAQ;QAEpE,MAAMC,qBAAqBL,KAAKM,YAAY,GACxCf,GAAGgB,gBAAgB,CAACP,KAAKM,YAAY,IACrCN,KAAKQ,MAAM;QAEf,IAAIR,KAAKQ,MAAM,CAACC,MAAM,GAAG,KAAKT,KAAKQ,MAAM,CAACC,MAAM,GAAGhB,oBAAoB;YACrE,MAAMI,mBAAmBa,SAAS,CAAC;gBACjCC,KAAKhB;gBACLiB,MAAMP;gBACNQ,QAAQjB;gBACRkB,aAAad,KAAKe,QAAQ;gBAC1BC,KAAKf;YACP;YAEA,OAAOF;QACT;QAEA,MAAMkB,mBAAmB,IAAI3B,OAAO;YAClC4B,QAAQrB;YACRsB,QAAQ;gBACNR,KAAKhB;gBACLiB,MAAMP;gBACNQ,QAAQjB;gBACRkB,aAAad,KAAKe,QAAQ;gBAC1BC,KAAKf;YACP;YACAmB,UAAU3B;YACV4B,WAAW;QACb;QAEA,MAAMJ,iBAAiBK,IAAI;QAE3B,OAAOvB;IACT;AACF,EAAC"}
|
package/dist/staticHandler.d.ts
DELETED
|
@@ -1,21 +0,0 @@
|
|
|
1
|
-
import type * as AWS from '@aws-sdk/client-s3';
|
|
2
|
-
import type { StaticHandler } from '@payloadcms/plugin-cloud-storage/types';
|
|
3
|
-
import type { CollectionConfig, PayloadRequest } from 'payload';
|
|
4
|
-
export type SignedDownloadsConfig = {
|
|
5
|
-
/** @default 7200 */
|
|
6
|
-
expiresIn?: number;
|
|
7
|
-
shouldUseSignedURL?(args: {
|
|
8
|
-
collection: CollectionConfig;
|
|
9
|
-
filename: string;
|
|
10
|
-
req: PayloadRequest;
|
|
11
|
-
}): boolean | Promise<boolean>;
|
|
12
|
-
} | boolean;
|
|
13
|
-
interface Args {
|
|
14
|
-
bucket: string;
|
|
15
|
-
collection: CollectionConfig;
|
|
16
|
-
getStorageClient: () => AWS.S3;
|
|
17
|
-
signedDownloads?: SignedDownloadsConfig;
|
|
18
|
-
}
|
|
19
|
-
export declare const getHandler: ({ bucket, collection, getStorageClient, signedDownloads, }: Args) => StaticHandler;
|
|
20
|
-
export {};
|
|
21
|
-
//# sourceMappingURL=staticHandler.d.ts.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"staticHandler.d.ts","sourceRoot":"","sources":["../src/staticHandler.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,KAAK,GAAG,MAAM,oBAAoB,CAAA;AAC9C,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,wCAAwC,CAAA;AAC3E,OAAO,KAAK,EAAE,gBAAgB,EAAE,cAAc,EAAE,MAAM,SAAS,CAAA;AAU/D,MAAM,MAAM,qBAAqB,GAC7B;IACE,oBAAoB;IACpB,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,kBAAkB,CAAC,CAAC,IAAI,EAAE;QACxB,UAAU,EAAE,gBAAgB,CAAA;QAC5B,QAAQ,EAAE,MAAM,CAAA;QAChB,GAAG,EAAE,cAAc,CAAA;KACpB,GAAG,OAAO,GAAG,OAAO,CAAC,OAAO,CAAC,CAAA;CAC/B,GACD,OAAO,CAAA;AAEX,UAAU,IAAI;IACZ,MAAM,EAAE,MAAM,CAAA;IACd,UAAU,EAAE,gBAAgB,CAAA;IAC5B,gBAAgB,EAAE,MAAM,GAAG,CAAC,EAAE,CAAA;IAC9B,eAAe,CAAC,EAAE,qBAAqB,CAAA;CACxC;AA8BD,eAAO,MAAM,UAAU,+DAKpB,IAAI,KAAG,aAmJT,CAAA"}
|
package/dist/staticHandler.js
DELETED
|
@@ -1,169 +0,0 @@
|
|
|
1
|
-
import { GetObjectCommand } from '@aws-sdk/client-s3';
|
|
2
|
-
import { getSignedUrl } from '@aws-sdk/s3-request-presigner';
|
|
3
|
-
import { getFilePrefix } from '@payloadcms/plugin-cloud-storage/utilities';
|
|
4
|
-
import path from 'path';
|
|
5
|
-
import { getRangeRequestInfo } from 'payload/internal';
|
|
6
|
-
import { sanitizeFilename } from 'payload/shared';
|
|
7
|
-
const isNodeReadableStream = (body)=>{
|
|
8
|
-
return typeof body === 'object' && body !== null && 'pipe' in body && typeof body.pipe === 'function' && 'destroy' in body && typeof body.destroy === 'function';
|
|
9
|
-
};
|
|
10
|
-
const abortRequestAndDestroyStream = ({ abortController, object })=>{
|
|
11
|
-
try {
|
|
12
|
-
abortController.abort();
|
|
13
|
-
} catch {
|
|
14
|
-
/* noop */ }
|
|
15
|
-
if (object?.Body && isNodeReadableStream(object.Body)) {
|
|
16
|
-
object.Body.destroy();
|
|
17
|
-
}
|
|
18
|
-
};
|
|
19
|
-
export const getHandler = ({ bucket, collection, getStorageClient, signedDownloads })=>{
|
|
20
|
-
return async (req, { headers: incomingHeaders, params: { clientUploadContext, filename } })=>{
|
|
21
|
-
let object = undefined;
|
|
22
|
-
let streamed = false;
|
|
23
|
-
const abortController = new AbortController();
|
|
24
|
-
if (req.signal) {
|
|
25
|
-
req.signal.addEventListener('abort', ()=>{
|
|
26
|
-
abortRequestAndDestroyStream({
|
|
27
|
-
abortController,
|
|
28
|
-
object
|
|
29
|
-
});
|
|
30
|
-
});
|
|
31
|
-
}
|
|
32
|
-
try {
|
|
33
|
-
const prefix = await getFilePrefix({
|
|
34
|
-
clientUploadContext,
|
|
35
|
-
collection,
|
|
36
|
-
filename,
|
|
37
|
-
req
|
|
38
|
-
});
|
|
39
|
-
const key = path.posix.join(prefix, sanitizeFilename(filename));
|
|
40
|
-
if (signedDownloads && !clientUploadContext) {
|
|
41
|
-
let useSignedURL = true;
|
|
42
|
-
if (typeof signedDownloads === 'object' && typeof signedDownloads.shouldUseSignedURL === 'function') {
|
|
43
|
-
useSignedURL = await signedDownloads.shouldUseSignedURL({
|
|
44
|
-
collection,
|
|
45
|
-
filename,
|
|
46
|
-
req
|
|
47
|
-
});
|
|
48
|
-
}
|
|
49
|
-
if (useSignedURL) {
|
|
50
|
-
const command = new GetObjectCommand({
|
|
51
|
-
Bucket: bucket,
|
|
52
|
-
Key: key
|
|
53
|
-
});
|
|
54
|
-
const signedUrl = await getSignedUrl(getStorageClient(), command, typeof signedDownloads === 'object' ? signedDownloads : {
|
|
55
|
-
expiresIn: 7200
|
|
56
|
-
});
|
|
57
|
-
return Response.redirect(signedUrl, 302);
|
|
58
|
-
}
|
|
59
|
-
}
|
|
60
|
-
// Get file size first for range validation
|
|
61
|
-
const headObject = await getStorageClient().headObject({
|
|
62
|
-
Bucket: bucket,
|
|
63
|
-
Key: key
|
|
64
|
-
});
|
|
65
|
-
const fileSize = headObject.ContentLength;
|
|
66
|
-
if (!fileSize) {
|
|
67
|
-
return new Response('Internal Server Error', {
|
|
68
|
-
status: 500
|
|
69
|
-
});
|
|
70
|
-
}
|
|
71
|
-
// Handle range request
|
|
72
|
-
const rangeHeader = req.headers.get('range');
|
|
73
|
-
const rangeResult = getRangeRequestInfo({
|
|
74
|
-
fileSize,
|
|
75
|
-
rangeHeader
|
|
76
|
-
});
|
|
77
|
-
if (rangeResult.type === 'invalid') {
|
|
78
|
-
return new Response(null, {
|
|
79
|
-
headers: new Headers(rangeResult.headers),
|
|
80
|
-
status: rangeResult.status
|
|
81
|
-
});
|
|
82
|
-
}
|
|
83
|
-
const rangeForS3 = rangeResult.type === 'partial' ? `bytes=${rangeResult.rangeStart}-${rangeResult.rangeEnd}` : undefined;
|
|
84
|
-
object = await getStorageClient().getObject({
|
|
85
|
-
Bucket: bucket,
|
|
86
|
-
Key: key,
|
|
87
|
-
Range: rangeForS3
|
|
88
|
-
}, {
|
|
89
|
-
abortSignal: abortController.signal
|
|
90
|
-
});
|
|
91
|
-
if (!object.Body) {
|
|
92
|
-
return new Response(null, {
|
|
93
|
-
status: 404,
|
|
94
|
-
statusText: 'Not Found'
|
|
95
|
-
});
|
|
96
|
-
}
|
|
97
|
-
let headers = new Headers(incomingHeaders);
|
|
98
|
-
// Add range-related headers from the result
|
|
99
|
-
for (const [key, value] of Object.entries(rangeResult.headers)){
|
|
100
|
-
headers.append(key, value);
|
|
101
|
-
}
|
|
102
|
-
headers.append('Content-Type', String(object.ContentType));
|
|
103
|
-
headers.append('ETag', String(object.ETag));
|
|
104
|
-
// Add Content-Security-Policy header for SVG files to prevent executable code
|
|
105
|
-
if (object.ContentType === 'image/svg+xml') {
|
|
106
|
-
headers.append('Content-Security-Policy', "script-src 'none'");
|
|
107
|
-
}
|
|
108
|
-
const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match');
|
|
109
|
-
const objectEtag = object.ETag;
|
|
110
|
-
if (collection.upload && typeof collection.upload === 'object' && typeof collection.upload.modifyResponseHeaders === 'function') {
|
|
111
|
-
headers = collection.upload.modifyResponseHeaders({
|
|
112
|
-
headers
|
|
113
|
-
}) || headers;
|
|
114
|
-
}
|
|
115
|
-
if (etagFromHeaders && etagFromHeaders === objectEtag) {
|
|
116
|
-
return new Response(null, {
|
|
117
|
-
headers,
|
|
118
|
-
status: 304
|
|
119
|
-
});
|
|
120
|
-
}
|
|
121
|
-
if (!isNodeReadableStream(object.Body)) {
|
|
122
|
-
req.payload.logger.error({
|
|
123
|
-
key,
|
|
124
|
-
msg: 'S3 object body is not a readable stream'
|
|
125
|
-
});
|
|
126
|
-
return new Response('Internal Server Error', {
|
|
127
|
-
status: 500
|
|
128
|
-
});
|
|
129
|
-
}
|
|
130
|
-
const stream = object.Body;
|
|
131
|
-
stream.on('error', (err)=>{
|
|
132
|
-
req.payload.logger.error({
|
|
133
|
-
err,
|
|
134
|
-
key,
|
|
135
|
-
msg: 'Error while streaming S3 object (aborting)'
|
|
136
|
-
});
|
|
137
|
-
abortRequestAndDestroyStream({
|
|
138
|
-
abortController,
|
|
139
|
-
object
|
|
140
|
-
});
|
|
141
|
-
});
|
|
142
|
-
streamed = true;
|
|
143
|
-
return new Response(stream, {
|
|
144
|
-
headers,
|
|
145
|
-
status: rangeResult.status
|
|
146
|
-
});
|
|
147
|
-
} catch (err) {
|
|
148
|
-
if (err && typeof err === 'object' && ('name' in err && (err.name === 'NoSuchKey' || err.name === 'NotFound') || 'httpStatusCode' in err && err.httpStatusCode === 404)) {
|
|
149
|
-
return new Response(null, {
|
|
150
|
-
status: 404,
|
|
151
|
-
statusText: 'Not Found'
|
|
152
|
-
});
|
|
153
|
-
}
|
|
154
|
-
req.payload.logger.error(err);
|
|
155
|
-
return new Response('Internal Server Error', {
|
|
156
|
-
status: 500
|
|
157
|
-
});
|
|
158
|
-
} finally{
|
|
159
|
-
if (!streamed) {
|
|
160
|
-
abortRequestAndDestroyStream({
|
|
161
|
-
abortController,
|
|
162
|
-
object
|
|
163
|
-
});
|
|
164
|
-
}
|
|
165
|
-
}
|
|
166
|
-
};
|
|
167
|
-
};
|
|
168
|
-
|
|
169
|
-
//# sourceMappingURL=staticHandler.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/staticHandler.ts"],"sourcesContent":["import type * as AWS from '@aws-sdk/client-s3'\nimport type { StaticHandler } from '@payloadcms/plugin-cloud-storage/types'\nimport type { CollectionConfig, PayloadRequest } from 'payload'\nimport type { Readable } from 'stream'\n\nimport { GetObjectCommand } from '@aws-sdk/client-s3'\nimport { getSignedUrl } from '@aws-sdk/s3-request-presigner'\nimport { getFilePrefix } from '@payloadcms/plugin-cloud-storage/utilities'\nimport path from 'path'\nimport { getRangeRequestInfo } from 'payload/internal'\nimport { sanitizeFilename } from 'payload/shared'\n\nexport type SignedDownloadsConfig =\n | {\n /** @default 7200 */\n expiresIn?: number\n shouldUseSignedURL?(args: {\n collection: CollectionConfig\n filename: string\n req: PayloadRequest\n }): boolean | Promise<boolean>\n }\n | boolean\n\ninterface Args {\n bucket: string\n collection: CollectionConfig\n getStorageClient: () => AWS.S3\n signedDownloads?: SignedDownloadsConfig\n}\n\nconst isNodeReadableStream = (body: AWS.GetObjectOutput['Body']): body is Readable => {\n return (\n typeof body === 'object' &&\n body !== null &&\n 'pipe' in body &&\n typeof body.pipe === 'function' &&\n 'destroy' in body &&\n typeof body.destroy === 'function'\n )\n}\n\nconst abortRequestAndDestroyStream = ({\n abortController,\n object,\n}: {\n abortController: AbortController\n object?: AWS.GetObjectOutput\n}) => {\n try {\n abortController.abort()\n } catch {\n /* noop */\n }\n if (object?.Body && isNodeReadableStream(object.Body)) {\n object.Body.destroy()\n }\n}\n\nexport const getHandler = ({\n bucket,\n collection,\n getStorageClient,\n signedDownloads,\n}: Args): StaticHandler => {\n return async (req, { headers: incomingHeaders, params: { clientUploadContext, filename } }) => {\n let object: AWS.GetObjectOutput | undefined = undefined\n let streamed = false\n\n const abortController = new AbortController()\n if (req.signal) {\n req.signal.addEventListener('abort', () => {\n abortRequestAndDestroyStream({ abortController, object })\n })\n }\n\n try {\n const prefix = await getFilePrefix({ clientUploadContext, collection, filename, req })\n\n const key = path.posix.join(prefix, sanitizeFilename(filename))\n\n if (signedDownloads && !clientUploadContext) {\n let useSignedURL = true\n if (\n typeof signedDownloads === 'object' &&\n typeof signedDownloads.shouldUseSignedURL === 'function'\n ) {\n useSignedURL = await signedDownloads.shouldUseSignedURL({ collection, filename, req })\n }\n\n if (useSignedURL) {\n const command = new GetObjectCommand({ Bucket: bucket, Key: key })\n const signedUrl = await getSignedUrl(\n getStorageClient(),\n command,\n typeof signedDownloads === 'object' ? signedDownloads : { expiresIn: 7200 },\n )\n return Response.redirect(signedUrl, 302)\n }\n }\n\n // Get file size first for range validation\n const headObject = await getStorageClient().headObject({\n Bucket: bucket,\n Key: key,\n })\n const fileSize = headObject.ContentLength\n\n if (!fileSize) {\n return new Response('Internal Server Error', { status: 500 })\n }\n\n // Handle range request\n const rangeHeader = req.headers.get('range')\n const rangeResult = getRangeRequestInfo({ fileSize, rangeHeader })\n\n if (rangeResult.type === 'invalid') {\n return new Response(null, {\n headers: new Headers(rangeResult.headers),\n status: rangeResult.status,\n })\n }\n\n const rangeForS3 =\n rangeResult.type === 'partial'\n ? `bytes=${rangeResult.rangeStart}-${rangeResult.rangeEnd}`\n : undefined\n\n object = await getStorageClient().getObject(\n {\n Bucket: bucket,\n Key: key,\n Range: rangeForS3,\n },\n { abortSignal: abortController.signal },\n )\n\n if (!object.Body) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n let headers = new Headers(incomingHeaders)\n\n // Add range-related headers from the result\n for (const [key, value] of Object.entries(rangeResult.headers)) {\n headers.append(key, value)\n }\n\n headers.append('Content-Type', String(object.ContentType))\n headers.append('ETag', String(object.ETag))\n\n // Add Content-Security-Policy header for SVG files to prevent executable code\n if (object.ContentType === 'image/svg+xml') {\n headers.append('Content-Security-Policy', \"script-src 'none'\")\n }\n\n const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match')\n const objectEtag = object.ETag\n\n if (\n collection.upload &&\n typeof collection.upload === 'object' &&\n typeof collection.upload.modifyResponseHeaders === 'function'\n ) {\n headers = collection.upload.modifyResponseHeaders({ headers }) || headers\n }\n\n if (etagFromHeaders && etagFromHeaders === objectEtag) {\n return new Response(null, {\n headers,\n status: 304,\n })\n }\n\n if (!isNodeReadableStream(object.Body)) {\n req.payload.logger.error({\n key,\n msg: 'S3 object body is not a readable stream',\n })\n return new Response('Internal Server Error', { status: 500 })\n }\n\n const stream = object.Body\n stream.on('error', (err: Error) => {\n req.payload.logger.error({\n err,\n key,\n msg: 'Error while streaming S3 object (aborting)',\n })\n abortRequestAndDestroyStream({ abortController, object })\n })\n\n streamed = true\n return new Response(stream, { headers, status: rangeResult.status })\n } catch (err) {\n if (\n err &&\n typeof err === 'object' &&\n (('name' in err && (err.name === 'NoSuchKey' || err.name === 'NotFound')) ||\n ('httpStatusCode' in err && err.httpStatusCode === 404))\n ) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n req.payload.logger.error(err)\n return new Response('Internal Server Error', { status: 500 })\n } finally {\n if (!streamed) {\n abortRequestAndDestroyStream({ abortController, object })\n }\n }\n }\n}\n"],"names":["GetObjectCommand","getSignedUrl","getFilePrefix","path","getRangeRequestInfo","sanitizeFilename","isNodeReadableStream","body","pipe","destroy","abortRequestAndDestroyStream","abortController","object","abort","Body","getHandler","bucket","collection","getStorageClient","signedDownloads","req","headers","incomingHeaders","params","clientUploadContext","filename","undefined","streamed","AbortController","signal","addEventListener","prefix","key","posix","join","useSignedURL","shouldUseSignedURL","command","Bucket","Key","signedUrl","expiresIn","Response","redirect","headObject","fileSize","ContentLength","status","rangeHeader","get","rangeResult","type","Headers","rangeForS3","rangeStart","rangeEnd","getObject","Range","abortSignal","statusText","value","Object","entries","append","String","ContentType","ETag","etagFromHeaders","objectEtag","upload","modifyResponseHeaders","payload","logger","error","msg","stream","on","err","name","httpStatusCode"],"mappings":"AAKA,SAASA,gBAAgB,QAAQ,qBAAoB;AACrD,SAASC,YAAY,QAAQ,gCAA+B;AAC5D,SAASC,aAAa,QAAQ,6CAA4C;AAC1E,OAAOC,UAAU,OAAM;AACvB,SAASC,mBAAmB,QAAQ,mBAAkB;AACtD,SAASC,gBAAgB,QAAQ,iBAAgB;AAqBjD,MAAMC,uBAAuB,CAACC;IAC5B,OACE,OAAOA,SAAS,YAChBA,SAAS,QACT,UAAUA,QACV,OAAOA,KAAKC,IAAI,KAAK,cACrB,aAAaD,QACb,OAAOA,KAAKE,OAAO,KAAK;AAE5B;AAEA,MAAMC,+BAA+B,CAAC,EACpCC,eAAe,EACfC,MAAM,EAIP;IACC,IAAI;QACFD,gBAAgBE,KAAK;IACvB,EAAE,OAAM;IACN,QAAQ,GACV;IACA,IAAID,QAAQE,QAAQR,qBAAqBM,OAAOE,IAAI,GAAG;QACrDF,OAAOE,IAAI,CAACL,OAAO;IACrB;AACF;AAEA,OAAO,MAAMM,aAAa,CAAC,EACzBC,MAAM,EACNC,UAAU,EACVC,gBAAgB,EAChBC,eAAe,EACV;IACL,OAAO,OAAOC,KAAK,EAAEC,SAASC,eAAe,EAAEC,QAAQ,EAAEC,mBAAmB,EAAEC,QAAQ,EAAE,EAAE;QACxF,IAAIb,SAA0Cc;QAC9C,IAAIC,WAAW;QAEf,MAAMhB,kBAAkB,IAAIiB;QAC5B,IAAIR,IAAIS,MAAM,EAAE;YACdT,IAAIS,MAAM,CAACC,gBAAgB,CAAC,SAAS;gBACnCpB,6BAA6B;oBAAEC;oBAAiBC;gBAAO;YACzD;QACF;QAEA,IAAI;YACF,MAAMmB,SAAS,MAAM7B,cAAc;gBAAEsB;gBAAqBP;gBAAYQ;gBAAUL;YAAI;YAEpF,MAAMY,MAAM7B,KAAK8B,KAAK,CAACC,IAAI,CAACH,QAAQ1B,iBAAiBoB;YAErD,IAAIN,mBAAmB,CAACK,qBAAqB;gBAC3C,IAAIW,eAAe;gBACnB,IACE,OAAOhB,oBAAoB,YAC3B,OAAOA,gBAAgBiB,kBAAkB,KAAK,YAC9C;oBACAD,eAAe,MAAMhB,gBAAgBiB,kBAAkB,CAAC;wBAAEnB;wBAAYQ;wBAAUL;oBAAI;gBACtF;gBAEA,IAAIe,cAAc;oBAChB,MAAME,UAAU,IAAIrC,iBAAiB;wBAAEsC,QAAQtB;wBAAQuB,KAAKP;oBAAI;oBAChE,MAAMQ,YAAY,MAAMvC,aACtBiB,oBACAmB,SACA,OAAOlB,oBAAoB,WAAWA,kBAAkB;wBAAEsB,WAAW;oBAAK;oBAE5E,OAAOC,SAASC,QAAQ,CAACH,WAAW;gBACtC;YACF;YAEA,2CAA2C;YAC3C,MAAMI,aAAa,MAAM1B,mBAAmB0B,UAAU,CAAC;gBACrDN,QAAQtB;gBACRuB,KAAKP;YACP;YACA,MAAMa,WAAWD,WAAWE,aAAa;YAEzC,IAAI,CAACD,UAAU;gBACb,OAAO,IAAIH,SAAS,yBAAyB;oBAAEK,QAAQ;gBAAI;YAC7D;YAEA,uBAAuB;YACvB,MAAMC,cAAc5B,IAAIC,OAAO,CAAC4B,GAAG,CAAC;YACpC,MAAMC,cAAc9C,oBAAoB;gBAAEyC;gBAAUG;YAAY;YAEhE,IAAIE,YAAYC,IAAI,KAAK,WAAW;gBAClC,OAAO,IAAIT,SAAS,MAAM;oBACxBrB,SAAS,IAAI+B,QAAQF,YAAY7B,OAAO;oBACxC0B,QAAQG,YAAYH,MAAM;gBAC5B;YACF;YAEA,MAAMM,aACJH,YAAYC,IAAI,KAAK,YACjB,CAAC,MAAM,EAAED,YAAYI,UAAU,CAAC,CAAC,EAAEJ,YAAYK,QAAQ,EAAE,GACzD7B;YAENd,SAAS,MAAMM,mBAAmBsC,SAAS,CACzC;gBACElB,QAAQtB;gBACRuB,KAAKP;gBACLyB,OAAOJ;YACT,GACA;gBAAEK,aAAa/C,gBAAgBkB,MAAM;YAAC;YAGxC,IAAI,CAACjB,OAAOE,IAAI,EAAE;gBAChB,OAAO,IAAI4B,SAAS,MAAM;oBAAEK,QAAQ;oBAAKY,YAAY;gBAAY;YACnE;YAEA,IAAItC,UAAU,IAAI+B,QAAQ9B;YAE1B,4CAA4C;YAC5C,KAAK,MAAM,CAACU,KAAK4B,MAAM,IAAIC,OAAOC,OAAO,CAACZ,YAAY7B,OAAO,EAAG;gBAC9DA,QAAQ0C,MAAM,CAAC/B,KAAK4B;YACtB;YAEAvC,QAAQ0C,MAAM,CAAC,gBAAgBC,OAAOpD,OAAOqD,WAAW;YACxD5C,QAAQ0C,MAAM,CAAC,QAAQC,OAAOpD,OAAOsD,IAAI;YAEzC,8EAA8E;YAC9E,IAAItD,OAAOqD,WAAW,KAAK,iBAAiB;gBAC1C5C,QAAQ0C,MAAM,CAAC,2BAA2B;YAC5C;YAEA,MAAMI,kBAAkB/C,IAAIC,OAAO,CAAC4B,GAAG,CAAC,WAAW7B,IAAIC,OAAO,CAAC4B,GAAG,CAAC;YACnE,MAAMmB,aAAaxD,OAAOsD,IAAI;YAE9B,IACEjD,WAAWoD,MAAM,IACjB,OAAOpD,WAAWoD,MAAM,KAAK,YAC7B,OAAOpD,WAAWoD,MAAM,CAACC,qBAAqB,KAAK,YACnD;gBACAjD,UAAUJ,WAAWoD,MAAM,CAACC,qBAAqB,CAAC;oBAAEjD;gBAAQ,MAAMA;YACpE;YAEA,IAAI8C,mBAAmBA,oBAAoBC,YAAY;gBACrD,OAAO,IAAI1B,SAAS,MAAM;oBACxBrB;oBACA0B,QAAQ;gBACV;YACF;YAEA,IAAI,CAACzC,qBAAqBM,OAAOE,IAAI,GAAG;gBACtCM,IAAImD,OAAO,CAACC,MAAM,CAACC,KAAK,CAAC;oBACvBzC;oBACA0C,KAAK;gBACP;gBACA,OAAO,IAAIhC,SAAS,yBAAyB;oBAAEK,QAAQ;gBAAI;YAC7D;YAEA,MAAM4B,SAAS/D,OAAOE,IAAI;YAC1B6D,OAAOC,EAAE,CAAC,SAAS,CAACC;gBAClBzD,IAAImD,OAAO,CAACC,MAAM,CAACC,KAAK,CAAC;oBACvBI;oBACA7C;oBACA0C,KAAK;gBACP;gBACAhE,6BAA6B;oBAAEC;oBAAiBC;gBAAO;YACzD;YAEAe,WAAW;YACX,OAAO,IAAIe,SAASiC,QAAQ;gBAAEtD;gBAAS0B,QAAQG,YAAYH,MAAM;YAAC;QACpE,EAAE,OAAO8B,KAAK;YACZ,IACEA,OACA,OAAOA,QAAQ,YACd,CAAA,AAAC,UAAUA,OAAQA,CAAAA,IAAIC,IAAI,KAAK,eAAeD,IAAIC,IAAI,KAAK,UAAS,KACnE,oBAAoBD,OAAOA,IAAIE,cAAc,KAAK,GAAG,GACxD;gBACA,OAAO,IAAIrC,SAAS,MAAM;oBAAEK,QAAQ;oBAAKY,YAAY;gBAAY;YACnE;YACAvC,IAAImD,OAAO,CAACC,MAAM,CAACC,KAAK,CAACI;YACzB,OAAO,IAAInC,SAAS,yBAAyB;gBAAEK,QAAQ;YAAI;QAC7D,SAAU;YACR,IAAI,CAACpB,UAAU;gBACbjB,6BAA6B;oBAAEC;oBAAiBC;gBAAO;YACzD;QACF;IACF;AACF,EAAC"}
|