@payloadcms/storage-r2 3.74.0-internal.bc57c0d → 3.75.0-internal.880861f

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,10 @@
1
+ import type { R2StorageClientUploadHandlerParams } from '../types.js';
2
+ export declare const R2ClientUploadHandler: ({ children, collectionSlug, enabled, extra, prefix, serverHandlerPath, }: {
3
+ children: import("react").ReactNode;
4
+ collectionSlug: import("packages/payload/src/index.js").UploadCollectionSlug;
5
+ enabled?: boolean;
6
+ extra: R2StorageClientUploadHandlerParams;
7
+ prefix?: string;
8
+ serverHandlerPath: `/${string}`;
9
+ }) => import("react").JSX.Element;
10
+ //# sourceMappingURL=R2ClientUploadHandler.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"R2ClientUploadHandler.d.ts","sourceRoot":"","sources":["../../src/client/R2ClientUploadHandler.ts"],"names":[],"mappings":"AAKA,OAAO,KAAK,EAGV,kCAAkC,EAGnC,MAAM,aAAa,CAAA;AAEpB,eAAO,MAAM,qBAAqB;;;;;;;aAe5B,OAAK,aAqDT,CAAA"}
@@ -0,0 +1,69 @@
1
+ 'use client';
2
+ import { createClientUploadHandler } from '@payloadcms/plugin-cloud-storage/client';
3
+ import { formatAdminURL } from 'payload/shared';
4
+ export const R2ClientUploadHandler = createClientUploadHandler({
5
+ handler: async ({ apiRoute, collectionSlug, extra: { chunkSize = 5 * 1024 * 1024, prefix = '' }, file, serverHandlerPath, serverURL })=>{
6
+ const params = {
7
+ collection: collectionSlug,
8
+ fileName: file.name,
9
+ fileType: file.type
10
+ };
11
+ const baseURL = formatAdminURL({
12
+ apiRoute,
13
+ path: serverHandlerPath,
14
+ serverURL
15
+ });
16
+ const endpoint = `${baseURL}?${String(new URLSearchParams(params))}`;
17
+ const multipart = await fetch(endpoint, {
18
+ method: 'POST'
19
+ });
20
+ if (!multipart.ok) {
21
+ throw new Error('Failed to initialize multipart upload');
22
+ }
23
+ const multipartUpload = await multipart.json();
24
+ const multipartUploadedParts = [];
25
+ params.multipartId = multipartUpload.uploadId;
26
+ params.multipartKey = multipartUpload.key;
27
+ const partTotal = Math.ceil(file.size / chunkSize);
28
+ for(let part = 1; part <= partTotal; part++){
29
+ const bytesEnd = Math.min(part * chunkSize, file.size);
30
+ const bytesStart = (part - 1) * chunkSize;
31
+ params.multipartNumber = String(part);
32
+ const body = file.slice(bytesStart, bytesEnd);
33
+ const headers = {
34
+ 'Content-Length': String(body.size),
35
+ 'Content-Type': 'application/octet-stream'
36
+ };
37
+ const uploaded = await fetch(endpoint, {
38
+ body,
39
+ headers,
40
+ method: 'POST'
41
+ });
42
+ if (!uploaded.ok) {
43
+ throw new Error(`Failed to upload part ${part} / ${partTotal}`);
44
+ }
45
+ multipartUploadedParts.push(await uploaded.json());
46
+ if (part === partTotal) {
47
+ delete params.multipartNumber;
48
+ const body = JSON.stringify(multipartUploadedParts);
49
+ const headers = {
50
+ 'Content-Type': 'application/json'
51
+ };
52
+ const complete = await fetch(endpoint, {
53
+ body,
54
+ headers,
55
+ method: 'POST'
56
+ });
57
+ if (!complete.ok) {
58
+ throw new Error(`Failed to complete multipart upload`);
59
+ }
60
+ const key = await complete.text();
61
+ return {
62
+ key
63
+ };
64
+ }
65
+ }
66
+ }
67
+ });
68
+
69
+ //# sourceMappingURL=R2ClientUploadHandler.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../src/client/R2ClientUploadHandler.ts"],"sourcesContent":["'use client'\n\nimport { createClientUploadHandler } from '@payloadcms/plugin-cloud-storage/client'\nimport { formatAdminURL } from 'payload/shared'\n\nimport type {\n R2MultipartUpload,\n R2StorageClientUploadContext,\n R2StorageClientUploadHandlerParams,\n R2StorageMultipartUploadHandlerParams,\n R2UploadedPart,\n} from '../types.js'\n\nexport const R2ClientUploadHandler = createClientUploadHandler<R2StorageClientUploadHandlerParams>({\n handler: async ({\n apiRoute,\n collectionSlug,\n extra: { chunkSize = 5 * 1024 * 1024, prefix = '' },\n file,\n serverHandlerPath,\n serverURL,\n }): Promise<R2StorageClientUploadContext | undefined> => {\n const params: R2StorageMultipartUploadHandlerParams = {\n collection: collectionSlug,\n fileName: file.name,\n fileType: file.type,\n }\n const baseURL = formatAdminURL({\n apiRoute,\n path: serverHandlerPath,\n serverURL,\n })\n\n const endpoint = `${baseURL}?${String(new URLSearchParams(params))}`\n\n const multipart = await fetch(endpoint, { method: 'POST' })\n if (!multipart.ok) {\n throw new Error('Failed to initialize multipart upload')\n }\n\n const multipartUpload = (await multipart.json()) as Pick<R2MultipartUpload, 'key' | 'uploadId'>\n const multipartUploadedParts: R2UploadedPart[] = []\n\n params.multipartId = multipartUpload.uploadId\n params.multipartKey = multipartUpload.key\n\n const partTotal = Math.ceil(file.size / chunkSize)\n\n for (let part = 1; part <= partTotal; part++) {\n const bytesEnd = Math.min(part * chunkSize, file.size)\n const bytesStart = (part - 1) * chunkSize\n\n params.multipartNumber = String(part)\n\n const body = file.slice(bytesStart, bytesEnd)\n const headers = {\n 'Content-Length': String(body.size),\n 'Content-Type': 'application/octet-stream',\n }\n const uploaded = await fetch(endpoint, { body, headers, method: 'POST' })\n if (!uploaded.ok) {\n throw new Error(`Failed to upload part ${part} / ${partTotal}`)\n }\n\n multipartUploadedParts.push((await uploaded.json()) as R2UploadedPart)\n\n if (part === partTotal) {\n delete params.multipartNumber\n\n const body = JSON.stringify(multipartUploadedParts)\n const headers = { 'Content-Type': 'application/json' }\n const complete = await fetch(endpoint, { body, headers, method: 'POST' })\n if (!complete.ok) {\n throw new Error(`Failed to complete multipart upload`)\n }\n\n const key = await complete.text()\n return { key }\n }\n }\n },\n})\n"],"names":["createClientUploadHandler","formatAdminURL","R2ClientUploadHandler","handler","apiRoute","collectionSlug","extra","chunkSize","prefix","file","serverHandlerPath","serverURL","params","collection","fileName","name","fileType","type","baseURL","path","endpoint","String","URLSearchParams","multipart","fetch","method","ok","Error","multipartUpload","json","multipartUploadedParts","multipartId","uploadId","multipartKey","key","partTotal","Math","ceil","size","part","bytesEnd","min","bytesStart","multipartNumber","body","slice","headers","uploaded","push","JSON","stringify","complete","text"],"mappings":"AAAA;AAEA,SAASA,yBAAyB,QAAQ,0CAAyC;AACnF,SAASC,cAAc,QAAQ,iBAAgB;AAU/C,OAAO,MAAMC,wBAAwBF,0BAA8D;IACjGG,SAAS,OAAO,EACdC,QAAQ,EACRC,cAAc,EACdC,OAAO,EAAEC,YAAY,IAAI,OAAO,IAAI,EAAEC,SAAS,EAAE,EAAE,EACnDC,IAAI,EACJC,iBAAiB,EACjBC,SAAS,EACV;QACC,MAAMC,SAAgD;YACpDC,YAAYR;YACZS,UAAUL,KAAKM,IAAI;YACnBC,UAAUP,KAAKQ,IAAI;QACrB;QACA,MAAMC,UAAUjB,eAAe;YAC7BG;YACAe,MAAMT;YACNC;QACF;QAEA,MAAMS,WAAW,GAAGF,QAAQ,CAAC,EAAEG,OAAO,IAAIC,gBAAgBV,UAAU;QAEpE,MAAMW,YAAY,MAAMC,MAAMJ,UAAU;YAAEK,QAAQ;QAAO;QACzD,IAAI,CAACF,UAAUG,EAAE,EAAE;YACjB,MAAM,IAAIC,MAAM;QAClB;QAEA,MAAMC,kBAAmB,MAAML,UAAUM,IAAI;QAC7C,MAAMC,yBAA2C,EAAE;QAEnDlB,OAAOmB,WAAW,GAAGH,gBAAgBI,QAAQ;QAC7CpB,OAAOqB,YAAY,GAAGL,gBAAgBM,GAAG;QAEzC,MAAMC,YAAYC,KAAKC,IAAI,CAAC5B,KAAK6B,IAAI,GAAG/B;QAExC,IAAK,IAAIgC,OAAO,GAAGA,QAAQJ,WAAWI,OAAQ;YAC5C,MAAMC,WAAWJ,KAAKK,GAAG,CAACF,OAAOhC,WAAWE,KAAK6B,IAAI;YACrD,MAAMI,aAAa,AAACH,CAAAA,OAAO,CAAA,IAAKhC;YAEhCK,OAAO+B,eAAe,GAAGtB,OAAOkB;YAEhC,MAAMK,OAAOnC,KAAKoC,KAAK,CAACH,YAAYF;YACpC,MAAMM,UAAU;gBACd,kBAAkBzB,OAAOuB,KAAKN,IAAI;gBAClC,gBAAgB;YAClB;YACA,MAAMS,WAAW,MAAMvB,MAAMJ,UAAU;gBAAEwB;gBAAME;gBAASrB,QAAQ;YAAO;YACvE,IAAI,CAACsB,SAASrB,EAAE,EAAE;gBAChB,MAAM,IAAIC,MAAM,CAAC,sBAAsB,EAAEY,KAAK,GAAG,EAAEJ,WAAW;YAChE;YAEAL,uBAAuBkB,IAAI,CAAE,MAAMD,SAASlB,IAAI;YAEhD,IAAIU,SAASJ,WAAW;gBACtB,OAAOvB,OAAO+B,eAAe;gBAE7B,MAAMC,OAAOK,KAAKC,SAAS,CAACpB;gBAC5B,MAAMgB,UAAU;oBAAE,gBAAgB;gBAAmB;gBACrD,MAAMK,WAAW,MAAM3B,MAAMJ,UAAU;oBAAEwB;oBAAME;oBAASrB,QAAQ;gBAAO;gBACvE,IAAI,CAAC0B,SAASzB,EAAE,EAAE;oBAChB,MAAM,IAAIC,MAAM,CAAC,mCAAmC,CAAC;gBACvD;gBAEA,MAAMO,MAAM,MAAMiB,SAASC,IAAI;gBAC/B,OAAO;oBAAElB;gBAAI;YACf;QACF;IACF;AACF,GAAE"}
@@ -0,0 +1,2 @@
1
+ export { R2ClientUploadHandler } from '../client/R2ClientUploadHandler.js';
2
+ //# sourceMappingURL=client.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"client.d.ts","sourceRoot":"","sources":["../../src/exports/client.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,qBAAqB,EAAE,MAAM,oCAAoC,CAAA"}
@@ -0,0 +1,3 @@
1
+ export { R2ClientUploadHandler } from '../client/R2ClientUploadHandler.js';
2
+
3
+ //# sourceMappingURL=client.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../src/exports/client.ts"],"sourcesContent":["export { R2ClientUploadHandler } from '../client/R2ClientUploadHandler.js'\n"],"names":["R2ClientUploadHandler"],"mappings":"AAAA,SAASA,qBAAqB,QAAQ,qCAAoC"}
@@ -0,0 +1,12 @@
1
+ import type { ClientUploadsAccess } from '@payloadcms/plugin-cloud-storage/types';
2
+ import type { PayloadHandler } from 'payload';
3
+ import type { R2StorageOptions } from './index.js';
4
+ import type { R2Bucket } from './types.js';
5
+ type Args = {
6
+ access?: ClientUploadsAccess;
7
+ bucket: R2Bucket;
8
+ collections: R2StorageOptions['collections'];
9
+ };
10
+ export declare const getHandleMultiPartUpload: ({ access, bucket, collections }: Args) => PayloadHandler;
11
+ export {};
12
+ //# sourceMappingURL=handleMultiPartUpload.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"handleMultiPartUpload.d.ts","sourceRoot":"","sources":["../src/handleMultiPartUpload.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AACjF,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,SAAS,CAAA;AAK7C,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAA;AAClD,OAAO,KAAK,EAAE,QAAQ,EAAyC,MAAM,YAAY,CAAA;AAEjF,KAAK,IAAI,GAAG;IACV,MAAM,CAAC,EAAE,mBAAmB,CAAA;IAC5B,MAAM,EAAE,QAAQ,CAAA;IAChB,WAAW,EAAE,gBAAgB,CAAC,aAAa,CAAC,CAAA;CAC7C,CAAA;AAGD,eAAO,MAAM,wBAAwB,oCACD,IAAI,KAAG,cAuExC,CAAA"}
@@ -0,0 +1,72 @@
1
+ import path from 'path';
2
+ import { APIError, Forbidden } from 'payload';
3
+ // Adapted from https://developers.cloudflare.com/r2/api/workers/workers-multipart-usage/
4
+ export const getHandleMultiPartUpload = ({ access, bucket, collections })=>async (req)=>{
5
+ const params = Object.fromEntries(req.searchParams);
6
+ const collectionSlug = params.collection;
7
+ const filetype = params.fileType;
8
+ const collectionConfig = collections[collectionSlug];
9
+ if (!collectionConfig) {
10
+ throw new APIError(`Collection ${collectionSlug} was not found in R2 Storage options`);
11
+ }
12
+ // Check custom access if provided, otherwise check collection's create access
13
+ if (access) {
14
+ if (!await access({
15
+ collectionSlug,
16
+ req
17
+ })) {
18
+ throw new Forbidden(req.t);
19
+ }
20
+ } else {
21
+ // Use the collection's create access control
22
+ const collection = req.payload.collections[collectionSlug];
23
+ if (!collection) {
24
+ throw new APIError(`Collection ${collectionSlug} not found`);
25
+ }
26
+ const createAccess = collection.config.access?.create;
27
+ if (createAccess) {
28
+ const hasAccess = await createAccess({
29
+ req
30
+ });
31
+ if (!hasAccess) {
32
+ throw new Forbidden(req.t);
33
+ }
34
+ } else if (!req.user) {
35
+ // No custom access and no user - deny by default
36
+ throw new Forbidden(req.t);
37
+ }
38
+ }
39
+ const prefix = typeof collectionConfig === 'object' && collectionConfig.prefix || '';
40
+ const fileKey = path.posix.join(prefix, params.fileName);
41
+ const multipartId = params.multipartId;
42
+ const multipartKey = params.multipartKey;
43
+ const multipartNumber = parseInt(params.multipartNumber || '');
44
+ if (multipartId && multipartKey) {
45
+ const multipartUpload = bucket.resumeMultipartUpload(multipartKey, multipartId);
46
+ const request = req;
47
+ if (isNaN(multipartNumber)) {
48
+ // Upload complete
49
+ const object = await multipartUpload.complete(await request.json());
50
+ return new Response(object.key, {
51
+ status: 200
52
+ });
53
+ } else {
54
+ // Upload part
55
+ const uploadedPart = await multipartUpload.uploadPart(multipartNumber, await request.arrayBuffer());
56
+ return Response.json(uploadedPart);
57
+ }
58
+ } else {
59
+ // Create multipart upload
60
+ const multipartUpload = await bucket.createMultipartUpload(fileKey, {
61
+ httpMetadata: {
62
+ contentType: filetype
63
+ }
64
+ });
65
+ return Response.json({
66
+ key: multipartUpload.key,
67
+ uploadId: multipartUpload.uploadId
68
+ });
69
+ }
70
+ };
71
+
72
+ //# sourceMappingURL=handleMultiPartUpload.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/handleMultiPartUpload.ts"],"sourcesContent":["import type { ClientUploadsAccess } from '@payloadcms/plugin-cloud-storage/types'\nimport type { PayloadHandler } from 'payload'\n\nimport path from 'path'\nimport { APIError, Forbidden } from 'payload'\n\nimport type { R2StorageOptions } from './index.js'\nimport type { R2Bucket, R2StorageMultipartUploadHandlerParams } from './types.js'\n\ntype Args = {\n access?: ClientUploadsAccess\n bucket: R2Bucket\n collections: R2StorageOptions['collections']\n}\n\n// Adapted from https://developers.cloudflare.com/r2/api/workers/workers-multipart-usage/\nexport const getHandleMultiPartUpload =\n ({ access, bucket, collections }: Args): PayloadHandler =>\n async (req) => {\n const params = Object.fromEntries(req.searchParams) as R2StorageMultipartUploadHandlerParams\n const collectionSlug = params.collection\n const filetype = params.fileType\n\n const collectionConfig = collections[collectionSlug]\n if (!collectionConfig) {\n throw new APIError(`Collection ${collectionSlug} was not found in R2 Storage options`)\n }\n\n // Check custom access if provided, otherwise check collection's create access\n if (access) {\n if (!(await access({ collectionSlug, req }))) {\n throw new Forbidden(req.t)\n }\n } else {\n // Use the collection's create access control\n const collection = req.payload.collections[collectionSlug]\n if (!collection) {\n throw new APIError(`Collection ${collectionSlug} not found`)\n }\n\n const createAccess = collection.config.access?.create\n if (createAccess) {\n const hasAccess = await createAccess({ req })\n if (!hasAccess) {\n throw new Forbidden(req.t)\n }\n } else if (!req.user) {\n // No custom access and no user - deny by default\n throw new Forbidden(req.t)\n }\n }\n\n const prefix = (typeof collectionConfig === 'object' && collectionConfig.prefix) || ''\n const fileKey = path.posix.join(prefix, params.fileName)\n\n const multipartId = params.multipartId\n const multipartKey = params.multipartKey\n const multipartNumber = parseInt(params.multipartNumber || '')\n\n if (multipartId && multipartKey) {\n const multipartUpload = bucket.resumeMultipartUpload(multipartKey, multipartId)\n const request = req as Request\n\n if (isNaN(multipartNumber)) {\n // Upload complete\n const object = await multipartUpload.complete((await request.json()) as any)\n return new Response(object.key, { status: 200 })\n } else {\n // Upload part\n const uploadedPart = await multipartUpload.uploadPart(\n multipartNumber,\n await request.arrayBuffer(),\n )\n return Response.json(uploadedPart)\n }\n } else {\n // Create multipart upload\n const multipartUpload = await bucket.createMultipartUpload(fileKey, {\n httpMetadata: {\n contentType: filetype,\n },\n })\n\n return Response.json({\n key: multipartUpload.key,\n uploadId: multipartUpload.uploadId,\n })\n }\n }\n"],"names":["path","APIError","Forbidden","getHandleMultiPartUpload","access","bucket","collections","req","params","Object","fromEntries","searchParams","collectionSlug","collection","filetype","fileType","collectionConfig","t","payload","createAccess","config","create","hasAccess","user","prefix","fileKey","posix","join","fileName","multipartId","multipartKey","multipartNumber","parseInt","multipartUpload","resumeMultipartUpload","request","isNaN","object","complete","json","Response","key","status","uploadedPart","uploadPart","arrayBuffer","createMultipartUpload","httpMetadata","contentType","uploadId"],"mappings":"AAGA,OAAOA,UAAU,OAAM;AACvB,SAASC,QAAQ,EAAEC,SAAS,QAAQ,UAAS;AAW7C,yFAAyF;AACzF,OAAO,MAAMC,2BACX,CAAC,EAAEC,MAAM,EAAEC,MAAM,EAAEC,WAAW,EAAQ,GACtC,OAAOC;QACL,MAAMC,SAASC,OAAOC,WAAW,CAACH,IAAII,YAAY;QAClD,MAAMC,iBAAiBJ,OAAOK,UAAU;QACxC,MAAMC,WAAWN,OAAOO,QAAQ;QAEhC,MAAMC,mBAAmBV,WAAW,CAACM,eAAe;QACpD,IAAI,CAACI,kBAAkB;YACrB,MAAM,IAAIf,SAAS,CAAC,WAAW,EAAEW,eAAe,oCAAoC,CAAC;QACvF;QAEA,8EAA8E;QAC9E,IAAIR,QAAQ;YACV,IAAI,CAAE,MAAMA,OAAO;gBAAEQ;gBAAgBL;YAAI,IAAK;gBAC5C,MAAM,IAAIL,UAAUK,IAAIU,CAAC;YAC3B;QACF,OAAO;YACL,6CAA6C;YAC7C,MAAMJ,aAAaN,IAAIW,OAAO,CAACZ,WAAW,CAACM,eAAe;YAC1D,IAAI,CAACC,YAAY;gBACf,MAAM,IAAIZ,SAAS,CAAC,WAAW,EAAEW,eAAe,UAAU,CAAC;YAC7D;YAEA,MAAMO,eAAeN,WAAWO,MAAM,CAAChB,MAAM,EAAEiB;YAC/C,IAAIF,cAAc;gBAChB,MAAMG,YAAY,MAAMH,aAAa;oBAAEZ;gBAAI;gBAC3C,IAAI,CAACe,WAAW;oBACd,MAAM,IAAIpB,UAAUK,IAAIU,CAAC;gBAC3B;YACF,OAAO,IAAI,CAACV,IAAIgB,IAAI,EAAE;gBACpB,iDAAiD;gBACjD,MAAM,IAAIrB,UAAUK,IAAIU,CAAC;YAC3B;QACF;QAEA,MAAMO,SAAS,AAAC,OAAOR,qBAAqB,YAAYA,iBAAiBQ,MAAM,IAAK;QACpF,MAAMC,UAAUzB,KAAK0B,KAAK,CAACC,IAAI,CAACH,QAAQhB,OAAOoB,QAAQ;QAEvD,MAAMC,cAAcrB,OAAOqB,WAAW;QACtC,MAAMC,eAAetB,OAAOsB,YAAY;QACxC,MAAMC,kBAAkBC,SAASxB,OAAOuB,eAAe,IAAI;QAE3D,IAAIF,eAAeC,cAAc;YAC/B,MAAMG,kBAAkB5B,OAAO6B,qBAAqB,CAACJ,cAAcD;YACnE,MAAMM,UAAU5B;YAEhB,IAAI6B,MAAML,kBAAkB;gBAC1B,kBAAkB;gBAClB,MAAMM,SAAS,MAAMJ,gBAAgBK,QAAQ,CAAE,MAAMH,QAAQI,IAAI;gBACjE,OAAO,IAAIC,SAASH,OAAOI,GAAG,EAAE;oBAAEC,QAAQ;gBAAI;YAChD,OAAO;gBACL,cAAc;gBACd,MAAMC,eAAe,MAAMV,gBAAgBW,UAAU,CACnDb,iBACA,MAAMI,QAAQU,WAAW;gBAE3B,OAAOL,SAASD,IAAI,CAACI;YACvB;QACF,OAAO;YACL,0BAA0B;YAC1B,MAAMV,kBAAkB,MAAM5B,OAAOyC,qBAAqB,CAACrB,SAAS;gBAClEsB,cAAc;oBACZC,aAAalC;gBACf;YACF;YAEA,OAAO0B,SAASD,IAAI,CAAC;gBACnBE,KAAKR,gBAAgBQ,GAAG;gBACxBQ,UAAUhB,gBAAgBgB,QAAQ;YACpC;QACF;IACF,EAAC"}
@@ -1 +1 @@
1
- {"version":3,"file":"handleUpload.d.ts","sourceRoot":"","sources":["../src/handleUpload.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,wCAAwC,CAAA;AAC1E,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,SAAS,CAAA;AAI/C,OAAO,KAAK,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAA;AAE1C,UAAU,IAAI;IACZ,MAAM,EAAE,QAAQ,CAAA;IAChB,UAAU,EAAE,gBAAgB,CAAA;IAC5B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAED,eAAO,MAAM,eAAe,uBAA6B,IAAI,KAAG,YAU/D,CAAA"}
1
+ {"version":3,"file":"handleUpload.d.ts","sourceRoot":"","sources":["../src/handleUpload.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,YAAY,EAAE,MAAM,wCAAwC,CAAA;AAC1E,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,SAAS,CAAA;AAI/C,OAAO,KAAK,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAA;AAE1C,UAAU,IAAI;IACZ,MAAM,EAAE,QAAQ,CAAA;IAChB,UAAU,EAAE,gBAAgB,CAAA;IAC5B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAED,eAAO,MAAM,eAAe,uBAA6B,IAAI,KAAG,YAQ/D,CAAA"}
@@ -10,7 +10,6 @@ export const getHandleUpload = ({ bucket, prefix = '' })=>{
10
10
  contentType: file.mimeType
11
11
  }
12
12
  });
13
- return data;
14
13
  };
15
14
  };
16
15
 
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/handleUpload.ts"],"sourcesContent":["import type { HandleUpload } from '@payloadcms/plugin-cloud-storage/types'\nimport type { CollectionConfig } from 'payload'\n\nimport path from 'path'\n\nimport type { R2Bucket } from './types.js'\n\ninterface Args {\n bucket: R2Bucket\n collection: CollectionConfig\n prefix?: string\n}\n\nexport const getHandleUpload = ({ bucket, prefix = '' }: Args): HandleUpload => {\n return async ({ data, file }) => {\n // Read more: https://github.com/cloudflare/workers-sdk/issues/6047#issuecomment-2691217843\n const buffer = process.env.NODE_ENV === 'development' ? new Blob([file.buffer]) : file.buffer\n await bucket.put(path.posix.join(data.prefix || prefix, file.filename), buffer, {\n httpMetadata: { contentType: file.mimeType },\n })\n\n return data\n }\n}\n"],"names":["path","getHandleUpload","bucket","prefix","data","file","buffer","process","env","NODE_ENV","Blob","put","posix","join","filename","httpMetadata","contentType","mimeType"],"mappings":"AAGA,OAAOA,UAAU,OAAM;AAUvB,OAAO,MAAMC,kBAAkB,CAAC,EAAEC,MAAM,EAAEC,SAAS,EAAE,EAAQ;IAC3D,OAAO,OAAO,EAAEC,IAAI,EAAEC,IAAI,EAAE;QAC1B,2FAA2F;QAC3F,MAAMC,SAASC,QAAQC,GAAG,CAACC,QAAQ,KAAK,gBAAgB,IAAIC,KAAK;YAACL,KAAKC,MAAM;SAAC,IAAID,KAAKC,MAAM;QAC7F,MAAMJ,OAAOS,GAAG,CAACX,KAAKY,KAAK,CAACC,IAAI,CAACT,KAAKD,MAAM,IAAIA,QAAQE,KAAKS,QAAQ,GAAGR,QAAQ;YAC9ES,cAAc;gBAAEC,aAAaX,KAAKY,QAAQ;YAAC;QAC7C;QAEA,OAAOb;IACT;AACF,EAAC"}
1
+ {"version":3,"sources":["../src/handleUpload.ts"],"sourcesContent":["import type { HandleUpload } from '@payloadcms/plugin-cloud-storage/types'\nimport type { CollectionConfig } from 'payload'\n\nimport path from 'path'\n\nimport type { R2Bucket } from './types.js'\n\ninterface Args {\n bucket: R2Bucket\n collection: CollectionConfig\n prefix?: string\n}\n\nexport const getHandleUpload = ({ bucket, prefix = '' }: Args): HandleUpload => {\n return async ({ data, file }) => {\n // Read more: https://github.com/cloudflare/workers-sdk/issues/6047#issuecomment-2691217843\n const buffer = process.env.NODE_ENV === 'development' ? new Blob([file.buffer]) : file.buffer\n await bucket.put(path.posix.join(data.prefix || prefix, file.filename), buffer, {\n httpMetadata: { contentType: file.mimeType },\n })\n }\n}\n"],"names":["path","getHandleUpload","bucket","prefix","data","file","buffer","process","env","NODE_ENV","Blob","put","posix","join","filename","httpMetadata","contentType","mimeType"],"mappings":"AAGA,OAAOA,UAAU,OAAM;AAUvB,OAAO,MAAMC,kBAAkB,CAAC,EAAEC,MAAM,EAAEC,SAAS,EAAE,EAAQ;IAC3D,OAAO,OAAO,EAAEC,IAAI,EAAEC,IAAI,EAAE;QAC1B,2FAA2F;QAC3F,MAAMC,SAASC,QAAQC,GAAG,CAACC,QAAQ,KAAK,gBAAgB,IAAIC,KAAK;YAACL,KAAKC,MAAM;SAAC,IAAID,KAAKC,MAAM;QAC7F,MAAMJ,OAAOS,GAAG,CAACX,KAAKY,KAAK,CAACC,IAAI,CAACT,KAAKD,MAAM,IAAIA,QAAQE,KAAKS,QAAQ,GAAGR,QAAQ;YAC9ES,cAAc;gBAAEC,aAAaX,KAAKY,QAAQ;YAAC;QAC7C;IACF;AACF,EAAC"}
package/dist/index.d.ts CHANGED
@@ -1,4 +1,4 @@
1
- import type { CollectionOptions } from '@payloadcms/plugin-cloud-storage/types';
1
+ import type { ClientUploadsConfig, CollectionOptions } from '@payloadcms/plugin-cloud-storage/types';
2
2
  import type { Plugin, UploadCollectionSlug } from 'payload';
3
3
  import type { R2Bucket } from './types.js';
4
4
  export interface R2StorageOptions {
@@ -13,6 +13,10 @@ export interface R2StorageOptions {
13
13
  */
14
14
  alwaysInsertFields?: boolean;
15
15
  bucket: R2Bucket;
16
+ /**
17
+ * Do uploads directly on the client, to bypass limits on Cloudflare/Vercel.
18
+ */
19
+ clientUploads?: ClientUploadsConfig;
16
20
  /**
17
21
  * Collection options to apply the R2 adapter to.
18
22
  */
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAGV,iBAAiB,EAElB,MAAM,wCAAwC,CAAA;AAC/C,OAAO,KAAK,EAAU,MAAM,EAAE,oBAAoB,EAAE,MAAM,SAAS,CAAA;AAInE,OAAO,KAAK,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAA;AAM1C,MAAM,WAAW,gBAAgB;IAC/B;;;;;;;;OAQG;IACH,kBAAkB,CAAC,EAAE,OAAO,CAAA;IAE5B,MAAM,EAAE,QAAQ,CAAA;IAChB;;OAEG;IACH,WAAW,EAAE,OAAO,CAAC,MAAM,CAAC,oBAAoB,EAAE,IAAI,CAAC,iBAAiB,EAAE,SAAS,CAAC,GAAG,IAAI,CAAC,CAAC,CAAA;IAC7F,OAAO,CAAC,EAAE,OAAO,CAAA;CAClB;AAED,KAAK,eAAe,GAAG,CAAC,aAAa,EAAE,gBAAgB,KAAK,MAAM,CAAA;AAElE,eAAO,MAAM,SAAS,EAAE,eA+CrB,CAAA"}
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAEV,mBAAmB,EAEnB,iBAAiB,EAElB,MAAM,wCAAwC,CAAA;AAC/C,OAAO,KAAK,EAAU,MAAM,EAAE,oBAAoB,EAAE,MAAM,SAAS,CAAA;AAKnE,OAAO,KAAK,EAAE,QAAQ,EAAsC,MAAM,YAAY,CAAA;AAO9E,MAAM,WAAW,gBAAgB;IAC/B;;;;;;;;OAQG;IACH,kBAAkB,CAAC,EAAE,OAAO,CAAA;IAE5B,MAAM,EAAE,QAAQ,CAAA;IAChB;;OAEG;IACH,aAAa,CAAC,EAAE,mBAAmB,CAAA;IACnC;;OAEG;IACH,WAAW,EAAE,OAAO,CAAC,MAAM,CAAC,oBAAoB,EAAE,IAAI,CAAC,iBAAiB,EAAE,SAAS,CAAC,GAAG,IAAI,CAAC,CAAC,CAAA;IAC7F,OAAO,CAAC,EAAE,OAAO,CAAA;CAClB;AAED,KAAK,eAAe,GAAG,CAAC,aAAa,EAAE,gBAAgB,KAAK,MAAM,CAAA;AAElE,eAAO,MAAM,SAAS,EAAE,eAsErB,CAAA"}
package/dist/index.js CHANGED
@@ -1,10 +1,27 @@
1
1
  import { cloudStoragePlugin } from '@payloadcms/plugin-cloud-storage';
2
+ import { initClientUploads } from '@payloadcms/plugin-cloud-storage/utilities';
2
3
  import { getHandleDelete } from './handleDelete.js';
4
+ import { getHandleMultiPartUpload } from './handleMultiPartUpload.js';
3
5
  import { getHandleUpload } from './handleUpload.js';
4
6
  import { getHandler } from './staticHandler.js';
5
7
  export const r2Storage = (r2StorageOptions)=>(incomingConfig)=>{
6
8
  const adapter = r2StorageInternal(r2StorageOptions);
7
9
  const isPluginDisabled = r2StorageOptions.enabled === false;
10
+ initClientUploads({
11
+ clientHandler: '@payloadcms/storage-r2/client#R2ClientUploadHandler',
12
+ collections: r2StorageOptions.collections,
13
+ config: incomingConfig,
14
+ enabled: !isPluginDisabled && Boolean(r2StorageOptions.clientUploads),
15
+ extraClientHandlerProps: (collection)=>({
16
+ prefix: typeof collection === 'object' && collection.prefix && `${collection.prefix}/` || ''
17
+ }),
18
+ serverHandler: getHandleMultiPartUpload({
19
+ access: typeof r2StorageOptions.clientUploads === 'object' ? r2StorageOptions.clientUploads.access : undefined,
20
+ bucket: r2StorageOptions.bucket,
21
+ collections: r2StorageOptions.collections
22
+ }),
23
+ serverHandlerPath: '/storage-r2-multi-part-upload'
24
+ });
8
25
  if (isPluginDisabled) {
9
26
  return incomingConfig;
10
27
  }
@@ -37,10 +54,11 @@ export const r2Storage = (r2StorageOptions)=>(incomingConfig)=>{
37
54
  collections: collectionsWithAdapter
38
55
  })(config);
39
56
  };
40
- function r2StorageInternal({ bucket }) {
57
+ function r2StorageInternal({ bucket, clientUploads }) {
41
58
  return ({ collection, prefix })=>{
42
59
  return {
43
60
  name: 'r2',
61
+ clientUploads,
44
62
  handleDelete: getHandleDelete({
45
63
  bucket
46
64
  }),
package/dist/index.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/index.ts"],"sourcesContent":["import type {\n Adapter,\n PluginOptions as CloudStoragePluginOptions,\n CollectionOptions,\n GeneratedAdapter,\n} from '@payloadcms/plugin-cloud-storage/types'\nimport type { Config, Plugin, UploadCollectionSlug } from 'payload'\n\nimport { cloudStoragePlugin } from '@payloadcms/plugin-cloud-storage'\n\nimport type { R2Bucket } from './types.js'\n\nimport { getHandleDelete } from './handleDelete.js'\nimport { getHandleUpload } from './handleUpload.js'\nimport { getHandler } from './staticHandler.js'\n\nexport interface R2StorageOptions {\n /**\n * When enabled, fields (like the prefix field) will always be inserted into\n * the collection schema regardless of whether the plugin is enabled. This\n * ensures a consistent schema across all environments.\n *\n * This will be enabled by default in Payload v4.\n *\n * @default false\n */\n alwaysInsertFields?: boolean\n\n bucket: R2Bucket\n /**\n * Collection options to apply the R2 adapter to.\n */\n collections: Partial<Record<UploadCollectionSlug, Omit<CollectionOptions, 'adapter'> | true>>\n enabled?: boolean\n}\n\ntype R2StoragePlugin = (r2StorageArgs: R2StorageOptions) => Plugin\n\nexport const r2Storage: R2StoragePlugin =\n (r2StorageOptions) =>\n (incomingConfig: Config): Config => {\n const adapter = r2StorageInternal(r2StorageOptions)\n\n const isPluginDisabled = r2StorageOptions.enabled === false\n\n if (isPluginDisabled) {\n return incomingConfig\n }\n\n // Add adapter to each collection option object\n const collectionsWithAdapter: CloudStoragePluginOptions['collections'] = Object.entries(\n r2StorageOptions.collections,\n ).reduce(\n (acc, [slug, collOptions]) => ({\n ...acc,\n [slug]: {\n ...(collOptions === true ? {} : collOptions),\n adapter,\n },\n }),\n {} as Record<string, CollectionOptions>,\n )\n\n // Set disableLocalStorage: true for collections specified in the plugin options\n const config = {\n ...incomingConfig,\n collections: (incomingConfig.collections || []).map((collection) => {\n if (!collectionsWithAdapter[collection.slug]) {\n return collection\n }\n\n return {\n ...collection,\n upload: {\n ...(typeof collection.upload === 'object' ? collection.upload : {}),\n disableLocalStorage: true,\n },\n }\n }),\n }\n\n return cloudStoragePlugin({\n alwaysInsertFields: r2StorageOptions.alwaysInsertFields,\n collections: collectionsWithAdapter,\n })(config)\n }\n\nfunction r2StorageInternal({ bucket }: R2StorageOptions): Adapter {\n return ({ collection, prefix }): GeneratedAdapter => {\n return {\n name: 'r2',\n handleDelete: getHandleDelete({ bucket }),\n handleUpload: getHandleUpload({\n bucket,\n collection,\n prefix,\n }),\n staticHandler: getHandler({ bucket, collection, prefix }),\n }\n }\n}\n"],"names":["cloudStoragePlugin","getHandleDelete","getHandleUpload","getHandler","r2Storage","r2StorageOptions","incomingConfig","adapter","r2StorageInternal","isPluginDisabled","enabled","collectionsWithAdapter","Object","entries","collections","reduce","acc","slug","collOptions","config","map","collection","upload","disableLocalStorage","alwaysInsertFields","bucket","prefix","name","handleDelete","handleUpload","staticHandler"],"mappings":"AAQA,SAASA,kBAAkB,QAAQ,mCAAkC;AAIrE,SAASC,eAAe,QAAQ,oBAAmB;AACnD,SAASC,eAAe,QAAQ,oBAAmB;AACnD,SAASC,UAAU,QAAQ,qBAAoB;AAwB/C,OAAO,MAAMC,YACX,CAACC,mBACD,CAACC;QACC,MAAMC,UAAUC,kBAAkBH;QAElC,MAAMI,mBAAmBJ,iBAAiBK,OAAO,KAAK;QAEtD,IAAID,kBAAkB;YACpB,OAAOH;QACT;QAEA,+CAA+C;QAC/C,MAAMK,yBAAmEC,OAAOC,OAAO,CACrFR,iBAAiBS,WAAW,EAC5BC,MAAM,CACN,CAACC,KAAK,CAACC,MAAMC,YAAY,GAAM,CAAA;gBAC7B,GAAGF,GAAG;gBACN,CAACC,KAAK,EAAE;oBACN,GAAIC,gBAAgB,OAAO,CAAC,IAAIA,WAAW;oBAC3CX;gBACF;YACF,CAAA,GACA,CAAC;QAGH,gFAAgF;QAChF,MAAMY,SAAS;YACb,GAAGb,cAAc;YACjBQ,aAAa,AAACR,CAAAA,eAAeQ,WAAW,IAAI,EAAE,AAAD,EAAGM,GAAG,CAAC,CAACC;gBACnD,IAAI,CAACV,sBAAsB,CAACU,WAAWJ,IAAI,CAAC,EAAE;oBAC5C,OAAOI;gBACT;gBAEA,OAAO;oBACL,GAAGA,UAAU;oBACbC,QAAQ;wBACN,GAAI,OAAOD,WAAWC,MAAM,KAAK,WAAWD,WAAWC,MAAM,GAAG,CAAC,CAAC;wBAClEC,qBAAqB;oBACvB;gBACF;YACF;QACF;QAEA,OAAOvB,mBAAmB;YACxBwB,oBAAoBnB,iBAAiBmB,kBAAkB;YACvDV,aAAaH;QACf,GAAGQ;IACL,EAAC;AAEH,SAASX,kBAAkB,EAAEiB,MAAM,EAAoB;IACrD,OAAO,CAAC,EAAEJ,UAAU,EAAEK,MAAM,EAAE;QAC5B,OAAO;YACLC,MAAM;YACNC,cAAc3B,gBAAgB;gBAAEwB;YAAO;YACvCI,cAAc3B,gBAAgB;gBAC5BuB;gBACAJ;gBACAK;YACF;YACAI,eAAe3B,WAAW;gBAAEsB;gBAAQJ;gBAAYK;YAAO;QACzD;IACF;AACF"}
1
+ {"version":3,"sources":["../src/index.ts"],"sourcesContent":["import type {\n Adapter,\n ClientUploadsConfig,\n PluginOptions as CloudStoragePluginOptions,\n CollectionOptions,\n GeneratedAdapter,\n} from '@payloadcms/plugin-cloud-storage/types'\nimport type { Config, Plugin, UploadCollectionSlug } from 'payload'\n\nimport { cloudStoragePlugin } from '@payloadcms/plugin-cloud-storage'\nimport { initClientUploads } from '@payloadcms/plugin-cloud-storage/utilities'\n\nimport type { R2Bucket, R2StorageClientUploadHandlerParams } from './types.js'\n\nimport { getHandleDelete } from './handleDelete.js'\nimport { getHandleMultiPartUpload } from './handleMultiPartUpload.js'\nimport { getHandleUpload } from './handleUpload.js'\nimport { getHandler } from './staticHandler.js'\n\nexport interface R2StorageOptions {\n /**\n * When enabled, fields (like the prefix field) will always be inserted into\n * the collection schema regardless of whether the plugin is enabled. This\n * ensures a consistent schema across all environments.\n *\n * This will be enabled by default in Payload v4.\n *\n * @default false\n */\n alwaysInsertFields?: boolean\n\n bucket: R2Bucket\n /**\n * Do uploads directly on the client, to bypass limits on Cloudflare/Vercel.\n */\n clientUploads?: ClientUploadsConfig\n /**\n * Collection options to apply the R2 adapter to.\n */\n collections: Partial<Record<UploadCollectionSlug, Omit<CollectionOptions, 'adapter'> | true>>\n enabled?: boolean\n}\n\ntype R2StoragePlugin = (r2StorageArgs: R2StorageOptions) => Plugin\n\nexport const r2Storage: R2StoragePlugin =\n (r2StorageOptions) =>\n (incomingConfig: Config): Config => {\n const adapter = r2StorageInternal(r2StorageOptions)\n\n const isPluginDisabled = r2StorageOptions.enabled === false\n\n initClientUploads<\n R2StorageClientUploadHandlerParams,\n R2StorageOptions['collections'][keyof R2StorageOptions['collections']]\n >({\n clientHandler: '@payloadcms/storage-r2/client#R2ClientUploadHandler',\n collections: r2StorageOptions.collections,\n config: incomingConfig,\n enabled: !isPluginDisabled && Boolean(r2StorageOptions.clientUploads),\n extraClientHandlerProps: (collection) => ({\n prefix:\n (typeof collection === 'object' && collection.prefix && `${collection.prefix}/`) || '',\n }),\n serverHandler: getHandleMultiPartUpload({\n access:\n typeof r2StorageOptions.clientUploads === 'object'\n ? r2StorageOptions.clientUploads.access\n : undefined,\n bucket: r2StorageOptions.bucket,\n collections: r2StorageOptions.collections,\n }),\n serverHandlerPath: '/storage-r2-multi-part-upload',\n })\n\n if (isPluginDisabled) {\n return incomingConfig\n }\n\n // Add adapter to each collection option object\n const collectionsWithAdapter: CloudStoragePluginOptions['collections'] = Object.entries(\n r2StorageOptions.collections,\n ).reduce(\n (acc, [slug, collOptions]) => ({\n ...acc,\n [slug]: {\n ...(collOptions === true ? {} : collOptions),\n adapter,\n },\n }),\n {} as Record<string, CollectionOptions>,\n )\n\n // Set disableLocalStorage: true for collections specified in the plugin options\n const config = {\n ...incomingConfig,\n collections: (incomingConfig.collections || []).map((collection) => {\n if (!collectionsWithAdapter[collection.slug]) {\n return collection\n }\n\n return {\n ...collection,\n upload: {\n ...(typeof collection.upload === 'object' ? collection.upload : {}),\n disableLocalStorage: true,\n },\n }\n }),\n }\n\n return cloudStoragePlugin({\n alwaysInsertFields: r2StorageOptions.alwaysInsertFields,\n collections: collectionsWithAdapter,\n })(config)\n }\n\nfunction r2StorageInternal({ bucket, clientUploads }: R2StorageOptions): Adapter {\n return ({ collection, prefix }): GeneratedAdapter => {\n return {\n name: 'r2',\n clientUploads,\n handleDelete: getHandleDelete({ bucket }),\n handleUpload: getHandleUpload({\n bucket,\n collection,\n prefix,\n }),\n staticHandler: getHandler({ bucket, collection, prefix }),\n }\n }\n}\n"],"names":["cloudStoragePlugin","initClientUploads","getHandleDelete","getHandleMultiPartUpload","getHandleUpload","getHandler","r2Storage","r2StorageOptions","incomingConfig","adapter","r2StorageInternal","isPluginDisabled","enabled","clientHandler","collections","config","Boolean","clientUploads","extraClientHandlerProps","collection","prefix","serverHandler","access","undefined","bucket","serverHandlerPath","collectionsWithAdapter","Object","entries","reduce","acc","slug","collOptions","map","upload","disableLocalStorage","alwaysInsertFields","name","handleDelete","handleUpload","staticHandler"],"mappings":"AASA,SAASA,kBAAkB,QAAQ,mCAAkC;AACrE,SAASC,iBAAiB,QAAQ,6CAA4C;AAI9E,SAASC,eAAe,QAAQ,oBAAmB;AACnD,SAASC,wBAAwB,QAAQ,6BAA4B;AACrE,SAASC,eAAe,QAAQ,oBAAmB;AACnD,SAASC,UAAU,QAAQ,qBAAoB;AA4B/C,OAAO,MAAMC,YACX,CAACC,mBACD,CAACC;QACC,MAAMC,UAAUC,kBAAkBH;QAElC,MAAMI,mBAAmBJ,iBAAiBK,OAAO,KAAK;QAEtDX,kBAGE;YACAY,eAAe;YACfC,aAAaP,iBAAiBO,WAAW;YACzCC,QAAQP;YACRI,SAAS,CAACD,oBAAoBK,QAAQT,iBAAiBU,aAAa;YACpEC,yBAAyB,CAACC,aAAgB,CAAA;oBACxCC,QACE,AAAC,OAAOD,eAAe,YAAYA,WAAWC,MAAM,IAAI,GAAGD,WAAWC,MAAM,CAAC,CAAC,CAAC,IAAK;gBACxF,CAAA;YACAC,eAAelB,yBAAyB;gBACtCmB,QACE,OAAOf,iBAAiBU,aAAa,KAAK,WACtCV,iBAAiBU,aAAa,CAACK,MAAM,GACrCC;gBACNC,QAAQjB,iBAAiBiB,MAAM;gBAC/BV,aAAaP,iBAAiBO,WAAW;YAC3C;YACAW,mBAAmB;QACrB;QAEA,IAAId,kBAAkB;YACpB,OAAOH;QACT;QAEA,+CAA+C;QAC/C,MAAMkB,yBAAmEC,OAAOC,OAAO,CACrFrB,iBAAiBO,WAAW,EAC5Be,MAAM,CACN,CAACC,KAAK,CAACC,MAAMC,YAAY,GAAM,CAAA;gBAC7B,GAAGF,GAAG;gBACN,CAACC,KAAK,EAAE;oBACN,GAAIC,gBAAgB,OAAO,CAAC,IAAIA,WAAW;oBAC3CvB;gBACF;YACF,CAAA,GACA,CAAC;QAGH,gFAAgF;QAChF,MAAMM,SAAS;YACb,GAAGP,cAAc;YACjBM,aAAa,AAACN,CAAAA,eAAeM,WAAW,IAAI,EAAE,AAAD,EAAGmB,GAAG,CAAC,CAACd;gBACnD,IAAI,CAACO,sBAAsB,CAACP,WAAWY,IAAI,CAAC,EAAE;oBAC5C,OAAOZ;gBACT;gBAEA,OAAO;oBACL,GAAGA,UAAU;oBACbe,QAAQ;wBACN,GAAI,OAAOf,WAAWe,MAAM,KAAK,WAAWf,WAAWe,MAAM,GAAG,CAAC,CAAC;wBAClEC,qBAAqB;oBACvB;gBACF;YACF;QACF;QAEA,OAAOnC,mBAAmB;YACxBoC,oBAAoB7B,iBAAiB6B,kBAAkB;YACvDtB,aAAaY;QACf,GAAGX;IACL,EAAC;AAEH,SAASL,kBAAkB,EAAEc,MAAM,EAAEP,aAAa,EAAoB;IACpE,OAAO,CAAC,EAAEE,UAAU,EAAEC,MAAM,EAAE;QAC5B,OAAO;YACLiB,MAAM;YACNpB;YACAqB,cAAcpC,gBAAgB;gBAAEsB;YAAO;YACvCe,cAAcnC,gBAAgB;gBAC5BoB;gBACAL;gBACAC;YACF;YACAoB,eAAenC,WAAW;gBAAEmB;gBAAQL;gBAAYC;YAAO;QACzD;IACF;AACF"}
@@ -1 +1 @@
1
- {"version":3,"file":"staticHandler.d.ts","sourceRoot":"","sources":["../src/staticHandler.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,wCAAwC,CAAA;AAC3E,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,SAAS,CAAA;AAK/C,OAAO,KAAK,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAA;AAE1C,UAAU,IAAI;IACZ,MAAM,EAAE,QAAQ,CAAA;IAChB,UAAU,EAAE,gBAAgB,CAAA;IAC5B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAID,eAAO,MAAM,UAAU,mCAAyC,IAAI,KAAG,aA8EtE,CAAA"}
1
+ {"version":3,"file":"staticHandler.d.ts","sourceRoot":"","sources":["../src/staticHandler.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,wCAAwC,CAAA;AAC3E,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,SAAS,CAAA;AAK/C,OAAO,KAAK,EAAE,QAAQ,EAAgB,MAAM,YAAY,CAAA;AAExD,UAAU,IAAI;IACZ,MAAM,EAAE,QAAQ,CAAA;IAChB,UAAU,EAAE,gBAAgB,CAAA;IAC5B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAID,eAAO,MAAM,UAAU,mCAAyC,IAAI,KAAG,aAqGtE,CAAA"}
@@ -2,7 +2,7 @@ import path from 'path';
2
2
  import { getRangeRequestInfo } from 'payload/internal';
3
3
  const isMiniflare = process.env.NODE_ENV === 'development';
4
4
  export const getHandler = ({ bucket, collection, prefix = '' })=>{
5
- return async (req, { headers: incomingHeaders, params: { filename } })=>{
5
+ return async (req, { headers: incomingHeaders, params: { clientUploadContext, filename } })=>{
6
6
  try {
7
7
  const key = path.posix.join(prefix, filename);
8
8
  // Get file size for range validation
@@ -14,6 +14,12 @@ export const getHandler = ({ bucket, collection, prefix = '' })=>{
14
14
  });
15
15
  }
16
16
  const fileSize = headObj.size;
17
+ // Don't return large file uploads back to the client, or the Worker will run out of memory
18
+ if (fileSize > 50 * 1024 * 1024 && clientUploadContext) {
19
+ return new Response(null, {
20
+ status: 200
21
+ });
22
+ }
17
23
  // Handle range request
18
24
  const rangeHeader = req.headers.get('range');
19
25
  const rangeResult = getRangeRequestInfo({
@@ -47,7 +53,25 @@ export const getHandler = ({ bucket, collection, prefix = '' })=>{
47
53
  headers.append(key, value);
48
54
  }
49
55
  // Add R2-specific headers
50
- if (!isMiniflare) {
56
+ if (isMiniflare) {
57
+ // In development with Miniflare, manually set headers from httpMetadata
58
+ const metadata = obj.httpMetadata;
59
+ if (metadata?.cacheControl) {
60
+ headers.set('Cache-Control', metadata.cacheControl);
61
+ }
62
+ if (metadata?.contentDisposition) {
63
+ headers.set('Content-Disposition', metadata.contentDisposition);
64
+ }
65
+ if (metadata?.contentEncoding) {
66
+ headers.set('Content-Encoding', metadata.contentEncoding);
67
+ }
68
+ if (metadata?.contentLanguage) {
69
+ headers.set('Content-Language', metadata.contentLanguage);
70
+ }
71
+ if (metadata?.contentType) {
72
+ headers.set('Content-Type', metadata.contentType);
73
+ }
74
+ } else {
51
75
  obj.writeHttpMetadata(headers);
52
76
  }
53
77
  const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match');
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/staticHandler.ts"],"sourcesContent":["import type { StaticHandler } from '@payloadcms/plugin-cloud-storage/types'\nimport type { CollectionConfig } from 'payload'\n\nimport path from 'path'\nimport { getRangeRequestInfo } from 'payload/internal'\n\nimport type { R2Bucket } from './types.js'\n\ninterface Args {\n bucket: R2Bucket\n collection: CollectionConfig\n prefix?: string\n}\n\nconst isMiniflare = process.env.NODE_ENV === 'development'\n\nexport const getHandler = ({ bucket, collection, prefix = '' }: Args): StaticHandler => {\n return async (req, { headers: incomingHeaders, params: { filename } }) => {\n try {\n const key = path.posix.join(prefix, filename)\n\n // Get file size for range validation\n const headObj = await bucket?.head(key)\n if (!headObj) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n const fileSize = headObj.size\n\n // Handle range request\n const rangeHeader = req.headers.get('range')\n const rangeResult = getRangeRequestInfo({ fileSize, rangeHeader })\n\n if (rangeResult.type === 'invalid') {\n return new Response(null, {\n headers: new Headers(rangeResult.headers),\n status: rangeResult.status,\n })\n }\n\n // Get object with range if needed\n // Due to https://github.com/cloudflare/workers-sdk/issues/6047\n // We cannot send a Headers instance to Miniflare\n const obj =\n rangeResult.type === 'partial' && !isMiniflare\n ? await bucket?.get(key, {\n range: {\n length: rangeResult.rangeEnd - rangeResult.rangeStart + 1,\n offset: rangeResult.rangeStart,\n },\n })\n : await bucket?.get(key)\n\n if (obj?.body == undefined) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n let headers = new Headers(incomingHeaders)\n\n // Add range-related headers from the result\n for (const [key, value] of Object.entries(rangeResult.headers)) {\n headers.append(key, value)\n }\n\n // Add R2-specific headers\n if (!isMiniflare) {\n obj.writeHttpMetadata(headers)\n }\n\n const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match')\n\n if (\n collection.upload &&\n typeof collection.upload === 'object' &&\n typeof collection.upload.modifyResponseHeaders === 'function'\n ) {\n headers = collection.upload.modifyResponseHeaders({ headers }) || headers\n }\n\n if (etagFromHeaders && etagFromHeaders === obj.etag) {\n return new Response(null, {\n headers,\n status: 304,\n })\n }\n\n return new Response(obj.body, {\n headers,\n status: rangeResult.status,\n })\n } catch (err: unknown) {\n return new Response('Internal Server Error', { status: 500 })\n }\n }\n}\n"],"names":["path","getRangeRequestInfo","isMiniflare","process","env","NODE_ENV","getHandler","bucket","collection","prefix","req","headers","incomingHeaders","params","filename","key","posix","join","headObj","head","Response","status","statusText","fileSize","size","rangeHeader","get","rangeResult","type","Headers","obj","range","length","rangeEnd","rangeStart","offset","body","undefined","value","Object","entries","append","writeHttpMetadata","etagFromHeaders","upload","modifyResponseHeaders","etag","err"],"mappings":"AAGA,OAAOA,UAAU,OAAM;AACvB,SAASC,mBAAmB,QAAQ,mBAAkB;AAUtD,MAAMC,cAAcC,QAAQC,GAAG,CAACC,QAAQ,KAAK;AAE7C,OAAO,MAAMC,aAAa,CAAC,EAAEC,MAAM,EAAEC,UAAU,EAAEC,SAAS,EAAE,EAAQ;IAClE,OAAO,OAAOC,KAAK,EAAEC,SAASC,eAAe,EAAEC,QAAQ,EAAEC,QAAQ,EAAE,EAAE;QACnE,IAAI;YACF,MAAMC,MAAMf,KAAKgB,KAAK,CAACC,IAAI,CAACR,QAAQK;YAEpC,qCAAqC;YACrC,MAAMI,UAAU,MAAMX,QAAQY,KAAKJ;YACnC,IAAI,CAACG,SAAS;gBACZ,OAAO,IAAIE,SAAS,MAAM;oBAAEC,QAAQ;oBAAKC,YAAY;gBAAY;YACnE;YAEA,MAAMC,WAAWL,QAAQM,IAAI;YAE7B,uBAAuB;YACvB,MAAMC,cAAcf,IAAIC,OAAO,CAACe,GAAG,CAAC;YACpC,MAAMC,cAAc1B,oBAAoB;gBAAEsB;gBAAUE;YAAY;YAEhE,IAAIE,YAAYC,IAAI,KAAK,WAAW;gBAClC,OAAO,IAAIR,SAAS,MAAM;oBACxBT,SAAS,IAAIkB,QAAQF,YAAYhB,OAAO;oBACxCU,QAAQM,YAAYN,MAAM;gBAC5B;YACF;YAEA,kCAAkC;YAClC,+DAA+D;YAC/D,iDAAiD;YACjD,MAAMS,MACJH,YAAYC,IAAI,KAAK,aAAa,CAAC1B,cAC/B,MAAMK,QAAQmB,IAAIX,KAAK;gBACrBgB,OAAO;oBACLC,QAAQL,YAAYM,QAAQ,GAAGN,YAAYO,UAAU,GAAG;oBACxDC,QAAQR,YAAYO,UAAU;gBAChC;YACF,KACA,MAAM3B,QAAQmB,IAAIX;YAExB,IAAIe,KAAKM,QAAQC,WAAW;gBAC1B,OAAO,IAAIjB,SAAS,MAAM;oBAAEC,QAAQ;oBAAKC,YAAY;gBAAY;YACnE;YAEA,IAAIX,UAAU,IAAIkB,QAAQjB;YAE1B,4CAA4C;YAC5C,KAAK,MAAM,CAACG,KAAKuB,MAAM,IAAIC,OAAOC,OAAO,CAACb,YAAYhB,OAAO,EAAG;gBAC9DA,QAAQ8B,MAAM,CAAC1B,KAAKuB;YACtB;YAEA,0BAA0B;YAC1B,IAAI,CAACpC,aAAa;gBAChB4B,IAAIY,iBAAiB,CAAC/B;YACxB;YAEA,MAAMgC,kBAAkBjC,IAAIC,OAAO,CAACe,GAAG,CAAC,WAAWhB,IAAIC,OAAO,CAACe,GAAG,CAAC;YAEnE,IACElB,WAAWoC,MAAM,IACjB,OAAOpC,WAAWoC,MAAM,KAAK,YAC7B,OAAOpC,WAAWoC,MAAM,CAACC,qBAAqB,KAAK,YACnD;gBACAlC,UAAUH,WAAWoC,MAAM,CAACC,qBAAqB,CAAC;oBAAElC;gBAAQ,MAAMA;YACpE;YAEA,IAAIgC,mBAAmBA,oBAAoBb,IAAIgB,IAAI,EAAE;gBACnD,OAAO,IAAI1B,SAAS,MAAM;oBACxBT;oBACAU,QAAQ;gBACV;YACF;YAEA,OAAO,IAAID,SAASU,IAAIM,IAAI,EAAE;gBAC5BzB;gBACAU,QAAQM,YAAYN,MAAM;YAC5B;QACF,EAAE,OAAO0B,KAAc;YACrB,OAAO,IAAI3B,SAAS,yBAAyB;gBAAEC,QAAQ;YAAI;QAC7D;IACF;AACF,EAAC"}
1
+ {"version":3,"sources":["../src/staticHandler.ts"],"sourcesContent":["import type { StaticHandler } from '@payloadcms/plugin-cloud-storage/types'\nimport type { CollectionConfig } from 'payload'\n\nimport path from 'path'\nimport { getRangeRequestInfo } from 'payload/internal'\n\nimport type { R2Bucket, R2ObjectBody } from './types.js'\n\ninterface Args {\n bucket: R2Bucket\n collection: CollectionConfig\n prefix?: string\n}\n\nconst isMiniflare = process.env.NODE_ENV === 'development'\n\nexport const getHandler = ({ bucket, collection, prefix = '' }: Args): StaticHandler => {\n return async (req, { headers: incomingHeaders, params: { clientUploadContext, filename } }) => {\n try {\n const key = path.posix.join(prefix, filename)\n\n // Get file size for range validation\n const headObj = await bucket?.head(key)\n if (!headObj) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n const fileSize = headObj.size\n\n // Don't return large file uploads back to the client, or the Worker will run out of memory\n if (fileSize > 50 * 1024 * 1024 && clientUploadContext) {\n return new Response(null, { status: 200 })\n }\n\n // Handle range request\n const rangeHeader = req.headers.get('range')\n const rangeResult = getRangeRequestInfo({ fileSize, rangeHeader })\n\n if (rangeResult.type === 'invalid') {\n return new Response(null, {\n headers: new Headers(rangeResult.headers),\n status: rangeResult.status,\n })\n }\n\n // Get object with range if needed\n // Due to https://github.com/cloudflare/workers-sdk/issues/6047\n // We cannot send a Headers instance to Miniflare\n const obj: R2ObjectBody =\n rangeResult.type === 'partial' && !isMiniflare\n ? await bucket?.get(key, {\n range: {\n length: rangeResult.rangeEnd - rangeResult.rangeStart + 1,\n offset: rangeResult.rangeStart,\n },\n })\n : await bucket?.get(key)\n\n if (obj?.body == undefined) {\n return new Response(null, { status: 404, statusText: 'Not Found' })\n }\n\n let headers = new Headers(incomingHeaders)\n\n // Add range-related headers from the result\n for (const [key, value] of Object.entries(rangeResult.headers)) {\n headers.append(key, value)\n }\n\n // Add R2-specific headers\n if (isMiniflare) {\n // In development with Miniflare, manually set headers from httpMetadata\n const metadata = obj.httpMetadata\n if (metadata?.cacheControl) {\n headers.set('Cache-Control', metadata.cacheControl)\n }\n if (metadata?.contentDisposition) {\n headers.set('Content-Disposition', metadata.contentDisposition)\n }\n if (metadata?.contentEncoding) {\n headers.set('Content-Encoding', metadata.contentEncoding)\n }\n if (metadata?.contentLanguage) {\n headers.set('Content-Language', metadata.contentLanguage)\n }\n if (metadata?.contentType) {\n headers.set('Content-Type', metadata.contentType)\n }\n } else {\n obj.writeHttpMetadata(headers)\n }\n\n const etagFromHeaders = req.headers.get('etag') || req.headers.get('if-none-match')\n\n if (\n collection.upload &&\n typeof collection.upload === 'object' &&\n typeof collection.upload.modifyResponseHeaders === 'function'\n ) {\n headers = collection.upload.modifyResponseHeaders({ headers }) || headers\n }\n\n if (etagFromHeaders && etagFromHeaders === obj.etag) {\n return new Response(null, {\n headers,\n status: 304,\n })\n }\n\n return new Response(obj.body, {\n headers,\n status: rangeResult.status,\n })\n } catch (err: unknown) {\n return new Response('Internal Server Error', { status: 500 })\n }\n }\n}\n"],"names":["path","getRangeRequestInfo","isMiniflare","process","env","NODE_ENV","getHandler","bucket","collection","prefix","req","headers","incomingHeaders","params","clientUploadContext","filename","key","posix","join","headObj","head","Response","status","statusText","fileSize","size","rangeHeader","get","rangeResult","type","Headers","obj","range","length","rangeEnd","rangeStart","offset","body","undefined","value","Object","entries","append","metadata","httpMetadata","cacheControl","set","contentDisposition","contentEncoding","contentLanguage","contentType","writeHttpMetadata","etagFromHeaders","upload","modifyResponseHeaders","etag","err"],"mappings":"AAGA,OAAOA,UAAU,OAAM;AACvB,SAASC,mBAAmB,QAAQ,mBAAkB;AAUtD,MAAMC,cAAcC,QAAQC,GAAG,CAACC,QAAQ,KAAK;AAE7C,OAAO,MAAMC,aAAa,CAAC,EAAEC,MAAM,EAAEC,UAAU,EAAEC,SAAS,EAAE,EAAQ;IAClE,OAAO,OAAOC,KAAK,EAAEC,SAASC,eAAe,EAAEC,QAAQ,EAAEC,mBAAmB,EAAEC,QAAQ,EAAE,EAAE;QACxF,IAAI;YACF,MAAMC,MAAMhB,KAAKiB,KAAK,CAACC,IAAI,CAACT,QAAQM;YAEpC,qCAAqC;YACrC,MAAMI,UAAU,MAAMZ,QAAQa,KAAKJ;YACnC,IAAI,CAACG,SAAS;gBACZ,OAAO,IAAIE,SAAS,MAAM;oBAAEC,QAAQ;oBAAKC,YAAY;gBAAY;YACnE;YAEA,MAAMC,WAAWL,QAAQM,IAAI;YAE7B,2FAA2F;YAC3F,IAAID,WAAW,KAAK,OAAO,QAAQV,qBAAqB;gBACtD,OAAO,IAAIO,SAAS,MAAM;oBAAEC,QAAQ;gBAAI;YAC1C;YAEA,uBAAuB;YACvB,MAAMI,cAAchB,IAAIC,OAAO,CAACgB,GAAG,CAAC;YACpC,MAAMC,cAAc3B,oBAAoB;gBAAEuB;gBAAUE;YAAY;YAEhE,IAAIE,YAAYC,IAAI,KAAK,WAAW;gBAClC,OAAO,IAAIR,SAAS,MAAM;oBACxBV,SAAS,IAAImB,QAAQF,YAAYjB,OAAO;oBACxCW,QAAQM,YAAYN,MAAM;gBAC5B;YACF;YAEA,kCAAkC;YAClC,+DAA+D;YAC/D,iDAAiD;YACjD,MAAMS,MACJH,YAAYC,IAAI,KAAK,aAAa,CAAC3B,cAC/B,MAAMK,QAAQoB,IAAIX,KAAK;gBACrBgB,OAAO;oBACLC,QAAQL,YAAYM,QAAQ,GAAGN,YAAYO,UAAU,GAAG;oBACxDC,QAAQR,YAAYO,UAAU;gBAChC;YACF,KACA,MAAM5B,QAAQoB,IAAIX;YAExB,IAAIe,KAAKM,QAAQC,WAAW;gBAC1B,OAAO,IAAIjB,SAAS,MAAM;oBAAEC,QAAQ;oBAAKC,YAAY;gBAAY;YACnE;YAEA,IAAIZ,UAAU,IAAImB,QAAQlB;YAE1B,4CAA4C;YAC5C,KAAK,MAAM,CAACI,KAAKuB,MAAM,IAAIC,OAAOC,OAAO,CAACb,YAAYjB,OAAO,EAAG;gBAC9DA,QAAQ+B,MAAM,CAAC1B,KAAKuB;YACtB;YAEA,0BAA0B;YAC1B,IAAIrC,aAAa;gBACf,wEAAwE;gBACxE,MAAMyC,WAAWZ,IAAIa,YAAY;gBACjC,IAAID,UAAUE,cAAc;oBAC1BlC,QAAQmC,GAAG,CAAC,iBAAiBH,SAASE,YAAY;gBACpD;gBACA,IAAIF,UAAUI,oBAAoB;oBAChCpC,QAAQmC,GAAG,CAAC,uBAAuBH,SAASI,kBAAkB;gBAChE;gBACA,IAAIJ,UAAUK,iBAAiB;oBAC7BrC,QAAQmC,GAAG,CAAC,oBAAoBH,SAASK,eAAe;gBAC1D;gBACA,IAAIL,UAAUM,iBAAiB;oBAC7BtC,QAAQmC,GAAG,CAAC,oBAAoBH,SAASM,eAAe;gBAC1D;gBACA,IAAIN,UAAUO,aAAa;oBACzBvC,QAAQmC,GAAG,CAAC,gBAAgBH,SAASO,WAAW;gBAClD;YACF,OAAO;gBACLnB,IAAIoB,iBAAiB,CAACxC;YACxB;YAEA,MAAMyC,kBAAkB1C,IAAIC,OAAO,CAACgB,GAAG,CAAC,WAAWjB,IAAIC,OAAO,CAACgB,GAAG,CAAC;YAEnE,IACEnB,WAAW6C,MAAM,IACjB,OAAO7C,WAAW6C,MAAM,KAAK,YAC7B,OAAO7C,WAAW6C,MAAM,CAACC,qBAAqB,KAAK,YACnD;gBACA3C,UAAUH,WAAW6C,MAAM,CAACC,qBAAqB,CAAC;oBAAE3C;gBAAQ,MAAMA;YACpE;YAEA,IAAIyC,mBAAmBA,oBAAoBrB,IAAIwB,IAAI,EAAE;gBACnD,OAAO,IAAIlC,SAAS,MAAM;oBACxBV;oBACAW,QAAQ;gBACV;YACF;YAEA,OAAO,IAAID,SAASU,IAAIM,IAAI,EAAE;gBAC5B1B;gBACAW,QAAQM,YAAYN,MAAM;YAC5B;QACF,EAAE,OAAOkC,KAAc;YACrB,OAAO,IAAInC,SAAS,yBAAyB;gBAAEC,QAAQ;YAAI;QAC7D;IACF;AACF,EAAC"}
package/dist/types.d.ts CHANGED
@@ -12,7 +12,7 @@ export interface R2GetOptions {
12
12
  range?: R2Range;
13
13
  }
14
14
  export interface R2Bucket {
15
- createMultipartUpload(key: string, options?: any): Promise<any>;
15
+ createMultipartUpload(key: string, options?: any): Promise<R2MultipartUpload>;
16
16
  delete(keys: string | string[]): Promise<void>;
17
17
  get(key: string, options?: R2GetOptions): Promise<any | null>;
18
18
  head(key: string): Promise<any>;
@@ -22,6 +22,51 @@ export interface R2Bucket {
22
22
  onlyIf: any;
23
23
  } & any): Promise<any | null>;
24
24
  put(key: string, value: ArrayBuffer | ArrayBufferView | Blob | null | ReadableStream | string, options?: any): Promise<any>;
25
- resumeMultipartUpload(key: string, uploadId: string): any;
25
+ resumeMultipartUpload(key: string, uploadId: string): R2MultipartUpload;
26
26
  }
27
+ interface R2HTTPMetadata {
28
+ cacheControl?: string;
29
+ cacheExpiry?: Date;
30
+ contentDisposition?: string;
31
+ contentEncoding?: string;
32
+ contentLanguage?: string;
33
+ contentType?: string;
34
+ }
35
+ export interface R2Object {
36
+ readonly etag: string;
37
+ readonly httpMetadata?: R2HTTPMetadata;
38
+ readonly key: string;
39
+ readonly size: number;
40
+ writeHttpMetadata(headers: Headers): void;
41
+ }
42
+ export interface R2ObjectBody extends R2Object {
43
+ get body(): ReadableStream;
44
+ }
45
+ export interface R2MultipartUpload {
46
+ abort(): Promise<void>;
47
+ complete(uploadedParts: R2UploadedPart[]): Promise<R2Object>;
48
+ readonly key: string;
49
+ readonly uploadId: string;
50
+ uploadPart(partNumber: number, value: (ArrayBuffer | ArrayBufferView) | Blob | ReadableStream | string, options?: any): Promise<R2UploadedPart>;
51
+ }
52
+ export interface R2StorageClientUploadContext {
53
+ key: string;
54
+ }
55
+ export type R2StorageClientUploadHandlerParams = {
56
+ chunkSize?: number;
57
+ prefix: string;
58
+ };
59
+ export type R2StorageMultipartUploadHandlerParams = {
60
+ collection: string;
61
+ fileName: string;
62
+ fileType: string;
63
+ multipartId?: string;
64
+ multipartKey?: string;
65
+ multipartNumber?: string;
66
+ };
67
+ export interface R2UploadedPart {
68
+ etag: string;
69
+ partNumber: number;
70
+ }
71
+ export {};
27
72
  //# sourceMappingURL=types.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"types.d.ts","sourceRoot":"","sources":["../src/types.ts"],"names":[],"mappings":"AAAA,MAAM,WAAW,OAAO;IACtB,oCAAoC;IACpC,MAAM,CAAC,EAAE,MAAM,CAAA;IACf,gDAAgD;IAChD,MAAM,CAAC,EAAE,MAAM,CAAA;IACf,8BAA8B;IAC9B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAED,MAAM,WAAW,YAAY;IAC3B,CAAC,GAAG,EAAE,MAAM,GAAG,GAAG,CAAA;IAClB,MAAM,CAAC,EAAE,GAAG,GAAG,OAAO,CAAA;IACtB,KAAK,CAAC,EAAE,OAAO,CAAA;CAChB;AAED,MAAM,WAAW,QAAQ;IACvB,qBAAqB,CAAC,GAAG,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE,GAAG,GAAG,OAAO,CAAC,GAAG,CAAC,CAAA;IAC/D,MAAM,CAAC,IAAI,EAAE,MAAM,GAAG,MAAM,EAAE,GAAG,OAAO,CAAC,IAAI,CAAC,CAAA;IAC9C,GAAG,CAAC,GAAG,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE,YAAY,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAAA;IAC7D,IAAI,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,CAAC,CAAA;IAC/B,IAAI,CAAC,OAAO,CAAC,EAAE,GAAG,GAAG,OAAO,CAAC,GAAG,CAAC,CAAA;IACjC,GAAG,CACD,GAAG,EAAE,MAAM,EACX,KAAK,EAAE,WAAW,GAAG,eAAe,GAAG,IAAI,GAAG,IAAI,GAAG,cAAc,GAAG,MAAM,EAC5E,OAAO,CAAC,EAAE;QACR,YAAY,CAAC,EAAE,GAAG,GAAG,OAAO,CAAA;QAC5B,MAAM,EAAE,GAAG,CAAA;KACZ,GAAG,GAAG,GACN,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAAA;IACtB,GAAG,CACD,GAAG,EAAE,MAAM,EACX,KAAK,EAAE,WAAW,GAAG,eAAe,GAAG,IAAI,GAAG,IAAI,GAAG,cAAc,GAAG,MAAM,EAC5E,OAAO,CAAC,EAAE,GAAG,GACZ,OAAO,CAAC,GAAG,CAAC,CAAA;IACf,qBAAqB,CAAC,GAAG,EAAE,MAAM,EAAE,QAAQ,EAAE,MAAM,GAAG,GAAG,CAAA;CAC1D"}
1
+ {"version":3,"file":"types.d.ts","sourceRoot":"","sources":["../src/types.ts"],"names":[],"mappings":"AAAA,MAAM,WAAW,OAAO;IACtB,oCAAoC;IACpC,MAAM,CAAC,EAAE,MAAM,CAAA;IACf,gDAAgD;IAChD,MAAM,CAAC,EAAE,MAAM,CAAA;IACf,8BAA8B;IAC9B,MAAM,CAAC,EAAE,MAAM,CAAA;CAChB;AAED,MAAM,WAAW,YAAY;IAC3B,CAAC,GAAG,EAAE,MAAM,GAAG,GAAG,CAAA;IAClB,MAAM,CAAC,EAAE,GAAG,GAAG,OAAO,CAAA;IACtB,KAAK,CAAC,EAAE,OAAO,CAAA;CAChB;AAED,MAAM,WAAW,QAAQ;IACvB,qBAAqB,CAAC,GAAG,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE,GAAG,GAAG,OAAO,CAAC,iBAAiB,CAAC,CAAA;IAC7E,MAAM,CAAC,IAAI,EAAE,MAAM,GAAG,MAAM,EAAE,GAAG,OAAO,CAAC,IAAI,CAAC,CAAA;IAC9C,GAAG,CAAC,GAAG,EAAE,MAAM,EAAE,OAAO,CAAC,EAAE,YAAY,GAAG,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAAA;IAC7D,IAAI,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC,GAAG,CAAC,CAAA;IAC/B,IAAI,CAAC,OAAO,CAAC,EAAE,GAAG,GAAG,OAAO,CAAC,GAAG,CAAC,CAAA;IACjC,GAAG,CACD,GAAG,EAAE,MAAM,EACX,KAAK,EAAE,WAAW,GAAG,eAAe,GAAG,IAAI,GAAG,IAAI,GAAG,cAAc,GAAG,MAAM,EAC5E,OAAO,CAAC,EAAE;QACR,YAAY,CAAC,EAAE,GAAG,GAAG,OAAO,CAAA;QAC5B,MAAM,EAAE,GAAG,CAAA;KACZ,GAAG,GAAG,GACN,OAAO,CAAC,GAAG,GAAG,IAAI,CAAC,CAAA;IACtB,GAAG,CACD,GAAG,EAAE,MAAM,EACX,KAAK,EAAE,WAAW,GAAG,eAAe,GAAG,IAAI,GAAG,IAAI,GAAG,cAAc,GAAG,MAAM,EAC5E,OAAO,CAAC,EAAE,GAAG,GACZ,OAAO,CAAC,GAAG,CAAC,CAAA;IACf,qBAAqB,CAAC,GAAG,EAAE,MAAM,EAAE,QAAQ,EAAE,MAAM,GAAG,iBAAiB,CAAA;CACxE;AAED,UAAU,cAAc;IACtB,YAAY,CAAC,EAAE,MAAM,CAAA;IACrB,WAAW,CAAC,EAAE,IAAI,CAAA;IAClB,kBAAkB,CAAC,EAAE,MAAM,CAAA;IAC3B,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,eAAe,CAAC,EAAE,MAAM,CAAA;IACxB,WAAW,CAAC,EAAE,MAAM,CAAA;CACrB;AAED,MAAM,WAAW,QAAQ;IACvB,QAAQ,CAAC,IAAI,EAAE,MAAM,CAAA;IACrB,QAAQ,CAAC,YAAY,CAAC,EAAE,cAAc,CAAA;IACtC,QAAQ,CAAC,GAAG,EAAE,MAAM,CAAA;IACpB,QAAQ,CAAC,IAAI,EAAE,MAAM,CAAA;IAErB,iBAAiB,CAAC,OAAO,EAAE,OAAO,GAAG,IAAI,CAAA;CAC1C;AACD,MAAM,WAAW,YAAa,SAAQ,QAAQ;IAC5C,IAAI,IAAI,IAAI,cAAc,CAAA;CAC3B;AAED,MAAM,WAAW,iBAAiB;IAChC,KAAK,IAAI,OAAO,CAAC,IAAI,CAAC,CAAA;IACtB,QAAQ,CAAC,aAAa,EAAE,cAAc,EAAE,GAAG,OAAO,CAAC,QAAQ,CAAC,CAAA;IAC5D,QAAQ,CAAC,GAAG,EAAE,MAAM,CAAA;IACpB,QAAQ,CAAC,QAAQ,EAAE,MAAM,CAAA;IACzB,UAAU,CACR,UAAU,EAAE,MAAM,EAClB,KAAK,EAAE,CAAC,WAAW,GAAG,eAAe,CAAC,GAAG,IAAI,GAAG,cAAc,GAAG,MAAM,EACvE,OAAO,CAAC,EAAE,GAAG,GACZ,OAAO,CAAC,cAAc,CAAC,CAAA;CAC3B;AAED,MAAM,WAAW,4BAA4B;IAC3C,GAAG,EAAE,MAAM,CAAA;CACZ;AACD,MAAM,MAAM,kCAAkC,GAAG;IAC/C,SAAS,CAAC,EAAE,MAAM,CAAA;IAClB,MAAM,EAAE,MAAM,CAAA;CACf,CAAA;AAED,MAAM,MAAM,qCAAqC,GAAG;IAClD,UAAU,EAAE,MAAM,CAAA;IAClB,QAAQ,EAAE,MAAM,CAAA;IAChB,QAAQ,EAAE,MAAM,CAAA;IAChB,WAAW,CAAC,EAAE,MAAM,CAAA;IACpB,YAAY,CAAC,EAAE,MAAM,CAAA;IACrB,eAAe,CAAC,EAAE,MAAM,CAAA;CACzB,CAAA;AAED,MAAM,WAAW,cAAc;IAC7B,IAAI,EAAE,MAAM,CAAA;IACZ,UAAU,EAAE,MAAM,CAAA;CACnB"}
package/dist/types.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/types.ts"],"sourcesContent":["export interface R2Range {\n /** The number of bytes to return */\n length?: number\n /** The byte offset to start from (inclusive) */\n offset?: number\n /** Return the last n bytes */\n suffix?: number\n}\n\nexport interface R2GetOptions {\n [key: string]: any\n onlyIf?: any | Headers\n range?: R2Range\n}\n\nexport interface R2Bucket {\n createMultipartUpload(key: string, options?: any): Promise<any>\n delete(keys: string | string[]): Promise<void>\n get(key: string, options?: R2GetOptions): Promise<any | null>\n head(key: string): Promise<any>\n list(options?: any): Promise<any>\n put(\n key: string,\n value: ArrayBuffer | ArrayBufferView | Blob | null | ReadableStream | string,\n options?: {\n httpMetadata?: any | Headers\n onlyIf: any\n } & any,\n ): Promise<any | null>\n put(\n key: string,\n value: ArrayBuffer | ArrayBufferView | Blob | null | ReadableStream | string,\n options?: any,\n ): Promise<any>\n resumeMultipartUpload(key: string, uploadId: string): any\n}\n"],"names":[],"mappings":"AAeA,WAoBC"}
1
+ {"version":3,"sources":["../src/types.ts"],"sourcesContent":["export interface R2Range {\n /** The number of bytes to return */\n length?: number\n /** The byte offset to start from (inclusive) */\n offset?: number\n /** Return the last n bytes */\n suffix?: number\n}\n\nexport interface R2GetOptions {\n [key: string]: any\n onlyIf?: any | Headers\n range?: R2Range\n}\n\nexport interface R2Bucket {\n createMultipartUpload(key: string, options?: any): Promise<R2MultipartUpload>\n delete(keys: string | string[]): Promise<void>\n get(key: string, options?: R2GetOptions): Promise<any | null>\n head(key: string): Promise<any>\n list(options?: any): Promise<any>\n put(\n key: string,\n value: ArrayBuffer | ArrayBufferView | Blob | null | ReadableStream | string,\n options?: {\n httpMetadata?: any | Headers\n onlyIf: any\n } & any,\n ): Promise<any | null>\n put(\n key: string,\n value: ArrayBuffer | ArrayBufferView | Blob | null | ReadableStream | string,\n options?: any,\n ): Promise<any>\n resumeMultipartUpload(key: string, uploadId: string): R2MultipartUpload\n}\n\ninterface R2HTTPMetadata {\n cacheControl?: string\n cacheExpiry?: Date\n contentDisposition?: string\n contentEncoding?: string\n contentLanguage?: string\n contentType?: string\n}\n\nexport interface R2Object {\n readonly etag: string\n readonly httpMetadata?: R2HTTPMetadata\n readonly key: string\n readonly size: number\n\n writeHttpMetadata(headers: Headers): void\n}\nexport interface R2ObjectBody extends R2Object {\n get body(): ReadableStream\n}\n\nexport interface R2MultipartUpload {\n abort(): Promise<void>\n complete(uploadedParts: R2UploadedPart[]): Promise<R2Object>\n readonly key: string\n readonly uploadId: string\n uploadPart(\n partNumber: number,\n value: (ArrayBuffer | ArrayBufferView) | Blob | ReadableStream | string,\n options?: any,\n ): Promise<R2UploadedPart>\n}\n\nexport interface R2StorageClientUploadContext {\n key: string\n}\nexport type R2StorageClientUploadHandlerParams = {\n chunkSize?: number\n prefix: string\n}\n\nexport type R2StorageMultipartUploadHandlerParams = {\n collection: string\n fileName: string\n fileType: string\n multipartId?: string\n multipartKey?: string\n multipartNumber?: string\n}\n\nexport interface R2UploadedPart {\n etag: string\n partNumber: number\n}\n"],"names":[],"mappings":"AAuFA,WAGC"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@payloadcms/storage-r2",
3
- "version": "3.74.0-internal.bc57c0d",
3
+ "version": "3.75.0-internal.880861f",
4
4
  "description": "Payload storage adapter for Cloudflare R2",
5
5
  "homepage": "https://payloadcms.com",
6
6
  "repository": {
@@ -37,13 +37,13 @@
37
37
  "dist"
38
38
  ],
39
39
  "dependencies": {
40
- "@payloadcms/plugin-cloud-storage": "3.74.0-internal.bc57c0d"
40
+ "@payloadcms/plugin-cloud-storage": "3.75.0-internal.880861f"
41
41
  },
42
42
  "devDependencies": {
43
- "payload": "3.74.0-internal.bc57c0d"
43
+ "payload": "3.75.0-internal.880861f"
44
44
  },
45
45
  "peerDependencies": {
46
- "payload": "3.74.0-internal.bc57c0d"
46
+ "payload": "3.75.0-internal.880861f"
47
47
  },
48
48
  "engines": {
49
49
  "node": "^18.20.2 || >=20.9.0"