@storm-software/cloudflare-tools 0.71.117 → 0.71.118

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -2,6 +2,18 @@
2
2
 
3
3
  # Changelog for Storm Ops - Cloudflare Tools
4
4
 
5
+ ## [0.71.117](https://github.com/storm-software/storm-ops/releases/tag/cloudflare-tools%400.71.117) (04/01/2026)
6
+
7
+ ### Miscellaneous
8
+
9
+ - **monorepo:** Regenerate `README.md` file ([b226844d5](https://github.com/storm-software/storm-ops/commit/b226844d5))
10
+
11
+ ### Updated Dependencies
12
+
13
+ - Updated **workspace-tools** to **v1.295.43**
14
+ - Updated **config-tools** to **v1.189.67**
15
+ - Updated **config** to **v1.137.21**
16
+
5
17
  ## [0.71.116](https://github.com/storm-software/storm-ops/releases/tag/cloudflare-tools%400.71.116) (04/01/2026)
6
18
 
7
19
  ### Miscellaneous
@@ -1,8 +1,5 @@
1
1
  "use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }
2
2
 
3
- var _chunkL3VPGV3Gjs = require('./chunk-L3VPGV3G.js');
4
-
5
-
6
3
 
7
4
  var _chunkOEP2XBU4js = require('./chunk-OEP2XBU4.js');
8
5
 
@@ -15,6 +12,9 @@ var _chunkOEP2XBU4js = require('./chunk-OEP2XBU4.js');
15
12
 
16
13
  var _chunkRIBK362Wjs = require('./chunk-RIBK362W.js');
17
14
 
15
+
16
+ var _chunkQ3UOHXUVjs = require('./chunk-Q3UOHXUV.js');
17
+
18
18
  // src/generators/worker/generator.ts
19
19
 
20
20
 
@@ -70,7 +70,7 @@ ${Object.keys(process.env).map((key) => ` - ${key}=${JSON.stringify(process.env[
70
70
  const options = await normalizeOptions(tree, schema, config);
71
71
  const tasks = [];
72
72
  tasks.push(
73
- await _chunkL3VPGV3Gjs.generator_default.call(void 0, tree, {
73
+ await _chunkQ3UOHXUVjs.generator_default.call(void 0, tree, {
74
74
  ...options,
75
75
  skipFormat: true
76
76
  })
@@ -0,0 +1,231 @@
1
+ "use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; } function _nullishCoalesce(lhs, rhsFn) { if (lhs != null) { return lhs; } else { return rhsFn(); } } function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }
2
+
3
+
4
+ var _chunkN7FW365Qjs = require('./chunk-N7FW365Q.js');
5
+
6
+
7
+
8
+ var _chunkOEP2XBU4js = require('./chunk-OEP2XBU4.js');
9
+
10
+
11
+ var _chunkKUGEZPUOjs = require('./chunk-KUGEZPUO.js');
12
+
13
+
14
+
15
+
16
+ var _chunkSIS5VQAUjs = require('./chunk-SIS5VQAU.js');
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+ var _chunkRIBK362Wjs = require('./chunk-RIBK362W.js');
25
+
26
+ // src/executors/r2-upload-publish/executor.ts
27
+
28
+
29
+
30
+
31
+ var _clients3 = require('@aws-sdk/client-s3');
32
+
33
+
34
+
35
+ var _devkit = require('@nx/devkit');
36
+ var _glob = require('glob');
37
+ var _mimetypes = require('mime-types'); var _mimetypes2 = _interopRequireDefault(_mimetypes);
38
+ var _child_process = require('child_process');
39
+ var _fs = require('fs');
40
+ var _promises = require('fs/promises');
41
+ async function runExecutor(options, context) {
42
+ const isDryRun = process.env.NX_DRY_RUN === "true" || options.dryRun || false;
43
+ if (!context.projectName) {
44
+ throw new Error("The executor requires a projectName.");
45
+ }
46
+ if (!options.path) {
47
+ throw new Error("The executor requires the `path` option to upload.");
48
+ }
49
+ console.info(
50
+ `\u{1F680} Running Storm Cloudflare Publish executor on the ${context.projectName} worker`
51
+ );
52
+ if (!context.projectName || !_optionalChain([context, 'access', _ => _.projectsConfigurations, 'optionalAccess', _2 => _2.projects]) || !context.projectsConfigurations.projects[context.projectName] || !_optionalChain([context, 'access', _3 => _3.projectsConfigurations, 'access', _4 => _4.projects, 'access', _5 => _5[context.projectName], 'optionalAccess', _6 => _6.root])) {
53
+ throw new Error("The executor requires projectsConfigurations.");
54
+ }
55
+ try {
56
+ const workspaceRoot = _chunkOEP2XBU4js.findWorkspaceRoot.call(void 0, );
57
+ const config = await _chunkOEP2XBU4js.getConfig.call(void 0, workspaceRoot);
58
+ const projectName = _nullishCoalesce(_optionalChain([context, 'access', _7 => _7.projectsConfigurations, 'access', _8 => _8.projects, 'access', _9 => _9[context.projectName], 'optionalAccess', _10 => _10.name]), () => ( context.projectName));
59
+ const projectDetails = _chunkN7FW365Qjs.getPackageInfo.call(void 0,
60
+ context.projectsConfigurations.projects[context.projectName]
61
+ );
62
+ const bucketId = options.bucketId;
63
+ const bucketPath = options.bucketPath || "/";
64
+ if (!bucketId) {
65
+ throw new Error("The executor requires a bucketId.");
66
+ }
67
+ const args = _chunkN7FW365Qjs.createCliOptions.call(void 0, { ...options });
68
+ if (isDryRun) {
69
+ args.push("--dry-run");
70
+ }
71
+ const cloudflareAccountId = process.env.CLOUDFLARE_ACCOUNT_ID || process.env.STORM_BOT_CLOUDFLARE_ACCOUNT;
72
+ if (!_optionalChain([options, 'optionalAccess', _11 => _11.registry]) && !cloudflareAccountId) {
73
+ throw new Error(
74
+ "The registry option and `CLOUDFLARE_ACCOUNT_ID` (or `STORM_BOT_CLOUDFLARE_ACCOUNT`) environment variable are not set. Please set one of these values to upload to the Cloudflare R2 bucket."
75
+ );
76
+ }
77
+ if (!process.env.STORM_BOT_ACCESS_KEY_ID && !process.env.ACCESS_KEY_ID && !process.env.CLOUDFLARE_ACCESS_KEY_ID && !process.env.AWS_ACCESS_KEY_ID || !process.env.STORM_BOT_SECRET_ACCESS_KEY && !process.env.CLOUDFLARE_SECRET_ACCESS_KEY && !process.env.SECRET_ACCESS_KEY && !process.env.AWS_SECRET_ACCESS_KEY) {
78
+ throw new Error(
79
+ "The `ACCESS_KEY_ID` (or `STORM_BOT_ACCESS_KEY_ID`) and `SECRET_ACCESS_KEY` (or `STORM_BOT_SECRET_ACCESS_KEY`) environment variables are not set. Please set these environment variables to upload to the Cloudflare R2 bucket."
80
+ );
81
+ }
82
+ const registry = _optionalChain([options, 'optionalAccess', _12 => _12.registry]) ? options.registry : `https://${cloudflareAccountId}.r2.cloudflarestorage.com`;
83
+ let projectGraph;
84
+ try {
85
+ projectGraph = _devkit.readCachedProjectGraph.call(void 0, );
86
+ } catch (e) {
87
+ await _devkit.createProjectGraphAsync.call(void 0, );
88
+ projectGraph = _devkit.readCachedProjectGraph.call(void 0, );
89
+ }
90
+ if (!projectGraph) {
91
+ throw new Error(
92
+ "The executor failed because the project graph is not available. Please run the build command again."
93
+ );
94
+ }
95
+ _chunkRIBK362Wjs.writeDebug.call(void 0,
96
+ `Publishing ${context.projectName} to the ${bucketId} R2 Bucket (at ${registry})`
97
+ );
98
+ const client = new (0, _clients3.S3Client)({
99
+ region: "auto",
100
+ endpoint: registry,
101
+ credentials: {
102
+ // eslint-disable-next-line @typescript-eslint/no-non-null-assertion
103
+ accessKeyId: process.env.STORM_BOT_ACCESS_KEY_ID || process.env.CLOUDFLARE_ACCESS_KEY_ID || process.env.AWS_ACCESS_KEY_ID || process.env.ACCESS_KEY_ID,
104
+ // eslint-disable-next-line @typescript-eslint/no-non-null-assertion
105
+ secretAccessKey: process.env.STORM_BOT_SECRET_ACCESS_KEY || process.env.CLOUDFLARE_SECRET_ACCESS_KEY || process.env.AWS_SECRET_ACCESS_KEY || process.env.SECRET_ACCESS_KEY
106
+ },
107
+ requestHandler: _chunkKUGEZPUOjs.createHttpHandler.call(void 0, )
108
+ });
109
+ const version = _optionalChain([projectDetails, 'optionalAccess', _13 => _13.content, 'optionalAccess', _14 => _14.version]);
110
+ if (version) {
111
+ _chunkRIBK362Wjs.writeDebug.call(void 0, `Starting upload version ${version}`);
112
+ }
113
+ const basePath = options.path;
114
+ const files = await _glob.glob.call(void 0, _chunkRIBK362Wjs.joinPaths.call(void 0, basePath, "**/*"), {
115
+ ignore: "**/{*.stories.tsx,*.stories.ts,*.spec.tsx,*.spec.ts}"
116
+ });
117
+ const internalDependencies = await _chunkSIS5VQAUjs.getInternalDependencies.call(void 0,
118
+ context.projectName,
119
+ projectGraph
120
+ );
121
+ const dependencies = internalDependencies.filter(
122
+ (projectNode) => !projectNode.data.tags || projectNode.data.tags.every((tag) => tag.toLowerCase() !== "component")
123
+ ).reduce((ret, dep) => {
124
+ if (!ret[dep.name]) {
125
+ ret[dep.name] = "latest";
126
+ }
127
+ return ret;
128
+ }, _nullishCoalesce(_optionalChain([projectDetails, 'optionalAccess', _15 => _15.content, 'access', _16 => _16.dependencies]), () => ( {})));
129
+ const release = _nullishCoalesce(options.tag, () => ( _child_process.execSync.call(void 0, "npm config get tag").toString().trim()));
130
+ if (options.clean === true) {
131
+ _chunkRIBK362Wjs.writeDebug.call(void 0, `Clearing out existing items in ${bucketPath}`);
132
+ if (!isDryRun) {
133
+ const response = await client.send(
134
+ new (0, _clients3.ListObjectsCommand)({
135
+ Bucket: bucketId,
136
+ Prefix: !bucketPath || bucketPath === "/" ? void 0 : bucketPath
137
+ })
138
+ );
139
+ if (_optionalChain([response, 'optionalAccess', _17 => _17.Contents]) && response.Contents.length > 0) {
140
+ _chunkRIBK362Wjs.writeTrace.call(void 0,
141
+ `Deleting the following existing items from the R2 bucket path ${bucketPath}: ${response.Contents.map((item) => item.Key).join(", ")}`
142
+ );
143
+ await client.send(
144
+ new (0, _clients3.DeleteObjectsCommand)({
145
+ Bucket: bucketId,
146
+ Delete: {
147
+ Objects: response.Contents.map((item) => ({
148
+ Key: item.Key
149
+ })),
150
+ Quiet: false
151
+ }
152
+ })
153
+ );
154
+ } else {
155
+ _chunkRIBK362Wjs.writeDebug.call(void 0,
156
+ `No existing items to delete in the R2 bucket path ${bucketPath}`
157
+ );
158
+ }
159
+ } else {
160
+ _chunkRIBK362Wjs.writeWarning.call(void 0, "[Dry run]: Skipping R2 bucket clean.");
161
+ }
162
+ }
163
+ if (options.writeMetaJson === true) {
164
+ const meta = {
165
+ name: context.projectName,
166
+ version,
167
+ release,
168
+ description: _optionalChain([projectDetails, 'optionalAccess', _18 => _18.content, 'optionalAccess', _19 => _19.description]),
169
+ tags: _optionalChain([projectDetails, 'optionalAccess', _20 => _20.content, 'optionalAccess', _21 => _21.keywords]),
170
+ dependencies,
171
+ devDependencies: null,
172
+ internalDependencies: internalDependencies.filter(
173
+ (projectNode) => projectNode.data.tags && projectNode.data.tags.some(
174
+ (tag) => tag.toLowerCase() === "component"
175
+ )
176
+ ).map((dep) => dep.name)
177
+ };
178
+ if (_optionalChain([projectDetails, 'optionalAccess', _22 => _22.type]) === "package.json") {
179
+ meta.devDependencies = _optionalChain([projectDetails, 'optionalAccess', _23 => _23.content, 'optionalAccess', _24 => _24.devDependencies]);
180
+ }
181
+ await _chunkSIS5VQAUjs.uploadFile.call(void 0,
182
+ client,
183
+ bucketId,
184
+ bucketPath,
185
+ "meta.json",
186
+ version,
187
+ JSON.stringify(meta),
188
+ "application/json",
189
+ isDryRun
190
+ );
191
+ }
192
+ await Promise.all(
193
+ files.map(async (file) => {
194
+ if (_optionalChain([_fs.statSync.call(void 0, file, {
195
+ throwIfNoEntry: false
196
+ }), 'optionalAccess', _25 => _25.isFile, 'call', _26 => _26()])) {
197
+ const name = _chunkRIBK362Wjs.correctPaths.call(void 0, file).replace(_chunkRIBK362Wjs.correctPaths.call(void 0, basePath), "");
198
+ const type = _mimetypes2.default.lookup(name) || "application/octet-stream";
199
+ await _chunkSIS5VQAUjs.uploadFile.call(void 0,
200
+ client,
201
+ bucketId,
202
+ bucketPath,
203
+ name,
204
+ version,
205
+ await _promises.readFile.call(void 0, file, _chunkSIS5VQAUjs.getEncoding.call(void 0, type)),
206
+ type,
207
+ isDryRun
208
+ );
209
+ }
210
+ })
211
+ );
212
+ _chunkRIBK362Wjs.writeSuccess.call(void 0,
213
+ `Successfully uploaded the ${projectName} project to the Cloudflare R2 bucket.`,
214
+ config
215
+ );
216
+ return {
217
+ success: true
218
+ };
219
+ } catch (error) {
220
+ console.error("Failed to publish to Cloudflare R2 bucket");
221
+ console.error(error);
222
+ console.log("");
223
+ return {
224
+ success: false
225
+ };
226
+ }
227
+ }
228
+
229
+
230
+
231
+ exports.runExecutor = runExecutor;
@@ -0,0 +1,231 @@
1
+ import {
2
+ createCliOptions,
3
+ getPackageInfo
4
+ } from "./chunk-3MAI3FU2.mjs";
5
+ import {
6
+ findWorkspaceRoot,
7
+ getConfig
8
+ } from "./chunk-5QM4JUN3.mjs";
9
+ import {
10
+ createHttpHandler
11
+ } from "./chunk-SWYYMID7.mjs";
12
+ import {
13
+ getEncoding,
14
+ getInternalDependencies,
15
+ uploadFile
16
+ } from "./chunk-M3VAGNJQ.mjs";
17
+ import {
18
+ correctPaths,
19
+ joinPaths,
20
+ writeDebug,
21
+ writeSuccess,
22
+ writeTrace,
23
+ writeWarning
24
+ } from "./chunk-ZDABHOZ2.mjs";
25
+
26
+ // src/executors/r2-upload-publish/executor.ts
27
+ import {
28
+ DeleteObjectsCommand,
29
+ ListObjectsCommand,
30
+ S3Client
31
+ } from "@aws-sdk/client-s3";
32
+ import {
33
+ createProjectGraphAsync,
34
+ readCachedProjectGraph
35
+ } from "@nx/devkit";
36
+ import { glob } from "glob";
37
+ import mime from "mime-types";
38
+ import { execSync } from "node:child_process";
39
+ import { statSync } from "node:fs";
40
+ import { readFile } from "node:fs/promises";
41
+ async function runExecutor(options, context) {
42
+ const isDryRun = process.env.NX_DRY_RUN === "true" || options.dryRun || false;
43
+ if (!context.projectName) {
44
+ throw new Error("The executor requires a projectName.");
45
+ }
46
+ if (!options.path) {
47
+ throw new Error("The executor requires the `path` option to upload.");
48
+ }
49
+ console.info(
50
+ `\u{1F680} Running Storm Cloudflare Publish executor on the ${context.projectName} worker`
51
+ );
52
+ if (!context.projectName || !context.projectsConfigurations?.projects || !context.projectsConfigurations.projects[context.projectName] || !context.projectsConfigurations.projects[context.projectName]?.root) {
53
+ throw new Error("The executor requires projectsConfigurations.");
54
+ }
55
+ try {
56
+ const workspaceRoot = findWorkspaceRoot();
57
+ const config = await getConfig(workspaceRoot);
58
+ const projectName = context.projectsConfigurations.projects[context.projectName]?.name ?? context.projectName;
59
+ const projectDetails = getPackageInfo(
60
+ context.projectsConfigurations.projects[context.projectName]
61
+ );
62
+ const bucketId = options.bucketId;
63
+ const bucketPath = options.bucketPath || "/";
64
+ if (!bucketId) {
65
+ throw new Error("The executor requires a bucketId.");
66
+ }
67
+ const args = createCliOptions({ ...options });
68
+ if (isDryRun) {
69
+ args.push("--dry-run");
70
+ }
71
+ const cloudflareAccountId = process.env.CLOUDFLARE_ACCOUNT_ID || process.env.STORM_BOT_CLOUDFLARE_ACCOUNT;
72
+ if (!options?.registry && !cloudflareAccountId) {
73
+ throw new Error(
74
+ "The registry option and `CLOUDFLARE_ACCOUNT_ID` (or `STORM_BOT_CLOUDFLARE_ACCOUNT`) environment variable are not set. Please set one of these values to upload to the Cloudflare R2 bucket."
75
+ );
76
+ }
77
+ if (!process.env.STORM_BOT_ACCESS_KEY_ID && !process.env.ACCESS_KEY_ID && !process.env.CLOUDFLARE_ACCESS_KEY_ID && !process.env.AWS_ACCESS_KEY_ID || !process.env.STORM_BOT_SECRET_ACCESS_KEY && !process.env.CLOUDFLARE_SECRET_ACCESS_KEY && !process.env.SECRET_ACCESS_KEY && !process.env.AWS_SECRET_ACCESS_KEY) {
78
+ throw new Error(
79
+ "The `ACCESS_KEY_ID` (or `STORM_BOT_ACCESS_KEY_ID`) and `SECRET_ACCESS_KEY` (or `STORM_BOT_SECRET_ACCESS_KEY`) environment variables are not set. Please set these environment variables to upload to the Cloudflare R2 bucket."
80
+ );
81
+ }
82
+ const registry = options?.registry ? options.registry : `https://${cloudflareAccountId}.r2.cloudflarestorage.com`;
83
+ let projectGraph;
84
+ try {
85
+ projectGraph = readCachedProjectGraph();
86
+ } catch {
87
+ await createProjectGraphAsync();
88
+ projectGraph = readCachedProjectGraph();
89
+ }
90
+ if (!projectGraph) {
91
+ throw new Error(
92
+ "The executor failed because the project graph is not available. Please run the build command again."
93
+ );
94
+ }
95
+ writeDebug(
96
+ `Publishing ${context.projectName} to the ${bucketId} R2 Bucket (at ${registry})`
97
+ );
98
+ const client = new S3Client({
99
+ region: "auto",
100
+ endpoint: registry,
101
+ credentials: {
102
+ // eslint-disable-next-line @typescript-eslint/no-non-null-assertion
103
+ accessKeyId: process.env.STORM_BOT_ACCESS_KEY_ID || process.env.CLOUDFLARE_ACCESS_KEY_ID || process.env.AWS_ACCESS_KEY_ID || process.env.ACCESS_KEY_ID,
104
+ // eslint-disable-next-line @typescript-eslint/no-non-null-assertion
105
+ secretAccessKey: process.env.STORM_BOT_SECRET_ACCESS_KEY || process.env.CLOUDFLARE_SECRET_ACCESS_KEY || process.env.AWS_SECRET_ACCESS_KEY || process.env.SECRET_ACCESS_KEY
106
+ },
107
+ requestHandler: createHttpHandler()
108
+ });
109
+ const version = projectDetails?.content?.version;
110
+ if (version) {
111
+ writeDebug(`Starting upload version ${version}`);
112
+ }
113
+ const basePath = options.path;
114
+ const files = await glob(joinPaths(basePath, "**/*"), {
115
+ ignore: "**/{*.stories.tsx,*.stories.ts,*.spec.tsx,*.spec.ts}"
116
+ });
117
+ const internalDependencies = await getInternalDependencies(
118
+ context.projectName,
119
+ projectGraph
120
+ );
121
+ const dependencies = internalDependencies.filter(
122
+ (projectNode) => !projectNode.data.tags || projectNode.data.tags.every((tag) => tag.toLowerCase() !== "component")
123
+ ).reduce((ret, dep) => {
124
+ if (!ret[dep.name]) {
125
+ ret[dep.name] = "latest";
126
+ }
127
+ return ret;
128
+ }, projectDetails?.content.dependencies ?? {});
129
+ const release = options.tag ?? execSync("npm config get tag").toString().trim();
130
+ if (options.clean === true) {
131
+ writeDebug(`Clearing out existing items in ${bucketPath}`);
132
+ if (!isDryRun) {
133
+ const response = await client.send(
134
+ new ListObjectsCommand({
135
+ Bucket: bucketId,
136
+ Prefix: !bucketPath || bucketPath === "/" ? void 0 : bucketPath
137
+ })
138
+ );
139
+ if (response?.Contents && response.Contents.length > 0) {
140
+ writeTrace(
141
+ `Deleting the following existing items from the R2 bucket path ${bucketPath}: ${response.Contents.map((item) => item.Key).join(", ")}`
142
+ );
143
+ await client.send(
144
+ new DeleteObjectsCommand({
145
+ Bucket: bucketId,
146
+ Delete: {
147
+ Objects: response.Contents.map((item) => ({
148
+ Key: item.Key
149
+ })),
150
+ Quiet: false
151
+ }
152
+ })
153
+ );
154
+ } else {
155
+ writeDebug(
156
+ `No existing items to delete in the R2 bucket path ${bucketPath}`
157
+ );
158
+ }
159
+ } else {
160
+ writeWarning("[Dry run]: Skipping R2 bucket clean.");
161
+ }
162
+ }
163
+ if (options.writeMetaJson === true) {
164
+ const meta = {
165
+ name: context.projectName,
166
+ version,
167
+ release,
168
+ description: projectDetails?.content?.description,
169
+ tags: projectDetails?.content?.keywords,
170
+ dependencies,
171
+ devDependencies: null,
172
+ internalDependencies: internalDependencies.filter(
173
+ (projectNode) => projectNode.data.tags && projectNode.data.tags.some(
174
+ (tag) => tag.toLowerCase() === "component"
175
+ )
176
+ ).map((dep) => dep.name)
177
+ };
178
+ if (projectDetails?.type === "package.json") {
179
+ meta.devDependencies = projectDetails?.content?.devDependencies;
180
+ }
181
+ await uploadFile(
182
+ client,
183
+ bucketId,
184
+ bucketPath,
185
+ "meta.json",
186
+ version,
187
+ JSON.stringify(meta),
188
+ "application/json",
189
+ isDryRun
190
+ );
191
+ }
192
+ await Promise.all(
193
+ files.map(async (file) => {
194
+ if (statSync(file, {
195
+ throwIfNoEntry: false
196
+ })?.isFile()) {
197
+ const name = correctPaths(file).replace(correctPaths(basePath), "");
198
+ const type = mime.lookup(name) || "application/octet-stream";
199
+ await uploadFile(
200
+ client,
201
+ bucketId,
202
+ bucketPath,
203
+ name,
204
+ version,
205
+ await readFile(file, getEncoding(type)),
206
+ type,
207
+ isDryRun
208
+ );
209
+ }
210
+ })
211
+ );
212
+ writeSuccess(
213
+ `Successfully uploaded the ${projectName} project to the Cloudflare R2 bucket.`,
214
+ config
215
+ );
216
+ return {
217
+ success: true
218
+ };
219
+ } catch (error) {
220
+ console.error("Failed to publish to Cloudflare R2 bucket");
221
+ console.error(error);
222
+ console.log("");
223
+ return {
224
+ success: false
225
+ };
226
+ }
227
+ }
228
+
229
+ export {
230
+ runExecutor
231
+ };