@nxtedition/lib 19.1.3 → 19.1.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/errors.js +14 -4
- package/package.json +1 -1
- package/s3.js +23 -8
package/errors.js
CHANGED
|
@@ -42,12 +42,16 @@ export function parseError(error) {
|
|
|
42
42
|
const kSeen = Symbol('kSeen')
|
|
43
43
|
|
|
44
44
|
export function serializeError(error) {
|
|
45
|
+
return _serializeError(error, { depth: 0 })
|
|
46
|
+
}
|
|
47
|
+
|
|
48
|
+
function _serializeError(error, { depth }) {
|
|
45
49
|
if (!error) {
|
|
46
50
|
return null
|
|
47
51
|
}
|
|
48
52
|
|
|
49
53
|
if (typeof error === 'string') {
|
|
50
|
-
return [
|
|
54
|
+
return [_serializeError({ message: error }, { depth })]
|
|
51
55
|
}
|
|
52
56
|
|
|
53
57
|
if (Buffer.isBuffer(error)) {
|
|
@@ -55,13 +59,17 @@ export function serializeError(error) {
|
|
|
55
59
|
}
|
|
56
60
|
|
|
57
61
|
if (Array.isArray(error)) {
|
|
58
|
-
return error.map(
|
|
62
|
+
return error.map((x) => _serializeError(x, { depth })).filter(Boolean)
|
|
59
63
|
}
|
|
60
64
|
|
|
61
65
|
if (Object.prototype.hasOwnProperty.call(error, kSeen)) {
|
|
62
66
|
return null
|
|
63
67
|
}
|
|
64
68
|
|
|
69
|
+
if (depth > 8) {
|
|
70
|
+
return [{ messages: 'Maximum error serialization depth exceeded' }]
|
|
71
|
+
}
|
|
72
|
+
|
|
65
73
|
error[kSeen] = undefined
|
|
66
74
|
|
|
67
75
|
const type =
|
|
@@ -104,8 +112,10 @@ export function serializeError(error) {
|
|
|
104
112
|
signalCode = SIGNALS[signalCode] ?? signalCode
|
|
105
113
|
}
|
|
106
114
|
|
|
107
|
-
errors = Array.isArray(errors)
|
|
108
|
-
|
|
115
|
+
errors = Array.isArray(errors)
|
|
116
|
+
? errors.map((x) => _serializeError(x, { depth: depth + 1 })).filter(Boolean)
|
|
117
|
+
: undefined
|
|
118
|
+
cause = cause ? _serializeError(cause, { depth: depth + 1 }) : undefined
|
|
109
119
|
|
|
110
120
|
delete error[kSeen]
|
|
111
121
|
|
package/package.json
CHANGED
package/s3.js
CHANGED
|
@@ -4,7 +4,8 @@ import tp from 'node:timers/promises'
|
|
|
4
4
|
import AWS from '@aws-sdk/client-s3'
|
|
5
5
|
import PQueue from 'p-queue'
|
|
6
6
|
|
|
7
|
-
const
|
|
7
|
+
const QUEUE = new PQueue({ concurrency: 8 })
|
|
8
|
+
const MD5_HEX_EXPR = /^[A-F0-9]{32}$/i
|
|
8
9
|
|
|
9
10
|
/**
|
|
10
11
|
* Uploads a file to S3 using multipart upload.
|
|
@@ -14,6 +15,7 @@ const queue = new PQueue({ concurrency: 8 })
|
|
|
14
15
|
* @param {AbortSignal} options.signal - The signal to abort the upload.
|
|
15
16
|
* @param {Object} options.logger - The logger to use.
|
|
16
17
|
* @param {number} [options.partSize=16e6] - The size of each part in the multipart upload.
|
|
18
|
+
* @param {PQueue} [options.queue] - The queue to use for part uploads.
|
|
17
19
|
* @param {Object} options.params - The parameters for the upload.
|
|
18
20
|
* @param {Buffer|NodeJS.ReadStream} options.params.Body - The data to upload.
|
|
19
21
|
* @param {string} options.params.Key - The key of the object.
|
|
@@ -22,7 +24,14 @@ const queue = new PQueue({ concurrency: 8 })
|
|
|
22
24
|
* @param {number} [options.params.ContentLength] - The length of the object.
|
|
23
25
|
* @returns {Promise<Object>} The result of the upload.
|
|
24
26
|
*/
|
|
25
|
-
export async function upload({
|
|
27
|
+
export async function upload({
|
|
28
|
+
client: s3,
|
|
29
|
+
queue = QUEUE,
|
|
30
|
+
signal,
|
|
31
|
+
logger,
|
|
32
|
+
partSize = 16e6,
|
|
33
|
+
params,
|
|
34
|
+
}) {
|
|
26
35
|
if (s3 == null) {
|
|
27
36
|
throw new Error('Invalid client')
|
|
28
37
|
}
|
|
@@ -35,9 +44,18 @@ export async function upload({ client: s3, signal, logger, partSize = 16e6, para
|
|
|
35
44
|
throw new Error('Invalid params')
|
|
36
45
|
}
|
|
37
46
|
|
|
38
|
-
const { Body, Key, Bucket, ContentMD5, ContentLength } = params
|
|
47
|
+
const { Body, Key, Bucket, ContentMD5, ContentLength } = params
|
|
39
48
|
|
|
40
|
-
|
|
49
|
+
const size = ContentLength != null ? Number(ContentLength) : null
|
|
50
|
+
|
|
51
|
+
if (size != null && (!Number.isFinite(size) || size < 0)) {
|
|
52
|
+
throw new Error('Invalid params.ContentLength')
|
|
53
|
+
}
|
|
54
|
+
|
|
55
|
+
const hash = ContentMD5 != null ? Buffer.from(ContentMD5, 'base64').toString('hex') : null
|
|
56
|
+
if (hash != null && !MD5_HEX_EXPR.test(hash)) {
|
|
57
|
+
throw new Error('Invalid params.ContentMD5')
|
|
58
|
+
}
|
|
41
59
|
|
|
42
60
|
const promises = []
|
|
43
61
|
|
|
@@ -123,7 +141,7 @@ export async function upload({ client: s3, signal, logger, partSize = 16e6, para
|
|
|
123
141
|
} catch (err) {
|
|
124
142
|
logger?.warn({ err }, 'part upload failed')
|
|
125
143
|
|
|
126
|
-
if (retryCount <
|
|
144
|
+
if (retryCount < 5) {
|
|
127
145
|
logger?.debug({ retryCount }, 'part upload retry')
|
|
128
146
|
await tp.setTimeout(1e3, undefined, { signal: uploader.signal })
|
|
129
147
|
} else {
|
|
@@ -196,9 +214,6 @@ export async function upload({ client: s3, signal, logger, partSize = 16e6, para
|
|
|
196
214
|
parts,
|
|
197
215
|
}
|
|
198
216
|
|
|
199
|
-
const size = ContentLength != null ? Number(ContentLength) : null
|
|
200
|
-
const hash = ContentMD5 ? Buffer.from(ContentMD5, 'base64').toString('hex') : null
|
|
201
|
-
|
|
202
217
|
if (size != null && size !== result.size) {
|
|
203
218
|
throw new Error(`Expected size ${size} but got ${result.size}`)
|
|
204
219
|
}
|