overtake 1.3.2 → 2.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +12 -15
- package/bin/overtake.js +1 -1
- package/build/executor.d.ts +8 -3
- package/build/index.d.ts +10 -11
- package/build/reporter.d.ts +10 -2
- package/build/runner.d.ts +1 -1
- package/build/types.d.ts +7 -7
- package/build/utils.d.ts +3 -17
- package/package.json +8 -26
- package/src/__tests__/assert-no-closure.ts +135 -0
- package/src/__tests__/benchmark-execute.ts +48 -0
- package/src/cli.ts +139 -144
- package/src/executor.ts +59 -24
- package/src/index.ts +135 -68
- package/src/reporter.ts +77 -125
- package/src/runner.ts +28 -25
- package/src/types.ts +9 -9
- package/src/utils.ts +62 -46
- package/src/worker.ts +13 -12
- package/tsconfig.json +3 -1
- package/build/cli.cjs +0 -179
- package/build/cli.cjs.map +0 -1
- package/build/cli.js +0 -134
- package/build/cli.js.map +0 -1
- package/build/executor.cjs +0 -116
- package/build/executor.cjs.map +0 -1
- package/build/executor.js +0 -106
- package/build/executor.js.map +0 -1
- package/build/gc-watcher.cjs +0 -30
- package/build/gc-watcher.cjs.map +0 -1
- package/build/gc-watcher.js +0 -20
- package/build/gc-watcher.js.map +0 -1
- package/build/index.cjs +0 -400
- package/build/index.cjs.map +0 -1
- package/build/index.js +0 -335
- package/build/index.js.map +0 -1
- package/build/reporter.cjs +0 -364
- package/build/reporter.cjs.map +0 -1
- package/build/reporter.js +0 -346
- package/build/reporter.js.map +0 -1
- package/build/runner.cjs +0 -528
- package/build/runner.cjs.map +0 -1
- package/build/runner.js +0 -518
- package/build/runner.js.map +0 -1
- package/build/types.cjs +0 -66
- package/build/types.cjs.map +0 -1
- package/build/types.js +0 -33
- package/build/types.js.map +0 -1
- package/build/utils.cjs +0 -121
- package/build/utils.cjs.map +0 -1
- package/build/utils.js +0 -85
- package/build/utils.js.map +0 -1
- package/build/worker.cjs +0 -158
- package/build/worker.cjs.map +0 -1
- package/build/worker.js +0 -113
- package/build/worker.js.map +0 -1
package/build/runner.js
DELETED
|
@@ -1,518 +0,0 @@
|
|
|
1
|
-
import { performance, PerformanceObserver } from 'node:perf_hooks';
|
|
2
|
-
import { Control, DURATION_SCALE } from "./types.js";
|
|
3
|
-
import { GCWatcher } from "./gc-watcher.js";
|
|
4
|
-
const COMPLETE_VALUE = 100_00;
|
|
5
|
-
const hr = process.hrtime.bigint.bind(process.hrtime);
|
|
6
|
-
const sink = new Int32Array(new SharedArrayBuffer(Int32Array.BYTES_PER_ELEMENT));
|
|
7
|
-
const consume = (value)=>{
|
|
8
|
-
let payload = 0;
|
|
9
|
-
switch(typeof value){
|
|
10
|
-
case 'number':
|
|
11
|
-
payload = Number.isFinite(value) ? Math.trunc(value) : 0;
|
|
12
|
-
break;
|
|
13
|
-
case 'bigint':
|
|
14
|
-
payload = Number(value & 0xffff_ffffn);
|
|
15
|
-
break;
|
|
16
|
-
case 'string':
|
|
17
|
-
payload = value.length;
|
|
18
|
-
break;
|
|
19
|
-
case 'boolean':
|
|
20
|
-
payload = value ? 1 : 0;
|
|
21
|
-
break;
|
|
22
|
-
case 'object':
|
|
23
|
-
payload = value === null ? 0 : 1;
|
|
24
|
-
break;
|
|
25
|
-
case 'function':
|
|
26
|
-
payload = 1;
|
|
27
|
-
break;
|
|
28
|
-
default:
|
|
29
|
-
payload = -1;
|
|
30
|
-
}
|
|
31
|
-
Atomics.xor(sink, 0, payload);
|
|
32
|
-
};
|
|
33
|
-
const runSync = (run, overhead)=>{
|
|
34
|
-
return (...args)=>{
|
|
35
|
-
const start = hr();
|
|
36
|
-
const result = run(...args);
|
|
37
|
-
consume(result);
|
|
38
|
-
const duration = hr() - start;
|
|
39
|
-
return duration > overhead ? duration - overhead : 0n;
|
|
40
|
-
};
|
|
41
|
-
};
|
|
42
|
-
const runAsync = (run)=>{
|
|
43
|
-
return async (...args)=>{
|
|
44
|
-
const start = hr();
|
|
45
|
-
const result = await run(...args);
|
|
46
|
-
consume(result);
|
|
47
|
-
return hr() - start;
|
|
48
|
-
};
|
|
49
|
-
};
|
|
50
|
-
const isThenable = (value)=>{
|
|
51
|
-
return value !== null && (typeof value === 'object' || typeof value === 'function') && typeof value.then === 'function';
|
|
52
|
-
};
|
|
53
|
-
const TARGET_SAMPLE_NS = 1_000_000n;
|
|
54
|
-
const MAX_BATCH = 1_048_576;
|
|
55
|
-
const PROGRESS_STRIDE = 16;
|
|
56
|
-
const GC_STRIDE = 32;
|
|
57
|
-
const OUTLIER_MULTIPLIER = 4;
|
|
58
|
-
const OUTLIER_IQR_MULTIPLIER = 3;
|
|
59
|
-
const OUTLIER_WINDOW = 64;
|
|
60
|
-
const OUTLIER_ABS_THRESHOLD = 10_000_000;
|
|
61
|
-
const BASELINE_SAMPLES = 16;
|
|
62
|
-
const OUTLIER_SCRATCH = new Float64Array(OUTLIER_WINDOW);
|
|
63
|
-
const measureTimerOverhead = ()=>{
|
|
64
|
-
let total = 0n;
|
|
65
|
-
for(let i = 0; i < BASELINE_SAMPLES; i++){
|
|
66
|
-
const start = hr();
|
|
67
|
-
consume(0);
|
|
68
|
-
total += hr() - start;
|
|
69
|
-
}
|
|
70
|
-
return total / BigInt(BASELINE_SAMPLES);
|
|
71
|
-
};
|
|
72
|
-
const collectSample = async ({ batchSize, run, runRaw, runIsAsync, pre, preIsAsync, post, postIsAsync, context, data, nextNonce })=>{
|
|
73
|
-
const canBatchTime = !runIsAsync && !pre && !post;
|
|
74
|
-
if (canBatchTime) {
|
|
75
|
-
const batchStart = hr();
|
|
76
|
-
if (nextNonce) {
|
|
77
|
-
for(let b = 0; b < batchSize; b++){
|
|
78
|
-
consume(runRaw(context, data, nextNonce()));
|
|
79
|
-
}
|
|
80
|
-
} else {
|
|
81
|
-
for(let b = 0; b < batchSize; b++){
|
|
82
|
-
consume(runRaw(context, data));
|
|
83
|
-
}
|
|
84
|
-
}
|
|
85
|
-
return (hr() - batchStart) * DURATION_SCALE / BigInt(batchSize);
|
|
86
|
-
}
|
|
87
|
-
let sampleDuration = 0n;
|
|
88
|
-
for(let b = 0; b < batchSize; b++){
|
|
89
|
-
if (pre) {
|
|
90
|
-
if (preIsAsync) {
|
|
91
|
-
await pre(context, data);
|
|
92
|
-
} else {
|
|
93
|
-
pre(context, data);
|
|
94
|
-
}
|
|
95
|
-
}
|
|
96
|
-
if (runIsAsync) {
|
|
97
|
-
const runAsyncFn = run;
|
|
98
|
-
const duration = nextNonce ? await runAsyncFn(context, data, nextNonce()) : await runAsyncFn(context, data);
|
|
99
|
-
sampleDuration += duration;
|
|
100
|
-
} else {
|
|
101
|
-
const runSyncFn = run;
|
|
102
|
-
const duration = nextNonce ? runSyncFn(context, data, nextNonce()) : runSyncFn(context, data);
|
|
103
|
-
sampleDuration += duration;
|
|
104
|
-
}
|
|
105
|
-
if (post) {
|
|
106
|
-
if (postIsAsync) {
|
|
107
|
-
await post(context, data);
|
|
108
|
-
} else {
|
|
109
|
-
post(context, data);
|
|
110
|
-
}
|
|
111
|
-
}
|
|
112
|
-
}
|
|
113
|
-
return sampleDuration * DURATION_SCALE / BigInt(batchSize);
|
|
114
|
-
};
|
|
115
|
-
const tuneParameters = async ({ initialBatch, run, runRaw, runIsAsync, pre, preIsAsync, post, postIsAsync, context, data, minCycles, relThreshold, maxCycles, nextNonce })=>{
|
|
116
|
-
let batchSize = initialBatch;
|
|
117
|
-
let bestCv = Number.POSITIVE_INFINITY;
|
|
118
|
-
let bestBatch = batchSize;
|
|
119
|
-
for(let attempt = 0; attempt < 3; attempt++){
|
|
120
|
-
const samples = [];
|
|
121
|
-
const sampleCount = Math.min(8, maxCycles);
|
|
122
|
-
for(let s = 0; s < sampleCount; s++){
|
|
123
|
-
const duration = await collectSample({
|
|
124
|
-
batchSize,
|
|
125
|
-
run,
|
|
126
|
-
runRaw,
|
|
127
|
-
runIsAsync,
|
|
128
|
-
pre,
|
|
129
|
-
preIsAsync,
|
|
130
|
-
post,
|
|
131
|
-
postIsAsync,
|
|
132
|
-
context,
|
|
133
|
-
data,
|
|
134
|
-
nextNonce
|
|
135
|
-
});
|
|
136
|
-
samples.push(Number(duration));
|
|
137
|
-
}
|
|
138
|
-
const mean = samples.reduce((acc, v)=>acc + v, 0) / samples.length;
|
|
139
|
-
const variance = samples.reduce((acc, v)=>acc + (v - mean) * (v - mean), 0) / Math.max(1, samples.length - 1);
|
|
140
|
-
const stddev = Math.sqrt(variance);
|
|
141
|
-
const cv = mean === 0 ? Number.POSITIVE_INFINITY : stddev / mean;
|
|
142
|
-
if (cv < bestCv) {
|
|
143
|
-
bestCv = cv;
|
|
144
|
-
bestBatch = batchSize;
|
|
145
|
-
}
|
|
146
|
-
if (cv <= relThreshold || batchSize >= MAX_BATCH) {
|
|
147
|
-
break;
|
|
148
|
-
}
|
|
149
|
-
batchSize = Math.min(MAX_BATCH, batchSize * 2);
|
|
150
|
-
}
|
|
151
|
-
const tunedRel = bestCv < relThreshold ? Math.max(bestCv * 1.5, relThreshold * 0.5) : relThreshold;
|
|
152
|
-
const tunedMin = Math.min(maxCycles, Math.max(minCycles, Math.ceil(minCycles * Math.max(1, bestCv / (relThreshold || 1e-6)))));
|
|
153
|
-
return {
|
|
154
|
-
batchSize: bestBatch,
|
|
155
|
-
relThreshold: tunedRel,
|
|
156
|
-
minCycles: tunedMin
|
|
157
|
-
};
|
|
158
|
-
};
|
|
159
|
-
const createGCTracker = ()=>{
|
|
160
|
-
if (process.env.OVERTAKE_GC_OBSERVER !== '1') {
|
|
161
|
-
return null;
|
|
162
|
-
}
|
|
163
|
-
if (typeof PerformanceObserver === 'undefined') {
|
|
164
|
-
return null;
|
|
165
|
-
}
|
|
166
|
-
const events = [];
|
|
167
|
-
const observer = new PerformanceObserver((list)=>{
|
|
168
|
-
for (const entry of list.getEntries()){
|
|
169
|
-
events.push({
|
|
170
|
-
start: entry.startTime,
|
|
171
|
-
end: entry.startTime + entry.duration
|
|
172
|
-
});
|
|
173
|
-
}
|
|
174
|
-
});
|
|
175
|
-
try {
|
|
176
|
-
observer.observe({
|
|
177
|
-
entryTypes: [
|
|
178
|
-
'gc'
|
|
179
|
-
]
|
|
180
|
-
});
|
|
181
|
-
} catch {
|
|
182
|
-
return null;
|
|
183
|
-
}
|
|
184
|
-
const overlaps = (start, end)=>{
|
|
185
|
-
let noisy = false;
|
|
186
|
-
for(let i = events.length - 1; i >= 0; i--){
|
|
187
|
-
const event = events[i];
|
|
188
|
-
if (event.end < start - 5_000) {
|
|
189
|
-
events.splice(i, 1);
|
|
190
|
-
continue;
|
|
191
|
-
}
|
|
192
|
-
if (event.start <= end && event.end >= start) {
|
|
193
|
-
noisy = true;
|
|
194
|
-
}
|
|
195
|
-
}
|
|
196
|
-
return noisy;
|
|
197
|
-
};
|
|
198
|
-
const dispose = ()=>observer.disconnect();
|
|
199
|
-
return {
|
|
200
|
-
overlaps,
|
|
201
|
-
dispose
|
|
202
|
-
};
|
|
203
|
-
};
|
|
204
|
-
const pushWindow = (arr, value, cap)=>{
|
|
205
|
-
if (arr.length === cap) {
|
|
206
|
-
arr.shift();
|
|
207
|
-
}
|
|
208
|
-
arr.push(value);
|
|
209
|
-
};
|
|
210
|
-
const medianAndIqr = (arr)=>{
|
|
211
|
-
if (arr.length === 0) return {
|
|
212
|
-
median: 0,
|
|
213
|
-
iqr: 0
|
|
214
|
-
};
|
|
215
|
-
for(let i = 0; i < arr.length; i++){
|
|
216
|
-
OUTLIER_SCRATCH[i] = arr[i];
|
|
217
|
-
}
|
|
218
|
-
const view = OUTLIER_SCRATCH.subarray(0, arr.length);
|
|
219
|
-
view.sort();
|
|
220
|
-
const mid = Math.floor(view.length / 2);
|
|
221
|
-
const median = view.length % 2 === 0 ? (view[mid - 1] + view[mid]) / 2 : view[mid];
|
|
222
|
-
const q1Idx = Math.floor(view.length * 0.25);
|
|
223
|
-
const q3Idx = Math.floor(view.length * 0.75);
|
|
224
|
-
const q1 = view[q1Idx];
|
|
225
|
-
const q3 = view[q3Idx];
|
|
226
|
-
return {
|
|
227
|
-
median,
|
|
228
|
-
iqr: q3 - q1
|
|
229
|
-
};
|
|
230
|
-
};
|
|
231
|
-
const windowCv = (arr)=>{
|
|
232
|
-
if (arr.length < 2) return Number.POSITIVE_INFINITY;
|
|
233
|
-
const mean = arr.reduce((a, v)=>a + v, 0) / arr.length;
|
|
234
|
-
const variance = arr.reduce((a, v)=>a + (v - mean) * (v - mean), 0) / (arr.length - 1);
|
|
235
|
-
const stddev = Math.sqrt(variance);
|
|
236
|
-
return mean === 0 ? Number.POSITIVE_INFINITY : stddev / mean;
|
|
237
|
-
};
|
|
238
|
-
export const benchmark = async ({ setup, teardown, pre, run: runRaw, post, data, warmupCycles, minCycles, absThreshold, relThreshold, gcObserver = false, durationsSAB, controlSAB })=>{
|
|
239
|
-
const durations = new BigUint64Array(durationsSAB);
|
|
240
|
-
const control = new Int32Array(controlSAB);
|
|
241
|
-
control[Control.INDEX] = 0;
|
|
242
|
-
control[Control.PROGRESS] = 0;
|
|
243
|
-
control[Control.COMPLETE] = 255;
|
|
244
|
-
control[Control.HEAP_USED] = 0;
|
|
245
|
-
const context = await setup?.();
|
|
246
|
-
const heapBefore = process.memoryUsage().heapUsed;
|
|
247
|
-
const input = data;
|
|
248
|
-
const maxCycles = durations.length;
|
|
249
|
-
const gcWatcher = gcObserver ? new GCWatcher() : null;
|
|
250
|
-
const gcTracker = gcObserver ? createGCTracker() : null;
|
|
251
|
-
try {
|
|
252
|
-
let preIsAsync = false;
|
|
253
|
-
if (pre) {
|
|
254
|
-
const preResult = pre(context, input);
|
|
255
|
-
preIsAsync = isThenable(preResult);
|
|
256
|
-
if (preIsAsync) {
|
|
257
|
-
await preResult;
|
|
258
|
-
}
|
|
259
|
-
}
|
|
260
|
-
const probeStart = hr();
|
|
261
|
-
const probeResult = runRaw(context, input);
|
|
262
|
-
const runIsAsync = isThenable(probeResult);
|
|
263
|
-
if (runIsAsync) {
|
|
264
|
-
const resolved = await probeResult;
|
|
265
|
-
consume(resolved);
|
|
266
|
-
} else {
|
|
267
|
-
consume(probeResult);
|
|
268
|
-
}
|
|
269
|
-
const durationProbeRaw = hr() - probeStart;
|
|
270
|
-
let postIsAsync = false;
|
|
271
|
-
if (post) {
|
|
272
|
-
const postResult = post(context, input);
|
|
273
|
-
postIsAsync = isThenable(postResult);
|
|
274
|
-
if (postIsAsync) {
|
|
275
|
-
await postResult;
|
|
276
|
-
}
|
|
277
|
-
}
|
|
278
|
-
const timerOverhead = runIsAsync ? 0n : measureTimerOverhead();
|
|
279
|
-
let durationProbe = runIsAsync ? durationProbeRaw : durationProbeRaw > timerOverhead ? durationProbeRaw - timerOverhead : 0n;
|
|
280
|
-
const shouldPerturbInput = process.env.OVERTAKE_PERTURB_INPUT === '1';
|
|
281
|
-
let nonce = 0;
|
|
282
|
-
const nextNonce = shouldPerturbInput ? ()=>{
|
|
283
|
-
nonce = nonce + 1 | 0;
|
|
284
|
-
return nonce;
|
|
285
|
-
} : null;
|
|
286
|
-
if (!runIsAsync && !pre && !post) {
|
|
287
|
-
const PROBE_TIME_LIMIT_NS = 1_000_000_000n;
|
|
288
|
-
const INITIAL_PROBE_SIZE = 10;
|
|
289
|
-
const MAX_PROBE_SIZE = 10_000;
|
|
290
|
-
const initialStart = hr();
|
|
291
|
-
if (nextNonce) {
|
|
292
|
-
for(let i = 0; i < INITIAL_PROBE_SIZE; i++){
|
|
293
|
-
consume(runRaw(context, input, nextNonce()));
|
|
294
|
-
}
|
|
295
|
-
} else {
|
|
296
|
-
for(let i = 0; i < INITIAL_PROBE_SIZE; i++){
|
|
297
|
-
consume(runRaw(context, input));
|
|
298
|
-
}
|
|
299
|
-
}
|
|
300
|
-
const initialDuration = hr() - initialStart;
|
|
301
|
-
const estimatedPerOp = initialDuration / BigInt(INITIAL_PROBE_SIZE);
|
|
302
|
-
const remainingBudget = PROBE_TIME_LIMIT_NS - initialDuration;
|
|
303
|
-
const additionalIterations = estimatedPerOp > 0n ? Number(remainingBudget / estimatedPerOp) : MAX_PROBE_SIZE - INITIAL_PROBE_SIZE;
|
|
304
|
-
const cappedAdditional = Math.min(Math.max(0, additionalIterations), MAX_PROBE_SIZE - INITIAL_PROBE_SIZE);
|
|
305
|
-
let totalIterations = INITIAL_PROBE_SIZE;
|
|
306
|
-
if (cappedAdditional > 0) {
|
|
307
|
-
if (nextNonce) {
|
|
308
|
-
for(let i = 0; i < cappedAdditional; i++){
|
|
309
|
-
consume(runRaw(context, input, nextNonce()));
|
|
310
|
-
}
|
|
311
|
-
} else {
|
|
312
|
-
for(let i = 0; i < cappedAdditional; i++){
|
|
313
|
-
consume(runRaw(context, input));
|
|
314
|
-
}
|
|
315
|
-
}
|
|
316
|
-
totalIterations += cappedAdditional;
|
|
317
|
-
}
|
|
318
|
-
durationProbe = (hr() - initialStart) / BigInt(totalIterations);
|
|
319
|
-
}
|
|
320
|
-
const runTimedSync = runIsAsync ? null : runSync(runRaw, timerOverhead);
|
|
321
|
-
const runTimedAsync = runIsAsync ? runAsync(runRaw) : null;
|
|
322
|
-
const run = runIsAsync ? runTimedAsync : runTimedSync;
|
|
323
|
-
const runOnceSync = runIsAsync ? null : nextNonce ? (ctx, dataValue)=>runTimedSync(ctx, dataValue, nextNonce()) : runTimedSync;
|
|
324
|
-
const runOnceAsync = runIsAsync ? nextNonce ? (ctx, dataValue)=>runTimedAsync(ctx, dataValue, nextNonce()) : runTimedAsync : null;
|
|
325
|
-
const preSync = preIsAsync ? null : pre;
|
|
326
|
-
const preAsync = preIsAsync ? pre : null;
|
|
327
|
-
const postSync = postIsAsync ? null : post;
|
|
328
|
-
const postAsync = postIsAsync ? post : null;
|
|
329
|
-
const durationPerRun = durationProbe === 0n ? 1n : durationProbe;
|
|
330
|
-
const suggestedBatch = Number(TARGET_SAMPLE_NS / durationPerRun);
|
|
331
|
-
const minBatchForFastOps = durationProbe < 100n ? 100_000 : 1;
|
|
332
|
-
const initialBatchSize = Math.min(MAX_BATCH, Math.max(minBatchForFastOps, suggestedBatch));
|
|
333
|
-
const tuned = await tuneParameters({
|
|
334
|
-
initialBatch: initialBatchSize,
|
|
335
|
-
run,
|
|
336
|
-
runRaw,
|
|
337
|
-
runIsAsync,
|
|
338
|
-
pre,
|
|
339
|
-
preIsAsync,
|
|
340
|
-
post,
|
|
341
|
-
postIsAsync,
|
|
342
|
-
context,
|
|
343
|
-
data: input,
|
|
344
|
-
minCycles,
|
|
345
|
-
relThreshold,
|
|
346
|
-
maxCycles,
|
|
347
|
-
nextNonce
|
|
348
|
-
});
|
|
349
|
-
let batchSize = tuned.batchSize;
|
|
350
|
-
minCycles = tuned.minCycles;
|
|
351
|
-
relThreshold = tuned.relThreshold;
|
|
352
|
-
const warmupStart = performance.now();
|
|
353
|
-
let warmupRemaining = warmupCycles;
|
|
354
|
-
const warmupWindow = [];
|
|
355
|
-
const warmupCap = Math.max(warmupCycles, Math.min(maxCycles, warmupCycles * 4 || 1000));
|
|
356
|
-
const canBatchTime = !runIsAsync && !preSync && !preAsync && !postSync && !postAsync;
|
|
357
|
-
const runWarmup = async ()=>{
|
|
358
|
-
if (canBatchTime) {
|
|
359
|
-
const batchStart = hr();
|
|
360
|
-
if (nextNonce) {
|
|
361
|
-
for(let b = 0; b < batchSize; b++){
|
|
362
|
-
consume(runRaw(context, input, nextNonce()));
|
|
363
|
-
}
|
|
364
|
-
} else {
|
|
365
|
-
for(let b = 0; b < batchSize; b++){
|
|
366
|
-
consume(runRaw(context, input));
|
|
367
|
-
}
|
|
368
|
-
}
|
|
369
|
-
return (hr() - batchStart) * DURATION_SCALE / BigInt(batchSize);
|
|
370
|
-
}
|
|
371
|
-
if (preSync) {
|
|
372
|
-
preSync(context, input);
|
|
373
|
-
} else if (preAsync) {
|
|
374
|
-
await preAsync(context, input);
|
|
375
|
-
}
|
|
376
|
-
const duration = runIsAsync ? await runOnceAsync(context, input) : runOnceSync(context, input);
|
|
377
|
-
if (postSync) {
|
|
378
|
-
postSync(context, input);
|
|
379
|
-
} else if (postAsync) {
|
|
380
|
-
await postAsync(context, input);
|
|
381
|
-
}
|
|
382
|
-
return duration * DURATION_SCALE;
|
|
383
|
-
};
|
|
384
|
-
while(performance.now() - warmupStart < 1_000 && warmupRemaining > 0){
|
|
385
|
-
const duration = await runWarmup();
|
|
386
|
-
pushWindow(warmupWindow, Number(duration), warmupCap);
|
|
387
|
-
warmupRemaining--;
|
|
388
|
-
}
|
|
389
|
-
let warmupDone = 0;
|
|
390
|
-
while(warmupDone < warmupRemaining){
|
|
391
|
-
const duration = await runWarmup();
|
|
392
|
-
pushWindow(warmupWindow, Number(duration), warmupCap);
|
|
393
|
-
warmupDone++;
|
|
394
|
-
if (global.gc && warmupDone % GC_STRIDE === 0) {
|
|
395
|
-
global.gc();
|
|
396
|
-
}
|
|
397
|
-
}
|
|
398
|
-
while(warmupWindow.length >= 8 && warmupWindow.length < warmupCap){
|
|
399
|
-
const cv = windowCv(warmupWindow);
|
|
400
|
-
if (cv <= relThreshold * 2) {
|
|
401
|
-
break;
|
|
402
|
-
}
|
|
403
|
-
const duration = await runWarmup();
|
|
404
|
-
pushWindow(warmupWindow, Number(duration), warmupCap);
|
|
405
|
-
}
|
|
406
|
-
let i = 0;
|
|
407
|
-
let mean = 0n;
|
|
408
|
-
let m2 = 0n;
|
|
409
|
-
const outlierWindow = [];
|
|
410
|
-
let skipped = 0;
|
|
411
|
-
const maxSkipped = maxCycles * 10;
|
|
412
|
-
let disableFiltering = false;
|
|
413
|
-
while(true){
|
|
414
|
-
if (i >= maxCycles) break;
|
|
415
|
-
if (!disableFiltering && skipped >= maxSkipped) {
|
|
416
|
-
console.error(`Warning: ${skipped} samples skipped due to noise/outlier detection. ` + `Disabling filtering for remaining samples. Results may have higher variance.`);
|
|
417
|
-
disableFiltering = true;
|
|
418
|
-
}
|
|
419
|
-
if (global.gc && i > 0 && i % GC_STRIDE === 0) {
|
|
420
|
-
global.gc();
|
|
421
|
-
}
|
|
422
|
-
const gcMarker = gcWatcher?.start();
|
|
423
|
-
const sampleStart = performance.now();
|
|
424
|
-
let sampleDuration = 0n;
|
|
425
|
-
if (canBatchTime) {
|
|
426
|
-
const batchStart = hr();
|
|
427
|
-
if (nextNonce) {
|
|
428
|
-
for(let b = 0; b < batchSize; b++){
|
|
429
|
-
consume(runRaw(context, input, nextNonce()));
|
|
430
|
-
}
|
|
431
|
-
} else {
|
|
432
|
-
for(let b = 0; b < batchSize; b++){
|
|
433
|
-
consume(runRaw(context, input));
|
|
434
|
-
}
|
|
435
|
-
}
|
|
436
|
-
const batchDuration = hr() - batchStart;
|
|
437
|
-
sampleDuration = batchDuration * DURATION_SCALE / BigInt(batchSize);
|
|
438
|
-
} else {
|
|
439
|
-
for(let b = 0; b < batchSize; b++){
|
|
440
|
-
if (preSync) {
|
|
441
|
-
preSync(context, input);
|
|
442
|
-
} else if (preAsync) {
|
|
443
|
-
await preAsync(context, input);
|
|
444
|
-
}
|
|
445
|
-
const duration = runIsAsync ? await runOnceAsync(context, input) : runOnceSync(context, input);
|
|
446
|
-
sampleDuration += duration;
|
|
447
|
-
if (postSync) {
|
|
448
|
-
postSync(context, input);
|
|
449
|
-
} else if (postAsync) {
|
|
450
|
-
await postAsync(context, input);
|
|
451
|
-
}
|
|
452
|
-
}
|
|
453
|
-
sampleDuration = sampleDuration * DURATION_SCALE / BigInt(batchSize);
|
|
454
|
-
}
|
|
455
|
-
const sampleEnd = performance.now();
|
|
456
|
-
if (!disableFiltering) {
|
|
457
|
-
const gcNoise = (gcMarker ? gcWatcher.seen(gcMarker) : false) || (gcTracker?.overlaps(sampleStart, sampleEnd) ?? false);
|
|
458
|
-
if (gcNoise) {
|
|
459
|
-
skipped++;
|
|
460
|
-
continue;
|
|
461
|
-
}
|
|
462
|
-
}
|
|
463
|
-
const durationNumber = Number(sampleDuration);
|
|
464
|
-
pushWindow(outlierWindow, durationNumber, OUTLIER_WINDOW);
|
|
465
|
-
if (!disableFiltering) {
|
|
466
|
-
const { median, iqr } = medianAndIqr(outlierWindow);
|
|
467
|
-
const maxAllowed = median + OUTLIER_IQR_MULTIPLIER * iqr || Number.POSITIVE_INFINITY;
|
|
468
|
-
if (outlierWindow.length >= 8 && durationNumber > maxAllowed && durationNumber - median > OUTLIER_ABS_THRESHOLD) {
|
|
469
|
-
skipped++;
|
|
470
|
-
continue;
|
|
471
|
-
}
|
|
472
|
-
const meanNumber = Number(mean);
|
|
473
|
-
if (i >= 8 && meanNumber > 0 && durationNumber > OUTLIER_MULTIPLIER * meanNumber && durationNumber - meanNumber > OUTLIER_ABS_THRESHOLD) {
|
|
474
|
-
skipped++;
|
|
475
|
-
continue;
|
|
476
|
-
}
|
|
477
|
-
}
|
|
478
|
-
durations[i++] = sampleDuration;
|
|
479
|
-
const delta = sampleDuration - mean;
|
|
480
|
-
mean += delta / BigInt(i);
|
|
481
|
-
m2 += delta * (sampleDuration - mean);
|
|
482
|
-
const progress = i / maxCycles * COMPLETE_VALUE;
|
|
483
|
-
if (i % PROGRESS_STRIDE === 0) {
|
|
484
|
-
control[Control.PROGRESS] = progress;
|
|
485
|
-
}
|
|
486
|
-
if (i >= minCycles) {
|
|
487
|
-
const variance = Number(m2) / (i - 1);
|
|
488
|
-
const stddev = Math.sqrt(variance);
|
|
489
|
-
if (stddev <= Number(absThreshold)) {
|
|
490
|
-
break;
|
|
491
|
-
}
|
|
492
|
-
const meanNum = Number(mean);
|
|
493
|
-
const cov = stddev / (meanNum || 1);
|
|
494
|
-
if (cov <= relThreshold) {
|
|
495
|
-
break;
|
|
496
|
-
}
|
|
497
|
-
}
|
|
498
|
-
}
|
|
499
|
-
control[Control.INDEX] = i;
|
|
500
|
-
control[Control.COMPLETE] = 0;
|
|
501
|
-
const heapAfter = process.memoryUsage().heapUsed;
|
|
502
|
-
control[Control.HEAP_USED] = Math.max(0, Math.round((heapAfter - heapBefore) / 1024));
|
|
503
|
-
} catch (e) {
|
|
504
|
-
console.error(e && typeof e === 'object' && 'stack' in e ? e.stack : e);
|
|
505
|
-
control[Control.COMPLETE] = 1;
|
|
506
|
-
} finally{
|
|
507
|
-
gcTracker?.dispose?.();
|
|
508
|
-
try {
|
|
509
|
-
await teardown?.(context);
|
|
510
|
-
} catch (e) {
|
|
511
|
-
control[Control.COMPLETE] = 2;
|
|
512
|
-
console.error(e && typeof e === 'object' && 'stack' in e ? e.stack : e);
|
|
513
|
-
}
|
|
514
|
-
}
|
|
515
|
-
return control[Control.COMPLETE];
|
|
516
|
-
};
|
|
517
|
-
|
|
518
|
-
//# sourceMappingURL=runner.js.map
|
package/build/runner.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../src/runner.ts"],"sourcesContent":["import { performance, PerformanceObserver } from 'node:perf_hooks';\nimport { Options, Control, DURATION_SCALE } from './types.js';\nimport { GCWatcher } from './gc-watcher.js';\nimport { StepFn } from './types.js';\n\nconst COMPLETE_VALUE = 100_00;\n\nconst hr = process.hrtime.bigint.bind(process.hrtime);\n\nconst sink = new Int32Array(new SharedArrayBuffer(Int32Array.BYTES_PER_ELEMENT));\nconst consume = (value: unknown) => {\n let payload = 0;\n switch (typeof value) {\n case 'number':\n payload = Number.isFinite(value) ? Math.trunc(value) : 0;\n break;\n case 'bigint':\n payload = Number(value & 0xffff_ffffn);\n break;\n case 'string':\n payload = value.length;\n break;\n case 'boolean':\n payload = value ? 1 : 0;\n break;\n case 'object':\n payload = value === null ? 0 : 1;\n break;\n case 'function':\n payload = 1;\n break;\n default:\n payload = -1;\n }\n Atomics.xor(sink, 0, payload);\n};\n\nconst runSync = (run: Function, overhead: bigint) => {\n return (...args: unknown[]) => {\n const start = hr();\n const result = run(...args);\n consume(result);\n const duration = hr() - start;\n return duration > overhead ? duration - overhead : 0n;\n };\n};\n\nconst runAsync = (run: Function) => {\n return async (...args: unknown[]) => {\n const start = hr();\n const result = await run(...args);\n consume(result);\n return hr() - start;\n };\n};\n\nconst isThenable = (value: unknown): value is PromiseLike<unknown> => {\n return value !== null && (typeof value === 'object' || typeof value === 'function') && typeof (value as PromiseLike<unknown>).then === 'function';\n};\n\nconst TARGET_SAMPLE_NS = 1_000_000n; // aim for ~1ms per measured sample\nconst MAX_BATCH = 1_048_576;\nconst PROGRESS_STRIDE = 16;\nconst GC_STRIDE = 32;\nconst OUTLIER_MULTIPLIER = 4;\nconst OUTLIER_IQR_MULTIPLIER = 3;\nconst OUTLIER_WINDOW = 64;\nconst OUTLIER_ABS_THRESHOLD = 10_000_000;\nconst BASELINE_SAMPLES = 16;\nconst OUTLIER_SCRATCH = new Float64Array(OUTLIER_WINDOW);\n\ntype GCEvent = { start: number; end: number };\ntype RunTimedSync<TContext, TInput> = (ctx: TContext, data: TInput, nonce?: number) => bigint;\ntype RunTimedAsync<TContext, TInput> = (ctx: TContext, data: TInput, nonce?: number) => Promise<bigint>;\n\nconst measureTimerOverhead = () => {\n let total = 0n;\n for (let i = 0; i < BASELINE_SAMPLES; i++) {\n const start = hr();\n consume(0);\n total += hr() - start;\n }\n return total / BigInt(BASELINE_SAMPLES);\n};\n\nconst collectSample = async <TContext, TInput>({\n batchSize,\n run,\n runRaw,\n runIsAsync,\n pre,\n preIsAsync,\n post,\n postIsAsync,\n context,\n data,\n nextNonce,\n}: {\n batchSize: number;\n run: RunTimedSync<TContext, TInput> | RunTimedAsync<TContext, TInput>;\n runRaw: StepFn<TContext, TInput>;\n runIsAsync: boolean;\n pre: StepFn<TContext, TInput> | undefined;\n preIsAsync: boolean;\n post: StepFn<TContext, TInput> | undefined;\n postIsAsync: boolean;\n context: TContext;\n data: TInput;\n nextNonce: (() => number) | null;\n}) => {\n const canBatchTime = !runIsAsync && !pre && !post;\n if (canBatchTime) {\n const batchStart = hr();\n if (nextNonce) {\n for (let b = 0; b < batchSize; b++) {\n consume((runRaw as Function)(context, data, nextNonce()));\n }\n } else {\n for (let b = 0; b < batchSize; b++) {\n consume(runRaw(context, data));\n }\n }\n return ((hr() - batchStart) * DURATION_SCALE) / BigInt(batchSize);\n }\n\n let sampleDuration = 0n;\n for (let b = 0; b < batchSize; b++) {\n if (pre) {\n if (preIsAsync) {\n await pre(context, data);\n } else {\n pre(context, data);\n }\n }\n\n if (runIsAsync) {\n const runAsyncFn = run as RunTimedAsync<TContext, TInput>;\n const duration = nextNonce ? await runAsyncFn(context, data, nextNonce()) : await runAsyncFn(context, data);\n sampleDuration += duration;\n } else {\n const runSyncFn = run as RunTimedSync<TContext, TInput>;\n const duration = nextNonce ? runSyncFn(context, data, nextNonce()) : runSyncFn(context, data);\n sampleDuration += duration;\n }\n\n if (post) {\n if (postIsAsync) {\n await post(context, data);\n } else {\n post(context, data);\n }\n }\n }\n return (sampleDuration * DURATION_SCALE) / BigInt(batchSize);\n};\n\nconst tuneParameters = async <TContext, TInput>({\n initialBatch,\n run,\n runRaw,\n runIsAsync,\n pre,\n preIsAsync,\n post,\n postIsAsync,\n context,\n data,\n minCycles,\n relThreshold,\n maxCycles,\n nextNonce,\n}: {\n initialBatch: number;\n run: RunTimedSync<TContext, TInput> | RunTimedAsync<TContext, TInput>;\n runRaw: StepFn<TContext, TInput>;\n runIsAsync: boolean;\n pre?: StepFn<TContext, TInput>;\n preIsAsync: boolean;\n post?: StepFn<TContext, TInput>;\n postIsAsync: boolean;\n context: TContext;\n data: TInput;\n minCycles: number;\n relThreshold: number;\n maxCycles: number;\n nextNonce: (() => number) | null;\n}) => {\n let batchSize = initialBatch;\n let bestCv = Number.POSITIVE_INFINITY;\n let bestBatch = batchSize;\n\n for (let attempt = 0; attempt < 3; attempt++) {\n const samples: number[] = [];\n const sampleCount = Math.min(8, maxCycles);\n for (let s = 0; s < sampleCount; s++) {\n const duration = await collectSample({\n batchSize,\n run,\n runRaw,\n runIsAsync,\n pre,\n preIsAsync,\n post,\n postIsAsync,\n context,\n data,\n nextNonce,\n });\n samples.push(Number(duration));\n }\n const mean = samples.reduce((acc, v) => acc + v, 0) / samples.length;\n const variance = samples.reduce((acc, v) => acc + (v - mean) * (v - mean), 0) / Math.max(1, samples.length - 1);\n const stddev = Math.sqrt(variance);\n const cv = mean === 0 ? Number.POSITIVE_INFINITY : stddev / mean;\n\n if (cv < bestCv) {\n bestCv = cv;\n bestBatch = batchSize;\n }\n\n if (cv <= relThreshold || batchSize >= MAX_BATCH) {\n break;\n }\n batchSize = Math.min(MAX_BATCH, batchSize * 2);\n }\n\n const tunedRel = bestCv < relThreshold ? Math.max(bestCv * 1.5, relThreshold * 0.5) : relThreshold;\n const tunedMin = Math.min(maxCycles, Math.max(minCycles, Math.ceil(minCycles * Math.max(1, bestCv / (relThreshold || 1e-6)))));\n\n return { batchSize: bestBatch, relThreshold: tunedRel, minCycles: tunedMin };\n};\n\nconst createGCTracker = () => {\n if (process.env.OVERTAKE_GC_OBSERVER !== '1') {\n return null;\n }\n if (typeof PerformanceObserver === 'undefined') {\n return null;\n }\n\n const events: GCEvent[] = [];\n const observer = new PerformanceObserver((list) => {\n for (const entry of list.getEntries()) {\n events.push({ start: entry.startTime, end: entry.startTime + entry.duration });\n }\n });\n\n try {\n observer.observe({ entryTypes: ['gc'] });\n } catch {\n return null;\n }\n\n const overlaps = (start: number, end: number) => {\n let noisy = false;\n for (let i = events.length - 1; i >= 0; i--) {\n const event = events[i];\n if (event.end < start - 5_000) {\n events.splice(i, 1);\n continue;\n }\n if (event.start <= end && event.end >= start) {\n noisy = true;\n }\n }\n return noisy;\n };\n\n const dispose = () => observer.disconnect();\n\n return { overlaps, dispose };\n};\n\nconst pushWindow = (arr: number[], value: number, cap: number) => {\n if (arr.length === cap) {\n arr.shift();\n }\n arr.push(value);\n};\n\nconst medianAndIqr = (arr: number[]) => {\n if (arr.length === 0) return { median: 0, iqr: 0 };\n for (let i = 0; i < arr.length; i++) {\n OUTLIER_SCRATCH[i] = arr[i];\n }\n const view = OUTLIER_SCRATCH.subarray(0, arr.length);\n view.sort();\n const mid = Math.floor(view.length / 2);\n const median = view.length % 2 === 0 ? (view[mid - 1] + view[mid]) / 2 : view[mid];\n const q1Idx = Math.floor(view.length * 0.25);\n const q3Idx = Math.floor(view.length * 0.75);\n const q1 = view[q1Idx];\n const q3 = view[q3Idx];\n return { median, iqr: q3 - q1 };\n};\n\nconst windowCv = (arr: number[]) => {\n if (arr.length < 2) return Number.POSITIVE_INFINITY;\n const mean = arr.reduce((a, v) => a + v, 0) / arr.length;\n const variance = arr.reduce((a, v) => a + (v - mean) * (v - mean), 0) / (arr.length - 1);\n const stddev = Math.sqrt(variance);\n return mean === 0 ? Number.POSITIVE_INFINITY : stddev / mean;\n};\n\nexport const benchmark = async <TContext, TInput>({\n setup,\n teardown,\n pre,\n run: runRaw,\n post,\n data,\n\n warmupCycles,\n minCycles,\n absThreshold,\n relThreshold,\n gcObserver = false,\n\n durationsSAB,\n controlSAB,\n}: Required<Options<TContext, TInput>>) => {\n const durations = new BigUint64Array(durationsSAB);\n const control = new Int32Array(controlSAB);\n\n control[Control.INDEX] = 0;\n control[Control.PROGRESS] = 0;\n control[Control.COMPLETE] = 255;\n control[Control.HEAP_USED] = 0;\n\n const context = (await setup?.()) as TContext;\n const heapBefore = process.memoryUsage().heapUsed;\n const input = data as TInput;\n const maxCycles = durations.length;\n const gcWatcher = gcObserver ? new GCWatcher() : null;\n const gcTracker = gcObserver ? createGCTracker() : null;\n\n try {\n // classify sync/async and capture initial duration\n let preIsAsync = false;\n if (pre) {\n const preResult = pre(context, input);\n preIsAsync = isThenable(preResult);\n if (preIsAsync) {\n await preResult;\n }\n }\n\n const probeStart = hr();\n const probeResult = runRaw(context, input);\n const runIsAsync = isThenable(probeResult);\n if (runIsAsync) {\n const resolved = await probeResult;\n consume(resolved);\n } else {\n consume(probeResult);\n }\n const durationProbeRaw = hr() - probeStart;\n\n let postIsAsync = false;\n if (post) {\n const postResult = post(context, input);\n postIsAsync = isThenable(postResult);\n if (postIsAsync) {\n await postResult;\n }\n }\n\n const timerOverhead = runIsAsync ? 0n : measureTimerOverhead();\n let durationProbe = runIsAsync ? durationProbeRaw : durationProbeRaw > timerOverhead ? durationProbeRaw - timerOverhead : 0n;\n\n const shouldPerturbInput = process.env.OVERTAKE_PERTURB_INPUT === '1';\n let nonce = 0;\n const nextNonce = shouldPerturbInput\n ? () => {\n nonce = (nonce + 1) | 0;\n return nonce;\n }\n : null;\n\n if (!runIsAsync && !pre && !post) {\n const PROBE_TIME_LIMIT_NS = 1_000_000_000n;\n const INITIAL_PROBE_SIZE = 10;\n const MAX_PROBE_SIZE = 10_000;\n\n const initialStart = hr();\n if (nextNonce) {\n for (let i = 0; i < INITIAL_PROBE_SIZE; i++) {\n consume((runRaw as Function)(context, input, nextNonce()));\n }\n } else {\n for (let i = 0; i < INITIAL_PROBE_SIZE; i++) {\n consume(runRaw(context, input));\n }\n }\n const initialDuration = hr() - initialStart;\n const estimatedPerOp = initialDuration / BigInt(INITIAL_PROBE_SIZE);\n\n const remainingBudget = PROBE_TIME_LIMIT_NS - initialDuration;\n const additionalIterations = estimatedPerOp > 0n ? Number(remainingBudget / estimatedPerOp) : MAX_PROBE_SIZE - INITIAL_PROBE_SIZE;\n const cappedAdditional = Math.min(Math.max(0, additionalIterations), MAX_PROBE_SIZE - INITIAL_PROBE_SIZE);\n\n let totalIterations = INITIAL_PROBE_SIZE;\n if (cappedAdditional > 0) {\n if (nextNonce) {\n for (let i = 0; i < cappedAdditional; i++) {\n consume((runRaw as Function)(context, input, nextNonce()));\n }\n } else {\n for (let i = 0; i < cappedAdditional; i++) {\n consume(runRaw(context, input));\n }\n }\n totalIterations += cappedAdditional;\n }\n\n durationProbe = (hr() - initialStart) / BigInt(totalIterations);\n }\n\n const runTimedSync = runIsAsync ? null : runSync(runRaw, timerOverhead);\n const runTimedAsync = runIsAsync ? runAsync(runRaw) : null;\n const run = runIsAsync ? runTimedAsync! : runTimedSync!;\n\n const runOnceSync: RunTimedSync<TContext, TInput> | null = runIsAsync ? null : nextNonce ? (ctx, dataValue) => runTimedSync!(ctx, dataValue, nextNonce()) : runTimedSync!;\n const runOnceAsync: RunTimedAsync<TContext, TInput> | null = runIsAsync ? (nextNonce ? (ctx, dataValue) => runTimedAsync!(ctx, dataValue, nextNonce()) : runTimedAsync!) : null;\n\n const preSync = preIsAsync ? null : pre;\n const preAsync = preIsAsync ? pre : null;\n const postSync = postIsAsync ? null : post;\n const postAsync = postIsAsync ? post : null;\n\n // choose batch size to amortize timer overhead\n const durationPerRun = durationProbe === 0n ? 1n : durationProbe;\n const suggestedBatch = Number(TARGET_SAMPLE_NS / durationPerRun);\n const minBatchForFastOps = durationProbe < 100n ? 100_000 : 1;\n const initialBatchSize = Math.min(MAX_BATCH, Math.max(minBatchForFastOps, suggestedBatch));\n\n // auto-tune based on warmup samples\n const tuned = await tuneParameters({\n initialBatch: initialBatchSize,\n run,\n runRaw,\n runIsAsync,\n pre,\n preIsAsync,\n post,\n postIsAsync,\n context,\n data: input,\n minCycles,\n relThreshold,\n maxCycles,\n nextNonce,\n });\n let batchSize = tuned.batchSize;\n minCycles = tuned.minCycles;\n relThreshold = tuned.relThreshold;\n\n // warmup: run until requested cycles, adapt if unstable\n const warmupStart = performance.now();\n let warmupRemaining = warmupCycles;\n const warmupWindow: number[] = [];\n const warmupCap = Math.max(warmupCycles, Math.min(maxCycles, warmupCycles * 4 || 1000));\n const canBatchTime = !runIsAsync && !preSync && !preAsync && !postSync && !postAsync;\n\n const runWarmup = async () => {\n if (canBatchTime) {\n const batchStart = hr();\n if (nextNonce) {\n for (let b = 0; b < batchSize; b++) {\n consume((runRaw as Function)(context, input, nextNonce()));\n }\n } else {\n for (let b = 0; b < batchSize; b++) {\n consume(runRaw(context, input));\n }\n }\n return ((hr() - batchStart) * DURATION_SCALE) / BigInt(batchSize);\n }\n\n if (preSync) {\n preSync(context, input);\n } else if (preAsync) {\n await preAsync(context, input);\n }\n\n const duration = runIsAsync ? await runOnceAsync!(context, input) : runOnceSync!(context, input);\n\n if (postSync) {\n postSync(context, input);\n } else if (postAsync) {\n await postAsync(context, input);\n }\n\n return duration * DURATION_SCALE;\n };\n\n while (performance.now() - warmupStart < 1_000 && warmupRemaining > 0) {\n const duration = await runWarmup();\n pushWindow(warmupWindow, Number(duration), warmupCap);\n warmupRemaining--;\n }\n let warmupDone = 0;\n while (warmupDone < warmupRemaining) {\n const duration = await runWarmup();\n pushWindow(warmupWindow, Number(duration), warmupCap);\n warmupDone++;\n if (global.gc && warmupDone % GC_STRIDE === 0) {\n global.gc();\n }\n }\n while (warmupWindow.length >= 8 && warmupWindow.length < warmupCap) {\n const cv = windowCv(warmupWindow);\n if (cv <= relThreshold * 2) {\n break;\n }\n const duration = await runWarmup();\n pushWindow(warmupWindow, Number(duration), warmupCap);\n }\n\n let i = 0;\n let mean = 0n;\n let m2 = 0n;\n const outlierWindow: number[] = [];\n let skipped = 0;\n const maxSkipped = maxCycles * 10;\n let disableFiltering = false;\n\n while (true) {\n if (i >= maxCycles) break;\n if (!disableFiltering && skipped >= maxSkipped) {\n console.error(`Warning: ${skipped} samples skipped due to noise/outlier detection. ` + `Disabling filtering for remaining samples. Results may have higher variance.`);\n disableFiltering = true;\n }\n\n if (global.gc && i > 0 && i % GC_STRIDE === 0) {\n global.gc();\n }\n\n const gcMarker = gcWatcher?.start();\n const sampleStart = performance.now();\n let sampleDuration = 0n;\n\n if (canBatchTime) {\n const batchStart = hr();\n if (nextNonce) {\n for (let b = 0; b < batchSize; b++) {\n consume((runRaw as Function)(context, input, nextNonce()));\n }\n } else {\n for (let b = 0; b < batchSize; b++) {\n consume(runRaw(context, input));\n }\n }\n const batchDuration = hr() - batchStart;\n sampleDuration = (batchDuration * DURATION_SCALE) / BigInt(batchSize);\n } else {\n for (let b = 0; b < batchSize; b++) {\n if (preSync) {\n preSync(context, input);\n } else if (preAsync) {\n await preAsync(context, input);\n }\n\n const duration = runIsAsync ? await runOnceAsync!(context, input) : runOnceSync!(context, input);\n sampleDuration += duration;\n\n if (postSync) {\n postSync(context, input);\n } else if (postAsync) {\n await postAsync(context, input);\n }\n }\n sampleDuration = (sampleDuration * DURATION_SCALE) / BigInt(batchSize);\n }\n\n const sampleEnd = performance.now();\n if (!disableFiltering) {\n const gcNoise = (gcMarker ? gcWatcher!.seen(gcMarker) : false) || (gcTracker?.overlaps(sampleStart, sampleEnd) ?? false);\n if (gcNoise) {\n skipped++;\n continue;\n }\n }\n\n const durationNumber = Number(sampleDuration);\n pushWindow(outlierWindow, durationNumber, OUTLIER_WINDOW);\n if (!disableFiltering) {\n const { median, iqr } = medianAndIqr(outlierWindow);\n const maxAllowed = median + OUTLIER_IQR_MULTIPLIER * iqr || Number.POSITIVE_INFINITY;\n if (outlierWindow.length >= 8 && durationNumber > maxAllowed && durationNumber - median > OUTLIER_ABS_THRESHOLD) {\n skipped++;\n continue;\n }\n\n const meanNumber = Number(mean);\n if (i >= 8 && meanNumber > 0 && durationNumber > OUTLIER_MULTIPLIER * meanNumber && durationNumber - meanNumber > OUTLIER_ABS_THRESHOLD) {\n skipped++;\n continue;\n }\n }\n\n durations[i++] = sampleDuration;\n const delta = sampleDuration - mean;\n mean += delta / BigInt(i);\n m2 += delta * (sampleDuration - mean);\n\n const progress = (i / maxCycles) * COMPLETE_VALUE;\n if (i % PROGRESS_STRIDE === 0) {\n control[Control.PROGRESS] = progress;\n }\n\n if (i >= minCycles) {\n const variance = Number(m2) / (i - 1);\n const stddev = Math.sqrt(variance);\n if (stddev <= Number(absThreshold)) {\n break;\n }\n\n const meanNum = Number(mean);\n const cov = stddev / (meanNum || 1);\n if (cov <= relThreshold) {\n break;\n }\n }\n }\n\n control[Control.INDEX] = i;\n control[Control.COMPLETE] = 0;\n const heapAfter = process.memoryUsage().heapUsed;\n control[Control.HEAP_USED] = Math.max(0, Math.round((heapAfter - heapBefore) / 1024));\n } catch (e) {\n console.error(e && typeof e === 'object' && 'stack' in e ? e.stack : e);\n control[Control.COMPLETE] = 1;\n } finally {\n gcTracker?.dispose?.();\n try {\n await teardown?.(context);\n } catch (e) {\n control[Control.COMPLETE] = 2;\n console.error(e && typeof e === 'object' && 'stack' in e ? e.stack : e);\n }\n }\n\n return control[Control.COMPLETE];\n};\n"],"names":["performance","PerformanceObserver","Control","DURATION_SCALE","GCWatcher","COMPLETE_VALUE","hr","process","hrtime","bigint","bind","sink","Int32Array","SharedArrayBuffer","BYTES_PER_ELEMENT","consume","value","payload","Number","isFinite","Math","trunc","length","Atomics","xor","runSync","run","overhead","args","start","result","duration","runAsync","isThenable","then","TARGET_SAMPLE_NS","MAX_BATCH","PROGRESS_STRIDE","GC_STRIDE","OUTLIER_MULTIPLIER","OUTLIER_IQR_MULTIPLIER","OUTLIER_WINDOW","OUTLIER_ABS_THRESHOLD","BASELINE_SAMPLES","OUTLIER_SCRATCH","Float64Array","measureTimerOverhead","total","i","BigInt","collectSample","batchSize","runRaw","runIsAsync","pre","preIsAsync","post","postIsAsync","context","data","nextNonce","canBatchTime","batchStart","b","sampleDuration","runAsyncFn","runSyncFn","tuneParameters","initialBatch","minCycles","relThreshold","maxCycles","bestCv","POSITIVE_INFINITY","bestBatch","attempt","samples","sampleCount","min","s","push","mean","reduce","acc","v","variance","max","stddev","sqrt","cv","tunedRel","tunedMin","ceil","createGCTracker","env","OVERTAKE_GC_OBSERVER","events","observer","list","entry","getEntries","startTime","end","observe","entryTypes","overlaps","noisy","event","splice","dispose","disconnect","pushWindow","arr","cap","shift","medianAndIqr","median","iqr","view","subarray","sort","mid","floor","q1Idx","q3Idx","q1","q3","windowCv","a","benchmark","setup","teardown","warmupCycles","absThreshold","gcObserver","durationsSAB","controlSAB","durations","BigUint64Array","control","INDEX","PROGRESS","COMPLETE","HEAP_USED","heapBefore","memoryUsage","heapUsed","input","gcWatcher","gcTracker","preResult","probeStart","probeResult","resolved","durationProbeRaw","postResult","timerOverhead","durationProbe","shouldPerturbInput","OVERTAKE_PERTURB_INPUT","nonce","PROBE_TIME_LIMIT_NS","INITIAL_PROBE_SIZE","MAX_PROBE_SIZE","initialStart","initialDuration","estimatedPerOp","remainingBudget","additionalIterations","cappedAdditional","totalIterations","runTimedSync","runTimedAsync","runOnceSync","ctx","dataValue","runOnceAsync","preSync","preAsync","postSync","postAsync","durationPerRun","suggestedBatch","minBatchForFastOps","initialBatchSize","tuned","warmupStart","now","warmupRemaining","warmupWindow","warmupCap","runWarmup","warmupDone","global","gc","m2","outlierWindow","skipped","maxSkipped","disableFiltering","console","error","gcMarker","sampleStart","batchDuration","sampleEnd","gcNoise","seen","durationNumber","maxAllowed","meanNumber","delta","progress","meanNum","cov","heapAfter","round","e","stack"],"mappings":"AAAA,SAASA,WAAW,EAAEC,mBAAmB,QAAQ,kBAAkB;AACnE,SAAkBC,OAAO,EAAEC,cAAc,QAAQ,aAAa;AAC9D,SAASC,SAAS,QAAQ,kBAAkB;AAG5C,MAAMC,iBAAiB;AAEvB,MAAMC,KAAKC,QAAQC,MAAM,CAACC,MAAM,CAACC,IAAI,CAACH,QAAQC,MAAM;AAEpD,MAAMG,OAAO,IAAIC,WAAW,IAAIC,kBAAkBD,WAAWE,iBAAiB;AAC9E,MAAMC,UAAU,CAACC;IACf,IAAIC,UAAU;IACd,OAAQ,OAAOD;QACb,KAAK;YACHC,UAAUC,OAAOC,QAAQ,CAACH,SAASI,KAAKC,KAAK,CAACL,SAAS;YACvD;QACF,KAAK;YACHC,UAAUC,OAAOF,QAAQ,YAAY;YACrC;QACF,KAAK;YACHC,UAAUD,MAAMM,MAAM;YACtB;QACF,KAAK;YACHL,UAAUD,QAAQ,IAAI;YACtB;QACF,KAAK;YACHC,UAAUD,UAAU,OAAO,IAAI;YAC/B;QACF,KAAK;YACHC,UAAU;YACV;QACF;YACEA,UAAU,CAAC;IACf;IACAM,QAAQC,GAAG,CAACb,MAAM,GAAGM;AACvB;AAEA,MAAMQ,UAAU,CAACC,KAAeC;IAC9B,OAAO,CAAC,GAAGC;QACT,MAAMC,QAAQvB;QACd,MAAMwB,SAASJ,OAAOE;QACtBb,QAAQe;QACR,MAAMC,WAAWzB,OAAOuB;QACxB,OAAOE,WAAWJ,WAAWI,WAAWJ,WAAW,EAAE;IACvD;AACF;AAEA,MAAMK,WAAW,CAACN;IAChB,OAAO,OAAO,GAAGE;QACf,MAAMC,QAAQvB;QACd,MAAMwB,SAAS,MAAMJ,OAAOE;QAC5Bb,QAAQe;QACR,OAAOxB,OAAOuB;IAChB;AACF;AAEA,MAAMI,aAAa,CAACjB;IAClB,OAAOA,UAAU,QAAS,CAAA,OAAOA,UAAU,YAAY,OAAOA,UAAU,UAAS,KAAM,OAAO,AAACA,MAA+BkB,IAAI,KAAK;AACzI;AAEA,MAAMC,mBAAmB,UAAU;AACnC,MAAMC,YAAY;AAClB,MAAMC,kBAAkB;AACxB,MAAMC,YAAY;AAClB,MAAMC,qBAAqB;AAC3B,MAAMC,yBAAyB;AAC/B,MAAMC,iBAAiB;AACvB,MAAMC,wBAAwB;AAC9B,MAAMC,mBAAmB;AACzB,MAAMC,kBAAkB,IAAIC,aAAaJ;AAMzC,MAAMK,uBAAuB;IAC3B,IAAIC,QAAQ,EAAE;IACd,IAAK,IAAIC,IAAI,GAAGA,IAAIL,kBAAkBK,IAAK;QACzC,MAAMnB,QAAQvB;QACdS,QAAQ;QACRgC,SAASzC,OAAOuB;IAClB;IACA,OAAOkB,QAAQE,OAAON;AACxB;AAEA,MAAMO,gBAAgB,OAAyB,EAC7CC,SAAS,EACTzB,GAAG,EACH0B,MAAM,EACNC,UAAU,EACVC,GAAG,EACHC,UAAU,EACVC,IAAI,EACJC,WAAW,EACXC,OAAO,EACPC,IAAI,EACJC,SAAS,EAaV;IACC,MAAMC,eAAe,CAACR,cAAc,CAACC,OAAO,CAACE;IAC7C,IAAIK,cAAc;QAChB,MAAMC,aAAaxD;QACnB,IAAIsD,WAAW;YACb,IAAK,IAAIG,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;gBAClChD,QAAQ,AAACqC,OAAoBM,SAASC,MAAMC;YAC9C;QACF,OAAO;YACL,IAAK,IAAIG,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;gBAClChD,QAAQqC,OAAOM,SAASC;YAC1B;QACF;QACA,OAAO,AAAErD,CAAAA,OAAOwD,UAAS,IAAK3D,iBAAkB8C,OAAOE;IACzD;IAEA,IAAIa,iBAAiB,EAAE;IACvB,IAAK,IAAID,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;QAClC,IAAIT,KAAK;YACP,IAAIC,YAAY;gBACd,MAAMD,IAAII,SAASC;YACrB,OAAO;gBACLL,IAAII,SAASC;YACf;QACF;QAEA,IAAIN,YAAY;YACd,MAAMY,aAAavC;YACnB,MAAMK,WAAW6B,YAAY,MAAMK,WAAWP,SAASC,MAAMC,eAAe,MAAMK,WAAWP,SAASC;YACtGK,kBAAkBjC;QACpB,OAAO;YACL,MAAMmC,YAAYxC;YAClB,MAAMK,WAAW6B,YAAYM,UAAUR,SAASC,MAAMC,eAAeM,UAAUR,SAASC;YACxFK,kBAAkBjC;QACpB;QAEA,IAAIyB,MAAM;YACR,IAAIC,aAAa;gBACf,MAAMD,KAAKE,SAASC;YACtB,OAAO;gBACLH,KAAKE,SAASC;YAChB;QACF;IACF;IACA,OAAO,AAACK,iBAAiB7D,iBAAkB8C,OAAOE;AACpD;AAEA,MAAMgB,iBAAiB,OAAyB,EAC9CC,YAAY,EACZ1C,GAAG,EACH0B,MAAM,EACNC,UAAU,EACVC,GAAG,EACHC,UAAU,EACVC,IAAI,EACJC,WAAW,EACXC,OAAO,EACPC,IAAI,EACJU,SAAS,EACTC,YAAY,EACZC,SAAS,EACTX,SAAS,EAgBV;IACC,IAAIT,YAAYiB;IAChB,IAAII,SAAStD,OAAOuD,iBAAiB;IACrC,IAAIC,YAAYvB;IAEhB,IAAK,IAAIwB,UAAU,GAAGA,UAAU,GAAGA,UAAW;QAC5C,MAAMC,UAAoB,EAAE;QAC5B,MAAMC,cAAczD,KAAK0D,GAAG,CAAC,GAAGP;QAChC,IAAK,IAAIQ,IAAI,GAAGA,IAAIF,aAAaE,IAAK;YACpC,MAAMhD,WAAW,MAAMmB,cAAc;gBACnCC;gBACAzB;gBACA0B;gBACAC;gBACAC;gBACAC;gBACAC;gBACAC;gBACAC;gBACAC;gBACAC;YACF;YACAgB,QAAQI,IAAI,CAAC9D,OAAOa;QACtB;QACA,MAAMkD,OAAOL,QAAQM,MAAM,CAAC,CAACC,KAAKC,IAAMD,MAAMC,GAAG,KAAKR,QAAQtD,MAAM;QACpE,MAAM+D,WAAWT,QAAQM,MAAM,CAAC,CAACC,KAAKC,IAAMD,MAAM,AAACC,CAAAA,IAAIH,IAAG,IAAMG,CAAAA,IAAIH,IAAG,GAAI,KAAK7D,KAAKkE,GAAG,CAAC,GAAGV,QAAQtD,MAAM,GAAG;QAC7G,MAAMiE,SAASnE,KAAKoE,IAAI,CAACH;QACzB,MAAMI,KAAKR,SAAS,IAAI/D,OAAOuD,iBAAiB,GAAGc,SAASN;QAE5D,IAAIQ,KAAKjB,QAAQ;YACfA,SAASiB;YACTf,YAAYvB;QACd;QAEA,IAAIsC,MAAMnB,gBAAgBnB,aAAaf,WAAW;YAChD;QACF;QACAe,YAAY/B,KAAK0D,GAAG,CAAC1C,WAAWe,YAAY;IAC9C;IAEA,MAAMuC,WAAWlB,SAASF,eAAelD,KAAKkE,GAAG,CAACd,SAAS,KAAKF,eAAe,OAAOA;IACtF,MAAMqB,WAAWvE,KAAK0D,GAAG,CAACP,WAAWnD,KAAKkE,GAAG,CAACjB,WAAWjD,KAAKwE,IAAI,CAACvB,YAAYjD,KAAKkE,GAAG,CAAC,GAAGd,SAAUF,CAAAA,gBAAgB,IAAG;IAExH,OAAO;QAAEnB,WAAWuB;QAAWJ,cAAcoB;QAAUrB,WAAWsB;IAAS;AAC7E;AAEA,MAAME,kBAAkB;IACtB,IAAItF,QAAQuF,GAAG,CAACC,oBAAoB,KAAK,KAAK;QAC5C,OAAO;IACT;IACA,IAAI,OAAO9F,wBAAwB,aAAa;QAC9C,OAAO;IACT;IAEA,MAAM+F,SAAoB,EAAE;IAC5B,MAAMC,WAAW,IAAIhG,oBAAoB,CAACiG;QACxC,KAAK,MAAMC,SAASD,KAAKE,UAAU,GAAI;YACrCJ,OAAOhB,IAAI,CAAC;gBAAEnD,OAAOsE,MAAME,SAAS;gBAAEC,KAAKH,MAAME,SAAS,GAAGF,MAAMpE,QAAQ;YAAC;QAC9E;IACF;IAEA,IAAI;QACFkE,SAASM,OAAO,CAAC;YAAEC,YAAY;gBAAC;aAAK;QAAC;IACxC,EAAE,OAAM;QACN,OAAO;IACT;IAEA,MAAMC,WAAW,CAAC5E,OAAeyE;QAC/B,IAAII,QAAQ;QACZ,IAAK,IAAI1D,IAAIgD,OAAO1E,MAAM,GAAG,GAAG0B,KAAK,GAAGA,IAAK;YAC3C,MAAM2D,QAAQX,MAAM,CAAChD,EAAE;YACvB,IAAI2D,MAAML,GAAG,GAAGzE,QAAQ,OAAO;gBAC7BmE,OAAOY,MAAM,CAAC5D,GAAG;gBACjB;YACF;YACA,IAAI2D,MAAM9E,KAAK,IAAIyE,OAAOK,MAAML,GAAG,IAAIzE,OAAO;gBAC5C6E,QAAQ;YACV;QACF;QACA,OAAOA;IACT;IAEA,MAAMG,UAAU,IAAMZ,SAASa,UAAU;IAEzC,OAAO;QAAEL;QAAUI;IAAQ;AAC7B;AAEA,MAAME,aAAa,CAACC,KAAehG,OAAeiG;IAChD,IAAID,IAAI1F,MAAM,KAAK2F,KAAK;QACtBD,IAAIE,KAAK;IACX;IACAF,IAAIhC,IAAI,CAAChE;AACX;AAEA,MAAMmG,eAAe,CAACH;IACpB,IAAIA,IAAI1F,MAAM,KAAK,GAAG,OAAO;QAAE8F,QAAQ;QAAGC,KAAK;IAAE;IACjD,IAAK,IAAIrE,IAAI,GAAGA,IAAIgE,IAAI1F,MAAM,EAAE0B,IAAK;QACnCJ,eAAe,CAACI,EAAE,GAAGgE,GAAG,CAAChE,EAAE;IAC7B;IACA,MAAMsE,OAAO1E,gBAAgB2E,QAAQ,CAAC,GAAGP,IAAI1F,MAAM;IACnDgG,KAAKE,IAAI;IACT,MAAMC,MAAMrG,KAAKsG,KAAK,CAACJ,KAAKhG,MAAM,GAAG;IACrC,MAAM8F,SAASE,KAAKhG,MAAM,GAAG,MAAM,IAAI,AAACgG,CAAAA,IAAI,CAACG,MAAM,EAAE,GAAGH,IAAI,CAACG,IAAI,AAAD,IAAK,IAAIH,IAAI,CAACG,IAAI;IAClF,MAAME,QAAQvG,KAAKsG,KAAK,CAACJ,KAAKhG,MAAM,GAAG;IACvC,MAAMsG,QAAQxG,KAAKsG,KAAK,CAACJ,KAAKhG,MAAM,GAAG;IACvC,MAAMuG,KAAKP,IAAI,CAACK,MAAM;IACtB,MAAMG,KAAKR,IAAI,CAACM,MAAM;IACtB,OAAO;QAAER;QAAQC,KAAKS,KAAKD;IAAG;AAChC;AAEA,MAAME,WAAW,CAACf;IAChB,IAAIA,IAAI1F,MAAM,GAAG,GAAG,OAAOJ,OAAOuD,iBAAiB;IACnD,MAAMQ,OAAO+B,IAAI9B,MAAM,CAAC,CAAC8C,GAAG5C,IAAM4C,IAAI5C,GAAG,KAAK4B,IAAI1F,MAAM;IACxD,MAAM+D,WAAW2B,IAAI9B,MAAM,CAAC,CAAC8C,GAAG5C,IAAM4C,IAAI,AAAC5C,CAAAA,IAAIH,IAAG,IAAMG,CAAAA,IAAIH,IAAG,GAAI,KAAM+B,CAAAA,IAAI1F,MAAM,GAAG,CAAA;IACtF,MAAMiE,SAASnE,KAAKoE,IAAI,CAACH;IACzB,OAAOJ,SAAS,IAAI/D,OAAOuD,iBAAiB,GAAGc,SAASN;AAC1D;AAEA,OAAO,MAAMgD,YAAY,OAAyB,EAChDC,KAAK,EACLC,QAAQ,EACR7E,GAAG,EACH5B,KAAK0B,MAAM,EACXI,IAAI,EACJG,IAAI,EAEJyE,YAAY,EACZ/D,SAAS,EACTgE,YAAY,EACZ/D,YAAY,EACZgE,aAAa,KAAK,EAElBC,YAAY,EACZC,UAAU,EAC0B;IACpC,MAAMC,YAAY,IAAIC,eAAeH;IACrC,MAAMI,UAAU,IAAI/H,WAAW4H;IAE/BG,OAAO,CAACzI,QAAQ0I,KAAK,CAAC,GAAG;IACzBD,OAAO,CAACzI,QAAQ2I,QAAQ,CAAC,GAAG;IAC5BF,OAAO,CAACzI,QAAQ4I,QAAQ,CAAC,GAAG;IAC5BH,OAAO,CAACzI,QAAQ6I,SAAS,CAAC,GAAG;IAE7B,MAAMrF,UAAW,MAAMwE;IACvB,MAAMc,aAAazI,QAAQ0I,WAAW,GAAGC,QAAQ;IACjD,MAAMC,QAAQxF;IACd,MAAMY,YAAYkE,UAAUnH,MAAM;IAClC,MAAM8H,YAAYd,aAAa,IAAIlI,cAAc;IACjD,MAAMiJ,YAAYf,aAAazC,oBAAoB;IAEnD,IAAI;QAEF,IAAItC,aAAa;QACjB,IAAID,KAAK;YACP,MAAMgG,YAAYhG,IAAII,SAASyF;YAC/B5F,aAAatB,WAAWqH;YACxB,IAAI/F,YAAY;gBACd,MAAM+F;YACR;QACF;QAEA,MAAMC,aAAajJ;QACnB,MAAMkJ,cAAcpG,OAAOM,SAASyF;QACpC,MAAM9F,aAAapB,WAAWuH;QAC9B,IAAInG,YAAY;YACd,MAAMoG,WAAW,MAAMD;YACvBzI,QAAQ0I;QACV,OAAO;YACL1I,QAAQyI;QACV;QACA,MAAME,mBAAmBpJ,OAAOiJ;QAEhC,IAAI9F,cAAc;QAClB,IAAID,MAAM;YACR,MAAMmG,aAAanG,KAAKE,SAASyF;YACjC1F,cAAcxB,WAAW0H;YACzB,IAAIlG,aAAa;gBACf,MAAMkG;YACR;QACF;QAEA,MAAMC,gBAAgBvG,aAAa,EAAE,GAAGP;QACxC,IAAI+G,gBAAgBxG,aAAaqG,mBAAmBA,mBAAmBE,gBAAgBF,mBAAmBE,gBAAgB,EAAE;QAE5H,MAAME,qBAAqBvJ,QAAQuF,GAAG,CAACiE,sBAAsB,KAAK;QAClE,IAAIC,QAAQ;QACZ,MAAMpG,YAAYkG,qBACd;YACEE,QAAQ,AAACA,QAAQ,IAAK;YACtB,OAAOA;QACT,IACA;QAEJ,IAAI,CAAC3G,cAAc,CAACC,OAAO,CAACE,MAAM;YAChC,MAAMyG,sBAAsB,cAAc;YAC1C,MAAMC,qBAAqB;YAC3B,MAAMC,iBAAiB;YAEvB,MAAMC,eAAe9J;YACrB,IAAIsD,WAAW;gBACb,IAAK,IAAIZ,IAAI,GAAGA,IAAIkH,oBAAoBlH,IAAK;oBAC3CjC,QAAQ,AAACqC,OAAoBM,SAASyF,OAAOvF;gBAC/C;YACF,OAAO;gBACL,IAAK,IAAIZ,IAAI,GAAGA,IAAIkH,oBAAoBlH,IAAK;oBAC3CjC,QAAQqC,OAAOM,SAASyF;gBAC1B;YACF;YACA,MAAMkB,kBAAkB/J,OAAO8J;YAC/B,MAAME,iBAAiBD,kBAAkBpH,OAAOiH;YAEhD,MAAMK,kBAAkBN,sBAAsBI;YAC9C,MAAMG,uBAAuBF,iBAAiB,EAAE,GAAGpJ,OAAOqJ,kBAAkBD,kBAAkBH,iBAAiBD;YAC/G,MAAMO,mBAAmBrJ,KAAK0D,GAAG,CAAC1D,KAAKkE,GAAG,CAAC,GAAGkF,uBAAuBL,iBAAiBD;YAEtF,IAAIQ,kBAAkBR;YACtB,IAAIO,mBAAmB,GAAG;gBACxB,IAAI7G,WAAW;oBACb,IAAK,IAAIZ,IAAI,GAAGA,IAAIyH,kBAAkBzH,IAAK;wBACzCjC,QAAQ,AAACqC,OAAoBM,SAASyF,OAAOvF;oBAC/C;gBACF,OAAO;oBACL,IAAK,IAAIZ,IAAI,GAAGA,IAAIyH,kBAAkBzH,IAAK;wBACzCjC,QAAQqC,OAAOM,SAASyF;oBAC1B;gBACF;gBACAuB,mBAAmBD;YACrB;YAEAZ,gBAAgB,AAACvJ,CAAAA,OAAO8J,YAAW,IAAKnH,OAAOyH;QACjD;QAEA,MAAMC,eAAetH,aAAa,OAAO5B,QAAQ2B,QAAQwG;QACzD,MAAMgB,gBAAgBvH,aAAarB,SAASoB,UAAU;QACtD,MAAM1B,MAAM2B,aAAauH,gBAAiBD;QAE1C,MAAME,cAAqDxH,aAAa,OAAOO,YAAY,CAACkH,KAAKC,YAAcJ,aAAcG,KAAKC,WAAWnH,eAAe+G;QAC5J,MAAMK,eAAuD3H,aAAcO,YAAY,CAACkH,KAAKC,YAAcH,cAAeE,KAAKC,WAAWnH,eAAegH,gBAAkB;QAE3K,MAAMK,UAAU1H,aAAa,OAAOD;QACpC,MAAM4H,WAAW3H,aAAaD,MAAM;QACpC,MAAM6H,WAAW1H,cAAc,OAAOD;QACtC,MAAM4H,YAAY3H,cAAcD,OAAO;QAGvC,MAAM6H,iBAAiBxB,kBAAkB,EAAE,GAAG,EAAE,GAAGA;QACnD,MAAMyB,iBAAiBpK,OAAOiB,mBAAmBkJ;QACjD,MAAME,qBAAqB1B,gBAAgB,IAAI,GAAG,UAAU;QAC5D,MAAM2B,mBAAmBpK,KAAK0D,GAAG,CAAC1C,WAAWhB,KAAKkE,GAAG,CAACiG,oBAAoBD;QAG1E,MAAMG,QAAQ,MAAMtH,eAAe;YACjCC,cAAcoH;YACd9J;YACA0B;YACAC;YACAC;YACAC;YACAC;YACAC;YACAC;YACAC,MAAMwF;YACN9E;YACAC;YACAC;YACAX;QACF;QACA,IAAIT,YAAYsI,MAAMtI,SAAS;QAC/BkB,YAAYoH,MAAMpH,SAAS;QAC3BC,eAAemH,MAAMnH,YAAY;QAGjC,MAAMoH,cAAc1L,YAAY2L,GAAG;QACnC,IAAIC,kBAAkBxD;QACtB,MAAMyD,eAAyB,EAAE;QACjC,MAAMC,YAAY1K,KAAKkE,GAAG,CAAC8C,cAAchH,KAAK0D,GAAG,CAACP,WAAW6D,eAAe,KAAK;QACjF,MAAMvE,eAAe,CAACR,cAAc,CAAC4H,WAAW,CAACC,YAAY,CAACC,YAAY,CAACC;QAE3E,MAAMW,YAAY;YAChB,IAAIlI,cAAc;gBAChB,MAAMC,aAAaxD;gBACnB,IAAIsD,WAAW;oBACb,IAAK,IAAIG,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;wBAClChD,QAAQ,AAACqC,OAAoBM,SAASyF,OAAOvF;oBAC/C;gBACF,OAAO;oBACL,IAAK,IAAIG,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;wBAClChD,QAAQqC,OAAOM,SAASyF;oBAC1B;gBACF;gBACA,OAAO,AAAE7I,CAAAA,OAAOwD,UAAS,IAAK3D,iBAAkB8C,OAAOE;YACzD;YAEA,IAAI8H,SAAS;gBACXA,QAAQvH,SAASyF;YACnB,OAAO,IAAI+B,UAAU;gBACnB,MAAMA,SAASxH,SAASyF;YAC1B;YAEA,MAAMpH,WAAWsB,aAAa,MAAM2H,aAActH,SAASyF,SAAS0B,YAAanH,SAASyF;YAE1F,IAAIgC,UAAU;gBACZA,SAASzH,SAASyF;YACpB,OAAO,IAAIiC,WAAW;gBACpB,MAAMA,UAAU1H,SAASyF;YAC3B;YAEA,OAAOpH,WAAW5B;QACpB;QAEA,MAAOH,YAAY2L,GAAG,KAAKD,cAAc,SAASE,kBAAkB,EAAG;YACrE,MAAM7J,WAAW,MAAMgK;YACvBhF,WAAW8E,cAAc3K,OAAOa,WAAW+J;YAC3CF;QACF;QACA,IAAII,aAAa;QACjB,MAAOA,aAAaJ,gBAAiB;YACnC,MAAM7J,WAAW,MAAMgK;YACvBhF,WAAW8E,cAAc3K,OAAOa,WAAW+J;YAC3CE;YACA,IAAIC,OAAOC,EAAE,IAAIF,aAAa1J,cAAc,GAAG;gBAC7C2J,OAAOC,EAAE;YACX;QACF;QACA,MAAOL,aAAavK,MAAM,IAAI,KAAKuK,aAAavK,MAAM,GAAGwK,UAAW;YAClE,MAAMrG,KAAKsC,SAAS8D;YACpB,IAAIpG,MAAMnB,eAAe,GAAG;gBAC1B;YACF;YACA,MAAMvC,WAAW,MAAMgK;YACvBhF,WAAW8E,cAAc3K,OAAOa,WAAW+J;QAC7C;QAEA,IAAI9I,IAAI;QACR,IAAIiC,OAAO,EAAE;QACb,IAAIkH,KAAK,EAAE;QACX,MAAMC,gBAA0B,EAAE;QAClC,IAAIC,UAAU;QACd,MAAMC,aAAa/H,YAAY;QAC/B,IAAIgI,mBAAmB;QAEvB,MAAO,KAAM;YACX,IAAIvJ,KAAKuB,WAAW;YACpB,IAAI,CAACgI,oBAAoBF,WAAWC,YAAY;gBAC9CE,QAAQC,KAAK,CAAC,CAAC,SAAS,EAAEJ,QAAQ,iDAAiD,CAAC,GAAG,CAAC,4EAA4E,CAAC;gBACrKE,mBAAmB;YACrB;YAEA,IAAIN,OAAOC,EAAE,IAAIlJ,IAAI,KAAKA,IAAIV,cAAc,GAAG;gBAC7C2J,OAAOC,EAAE;YACX;YAEA,MAAMQ,WAAWtD,WAAWvH;YAC5B,MAAM8K,cAAc3M,YAAY2L,GAAG;YACnC,IAAI3H,iBAAiB,EAAE;YAEvB,IAAIH,cAAc;gBAChB,MAAMC,aAAaxD;gBACnB,IAAIsD,WAAW;oBACb,IAAK,IAAIG,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;wBAClChD,QAAQ,AAACqC,OAAoBM,SAASyF,OAAOvF;oBAC/C;gBACF,OAAO;oBACL,IAAK,IAAIG,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;wBAClChD,QAAQqC,OAAOM,SAASyF;oBAC1B;gBACF;gBACA,MAAMyD,gBAAgBtM,OAAOwD;gBAC7BE,iBAAiB,AAAC4I,gBAAgBzM,iBAAkB8C,OAAOE;YAC7D,OAAO;gBACL,IAAK,IAAIY,IAAI,GAAGA,IAAIZ,WAAWY,IAAK;oBAClC,IAAIkH,SAAS;wBACXA,QAAQvH,SAASyF;oBACnB,OAAO,IAAI+B,UAAU;wBACnB,MAAMA,SAASxH,SAASyF;oBAC1B;oBAEA,MAAMpH,WAAWsB,aAAa,MAAM2H,aAActH,SAASyF,SAAS0B,YAAanH,SAASyF;oBAC1FnF,kBAAkBjC;oBAElB,IAAIoJ,UAAU;wBACZA,SAASzH,SAASyF;oBACpB,OAAO,IAAIiC,WAAW;wBACpB,MAAMA,UAAU1H,SAASyF;oBAC3B;gBACF;gBACAnF,iBAAiB,AAACA,iBAAiB7D,iBAAkB8C,OAAOE;YAC9D;YAEA,MAAM0J,YAAY7M,YAAY2L,GAAG;YACjC,IAAI,CAACY,kBAAkB;gBACrB,MAAMO,UAAU,AAACJ,CAAAA,WAAWtD,UAAW2D,IAAI,CAACL,YAAY,KAAI,KAAOrD,CAAAA,WAAW5C,SAASkG,aAAaE,cAAc,KAAI;gBACtH,IAAIC,SAAS;oBACXT;oBACA;gBACF;YACF;YAEA,MAAMW,iBAAiB9L,OAAO8C;YAC9B+C,WAAWqF,eAAeY,gBAAgBvK;YAC1C,IAAI,CAAC8J,kBAAkB;gBACrB,MAAM,EAAEnF,MAAM,EAAEC,GAAG,EAAE,GAAGF,aAAaiF;gBACrC,MAAMa,aAAa7F,SAAS5E,yBAAyB6E,OAAOnG,OAAOuD,iBAAiB;gBACpF,IAAI2H,cAAc9K,MAAM,IAAI,KAAK0L,iBAAiBC,cAAcD,iBAAiB5F,SAAS1E,uBAAuB;oBAC/G2J;oBACA;gBACF;gBAEA,MAAMa,aAAahM,OAAO+D;gBAC1B,IAAIjC,KAAK,KAAKkK,aAAa,KAAKF,iBAAiBzK,qBAAqB2K,cAAcF,iBAAiBE,aAAaxK,uBAAuB;oBACvI2J;oBACA;gBACF;YACF;YAEA5D,SAAS,CAACzF,IAAI,GAAGgB;YACjB,MAAMmJ,QAAQnJ,iBAAiBiB;YAC/BA,QAAQkI,QAAQlK,OAAOD;YACvBmJ,MAAMgB,QAASnJ,CAAAA,iBAAiBiB,IAAG;YAEnC,MAAMmI,WAAW,AAACpK,IAAIuB,YAAalE;YACnC,IAAI2C,IAAIX,oBAAoB,GAAG;gBAC7BsG,OAAO,CAACzI,QAAQ2I,QAAQ,CAAC,GAAGuE;YAC9B;YAEA,IAAIpK,KAAKqB,WAAW;gBAClB,MAAMgB,WAAWnE,OAAOiL,MAAOnJ,CAAAA,IAAI,CAAA;gBACnC,MAAMuC,SAASnE,KAAKoE,IAAI,CAACH;gBACzB,IAAIE,UAAUrE,OAAOmH,eAAe;oBAClC;gBACF;gBAEA,MAAMgF,UAAUnM,OAAO+D;gBACvB,MAAMqI,MAAM/H,SAAU8H,CAAAA,WAAW,CAAA;gBACjC,IAAIC,OAAOhJ,cAAc;oBACvB;gBACF;YACF;QACF;QAEAqE,OAAO,CAACzI,QAAQ0I,KAAK,CAAC,GAAG5F;QACzB2F,OAAO,CAACzI,QAAQ4I,QAAQ,CAAC,GAAG;QAC5B,MAAMyE,YAAYhN,QAAQ0I,WAAW,GAAGC,QAAQ;QAChDP,OAAO,CAACzI,QAAQ6I,SAAS,CAAC,GAAG3H,KAAKkE,GAAG,CAAC,GAAGlE,KAAKoM,KAAK,CAAC,AAACD,CAAAA,YAAYvE,UAAS,IAAK;IACjF,EAAE,OAAOyE,GAAG;QACVjB,QAAQC,KAAK,CAACgB,KAAK,OAAOA,MAAM,YAAY,WAAWA,IAAIA,EAAEC,KAAK,GAAGD;QACrE9E,OAAO,CAACzI,QAAQ4I,QAAQ,CAAC,GAAG;IAC9B,SAAU;QACRO,WAAWxC;QACX,IAAI;YACF,MAAMsB,WAAWzE;QACnB,EAAE,OAAO+J,GAAG;YACV9E,OAAO,CAACzI,QAAQ4I,QAAQ,CAAC,GAAG;YAC5B0D,QAAQC,KAAK,CAACgB,KAAK,OAAOA,MAAM,YAAY,WAAWA,IAAIA,EAAEC,KAAK,GAAGD;QACvE;IACF;IAEA,OAAO9E,OAAO,CAACzI,QAAQ4I,QAAQ,CAAC;AAClC,EAAE"}
|
package/build/types.cjs
DELETED
|
@@ -1,66 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
Object.defineProperty(exports, "__esModule", {
|
|
3
|
-
value: true
|
|
4
|
-
});
|
|
5
|
-
function _export(target, all) {
|
|
6
|
-
for(var name in all)Object.defineProperty(target, name, {
|
|
7
|
-
enumerable: true,
|
|
8
|
-
get: Object.getOwnPropertyDescriptor(all, name).get
|
|
9
|
-
});
|
|
10
|
-
}
|
|
11
|
-
_export(exports, {
|
|
12
|
-
get COMPLETE_VALUE () {
|
|
13
|
-
return COMPLETE_VALUE;
|
|
14
|
-
},
|
|
15
|
-
get CONTROL_SLOTS () {
|
|
16
|
-
return CONTROL_SLOTS;
|
|
17
|
-
},
|
|
18
|
-
get Control () {
|
|
19
|
-
return Control;
|
|
20
|
-
},
|
|
21
|
-
get DEFAULT_CYCLES () {
|
|
22
|
-
return DEFAULT_CYCLES;
|
|
23
|
-
},
|
|
24
|
-
get DURATION_SCALE () {
|
|
25
|
-
return DURATION_SCALE;
|
|
26
|
-
},
|
|
27
|
-
get REPORT_TYPES () {
|
|
28
|
-
return REPORT_TYPES;
|
|
29
|
-
},
|
|
30
|
-
get Z95 () {
|
|
31
|
-
return Z95;
|
|
32
|
-
}
|
|
33
|
-
});
|
|
34
|
-
const REPORT_TYPES = Array.from({
|
|
35
|
-
length: 99
|
|
36
|
-
}, (_, idx)=>`p${idx + 1}`).concat([
|
|
37
|
-
'ops',
|
|
38
|
-
'mean',
|
|
39
|
-
'min',
|
|
40
|
-
'max',
|
|
41
|
-
'median',
|
|
42
|
-
'mode',
|
|
43
|
-
'variance',
|
|
44
|
-
'sd',
|
|
45
|
-
'sem',
|
|
46
|
-
'moe',
|
|
47
|
-
'rme',
|
|
48
|
-
'mad',
|
|
49
|
-
'iqr',
|
|
50
|
-
'ci_lower',
|
|
51
|
-
'ci_upper'
|
|
52
|
-
]);
|
|
53
|
-
var Control = /*#__PURE__*/ function(Control) {
|
|
54
|
-
Control[Control["INDEX"] = 0] = "INDEX";
|
|
55
|
-
Control[Control["PROGRESS"] = 1] = "PROGRESS";
|
|
56
|
-
Control[Control["COMPLETE"] = 2] = "COMPLETE";
|
|
57
|
-
Control[Control["HEAP_USED"] = 3] = "HEAP_USED";
|
|
58
|
-
return Control;
|
|
59
|
-
}({});
|
|
60
|
-
const CONTROL_SLOTS = Object.values(Control).length / 2;
|
|
61
|
-
const DEFAULT_CYCLES = 1_000;
|
|
62
|
-
const Z95 = 1.96;
|
|
63
|
-
const DURATION_SCALE = 1000n;
|
|
64
|
-
const COMPLETE_VALUE = 100_00;
|
|
65
|
-
|
|
66
|
-
//# sourceMappingURL=types.cjs.map
|