@aztec/prover-node 0.0.1-commit.b655e406 → 0.0.1-commit.c31f2472
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/actions/download-epoch-proving-job.d.ts +4 -4
- package/dest/actions/index.d.ts +1 -1
- package/dest/actions/rerun-epoch-proving-job.d.ts +3 -2
- package/dest/actions/rerun-epoch-proving-job.d.ts.map +1 -1
- package/dest/actions/rerun-epoch-proving-job.js +5 -3
- package/dest/actions/upload-epoch-proof-failure.d.ts +1 -1
- package/dest/bin/run-failed-epoch.d.ts +1 -1
- package/dest/config.d.ts +5 -4
- package/dest/config.d.ts.map +1 -1
- package/dest/config.js +4 -3
- package/dest/factory.d.ts +2 -4
- package/dest/factory.d.ts.map +1 -1
- package/dest/factory.js +21 -16
- package/dest/index.d.ts +2 -1
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +1 -0
- package/dest/job/epoch-proving-job-data.d.ts +8 -6
- package/dest/job/epoch-proving-job-data.d.ts.map +1 -1
- package/dest/job/epoch-proving-job-data.js +25 -18
- package/dest/job/epoch-proving-job.d.ts +7 -13
- package/dest/job/epoch-proving-job.d.ts.map +1 -1
- package/dest/job/epoch-proving-job.js +486 -99
- package/dest/metrics.d.ts +4 -3
- package/dest/metrics.d.ts.map +1 -1
- package/dest/metrics.js +30 -98
- package/dest/monitors/epoch-monitor.d.ts +3 -2
- package/dest/monitors/epoch-monitor.d.ts.map +1 -1
- package/dest/monitors/epoch-monitor.js +3 -11
- package/dest/monitors/index.d.ts +1 -1
- package/dest/prover-node-publisher.d.ts +12 -9
- package/dest/prover-node-publisher.d.ts.map +1 -1
- package/dest/prover-node-publisher.js +47 -40
- package/dest/prover-node.d.ts +9 -8
- package/dest/prover-node.d.ts.map +1 -1
- package/dest/prover-node.js +432 -51
- package/dest/prover-publisher-factory.d.ts +7 -3
- package/dest/prover-publisher-factory.d.ts.map +1 -1
- package/dest/prover-publisher-factory.js +4 -2
- package/dest/test/index.d.ts +1 -1
- package/dest/test/index.d.ts.map +1 -1
- package/package.json +26 -25
- package/src/actions/rerun-epoch-proving-job.ts +5 -3
- package/src/bin/run-failed-epoch.ts +1 -1
- package/src/config.ts +6 -4
- package/src/factory.ts +31 -18
- package/src/index.ts +1 -0
- package/src/job/epoch-proving-job-data.ts +31 -25
- package/src/job/epoch-proving-job.ts +114 -99
- package/src/metrics.ts +31 -82
- package/src/monitors/epoch-monitor.ts +5 -11
- package/src/prover-node-publisher.ts +68 -55
- package/src/prover-node.ts +49 -43
- package/src/prover-publisher-factory.ts +14 -6
|
@@ -1,15 +1,18 @@
|
|
|
1
1
|
import { NUMBER_OF_L1_L2_MESSAGES_PER_ROLLUP } from '@aztec/constants';
|
|
2
2
|
import { asyncPool } from '@aztec/foundation/async-pool';
|
|
3
|
+
import { BlockNumber, EpochNumber } from '@aztec/foundation/branded-types';
|
|
3
4
|
import { padArrayEnd } from '@aztec/foundation/collection';
|
|
4
|
-
import { Fr } from '@aztec/foundation/
|
|
5
|
-
import { createLogger } from '@aztec/foundation/log';
|
|
5
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
6
|
+
import { type Logger, type LoggerBindings, createLogger } from '@aztec/foundation/log';
|
|
6
7
|
import { RunningPromise, promiseWithResolvers } from '@aztec/foundation/promise';
|
|
7
8
|
import { Timer } from '@aztec/foundation/timer';
|
|
8
9
|
import { getVKTreeRoot } from '@aztec/noir-protocol-circuits-types/vk-tree';
|
|
9
10
|
import { protocolContractsHash } from '@aztec/protocol-contracts';
|
|
10
11
|
import { buildFinalBlobChallenges } from '@aztec/prover-client/helpers';
|
|
11
12
|
import type { PublicProcessor, PublicProcessorFactory } from '@aztec/simulator/server';
|
|
13
|
+
import { PublicSimulatorConfig } from '@aztec/stdlib/avm';
|
|
12
14
|
import type { L2Block, L2BlockSource } from '@aztec/stdlib/block';
|
|
15
|
+
import type { Checkpoint } from '@aztec/stdlib/checkpoint';
|
|
13
16
|
import {
|
|
14
17
|
type EpochProver,
|
|
15
18
|
type EpochProvingJobState,
|
|
@@ -40,7 +43,7 @@ export type EpochProvingJobOptions = {
|
|
|
40
43
|
*/
|
|
41
44
|
export class EpochProvingJob implements Traceable {
|
|
42
45
|
private state: EpochProvingJobState = 'initialized';
|
|
43
|
-
private log
|
|
46
|
+
private log: Logger;
|
|
44
47
|
private uuid: string;
|
|
45
48
|
|
|
46
49
|
private runPromise: Promise<void> | undefined;
|
|
@@ -59,9 +62,14 @@ export class EpochProvingJob implements Traceable {
|
|
|
59
62
|
private metrics: ProverNodeJobMetrics,
|
|
60
63
|
private deadline: Date | undefined,
|
|
61
64
|
private config: EpochProvingJobOptions,
|
|
65
|
+
bindings?: LoggerBindings,
|
|
62
66
|
) {
|
|
63
67
|
validateEpochProvingJobData(data);
|
|
64
68
|
this.uuid = crypto.randomUUID();
|
|
69
|
+
this.log = createLogger('prover-node:epoch-proving-job', {
|
|
70
|
+
...bindings,
|
|
71
|
+
instanceId: `epoch-${data.epochNumber}`,
|
|
72
|
+
});
|
|
65
73
|
this.tracer = metrics.tracer;
|
|
66
74
|
}
|
|
67
75
|
|
|
@@ -73,7 +81,7 @@ export class EpochProvingJob implements Traceable {
|
|
|
73
81
|
return this.state;
|
|
74
82
|
}
|
|
75
83
|
|
|
76
|
-
public getEpochNumber():
|
|
84
|
+
public getEpochNumber(): EpochNumber {
|
|
77
85
|
return this.data.epochNumber;
|
|
78
86
|
}
|
|
79
87
|
|
|
@@ -89,8 +97,8 @@ export class EpochProvingJob implements Traceable {
|
|
|
89
97
|
return this.data.epochNumber;
|
|
90
98
|
}
|
|
91
99
|
|
|
92
|
-
private get
|
|
93
|
-
return this.data.
|
|
100
|
+
private get checkpoints() {
|
|
101
|
+
return this.data.checkpoints;
|
|
94
102
|
}
|
|
95
103
|
|
|
96
104
|
private get txs() {
|
|
@@ -105,7 +113,7 @@ export class EpochProvingJob implements Traceable {
|
|
|
105
113
|
* Proves the given epoch and submits the proof to L1.
|
|
106
114
|
*/
|
|
107
115
|
@trackSpan('EpochProvingJob.run', function () {
|
|
108
|
-
return { [Attributes.EPOCH_NUMBER]:
|
|
116
|
+
return { [Attributes.EPOCH_NUMBER]: this.data.epochNumber };
|
|
109
117
|
})
|
|
110
118
|
public async run() {
|
|
111
119
|
this.scheduleDeadlineStop();
|
|
@@ -114,14 +122,22 @@ export class EpochProvingJob implements Traceable {
|
|
|
114
122
|
}
|
|
115
123
|
|
|
116
124
|
const attestations = this.attestations.map(attestation => attestation.toViem());
|
|
117
|
-
const epochNumber =
|
|
118
|
-
const
|
|
119
|
-
const
|
|
120
|
-
const
|
|
121
|
-
|
|
125
|
+
const epochNumber = this.epochNumber;
|
|
126
|
+
const epochSizeCheckpoints = this.checkpoints.length;
|
|
127
|
+
const epochSizeBlocks = this.checkpoints.reduce((accum, checkpoint) => accum + checkpoint.blocks.length, 0);
|
|
128
|
+
const epochSizeTxs = this.checkpoints.reduce(
|
|
129
|
+
(accum, checkpoint) =>
|
|
130
|
+
accum + checkpoint.blocks.reduce((accumC, block) => accumC + block.body.txEffects.length, 0),
|
|
131
|
+
0,
|
|
132
|
+
);
|
|
133
|
+
const fromCheckpoint = this.checkpoints[0].number;
|
|
134
|
+
const toCheckpoint = this.checkpoints.at(-1)!.number;
|
|
135
|
+
const fromBlock = this.checkpoints[0].blocks[0].number;
|
|
136
|
+
const toBlock = this.checkpoints.at(-1)!.blocks.at(-1)!.number;
|
|
137
|
+
this.log.info(`Starting epoch ${epochNumber} proving job with checkpoints ${fromCheckpoint} to ${toCheckpoint}`, {
|
|
122
138
|
fromBlock,
|
|
123
139
|
toBlock,
|
|
124
|
-
|
|
140
|
+
epochSizeTxs,
|
|
125
141
|
epochNumber,
|
|
126
142
|
uuid: this.uuid,
|
|
127
143
|
});
|
|
@@ -132,83 +148,93 @@ export class EpochProvingJob implements Traceable {
|
|
|
132
148
|
this.runPromise = promise;
|
|
133
149
|
|
|
134
150
|
try {
|
|
135
|
-
const blobFieldsPerCheckpoint = this.
|
|
151
|
+
const blobFieldsPerCheckpoint = this.checkpoints.map(checkpoint => checkpoint.toBlobFields());
|
|
136
152
|
const finalBlobBatchingChallenges = await buildFinalBlobChallenges(blobFieldsPerCheckpoint);
|
|
137
153
|
|
|
138
|
-
|
|
139
|
-
// Total number of checkpoints equals number of blocks because we currently build a checkpoint with only one block.
|
|
140
|
-
const totalNumCheckpoints = epochSizeBlocks;
|
|
141
|
-
|
|
142
|
-
this.prover.startNewEpoch(epochNumber, totalNumCheckpoints, finalBlobBatchingChallenges);
|
|
154
|
+
this.prover.startNewEpoch(epochNumber, epochSizeCheckpoints, finalBlobBatchingChallenges);
|
|
143
155
|
await this.prover.startChonkVerifierCircuits(Array.from(this.txs.values()));
|
|
144
156
|
|
|
145
|
-
|
|
146
|
-
|
|
157
|
+
// Everything in the epoch should have the same chainId and version.
|
|
158
|
+
const { chainId, version } = this.checkpoints[0].blocks[0].header.globalVariables;
|
|
147
159
|
|
|
148
|
-
|
|
149
|
-
const txs = this.getTxs(block);
|
|
150
|
-
const l1ToL2Messages = this.getL1ToL2Messages(block);
|
|
151
|
-
const previousHeader = this.getBlockHeader(block.number - 1)!;
|
|
152
|
-
|
|
153
|
-
this.log.verbose(`Starting processing block ${block.number}`, {
|
|
154
|
-
number: block.number,
|
|
155
|
-
blockHash: (await block.hash()).toString(),
|
|
156
|
-
lastArchive: block.header.lastArchive.root,
|
|
157
|
-
noteHashTreeRoot: block.header.state.partial.noteHashTree.root,
|
|
158
|
-
nullifierTreeRoot: block.header.state.partial.nullifierTree.root,
|
|
159
|
-
publicDataTreeRoot: block.header.state.partial.publicDataTree.root,
|
|
160
|
-
previousHeader: previousHeader.hash(),
|
|
161
|
-
uuid: this.uuid,
|
|
162
|
-
...globalVariables,
|
|
163
|
-
});
|
|
160
|
+
const previousBlockHeaders = this.gatherPreviousBlockHeaders();
|
|
164
161
|
|
|
162
|
+
await asyncPool(this.config.parallelBlockLimit ?? 32, this.checkpoints, async checkpoint => {
|
|
163
|
+
this.checkState();
|
|
164
|
+
|
|
165
|
+
const checkpointIndex = checkpoint.number - fromCheckpoint;
|
|
165
166
|
const checkpointConstants = CheckpointConstantData.from({
|
|
166
|
-
chainId
|
|
167
|
-
version
|
|
167
|
+
chainId,
|
|
168
|
+
version,
|
|
168
169
|
vkTreeRoot: getVKTreeRoot(),
|
|
169
170
|
protocolContractsHash: protocolContractsHash,
|
|
170
171
|
proverId: this.prover.getProverId().toField(),
|
|
171
|
-
slotNumber:
|
|
172
|
-
coinbase:
|
|
173
|
-
feeRecipient:
|
|
174
|
-
gasFees:
|
|
172
|
+
slotNumber: checkpoint.header.slotNumber,
|
|
173
|
+
coinbase: checkpoint.header.coinbase,
|
|
174
|
+
feeRecipient: checkpoint.header.feeRecipient,
|
|
175
|
+
gasFees: checkpoint.header.gasFees,
|
|
176
|
+
});
|
|
177
|
+
const previousHeader = previousBlockHeaders[checkpointIndex];
|
|
178
|
+
const l1ToL2Messages = this.getL1ToL2Messages(checkpoint);
|
|
179
|
+
|
|
180
|
+
this.log.verbose(`Starting processing checkpoint ${checkpoint.number}`, {
|
|
181
|
+
number: checkpoint.number,
|
|
182
|
+
checkpointHash: checkpoint.hash().toString(),
|
|
183
|
+
lastArchive: checkpoint.header.lastArchiveRoot,
|
|
184
|
+
previousHeader: previousHeader.hash(),
|
|
185
|
+
uuid: this.uuid,
|
|
175
186
|
});
|
|
176
187
|
|
|
177
|
-
// TODO(#17027): Enable multiple blocks per checkpoint.
|
|
178
|
-
// Each checkpoint has only one block.
|
|
179
|
-
const totalNumBlocks = 1;
|
|
180
|
-
const checkpointIndex = block.number - fromBlock;
|
|
181
188
|
await this.prover.startNewCheckpoint(
|
|
182
189
|
checkpointIndex,
|
|
183
190
|
checkpointConstants,
|
|
184
191
|
l1ToL2Messages,
|
|
185
|
-
|
|
186
|
-
blobFieldsPerCheckpoint[checkpointIndex].length,
|
|
192
|
+
checkpoint.blocks.length,
|
|
187
193
|
previousHeader,
|
|
188
194
|
);
|
|
189
195
|
|
|
190
|
-
|
|
191
|
-
|
|
196
|
+
for (const block of checkpoint.blocks) {
|
|
197
|
+
const globalVariables = block.header.globalVariables;
|
|
198
|
+
const txs = this.getTxs(block);
|
|
199
|
+
|
|
200
|
+
this.log.verbose(`Starting processing block ${block.number}`, {
|
|
201
|
+
number: block.number,
|
|
202
|
+
blockHash: (await block.hash()).toString(),
|
|
203
|
+
lastArchive: block.header.lastArchive.root,
|
|
204
|
+
noteHashTreeRoot: block.header.state.partial.noteHashTree.root,
|
|
205
|
+
nullifierTreeRoot: block.header.state.partial.nullifierTree.root,
|
|
206
|
+
publicDataTreeRoot: block.header.state.partial.publicDataTree.root,
|
|
207
|
+
...globalVariables,
|
|
208
|
+
numTxs: txs.length,
|
|
209
|
+
});
|
|
192
210
|
|
|
193
|
-
|
|
194
|
-
|
|
195
|
-
|
|
196
|
-
|
|
197
|
-
|
|
198
|
-
|
|
199
|
-
|
|
200
|
-
|
|
201
|
-
|
|
202
|
-
|
|
203
|
-
|
|
204
|
-
|
|
205
|
-
|
|
206
|
-
|
|
207
|
-
|
|
211
|
+
// Start block proving
|
|
212
|
+
await this.prover.startNewBlock(block.number, globalVariables.timestamp, txs.length);
|
|
213
|
+
|
|
214
|
+
// Process public fns
|
|
215
|
+
const db = await this.createFork(BlockNumber(block.number - 1), l1ToL2Messages);
|
|
216
|
+
const config = PublicSimulatorConfig.from({
|
|
217
|
+
proverId: this.prover.getProverId().toField(),
|
|
218
|
+
skipFeeEnforcement: false,
|
|
219
|
+
collectDebugLogs: false,
|
|
220
|
+
collectHints: true,
|
|
221
|
+
collectPublicInputs: true,
|
|
222
|
+
collectStatistics: false,
|
|
223
|
+
});
|
|
224
|
+
const publicProcessor = this.publicProcessorFactory.create(db, globalVariables, config);
|
|
225
|
+
const processed = await this.processTxs(publicProcessor, txs);
|
|
226
|
+
await this.prover.addTxs(processed);
|
|
227
|
+
await db.close();
|
|
228
|
+
this.log.verbose(`Processed all ${txs.length} txs for block ${block.number}`, {
|
|
229
|
+
blockNumber: block.number,
|
|
230
|
+
blockHash: (await block.hash()).toString(),
|
|
231
|
+
uuid: this.uuid,
|
|
232
|
+
});
|
|
208
233
|
|
|
209
|
-
|
|
210
|
-
|
|
211
|
-
|
|
234
|
+
// Mark block as completed to pad it
|
|
235
|
+
const expectedBlockHeader = block.header;
|
|
236
|
+
await this.prover.setBlockCompleted(block.number, expectedBlockHeader);
|
|
237
|
+
}
|
|
212
238
|
});
|
|
213
239
|
|
|
214
240
|
const executionTime = timer.ms();
|
|
@@ -221,16 +247,16 @@ export class EpochProvingJob implements Traceable {
|
|
|
221
247
|
|
|
222
248
|
if (this.config.skipSubmitProof) {
|
|
223
249
|
this.log.info(
|
|
224
|
-
`Proof publishing is disabled. Dropping valid proof for epoch ${epochNumber} (
|
|
250
|
+
`Proof publishing is disabled. Dropping valid proof for epoch ${epochNumber} (checkpoints ${fromCheckpoint} to ${toCheckpoint})`,
|
|
225
251
|
);
|
|
226
252
|
this.state = 'completed';
|
|
227
|
-
this.metrics.recordProvingJob(executionTime, timer.ms(), epochSizeBlocks, epochSizeTxs);
|
|
253
|
+
this.metrics.recordProvingJob(executionTime, timer.ms(), epochSizeCheckpoints, epochSizeBlocks, epochSizeTxs);
|
|
228
254
|
return;
|
|
229
255
|
}
|
|
230
256
|
|
|
231
257
|
const success = await this.publisher.submitEpochProof({
|
|
232
|
-
|
|
233
|
-
|
|
258
|
+
fromCheckpoint,
|
|
259
|
+
toCheckpoint,
|
|
234
260
|
epochNumber,
|
|
235
261
|
publicInputs,
|
|
236
262
|
proof,
|
|
@@ -241,12 +267,12 @@ export class EpochProvingJob implements Traceable {
|
|
|
241
267
|
throw new Error('Failed to submit epoch proof to L1');
|
|
242
268
|
}
|
|
243
269
|
|
|
244
|
-
this.log.info(`Submitted proof for epoch ${epochNumber} (
|
|
270
|
+
this.log.info(`Submitted proof for epoch ${epochNumber} (checkpoints ${fromCheckpoint} to ${toCheckpoint})`, {
|
|
245
271
|
epochNumber,
|
|
246
272
|
uuid: this.uuid,
|
|
247
273
|
});
|
|
248
274
|
this.state = 'completed';
|
|
249
|
-
this.metrics.recordProvingJob(executionTime, timer.ms(), epochSizeBlocks, epochSizeTxs);
|
|
275
|
+
this.metrics.recordProvingJob(executionTime, timer.ms(), epochSizeCheckpoints, epochSizeBlocks, epochSizeTxs);
|
|
250
276
|
} catch (err: any) {
|
|
251
277
|
if (err && err.name === 'HaltExecutionError') {
|
|
252
278
|
this.log.warn(`Halted execution of epoch ${epochNumber} prover job`, {
|
|
@@ -272,7 +298,7 @@ export class EpochProvingJob implements Traceable {
|
|
|
272
298
|
* Create a new db fork for tx processing, inserting all L1 to L2.
|
|
273
299
|
* REFACTOR: The prover already spawns a db fork of its own for each block, so we may be able to do away with just one fork.
|
|
274
300
|
*/
|
|
275
|
-
private async createFork(blockNumber:
|
|
301
|
+
private async createFork(blockNumber: BlockNumber, l1ToL2Messages: Fr[]) {
|
|
276
302
|
const db = await this.dbProvider.fork(blockNumber);
|
|
277
303
|
const l1ToL2MessagesPadded = padArrayEnd<Fr, number>(
|
|
278
304
|
l1ToL2Messages,
|
|
@@ -282,7 +308,7 @@ export class EpochProvingJob implements Traceable {
|
|
|
282
308
|
);
|
|
283
309
|
this.log.verbose(`Creating fork at ${blockNumber} with ${l1ToL2Messages.length} L1 to L2 messages`, {
|
|
284
310
|
blockNumber,
|
|
285
|
-
l1ToL2Messages:
|
|
311
|
+
l1ToL2Messages: l1ToL2Messages.map(m => m.toString()),
|
|
286
312
|
});
|
|
287
313
|
await db.appendLeaves(MerkleTreeId.L1_TO_L2_MESSAGE_TREE, l1ToL2MessagesPadded);
|
|
288
314
|
return db;
|
|
@@ -341,11 +367,12 @@ export class EpochProvingJob implements Traceable {
|
|
|
341
367
|
const intervalMs = Math.ceil((await l2BlockSource.getL1Constants()).ethereumSlotDuration / 2) * 1000;
|
|
342
368
|
this.epochCheckPromise = new RunningPromise(
|
|
343
369
|
async () => {
|
|
344
|
-
const
|
|
345
|
-
const blockHashes = await Promise.all(
|
|
346
|
-
const
|
|
370
|
+
const blockHeaders = await l2BlockSource.getCheckpointedBlockHeadersForEpoch(this.epochNumber);
|
|
371
|
+
const blockHashes = await Promise.all(blockHeaders.map(async header => (await header.hash()).toField()));
|
|
372
|
+
const thisBlocks = this.checkpoints.flatMap(checkpoint => checkpoint.blocks);
|
|
373
|
+
const thisBlockHashes = await Promise.all(thisBlocks.map(block => block.hash()));
|
|
347
374
|
if (
|
|
348
|
-
|
|
375
|
+
blockHeaders.length !== thisBlocks.length ||
|
|
349
376
|
!blockHashes.every((block, i) => block.equals(thisBlockHashes[i]))
|
|
350
377
|
) {
|
|
351
378
|
this.log.warn('Epoch blocks changed underfoot', {
|
|
@@ -363,30 +390,18 @@ export class EpochProvingJob implements Traceable {
|
|
|
363
390
|
this.log.verbose(`Scheduled epoch check for epoch ${this.epochNumber} every ${intervalMs}ms`);
|
|
364
391
|
}
|
|
365
392
|
|
|
366
|
-
/* Returns the header
|
|
367
|
-
private
|
|
368
|
-
const
|
|
369
|
-
|
|
370
|
-
return block.getBlockHeader();
|
|
371
|
-
}
|
|
372
|
-
|
|
373
|
-
if (blockNumber === Number(this.data.previousBlockHeader.getBlockNumber())) {
|
|
374
|
-
return this.data.previousBlockHeader;
|
|
375
|
-
}
|
|
376
|
-
|
|
377
|
-
throw new Error(
|
|
378
|
-
`Block header not found for block number ${blockNumber} (got ${this.blocks
|
|
379
|
-
.map(b => b.number)
|
|
380
|
-
.join(', ')} and previous header ${this.data.previousBlockHeader.getBlockNumber()})`,
|
|
381
|
-
);
|
|
393
|
+
/* Returns the last block header in the previous checkpoint for all checkpoints in the epoch */
|
|
394
|
+
private gatherPreviousBlockHeaders() {
|
|
395
|
+
const lastBlocks = this.checkpoints.map(checkpoint => checkpoint.blocks.at(-1)!);
|
|
396
|
+
return [this.data.previousBlockHeader, ...lastBlocks.map(block => block.header).slice(0, -1)];
|
|
382
397
|
}
|
|
383
398
|
|
|
384
399
|
private getTxs(block: L2Block): Tx[] {
|
|
385
400
|
return block.body.txEffects.map(txEffect => this.txs.get(txEffect.txHash.toString())!);
|
|
386
401
|
}
|
|
387
402
|
|
|
388
|
-
private getL1ToL2Messages(
|
|
389
|
-
return this.data.l1ToL2Messages[
|
|
403
|
+
private getL1ToL2Messages(checkpoint: Checkpoint) {
|
|
404
|
+
return this.data.l1ToL2Messages[checkpoint.number];
|
|
390
405
|
}
|
|
391
406
|
|
|
392
407
|
private async processTxs(publicProcessor: PublicProcessor, txs: Tx[]): Promise<ProcessedTx[]> {
|
package/src/metrics.ts
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import type { RollupContract } from '@aztec/ethereum';
|
|
1
|
+
import type { RollupContract } from '@aztec/ethereum/contracts';
|
|
2
2
|
import type { EthAddress } from '@aztec/foundation/eth-address';
|
|
3
3
|
import { createLogger } from '@aztec/foundation/log';
|
|
4
4
|
import type { L1PublishProofStats, L1PublishStats } from '@aztec/stdlib/stats';
|
|
@@ -13,7 +13,7 @@ import {
|
|
|
13
13
|
type TelemetryClient,
|
|
14
14
|
type Tracer,
|
|
15
15
|
type UpDownCounter,
|
|
16
|
-
|
|
16
|
+
createUpDownCounterWithDefault,
|
|
17
17
|
} from '@aztec/telemetry-client';
|
|
18
18
|
|
|
19
19
|
import { formatEther, formatUnits } from 'viem';
|
|
@@ -21,6 +21,7 @@ import { formatEther, formatUnits } from 'viem';
|
|
|
21
21
|
export class ProverNodeJobMetrics {
|
|
22
22
|
proverEpochExecutionDuration: Histogram;
|
|
23
23
|
provingJobDuration: Histogram;
|
|
24
|
+
provingJobCheckpoints: Gauge;
|
|
24
25
|
provingJobBlocks: Gauge;
|
|
25
26
|
provingJobTransactions: Gauge;
|
|
26
27
|
|
|
@@ -29,29 +30,23 @@ export class ProverNodeJobMetrics {
|
|
|
29
30
|
public readonly tracer: Tracer,
|
|
30
31
|
private logger = createLogger('prover-node:publisher:metrics'),
|
|
31
32
|
) {
|
|
32
|
-
this.proverEpochExecutionDuration = this.meter.createHistogram(Metrics.PROVER_NODE_EXECUTION_DURATION
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
this.provingJobDuration = this.meter.createHistogram(Metrics.PROVER_NODE_JOB_DURATION, {
|
|
38
|
-
description: 'Duration of proving job',
|
|
39
|
-
unit: 's',
|
|
40
|
-
valueType: ValueType.DOUBLE,
|
|
41
|
-
});
|
|
42
|
-
this.provingJobBlocks = this.meter.createGauge(Metrics.PROVER_NODE_JOB_BLOCKS, {
|
|
43
|
-
description: 'Number of blocks in a proven epoch',
|
|
44
|
-
valueType: ValueType.INT,
|
|
45
|
-
});
|
|
46
|
-
this.provingJobTransactions = this.meter.createGauge(Metrics.PROVER_NODE_JOB_TRANSACTIONS, {
|
|
47
|
-
description: 'Number of transactions in a proven epoch',
|
|
48
|
-
valueType: ValueType.INT,
|
|
49
|
-
});
|
|
33
|
+
this.proverEpochExecutionDuration = this.meter.createHistogram(Metrics.PROVER_NODE_EXECUTION_DURATION);
|
|
34
|
+
this.provingJobDuration = this.meter.createHistogram(Metrics.PROVER_NODE_JOB_DURATION);
|
|
35
|
+
this.provingJobCheckpoints = this.meter.createGauge(Metrics.PROVER_NODE_JOB_CHECKPOINTS);
|
|
36
|
+
this.provingJobBlocks = this.meter.createGauge(Metrics.PROVER_NODE_JOB_BLOCKS);
|
|
37
|
+
this.provingJobTransactions = this.meter.createGauge(Metrics.PROVER_NODE_JOB_TRANSACTIONS);
|
|
50
38
|
}
|
|
51
39
|
|
|
52
|
-
public recordProvingJob(
|
|
40
|
+
public recordProvingJob(
|
|
41
|
+
executionTimeMs: number,
|
|
42
|
+
totalTimeMs: number,
|
|
43
|
+
numCheckpoints: number,
|
|
44
|
+
numBlocks: number,
|
|
45
|
+
numTxs: number,
|
|
46
|
+
) {
|
|
53
47
|
this.proverEpochExecutionDuration.record(Math.ceil(executionTimeMs));
|
|
54
48
|
this.provingJobDuration.record(totalTimeMs / 1000);
|
|
49
|
+
this.provingJobCheckpoints.record(Math.floor(numCheckpoints));
|
|
55
50
|
this.provingJobBlocks.record(Math.floor(numBlocks));
|
|
56
51
|
this.provingJobTransactions.record(Math.floor(numTxs));
|
|
57
52
|
}
|
|
@@ -69,15 +64,9 @@ export class ProverNodeRewardsMetrics {
|
|
|
69
64
|
private rollup: RollupContract,
|
|
70
65
|
private logger = createLogger('prover-node:publisher:metrics'),
|
|
71
66
|
) {
|
|
72
|
-
this.rewards = this.meter.createObservableGauge(Metrics.PROVER_NODE_REWARDS_PER_EPOCH
|
|
73
|
-
valueType: ValueType.DOUBLE,
|
|
74
|
-
description: 'The rewards earned',
|
|
75
|
-
});
|
|
67
|
+
this.rewards = this.meter.createObservableGauge(Metrics.PROVER_NODE_REWARDS_PER_EPOCH);
|
|
76
68
|
|
|
77
|
-
this.accumulatedRewards = this.meter
|
|
78
|
-
valueType: ValueType.DOUBLE,
|
|
79
|
-
description: 'The rewards earned (total)',
|
|
80
|
-
});
|
|
69
|
+
this.accumulatedRewards = createUpDownCounterWithDefault(this.meter, Metrics.PROVER_NODE_REWARDS_TOTAL);
|
|
81
70
|
}
|
|
82
71
|
|
|
83
72
|
public async start() {
|
|
@@ -97,7 +86,7 @@ export class ProverNodeRewardsMetrics {
|
|
|
97
86
|
// look at the prev epoch so that we get an accurate value, after proof submission window has closed
|
|
98
87
|
// For example, if proof submission window is 1 epoch, and we are in epoch 2, we should be looking at epoch 0.
|
|
99
88
|
// Similarly, if the proof submission window is 0, and we are in epoch 1, we should be looking at epoch 0.
|
|
100
|
-
const closedEpoch = epoch - BigInt(this.proofSubmissionEpochs) - 1n;
|
|
89
|
+
const closedEpoch = BigInt(epoch) - BigInt(this.proofSubmissionEpochs) - 1n;
|
|
101
90
|
const rewards = await this.rollup.getSpecificProverRewardsForEpoch(closedEpoch, this.coinbase);
|
|
102
91
|
|
|
103
92
|
const fmt = parseFloat(formatUnits(rewards, 18));
|
|
@@ -138,68 +127,28 @@ export class ProverNodePublisherMetrics {
|
|
|
138
127
|
) {
|
|
139
128
|
this.meter = client.getMeter(name);
|
|
140
129
|
|
|
141
|
-
this.gasPrice = this.meter.createHistogram(Metrics.L1_PUBLISHER_GAS_PRICE
|
|
142
|
-
description: 'The gas price used for transactions',
|
|
143
|
-
unit: 'gwei',
|
|
144
|
-
valueType: ValueType.DOUBLE,
|
|
145
|
-
});
|
|
130
|
+
this.gasPrice = this.meter.createHistogram(Metrics.L1_PUBLISHER_GAS_PRICE);
|
|
146
131
|
|
|
147
|
-
this.txCount = this.meter
|
|
148
|
-
|
|
132
|
+
this.txCount = createUpDownCounterWithDefault(this.meter, Metrics.L1_PUBLISHER_TX_COUNT, {
|
|
133
|
+
[Attributes.L1_TX_TYPE]: ['submitProof'],
|
|
134
|
+
[Attributes.OK]: [true, false],
|
|
149
135
|
});
|
|
150
136
|
|
|
151
|
-
this.txDuration = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_DURATION
|
|
152
|
-
description: 'The duration of transaction processing',
|
|
153
|
-
unit: 'ms',
|
|
154
|
-
valueType: ValueType.INT,
|
|
155
|
-
});
|
|
137
|
+
this.txDuration = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_DURATION);
|
|
156
138
|
|
|
157
|
-
this.txGas = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_GAS
|
|
158
|
-
description: 'The gas consumed by transactions',
|
|
159
|
-
unit: 'gas',
|
|
160
|
-
valueType: ValueType.INT,
|
|
161
|
-
});
|
|
139
|
+
this.txGas = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_GAS);
|
|
162
140
|
|
|
163
|
-
this.txCalldataSize = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_CALLDATA_SIZE
|
|
164
|
-
description: 'The size of the calldata in transactions',
|
|
165
|
-
unit: 'By',
|
|
166
|
-
valueType: ValueType.INT,
|
|
167
|
-
});
|
|
141
|
+
this.txCalldataSize = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_CALLDATA_SIZE);
|
|
168
142
|
|
|
169
|
-
this.txCalldataGas = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_CALLDATA_GAS
|
|
170
|
-
description: 'The gas consumed by the calldata in transactions',
|
|
171
|
-
unit: 'gas',
|
|
172
|
-
valueType: ValueType.INT,
|
|
173
|
-
});
|
|
143
|
+
this.txCalldataGas = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_CALLDATA_GAS);
|
|
174
144
|
|
|
175
|
-
this.txBlobDataGasUsed = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_BLOBDATA_GAS_USED
|
|
176
|
-
description: 'The amount of blob gas used in transactions',
|
|
177
|
-
unit: 'gas',
|
|
178
|
-
valueType: ValueType.INT,
|
|
179
|
-
});
|
|
145
|
+
this.txBlobDataGasUsed = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_BLOBDATA_GAS_USED);
|
|
180
146
|
|
|
181
|
-
this.txBlobDataGasCost = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_BLOBDATA_GAS_COST
|
|
182
|
-
description: 'The gas cost of blobs in transactions',
|
|
183
|
-
unit: 'gwei',
|
|
184
|
-
valueType: ValueType.INT,
|
|
185
|
-
});
|
|
147
|
+
this.txBlobDataGasCost = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_BLOBDATA_GAS_COST);
|
|
186
148
|
|
|
187
|
-
this.txTotalFee = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_TOTAL_FEE
|
|
188
|
-
description: 'How much L1 tx costs',
|
|
189
|
-
unit: 'gwei',
|
|
190
|
-
valueType: ValueType.DOUBLE,
|
|
191
|
-
advice: {
|
|
192
|
-
explicitBucketBoundaries: [
|
|
193
|
-
0.001, 0.002, 0.004, 0.008, 0.01, 0.02, 0.04, 0.08, 0.1, 0.2, 0.4, 0.8, 1, 1.2, 1.4, 1.8, 2,
|
|
194
|
-
],
|
|
195
|
-
},
|
|
196
|
-
});
|
|
149
|
+
this.txTotalFee = this.meter.createHistogram(Metrics.L1_PUBLISHER_TX_TOTAL_FEE);
|
|
197
150
|
|
|
198
|
-
this.senderBalance = this.meter.createGauge(Metrics.L1_PUBLISHER_BALANCE
|
|
199
|
-
unit: 'eth',
|
|
200
|
-
description: 'The balance of the sender address',
|
|
201
|
-
valueType: ValueType.DOUBLE,
|
|
202
|
-
});
|
|
151
|
+
this.senderBalance = this.meter.createGauge(Metrics.L1_PUBLISHER_BALANCE);
|
|
203
152
|
}
|
|
204
153
|
|
|
205
154
|
recordFailedTx() {
|
|
@@ -1,18 +1,13 @@
|
|
|
1
|
+
import { BlockNumber, EpochNumber } from '@aztec/foundation/branded-types';
|
|
1
2
|
import { createLogger } from '@aztec/foundation/log';
|
|
2
3
|
import { RunningPromise } from '@aztec/foundation/running-promise';
|
|
3
4
|
import { sleep } from '@aztec/foundation/sleep';
|
|
4
5
|
import type { L2BlockSource } from '@aztec/stdlib/block';
|
|
5
6
|
import { type L1RollupConstants, getEpochAtSlot } from '@aztec/stdlib/epoch-helpers';
|
|
6
|
-
import {
|
|
7
|
-
type TelemetryClient,
|
|
8
|
-
type Traceable,
|
|
9
|
-
type Tracer,
|
|
10
|
-
getTelemetryClient,
|
|
11
|
-
trackSpan,
|
|
12
|
-
} from '@aztec/telemetry-client';
|
|
7
|
+
import { type TelemetryClient, type Traceable, type Tracer, getTelemetryClient } from '@aztec/telemetry-client';
|
|
13
8
|
|
|
14
9
|
export interface EpochMonitorHandler {
|
|
15
|
-
handleEpochReadyToProve(epochNumber:
|
|
10
|
+
handleEpochReadyToProve(epochNumber: EpochNumber): Promise<boolean>;
|
|
16
11
|
}
|
|
17
12
|
|
|
18
13
|
/**
|
|
@@ -32,7 +27,7 @@ export class EpochMonitor implements Traceable {
|
|
|
32
27
|
public readonly tracer: Tracer;
|
|
33
28
|
|
|
34
29
|
private handler: EpochMonitorHandler | undefined;
|
|
35
|
-
private latestEpochNumber:
|
|
30
|
+
private latestEpochNumber: EpochNumber | undefined;
|
|
36
31
|
|
|
37
32
|
constructor(
|
|
38
33
|
private readonly l2BlockSource: L2BlockSource,
|
|
@@ -72,7 +67,6 @@ export class EpochMonitor implements Traceable {
|
|
|
72
67
|
this.log.info('Stopped EpochMonitor');
|
|
73
68
|
}
|
|
74
69
|
|
|
75
|
-
@trackSpan('EpochMonitor.work')
|
|
76
70
|
public async work() {
|
|
77
71
|
const { epochToProve, blockNumber, slotNumber } = await this.getEpochNumberToProve();
|
|
78
72
|
this.log.debug(`Epoch to prove: ${epochToProve}`, { blockNumber, slotNumber });
|
|
@@ -104,7 +98,7 @@ export class EpochMonitor implements Traceable {
|
|
|
104
98
|
|
|
105
99
|
private async getEpochNumberToProve() {
|
|
106
100
|
const lastBlockProven = await this.l2BlockSource.getProvenBlockNumber();
|
|
107
|
-
const firstBlockToProve = lastBlockProven + 1;
|
|
101
|
+
const firstBlockToProve = BlockNumber(lastBlockProven + 1);
|
|
108
102
|
const firstBlockHeaderToProve = await this.l2BlockSource.getBlockHeader(firstBlockToProve);
|
|
109
103
|
if (!firstBlockHeaderToProve) {
|
|
110
104
|
return { epochToProve: undefined, blockNumber: firstBlockToProve };
|