@aztec/prover-node 0.0.1-commit.f146247c → 0.0.1-commit.f1b29a41e

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (42) hide show
  1. package/dest/actions/rerun-epoch-proving-job.d.ts +2 -2
  2. package/dest/actions/rerun-epoch-proving-job.d.ts.map +1 -1
  3. package/dest/actions/rerun-epoch-proving-job.js +1 -3
  4. package/dest/actions/upload-epoch-proof-failure.d.ts +2 -2
  5. package/dest/actions/upload-epoch-proof-failure.d.ts.map +1 -1
  6. package/dest/bin/run-failed-epoch.js +5 -2
  7. package/dest/config.d.ts +5 -8
  8. package/dest/config.d.ts.map +1 -1
  9. package/dest/config.js +16 -19
  10. package/dest/factory.d.ts +19 -13
  11. package/dest/factory.d.ts.map +1 -1
  12. package/dest/factory.js +39 -59
  13. package/dest/job/epoch-proving-job.d.ts +1 -1
  14. package/dest/job/epoch-proving-job.d.ts.map +1 -1
  15. package/dest/job/epoch-proving-job.js +32 -11
  16. package/dest/metrics.d.ts +11 -1
  17. package/dest/metrics.d.ts.map +1 -1
  18. package/dest/metrics.js +25 -0
  19. package/dest/monitors/epoch-monitor.d.ts +1 -1
  20. package/dest/monitors/epoch-monitor.d.ts.map +1 -1
  21. package/dest/monitors/epoch-monitor.js +5 -7
  22. package/dest/prover-node-publisher.d.ts +3 -3
  23. package/dest/prover-node-publisher.d.ts.map +1 -1
  24. package/dest/prover-node-publisher.js +3 -3
  25. package/dest/prover-node.d.ts +19 -9
  26. package/dest/prover-node.d.ts.map +1 -1
  27. package/dest/prover-node.js +14 -9
  28. package/dest/prover-publisher-factory.d.ts +4 -4
  29. package/dest/prover-publisher-factory.d.ts.map +1 -1
  30. package/dest/prover-publisher-factory.js +3 -3
  31. package/package.json +23 -22
  32. package/src/actions/rerun-epoch-proving-job.ts +2 -2
  33. package/src/actions/upload-epoch-proof-failure.ts +1 -1
  34. package/src/bin/run-failed-epoch.ts +4 -1
  35. package/src/config.ts +23 -31
  36. package/src/factory.ts +67 -102
  37. package/src/job/epoch-proving-job.ts +47 -17
  38. package/src/metrics.ts +34 -0
  39. package/src/monitors/epoch-monitor.ts +3 -4
  40. package/src/prover-node-publisher.ts +4 -4
  41. package/src/prover-node.ts +17 -11
  42. package/src/prover-publisher-factory.ts +5 -5
package/src/config.ts CHANGED
@@ -1,18 +1,15 @@
1
- import { type ArchiverConfig, archiverConfigMappings } from '@aztec/archiver/config';
2
1
  import type { ACVMConfig, BBConfig } from '@aztec/bb-prover/config';
3
- import { type GenesisStateConfig, genesisStateConfigMappings } from '@aztec/ethereum/config';
4
2
  import {
5
3
  type ConfigMappingsType,
6
4
  booleanConfigHelper,
7
5
  getConfigFromMappings,
8
6
  numberConfigHelper,
7
+ pickConfigMappings,
9
8
  } from '@aztec/foundation/config';
10
- import { type DataStoreConfig, dataConfigMappings } from '@aztec/kv-store/config';
11
9
  import { type KeyStoreConfig, keyStoreConfigMappings } from '@aztec/node-keystore/config';
12
10
  import { ethPrivateKeySchema } from '@aztec/node-keystore/schemas';
13
11
  import type { KeyStore } from '@aztec/node-keystore/types';
14
12
  import { type SharedNodeConfig, sharedNodeConfigMappings } from '@aztec/node-lib/config';
15
- import { type P2PConfig, p2pConfigMappings } from '@aztec/p2p/config';
16
13
  import {
17
14
  type ProverAgentConfig,
18
15
  type ProverBrokerConfig,
@@ -21,24 +18,20 @@ import {
21
18
  } from '@aztec/prover-client/broker/config';
22
19
  import { type ProverClientUserConfig, bbConfigMappings, proverClientConfigMappings } from '@aztec/prover-client/config';
23
20
  import {
24
- type PublisherConfig,
25
- type TxSenderConfig,
26
- getPublisherConfigMappings,
27
- getTxSenderConfigMappings,
21
+ type ProverPublisherConfig,
22
+ type ProverTxSenderConfig,
23
+ proverPublisherConfigMappings,
24
+ proverTxSenderConfigMappings,
28
25
  } from '@aztec/sequencer-client/config';
29
- import { type WorldStateConfig, worldStateConfigMappings } from '@aztec/world-state/config';
30
-
31
- export type ProverNodeConfig = ArchiverConfig &
32
- ProverClientUserConfig &
33
- P2PConfig &
34
- WorldStateConfig &
35
- PublisherConfig &
36
- TxSenderConfig &
26
+ import { type DataStoreConfig, dataConfigMappings } from '@aztec/stdlib/kv-store';
27
+
28
+ export type ProverNodeConfig = ProverClientUserConfig &
29
+ ProverPublisherConfig &
30
+ ProverTxSenderConfig &
37
31
  DataStoreConfig &
38
32
  KeyStoreConfig &
39
- SharedNodeConfig &
40
33
  SpecificProverNodeConfig &
41
- GenesisStateConfig;
34
+ Pick<SharedNodeConfig, 'web3SignerUrl'>;
42
35
 
43
36
  export type SpecificProverNodeConfig = {
44
37
  proverNodeMaxPendingJobs: number;
@@ -53,7 +46,7 @@ export type SpecificProverNodeConfig = {
53
46
  txGatheringMaxParallelRequestsPerNode: number;
54
47
  };
55
48
 
56
- const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeConfig> = {
49
+ export const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeConfig> = {
57
50
  proverNodeMaxPendingJobs: {
58
51
  env: 'PROVER_NODE_MAX_PENDING_JOBS',
59
52
  description: 'The maximum number of pending jobs for the prover node',
@@ -67,7 +60,7 @@ const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeCon
67
60
  proverNodeMaxParallelBlocksPerEpoch: {
68
61
  env: 'PROVER_NODE_MAX_PARALLEL_BLOCKS_PER_EPOCH',
69
62
  description: 'The Maximum number of blocks to process in parallel while proving an epoch',
70
- ...numberConfigHelper(32),
63
+ ...numberConfigHelper(0),
71
64
  },
72
65
  proverNodeFailedEpochStore: {
73
66
  env: 'PROVER_NODE_FAILED_EPOCH_STORE',
@@ -108,15 +101,11 @@ const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeCon
108
101
  export const proverNodeConfigMappings: ConfigMappingsType<ProverNodeConfig> = {
109
102
  ...dataConfigMappings,
110
103
  ...keyStoreConfigMappings,
111
- ...archiverConfigMappings,
112
104
  ...proverClientConfigMappings,
113
- ...p2pConfigMappings,
114
- ...worldStateConfigMappings,
115
- ...getPublisherConfigMappings('PROVER'),
116
- ...getTxSenderConfigMappings('PROVER'),
105
+ ...proverPublisherConfigMappings,
106
+ ...proverTxSenderConfigMappings,
117
107
  ...specificProverNodeConfigMappings,
118
- ...genesisStateConfigMappings,
119
- ...sharedNodeConfigMappings,
108
+ ...pickConfigMappings(sharedNodeConfigMappings, ['web3SignerUrl']),
120
109
  };
121
110
 
122
111
  export function getProverNodeConfigFromEnv(): ProverNodeConfig {
@@ -143,7 +132,7 @@ function createKeyStoreFromWeb3Signer(config: ProverNodeConfig): KeyStore | unde
143
132
  }
144
133
 
145
134
  // Also, we need at least one publisher address.
146
- const publishers = config.publisherAddresses ?? [];
135
+ const publishers = config.proverPublisherAddresses ?? [];
147
136
 
148
137
  if (publishers.length === 0) {
149
138
  return undefined;
@@ -164,8 +153,8 @@ function createKeyStoreFromWeb3Signer(config: ProverNodeConfig): KeyStore | unde
164
153
 
165
154
  function createKeyStoreFromPublisherKeys(config: ProverNodeConfig): KeyStore | undefined {
166
155
  // Extract the publisher keys from the provided config.
167
- const publisherKeys = config.publisherPrivateKeys
168
- ? config.publisherPrivateKeys.map((k: { getValue: () => string }) => ethPrivateKeySchema.parse(k.getValue()))
156
+ const publisherKeys = config.proverPublisherPrivateKeys
157
+ ? config.proverPublisherPrivateKeys.map((k: { getValue: () => string }) => ethPrivateKeySchema.parse(k.getValue()))
169
158
  : [];
170
159
 
171
160
  // There must be at least 1.
@@ -194,7 +183,10 @@ function createKeyStoreFromPublisherKeys(config: ProverNodeConfig): KeyStore | u
194
183
 
195
184
  export function createKeyStoreForProver(config: ProverNodeConfig): KeyStore | undefined {
196
185
  if (config.web3SignerUrl !== undefined && config.web3SignerUrl.length > 0) {
197
- return createKeyStoreFromWeb3Signer(config);
186
+ const keyStore = createKeyStoreFromWeb3Signer(config);
187
+ if (keyStore) {
188
+ return keyStore;
189
+ }
198
190
  }
199
191
 
200
192
  return createKeyStoreFromPublisherKeys(config);
package/src/factory.ts CHANGED
@@ -1,34 +1,36 @@
1
- import { type Archiver, createArchiver } from '@aztec/archiver';
2
- import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
3
- import { createBlobClientWithFileStores } from '@aztec/blob-client/client';
4
- import { EpochCache } from '@aztec/epoch-cache';
1
+ import type { Archiver } from '@aztec/archiver';
2
+ import type { BlobClientInterface } from '@aztec/blob-client/client';
3
+ import { Blob } from '@aztec/blob-lib';
4
+ import type { EpochCacheInterface } from '@aztec/epoch-cache';
5
5
  import { createEthereumChain } from '@aztec/ethereum/chain';
6
+ import { makeL1HttpTransport } from '@aztec/ethereum/client';
6
7
  import { RollupContract } from '@aztec/ethereum/contracts';
7
8
  import { L1TxUtils } from '@aztec/ethereum/l1-tx-utils';
8
9
  import { PublisherManager } from '@aztec/ethereum/publisher-manager';
9
10
  import { pick } from '@aztec/foundation/collection';
10
11
  import { type Logger, createLogger } from '@aztec/foundation/log';
11
12
  import { DateProvider } from '@aztec/foundation/timer';
12
- import type { DataStoreConfig } from '@aztec/kv-store/config';
13
- import { type KeyStoreConfig, KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
14
- import { trySnapshotSync } from '@aztec/node-lib/actions';
15
- import {
16
- createForwarderL1TxUtilsFromEthSigner,
17
- createL1TxUtilsFromEthSignerWithStore,
18
- } from '@aztec/node-lib/factories';
19
- import { NodeRpcTxSource, createP2PClient } from '@aztec/p2p';
20
- import { type ProverClientConfig, createProverClient } from '@aztec/prover-client';
13
+ import { KeystoreManager } from '@aztec/node-keystore';
14
+ import { createForwarderL1TxUtilsFromSigners, createL1TxUtilsFromSigners } from '@aztec/node-lib/factories';
15
+ import { type ProverClientConfig, type ProverClientUserConfig, createProverClient } from '@aztec/prover-client';
21
16
  import { createAndStartProvingBroker } from '@aztec/prover-client/broker';
22
- import type { AztecNode, ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
23
- import { P2PClientType } from '@aztec/stdlib/p2p';
24
- import type { PublicDataTreeLeaf } from '@aztec/stdlib/trees';
25
- import { getPackageVersion } from '@aztec/stdlib/update-checker';
17
+ import {
18
+ type ProverPublisherConfig,
19
+ type ProverTxSenderConfig,
20
+ getPublisherConfigFromProverConfig,
21
+ } from '@aztec/sequencer-client';
22
+ import type {
23
+ ITxProvider,
24
+ ProverConfig,
25
+ ProvingJobBroker,
26
+ Service,
27
+ WorldStateSynchronizer,
28
+ } from '@aztec/stdlib/interfaces/server';
26
29
  import { L1Metrics, type TelemetryClient, getTelemetryClient } from '@aztec/telemetry-client';
27
- import { createWorldStateSynchronizer } from '@aztec/world-state';
28
30
 
29
- import { createPublicClient, fallback, http } from 'viem';
31
+ import { createPublicClient } from 'viem';
30
32
 
31
- import { type ProverNodeConfig, createKeyStoreForProver } from './config.js';
33
+ import type { SpecificProverNodeConfig } from './config.js';
32
34
  import { EpochMonitor } from './monitors/epoch-monitor.js';
33
35
  import { ProverNode } from './prover-node.js';
34
36
  import { ProverPublisherFactory } from './prover-publisher-factory.js';
@@ -36,54 +38,42 @@ import { ProverPublisherFactory } from './prover-publisher-factory.js';
36
38
  export type ProverNodeDeps = {
37
39
  telemetry?: TelemetryClient;
38
40
  log?: Logger;
39
- aztecNodeTxProvider?: Pick<AztecNode, 'getTxsByHash'>;
40
- archiver?: Archiver;
41
+ archiver: Archiver;
41
42
  publisherFactory?: ProverPublisherFactory;
42
43
  broker?: ProvingJobBroker;
43
44
  l1TxUtils?: L1TxUtils;
44
45
  dateProvider?: DateProvider;
46
+ worldStateSynchronizer: WorldStateSynchronizer;
47
+ p2pClient: { getTxProvider(): ITxProvider } & Partial<Service>;
48
+ epochCache: EpochCacheInterface;
49
+ blobClient: BlobClientInterface;
50
+ keyStoreManager?: KeystoreManager;
45
51
  };
46
52
 
47
- /** Creates a new prover node given a config. */
53
+ /** Creates a new prover node subsystem given a config and dependencies */
48
54
  export async function createProverNode(
49
- userConfig: ProverNodeConfig & DataStoreConfig & KeyStoreConfig,
50
- deps: ProverNodeDeps = {},
51
- options: {
52
- prefilledPublicData?: PublicDataTreeLeaf[];
53
- } = {},
55
+ userConfig: SpecificProverNodeConfig &
56
+ ProverConfig &
57
+ ProverClientUserConfig &
58
+ ProverPublisherConfig &
59
+ ProverTxSenderConfig,
60
+ deps: ProverNodeDeps,
54
61
  ) {
55
62
  const config = { ...userConfig };
56
63
  const telemetry = deps.telemetry ?? getTelemetryClient();
57
64
  const dateProvider = deps.dateProvider ?? new DateProvider();
58
- const blobClient = await createBlobClientWithFileStores(config, createLogger('prover-node:blob-client:client'));
59
- const log = deps.log ?? createLogger('prover-node');
60
-
61
- // Build a key store from file if given or from environment otherwise
62
- let keyStoreManager: KeystoreManager | undefined;
63
- const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
64
- if (keyStoreProvided) {
65
- const keyStores = loadKeystores(config.keyStoreDirectory!);
66
- keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
67
- } else {
68
- const keyStore = createKeyStoreForProver(config);
69
- if (keyStore) {
70
- keyStoreManager = new KeystoreManager(keyStore);
71
- }
72
- }
65
+ const log = deps.log ?? createLogger('prover');
73
66
 
74
- await keyStoreManager?.validateSigners();
67
+ const { p2pClient, archiver, keyStoreManager, worldStateSynchronizer } = deps;
75
68
 
76
69
  // Extract the prover signers from the key store and verify that we have one.
70
+ await keyStoreManager?.validateSigners();
77
71
  const proverSigners = keyStoreManager?.createProverSigners();
78
72
 
79
73
  if (proverSigners === undefined) {
80
74
  throw new Error('Failed to create prover key store configuration');
81
75
  } else if (proverSigners.signers.length === 0) {
82
76
  throw new Error('No prover signers found in the key store');
83
- } else if (!keyStoreProvided) {
84
- log.warn(
85
- 'KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS',
86
- );
87
77
  }
88
78
 
89
79
  log.info(`Creating prover with publishers ${proverSigners.signers.map(signer => signer.address.toString()).join()}`);
@@ -95,27 +85,7 @@ export async function createProverNode(
95
85
  const proverId = proverSigners.id ?? proverIdInUserConfig ?? proverSigners.signers[0].address;
96
86
 
97
87
  // Now create the prover client configuration from this.
98
- const proverClientConfig: ProverClientConfig = {
99
- ...config,
100
- proverId,
101
- };
102
-
103
- await trySnapshotSync(config, log);
104
-
105
- const epochCache = await EpochCache.create(config.l1Contracts.rollupAddress, config);
106
-
107
- const archiver =
108
- deps.archiver ??
109
- (await createArchiver(config, { blobClient, epochCache, telemetry, dateProvider }, { blockUntilSync: true }));
110
- log.verbose(`Created archiver and synced to block ${await archiver.getBlockNumber()}`);
111
-
112
- const worldStateSynchronizer = await createWorldStateSynchronizer(
113
- config,
114
- archiver,
115
- options.prefilledPublicData,
116
- telemetry,
117
- );
118
- await worldStateSynchronizer.start();
88
+ const proverClientConfig: ProverClientConfig = { ...config, proverId };
119
89
 
120
90
  const broker = deps.broker ?? (await createAndStartProvingBroker(config, telemetry));
121
91
 
@@ -126,7 +96,7 @@ export async function createProverNode(
126
96
 
127
97
  const publicClient = createPublicClient({
128
98
  chain: chain.chainInfo,
129
- transport: fallback(config.l1RpcUrls.map((url: string) => http(url, { batch: false }))),
99
+ transport: makeL1HttpTransport(config.l1RpcUrls, { timeout: config.l1HttpTimeoutMS }),
130
100
  pollingInterval: config.viemPollingIntervalMS,
131
101
  });
132
102
 
@@ -134,55 +104,45 @@ export async function createProverNode(
134
104
 
135
105
  const l1TxUtils = deps.l1TxUtils
136
106
  ? [deps.l1TxUtils]
137
- : config.publisherForwarderAddress
138
- ? await createForwarderL1TxUtilsFromEthSigner(
107
+ : config.proverPublisherForwarderAddress
108
+ ? await createForwarderL1TxUtilsFromSigners(
139
109
  publicClient,
140
110
  proverSigners.signers,
141
- config.publisherForwarderAddress,
111
+ config.proverPublisherForwarderAddress,
142
112
  { ...config, scope: 'prover' },
143
- { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
113
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
144
114
  )
145
- : await createL1TxUtilsFromEthSignerWithStore(
115
+ : await createL1TxUtilsFromSigners(
146
116
  publicClient,
147
117
  proverSigners.signers,
148
118
  { ...config, scope: 'prover' },
149
119
  { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
150
120
  );
151
121
 
122
+ // Create a funder L1TxUtils from the keystore funding account (if configured)
123
+ const fundingSigner = keyStoreManager?.createFundingSigner();
124
+ let funderL1TxUtils: L1TxUtils | undefined;
125
+ if (fundingSigner) {
126
+ const [funder] = await createL1TxUtilsFromSigners(
127
+ publicClient,
128
+ [fundingSigner],
129
+ { ...config, scope: 'prover' },
130
+ { telemetry, logger: log.createChild('l1-tx-utils:funder'), dateProvider },
131
+ );
132
+ funderL1TxUtils = funder;
133
+ }
134
+
152
135
  const publisherFactory =
153
136
  deps.publisherFactory ??
154
137
  new ProverPublisherFactory(config, {
155
138
  rollupContract,
156
- publisherManager: new PublisherManager(l1TxUtils, config, log.getBindings()),
139
+ publisherManager: new PublisherManager(l1TxUtils, getPublisherConfigFromProverConfig(config), {
140
+ bindings: log.getBindings(),
141
+ funder: funderL1TxUtils,
142
+ }),
157
143
  telemetry,
158
144
  });
159
145
 
160
- const proofVerifier = new QueuedIVCVerifier(
161
- config,
162
- config.realProofs || config.debugForceTxProofVerification
163
- ? await BBCircuitVerifier.new(config)
164
- : new TestCircuitVerifier(config.proverTestVerificationDelayMs),
165
- );
166
-
167
- const p2pClient = await createP2PClient(
168
- P2PClientType.Prover,
169
- config,
170
- archiver,
171
- proofVerifier,
172
- worldStateSynchronizer,
173
- epochCache,
174
- getPackageVersion() ?? '',
175
- dateProvider,
176
- telemetry,
177
- {
178
- txCollectionNodeSources: deps.aztecNodeTxProvider
179
- ? [new NodeRpcTxSource(deps.aztecNodeTxProvider, 'TestNode')]
180
- : [],
181
- },
182
- );
183
-
184
- await p2pClient.start();
185
-
186
146
  const proverNodeConfig = {
187
147
  ...pick(
188
148
  config,
@@ -213,6 +173,9 @@ export async function createProverNode(
213
173
  l1TxUtils.map(utils => utils.getSenderAddress()),
214
174
  );
215
175
 
176
+ // Extract the shared delayer from the first L1TxUtils instance (all instances share the same delayer)
177
+ const delayer = l1TxUtils[0]?.delayer;
178
+
216
179
  return new ProverNode(
217
180
  prover,
218
181
  publisherFactory,
@@ -226,5 +189,7 @@ export async function createProverNode(
226
189
  l1Metrics,
227
190
  proverNodeConfig,
228
191
  telemetry,
192
+ delayer,
193
+ dateProvider,
229
194
  );
230
195
  }
@@ -6,6 +6,7 @@ import { Fr } from '@aztec/foundation/curves/bn254';
6
6
  import { type Logger, type LoggerBindings, createLogger } from '@aztec/foundation/log';
7
7
  import { RunningPromise, promiseWithResolvers } from '@aztec/foundation/promise';
8
8
  import { Timer } from '@aztec/foundation/timer';
9
+ import { AVM_MAX_CONCURRENT_SIMULATIONS } from '@aztec/native';
9
10
  import { getVKTreeRoot } from '@aztec/noir-protocol-circuits-types/vk-tree';
10
11
  import { protocolContractsHash } from '@aztec/protocol-contracts';
11
12
  import { buildFinalBlobChallenges } from '@aztec/prover-client/helpers';
@@ -148,19 +149,32 @@ export class EpochProvingJob implements Traceable {
148
149
  this.runPromise = promise;
149
150
 
150
151
  try {
152
+ const blobTimer = new Timer();
151
153
  const blobFieldsPerCheckpoint = this.checkpoints.map(checkpoint => checkpoint.toBlobFields());
152
154
  const finalBlobBatchingChallenges = await buildFinalBlobChallenges(blobFieldsPerCheckpoint);
155
+ this.metrics.recordBlobProcessing(blobTimer.ms());
153
156
 
154
157
  this.prover.startNewEpoch(epochNumber, epochSizeCheckpoints, finalBlobBatchingChallenges);
158
+ const chonkTimer = new Timer();
155
159
  await this.prover.startChonkVerifierCircuits(Array.from(this.txs.values()));
160
+ this.metrics.recordChonkVerifier(chonkTimer.ms());
156
161
 
157
162
  // Everything in the epoch should have the same chainId and version.
158
163
  const { chainId, version } = this.checkpoints[0].blocks[0].header.globalVariables;
159
164
 
160
165
  const previousBlockHeaders = this.gatherPreviousBlockHeaders();
161
166
 
162
- await asyncPool(this.config.parallelBlockLimit ?? 32, this.checkpoints, async checkpoint => {
167
+ const allCheckpointsTimer = new Timer();
168
+
169
+ const parallelism = this.config.parallelBlockLimit
170
+ ? this.config.parallelBlockLimit
171
+ : AVM_MAX_CONCURRENT_SIMULATIONS > 0
172
+ ? AVM_MAX_CONCURRENT_SIMULATIONS
173
+ : this.checkpoints.length;
174
+
175
+ await asyncPool(parallelism, this.checkpoints, async checkpoint => {
163
176
  this.checkState();
177
+ const checkpointTimer = new Timer();
164
178
 
165
179
  const checkpointIndex = checkpoint.number - fromCheckpoint;
166
180
  const checkpointConstants = CheckpointConstantData.from({
@@ -193,7 +207,9 @@ export class EpochProvingJob implements Traceable {
193
207
  previousHeader,
194
208
  );
195
209
 
196
- for (const block of checkpoint.blocks) {
210
+ for (let blockIndex = 0; blockIndex < checkpoint.blocks.length; blockIndex++) {
211
+ const blockTimer = new Timer();
212
+ const block = checkpoint.blocks[blockIndex];
197
213
  const globalVariables = block.header.globalVariables;
198
214
  const txs = this.getTxs(block);
199
215
 
@@ -211,8 +227,12 @@ export class EpochProvingJob implements Traceable {
211
227
  // Start block proving
212
228
  await this.prover.startNewBlock(block.number, globalVariables.timestamp, txs.length);
213
229
 
214
- // Process public fns
215
- const db = await this.createFork(BlockNumber(block.number - 1), l1ToL2Messages);
230
+ // Process public fns. L1 to L2 messages are only inserted for the first block of a checkpoint,
231
+ // as the fork for subsequent blocks already includes them from the previous block's synced state.
232
+ const db = await this.createFork(
233
+ BlockNumber(block.number - 1),
234
+ blockIndex === 0 ? l1ToL2Messages : undefined,
235
+ );
216
236
  const config = PublicSimulatorConfig.from({
217
237
  proverId: this.prover.getProverId().toField(),
218
238
  skipFeeEnforcement: false,
@@ -234,8 +254,11 @@ export class EpochProvingJob implements Traceable {
234
254
  // Mark block as completed to pad it
235
255
  const expectedBlockHeader = block.header;
236
256
  await this.prover.setBlockCompleted(block.number, expectedBlockHeader);
257
+ this.metrics.recordBlockProcessing(blockTimer.ms());
237
258
  }
259
+ this.metrics.recordCheckpointProcessing(checkpointTimer.ms());
238
260
  });
261
+ this.metrics.recordAllCheckpointsProcessing(allCheckpointsTimer.ms());
239
262
 
240
263
  const executionTime = timer.ms();
241
264
 
@@ -295,22 +318,29 @@ export class EpochProvingJob implements Traceable {
295
318
  }
296
319
 
297
320
  /**
298
- * Create a new db fork for tx processing, inserting all L1 to L2.
321
+ * Create a new db fork for tx processing, optionally inserting L1 to L2 messages.
322
+ * L1 to L2 messages should only be inserted for the first block in a checkpoint,
323
+ * as subsequent blocks' synced state already includes them.
299
324
  * REFACTOR: The prover already spawns a db fork of its own for each block, so we may be able to do away with just one fork.
300
325
  */
301
- private async createFork(blockNumber: BlockNumber, l1ToL2Messages: Fr[]) {
326
+ private async createFork(blockNumber: BlockNumber, l1ToL2Messages: Fr[] | undefined) {
327
+ this.log.verbose(`Creating fork at ${blockNumber}`, { blockNumber });
302
328
  const db = await this.dbProvider.fork(blockNumber);
303
- const l1ToL2MessagesPadded = padArrayEnd<Fr, number>(
304
- l1ToL2Messages,
305
- Fr.ZERO,
306
- NUMBER_OF_L1_L2_MESSAGES_PER_ROLLUP,
307
- 'Too many L1 to L2 messages',
308
- );
309
- this.log.verbose(`Creating fork at ${blockNumber} with ${l1ToL2Messages.length} L1 to L2 messages`, {
310
- blockNumber,
311
- l1ToL2Messages: l1ToL2Messages.map(m => m.toString()),
312
- });
313
- await db.appendLeaves(MerkleTreeId.L1_TO_L2_MESSAGE_TREE, l1ToL2MessagesPadded);
329
+
330
+ if (l1ToL2Messages !== undefined) {
331
+ this.log.verbose(`Inserting ${l1ToL2Messages.length} L1 to L2 messages in fork`, {
332
+ blockNumber,
333
+ l1ToL2Messages: l1ToL2Messages.map(m => m.toString()),
334
+ });
335
+ const l1ToL2MessagesPadded = padArrayEnd<Fr, number>(
336
+ l1ToL2Messages,
337
+ Fr.ZERO,
338
+ NUMBER_OF_L1_L2_MESSAGES_PER_ROLLUP,
339
+ 'Too many L1 to L2 messages',
340
+ );
341
+ await db.appendLeaves(MerkleTreeId.L1_TO_L2_MESSAGE_TREE, l1ToL2MessagesPadded);
342
+ }
343
+
314
344
  return db;
315
345
  }
316
346
 
package/src/metrics.ts CHANGED
@@ -25,6 +25,12 @@ export class ProverNodeJobMetrics {
25
25
  provingJobBlocks: Gauge;
26
26
  provingJobTransactions: Gauge;
27
27
 
28
+ private blobProcessingDuration: Gauge;
29
+ private chonkVerifierDuration: Gauge;
30
+ private blockProcessingDuration: Histogram;
31
+ private checkpointProcessingDuration: Histogram;
32
+ private allCheckpointsProcessingDuration: Gauge;
33
+
28
34
  constructor(
29
35
  private meter: Meter,
30
36
  public readonly tracer: Tracer,
@@ -35,6 +41,14 @@ export class ProverNodeJobMetrics {
35
41
  this.provingJobCheckpoints = this.meter.createGauge(Metrics.PROVER_NODE_JOB_CHECKPOINTS);
36
42
  this.provingJobBlocks = this.meter.createGauge(Metrics.PROVER_NODE_JOB_BLOCKS);
37
43
  this.provingJobTransactions = this.meter.createGauge(Metrics.PROVER_NODE_JOB_TRANSACTIONS);
44
+
45
+ this.blobProcessingDuration = this.meter.createGauge(Metrics.PROVER_NODE_BLOB_PROCESSING_LAST_DURATION);
46
+ this.chonkVerifierDuration = this.meter.createGauge(Metrics.PROVER_NODE_CHONK_VERIFIER_LAST_DURATION);
47
+ this.blockProcessingDuration = this.meter.createHistogram(Metrics.PROVER_NODE_BLOCK_PROCESSING_DURATION);
48
+ this.checkpointProcessingDuration = this.meter.createHistogram(Metrics.PROVER_NODE_CHECKPOINT_PROCESSING_DURATION);
49
+ this.allCheckpointsProcessingDuration = this.meter.createGauge(
50
+ Metrics.PROVER_NODE_ALL_CHECKPOINTS_PROCESSING_LAST_DURATION,
51
+ );
38
52
  }
39
53
 
40
54
  public recordProvingJob(
@@ -50,6 +64,26 @@ export class ProverNodeJobMetrics {
50
64
  this.provingJobBlocks.record(Math.floor(numBlocks));
51
65
  this.provingJobTransactions.record(Math.floor(numTxs));
52
66
  }
67
+
68
+ public recordBlobProcessing(durationMs: number) {
69
+ this.blobProcessingDuration.record(Math.ceil(durationMs));
70
+ }
71
+
72
+ public recordChonkVerifier(durationMs: number) {
73
+ this.chonkVerifierDuration.record(Math.ceil(durationMs));
74
+ }
75
+
76
+ public recordBlockProcessing(durationMs: number) {
77
+ this.blockProcessingDuration.record(Math.ceil(durationMs));
78
+ }
79
+
80
+ public recordCheckpointProcessing(durationMs: number) {
81
+ this.checkpointProcessingDuration.record(Math.ceil(durationMs));
82
+ }
83
+
84
+ public recordAllCheckpointsProcessing(durationMs: number) {
85
+ this.allCheckpointsProcessingDuration.record(Math.ceil(durationMs));
86
+ }
53
87
  }
54
88
 
55
89
  export class ProverNodeRewardsMetrics {
@@ -69,9 +69,8 @@ export class EpochMonitor implements Traceable {
69
69
 
70
70
  public async work() {
71
71
  const { epochToProve, blockNumber, slotNumber } = await this.getEpochNumberToProve();
72
- this.log.debug(`Epoch to prove: ${epochToProve}`, { blockNumber, slotNumber });
73
72
  if (epochToProve === undefined) {
74
- this.log.trace(`Next block to prove ${blockNumber} not yet mined`, { blockNumber });
73
+ this.log.trace(`Next block to prove ${blockNumber} not yet mined`, { epochToProve, blockNumber, slotNumber });
75
74
  return;
76
75
  }
77
76
  if (this.latestEpochNumber !== undefined && epochToProve <= this.latestEpochNumber) {
@@ -86,11 +85,11 @@ export class EpochMonitor implements Traceable {
86
85
  }
87
86
 
88
87
  if (this.options.provingDelayMs) {
89
- this.log.debug(`Waiting ${this.options.provingDelayMs}ms before proving epoch ${epochToProve}`);
88
+ this.log.warn(`Waiting ${this.options.provingDelayMs}ms before proving epoch ${epochToProve}`);
90
89
  await sleep(this.options.provingDelayMs);
91
90
  }
92
91
 
93
- this.log.debug(`Epoch ${epochToProve} is ready to be proven`);
92
+ this.log.verbose(`Epoch ${epochToProve} is ready to be proven`);
94
93
  if (await this.handler?.handleEpochReadyToProve(epochToProve)) {
95
94
  this.latestEpochNumber = epochToProve;
96
95
  }
@@ -1,5 +1,5 @@
1
1
  import { BatchedBlob, getEthBlobEvaluationInputs } from '@aztec/blob-lib';
2
- import { AZTEC_MAX_EPOCH_DURATION } from '@aztec/constants';
2
+ import { MAX_CHECKPOINTS_PER_EPOCH } from '@aztec/constants';
3
3
  import type { RollupContract, ViemCommitteeAttestation } from '@aztec/ethereum/contracts';
4
4
  import type { L1TxUtils } from '@aztec/ethereum/l1-tx-utils';
5
5
  import { makeTuple } from '@aztec/foundation/array';
@@ -31,7 +31,7 @@ export type L1SubmitEpochProofArgs = {
31
31
  endTimestamp: Fr;
32
32
  outHash: Fr;
33
33
  proverId: Fr;
34
- fees: Tuple<FeeRecipient, typeof AZTEC_MAX_EPOCH_DURATION>;
34
+ fees: Tuple<FeeRecipient, typeof MAX_CHECKPOINTS_PER_EPOCH>;
35
35
  proof: Proof;
36
36
  };
37
37
 
@@ -168,7 +168,7 @@ export class ProverNodePublisher {
168
168
  // toCheckpoint can't be greater than pending
169
169
  if (toCheckpoint > pending) {
170
170
  throw new Error(
171
- `Cannot submit epoch proof for ${fromCheckpoint}-${toCheckpoint} as pending checkpoint is ${pending}`,
171
+ `Cannot submit epoch proof for ${fromCheckpoint}-${toCheckpoint} as proposed checkpoint is ${pending}`,
172
172
  );
173
173
  }
174
174
 
@@ -271,7 +271,7 @@ export class ProverNodePublisher {
271
271
  outHash: args.publicInputs.outHash.toString(),
272
272
  proverId: EthAddress.fromField(args.publicInputs.constants.proverId).toString(),
273
273
  } /*_args*/,
274
- makeTuple(AZTEC_MAX_EPOCH_DURATION * 2, i =>
274
+ makeTuple(MAX_CHECKPOINTS_PER_EPOCH * 2, i =>
275
275
  i % 2 === 0
276
276
  ? args.publicInputs.fees[i / 2].recipient.toField().toString()
277
277
  : args.publicInputs.fees[(i - 1) / 2].value.toString(),