@aztec/aztec-node 0.0.1-commit.03f7ef2 → 0.0.1-commit.04852196a

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,26 +1,15 @@
1
1
  import { Archiver, createArchiver } from '@aztec/archiver';
2
2
  import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
3
- import { type BlobClientInterface, createBlobClient } from '@aztec/blob-client/client';
4
- import {
5
- type BlobFileStoreMetadata,
6
- createReadOnlyFileStoreBlobClients,
7
- createWritableFileStoreBlobClient,
8
- } from '@aztec/blob-client/filestore';
9
- import {
10
- ARCHIVE_HEIGHT,
11
- INITIAL_L2_BLOCK_NUM,
12
- type L1_TO_L2_MSG_TREE_HEIGHT,
13
- type NOTE_HASH_TREE_HEIGHT,
14
- type NULLIFIER_TREE_HEIGHT,
15
- type PUBLIC_DATA_TREE_HEIGHT,
16
- } from '@aztec/constants';
3
+ import { type BlobClientInterface, createBlobClientWithFileStores } from '@aztec/blob-client/client';
4
+ import { Blob } from '@aztec/blob-lib';
5
+ import { ARCHIVE_HEIGHT, type L1_TO_L2_MSG_TREE_HEIGHT, type NOTE_HASH_TREE_HEIGHT } from '@aztec/constants';
17
6
  import { EpochCache, type EpochCacheInterface } from '@aztec/epoch-cache';
18
7
  import { createEthereumChain } from '@aztec/ethereum/chain';
19
8
  import { getPublicClient } from '@aztec/ethereum/client';
20
9
  import { RegistryContract, RollupContract } from '@aztec/ethereum/contracts';
21
10
  import type { L1ContractAddresses } from '@aztec/ethereum/l1-contract-addresses';
22
- import { BlockNumber, SlotNumber } from '@aztec/foundation/branded-types';
23
- import { compactArray, pick } from '@aztec/foundation/collection';
11
+ import { BlockNumber, CheckpointNumber, EpochNumber, SlotNumber } from '@aztec/foundation/branded-types';
12
+ import { compactArray, pick, unique } from '@aztec/foundation/collection';
24
13
  import { Fr } from '@aztec/foundation/curves/bn254';
25
14
  import { EthAddress } from '@aztec/foundation/eth-address';
26
15
  import { BadRequestError } from '@aztec/foundation/json-rpc';
@@ -28,22 +17,20 @@ import { type Logger, createLogger } from '@aztec/foundation/log';
28
17
  import { count } from '@aztec/foundation/string';
29
18
  import { DateProvider, Timer } from '@aztec/foundation/timer';
30
19
  import { MembershipWitness, SiblingPath } from '@aztec/foundation/trees';
31
- import { KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
20
+ import { type KeyStore, KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
32
21
  import { trySnapshotSync, uploadSnapshot } from '@aztec/node-lib/actions';
22
+ import { createForwarderL1TxUtilsFromSigners, createL1TxUtilsFromSigners } from '@aztec/node-lib/factories';
33
23
  import {
34
- createForwarderL1TxUtilsFromEthSigner,
35
- createL1TxUtilsWithBlobsFromEthSigner,
36
- } from '@aztec/node-lib/factories';
37
- import { type P2P, type P2PClientDeps, createP2PClient, getDefaultAllowedSetupFunctions } from '@aztec/p2p';
24
+ type P2P,
25
+ type P2PClientDeps,
26
+ createP2PClient,
27
+ createTxValidatorForAcceptingTxsOverRPC,
28
+ getDefaultAllowedSetupFunctions,
29
+ } from '@aztec/p2p';
38
30
  import { ProtocolContractAddress } from '@aztec/protocol-contracts';
39
- import {
40
- BlockBuilder,
41
- GlobalVariableBuilder,
42
- SequencerClient,
43
- type SequencerPublisher,
44
- createValidatorForAcceptingTxs,
45
- } from '@aztec/sequencer-client';
46
- import { CheckpointsBuilder } from '@aztec/sequencer-client';
31
+ import { type ProverNode, type ProverNodeDeps, createProverNode } from '@aztec/prover-node';
32
+ import { createKeyStoreForProver } from '@aztec/prover-node/config';
33
+ import { GlobalVariableBuilder, SequencerClient, type SequencerPublisher } from '@aztec/sequencer-client';
47
34
  import { PublicProcessorFactory } from '@aztec/simulator/server';
48
35
  import {
49
36
  AttestationsBlockWatcher,
@@ -55,13 +42,14 @@ import {
55
42
  import { CollectionLimitsConfig, PublicSimulatorConfig } from '@aztec/stdlib/avm';
56
43
  import { AztecAddress } from '@aztec/stdlib/aztec-address';
57
44
  import {
45
+ type BlockData,
46
+ BlockHash,
58
47
  type BlockParameter,
59
48
  type DataInBlock,
60
- type L2Block,
61
- L2BlockHash,
49
+ L2Block,
62
50
  type L2BlockSource,
63
- type PublishedL2Block,
64
51
  } from '@aztec/stdlib/block';
52
+ import type { PublishedCheckpoint } from '@aztec/stdlib/checkpoint';
65
53
  import type {
66
54
  ContractClassPublic,
67
55
  ContractDataSource,
@@ -88,9 +76,9 @@ import {
88
76
  type WorldStateSynchronizer,
89
77
  tryStop,
90
78
  } from '@aztec/stdlib/interfaces/server';
91
- import type { LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
79
+ import type { DebugLogStore, LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
80
+ import { InMemoryDebugLogStore, NullDebugLogStore } from '@aztec/stdlib/logs';
92
81
  import { InboxLeaf, type L1ToL2MessageSource } from '@aztec/stdlib/messaging';
93
- import { P2PClientType } from '@aztec/stdlib/p2p';
94
82
  import type { Offense, SlashPayloadRound } from '@aztec/stdlib/slashing';
95
83
  import type { NullifierLeafPreimage, PublicDataTreeLeaf, PublicDataTreeLeafPreimage } from '@aztec/stdlib/trees';
96
84
  import { MerkleTreeId, NullifierMembershipWitness, PublicDataWitness } from '@aztec/stdlib/trees';
@@ -116,6 +104,8 @@ import {
116
104
  trackSpan,
117
105
  } from '@aztec/telemetry-client';
118
106
  import {
107
+ FullNodeCheckpointsBuilder as CheckpointsBuilder,
108
+ FullNodeCheckpointsBuilder,
119
109
  NodeKeystoreAdapter,
120
110
  ValidatorClient,
121
111
  createBlockProposalHandler,
@@ -135,6 +125,7 @@ import { NodeMetrics } from './node_metrics.js';
135
125
  */
136
126
  export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
137
127
  private metrics: NodeMetrics;
128
+ private initialHeaderHashPromise: Promise<BlockHash> | undefined = undefined;
138
129
 
139
130
  // Prevent two snapshot operations to happen simultaneously
140
131
  private isUploadingSnapshot = false;
@@ -150,6 +141,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
150
141
  protected readonly l1ToL2MessageSource: L1ToL2MessageSource,
151
142
  protected readonly worldStateSynchronizer: WorldStateSynchronizer,
152
143
  protected readonly sequencer: SequencerClient | undefined,
144
+ protected readonly proverNode: ProverNode | undefined,
153
145
  protected readonly slasherClient: SlasherClientInterface | undefined,
154
146
  protected readonly validatorsSentinel: Sentinel | undefined,
155
147
  protected readonly epochPruneWatcher: EpochPruneWatcher | undefined,
@@ -161,12 +153,23 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
161
153
  private proofVerifier: ClientProtocolCircuitVerifier,
162
154
  private telemetry: TelemetryClient = getTelemetryClient(),
163
155
  private log = createLogger('node'),
156
+ private blobClient?: BlobClientInterface,
157
+ private validatorClient?: ValidatorClient,
158
+ private keyStoreManager?: KeystoreManager,
159
+ private debugLogStore: DebugLogStore = new NullDebugLogStore(),
164
160
  ) {
165
161
  this.metrics = new NodeMetrics(telemetry, 'AztecNodeService');
166
162
  this.tracer = telemetry.getTracer('AztecNodeService');
167
163
 
168
164
  this.log.info(`Aztec Node version: ${this.packageVersion}`);
169
165
  this.log.info(`Aztec Node started on chain 0x${l1ChainId.toString(16)}`, config.l1Contracts);
166
+
167
+ // A defensive check that protects us against introducing a bug in the complex `createAndSync` function. We must
168
+ // never have debugLogStore enabled when not in test mode because then we would be accumulating debug logs in
169
+ // memory which could be a DoS vector on the sequencer (since no fees are paid for debug logs).
170
+ if (debugLogStore.isEnabled && config.realProofs) {
171
+ throw new Error('debugLogStore should never be enabled when realProofs are set');
172
+ }
170
173
  }
171
174
 
172
175
  public async getWorldStateSyncStatus(): Promise<WorldStateSyncStatus> {
@@ -190,12 +193,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
190
193
  logger?: Logger;
191
194
  publisher?: SequencerPublisher;
192
195
  dateProvider?: DateProvider;
193
- blobClient?: BlobClientInterface;
194
- p2pClientDeps?: P2PClientDeps<P2PClientType.Full>;
196
+ p2pClientDeps?: P2PClientDeps;
197
+ proverNodeDeps?: Partial<ProverNodeDeps>;
195
198
  } = {},
196
199
  options: {
197
200
  prefilledPublicData?: PublicDataTreeLeaf[];
198
201
  dontStartSequencer?: boolean;
202
+ dontStartProverNode?: boolean;
199
203
  } = {},
200
204
  ): Promise<AztecNodeService> {
201
205
  const config = { ...inputConfig }; // Copy the config so we dont mutate the input object
@@ -205,16 +209,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
205
209
  const dateProvider = deps.dateProvider ?? new DateProvider();
206
210
  const ethereumChain = createEthereumChain(config.l1RpcUrls, config.l1ChainId);
207
211
 
208
- // Build a key store from file if given or from environment otherwise
212
+ // Build a key store from file if given or from environment otherwise.
213
+ // We keep the raw KeyStore available so we can merge with prover keys if enableProverNode is set.
209
214
  let keyStoreManager: KeystoreManager | undefined;
210
215
  const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
211
216
  if (keyStoreProvided) {
212
217
  const keyStores = loadKeystores(config.keyStoreDirectory!);
213
218
  keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
214
219
  } else {
215
- const keyStore = createKeyStoreForValidator(config);
216
- if (keyStore) {
217
- keyStoreManager = new KeystoreManager(keyStore);
220
+ const rawKeyStores: KeyStore[] = [];
221
+ const validatorKeyStore = createKeyStoreForValidator(config);
222
+ if (validatorKeyStore) {
223
+ rawKeyStores.push(validatorKeyStore);
224
+ }
225
+ if (config.enableProverNode) {
226
+ const proverKeyStore = createKeyStoreForProver(config);
227
+ if (proverKeyStore) {
228
+ rawKeyStores.push(proverKeyStore);
229
+ }
230
+ }
231
+ if (rawKeyStores.length > 0) {
232
+ keyStoreManager = new KeystoreManager(
233
+ rawKeyStores.length === 1 ? rawKeyStores[0] : mergeKeystores(rawKeyStores),
234
+ );
218
235
  }
219
236
  }
220
237
 
@@ -225,10 +242,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
225
242
  if (keyStoreManager === undefined) {
226
243
  throw new Error('Failed to create key store, a requirement for running a validator');
227
244
  }
228
- if (!keyStoreProvided) {
229
- log.warn(
230
- 'KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS',
231
- );
245
+ if (!keyStoreProvided && process.env.NODE_ENV !== 'test') {
246
+ log.warn("Keystore created from env: it's recommended to use a file-based key store for production");
232
247
  }
233
248
  ValidatorClient.validateKeyStoreConfiguration(keyStoreManager, log);
234
249
  }
@@ -256,10 +271,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
256
271
  config.l1Contracts = { ...config.l1Contracts, ...l1ContractsAddresses };
257
272
 
258
273
  const rollupContract = new RollupContract(publicClient, config.l1Contracts.rollupAddress.toString());
259
- const [l1GenesisTime, slotDuration, rollupVersionFromRollup] = await Promise.all([
274
+ const [l1GenesisTime, slotDuration, rollupVersionFromRollup, rollupManaLimit] = await Promise.all([
260
275
  rollupContract.getL1GenesisTime(),
261
276
  rollupContract.getSlotDuration(),
262
277
  rollupContract.getVersion(),
278
+ rollupContract.getManaLimit().then(Number),
263
279
  ] as const);
264
280
 
265
281
  config.rollupVersion ??= Number(rollupVersionFromRollup);
@@ -270,24 +286,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
270
286
  );
271
287
  }
272
288
 
273
- const blobFileStoreMetadata: BlobFileStoreMetadata = {
274
- l1ChainId: config.l1ChainId,
275
- rollupVersion: config.rollupVersion,
276
- rollupAddress: config.l1Contracts.rollupAddress.toString(),
277
- };
278
-
279
- const [fileStoreClients, fileStoreUploadClient] = await Promise.all([
280
- createReadOnlyFileStoreBlobClients(config.blobFileStoreUrls, blobFileStoreMetadata, log),
281
- createWritableFileStoreBlobClient(config.blobFileStoreUploadUrl, blobFileStoreMetadata, log),
282
- ]);
283
-
284
- const blobClient =
285
- deps.blobClient ??
286
- createBlobClient(config, {
287
- logger: createLogger('node:blob-client:client'),
288
- fileStoreClients,
289
- fileStoreUploadClient,
290
- });
289
+ const blobClient = await createBlobClientWithFileStores(config, log.createChild('blob-client'));
291
290
 
292
291
  // attempt snapshot sync if possible
293
292
  await trySnapshotSync(config, log);
@@ -311,14 +310,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
311
310
  config.realProofs || config.debugForceTxProofVerification
312
311
  ? await BBCircuitVerifier.new(config)
313
312
  : new TestCircuitVerifier(config.proverTestVerificationDelayMs);
313
+
314
+ let debugLogStore: DebugLogStore;
314
315
  if (!config.realProofs) {
315
316
  log.warn(`Aztec node is accepting fake proofs`);
317
+
318
+ debugLogStore = new InMemoryDebugLogStore();
319
+ log.info(
320
+ 'Aztec node started in test mode (realProofs set to false) hence debug logs from public functions will be collected and served',
321
+ );
322
+ } else {
323
+ debugLogStore = new NullDebugLogStore();
316
324
  }
325
+
317
326
  const proofVerifier = new QueuedIVCVerifier(config, circuitVerifier);
318
327
 
328
+ const proverOnly = config.enableProverNode && config.disableValidator;
329
+ if (proverOnly) {
330
+ log.info('Starting in prover-only mode: skipping validator, sequencer, sentinel, and slasher subsystems');
331
+ }
332
+
319
333
  // create the tx pool and the p2p client, which will need the l2 block source
320
334
  const p2pClient = await createP2PClient(
321
- P2PClientType.Full,
322
335
  config,
323
336
  archiver,
324
337
  proofVerifier,
@@ -330,56 +343,69 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
330
343
  deps.p2pClientDeps,
331
344
  );
332
345
 
333
- // We should really not be modifying the config object
334
- config.txPublicSetupAllowList = config.txPublicSetupAllowList ?? (await getDefaultAllowedSetupFunctions());
346
+ // We'll accumulate sentinel watchers here
347
+ const watchers: Watcher[] = [];
335
348
 
336
- const blockBuilder = new BlockBuilder(
337
- { ...config, l1GenesisTime, slotDuration: Number(slotDuration) },
349
+ // Create FullNodeCheckpointsBuilder for block proposal handling and tx validation.
350
+ // Override maxTxsPerCheckpoint with the validator-specific limit if set.
351
+ const validatorCheckpointsBuilder = new FullNodeCheckpointsBuilder(
352
+ {
353
+ ...config,
354
+ l1GenesisTime,
355
+ slotDuration: Number(slotDuration),
356
+ rollupManaLimit,
357
+ maxTxsPerCheckpoint: config.validateMaxTxsPerCheckpoint,
358
+ },
338
359
  worldStateSynchronizer,
339
360
  archiver,
340
361
  dateProvider,
341
362
  telemetry,
342
363
  );
343
364
 
344
- // We'll accumulate sentinel watchers here
345
- const watchers: Watcher[] = [];
365
+ let validatorClient: ValidatorClient | undefined;
346
366
 
347
- // Create validator client if required
348
- const validatorClient = createValidatorClient(config, {
349
- p2pClient,
350
- telemetry,
351
- dateProvider,
352
- epochCache,
353
- blockBuilder,
354
- blockSource: archiver,
355
- l1ToL2MessageSource: archiver,
356
- keyStoreManager,
357
- fileStoreBlobUploadClient: fileStoreUploadClient,
358
- });
367
+ if (!proverOnly) {
368
+ // Create validator client if required
369
+ validatorClient = await createValidatorClient(config, {
370
+ checkpointsBuilder: validatorCheckpointsBuilder,
371
+ worldState: worldStateSynchronizer,
372
+ p2pClient,
373
+ telemetry,
374
+ dateProvider,
375
+ epochCache,
376
+ blockSource: archiver,
377
+ l1ToL2MessageSource: archiver,
378
+ keyStoreManager,
379
+ blobClient,
380
+ });
359
381
 
360
- // If we have a validator client, register it as a source of offenses for the slasher,
361
- // and have it register callbacks on the p2p client *before* we start it, otherwise messages
362
- // like attestations or auths will fail.
363
- if (validatorClient) {
364
- watchers.push(validatorClient);
365
- if (!options.dontStartSequencer) {
366
- await validatorClient.registerHandlers();
382
+ // If we have a validator client, register it as a source of offenses for the slasher,
383
+ // and have it register callbacks on the p2p client *before* we start it, otherwise messages
384
+ // like attestations or auths will fail.
385
+ if (validatorClient) {
386
+ watchers.push(validatorClient);
387
+ if (!options.dontStartSequencer) {
388
+ await validatorClient.registerHandlers();
389
+ }
367
390
  }
368
391
  }
369
392
 
370
- // If there's no validator client but alwaysReexecuteBlockProposals is enabled,
371
- // create a BlockProposalHandler to reexecute block proposals for monitoring
372
- if (!validatorClient && config.alwaysReexecuteBlockProposals) {
373
- log.info('Setting up block proposal reexecution for monitoring');
393
+ // If there's no validator client, create a BlockProposalHandler to handle block proposals
394
+ // for monitoring or reexecution. Reexecution (default) allows us to follow the pending chain,
395
+ // while non-reexecution is used for validating the proposals and collecting their txs.
396
+ if (!validatorClient) {
397
+ const reexecute = !!config.alwaysReexecuteBlockProposals;
398
+ log.info(`Setting up block proposal handler` + (reexecute ? ' with reexecution of proposals' : ''));
374
399
  createBlockProposalHandler(config, {
375
- blockBuilder,
400
+ checkpointsBuilder: validatorCheckpointsBuilder,
401
+ worldState: worldStateSynchronizer,
376
402
  epochCache,
377
403
  blockSource: archiver,
378
404
  l1ToL2MessageSource: archiver,
379
405
  p2pClient,
380
406
  dateProvider,
381
407
  telemetry,
382
- }).registerForReexecution(p2pClient);
408
+ }).register(p2pClient, reexecute);
383
409
  }
384
410
 
385
411
  // Start world state and wait for it to sync to the archiver.
@@ -388,29 +414,33 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
388
414
  // Start p2p. Note that it depends on world state to be running.
389
415
  await p2pClient.start();
390
416
 
391
- const validatorsSentinel = await createSentinel(epochCache, archiver, p2pClient, config);
392
- if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
393
- watchers.push(validatorsSentinel);
394
- }
395
-
417
+ let validatorsSentinel: Awaited<ReturnType<typeof createSentinel>> | undefined;
396
418
  let epochPruneWatcher: EpochPruneWatcher | undefined;
397
- if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
398
- epochPruneWatcher = new EpochPruneWatcher(
399
- archiver,
400
- archiver,
401
- epochCache,
402
- p2pClient.getTxProvider(),
403
- blockBuilder,
404
- config,
405
- );
406
- watchers.push(epochPruneWatcher);
407
- }
408
-
409
- // We assume we want to slash for invalid attestations unless all max penalties are set to 0
410
419
  let attestationsBlockWatcher: AttestationsBlockWatcher | undefined;
411
- if (config.slashProposeInvalidAttestationsPenalty > 0n || config.slashAttestDescendantOfInvalidPenalty > 0n) {
412
- attestationsBlockWatcher = new AttestationsBlockWatcher(archiver, epochCache, config);
413
- watchers.push(attestationsBlockWatcher);
420
+
421
+ if (!proverOnly) {
422
+ validatorsSentinel = await createSentinel(epochCache, archiver, p2pClient, config);
423
+ if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
424
+ watchers.push(validatorsSentinel);
425
+ }
426
+
427
+ if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
428
+ epochPruneWatcher = new EpochPruneWatcher(
429
+ archiver,
430
+ archiver,
431
+ epochCache,
432
+ p2pClient.getTxProvider(),
433
+ validatorCheckpointsBuilder,
434
+ config,
435
+ );
436
+ watchers.push(epochPruneWatcher);
437
+ }
438
+
439
+ // We assume we want to slash for invalid attestations unless all max penalties are set to 0
440
+ if (config.slashProposeInvalidAttestationsPenalty > 0n || config.slashAttestDescendantOfInvalidPenalty > 0n) {
441
+ attestationsBlockWatcher = new AttestationsBlockWatcher(archiver, epochCache, config);
442
+ watchers.push(attestationsBlockWatcher);
443
+ }
414
444
  }
415
445
 
416
446
  // Start p2p-related services once the archiver has completed sync
@@ -447,27 +477,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
447
477
  );
448
478
  await slasherClient.start();
449
479
 
450
- const l1TxUtils = config.publisherForwarderAddress
451
- ? await createForwarderL1TxUtilsFromEthSigner(
480
+ const l1TxUtils = config.sequencerPublisherForwarderAddress
481
+ ? await createForwarderL1TxUtilsFromSigners(
452
482
  publicClient,
453
483
  keyStoreManager!.createAllValidatorPublisherSigners(),
454
- config.publisherForwarderAddress,
484
+ config.sequencerPublisherForwarderAddress,
455
485
  { ...config, scope: 'sequencer' },
456
- { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
486
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
457
487
  )
458
- : await createL1TxUtilsWithBlobsFromEthSigner(
488
+ : await createL1TxUtilsFromSigners(
459
489
  publicClient,
460
490
  keyStoreManager!.createAllValidatorPublisherSigners(),
461
491
  { ...config, scope: 'sequencer' },
462
- { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
492
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
463
493
  );
464
494
 
465
495
  // Create and start the sequencer client
466
496
  const checkpointsBuilder = new CheckpointsBuilder(
467
- { ...config, l1GenesisTime, slotDuration: Number(slotDuration) },
497
+ { ...config, l1GenesisTime, slotDuration: Number(slotDuration), rollupManaLimit },
498
+ worldStateSynchronizer,
468
499
  archiver,
469
500
  dateProvider,
470
501
  telemetry,
502
+ debugLogStore,
471
503
  );
472
504
 
473
505
  sequencer = await SequencerClient.new(config, {
@@ -495,6 +527,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
495
527
  log.warn(`Sequencer created but not started`);
496
528
  }
497
529
 
530
+ // Create prover node subsystem if enabled
531
+ let proverNode: ProverNode | undefined;
532
+ if (config.enableProverNode) {
533
+ proverNode = await createProverNode(config, {
534
+ ...deps.proverNodeDeps,
535
+ telemetry,
536
+ dateProvider,
537
+ archiver,
538
+ worldStateSynchronizer,
539
+ p2pClient,
540
+ epochCache,
541
+ blobClient,
542
+ keyStoreManager,
543
+ });
544
+
545
+ if (!options.dontStartProverNode) {
546
+ await proverNode.start();
547
+ log.info(`Prover node subsystem started`);
548
+ } else {
549
+ log.info(`Prover node subsystem created but not started`);
550
+ }
551
+ }
552
+
498
553
  const globalVariableBuilder = new GlobalVariableBuilder({
499
554
  ...config,
500
555
  rollupVersion: BigInt(config.rollupVersion),
@@ -502,7 +557,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
502
557
  slotDuration: Number(slotDuration),
503
558
  });
504
559
 
505
- return new AztecNodeService(
560
+ const node = new AztecNodeService(
506
561
  config,
507
562
  p2pClient,
508
563
  archiver,
@@ -511,6 +566,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
511
566
  archiver,
512
567
  worldStateSynchronizer,
513
568
  sequencer,
569
+ proverNode,
514
570
  slasherClient,
515
571
  validatorsSentinel,
516
572
  epochPruneWatcher,
@@ -522,7 +578,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
522
578
  proofVerifier,
523
579
  telemetry,
524
580
  log,
581
+ blobClient,
582
+ validatorClient,
583
+ keyStoreManager,
584
+ debugLogStore,
525
585
  );
586
+
587
+ return node;
526
588
  }
527
589
 
528
590
  /**
@@ -533,6 +595,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
533
595
  return this.sequencer;
534
596
  }
535
597
 
598
+ /** Returns the prover node subsystem, if enabled. */
599
+ public getProverNode(): ProverNode | undefined {
600
+ return this.proverNode;
601
+ }
602
+
536
603
  public getBlockSource(): L2BlockSource {
537
604
  return this.blockSource;
538
605
  }
@@ -558,7 +625,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
558
625
  }
559
626
 
560
627
  public async getAllowedPublicSetup(): Promise<AllowedElement[]> {
561
- return this.config.txPublicSetupAllowList ?? (await getDefaultAllowedSetupFunctions());
628
+ return [...(await getDefaultAllowedSetupFunctions()), ...(this.config.txPublicSetupAllowListExtend ?? [])];
562
629
  }
563
630
 
564
631
  /**
@@ -586,19 +653,26 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
586
653
  enr,
587
654
  l1ContractAddresses: contractAddresses,
588
655
  protocolContractAddresses: protocolContractAddresses,
656
+ realProofs: !!this.config.realProofs,
589
657
  };
590
658
 
591
659
  return nodeInfo;
592
660
  }
593
661
 
594
662
  /**
595
- * Get a block specified by its number.
596
- * @param number - The block number being requested.
663
+ * Get a block specified by its block number, block hash, or 'latest'.
664
+ * @param block - The block parameter (block number, block hash, or 'latest').
597
665
  * @returns The requested block.
598
666
  */
599
- public async getBlock(number: BlockParameter): Promise<L2Block | undefined> {
600
- const blockNumber = number === 'latest' ? await this.getBlockNumber() : (number as BlockNumber);
601
- return await this.blockSource.getBlock(blockNumber);
667
+ public async getBlock(block: BlockParameter): Promise<L2Block | undefined> {
668
+ if (BlockHash.isBlockHash(block)) {
669
+ return this.getBlockByHash(block);
670
+ }
671
+ const blockNumber = block === 'latest' ? await this.getBlockNumber() : (block as BlockNumber);
672
+ if (blockNumber === BlockNumber.ZERO) {
673
+ return this.buildInitialBlock();
674
+ }
675
+ return await this.blockSource.getL2Block(blockNumber);
602
676
  }
603
677
 
604
678
  /**
@@ -606,9 +680,17 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
606
680
  * @param blockHash - The block hash being requested.
607
681
  * @returns The requested block.
608
682
  */
609
- public async getBlockByHash(blockHash: Fr): Promise<L2Block | undefined> {
610
- const publishedBlock = await this.blockSource.getPublishedBlockByHash(blockHash);
611
- return publishedBlock?.block;
683
+ public async getBlockByHash(blockHash: BlockHash): Promise<L2Block | undefined> {
684
+ const initialBlockHash = await this.#getInitialHeaderHash();
685
+ if (blockHash.equals(initialBlockHash)) {
686
+ return this.buildInitialBlock();
687
+ }
688
+ return await this.blockSource.getL2BlockByHash(blockHash);
689
+ }
690
+
691
+ private buildInitialBlock(): L2Block {
692
+ const initialHeader = this.worldStateSynchronizer.getCommitted().getInitialHeader();
693
+ return L2Block.empty(initialHeader);
612
694
  }
613
695
 
614
696
  /**
@@ -617,8 +699,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
617
699
  * @returns The requested block.
618
700
  */
619
701
  public async getBlockByArchive(archive: Fr): Promise<L2Block | undefined> {
620
- const publishedBlock = await this.blockSource.getPublishedBlockByArchive(archive);
621
- return publishedBlock?.block;
702
+ return await this.blockSource.getL2BlockByArchive(archive);
622
703
  }
623
704
 
624
705
  /**
@@ -628,19 +709,23 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
628
709
  * @returns The blocks requested.
629
710
  */
630
711
  public async getBlocks(from: BlockNumber, limit: number): Promise<L2Block[]> {
631
- return (await this.blockSource.getBlocks(from, limit)) ?? [];
712
+ return (await this.blockSource.getBlocks(from, BlockNumber(limit))) ?? [];
713
+ }
714
+
715
+ public async getCheckpoints(from: CheckpointNumber, limit: number): Promise<PublishedCheckpoint[]> {
716
+ return (await this.blockSource.getCheckpoints(from, limit)) ?? [];
632
717
  }
633
718
 
634
- public async getPublishedBlocks(from: BlockNumber, limit: number): Promise<PublishedL2Block[]> {
635
- return (await this.blockSource.getPublishedBlocks(from, limit)) ?? [];
719
+ public async getCheckpointedBlocks(from: BlockNumber, limit: number) {
720
+ return (await this.blockSource.getCheckpointedBlocks(from, limit)) ?? [];
636
721
  }
637
722
 
638
723
  /**
639
- * Method to fetch the current base fees.
640
- * @returns The current base fees.
724
+ * Method to fetch the current min L2 fees.
725
+ * @returns The current min L2 fees.
641
726
  */
642
- public async getCurrentBaseFees(): Promise<GasFees> {
643
- return await this.globalVariableBuilder.getCurrentBaseFees();
727
+ public async getCurrentMinFees(): Promise<GasFees> {
728
+ return await this.globalVariableBuilder.getCurrentMinFees();
644
729
  }
645
730
 
646
731
  public async getMaxPriorityFees(): Promise<GasFees> {
@@ -663,6 +748,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
663
748
  return await this.blockSource.getProvenBlockNumber();
664
749
  }
665
750
 
751
+ public async getCheckpointedBlockNumber(): Promise<BlockNumber> {
752
+ return await this.blockSource.getCheckpointedL2BlockNumber();
753
+ }
754
+
755
+ public getCheckpointNumber(): Promise<CheckpointNumber> {
756
+ return this.blockSource.getCheckpointNumber();
757
+ }
758
+
666
759
  /**
667
760
  * Method to fetch the version of the package.
668
761
  * @returns The node package version
@@ -695,12 +788,43 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
695
788
  return this.contractDataSource.getContract(address);
696
789
  }
697
790
 
698
- public getPrivateLogsByTags(tags: SiloedTag[]): Promise<TxScopedL2Log[][]> {
699
- return this.logsSource.getPrivateLogsByTags(tags);
700
- }
701
-
702
- public getPublicLogsByTagsFromContract(contractAddress: AztecAddress, tags: Tag[]): Promise<TxScopedL2Log[][]> {
703
- return this.logsSource.getPublicLogsByTagsFromContract(contractAddress, tags);
791
+ public async getPrivateLogsByTags(
792
+ tags: SiloedTag[],
793
+ page?: number,
794
+ referenceBlock?: BlockHash,
795
+ ): Promise<TxScopedL2Log[][]> {
796
+ if (referenceBlock) {
797
+ const initialBlockHash = await this.#getInitialHeaderHash();
798
+ if (!referenceBlock.equals(initialBlockHash)) {
799
+ const header = await this.blockSource.getBlockHeaderByHash(referenceBlock);
800
+ if (!header) {
801
+ throw new Error(
802
+ `Block ${referenceBlock.toString()} not found in the node. This might indicate a reorg has occurred.`,
803
+ );
804
+ }
805
+ }
806
+ }
807
+ return this.logsSource.getPrivateLogsByTags(tags, page);
808
+ }
809
+
810
+ public async getPublicLogsByTagsFromContract(
811
+ contractAddress: AztecAddress,
812
+ tags: Tag[],
813
+ page?: number,
814
+ referenceBlock?: BlockHash,
815
+ ): Promise<TxScopedL2Log[][]> {
816
+ if (referenceBlock) {
817
+ const initialBlockHash = await this.#getInitialHeaderHash();
818
+ if (!referenceBlock.equals(initialBlockHash)) {
819
+ const header = await this.blockSource.getBlockHeaderByHash(referenceBlock);
820
+ if (!header) {
821
+ throw new Error(
822
+ `Block ${referenceBlock.toString()} not found in the node. This might indicate a reorg has occurred.`,
823
+ );
824
+ }
825
+ }
826
+ }
827
+ return this.logsSource.getPublicLogsByTagsFromContract(contractAddress, tags, page);
704
828
  }
705
829
 
706
830
  /**
@@ -742,26 +866,36 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
742
866
  }
743
867
 
744
868
  await this.p2pClient!.sendTx(tx);
745
- this.metrics.receivedTx(timer.ms(), true);
746
- this.log.info(`Received tx ${txHash}`, { txHash });
869
+ const duration = timer.ms();
870
+ this.metrics.receivedTx(duration, true);
871
+ this.log.info(`Received tx ${txHash} in ${duration}ms`, { txHash });
747
872
  }
748
873
 
749
874
  public async getTxReceipt(txHash: TxHash): Promise<TxReceipt> {
750
- let txReceipt = new TxReceipt(txHash, TxStatus.DROPPED, 'Tx dropped by P2P node.');
751
-
752
- // We first check if the tx is in pending (instead of first checking if it is mined) because if we first check
753
- // for mined and then for pending there could be a race condition where the tx is mined between the two checks
754
- // and we would incorrectly return a TxReceipt with status DROPPED
755
- if ((await this.p2pClient.getTxStatus(txHash)) === 'pending') {
756
- txReceipt = new TxReceipt(txHash, TxStatus.PENDING, '');
757
- }
875
+ // Check the tx pool status first. If the tx is known to the pool (pending or mined), we'll use that
876
+ // as a fallback if we don't find a settled receipt in the archiver.
877
+ const txPoolStatus = await this.p2pClient.getTxStatus(txHash);
878
+ const isKnownToPool = txPoolStatus === 'pending' || txPoolStatus === 'mined';
758
879
 
880
+ // Then get the actual tx from the archiver, which tracks every tx in a mined block.
759
881
  const settledTxReceipt = await this.blockSource.getSettledTxReceipt(txHash);
882
+
883
+ let receipt: TxReceipt;
760
884
  if (settledTxReceipt) {
761
- txReceipt = settledTxReceipt;
885
+ receipt = settledTxReceipt;
886
+ } else if (isKnownToPool) {
887
+ // If the tx is in the pool but not in the archiver, it's pending.
888
+ // This handles race conditions between archiver and p2p, where the archiver
889
+ // has pruned the block in which a tx was mined, but p2p has not caught up yet.
890
+ receipt = new TxReceipt(txHash, TxStatus.PENDING, undefined, undefined);
891
+ } else {
892
+ // Otherwise, if we don't know the tx, we consider it dropped.
893
+ receipt = new TxReceipt(txHash, TxStatus.DROPPED, undefined, 'Tx dropped by P2P node');
762
894
  }
763
895
 
764
- return txReceipt;
896
+ this.debugLogStore.decorateReceiptWithLogs(txHash.toString(), receipt);
897
+
898
+ return receipt;
765
899
  }
766
900
 
767
901
  public getTxEffect(txHash: TxHash): Promise<IndexedTxEffect | undefined> {
@@ -778,13 +912,23 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
778
912
  await tryStop(this.slasherClient);
779
913
  await tryStop(this.proofVerifier);
780
914
  await tryStop(this.sequencer);
915
+ await tryStop(this.proverNode);
781
916
  await tryStop(this.p2pClient);
782
917
  await tryStop(this.worldStateSynchronizer);
783
918
  await tryStop(this.blockSource);
919
+ await tryStop(this.blobClient);
784
920
  await tryStop(this.telemetry);
785
921
  this.log.info(`Stopped Aztec Node`);
786
922
  }
787
923
 
924
+ /**
925
+ * Returns the blob client used by this node.
926
+ * @internal - Exposed for testing purposes only.
927
+ */
928
+ public getBlobClient(): BlobClientInterface | undefined {
929
+ return this.blobClient;
930
+ }
931
+
788
932
  /**
789
933
  * Method to retrieve pending txs.
790
934
  * @param limit - The number of items to returns
@@ -817,20 +961,12 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
817
961
  return compactArray(await Promise.all(txHashes.map(txHash => this.getTxByHash(txHash))));
818
962
  }
819
963
 
820
- /**
821
- * Find the indexes of the given leaves in the given tree along with a block metadata pointing to the block in which
822
- * the leaves were inserted.
823
- * @param blockNumber - The block number at which to get the data or 'latest' for latest data.
824
- * @param treeId - The tree to search in.
825
- * @param leafValues - The values to search for.
826
- * @returns The indices of leaves and the block metadata of a block in which the leaves were inserted.
827
- */
828
964
  public async findLeavesIndexes(
829
- blockNumber: BlockParameter,
965
+ referenceBlock: BlockParameter,
830
966
  treeId: MerkleTreeId,
831
967
  leafValues: Fr[],
832
968
  ): Promise<(DataInBlock<bigint> | undefined)[]> {
833
- const committedDb = await this.#getWorldState(blockNumber);
969
+ const committedDb = await this.#getWorldState(referenceBlock);
834
970
  const maybeIndices = await committedDb.findLeafIndices(
835
971
  treeId,
836
972
  leafValues.map(x => x.toBuffer()),
@@ -882,56 +1018,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
882
1018
  }
883
1019
  return {
884
1020
  l2BlockNumber: BlockNumber(Number(blockNumber)),
885
- l2BlockHash: L2BlockHash.fromField(blockHash),
1021
+ l2BlockHash: new BlockHash(blockHash),
886
1022
  data: index,
887
1023
  };
888
1024
  });
889
1025
  }
890
1026
 
891
- /**
892
- * Returns a sibling path for the given index in the nullifier tree.
893
- * @param blockNumber - The block number at which to get the data.
894
- * @param leafIndex - The index of the leaf for which the sibling path is required.
895
- * @returns The sibling path for the leaf index.
896
- */
897
- public async getNullifierSiblingPath(
898
- blockNumber: BlockParameter,
899
- leafIndex: bigint,
900
- ): Promise<SiblingPath<typeof NULLIFIER_TREE_HEIGHT>> {
901
- const committedDb = await this.#getWorldState(blockNumber);
902
- return committedDb.getSiblingPath(MerkleTreeId.NULLIFIER_TREE, leafIndex);
903
- }
904
-
905
- /**
906
- * Returns a sibling path for the given index in the data tree.
907
- * @param blockNumber - The block number at which to get the data.
908
- * @param leafIndex - The index of the leaf for which the sibling path is required.
909
- * @returns The sibling path for the leaf index.
910
- */
911
- public async getNoteHashSiblingPath(
912
- blockNumber: BlockParameter,
913
- leafIndex: bigint,
914
- ): Promise<SiblingPath<typeof NOTE_HASH_TREE_HEIGHT>> {
915
- const committedDb = await this.#getWorldState(blockNumber);
916
- return committedDb.getSiblingPath(MerkleTreeId.NOTE_HASH_TREE, leafIndex);
917
- }
918
-
919
- public async getArchiveMembershipWitness(
920
- blockNumber: BlockParameter,
921
- archive: Fr,
1027
+ public async getBlockHashMembershipWitness(
1028
+ referenceBlock: BlockParameter,
1029
+ blockHash: BlockHash,
922
1030
  ): Promise<MembershipWitness<typeof ARCHIVE_HEIGHT> | undefined> {
923
- const committedDb = await this.#getWorldState(blockNumber);
924
- const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.ARCHIVE>(MerkleTreeId.ARCHIVE, [archive]);
1031
+ const committedDb = await this.#getWorldState(referenceBlock);
1032
+ const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.ARCHIVE>(MerkleTreeId.ARCHIVE, [blockHash]);
925
1033
  return pathAndIndex === undefined
926
1034
  ? undefined
927
1035
  : MembershipWitness.fromSiblingPath(pathAndIndex.index, pathAndIndex.path);
928
1036
  }
929
1037
 
930
1038
  public async getNoteHashMembershipWitness(
931
- blockNumber: BlockParameter,
1039
+ referenceBlock: BlockParameter,
932
1040
  noteHash: Fr,
933
1041
  ): Promise<MembershipWitness<typeof NOTE_HASH_TREE_HEIGHT> | undefined> {
934
- const committedDb = await this.#getWorldState(blockNumber);
1042
+ const committedDb = await this.#getWorldState(referenceBlock);
935
1043
  const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.NOTE_HASH_TREE>(
936
1044
  MerkleTreeId.NOTE_HASH_TREE,
937
1045
  [noteHash],
@@ -941,17 +1049,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
941
1049
  : MembershipWitness.fromSiblingPath(pathAndIndex.index, pathAndIndex.path);
942
1050
  }
943
1051
 
944
- /**
945
- * Returns the index and a sibling path for a leaf in the committed l1 to l2 data tree.
946
- * @param blockNumber - The block number at which to get the data.
947
- * @param l1ToL2Message - The l1ToL2Message to get the index / sibling path for.
948
- * @returns A tuple of the index and the sibling path of the L1ToL2Message (undefined if not found).
949
- */
950
1052
  public async getL1ToL2MessageMembershipWitness(
951
- blockNumber: BlockParameter,
1053
+ referenceBlock: BlockParameter,
952
1054
  l1ToL2Message: Fr,
953
1055
  ): Promise<[bigint, SiblingPath<typeof L1_TO_L2_MSG_TREE_HEIGHT>] | undefined> {
954
- const db = await this.#getWorldState(blockNumber);
1056
+ const db = await this.#getWorldState(referenceBlock);
955
1057
  const [witness] = await db.findSiblingPaths(MerkleTreeId.L1_TO_L2_MESSAGE_TREE, [l1ToL2Message]);
956
1058
  if (!witness) {
957
1059
  return undefined;
@@ -961,11 +1063,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
961
1063
  return [witness.index, witness.path];
962
1064
  }
963
1065
 
964
- public async getL1ToL2MessageBlock(l1ToL2Message: Fr): Promise<BlockNumber | undefined> {
1066
+ public async getL1ToL2MessageCheckpoint(l1ToL2Message: Fr): Promise<CheckpointNumber | undefined> {
965
1067
  const messageIndex = await this.l1ToL2MessageSource.getL1ToL2MessageIndex(l1ToL2Message);
966
- return messageIndex
967
- ? BlockNumber.fromCheckpointNumber(InboxLeaf.checkpointNumberFromIndex(messageIndex))
968
- : undefined;
1068
+ return messageIndex ? InboxLeaf.checkpointNumberFromIndex(messageIndex) : undefined;
969
1069
  }
970
1070
 
971
1071
  /**
@@ -979,56 +1079,36 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
979
1079
  }
980
1080
 
981
1081
  /**
982
- * Returns all the L2 to L1 messages in a block.
983
- * @param blockNumber - The block number at which to get the data.
984
- * @returns The L2 to L1 messages (undefined if the block number is not found).
1082
+ * Returns all the L2 to L1 messages in an epoch.
1083
+ * @param epoch - The epoch at which to get the data.
1084
+ * @returns The L2 to L1 messages (empty array if the epoch is not found).
985
1085
  */
986
- public async getL2ToL1Messages(blockNumber: BlockParameter): Promise<Fr[][] | undefined> {
987
- const block = await this.blockSource.getBlock(
988
- blockNumber === 'latest' ? await this.getBlockNumber() : (blockNumber as BlockNumber),
1086
+ public async getL2ToL1Messages(epoch: EpochNumber): Promise<Fr[][][][]> {
1087
+ // Assumes `getCheckpointedBlocksForEpoch` returns blocks in ascending order of block number.
1088
+ const checkpointedBlocks = await this.blockSource.getCheckpointedBlocksForEpoch(epoch);
1089
+ const blocksInCheckpoints: L2Block[][] = [];
1090
+ let previousSlotNumber = SlotNumber.ZERO;
1091
+ let checkpointIndex = -1;
1092
+ for (const checkpointedBlock of checkpointedBlocks) {
1093
+ const block = checkpointedBlock.block;
1094
+ const slotNumber = block.header.globalVariables.slotNumber;
1095
+ if (slotNumber !== previousSlotNumber) {
1096
+ checkpointIndex++;
1097
+ blocksInCheckpoints.push([]);
1098
+ previousSlotNumber = slotNumber;
1099
+ }
1100
+ blocksInCheckpoints[checkpointIndex].push(block);
1101
+ }
1102
+ return blocksInCheckpoints.map(blocks =>
1103
+ blocks.map(block => block.body.txEffects.map(txEffect => txEffect.l2ToL1Msgs)),
989
1104
  );
990
- return block?.body.txEffects.map(txEffect => txEffect.l2ToL1Msgs);
991
- }
992
-
993
- /**
994
- * Returns a sibling path for a leaf in the committed blocks tree.
995
- * @param blockNumber - The block number at which to get the data.
996
- * @param leafIndex - Index of the leaf in the tree.
997
- * @returns The sibling path.
998
- */
999
- public async getArchiveSiblingPath(
1000
- blockNumber: BlockParameter,
1001
- leafIndex: bigint,
1002
- ): Promise<SiblingPath<typeof ARCHIVE_HEIGHT>> {
1003
- const committedDb = await this.#getWorldState(blockNumber);
1004
- return committedDb.getSiblingPath(MerkleTreeId.ARCHIVE, leafIndex);
1005
- }
1006
-
1007
- /**
1008
- * Returns a sibling path for a leaf in the committed public data tree.
1009
- * @param blockNumber - The block number at which to get the data.
1010
- * @param leafIndex - Index of the leaf in the tree.
1011
- * @returns The sibling path.
1012
- */
1013
- public async getPublicDataSiblingPath(
1014
- blockNumber: BlockParameter,
1015
- leafIndex: bigint,
1016
- ): Promise<SiblingPath<typeof PUBLIC_DATA_TREE_HEIGHT>> {
1017
- const committedDb = await this.#getWorldState(blockNumber);
1018
- return committedDb.getSiblingPath(MerkleTreeId.PUBLIC_DATA_TREE, leafIndex);
1019
1105
  }
1020
1106
 
1021
- /**
1022
- * Returns a nullifier membership witness for a given nullifier at a given block.
1023
- * @param blockNumber - The block number at which to get the index.
1024
- * @param nullifier - Nullifier we try to find witness for.
1025
- * @returns The nullifier membership witness (if found).
1026
- */
1027
1107
  public async getNullifierMembershipWitness(
1028
- blockNumber: BlockParameter,
1108
+ referenceBlock: BlockParameter,
1029
1109
  nullifier: Fr,
1030
1110
  ): Promise<NullifierMembershipWitness | undefined> {
1031
- const db = await this.#getWorldState(blockNumber);
1111
+ const db = await this.#getWorldState(referenceBlock);
1032
1112
  const [witness] = await db.findSiblingPaths(MerkleTreeId.NULLIFIER_TREE, [nullifier.toBuffer()]);
1033
1113
  if (!witness) {
1034
1114
  return undefined;
@@ -1045,7 +1125,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1045
1125
 
1046
1126
  /**
1047
1127
  * Returns a low nullifier membership witness for a given nullifier at a given block.
1048
- * @param blockNumber - The block number at which to get the index.
1128
+ * @param referenceBlock - The block parameter (block number, block hash, or 'latest') at which to get the data
1129
+ * (which contains the root of the nullifier tree in which we are searching for the nullifier).
1049
1130
  * @param nullifier - Nullifier we try to find the low nullifier witness for.
1050
1131
  * @returns The low nullifier membership witness (if found).
1051
1132
  * @remarks Low nullifier witness can be used to perform a nullifier non-inclusion proof by leveraging the "linked
@@ -1058,10 +1139,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1058
1139
  * TODO: This is a confusing behavior and we should eventually address that.
1059
1140
  */
1060
1141
  public async getLowNullifierMembershipWitness(
1061
- blockNumber: BlockParameter,
1142
+ referenceBlock: BlockParameter,
1062
1143
  nullifier: Fr,
1063
1144
  ): Promise<NullifierMembershipWitness | undefined> {
1064
- const committedDb = await this.#getWorldState(blockNumber);
1145
+ const committedDb = await this.#getWorldState(referenceBlock);
1065
1146
  const findResult = await committedDb.getPreviousValueIndex(MerkleTreeId.NULLIFIER_TREE, nullifier.toBigInt());
1066
1147
  if (!findResult) {
1067
1148
  return undefined;
@@ -1076,8 +1157,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1076
1157
  return new NullifierMembershipWitness(BigInt(index), preimageData as NullifierLeafPreimage, siblingPath);
1077
1158
  }
1078
1159
 
1079
- async getPublicDataWitness(blockNumber: BlockParameter, leafSlot: Fr): Promise<PublicDataWitness | undefined> {
1080
- const committedDb = await this.#getWorldState(blockNumber);
1160
+ async getPublicDataWitness(referenceBlock: BlockParameter, leafSlot: Fr): Promise<PublicDataWitness | undefined> {
1161
+ const committedDb = await this.#getWorldState(referenceBlock);
1081
1162
  const lowLeafResult = await committedDb.getPreviousValueIndex(MerkleTreeId.PUBLIC_DATA_TREE, leafSlot.toBigInt());
1082
1163
  if (!lowLeafResult) {
1083
1164
  return undefined;
@@ -1091,19 +1172,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1091
1172
  }
1092
1173
  }
1093
1174
 
1094
- /**
1095
- * Gets the storage value at the given contract storage slot.
1096
- *
1097
- * @remarks The storage slot here refers to the slot as it is defined in Noir not the index in the merkle tree.
1098
- * Aztec's version of `eth_getStorageAt`.
1099
- *
1100
- * @param contract - Address of the contract to query.
1101
- * @param slot - Slot to query.
1102
- * @param blockNumber - The block number at which to get the data or 'latest'.
1103
- * @returns Storage value at the given contract slot.
1104
- */
1105
- public async getPublicStorageAt(blockNumber: BlockParameter, contract: AztecAddress, slot: Fr): Promise<Fr> {
1106
- const committedDb = await this.#getWorldState(blockNumber);
1175
+ public async getPublicStorageAt(referenceBlock: BlockParameter, contract: AztecAddress, slot: Fr): Promise<Fr> {
1176
+ const committedDb = await this.#getWorldState(referenceBlock);
1107
1177
  const leafSlot = await computePublicDataTreeLeafSlot(contract, slot);
1108
1178
 
1109
1179
  const lowLeafResult = await committedDb.getPreviousValueIndex(MerkleTreeId.PUBLIC_DATA_TREE, leafSlot.toBigInt());
@@ -1117,24 +1187,22 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1117
1187
  return preimage.leaf.value;
1118
1188
  }
1119
1189
 
1120
- /**
1121
- * Returns the currently committed block header, or the initial header if no blocks have been produced.
1122
- * @returns The current committed block header.
1123
- */
1124
- public async getBlockHeader(blockNumber: BlockParameter = 'latest'): Promise<BlockHeader | undefined> {
1125
- return blockNumber === BlockNumber.ZERO ||
1126
- (blockNumber === 'latest' && (await this.blockSource.getBlockNumber()) === BlockNumber.ZERO)
1127
- ? this.worldStateSynchronizer.getCommitted().getInitialHeader()
1128
- : this.blockSource.getBlockHeader(blockNumber === 'latest' ? blockNumber : (blockNumber as BlockNumber));
1129
- }
1130
-
1131
- /**
1132
- * Get a block header specified by its hash.
1133
- * @param blockHash - The block hash being requested.
1134
- * @returns The requested block header.
1135
- */
1136
- public async getBlockHeaderByHash(blockHash: Fr): Promise<BlockHeader | undefined> {
1137
- return await this.blockSource.getBlockHeaderByHash(blockHash);
1190
+ public async getBlockHeader(block: BlockParameter = 'latest'): Promise<BlockHeader | undefined> {
1191
+ if (BlockHash.isBlockHash(block)) {
1192
+ const initialBlockHash = await this.#getInitialHeaderHash();
1193
+ if (block.equals(initialBlockHash)) {
1194
+ // Block source doesn't handle initial header so we need to handle the case separately.
1195
+ return this.worldStateSynchronizer.getCommitted().getInitialHeader();
1196
+ }
1197
+ return this.blockSource.getBlockHeaderByHash(block);
1198
+ } else {
1199
+ // Block source doesn't handle initial header so we need to handle the case separately.
1200
+ const blockNumber = block === 'latest' ? await this.getBlockNumber() : (block as BlockNumber);
1201
+ if (blockNumber === BlockNumber.ZERO) {
1202
+ return this.worldStateSynchronizer.getCommitted().getInitialHeader();
1203
+ }
1204
+ return this.blockSource.getBlockHeader(block);
1205
+ }
1138
1206
  }
1139
1207
 
1140
1208
  /**
@@ -1146,6 +1214,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1146
1214
  return await this.blockSource.getBlockHeaderByArchive(archive);
1147
1215
  }
1148
1216
 
1217
+ public getBlockData(number: BlockNumber): Promise<BlockData | undefined> {
1218
+ return this.blockSource.getBlockData(number);
1219
+ }
1220
+
1221
+ public getBlockDataByArchive(archive: Fr): Promise<BlockData | undefined> {
1222
+ return this.blockSource.getBlockDataByArchive(archive);
1223
+ }
1224
+
1149
1225
  /**
1150
1226
  * Simulates the public part of a transaction with the current state.
1151
1227
  * @param tx - The transaction to simulate.
@@ -1169,7 +1245,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1169
1245
  }
1170
1246
 
1171
1247
  const txHash = tx.getTxHash();
1172
- const blockNumber = BlockNumber((await this.blockSource.getBlockNumber()) + 1);
1248
+ const latestBlockNumber = await this.blockSource.getBlockNumber();
1249
+ const blockNumber = BlockNumber.add(latestBlockNumber, 1);
1173
1250
 
1174
1251
  // If sequencer is not initialized, we just set these values to zero for simulation.
1175
1252
  const coinbase = EthAddress.ZERO;
@@ -1184,6 +1261,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1184
1261
  this.contractDataSource,
1185
1262
  new DateProvider(),
1186
1263
  this.telemetry,
1264
+ this.log.getBindings(),
1187
1265
  );
1188
1266
 
1189
1267
  this.log.verbose(`Simulating public calls for tx ${txHash}`, {
@@ -1192,6 +1270,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1192
1270
  blockNumber,
1193
1271
  });
1194
1272
 
1273
+ // Ensure world-state has caught up with the latest block we loaded from the archiver
1274
+ await this.worldStateSynchronizer.syncImmediate(latestBlockNumber);
1195
1275
  const merkleTreeFork = await this.worldStateSynchronizer.fork();
1196
1276
  try {
1197
1277
  const config = PublicSimulatorConfig.from({
@@ -1207,7 +1287,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1207
1287
  const processor = publicProcessorFactory.create(merkleTreeFork, newGlobalVariables, config);
1208
1288
 
1209
1289
  // REFACTOR: Consider merging ProcessReturnValues into ProcessedTx
1210
- const [processedTxs, failedTxs, _usedTxs, returns] = await processor.process([tx]);
1290
+ const [processedTxs, failedTxs, _usedTxs, returns, debugLogs] = await processor.process([tx]);
1211
1291
  // REFACTOR: Consider returning the error rather than throwing
1212
1292
  if (failedTxs.length) {
1213
1293
  this.log.warn(`Simulated tx ${txHash} fails: ${failedTxs[0].error}`, { txHash });
@@ -1221,6 +1301,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1221
1301
  processedTx.txEffect,
1222
1302
  returns,
1223
1303
  processedTx.gasUsed,
1304
+ debugLogs,
1224
1305
  );
1225
1306
  } finally {
1226
1307
  await merkleTreeFork.close();
@@ -1234,19 +1315,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1234
1315
  const db = this.worldStateSynchronizer.getCommitted();
1235
1316
  const verifier = isSimulation ? undefined : this.proofVerifier;
1236
1317
 
1237
- // We accept transactions if they are not expired by the next slot (checked based on the IncludeByTimestamp field)
1318
+ // We accept transactions if they are not expired by the next slot (checked based on the ExpirationTimestamp field)
1238
1319
  const { ts: nextSlotTimestamp } = this.epochCache.getEpochAndSlotInNextL1Slot();
1239
1320
  const blockNumber = BlockNumber((await this.blockSource.getBlockNumber()) + 1);
1240
- const validator = createValidatorForAcceptingTxs(db, this.contractDataSource, verifier, {
1241
- timestamp: nextSlotTimestamp,
1242
- blockNumber,
1243
- l1ChainId: this.l1ChainId,
1244
- rollupVersion: this.version,
1245
- setupAllowList: this.config.txPublicSetupAllowList ?? (await getDefaultAllowedSetupFunctions()),
1246
- gasFees: await this.getCurrentBaseFees(),
1247
- skipFeeEnforcement,
1248
- txsPermitted: !this.config.disableTransactions,
1249
- });
1321
+ const validator = createTxValidatorForAcceptingTxsOverRPC(
1322
+ db,
1323
+ this.contractDataSource,
1324
+ verifier,
1325
+ {
1326
+ timestamp: nextSlotTimestamp,
1327
+ blockNumber,
1328
+ l1ChainId: this.l1ChainId,
1329
+ rollupVersion: this.version,
1330
+ setupAllowList: [
1331
+ ...(await getDefaultAllowedSetupFunctions()),
1332
+ ...(this.config.txPublicSetupAllowListExtend ?? []),
1333
+ ],
1334
+ gasFees: await this.getCurrentMinFees(),
1335
+ skipFeeEnforcement,
1336
+ txsPermitted: !this.config.disableTransactions,
1337
+ },
1338
+ this.log.getBindings(),
1339
+ );
1250
1340
 
1251
1341
  return await validator.validateTx(tx);
1252
1342
  }
@@ -1315,7 +1405,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1315
1405
  }
1316
1406
 
1317
1407
  // And it has an L2 block hash
1318
- const l2BlockHash = await archiver.getL2Tips().then(tips => tips.latest.hash);
1408
+ const l2BlockHash = await archiver.getL2Tips().then(tips => tips.proposed.hash);
1319
1409
  if (!l2BlockHash) {
1320
1410
  this.metrics.recordSnapshotError();
1321
1411
  throw new Error(`Archiver has no latest L2 block hash downloaded. Cannot start snapshot.`);
@@ -1349,7 +1439,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1349
1439
  throw new Error('Archiver implementation does not support rollbacks.');
1350
1440
  }
1351
1441
 
1352
- const finalizedBlock = await archiver.getL2Tips().then(tips => tips.finalized.number);
1442
+ const finalizedBlock = await archiver.getL2Tips().then(tips => tips.finalized.block.number);
1353
1443
  if (targetBlock < finalizedBlock) {
1354
1444
  if (force) {
1355
1445
  this.log.warn(`Clearing world state database to allow rolling back behind finalized block ${finalizedBlock}`);
@@ -1410,16 +1500,107 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1410
1500
  }
1411
1501
  }
1412
1502
 
1503
+ public async reloadKeystore(): Promise<void> {
1504
+ if (!this.config.keyStoreDirectory?.length) {
1505
+ throw new BadRequestError(
1506
+ 'Cannot reload keystore: node is not using a file-based keystore. ' +
1507
+ 'Set KEY_STORE_DIRECTORY to use file-based keystores.',
1508
+ );
1509
+ }
1510
+ if (!this.validatorClient) {
1511
+ throw new BadRequestError('Cannot reload keystore: validator is not enabled.');
1512
+ }
1513
+
1514
+ this.log.info('Reloading keystore from disk');
1515
+
1516
+ // Re-read and validate keystore files
1517
+ const keyStores = loadKeystores(this.config.keyStoreDirectory);
1518
+ const newManager = new KeystoreManager(mergeKeystores(keyStores));
1519
+ await newManager.validateSigners();
1520
+ ValidatorClient.validateKeyStoreConfiguration(newManager, this.log);
1521
+
1522
+ // Validate that every validator's publisher keys overlap with the L1 signers
1523
+ // that were initialized at startup. Publishers cannot be hot-reloaded, so a
1524
+ // validator with a publisher key that doesn't match any existing L1 signer
1525
+ // would silently fail on every proposer slot.
1526
+ if (this.keyStoreManager && this.sequencer) {
1527
+ const oldAdapter = NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager);
1528
+ const availablePublishers = new Set(
1529
+ oldAdapter
1530
+ .getAttesterAddresses()
1531
+ .flatMap(a => oldAdapter.getPublisherAddresses(a).map(p => p.toString().toLowerCase())),
1532
+ );
1533
+
1534
+ const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
1535
+ for (const attester of newAdapter.getAttesterAddresses()) {
1536
+ const pubs = newAdapter.getPublisherAddresses(attester);
1537
+ if (pubs.length > 0 && !pubs.some(p => availablePublishers.has(p.toString().toLowerCase()))) {
1538
+ throw new BadRequestError(
1539
+ `Cannot reload keystore: validator ${attester} has publisher keys ` +
1540
+ `[${pubs.map(p => p.toString()).join(', ')}] but none match the L1 signers initialized at startup ` +
1541
+ `[${[...availablePublishers].join(', ')}]. Publishers cannot be hot-reloaded — ` +
1542
+ `use an existing publisher key or restart the node.`,
1543
+ );
1544
+ }
1545
+ }
1546
+ }
1547
+
1548
+ // Build adapters for old and new keystores to compute diff
1549
+ const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
1550
+ const newAddresses = newAdapter.getAttesterAddresses();
1551
+ const oldAddresses = this.keyStoreManager
1552
+ ? NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager).getAttesterAddresses()
1553
+ : [];
1554
+
1555
+ const oldSet = new Set(oldAddresses.map(a => a.toString()));
1556
+ const newSet = new Set(newAddresses.map(a => a.toString()));
1557
+ const added = newAddresses.filter(a => !oldSet.has(a.toString()));
1558
+ const removed = oldAddresses.filter(a => !newSet.has(a.toString()));
1559
+
1560
+ if (added.length > 0) {
1561
+ this.log.info(`Keystore reload: adding attester keys: ${added.map(a => a.toString()).join(', ')}`);
1562
+ }
1563
+ if (removed.length > 0) {
1564
+ this.log.info(`Keystore reload: removing attester keys: ${removed.map(a => a.toString()).join(', ')}`);
1565
+ }
1566
+ if (added.length === 0 && removed.length === 0) {
1567
+ this.log.info('Keystore reload: attester keys unchanged');
1568
+ }
1569
+
1570
+ // Update the validator client (coinbase, feeRecipient, attester keys)
1571
+ this.validatorClient.reloadKeystore(newManager);
1572
+
1573
+ // Update the publisher factory's keystore so newly-added validators
1574
+ // can be matched to existing publisher keys when proposing blocks.
1575
+ if (this.sequencer) {
1576
+ this.sequencer.updatePublisherNodeKeyStore(newAdapter);
1577
+ }
1578
+
1579
+ // Update slasher's "don't-slash-self" list with new validator addresses
1580
+ if (this.slasherClient && !this.config.slashSelfAllowed) {
1581
+ const slashValidatorsNever = unique(
1582
+ [...(this.config.slashValidatorsNever ?? []), ...newAddresses].map(a => a.toString()),
1583
+ ).map(EthAddress.fromString);
1584
+ this.slasherClient.updateConfig({ slashValidatorsNever });
1585
+ }
1586
+
1587
+ this.keyStoreManager = newManager;
1588
+ this.log.info('Keystore reloaded: coinbase, feeRecipient, and attester keys updated');
1589
+ }
1590
+
1591
+ #getInitialHeaderHash(): Promise<BlockHash> {
1592
+ if (!this.initialHeaderHashPromise) {
1593
+ this.initialHeaderHashPromise = this.worldStateSynchronizer.getCommitted().getInitialHeader().hash();
1594
+ }
1595
+ return this.initialHeaderHashPromise;
1596
+ }
1597
+
1413
1598
  /**
1414
1599
  * Returns an instance of MerkleTreeOperations having first ensured the world state is fully synched
1415
- * @param blockNumber - The block number at which to get the data.
1600
+ * @param block - The block parameter (block number, block hash, or 'latest') at which to get the data.
1416
1601
  * @returns An instance of a committed MerkleTreeOperations
1417
1602
  */
1418
- async #getWorldState(blockNumber: BlockParameter) {
1419
- if (typeof blockNumber === 'number' && blockNumber < INITIAL_L2_BLOCK_NUM - 1) {
1420
- throw new Error('Invalid block number to get world state for: ' + blockNumber);
1421
- }
1422
-
1603
+ async #getWorldState(block: BlockParameter) {
1423
1604
  let blockSyncedTo: BlockNumber = BlockNumber.ZERO;
1424
1605
  try {
1425
1606
  // Attempt to sync the world state if necessary
@@ -1428,15 +1609,39 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1428
1609
  this.log.error(`Error getting world state: ${err}`);
1429
1610
  }
1430
1611
 
1431
- // using a snapshot could be less efficient than using the committed db
1432
- if (blockNumber === 'latest' /*|| blockNumber === blockSyncedTo*/) {
1433
- this.log.debug(`Using committed db for block ${blockNumber}, world state synced upto ${blockSyncedTo}`);
1612
+ if (block === 'latest') {
1613
+ this.log.debug(`Using committed db for block 'latest', world state synced upto ${blockSyncedTo}`);
1434
1614
  return this.worldStateSynchronizer.getCommitted();
1435
- } else if (blockNumber <= blockSyncedTo) {
1615
+ }
1616
+
1617
+ if (BlockHash.isBlockHash(block)) {
1618
+ const initialBlockHash = await this.#getInitialHeaderHash();
1619
+ if (block.equals(initialBlockHash)) {
1620
+ // Block source doesn't handle initial header so we need to handle the case separately.
1621
+ return this.worldStateSynchronizer.getSnapshot(BlockNumber.ZERO);
1622
+ }
1623
+
1624
+ const header = await this.blockSource.getBlockHeaderByHash(block);
1625
+ if (!header) {
1626
+ throw new Error(
1627
+ `Block hash ${block.toString()} not found when querying world state. If the node API has been queried with anchor block hash possibly a reorg has occurred.`,
1628
+ );
1629
+ }
1630
+ const blockNumber = header.getBlockNumber();
1436
1631
  this.log.debug(`Using snapshot for block ${blockNumber}, world state synced upto ${blockSyncedTo}`);
1437
- return this.worldStateSynchronizer.getSnapshot(blockNumber as BlockNumber);
1438
- } else {
1439
- throw new Error(`Block ${blockNumber} not yet synced`);
1632
+ return this.worldStateSynchronizer.getSnapshot(blockNumber);
1633
+ }
1634
+
1635
+ // Block number provided
1636
+ {
1637
+ const blockNumber = block as BlockNumber;
1638
+
1639
+ if (blockNumber > blockSyncedTo) {
1640
+ throw new Error(`Queried block ${block} not yet synced by the node (node is synced upto ${blockSyncedTo}).`);
1641
+ }
1642
+
1643
+ this.log.debug(`Using snapshot for block ${blockNumber}, world state synced upto ${blockSyncedTo}`);
1644
+ return this.worldStateSynchronizer.getSnapshot(blockNumber);
1440
1645
  }
1441
1646
  }
1442
1647