@aztec/aztec-node 0.0.1-commit.0c875d939 → 0.0.1-commit.10bd49492

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,7 @@
1
1
  import { Archiver, createArchiver } from '@aztec/archiver';
2
2
  import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
3
3
  import { type BlobClientInterface, createBlobClientWithFileStores } from '@aztec/blob-client/client';
4
+ import { Blob } from '@aztec/blob-lib';
4
5
  import { ARCHIVE_HEIGHT, type L1_TO_L2_MSG_TREE_HEIGHT, type NOTE_HASH_TREE_HEIGHT } from '@aztec/constants';
5
6
  import { EpochCache, type EpochCacheInterface } from '@aztec/epoch-cache';
6
7
  import { createEthereumChain } from '@aztec/ethereum/chain';
@@ -8,7 +9,7 @@ import { getPublicClient } from '@aztec/ethereum/client';
8
9
  import { RegistryContract, RollupContract } from '@aztec/ethereum/contracts';
9
10
  import type { L1ContractAddresses } from '@aztec/ethereum/l1-contract-addresses';
10
11
  import { BlockNumber, CheckpointNumber, EpochNumber, SlotNumber } from '@aztec/foundation/branded-types';
11
- import { compactArray, pick } from '@aztec/foundation/collection';
12
+ import { compactArray, pick, unique } from '@aztec/foundation/collection';
12
13
  import { Fr } from '@aztec/foundation/curves/bn254';
13
14
  import { EthAddress } from '@aztec/foundation/eth-address';
14
15
  import { BadRequestError } from '@aztec/foundation/json-rpc';
@@ -16,14 +17,19 @@ import { type Logger, createLogger } from '@aztec/foundation/log';
16
17
  import { count } from '@aztec/foundation/string';
17
18
  import { DateProvider, Timer } from '@aztec/foundation/timer';
18
19
  import { MembershipWitness, SiblingPath } from '@aztec/foundation/trees';
19
- import { KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
20
+ import { type KeyStore, KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
20
21
  import { trySnapshotSync, uploadSnapshot } from '@aztec/node-lib/actions';
22
+ import { createForwarderL1TxUtilsFromSigners, createL1TxUtilsFromSigners } from '@aztec/node-lib/factories';
21
23
  import {
22
- createForwarderL1TxUtilsFromEthSigner,
23
- createL1TxUtilsWithBlobsFromEthSigner,
24
- } from '@aztec/node-lib/factories';
25
- import { type P2P, type P2PClientDeps, createP2PClient, getDefaultAllowedSetupFunctions } from '@aztec/p2p';
24
+ type P2P,
25
+ type P2PClientDeps,
26
+ createP2PClient,
27
+ createTxValidatorForAcceptingTxsOverRPC,
28
+ getDefaultAllowedSetupFunctions,
29
+ } from '@aztec/p2p';
26
30
  import { ProtocolContractAddress } from '@aztec/protocol-contracts';
31
+ import { type ProverNode, type ProverNodeDeps, createProverNode } from '@aztec/prover-node';
32
+ import { createKeyStoreForProver } from '@aztec/prover-node/config';
27
33
  import { GlobalVariableBuilder, SequencerClient, type SequencerPublisher } from '@aztec/sequencer-client';
28
34
  import { PublicProcessorFactory } from '@aztec/simulator/server';
29
35
  import {
@@ -35,7 +41,14 @@ import {
35
41
  } from '@aztec/slasher';
36
42
  import { CollectionLimitsConfig, PublicSimulatorConfig } from '@aztec/stdlib/avm';
37
43
  import { AztecAddress } from '@aztec/stdlib/aztec-address';
38
- import { BlockHash, type BlockParameter, type DataInBlock, L2Block, type L2BlockSource } from '@aztec/stdlib/block';
44
+ import {
45
+ type BlockData,
46
+ BlockHash,
47
+ type BlockParameter,
48
+ type DataInBlock,
49
+ L2Block,
50
+ type L2BlockSource,
51
+ } from '@aztec/stdlib/block';
39
52
  import type { PublishedCheckpoint } from '@aztec/stdlib/checkpoint';
40
53
  import type {
41
54
  ContractClassPublic,
@@ -63,9 +76,9 @@ import {
63
76
  type WorldStateSynchronizer,
64
77
  tryStop,
65
78
  } from '@aztec/stdlib/interfaces/server';
66
- import type { LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
79
+ import type { DebugLogStore, LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
80
+ import { InMemoryDebugLogStore, NullDebugLogStore } from '@aztec/stdlib/logs';
67
81
  import { InboxLeaf, type L1ToL2MessageSource } from '@aztec/stdlib/messaging';
68
- import { P2PClientType } from '@aztec/stdlib/p2p';
69
82
  import type { Offense, SlashPayloadRound } from '@aztec/stdlib/slashing';
70
83
  import type { NullifierLeafPreimage, PublicDataTreeLeaf, PublicDataTreeLeafPreimage } from '@aztec/stdlib/trees';
71
84
  import { MerkleTreeId, NullifierMembershipWitness, PublicDataWitness } from '@aztec/stdlib/trees';
@@ -97,7 +110,6 @@ import {
97
110
  ValidatorClient,
98
111
  createBlockProposalHandler,
99
112
  createValidatorClient,
100
- createValidatorForAcceptingTxs,
101
113
  } from '@aztec/validator-client';
102
114
  import { createWorldStateSynchronizer } from '@aztec/world-state';
103
115
 
@@ -129,6 +141,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
129
141
  protected readonly l1ToL2MessageSource: L1ToL2MessageSource,
130
142
  protected readonly worldStateSynchronizer: WorldStateSynchronizer,
131
143
  protected readonly sequencer: SequencerClient | undefined,
144
+ protected readonly proverNode: ProverNode | undefined,
132
145
  protected readonly slasherClient: SlasherClientInterface | undefined,
133
146
  protected readonly validatorsSentinel: Sentinel | undefined,
134
147
  protected readonly epochPruneWatcher: EpochPruneWatcher | undefined,
@@ -141,12 +154,22 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
141
154
  private telemetry: TelemetryClient = getTelemetryClient(),
142
155
  private log = createLogger('node'),
143
156
  private blobClient?: BlobClientInterface,
157
+ private validatorClient?: ValidatorClient,
158
+ private keyStoreManager?: KeystoreManager,
159
+ private debugLogStore: DebugLogStore = new NullDebugLogStore(),
144
160
  ) {
145
161
  this.metrics = new NodeMetrics(telemetry, 'AztecNodeService');
146
162
  this.tracer = telemetry.getTracer('AztecNodeService');
147
163
 
148
164
  this.log.info(`Aztec Node version: ${this.packageVersion}`);
149
165
  this.log.info(`Aztec Node started on chain 0x${l1ChainId.toString(16)}`, config.l1Contracts);
166
+
167
+ // A defensive check that protects us against introducing a bug in the complex `createAndSync` function. We must
168
+ // never have debugLogStore enabled when not in test mode because then we would be accumulating debug logs in
169
+ // memory which could be a DoS vector on the sequencer (since no fees are paid for debug logs).
170
+ if (debugLogStore.isEnabled && config.realProofs) {
171
+ throw new Error('debugLogStore should never be enabled when realProofs are set');
172
+ }
150
173
  }
151
174
 
152
175
  public async getWorldStateSyncStatus(): Promise<WorldStateSyncStatus> {
@@ -170,11 +193,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
170
193
  logger?: Logger;
171
194
  publisher?: SequencerPublisher;
172
195
  dateProvider?: DateProvider;
173
- p2pClientDeps?: P2PClientDeps<P2PClientType.Full>;
196
+ p2pClientDeps?: P2PClientDeps;
197
+ proverNodeDeps?: Partial<ProverNodeDeps>;
174
198
  } = {},
175
199
  options: {
176
200
  prefilledPublicData?: PublicDataTreeLeaf[];
177
201
  dontStartSequencer?: boolean;
202
+ dontStartProverNode?: boolean;
178
203
  } = {},
179
204
  ): Promise<AztecNodeService> {
180
205
  const config = { ...inputConfig }; // Copy the config so we dont mutate the input object
@@ -184,16 +209,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
184
209
  const dateProvider = deps.dateProvider ?? new DateProvider();
185
210
  const ethereumChain = createEthereumChain(config.l1RpcUrls, config.l1ChainId);
186
211
 
187
- // Build a key store from file if given or from environment otherwise
212
+ // Build a key store from file if given or from environment otherwise.
213
+ // We keep the raw KeyStore available so we can merge with prover keys if enableProverNode is set.
188
214
  let keyStoreManager: KeystoreManager | undefined;
189
215
  const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
190
216
  if (keyStoreProvided) {
191
217
  const keyStores = loadKeystores(config.keyStoreDirectory!);
192
218
  keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
193
219
  } else {
194
- const keyStore = createKeyStoreForValidator(config);
195
- if (keyStore) {
196
- keyStoreManager = new KeystoreManager(keyStore);
220
+ const rawKeyStores: KeyStore[] = [];
221
+ const validatorKeyStore = createKeyStoreForValidator(config);
222
+ if (validatorKeyStore) {
223
+ rawKeyStores.push(validatorKeyStore);
224
+ }
225
+ if (config.enableProverNode) {
226
+ const proverKeyStore = createKeyStoreForProver(config);
227
+ if (proverKeyStore) {
228
+ rawKeyStores.push(proverKeyStore);
229
+ }
230
+ }
231
+ if (rawKeyStores.length > 0) {
232
+ keyStoreManager = new KeystoreManager(
233
+ rawKeyStores.length === 1 ? rawKeyStores[0] : mergeKeystores(rawKeyStores),
234
+ );
197
235
  }
198
236
  }
199
237
 
@@ -204,10 +242,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
204
242
  if (keyStoreManager === undefined) {
205
243
  throw new Error('Failed to create key store, a requirement for running a validator');
206
244
  }
207
- if (!keyStoreProvided) {
208
- log.warn(
209
- 'KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS',
210
- );
245
+ if (!keyStoreProvided && process.env.NODE_ENV !== 'test') {
246
+ log.warn("Keystore created from env: it's recommended to use a file-based key store for production");
211
247
  }
212
248
  ValidatorClient.validateKeyStoreConfiguration(keyStoreManager, log);
213
249
  }
@@ -235,10 +271,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
235
271
  config.l1Contracts = { ...config.l1Contracts, ...l1ContractsAddresses };
236
272
 
237
273
  const rollupContract = new RollupContract(publicClient, config.l1Contracts.rollupAddress.toString());
238
- const [l1GenesisTime, slotDuration, rollupVersionFromRollup] = await Promise.all([
274
+ const [l1GenesisTime, slotDuration, rollupVersionFromRollup, rollupManaLimit] = await Promise.all([
239
275
  rollupContract.getL1GenesisTime(),
240
276
  rollupContract.getSlotDuration(),
241
277
  rollupContract.getVersion(),
278
+ rollupContract.getManaLimit().then(Number),
242
279
  ] as const);
243
280
 
244
281
  config.rollupVersion ??= Number(rollupVersionFromRollup);
@@ -249,7 +286,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
249
286
  );
250
287
  }
251
288
 
252
- const blobClient = await createBlobClientWithFileStores(config, createLogger('node:blob-client:client'));
289
+ const blobClient = await createBlobClientWithFileStores(config, log.createChild('blob-client'));
253
290
 
254
291
  // attempt snapshot sync if possible
255
292
  await trySnapshotSync(config, log);
@@ -273,14 +310,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
273
310
  config.realProofs || config.debugForceTxProofVerification
274
311
  ? await BBCircuitVerifier.new(config)
275
312
  : new TestCircuitVerifier(config.proverTestVerificationDelayMs);
313
+
314
+ let debugLogStore: DebugLogStore;
276
315
  if (!config.realProofs) {
277
316
  log.warn(`Aztec node is accepting fake proofs`);
317
+
318
+ debugLogStore = new InMemoryDebugLogStore();
319
+ log.info(
320
+ 'Aztec node started in test mode (realProofs set to false) hence debug logs from public functions will be collected and served',
321
+ );
322
+ } else {
323
+ debugLogStore = new NullDebugLogStore();
278
324
  }
325
+
279
326
  const proofVerifier = new QueuedIVCVerifier(config, circuitVerifier);
280
327
 
328
+ const proverOnly = config.enableProverNode && config.disableValidator;
329
+ if (proverOnly) {
330
+ log.info('Starting in prover-only mode: skipping validator, sequencer, sentinel, and slasher subsystems');
331
+ }
332
+
281
333
  // create the tx pool and the p2p client, which will need the l2 block source
282
334
  const p2pClient = await createP2PClient(
283
- P2PClientType.Full,
284
335
  config,
285
336
  archiver,
286
337
  proofVerifier,
@@ -292,49 +343,59 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
292
343
  deps.p2pClientDeps,
293
344
  );
294
345
 
295
- // We should really not be modifying the config object
296
- config.txPublicSetupAllowList = config.txPublicSetupAllowList ?? (await getDefaultAllowedSetupFunctions());
346
+ // We'll accumulate sentinel watchers here
347
+ const watchers: Watcher[] = [];
297
348
 
298
- // Create FullNodeCheckpointsBuilder for validator and non-validator block proposal handling
349
+ // Create FullNodeCheckpointsBuilder for block proposal handling and tx validation.
350
+ // Override maxTxsPerCheckpoint with the validator-specific limit if set.
299
351
  const validatorCheckpointsBuilder = new FullNodeCheckpointsBuilder(
300
- { ...config, l1GenesisTime, slotDuration: Number(slotDuration) },
352
+ {
353
+ ...config,
354
+ l1GenesisTime,
355
+ slotDuration: Number(slotDuration),
356
+ rollupManaLimit,
357
+ maxTxsPerCheckpoint: config.validateMaxTxsPerCheckpoint,
358
+ },
301
359
  worldStateSynchronizer,
302
360
  archiver,
303
361
  dateProvider,
304
362
  telemetry,
305
363
  );
306
364
 
307
- // We'll accumulate sentinel watchers here
308
- const watchers: Watcher[] = [];
365
+ let validatorClient: ValidatorClient | undefined;
309
366
 
310
- // Create validator client if required
311
- const validatorClient = await createValidatorClient(config, {
312
- checkpointsBuilder: validatorCheckpointsBuilder,
313
- worldState: worldStateSynchronizer,
314
- p2pClient,
315
- telemetry,
316
- dateProvider,
317
- epochCache,
318
- blockSource: archiver,
319
- l1ToL2MessageSource: archiver,
320
- keyStoreManager,
321
- blobClient,
322
- });
367
+ if (!proverOnly) {
368
+ // Create validator client if required
369
+ validatorClient = await createValidatorClient(config, {
370
+ checkpointsBuilder: validatorCheckpointsBuilder,
371
+ worldState: worldStateSynchronizer,
372
+ p2pClient,
373
+ telemetry,
374
+ dateProvider,
375
+ epochCache,
376
+ blockSource: archiver,
377
+ l1ToL2MessageSource: archiver,
378
+ keyStoreManager,
379
+ blobClient,
380
+ });
323
381
 
324
- // If we have a validator client, register it as a source of offenses for the slasher,
325
- // and have it register callbacks on the p2p client *before* we start it, otherwise messages
326
- // like attestations or auths will fail.
327
- if (validatorClient) {
328
- watchers.push(validatorClient);
329
- if (!options.dontStartSequencer) {
330
- await validatorClient.registerHandlers();
382
+ // If we have a validator client, register it as a source of offenses for the slasher,
383
+ // and have it register callbacks on the p2p client *before* we start it, otherwise messages
384
+ // like attestations or auths will fail.
385
+ if (validatorClient) {
386
+ watchers.push(validatorClient);
387
+ if (!options.dontStartSequencer) {
388
+ await validatorClient.registerHandlers();
389
+ }
331
390
  }
332
391
  }
333
392
 
334
- // If there's no validator client but alwaysReexecuteBlockProposals is enabled,
335
- // create a BlockProposalHandler to reexecute block proposals for monitoring
336
- if (!validatorClient && config.alwaysReexecuteBlockProposals) {
337
- log.info('Setting up block proposal reexecution for monitoring');
393
+ // If there's no validator client, create a BlockProposalHandler to handle block proposals
394
+ // for monitoring or reexecution. Reexecution (default) allows us to follow the pending chain,
395
+ // while non-reexecution is used for validating the proposals and collecting their txs.
396
+ if (!validatorClient) {
397
+ const reexecute = !!config.alwaysReexecuteBlockProposals;
398
+ log.info(`Setting up block proposal handler` + (reexecute ? ' with reexecution of proposals' : ''));
338
399
  createBlockProposalHandler(config, {
339
400
  checkpointsBuilder: validatorCheckpointsBuilder,
340
401
  worldState: worldStateSynchronizer,
@@ -344,7 +405,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
344
405
  p2pClient,
345
406
  dateProvider,
346
407
  telemetry,
347
- }).registerForReexecution(p2pClient);
408
+ }).register(p2pClient, reexecute);
348
409
  }
349
410
 
350
411
  // Start world state and wait for it to sync to the archiver.
@@ -353,29 +414,33 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
353
414
  // Start p2p. Note that it depends on world state to be running.
354
415
  await p2pClient.start();
355
416
 
356
- const validatorsSentinel = await createSentinel(epochCache, archiver, p2pClient, config);
357
- if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
358
- watchers.push(validatorsSentinel);
359
- }
360
-
417
+ let validatorsSentinel: Awaited<ReturnType<typeof createSentinel>> | undefined;
361
418
  let epochPruneWatcher: EpochPruneWatcher | undefined;
362
- if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
363
- epochPruneWatcher = new EpochPruneWatcher(
364
- archiver,
365
- archiver,
366
- epochCache,
367
- p2pClient.getTxProvider(),
368
- validatorCheckpointsBuilder,
369
- config,
370
- );
371
- watchers.push(epochPruneWatcher);
372
- }
373
-
374
- // We assume we want to slash for invalid attestations unless all max penalties are set to 0
375
419
  let attestationsBlockWatcher: AttestationsBlockWatcher | undefined;
376
- if (config.slashProposeInvalidAttestationsPenalty > 0n || config.slashAttestDescendantOfInvalidPenalty > 0n) {
377
- attestationsBlockWatcher = new AttestationsBlockWatcher(archiver, epochCache, config);
378
- watchers.push(attestationsBlockWatcher);
420
+
421
+ if (!proverOnly) {
422
+ validatorsSentinel = await createSentinel(epochCache, archiver, p2pClient, config);
423
+ if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
424
+ watchers.push(validatorsSentinel);
425
+ }
426
+
427
+ if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
428
+ epochPruneWatcher = new EpochPruneWatcher(
429
+ archiver,
430
+ archiver,
431
+ epochCache,
432
+ p2pClient.getTxProvider(),
433
+ validatorCheckpointsBuilder,
434
+ config,
435
+ );
436
+ watchers.push(epochPruneWatcher);
437
+ }
438
+
439
+ // We assume we want to slash for invalid attestations unless all max penalties are set to 0
440
+ if (config.slashProposeInvalidAttestationsPenalty > 0n || config.slashAttestDescendantOfInvalidPenalty > 0n) {
441
+ attestationsBlockWatcher = new AttestationsBlockWatcher(archiver, epochCache, config);
442
+ watchers.push(attestationsBlockWatcher);
443
+ }
379
444
  }
380
445
 
381
446
  // Start p2p-related services once the archiver has completed sync
@@ -412,28 +477,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
412
477
  );
413
478
  await slasherClient.start();
414
479
 
415
- const l1TxUtils = config.publisherForwarderAddress
416
- ? await createForwarderL1TxUtilsFromEthSigner(
480
+ const l1TxUtils = config.sequencerPublisherForwarderAddress
481
+ ? await createForwarderL1TxUtilsFromSigners(
417
482
  publicClient,
418
483
  keyStoreManager!.createAllValidatorPublisherSigners(),
419
- config.publisherForwarderAddress,
484
+ config.sequencerPublisherForwarderAddress,
420
485
  { ...config, scope: 'sequencer' },
421
- { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
486
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
422
487
  )
423
- : await createL1TxUtilsWithBlobsFromEthSigner(
488
+ : await createL1TxUtilsFromSigners(
424
489
  publicClient,
425
490
  keyStoreManager!.createAllValidatorPublisherSigners(),
426
491
  { ...config, scope: 'sequencer' },
427
- { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
492
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
428
493
  );
429
494
 
430
495
  // Create and start the sequencer client
431
496
  const checkpointsBuilder = new CheckpointsBuilder(
432
- { ...config, l1GenesisTime, slotDuration: Number(slotDuration) },
497
+ { ...config, l1GenesisTime, slotDuration: Number(slotDuration), rollupManaLimit },
433
498
  worldStateSynchronizer,
434
499
  archiver,
435
500
  dateProvider,
436
501
  telemetry,
502
+ debugLogStore,
437
503
  );
438
504
 
439
505
  sequencer = await SequencerClient.new(config, {
@@ -461,6 +527,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
461
527
  log.warn(`Sequencer created but not started`);
462
528
  }
463
529
 
530
+ // Create prover node subsystem if enabled
531
+ let proverNode: ProverNode | undefined;
532
+ if (config.enableProverNode) {
533
+ proverNode = await createProverNode(config, {
534
+ ...deps.proverNodeDeps,
535
+ telemetry,
536
+ dateProvider,
537
+ archiver,
538
+ worldStateSynchronizer,
539
+ p2pClient,
540
+ epochCache,
541
+ blobClient,
542
+ keyStoreManager,
543
+ });
544
+
545
+ if (!options.dontStartProverNode) {
546
+ await proverNode.start();
547
+ log.info(`Prover node subsystem started`);
548
+ } else {
549
+ log.info(`Prover node subsystem created but not started`);
550
+ }
551
+ }
552
+
464
553
  const globalVariableBuilder = new GlobalVariableBuilder({
465
554
  ...config,
466
555
  rollupVersion: BigInt(config.rollupVersion),
@@ -468,7 +557,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
468
557
  slotDuration: Number(slotDuration),
469
558
  });
470
559
 
471
- return new AztecNodeService(
560
+ const node = new AztecNodeService(
472
561
  config,
473
562
  p2pClient,
474
563
  archiver,
@@ -477,6 +566,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
477
566
  archiver,
478
567
  worldStateSynchronizer,
479
568
  sequencer,
569
+ proverNode,
480
570
  slasherClient,
481
571
  validatorsSentinel,
482
572
  epochPruneWatcher,
@@ -489,7 +579,12 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
489
579
  telemetry,
490
580
  log,
491
581
  blobClient,
582
+ validatorClient,
583
+ keyStoreManager,
584
+ debugLogStore,
492
585
  );
586
+
587
+ return node;
493
588
  }
494
589
 
495
590
  /**
@@ -500,6 +595,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
500
595
  return this.sequencer;
501
596
  }
502
597
 
598
+ /** Returns the prover node subsystem, if enabled. */
599
+ public getProverNode(): ProverNode | undefined {
600
+ return this.proverNode;
601
+ }
602
+
503
603
  public getBlockSource(): L2BlockSource {
504
604
  return this.blockSource;
505
605
  }
@@ -525,7 +625,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
525
625
  }
526
626
 
527
627
  public async getAllowedPublicSetup(): Promise<AllowedElement[]> {
528
- return this.config.txPublicSetupAllowList ?? (await getDefaultAllowedSetupFunctions());
628
+ return [...(await getDefaultAllowedSetupFunctions()), ...(this.config.txPublicSetupAllowListExtend ?? [])];
529
629
  }
530
630
 
531
631
  /**
@@ -620,6 +720,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
620
720
  return (await this.blockSource.getCheckpointedBlocks(from, limit)) ?? [];
621
721
  }
622
722
 
723
+ public getCheckpointsDataForEpoch(epochNumber: EpochNumber) {
724
+ return this.blockSource.getCheckpointsDataForEpoch(epochNumber);
725
+ }
726
+
623
727
  /**
624
728
  * Method to fetch the current min L2 fees.
625
729
  * @returns The current min L2 fees.
@@ -652,6 +756,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
652
756
  return await this.blockSource.getCheckpointedL2BlockNumber();
653
757
  }
654
758
 
759
+ public getCheckpointNumber(): Promise<CheckpointNumber> {
760
+ return this.blockSource.getCheckpointNumber();
761
+ }
762
+
655
763
  /**
656
764
  * Method to fetch the version of the package.
657
765
  * @returns The node package version
@@ -762,8 +870,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
762
870
  }
763
871
 
764
872
  await this.p2pClient!.sendTx(tx);
765
- this.metrics.receivedTx(timer.ms(), true);
766
- this.log.info(`Received tx ${txHash}`, { txHash });
873
+ const duration = timer.ms();
874
+ this.metrics.receivedTx(duration, true);
875
+ this.log.info(`Received tx ${txHash} in ${duration}ms`, { txHash });
767
876
  }
768
877
 
769
878
  public async getTxReceipt(txHash: TxHash): Promise<TxReceipt> {
@@ -775,18 +884,22 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
775
884
  // Then get the actual tx from the archiver, which tracks every tx in a mined block.
776
885
  const settledTxReceipt = await this.blockSource.getSettledTxReceipt(txHash);
777
886
 
887
+ let receipt: TxReceipt;
778
888
  if (settledTxReceipt) {
779
- // If the archiver has the receipt then return it.
780
- return settledTxReceipt;
889
+ receipt = settledTxReceipt;
781
890
  } else if (isKnownToPool) {
782
891
  // If the tx is in the pool but not in the archiver, it's pending.
783
892
  // This handles race conditions between archiver and p2p, where the archiver
784
893
  // has pruned the block in which a tx was mined, but p2p has not caught up yet.
785
- return new TxReceipt(txHash, TxStatus.PENDING, undefined, undefined);
894
+ receipt = new TxReceipt(txHash, TxStatus.PENDING, undefined, undefined);
786
895
  } else {
787
896
  // Otherwise, if we don't know the tx, we consider it dropped.
788
- return new TxReceipt(txHash, TxStatus.DROPPED, undefined, 'Tx dropped by P2P node');
897
+ receipt = new TxReceipt(txHash, TxStatus.DROPPED, undefined, 'Tx dropped by P2P node');
789
898
  }
899
+
900
+ this.debugLogStore.decorateReceiptWithLogs(txHash.toString(), receipt);
901
+
902
+ return receipt;
790
903
  }
791
904
 
792
905
  public getTxEffect(txHash: TxHash): Promise<IndexedTxEffect | undefined> {
@@ -803,6 +916,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
803
916
  await tryStop(this.slasherClient);
804
917
  await tryStop(this.proofVerifier);
805
918
  await tryStop(this.sequencer);
919
+ await tryStop(this.proverNode);
806
920
  await tryStop(this.p2pClient);
807
921
  await tryStop(this.worldStateSynchronizer);
808
922
  await tryStop(this.blockSource);
@@ -861,53 +975,59 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
861
975
  treeId,
862
976
  leafValues.map(x => x.toBuffer()),
863
977
  );
864
- // We filter out undefined values
865
- const indices = maybeIndices.filter(x => x !== undefined) as bigint[];
978
+ // Filter out undefined values to query block numbers only for found leaves
979
+ const definedIndices = maybeIndices.filter(x => x !== undefined);
866
980
 
867
- // Now we find the block numbers for the indices
868
- const blockNumbers = await committedDb.getBlockNumbersForLeafIndices(treeId, indices);
981
+ // Now we find the block numbers for the defined indices
982
+ const blockNumbers = await committedDb.getBlockNumbersForLeafIndices(treeId, definedIndices);
869
983
 
870
- // If any of the block numbers are undefined, we throw an error.
871
- for (let i = 0; i < indices.length; i++) {
872
- if (blockNumbers[i] === undefined) {
873
- throw new Error(`Block number is undefined for leaf index ${indices[i]} in tree ${MerkleTreeId[treeId]}`);
984
+ // Build a map from leaf index to block number
985
+ const indexToBlockNumber = new Map<bigint, BlockNumber>();
986
+ for (let i = 0; i < definedIndices.length; i++) {
987
+ const blockNumber = blockNumbers[i];
988
+ if (blockNumber === undefined) {
989
+ throw new Error(
990
+ `Block number is undefined for leaf index ${definedIndices[i]} in tree ${MerkleTreeId[treeId]}`,
991
+ );
874
992
  }
993
+ indexToBlockNumber.set(definedIndices[i], blockNumber);
875
994
  }
876
995
 
877
996
  // Get unique block numbers in order to optimize num calls to getLeafValue function.
878
- const uniqueBlockNumbers = [...new Set(blockNumbers.filter(x => x !== undefined))];
997
+ const uniqueBlockNumbers = [...new Set(indexToBlockNumber.values())];
879
998
 
880
- // Now we obtain the block hashes from the archive tree by calling await `committedDb.getLeafValue(treeId, index)`
881
- // (note that block number corresponds to the leaf index in the archive tree).
999
+ // Now we obtain the block hashes from the archive tree (block number = leaf index in archive tree).
882
1000
  const blockHashes = await Promise.all(
883
1001
  uniqueBlockNumbers.map(blockNumber => {
884
1002
  return committedDb.getLeafValue(MerkleTreeId.ARCHIVE, BigInt(blockNumber));
885
1003
  }),
886
1004
  );
887
1005
 
888
- // If any of the block hashes are undefined, we throw an error.
1006
+ // Build a map from block number to block hash
1007
+ const blockNumberToHash = new Map<BlockNumber, Fr>();
889
1008
  for (let i = 0; i < uniqueBlockNumbers.length; i++) {
890
- if (blockHashes[i] === undefined) {
1009
+ const blockHash = blockHashes[i];
1010
+ if (blockHash === undefined) {
891
1011
  throw new Error(`Block hash is undefined for block number ${uniqueBlockNumbers[i]}`);
892
1012
  }
1013
+ blockNumberToHash.set(uniqueBlockNumbers[i], blockHash);
893
1014
  }
894
1015
 
895
1016
  // Create DataInBlock objects by combining indices, blockNumbers and blockHashes and return them.
896
- return maybeIndices.map((index, i) => {
1017
+ return maybeIndices.map(index => {
897
1018
  if (index === undefined) {
898
1019
  return undefined;
899
1020
  }
900
- const blockNumber = blockNumbers[i];
1021
+ const blockNumber = indexToBlockNumber.get(index);
901
1022
  if (blockNumber === undefined) {
902
- return undefined;
1023
+ throw new Error(`Block number not found for leaf index ${index} in tree ${MerkleTreeId[treeId]}`);
903
1024
  }
904
- const blockHashIndex = uniqueBlockNumbers.indexOf(blockNumber);
905
- const blockHash = blockHashes[blockHashIndex];
906
- if (!blockHash) {
907
- return undefined;
1025
+ const blockHash = blockNumberToHash.get(blockNumber);
1026
+ if (blockHash === undefined) {
1027
+ throw new Error(`Block hash not found for block number ${blockNumber}`);
908
1028
  }
909
1029
  return {
910
- l2BlockNumber: BlockNumber(Number(blockNumber)),
1030
+ l2BlockNumber: blockNumber,
911
1031
  l2BlockHash: new BlockHash(blockHash),
912
1032
  data: index,
913
1033
  };
@@ -953,11 +1073,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
953
1073
  return [witness.index, witness.path];
954
1074
  }
955
1075
 
956
- public async getL1ToL2MessageBlock(l1ToL2Message: Fr): Promise<BlockNumber | undefined> {
1076
+ public async getL1ToL2MessageCheckpoint(l1ToL2Message: Fr): Promise<CheckpointNumber | undefined> {
957
1077
  const messageIndex = await this.l1ToL2MessageSource.getL1ToL2MessageIndex(l1ToL2Message);
958
- return messageIndex
959
- ? BlockNumber.fromCheckpointNumber(InboxLeaf.checkpointNumberFromIndex(messageIndex))
960
- : undefined;
1078
+ return messageIndex ? InboxLeaf.checkpointNumberFromIndex(messageIndex) : undefined;
961
1079
  }
962
1080
 
963
1081
  /**
@@ -1106,6 +1224,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1106
1224
  return await this.blockSource.getBlockHeaderByArchive(archive);
1107
1225
  }
1108
1226
 
1227
+ public getBlockData(number: BlockNumber): Promise<BlockData | undefined> {
1228
+ return this.blockSource.getBlockData(number);
1229
+ }
1230
+
1231
+ public getBlockDataByArchive(archive: Fr): Promise<BlockData | undefined> {
1232
+ return this.blockSource.getBlockDataByArchive(archive);
1233
+ }
1234
+
1109
1235
  /**
1110
1236
  * Simulates the public part of a transaction with the current state.
1111
1237
  * @param tx - The transaction to simulate.
@@ -1129,7 +1255,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1129
1255
  }
1130
1256
 
1131
1257
  const txHash = tx.getTxHash();
1132
- const blockNumber = BlockNumber((await this.blockSource.getBlockNumber()) + 1);
1258
+ const latestBlockNumber = await this.blockSource.getBlockNumber();
1259
+ const blockNumber = BlockNumber.add(latestBlockNumber, 1);
1133
1260
 
1134
1261
  // If sequencer is not initialized, we just set these values to zero for simulation.
1135
1262
  const coinbase = EthAddress.ZERO;
@@ -1153,6 +1280,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1153
1280
  blockNumber,
1154
1281
  });
1155
1282
 
1283
+ // Ensure world-state has caught up with the latest block we loaded from the archiver
1284
+ await this.worldStateSynchronizer.syncImmediate(latestBlockNumber);
1156
1285
  const merkleTreeFork = await this.worldStateSynchronizer.fork();
1157
1286
  try {
1158
1287
  const config = PublicSimulatorConfig.from({
@@ -1168,7 +1297,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1168
1297
  const processor = publicProcessorFactory.create(merkleTreeFork, newGlobalVariables, config);
1169
1298
 
1170
1299
  // REFACTOR: Consider merging ProcessReturnValues into ProcessedTx
1171
- const [processedTxs, failedTxs, _usedTxs, returns] = await processor.process([tx]);
1300
+ const [processedTxs, failedTxs, _usedTxs, returns, debugLogs] = await processor.process([tx]);
1172
1301
  // REFACTOR: Consider returning the error rather than throwing
1173
1302
  if (failedTxs.length) {
1174
1303
  this.log.warn(`Simulated tx ${txHash} fails: ${failedTxs[0].error}`, { txHash });
@@ -1182,6 +1311,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1182
1311
  processedTx.txEffect,
1183
1312
  returns,
1184
1313
  processedTx.gasUsed,
1314
+ debugLogs,
1185
1315
  );
1186
1316
  } finally {
1187
1317
  await merkleTreeFork.close();
@@ -1195,10 +1325,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1195
1325
  const db = this.worldStateSynchronizer.getCommitted();
1196
1326
  const verifier = isSimulation ? undefined : this.proofVerifier;
1197
1327
 
1198
- // We accept transactions if they are not expired by the next slot (checked based on the IncludeByTimestamp field)
1328
+ // We accept transactions if they are not expired by the next slot (checked based on the ExpirationTimestamp field)
1199
1329
  const { ts: nextSlotTimestamp } = this.epochCache.getEpochAndSlotInNextL1Slot();
1200
1330
  const blockNumber = BlockNumber((await this.blockSource.getBlockNumber()) + 1);
1201
- const validator = createValidatorForAcceptingTxs(
1331
+ const l1Constants = await this.blockSource.getL1Constants();
1332
+ const validator = createTxValidatorForAcceptingTxsOverRPC(
1202
1333
  db,
1203
1334
  this.contractDataSource,
1204
1335
  verifier,
@@ -1207,10 +1338,16 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1207
1338
  blockNumber,
1208
1339
  l1ChainId: this.l1ChainId,
1209
1340
  rollupVersion: this.version,
1210
- setupAllowList: this.config.txPublicSetupAllowList ?? (await getDefaultAllowedSetupFunctions()),
1341
+ setupAllowList: [
1342
+ ...(await getDefaultAllowedSetupFunctions()),
1343
+ ...(this.config.txPublicSetupAllowListExtend ?? []),
1344
+ ],
1211
1345
  gasFees: await this.getCurrentMinFees(),
1212
1346
  skipFeeEnforcement,
1213
1347
  txsPermitted: !this.config.disableTransactions,
1348
+ rollupManaLimit: l1Constants.rollupManaLimit,
1349
+ maxBlockL2Gas: this.config.validateMaxL2BlockGas,
1350
+ maxBlockDAGas: this.config.validateMaxDABlockGas,
1214
1351
  },
1215
1352
  this.log.getBindings(),
1216
1353
  );
@@ -1377,6 +1514,94 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
1377
1514
  }
1378
1515
  }
1379
1516
 
1517
+ public async reloadKeystore(): Promise<void> {
1518
+ if (!this.config.keyStoreDirectory?.length) {
1519
+ throw new BadRequestError(
1520
+ 'Cannot reload keystore: node is not using a file-based keystore. ' +
1521
+ 'Set KEY_STORE_DIRECTORY to use file-based keystores.',
1522
+ );
1523
+ }
1524
+ if (!this.validatorClient) {
1525
+ throw new BadRequestError('Cannot reload keystore: validator is not enabled.');
1526
+ }
1527
+
1528
+ this.log.info('Reloading keystore from disk');
1529
+
1530
+ // Re-read and validate keystore files
1531
+ const keyStores = loadKeystores(this.config.keyStoreDirectory);
1532
+ const newManager = new KeystoreManager(mergeKeystores(keyStores));
1533
+ await newManager.validateSigners();
1534
+ ValidatorClient.validateKeyStoreConfiguration(newManager, this.log);
1535
+
1536
+ // Validate that every validator's publisher keys overlap with the L1 signers
1537
+ // that were initialized at startup. Publishers cannot be hot-reloaded, so a
1538
+ // validator with a publisher key that doesn't match any existing L1 signer
1539
+ // would silently fail on every proposer slot.
1540
+ if (this.keyStoreManager && this.sequencer) {
1541
+ const oldAdapter = NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager);
1542
+ const availablePublishers = new Set(
1543
+ oldAdapter
1544
+ .getAttesterAddresses()
1545
+ .flatMap(a => oldAdapter.getPublisherAddresses(a).map(p => p.toString().toLowerCase())),
1546
+ );
1547
+
1548
+ const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
1549
+ for (const attester of newAdapter.getAttesterAddresses()) {
1550
+ const pubs = newAdapter.getPublisherAddresses(attester);
1551
+ if (pubs.length > 0 && !pubs.some(p => availablePublishers.has(p.toString().toLowerCase()))) {
1552
+ throw new BadRequestError(
1553
+ `Cannot reload keystore: validator ${attester} has publisher keys ` +
1554
+ `[${pubs.map(p => p.toString()).join(', ')}] but none match the L1 signers initialized at startup ` +
1555
+ `[${[...availablePublishers].join(', ')}]. Publishers cannot be hot-reloaded — ` +
1556
+ `use an existing publisher key or restart the node.`,
1557
+ );
1558
+ }
1559
+ }
1560
+ }
1561
+
1562
+ // Build adapters for old and new keystores to compute diff
1563
+ const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
1564
+ const newAddresses = newAdapter.getAttesterAddresses();
1565
+ const oldAddresses = this.keyStoreManager
1566
+ ? NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager).getAttesterAddresses()
1567
+ : [];
1568
+
1569
+ const oldSet = new Set(oldAddresses.map(a => a.toString()));
1570
+ const newSet = new Set(newAddresses.map(a => a.toString()));
1571
+ const added = newAddresses.filter(a => !oldSet.has(a.toString()));
1572
+ const removed = oldAddresses.filter(a => !newSet.has(a.toString()));
1573
+
1574
+ if (added.length > 0) {
1575
+ this.log.info(`Keystore reload: adding attester keys: ${added.map(a => a.toString()).join(', ')}`);
1576
+ }
1577
+ if (removed.length > 0) {
1578
+ this.log.info(`Keystore reload: removing attester keys: ${removed.map(a => a.toString()).join(', ')}`);
1579
+ }
1580
+ if (added.length === 0 && removed.length === 0) {
1581
+ this.log.info('Keystore reload: attester keys unchanged');
1582
+ }
1583
+
1584
+ // Update the validator client (coinbase, feeRecipient, attester keys)
1585
+ this.validatorClient.reloadKeystore(newManager);
1586
+
1587
+ // Update the publisher factory's keystore so newly-added validators
1588
+ // can be matched to existing publisher keys when proposing blocks.
1589
+ if (this.sequencer) {
1590
+ this.sequencer.updatePublisherNodeKeyStore(newAdapter);
1591
+ }
1592
+
1593
+ // Update slasher's "don't-slash-self" list with new validator addresses
1594
+ if (this.slasherClient && !this.config.slashSelfAllowed) {
1595
+ const slashValidatorsNever = unique(
1596
+ [...(this.config.slashValidatorsNever ?? []), ...newAddresses].map(a => a.toString()),
1597
+ ).map(EthAddress.fromString);
1598
+ this.slasherClient.updateConfig({ slashValidatorsNever });
1599
+ }
1600
+
1601
+ this.keyStoreManager = newManager;
1602
+ this.log.info('Keystore reloaded: coinbase, feeRecipient, and attester keys updated');
1603
+ }
1604
+
1380
1605
  #getInitialHeaderHash(): Promise<BlockHash> {
1381
1606
  if (!this.initialHeaderHashPromise) {
1382
1607
  this.initialHeaderHashPromise = this.worldStateSynchronizer.getCommitted().getInitialHeader().hash();