@aztec/aztec-node 0.0.1-commit.03f7ef2 → 0.0.1-commit.04852196a
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/aztec-node/config.d.ts +8 -5
- package/dest/aztec-node/config.d.ts.map +1 -1
- package/dest/aztec-node/config.js +11 -3
- package/dest/aztec-node/node_metrics.d.ts +1 -1
- package/dest/aztec-node/node_metrics.d.ts.map +1 -1
- package/dest/aztec-node/node_metrics.js +9 -16
- package/dest/aztec-node/server.d.ts +59 -103
- package/dest/aztec-node/server.d.ts.map +1 -1
- package/dest/aztec-node/server.js +809 -254
- package/dest/sentinel/factory.d.ts +2 -2
- package/dest/sentinel/factory.d.ts.map +1 -1
- package/dest/sentinel/factory.js +1 -1
- package/dest/sentinel/sentinel.d.ts +6 -5
- package/dest/sentinel/sentinel.d.ts.map +1 -1
- package/dest/sentinel/sentinel.js +81 -50
- package/dest/sentinel/store.d.ts +2 -2
- package/dest/sentinel/store.d.ts.map +1 -1
- package/dest/sentinel/store.js +11 -7
- package/package.json +28 -26
- package/src/aztec-node/config.ts +25 -9
- package/src/aztec-node/node_metrics.ts +6 -17
- package/src/aztec-node/server.ts +530 -325
- package/src/sentinel/factory.ts +2 -7
- package/src/sentinel/sentinel.ts +94 -52
- package/src/sentinel/store.ts +12 -12
package/src/aztec-node/server.ts
CHANGED
|
@@ -1,26 +1,15 @@
|
|
|
1
1
|
import { Archiver, createArchiver } from '@aztec/archiver';
|
|
2
2
|
import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
|
|
3
|
-
import { type BlobClientInterface,
|
|
4
|
-
import {
|
|
5
|
-
|
|
6
|
-
createReadOnlyFileStoreBlobClients,
|
|
7
|
-
createWritableFileStoreBlobClient,
|
|
8
|
-
} from '@aztec/blob-client/filestore';
|
|
9
|
-
import {
|
|
10
|
-
ARCHIVE_HEIGHT,
|
|
11
|
-
INITIAL_L2_BLOCK_NUM,
|
|
12
|
-
type L1_TO_L2_MSG_TREE_HEIGHT,
|
|
13
|
-
type NOTE_HASH_TREE_HEIGHT,
|
|
14
|
-
type NULLIFIER_TREE_HEIGHT,
|
|
15
|
-
type PUBLIC_DATA_TREE_HEIGHT,
|
|
16
|
-
} from '@aztec/constants';
|
|
3
|
+
import { type BlobClientInterface, createBlobClientWithFileStores } from '@aztec/blob-client/client';
|
|
4
|
+
import { Blob } from '@aztec/blob-lib';
|
|
5
|
+
import { ARCHIVE_HEIGHT, type L1_TO_L2_MSG_TREE_HEIGHT, type NOTE_HASH_TREE_HEIGHT } from '@aztec/constants';
|
|
17
6
|
import { EpochCache, type EpochCacheInterface } from '@aztec/epoch-cache';
|
|
18
7
|
import { createEthereumChain } from '@aztec/ethereum/chain';
|
|
19
8
|
import { getPublicClient } from '@aztec/ethereum/client';
|
|
20
9
|
import { RegistryContract, RollupContract } from '@aztec/ethereum/contracts';
|
|
21
10
|
import type { L1ContractAddresses } from '@aztec/ethereum/l1-contract-addresses';
|
|
22
|
-
import { BlockNumber, SlotNumber } from '@aztec/foundation/branded-types';
|
|
23
|
-
import { compactArray, pick } from '@aztec/foundation/collection';
|
|
11
|
+
import { BlockNumber, CheckpointNumber, EpochNumber, SlotNumber } from '@aztec/foundation/branded-types';
|
|
12
|
+
import { compactArray, pick, unique } from '@aztec/foundation/collection';
|
|
24
13
|
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
25
14
|
import { EthAddress } from '@aztec/foundation/eth-address';
|
|
26
15
|
import { BadRequestError } from '@aztec/foundation/json-rpc';
|
|
@@ -28,22 +17,20 @@ import { type Logger, createLogger } from '@aztec/foundation/log';
|
|
|
28
17
|
import { count } from '@aztec/foundation/string';
|
|
29
18
|
import { DateProvider, Timer } from '@aztec/foundation/timer';
|
|
30
19
|
import { MembershipWitness, SiblingPath } from '@aztec/foundation/trees';
|
|
31
|
-
import { KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
|
|
20
|
+
import { type KeyStore, KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
|
|
32
21
|
import { trySnapshotSync, uploadSnapshot } from '@aztec/node-lib/actions';
|
|
22
|
+
import { createForwarderL1TxUtilsFromSigners, createL1TxUtilsFromSigners } from '@aztec/node-lib/factories';
|
|
33
23
|
import {
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
24
|
+
type P2P,
|
|
25
|
+
type P2PClientDeps,
|
|
26
|
+
createP2PClient,
|
|
27
|
+
createTxValidatorForAcceptingTxsOverRPC,
|
|
28
|
+
getDefaultAllowedSetupFunctions,
|
|
29
|
+
} from '@aztec/p2p';
|
|
38
30
|
import { ProtocolContractAddress } from '@aztec/protocol-contracts';
|
|
39
|
-
import {
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
SequencerClient,
|
|
43
|
-
type SequencerPublisher,
|
|
44
|
-
createValidatorForAcceptingTxs,
|
|
45
|
-
} from '@aztec/sequencer-client';
|
|
46
|
-
import { CheckpointsBuilder } from '@aztec/sequencer-client';
|
|
31
|
+
import { type ProverNode, type ProverNodeDeps, createProverNode } from '@aztec/prover-node';
|
|
32
|
+
import { createKeyStoreForProver } from '@aztec/prover-node/config';
|
|
33
|
+
import { GlobalVariableBuilder, SequencerClient, type SequencerPublisher } from '@aztec/sequencer-client';
|
|
47
34
|
import { PublicProcessorFactory } from '@aztec/simulator/server';
|
|
48
35
|
import {
|
|
49
36
|
AttestationsBlockWatcher,
|
|
@@ -55,13 +42,14 @@ import {
|
|
|
55
42
|
import { CollectionLimitsConfig, PublicSimulatorConfig } from '@aztec/stdlib/avm';
|
|
56
43
|
import { AztecAddress } from '@aztec/stdlib/aztec-address';
|
|
57
44
|
import {
|
|
45
|
+
type BlockData,
|
|
46
|
+
BlockHash,
|
|
58
47
|
type BlockParameter,
|
|
59
48
|
type DataInBlock,
|
|
60
|
-
|
|
61
|
-
L2BlockHash,
|
|
49
|
+
L2Block,
|
|
62
50
|
type L2BlockSource,
|
|
63
|
-
type PublishedL2Block,
|
|
64
51
|
} from '@aztec/stdlib/block';
|
|
52
|
+
import type { PublishedCheckpoint } from '@aztec/stdlib/checkpoint';
|
|
65
53
|
import type {
|
|
66
54
|
ContractClassPublic,
|
|
67
55
|
ContractDataSource,
|
|
@@ -88,9 +76,9 @@ import {
|
|
|
88
76
|
type WorldStateSynchronizer,
|
|
89
77
|
tryStop,
|
|
90
78
|
} from '@aztec/stdlib/interfaces/server';
|
|
91
|
-
import type { LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
|
|
79
|
+
import type { DebugLogStore, LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
|
|
80
|
+
import { InMemoryDebugLogStore, NullDebugLogStore } from '@aztec/stdlib/logs';
|
|
92
81
|
import { InboxLeaf, type L1ToL2MessageSource } from '@aztec/stdlib/messaging';
|
|
93
|
-
import { P2PClientType } from '@aztec/stdlib/p2p';
|
|
94
82
|
import type { Offense, SlashPayloadRound } from '@aztec/stdlib/slashing';
|
|
95
83
|
import type { NullifierLeafPreimage, PublicDataTreeLeaf, PublicDataTreeLeafPreimage } from '@aztec/stdlib/trees';
|
|
96
84
|
import { MerkleTreeId, NullifierMembershipWitness, PublicDataWitness } from '@aztec/stdlib/trees';
|
|
@@ -116,6 +104,8 @@ import {
|
|
|
116
104
|
trackSpan,
|
|
117
105
|
} from '@aztec/telemetry-client';
|
|
118
106
|
import {
|
|
107
|
+
FullNodeCheckpointsBuilder as CheckpointsBuilder,
|
|
108
|
+
FullNodeCheckpointsBuilder,
|
|
119
109
|
NodeKeystoreAdapter,
|
|
120
110
|
ValidatorClient,
|
|
121
111
|
createBlockProposalHandler,
|
|
@@ -135,6 +125,7 @@ import { NodeMetrics } from './node_metrics.js';
|
|
|
135
125
|
*/
|
|
136
126
|
export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
137
127
|
private metrics: NodeMetrics;
|
|
128
|
+
private initialHeaderHashPromise: Promise<BlockHash> | undefined = undefined;
|
|
138
129
|
|
|
139
130
|
// Prevent two snapshot operations to happen simultaneously
|
|
140
131
|
private isUploadingSnapshot = false;
|
|
@@ -150,6 +141,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
150
141
|
protected readonly l1ToL2MessageSource: L1ToL2MessageSource,
|
|
151
142
|
protected readonly worldStateSynchronizer: WorldStateSynchronizer,
|
|
152
143
|
protected readonly sequencer: SequencerClient | undefined,
|
|
144
|
+
protected readonly proverNode: ProverNode | undefined,
|
|
153
145
|
protected readonly slasherClient: SlasherClientInterface | undefined,
|
|
154
146
|
protected readonly validatorsSentinel: Sentinel | undefined,
|
|
155
147
|
protected readonly epochPruneWatcher: EpochPruneWatcher | undefined,
|
|
@@ -161,12 +153,23 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
161
153
|
private proofVerifier: ClientProtocolCircuitVerifier,
|
|
162
154
|
private telemetry: TelemetryClient = getTelemetryClient(),
|
|
163
155
|
private log = createLogger('node'),
|
|
156
|
+
private blobClient?: BlobClientInterface,
|
|
157
|
+
private validatorClient?: ValidatorClient,
|
|
158
|
+
private keyStoreManager?: KeystoreManager,
|
|
159
|
+
private debugLogStore: DebugLogStore = new NullDebugLogStore(),
|
|
164
160
|
) {
|
|
165
161
|
this.metrics = new NodeMetrics(telemetry, 'AztecNodeService');
|
|
166
162
|
this.tracer = telemetry.getTracer('AztecNodeService');
|
|
167
163
|
|
|
168
164
|
this.log.info(`Aztec Node version: ${this.packageVersion}`);
|
|
169
165
|
this.log.info(`Aztec Node started on chain 0x${l1ChainId.toString(16)}`, config.l1Contracts);
|
|
166
|
+
|
|
167
|
+
// A defensive check that protects us against introducing a bug in the complex `createAndSync` function. We must
|
|
168
|
+
// never have debugLogStore enabled when not in test mode because then we would be accumulating debug logs in
|
|
169
|
+
// memory which could be a DoS vector on the sequencer (since no fees are paid for debug logs).
|
|
170
|
+
if (debugLogStore.isEnabled && config.realProofs) {
|
|
171
|
+
throw new Error('debugLogStore should never be enabled when realProofs are set');
|
|
172
|
+
}
|
|
170
173
|
}
|
|
171
174
|
|
|
172
175
|
public async getWorldStateSyncStatus(): Promise<WorldStateSyncStatus> {
|
|
@@ -190,12 +193,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
190
193
|
logger?: Logger;
|
|
191
194
|
publisher?: SequencerPublisher;
|
|
192
195
|
dateProvider?: DateProvider;
|
|
193
|
-
|
|
194
|
-
|
|
196
|
+
p2pClientDeps?: P2PClientDeps;
|
|
197
|
+
proverNodeDeps?: Partial<ProverNodeDeps>;
|
|
195
198
|
} = {},
|
|
196
199
|
options: {
|
|
197
200
|
prefilledPublicData?: PublicDataTreeLeaf[];
|
|
198
201
|
dontStartSequencer?: boolean;
|
|
202
|
+
dontStartProverNode?: boolean;
|
|
199
203
|
} = {},
|
|
200
204
|
): Promise<AztecNodeService> {
|
|
201
205
|
const config = { ...inputConfig }; // Copy the config so we dont mutate the input object
|
|
@@ -205,16 +209,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
205
209
|
const dateProvider = deps.dateProvider ?? new DateProvider();
|
|
206
210
|
const ethereumChain = createEthereumChain(config.l1RpcUrls, config.l1ChainId);
|
|
207
211
|
|
|
208
|
-
// Build a key store from file if given or from environment otherwise
|
|
212
|
+
// Build a key store from file if given or from environment otherwise.
|
|
213
|
+
// We keep the raw KeyStore available so we can merge with prover keys if enableProverNode is set.
|
|
209
214
|
let keyStoreManager: KeystoreManager | undefined;
|
|
210
215
|
const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
|
|
211
216
|
if (keyStoreProvided) {
|
|
212
217
|
const keyStores = loadKeystores(config.keyStoreDirectory!);
|
|
213
218
|
keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
|
|
214
219
|
} else {
|
|
215
|
-
const
|
|
216
|
-
|
|
217
|
-
|
|
220
|
+
const rawKeyStores: KeyStore[] = [];
|
|
221
|
+
const validatorKeyStore = createKeyStoreForValidator(config);
|
|
222
|
+
if (validatorKeyStore) {
|
|
223
|
+
rawKeyStores.push(validatorKeyStore);
|
|
224
|
+
}
|
|
225
|
+
if (config.enableProverNode) {
|
|
226
|
+
const proverKeyStore = createKeyStoreForProver(config);
|
|
227
|
+
if (proverKeyStore) {
|
|
228
|
+
rawKeyStores.push(proverKeyStore);
|
|
229
|
+
}
|
|
230
|
+
}
|
|
231
|
+
if (rawKeyStores.length > 0) {
|
|
232
|
+
keyStoreManager = new KeystoreManager(
|
|
233
|
+
rawKeyStores.length === 1 ? rawKeyStores[0] : mergeKeystores(rawKeyStores),
|
|
234
|
+
);
|
|
218
235
|
}
|
|
219
236
|
}
|
|
220
237
|
|
|
@@ -225,10 +242,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
225
242
|
if (keyStoreManager === undefined) {
|
|
226
243
|
throw new Error('Failed to create key store, a requirement for running a validator');
|
|
227
244
|
}
|
|
228
|
-
if (!keyStoreProvided) {
|
|
229
|
-
log.warn(
|
|
230
|
-
'KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS',
|
|
231
|
-
);
|
|
245
|
+
if (!keyStoreProvided && process.env.NODE_ENV !== 'test') {
|
|
246
|
+
log.warn("Keystore created from env: it's recommended to use a file-based key store for production");
|
|
232
247
|
}
|
|
233
248
|
ValidatorClient.validateKeyStoreConfiguration(keyStoreManager, log);
|
|
234
249
|
}
|
|
@@ -256,10 +271,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
256
271
|
config.l1Contracts = { ...config.l1Contracts, ...l1ContractsAddresses };
|
|
257
272
|
|
|
258
273
|
const rollupContract = new RollupContract(publicClient, config.l1Contracts.rollupAddress.toString());
|
|
259
|
-
const [l1GenesisTime, slotDuration, rollupVersionFromRollup] = await Promise.all([
|
|
274
|
+
const [l1GenesisTime, slotDuration, rollupVersionFromRollup, rollupManaLimit] = await Promise.all([
|
|
260
275
|
rollupContract.getL1GenesisTime(),
|
|
261
276
|
rollupContract.getSlotDuration(),
|
|
262
277
|
rollupContract.getVersion(),
|
|
278
|
+
rollupContract.getManaLimit().then(Number),
|
|
263
279
|
] as const);
|
|
264
280
|
|
|
265
281
|
config.rollupVersion ??= Number(rollupVersionFromRollup);
|
|
@@ -270,24 +286,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
270
286
|
);
|
|
271
287
|
}
|
|
272
288
|
|
|
273
|
-
const
|
|
274
|
-
l1ChainId: config.l1ChainId,
|
|
275
|
-
rollupVersion: config.rollupVersion,
|
|
276
|
-
rollupAddress: config.l1Contracts.rollupAddress.toString(),
|
|
277
|
-
};
|
|
278
|
-
|
|
279
|
-
const [fileStoreClients, fileStoreUploadClient] = await Promise.all([
|
|
280
|
-
createReadOnlyFileStoreBlobClients(config.blobFileStoreUrls, blobFileStoreMetadata, log),
|
|
281
|
-
createWritableFileStoreBlobClient(config.blobFileStoreUploadUrl, blobFileStoreMetadata, log),
|
|
282
|
-
]);
|
|
283
|
-
|
|
284
|
-
const blobClient =
|
|
285
|
-
deps.blobClient ??
|
|
286
|
-
createBlobClient(config, {
|
|
287
|
-
logger: createLogger('node:blob-client:client'),
|
|
288
|
-
fileStoreClients,
|
|
289
|
-
fileStoreUploadClient,
|
|
290
|
-
});
|
|
289
|
+
const blobClient = await createBlobClientWithFileStores(config, log.createChild('blob-client'));
|
|
291
290
|
|
|
292
291
|
// attempt snapshot sync if possible
|
|
293
292
|
await trySnapshotSync(config, log);
|
|
@@ -311,14 +310,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
311
310
|
config.realProofs || config.debugForceTxProofVerification
|
|
312
311
|
? await BBCircuitVerifier.new(config)
|
|
313
312
|
: new TestCircuitVerifier(config.proverTestVerificationDelayMs);
|
|
313
|
+
|
|
314
|
+
let debugLogStore: DebugLogStore;
|
|
314
315
|
if (!config.realProofs) {
|
|
315
316
|
log.warn(`Aztec node is accepting fake proofs`);
|
|
317
|
+
|
|
318
|
+
debugLogStore = new InMemoryDebugLogStore();
|
|
319
|
+
log.info(
|
|
320
|
+
'Aztec node started in test mode (realProofs set to false) hence debug logs from public functions will be collected and served',
|
|
321
|
+
);
|
|
322
|
+
} else {
|
|
323
|
+
debugLogStore = new NullDebugLogStore();
|
|
316
324
|
}
|
|
325
|
+
|
|
317
326
|
const proofVerifier = new QueuedIVCVerifier(config, circuitVerifier);
|
|
318
327
|
|
|
328
|
+
const proverOnly = config.enableProverNode && config.disableValidator;
|
|
329
|
+
if (proverOnly) {
|
|
330
|
+
log.info('Starting in prover-only mode: skipping validator, sequencer, sentinel, and slasher subsystems');
|
|
331
|
+
}
|
|
332
|
+
|
|
319
333
|
// create the tx pool and the p2p client, which will need the l2 block source
|
|
320
334
|
const p2pClient = await createP2PClient(
|
|
321
|
-
P2PClientType.Full,
|
|
322
335
|
config,
|
|
323
336
|
archiver,
|
|
324
337
|
proofVerifier,
|
|
@@ -330,56 +343,69 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
330
343
|
deps.p2pClientDeps,
|
|
331
344
|
);
|
|
332
345
|
|
|
333
|
-
// We
|
|
334
|
-
|
|
346
|
+
// We'll accumulate sentinel watchers here
|
|
347
|
+
const watchers: Watcher[] = [];
|
|
335
348
|
|
|
336
|
-
|
|
337
|
-
|
|
349
|
+
// Create FullNodeCheckpointsBuilder for block proposal handling and tx validation.
|
|
350
|
+
// Override maxTxsPerCheckpoint with the validator-specific limit if set.
|
|
351
|
+
const validatorCheckpointsBuilder = new FullNodeCheckpointsBuilder(
|
|
352
|
+
{
|
|
353
|
+
...config,
|
|
354
|
+
l1GenesisTime,
|
|
355
|
+
slotDuration: Number(slotDuration),
|
|
356
|
+
rollupManaLimit,
|
|
357
|
+
maxTxsPerCheckpoint: config.validateMaxTxsPerCheckpoint,
|
|
358
|
+
},
|
|
338
359
|
worldStateSynchronizer,
|
|
339
360
|
archiver,
|
|
340
361
|
dateProvider,
|
|
341
362
|
telemetry,
|
|
342
363
|
);
|
|
343
364
|
|
|
344
|
-
|
|
345
|
-
const watchers: Watcher[] = [];
|
|
365
|
+
let validatorClient: ValidatorClient | undefined;
|
|
346
366
|
|
|
347
|
-
|
|
348
|
-
|
|
349
|
-
|
|
350
|
-
|
|
351
|
-
|
|
352
|
-
|
|
353
|
-
|
|
354
|
-
|
|
355
|
-
|
|
356
|
-
|
|
357
|
-
|
|
358
|
-
|
|
367
|
+
if (!proverOnly) {
|
|
368
|
+
// Create validator client if required
|
|
369
|
+
validatorClient = await createValidatorClient(config, {
|
|
370
|
+
checkpointsBuilder: validatorCheckpointsBuilder,
|
|
371
|
+
worldState: worldStateSynchronizer,
|
|
372
|
+
p2pClient,
|
|
373
|
+
telemetry,
|
|
374
|
+
dateProvider,
|
|
375
|
+
epochCache,
|
|
376
|
+
blockSource: archiver,
|
|
377
|
+
l1ToL2MessageSource: archiver,
|
|
378
|
+
keyStoreManager,
|
|
379
|
+
blobClient,
|
|
380
|
+
});
|
|
359
381
|
|
|
360
|
-
|
|
361
|
-
|
|
362
|
-
|
|
363
|
-
|
|
364
|
-
|
|
365
|
-
|
|
366
|
-
|
|
382
|
+
// If we have a validator client, register it as a source of offenses for the slasher,
|
|
383
|
+
// and have it register callbacks on the p2p client *before* we start it, otherwise messages
|
|
384
|
+
// like attestations or auths will fail.
|
|
385
|
+
if (validatorClient) {
|
|
386
|
+
watchers.push(validatorClient);
|
|
387
|
+
if (!options.dontStartSequencer) {
|
|
388
|
+
await validatorClient.registerHandlers();
|
|
389
|
+
}
|
|
367
390
|
}
|
|
368
391
|
}
|
|
369
392
|
|
|
370
|
-
// If there's no validator client
|
|
371
|
-
//
|
|
372
|
-
|
|
373
|
-
|
|
393
|
+
// If there's no validator client, create a BlockProposalHandler to handle block proposals
|
|
394
|
+
// for monitoring or reexecution. Reexecution (default) allows us to follow the pending chain,
|
|
395
|
+
// while non-reexecution is used for validating the proposals and collecting their txs.
|
|
396
|
+
if (!validatorClient) {
|
|
397
|
+
const reexecute = !!config.alwaysReexecuteBlockProposals;
|
|
398
|
+
log.info(`Setting up block proposal handler` + (reexecute ? ' with reexecution of proposals' : ''));
|
|
374
399
|
createBlockProposalHandler(config, {
|
|
375
|
-
|
|
400
|
+
checkpointsBuilder: validatorCheckpointsBuilder,
|
|
401
|
+
worldState: worldStateSynchronizer,
|
|
376
402
|
epochCache,
|
|
377
403
|
blockSource: archiver,
|
|
378
404
|
l1ToL2MessageSource: archiver,
|
|
379
405
|
p2pClient,
|
|
380
406
|
dateProvider,
|
|
381
407
|
telemetry,
|
|
382
|
-
}).
|
|
408
|
+
}).register(p2pClient, reexecute);
|
|
383
409
|
}
|
|
384
410
|
|
|
385
411
|
// Start world state and wait for it to sync to the archiver.
|
|
@@ -388,29 +414,33 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
388
414
|
// Start p2p. Note that it depends on world state to be running.
|
|
389
415
|
await p2pClient.start();
|
|
390
416
|
|
|
391
|
-
|
|
392
|
-
if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
|
|
393
|
-
watchers.push(validatorsSentinel);
|
|
394
|
-
}
|
|
395
|
-
|
|
417
|
+
let validatorsSentinel: Awaited<ReturnType<typeof createSentinel>> | undefined;
|
|
396
418
|
let epochPruneWatcher: EpochPruneWatcher | undefined;
|
|
397
|
-
if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
|
|
398
|
-
epochPruneWatcher = new EpochPruneWatcher(
|
|
399
|
-
archiver,
|
|
400
|
-
archiver,
|
|
401
|
-
epochCache,
|
|
402
|
-
p2pClient.getTxProvider(),
|
|
403
|
-
blockBuilder,
|
|
404
|
-
config,
|
|
405
|
-
);
|
|
406
|
-
watchers.push(epochPruneWatcher);
|
|
407
|
-
}
|
|
408
|
-
|
|
409
|
-
// We assume we want to slash for invalid attestations unless all max penalties are set to 0
|
|
410
419
|
let attestationsBlockWatcher: AttestationsBlockWatcher | undefined;
|
|
411
|
-
|
|
412
|
-
|
|
413
|
-
|
|
420
|
+
|
|
421
|
+
if (!proverOnly) {
|
|
422
|
+
validatorsSentinel = await createSentinel(epochCache, archiver, p2pClient, config);
|
|
423
|
+
if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
|
|
424
|
+
watchers.push(validatorsSentinel);
|
|
425
|
+
}
|
|
426
|
+
|
|
427
|
+
if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
|
|
428
|
+
epochPruneWatcher = new EpochPruneWatcher(
|
|
429
|
+
archiver,
|
|
430
|
+
archiver,
|
|
431
|
+
epochCache,
|
|
432
|
+
p2pClient.getTxProvider(),
|
|
433
|
+
validatorCheckpointsBuilder,
|
|
434
|
+
config,
|
|
435
|
+
);
|
|
436
|
+
watchers.push(epochPruneWatcher);
|
|
437
|
+
}
|
|
438
|
+
|
|
439
|
+
// We assume we want to slash for invalid attestations unless all max penalties are set to 0
|
|
440
|
+
if (config.slashProposeInvalidAttestationsPenalty > 0n || config.slashAttestDescendantOfInvalidPenalty > 0n) {
|
|
441
|
+
attestationsBlockWatcher = new AttestationsBlockWatcher(archiver, epochCache, config);
|
|
442
|
+
watchers.push(attestationsBlockWatcher);
|
|
443
|
+
}
|
|
414
444
|
}
|
|
415
445
|
|
|
416
446
|
// Start p2p-related services once the archiver has completed sync
|
|
@@ -447,27 +477,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
447
477
|
);
|
|
448
478
|
await slasherClient.start();
|
|
449
479
|
|
|
450
|
-
const l1TxUtils = config.
|
|
451
|
-
? await
|
|
480
|
+
const l1TxUtils = config.sequencerPublisherForwarderAddress
|
|
481
|
+
? await createForwarderL1TxUtilsFromSigners(
|
|
452
482
|
publicClient,
|
|
453
483
|
keyStoreManager!.createAllValidatorPublisherSigners(),
|
|
454
|
-
config.
|
|
484
|
+
config.sequencerPublisherForwarderAddress,
|
|
455
485
|
{ ...config, scope: 'sequencer' },
|
|
456
|
-
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
|
|
486
|
+
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
|
|
457
487
|
)
|
|
458
|
-
: await
|
|
488
|
+
: await createL1TxUtilsFromSigners(
|
|
459
489
|
publicClient,
|
|
460
490
|
keyStoreManager!.createAllValidatorPublisherSigners(),
|
|
461
491
|
{ ...config, scope: 'sequencer' },
|
|
462
|
-
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
|
|
492
|
+
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
|
|
463
493
|
);
|
|
464
494
|
|
|
465
495
|
// Create and start the sequencer client
|
|
466
496
|
const checkpointsBuilder = new CheckpointsBuilder(
|
|
467
|
-
{ ...config, l1GenesisTime, slotDuration: Number(slotDuration) },
|
|
497
|
+
{ ...config, l1GenesisTime, slotDuration: Number(slotDuration), rollupManaLimit },
|
|
498
|
+
worldStateSynchronizer,
|
|
468
499
|
archiver,
|
|
469
500
|
dateProvider,
|
|
470
501
|
telemetry,
|
|
502
|
+
debugLogStore,
|
|
471
503
|
);
|
|
472
504
|
|
|
473
505
|
sequencer = await SequencerClient.new(config, {
|
|
@@ -495,6 +527,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
495
527
|
log.warn(`Sequencer created but not started`);
|
|
496
528
|
}
|
|
497
529
|
|
|
530
|
+
// Create prover node subsystem if enabled
|
|
531
|
+
let proverNode: ProverNode | undefined;
|
|
532
|
+
if (config.enableProverNode) {
|
|
533
|
+
proverNode = await createProverNode(config, {
|
|
534
|
+
...deps.proverNodeDeps,
|
|
535
|
+
telemetry,
|
|
536
|
+
dateProvider,
|
|
537
|
+
archiver,
|
|
538
|
+
worldStateSynchronizer,
|
|
539
|
+
p2pClient,
|
|
540
|
+
epochCache,
|
|
541
|
+
blobClient,
|
|
542
|
+
keyStoreManager,
|
|
543
|
+
});
|
|
544
|
+
|
|
545
|
+
if (!options.dontStartProverNode) {
|
|
546
|
+
await proverNode.start();
|
|
547
|
+
log.info(`Prover node subsystem started`);
|
|
548
|
+
} else {
|
|
549
|
+
log.info(`Prover node subsystem created but not started`);
|
|
550
|
+
}
|
|
551
|
+
}
|
|
552
|
+
|
|
498
553
|
const globalVariableBuilder = new GlobalVariableBuilder({
|
|
499
554
|
...config,
|
|
500
555
|
rollupVersion: BigInt(config.rollupVersion),
|
|
@@ -502,7 +557,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
502
557
|
slotDuration: Number(slotDuration),
|
|
503
558
|
});
|
|
504
559
|
|
|
505
|
-
|
|
560
|
+
const node = new AztecNodeService(
|
|
506
561
|
config,
|
|
507
562
|
p2pClient,
|
|
508
563
|
archiver,
|
|
@@ -511,6 +566,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
511
566
|
archiver,
|
|
512
567
|
worldStateSynchronizer,
|
|
513
568
|
sequencer,
|
|
569
|
+
proverNode,
|
|
514
570
|
slasherClient,
|
|
515
571
|
validatorsSentinel,
|
|
516
572
|
epochPruneWatcher,
|
|
@@ -522,7 +578,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
522
578
|
proofVerifier,
|
|
523
579
|
telemetry,
|
|
524
580
|
log,
|
|
581
|
+
blobClient,
|
|
582
|
+
validatorClient,
|
|
583
|
+
keyStoreManager,
|
|
584
|
+
debugLogStore,
|
|
525
585
|
);
|
|
586
|
+
|
|
587
|
+
return node;
|
|
526
588
|
}
|
|
527
589
|
|
|
528
590
|
/**
|
|
@@ -533,6 +595,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
533
595
|
return this.sequencer;
|
|
534
596
|
}
|
|
535
597
|
|
|
598
|
+
/** Returns the prover node subsystem, if enabled. */
|
|
599
|
+
public getProverNode(): ProverNode | undefined {
|
|
600
|
+
return this.proverNode;
|
|
601
|
+
}
|
|
602
|
+
|
|
536
603
|
public getBlockSource(): L2BlockSource {
|
|
537
604
|
return this.blockSource;
|
|
538
605
|
}
|
|
@@ -558,7 +625,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
558
625
|
}
|
|
559
626
|
|
|
560
627
|
public async getAllowedPublicSetup(): Promise<AllowedElement[]> {
|
|
561
|
-
return this.config.
|
|
628
|
+
return [...(await getDefaultAllowedSetupFunctions()), ...(this.config.txPublicSetupAllowListExtend ?? [])];
|
|
562
629
|
}
|
|
563
630
|
|
|
564
631
|
/**
|
|
@@ -586,19 +653,26 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
586
653
|
enr,
|
|
587
654
|
l1ContractAddresses: contractAddresses,
|
|
588
655
|
protocolContractAddresses: protocolContractAddresses,
|
|
656
|
+
realProofs: !!this.config.realProofs,
|
|
589
657
|
};
|
|
590
658
|
|
|
591
659
|
return nodeInfo;
|
|
592
660
|
}
|
|
593
661
|
|
|
594
662
|
/**
|
|
595
|
-
* Get a block specified by its number.
|
|
596
|
-
* @param
|
|
663
|
+
* Get a block specified by its block number, block hash, or 'latest'.
|
|
664
|
+
* @param block - The block parameter (block number, block hash, or 'latest').
|
|
597
665
|
* @returns The requested block.
|
|
598
666
|
*/
|
|
599
|
-
public async getBlock(
|
|
600
|
-
|
|
601
|
-
|
|
667
|
+
public async getBlock(block: BlockParameter): Promise<L2Block | undefined> {
|
|
668
|
+
if (BlockHash.isBlockHash(block)) {
|
|
669
|
+
return this.getBlockByHash(block);
|
|
670
|
+
}
|
|
671
|
+
const blockNumber = block === 'latest' ? await this.getBlockNumber() : (block as BlockNumber);
|
|
672
|
+
if (blockNumber === BlockNumber.ZERO) {
|
|
673
|
+
return this.buildInitialBlock();
|
|
674
|
+
}
|
|
675
|
+
return await this.blockSource.getL2Block(blockNumber);
|
|
602
676
|
}
|
|
603
677
|
|
|
604
678
|
/**
|
|
@@ -606,9 +680,17 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
606
680
|
* @param blockHash - The block hash being requested.
|
|
607
681
|
* @returns The requested block.
|
|
608
682
|
*/
|
|
609
|
-
public async getBlockByHash(blockHash:
|
|
610
|
-
const
|
|
611
|
-
|
|
683
|
+
public async getBlockByHash(blockHash: BlockHash): Promise<L2Block | undefined> {
|
|
684
|
+
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
685
|
+
if (blockHash.equals(initialBlockHash)) {
|
|
686
|
+
return this.buildInitialBlock();
|
|
687
|
+
}
|
|
688
|
+
return await this.blockSource.getL2BlockByHash(blockHash);
|
|
689
|
+
}
|
|
690
|
+
|
|
691
|
+
private buildInitialBlock(): L2Block {
|
|
692
|
+
const initialHeader = this.worldStateSynchronizer.getCommitted().getInitialHeader();
|
|
693
|
+
return L2Block.empty(initialHeader);
|
|
612
694
|
}
|
|
613
695
|
|
|
614
696
|
/**
|
|
@@ -617,8 +699,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
617
699
|
* @returns The requested block.
|
|
618
700
|
*/
|
|
619
701
|
public async getBlockByArchive(archive: Fr): Promise<L2Block | undefined> {
|
|
620
|
-
|
|
621
|
-
return publishedBlock?.block;
|
|
702
|
+
return await this.blockSource.getL2BlockByArchive(archive);
|
|
622
703
|
}
|
|
623
704
|
|
|
624
705
|
/**
|
|
@@ -628,19 +709,23 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
628
709
|
* @returns The blocks requested.
|
|
629
710
|
*/
|
|
630
711
|
public async getBlocks(from: BlockNumber, limit: number): Promise<L2Block[]> {
|
|
631
|
-
return (await this.blockSource.getBlocks(from, limit)) ?? [];
|
|
712
|
+
return (await this.blockSource.getBlocks(from, BlockNumber(limit))) ?? [];
|
|
713
|
+
}
|
|
714
|
+
|
|
715
|
+
public async getCheckpoints(from: CheckpointNumber, limit: number): Promise<PublishedCheckpoint[]> {
|
|
716
|
+
return (await this.blockSource.getCheckpoints(from, limit)) ?? [];
|
|
632
717
|
}
|
|
633
718
|
|
|
634
|
-
public async
|
|
635
|
-
return (await this.blockSource.
|
|
719
|
+
public async getCheckpointedBlocks(from: BlockNumber, limit: number) {
|
|
720
|
+
return (await this.blockSource.getCheckpointedBlocks(from, limit)) ?? [];
|
|
636
721
|
}
|
|
637
722
|
|
|
638
723
|
/**
|
|
639
|
-
* Method to fetch the current
|
|
640
|
-
* @returns The current
|
|
724
|
+
* Method to fetch the current min L2 fees.
|
|
725
|
+
* @returns The current min L2 fees.
|
|
641
726
|
*/
|
|
642
|
-
public async
|
|
643
|
-
return await this.globalVariableBuilder.
|
|
727
|
+
public async getCurrentMinFees(): Promise<GasFees> {
|
|
728
|
+
return await this.globalVariableBuilder.getCurrentMinFees();
|
|
644
729
|
}
|
|
645
730
|
|
|
646
731
|
public async getMaxPriorityFees(): Promise<GasFees> {
|
|
@@ -663,6 +748,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
663
748
|
return await this.blockSource.getProvenBlockNumber();
|
|
664
749
|
}
|
|
665
750
|
|
|
751
|
+
public async getCheckpointedBlockNumber(): Promise<BlockNumber> {
|
|
752
|
+
return await this.blockSource.getCheckpointedL2BlockNumber();
|
|
753
|
+
}
|
|
754
|
+
|
|
755
|
+
public getCheckpointNumber(): Promise<CheckpointNumber> {
|
|
756
|
+
return this.blockSource.getCheckpointNumber();
|
|
757
|
+
}
|
|
758
|
+
|
|
666
759
|
/**
|
|
667
760
|
* Method to fetch the version of the package.
|
|
668
761
|
* @returns The node package version
|
|
@@ -695,12 +788,43 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
695
788
|
return this.contractDataSource.getContract(address);
|
|
696
789
|
}
|
|
697
790
|
|
|
698
|
-
public getPrivateLogsByTags(
|
|
699
|
-
|
|
700
|
-
|
|
701
|
-
|
|
702
|
-
|
|
703
|
-
|
|
791
|
+
public async getPrivateLogsByTags(
|
|
792
|
+
tags: SiloedTag[],
|
|
793
|
+
page?: number,
|
|
794
|
+
referenceBlock?: BlockHash,
|
|
795
|
+
): Promise<TxScopedL2Log[][]> {
|
|
796
|
+
if (referenceBlock) {
|
|
797
|
+
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
798
|
+
if (!referenceBlock.equals(initialBlockHash)) {
|
|
799
|
+
const header = await this.blockSource.getBlockHeaderByHash(referenceBlock);
|
|
800
|
+
if (!header) {
|
|
801
|
+
throw new Error(
|
|
802
|
+
`Block ${referenceBlock.toString()} not found in the node. This might indicate a reorg has occurred.`,
|
|
803
|
+
);
|
|
804
|
+
}
|
|
805
|
+
}
|
|
806
|
+
}
|
|
807
|
+
return this.logsSource.getPrivateLogsByTags(tags, page);
|
|
808
|
+
}
|
|
809
|
+
|
|
810
|
+
public async getPublicLogsByTagsFromContract(
|
|
811
|
+
contractAddress: AztecAddress,
|
|
812
|
+
tags: Tag[],
|
|
813
|
+
page?: number,
|
|
814
|
+
referenceBlock?: BlockHash,
|
|
815
|
+
): Promise<TxScopedL2Log[][]> {
|
|
816
|
+
if (referenceBlock) {
|
|
817
|
+
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
818
|
+
if (!referenceBlock.equals(initialBlockHash)) {
|
|
819
|
+
const header = await this.blockSource.getBlockHeaderByHash(referenceBlock);
|
|
820
|
+
if (!header) {
|
|
821
|
+
throw new Error(
|
|
822
|
+
`Block ${referenceBlock.toString()} not found in the node. This might indicate a reorg has occurred.`,
|
|
823
|
+
);
|
|
824
|
+
}
|
|
825
|
+
}
|
|
826
|
+
}
|
|
827
|
+
return this.logsSource.getPublicLogsByTagsFromContract(contractAddress, tags, page);
|
|
704
828
|
}
|
|
705
829
|
|
|
706
830
|
/**
|
|
@@ -742,26 +866,36 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
742
866
|
}
|
|
743
867
|
|
|
744
868
|
await this.p2pClient!.sendTx(tx);
|
|
745
|
-
|
|
746
|
-
this.
|
|
869
|
+
const duration = timer.ms();
|
|
870
|
+
this.metrics.receivedTx(duration, true);
|
|
871
|
+
this.log.info(`Received tx ${txHash} in ${duration}ms`, { txHash });
|
|
747
872
|
}
|
|
748
873
|
|
|
749
874
|
public async getTxReceipt(txHash: TxHash): Promise<TxReceipt> {
|
|
750
|
-
|
|
751
|
-
|
|
752
|
-
|
|
753
|
-
|
|
754
|
-
// and we would incorrectly return a TxReceipt with status DROPPED
|
|
755
|
-
if ((await this.p2pClient.getTxStatus(txHash)) === 'pending') {
|
|
756
|
-
txReceipt = new TxReceipt(txHash, TxStatus.PENDING, '');
|
|
757
|
-
}
|
|
875
|
+
// Check the tx pool status first. If the tx is known to the pool (pending or mined), we'll use that
|
|
876
|
+
// as a fallback if we don't find a settled receipt in the archiver.
|
|
877
|
+
const txPoolStatus = await this.p2pClient.getTxStatus(txHash);
|
|
878
|
+
const isKnownToPool = txPoolStatus === 'pending' || txPoolStatus === 'mined';
|
|
758
879
|
|
|
880
|
+
// Then get the actual tx from the archiver, which tracks every tx in a mined block.
|
|
759
881
|
const settledTxReceipt = await this.blockSource.getSettledTxReceipt(txHash);
|
|
882
|
+
|
|
883
|
+
let receipt: TxReceipt;
|
|
760
884
|
if (settledTxReceipt) {
|
|
761
|
-
|
|
885
|
+
receipt = settledTxReceipt;
|
|
886
|
+
} else if (isKnownToPool) {
|
|
887
|
+
// If the tx is in the pool but not in the archiver, it's pending.
|
|
888
|
+
// This handles race conditions between archiver and p2p, where the archiver
|
|
889
|
+
// has pruned the block in which a tx was mined, but p2p has not caught up yet.
|
|
890
|
+
receipt = new TxReceipt(txHash, TxStatus.PENDING, undefined, undefined);
|
|
891
|
+
} else {
|
|
892
|
+
// Otherwise, if we don't know the tx, we consider it dropped.
|
|
893
|
+
receipt = new TxReceipt(txHash, TxStatus.DROPPED, undefined, 'Tx dropped by P2P node');
|
|
762
894
|
}
|
|
763
895
|
|
|
764
|
-
|
|
896
|
+
this.debugLogStore.decorateReceiptWithLogs(txHash.toString(), receipt);
|
|
897
|
+
|
|
898
|
+
return receipt;
|
|
765
899
|
}
|
|
766
900
|
|
|
767
901
|
public getTxEffect(txHash: TxHash): Promise<IndexedTxEffect | undefined> {
|
|
@@ -778,13 +912,23 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
778
912
|
await tryStop(this.slasherClient);
|
|
779
913
|
await tryStop(this.proofVerifier);
|
|
780
914
|
await tryStop(this.sequencer);
|
|
915
|
+
await tryStop(this.proverNode);
|
|
781
916
|
await tryStop(this.p2pClient);
|
|
782
917
|
await tryStop(this.worldStateSynchronizer);
|
|
783
918
|
await tryStop(this.blockSource);
|
|
919
|
+
await tryStop(this.blobClient);
|
|
784
920
|
await tryStop(this.telemetry);
|
|
785
921
|
this.log.info(`Stopped Aztec Node`);
|
|
786
922
|
}
|
|
787
923
|
|
|
924
|
+
/**
|
|
925
|
+
* Returns the blob client used by this node.
|
|
926
|
+
* @internal - Exposed for testing purposes only.
|
|
927
|
+
*/
|
|
928
|
+
public getBlobClient(): BlobClientInterface | undefined {
|
|
929
|
+
return this.blobClient;
|
|
930
|
+
}
|
|
931
|
+
|
|
788
932
|
/**
|
|
789
933
|
* Method to retrieve pending txs.
|
|
790
934
|
* @param limit - The number of items to returns
|
|
@@ -817,20 +961,12 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
817
961
|
return compactArray(await Promise.all(txHashes.map(txHash => this.getTxByHash(txHash))));
|
|
818
962
|
}
|
|
819
963
|
|
|
820
|
-
/**
|
|
821
|
-
* Find the indexes of the given leaves in the given tree along with a block metadata pointing to the block in which
|
|
822
|
-
* the leaves were inserted.
|
|
823
|
-
* @param blockNumber - The block number at which to get the data or 'latest' for latest data.
|
|
824
|
-
* @param treeId - The tree to search in.
|
|
825
|
-
* @param leafValues - The values to search for.
|
|
826
|
-
* @returns The indices of leaves and the block metadata of a block in which the leaves were inserted.
|
|
827
|
-
*/
|
|
828
964
|
public async findLeavesIndexes(
|
|
829
|
-
|
|
965
|
+
referenceBlock: BlockParameter,
|
|
830
966
|
treeId: MerkleTreeId,
|
|
831
967
|
leafValues: Fr[],
|
|
832
968
|
): Promise<(DataInBlock<bigint> | undefined)[]> {
|
|
833
|
-
const committedDb = await this.#getWorldState(
|
|
969
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
834
970
|
const maybeIndices = await committedDb.findLeafIndices(
|
|
835
971
|
treeId,
|
|
836
972
|
leafValues.map(x => x.toBuffer()),
|
|
@@ -882,56 +1018,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
882
1018
|
}
|
|
883
1019
|
return {
|
|
884
1020
|
l2BlockNumber: BlockNumber(Number(blockNumber)),
|
|
885
|
-
l2BlockHash:
|
|
1021
|
+
l2BlockHash: new BlockHash(blockHash),
|
|
886
1022
|
data: index,
|
|
887
1023
|
};
|
|
888
1024
|
});
|
|
889
1025
|
}
|
|
890
1026
|
|
|
891
|
-
|
|
892
|
-
|
|
893
|
-
|
|
894
|
-
* @param leafIndex - The index of the leaf for which the sibling path is required.
|
|
895
|
-
* @returns The sibling path for the leaf index.
|
|
896
|
-
*/
|
|
897
|
-
public async getNullifierSiblingPath(
|
|
898
|
-
blockNumber: BlockParameter,
|
|
899
|
-
leafIndex: bigint,
|
|
900
|
-
): Promise<SiblingPath<typeof NULLIFIER_TREE_HEIGHT>> {
|
|
901
|
-
const committedDb = await this.#getWorldState(blockNumber);
|
|
902
|
-
return committedDb.getSiblingPath(MerkleTreeId.NULLIFIER_TREE, leafIndex);
|
|
903
|
-
}
|
|
904
|
-
|
|
905
|
-
/**
|
|
906
|
-
* Returns a sibling path for the given index in the data tree.
|
|
907
|
-
* @param blockNumber - The block number at which to get the data.
|
|
908
|
-
* @param leafIndex - The index of the leaf for which the sibling path is required.
|
|
909
|
-
* @returns The sibling path for the leaf index.
|
|
910
|
-
*/
|
|
911
|
-
public async getNoteHashSiblingPath(
|
|
912
|
-
blockNumber: BlockParameter,
|
|
913
|
-
leafIndex: bigint,
|
|
914
|
-
): Promise<SiblingPath<typeof NOTE_HASH_TREE_HEIGHT>> {
|
|
915
|
-
const committedDb = await this.#getWorldState(blockNumber);
|
|
916
|
-
return committedDb.getSiblingPath(MerkleTreeId.NOTE_HASH_TREE, leafIndex);
|
|
917
|
-
}
|
|
918
|
-
|
|
919
|
-
public async getArchiveMembershipWitness(
|
|
920
|
-
blockNumber: BlockParameter,
|
|
921
|
-
archive: Fr,
|
|
1027
|
+
public async getBlockHashMembershipWitness(
|
|
1028
|
+
referenceBlock: BlockParameter,
|
|
1029
|
+
blockHash: BlockHash,
|
|
922
1030
|
): Promise<MembershipWitness<typeof ARCHIVE_HEIGHT> | undefined> {
|
|
923
|
-
const committedDb = await this.#getWorldState(
|
|
924
|
-
const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.ARCHIVE>(MerkleTreeId.ARCHIVE, [
|
|
1031
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1032
|
+
const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.ARCHIVE>(MerkleTreeId.ARCHIVE, [blockHash]);
|
|
925
1033
|
return pathAndIndex === undefined
|
|
926
1034
|
? undefined
|
|
927
1035
|
: MembershipWitness.fromSiblingPath(pathAndIndex.index, pathAndIndex.path);
|
|
928
1036
|
}
|
|
929
1037
|
|
|
930
1038
|
public async getNoteHashMembershipWitness(
|
|
931
|
-
|
|
1039
|
+
referenceBlock: BlockParameter,
|
|
932
1040
|
noteHash: Fr,
|
|
933
1041
|
): Promise<MembershipWitness<typeof NOTE_HASH_TREE_HEIGHT> | undefined> {
|
|
934
|
-
const committedDb = await this.#getWorldState(
|
|
1042
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
935
1043
|
const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.NOTE_HASH_TREE>(
|
|
936
1044
|
MerkleTreeId.NOTE_HASH_TREE,
|
|
937
1045
|
[noteHash],
|
|
@@ -941,17 +1049,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
941
1049
|
: MembershipWitness.fromSiblingPath(pathAndIndex.index, pathAndIndex.path);
|
|
942
1050
|
}
|
|
943
1051
|
|
|
944
|
-
/**
|
|
945
|
-
* Returns the index and a sibling path for a leaf in the committed l1 to l2 data tree.
|
|
946
|
-
* @param blockNumber - The block number at which to get the data.
|
|
947
|
-
* @param l1ToL2Message - The l1ToL2Message to get the index / sibling path for.
|
|
948
|
-
* @returns A tuple of the index and the sibling path of the L1ToL2Message (undefined if not found).
|
|
949
|
-
*/
|
|
950
1052
|
public async getL1ToL2MessageMembershipWitness(
|
|
951
|
-
|
|
1053
|
+
referenceBlock: BlockParameter,
|
|
952
1054
|
l1ToL2Message: Fr,
|
|
953
1055
|
): Promise<[bigint, SiblingPath<typeof L1_TO_L2_MSG_TREE_HEIGHT>] | undefined> {
|
|
954
|
-
const db = await this.#getWorldState(
|
|
1056
|
+
const db = await this.#getWorldState(referenceBlock);
|
|
955
1057
|
const [witness] = await db.findSiblingPaths(MerkleTreeId.L1_TO_L2_MESSAGE_TREE, [l1ToL2Message]);
|
|
956
1058
|
if (!witness) {
|
|
957
1059
|
return undefined;
|
|
@@ -961,11 +1063,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
961
1063
|
return [witness.index, witness.path];
|
|
962
1064
|
}
|
|
963
1065
|
|
|
964
|
-
public async
|
|
1066
|
+
public async getL1ToL2MessageCheckpoint(l1ToL2Message: Fr): Promise<CheckpointNumber | undefined> {
|
|
965
1067
|
const messageIndex = await this.l1ToL2MessageSource.getL1ToL2MessageIndex(l1ToL2Message);
|
|
966
|
-
return messageIndex
|
|
967
|
-
? BlockNumber.fromCheckpointNumber(InboxLeaf.checkpointNumberFromIndex(messageIndex))
|
|
968
|
-
: undefined;
|
|
1068
|
+
return messageIndex ? InboxLeaf.checkpointNumberFromIndex(messageIndex) : undefined;
|
|
969
1069
|
}
|
|
970
1070
|
|
|
971
1071
|
/**
|
|
@@ -979,56 +1079,36 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
979
1079
|
}
|
|
980
1080
|
|
|
981
1081
|
/**
|
|
982
|
-
* Returns all the L2 to L1 messages in
|
|
983
|
-
* @param
|
|
984
|
-
* @returns The L2 to L1 messages (
|
|
1082
|
+
* Returns all the L2 to L1 messages in an epoch.
|
|
1083
|
+
* @param epoch - The epoch at which to get the data.
|
|
1084
|
+
* @returns The L2 to L1 messages (empty array if the epoch is not found).
|
|
985
1085
|
*/
|
|
986
|
-
public async getL2ToL1Messages(
|
|
987
|
-
|
|
988
|
-
|
|
1086
|
+
public async getL2ToL1Messages(epoch: EpochNumber): Promise<Fr[][][][]> {
|
|
1087
|
+
// Assumes `getCheckpointedBlocksForEpoch` returns blocks in ascending order of block number.
|
|
1088
|
+
const checkpointedBlocks = await this.blockSource.getCheckpointedBlocksForEpoch(epoch);
|
|
1089
|
+
const blocksInCheckpoints: L2Block[][] = [];
|
|
1090
|
+
let previousSlotNumber = SlotNumber.ZERO;
|
|
1091
|
+
let checkpointIndex = -1;
|
|
1092
|
+
for (const checkpointedBlock of checkpointedBlocks) {
|
|
1093
|
+
const block = checkpointedBlock.block;
|
|
1094
|
+
const slotNumber = block.header.globalVariables.slotNumber;
|
|
1095
|
+
if (slotNumber !== previousSlotNumber) {
|
|
1096
|
+
checkpointIndex++;
|
|
1097
|
+
blocksInCheckpoints.push([]);
|
|
1098
|
+
previousSlotNumber = slotNumber;
|
|
1099
|
+
}
|
|
1100
|
+
blocksInCheckpoints[checkpointIndex].push(block);
|
|
1101
|
+
}
|
|
1102
|
+
return blocksInCheckpoints.map(blocks =>
|
|
1103
|
+
blocks.map(block => block.body.txEffects.map(txEffect => txEffect.l2ToL1Msgs)),
|
|
989
1104
|
);
|
|
990
|
-
return block?.body.txEffects.map(txEffect => txEffect.l2ToL1Msgs);
|
|
991
|
-
}
|
|
992
|
-
|
|
993
|
-
/**
|
|
994
|
-
* Returns a sibling path for a leaf in the committed blocks tree.
|
|
995
|
-
* @param blockNumber - The block number at which to get the data.
|
|
996
|
-
* @param leafIndex - Index of the leaf in the tree.
|
|
997
|
-
* @returns The sibling path.
|
|
998
|
-
*/
|
|
999
|
-
public async getArchiveSiblingPath(
|
|
1000
|
-
blockNumber: BlockParameter,
|
|
1001
|
-
leafIndex: bigint,
|
|
1002
|
-
): Promise<SiblingPath<typeof ARCHIVE_HEIGHT>> {
|
|
1003
|
-
const committedDb = await this.#getWorldState(blockNumber);
|
|
1004
|
-
return committedDb.getSiblingPath(MerkleTreeId.ARCHIVE, leafIndex);
|
|
1005
|
-
}
|
|
1006
|
-
|
|
1007
|
-
/**
|
|
1008
|
-
* Returns a sibling path for a leaf in the committed public data tree.
|
|
1009
|
-
* @param blockNumber - The block number at which to get the data.
|
|
1010
|
-
* @param leafIndex - Index of the leaf in the tree.
|
|
1011
|
-
* @returns The sibling path.
|
|
1012
|
-
*/
|
|
1013
|
-
public async getPublicDataSiblingPath(
|
|
1014
|
-
blockNumber: BlockParameter,
|
|
1015
|
-
leafIndex: bigint,
|
|
1016
|
-
): Promise<SiblingPath<typeof PUBLIC_DATA_TREE_HEIGHT>> {
|
|
1017
|
-
const committedDb = await this.#getWorldState(blockNumber);
|
|
1018
|
-
return committedDb.getSiblingPath(MerkleTreeId.PUBLIC_DATA_TREE, leafIndex);
|
|
1019
1105
|
}
|
|
1020
1106
|
|
|
1021
|
-
/**
|
|
1022
|
-
* Returns a nullifier membership witness for a given nullifier at a given block.
|
|
1023
|
-
* @param blockNumber - The block number at which to get the index.
|
|
1024
|
-
* @param nullifier - Nullifier we try to find witness for.
|
|
1025
|
-
* @returns The nullifier membership witness (if found).
|
|
1026
|
-
*/
|
|
1027
1107
|
public async getNullifierMembershipWitness(
|
|
1028
|
-
|
|
1108
|
+
referenceBlock: BlockParameter,
|
|
1029
1109
|
nullifier: Fr,
|
|
1030
1110
|
): Promise<NullifierMembershipWitness | undefined> {
|
|
1031
|
-
const db = await this.#getWorldState(
|
|
1111
|
+
const db = await this.#getWorldState(referenceBlock);
|
|
1032
1112
|
const [witness] = await db.findSiblingPaths(MerkleTreeId.NULLIFIER_TREE, [nullifier.toBuffer()]);
|
|
1033
1113
|
if (!witness) {
|
|
1034
1114
|
return undefined;
|
|
@@ -1045,7 +1125,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1045
1125
|
|
|
1046
1126
|
/**
|
|
1047
1127
|
* Returns a low nullifier membership witness for a given nullifier at a given block.
|
|
1048
|
-
* @param
|
|
1128
|
+
* @param referenceBlock - The block parameter (block number, block hash, or 'latest') at which to get the data
|
|
1129
|
+
* (which contains the root of the nullifier tree in which we are searching for the nullifier).
|
|
1049
1130
|
* @param nullifier - Nullifier we try to find the low nullifier witness for.
|
|
1050
1131
|
* @returns The low nullifier membership witness (if found).
|
|
1051
1132
|
* @remarks Low nullifier witness can be used to perform a nullifier non-inclusion proof by leveraging the "linked
|
|
@@ -1058,10 +1139,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1058
1139
|
* TODO: This is a confusing behavior and we should eventually address that.
|
|
1059
1140
|
*/
|
|
1060
1141
|
public async getLowNullifierMembershipWitness(
|
|
1061
|
-
|
|
1142
|
+
referenceBlock: BlockParameter,
|
|
1062
1143
|
nullifier: Fr,
|
|
1063
1144
|
): Promise<NullifierMembershipWitness | undefined> {
|
|
1064
|
-
const committedDb = await this.#getWorldState(
|
|
1145
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1065
1146
|
const findResult = await committedDb.getPreviousValueIndex(MerkleTreeId.NULLIFIER_TREE, nullifier.toBigInt());
|
|
1066
1147
|
if (!findResult) {
|
|
1067
1148
|
return undefined;
|
|
@@ -1076,8 +1157,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1076
1157
|
return new NullifierMembershipWitness(BigInt(index), preimageData as NullifierLeafPreimage, siblingPath);
|
|
1077
1158
|
}
|
|
1078
1159
|
|
|
1079
|
-
async getPublicDataWitness(
|
|
1080
|
-
const committedDb = await this.#getWorldState(
|
|
1160
|
+
async getPublicDataWitness(referenceBlock: BlockParameter, leafSlot: Fr): Promise<PublicDataWitness | undefined> {
|
|
1161
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1081
1162
|
const lowLeafResult = await committedDb.getPreviousValueIndex(MerkleTreeId.PUBLIC_DATA_TREE, leafSlot.toBigInt());
|
|
1082
1163
|
if (!lowLeafResult) {
|
|
1083
1164
|
return undefined;
|
|
@@ -1091,19 +1172,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1091
1172
|
}
|
|
1092
1173
|
}
|
|
1093
1174
|
|
|
1094
|
-
|
|
1095
|
-
|
|
1096
|
-
*
|
|
1097
|
-
* @remarks The storage slot here refers to the slot as it is defined in Noir not the index in the merkle tree.
|
|
1098
|
-
* Aztec's version of `eth_getStorageAt`.
|
|
1099
|
-
*
|
|
1100
|
-
* @param contract - Address of the contract to query.
|
|
1101
|
-
* @param slot - Slot to query.
|
|
1102
|
-
* @param blockNumber - The block number at which to get the data or 'latest'.
|
|
1103
|
-
* @returns Storage value at the given contract slot.
|
|
1104
|
-
*/
|
|
1105
|
-
public async getPublicStorageAt(blockNumber: BlockParameter, contract: AztecAddress, slot: Fr): Promise<Fr> {
|
|
1106
|
-
const committedDb = await this.#getWorldState(blockNumber);
|
|
1175
|
+
public async getPublicStorageAt(referenceBlock: BlockParameter, contract: AztecAddress, slot: Fr): Promise<Fr> {
|
|
1176
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1107
1177
|
const leafSlot = await computePublicDataTreeLeafSlot(contract, slot);
|
|
1108
1178
|
|
|
1109
1179
|
const lowLeafResult = await committedDb.getPreviousValueIndex(MerkleTreeId.PUBLIC_DATA_TREE, leafSlot.toBigInt());
|
|
@@ -1117,24 +1187,22 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1117
1187
|
return preimage.leaf.value;
|
|
1118
1188
|
}
|
|
1119
1189
|
|
|
1120
|
-
|
|
1121
|
-
|
|
1122
|
-
|
|
1123
|
-
|
|
1124
|
-
|
|
1125
|
-
|
|
1126
|
-
|
|
1127
|
-
|
|
1128
|
-
|
|
1129
|
-
|
|
1130
|
-
|
|
1131
|
-
|
|
1132
|
-
|
|
1133
|
-
|
|
1134
|
-
|
|
1135
|
-
|
|
1136
|
-
public async getBlockHeaderByHash(blockHash: Fr): Promise<BlockHeader | undefined> {
|
|
1137
|
-
return await this.blockSource.getBlockHeaderByHash(blockHash);
|
|
1190
|
+
public async getBlockHeader(block: BlockParameter = 'latest'): Promise<BlockHeader | undefined> {
|
|
1191
|
+
if (BlockHash.isBlockHash(block)) {
|
|
1192
|
+
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
1193
|
+
if (block.equals(initialBlockHash)) {
|
|
1194
|
+
// Block source doesn't handle initial header so we need to handle the case separately.
|
|
1195
|
+
return this.worldStateSynchronizer.getCommitted().getInitialHeader();
|
|
1196
|
+
}
|
|
1197
|
+
return this.blockSource.getBlockHeaderByHash(block);
|
|
1198
|
+
} else {
|
|
1199
|
+
// Block source doesn't handle initial header so we need to handle the case separately.
|
|
1200
|
+
const blockNumber = block === 'latest' ? await this.getBlockNumber() : (block as BlockNumber);
|
|
1201
|
+
if (blockNumber === BlockNumber.ZERO) {
|
|
1202
|
+
return this.worldStateSynchronizer.getCommitted().getInitialHeader();
|
|
1203
|
+
}
|
|
1204
|
+
return this.blockSource.getBlockHeader(block);
|
|
1205
|
+
}
|
|
1138
1206
|
}
|
|
1139
1207
|
|
|
1140
1208
|
/**
|
|
@@ -1146,6 +1214,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1146
1214
|
return await this.blockSource.getBlockHeaderByArchive(archive);
|
|
1147
1215
|
}
|
|
1148
1216
|
|
|
1217
|
+
public getBlockData(number: BlockNumber): Promise<BlockData | undefined> {
|
|
1218
|
+
return this.blockSource.getBlockData(number);
|
|
1219
|
+
}
|
|
1220
|
+
|
|
1221
|
+
public getBlockDataByArchive(archive: Fr): Promise<BlockData | undefined> {
|
|
1222
|
+
return this.blockSource.getBlockDataByArchive(archive);
|
|
1223
|
+
}
|
|
1224
|
+
|
|
1149
1225
|
/**
|
|
1150
1226
|
* Simulates the public part of a transaction with the current state.
|
|
1151
1227
|
* @param tx - The transaction to simulate.
|
|
@@ -1169,7 +1245,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1169
1245
|
}
|
|
1170
1246
|
|
|
1171
1247
|
const txHash = tx.getTxHash();
|
|
1172
|
-
const
|
|
1248
|
+
const latestBlockNumber = await this.blockSource.getBlockNumber();
|
|
1249
|
+
const blockNumber = BlockNumber.add(latestBlockNumber, 1);
|
|
1173
1250
|
|
|
1174
1251
|
// If sequencer is not initialized, we just set these values to zero for simulation.
|
|
1175
1252
|
const coinbase = EthAddress.ZERO;
|
|
@@ -1184,6 +1261,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1184
1261
|
this.contractDataSource,
|
|
1185
1262
|
new DateProvider(),
|
|
1186
1263
|
this.telemetry,
|
|
1264
|
+
this.log.getBindings(),
|
|
1187
1265
|
);
|
|
1188
1266
|
|
|
1189
1267
|
this.log.verbose(`Simulating public calls for tx ${txHash}`, {
|
|
@@ -1192,6 +1270,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1192
1270
|
blockNumber,
|
|
1193
1271
|
});
|
|
1194
1272
|
|
|
1273
|
+
// Ensure world-state has caught up with the latest block we loaded from the archiver
|
|
1274
|
+
await this.worldStateSynchronizer.syncImmediate(latestBlockNumber);
|
|
1195
1275
|
const merkleTreeFork = await this.worldStateSynchronizer.fork();
|
|
1196
1276
|
try {
|
|
1197
1277
|
const config = PublicSimulatorConfig.from({
|
|
@@ -1207,7 +1287,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1207
1287
|
const processor = publicProcessorFactory.create(merkleTreeFork, newGlobalVariables, config);
|
|
1208
1288
|
|
|
1209
1289
|
// REFACTOR: Consider merging ProcessReturnValues into ProcessedTx
|
|
1210
|
-
const [processedTxs, failedTxs, _usedTxs, returns] = await processor.process([tx]);
|
|
1290
|
+
const [processedTxs, failedTxs, _usedTxs, returns, debugLogs] = await processor.process([tx]);
|
|
1211
1291
|
// REFACTOR: Consider returning the error rather than throwing
|
|
1212
1292
|
if (failedTxs.length) {
|
|
1213
1293
|
this.log.warn(`Simulated tx ${txHash} fails: ${failedTxs[0].error}`, { txHash });
|
|
@@ -1221,6 +1301,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1221
1301
|
processedTx.txEffect,
|
|
1222
1302
|
returns,
|
|
1223
1303
|
processedTx.gasUsed,
|
|
1304
|
+
debugLogs,
|
|
1224
1305
|
);
|
|
1225
1306
|
} finally {
|
|
1226
1307
|
await merkleTreeFork.close();
|
|
@@ -1234,19 +1315,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1234
1315
|
const db = this.worldStateSynchronizer.getCommitted();
|
|
1235
1316
|
const verifier = isSimulation ? undefined : this.proofVerifier;
|
|
1236
1317
|
|
|
1237
|
-
// We accept transactions if they are not expired by the next slot (checked based on the
|
|
1318
|
+
// We accept transactions if they are not expired by the next slot (checked based on the ExpirationTimestamp field)
|
|
1238
1319
|
const { ts: nextSlotTimestamp } = this.epochCache.getEpochAndSlotInNextL1Slot();
|
|
1239
1320
|
const blockNumber = BlockNumber((await this.blockSource.getBlockNumber()) + 1);
|
|
1240
|
-
const validator =
|
|
1241
|
-
|
|
1242
|
-
|
|
1243
|
-
|
|
1244
|
-
|
|
1245
|
-
|
|
1246
|
-
|
|
1247
|
-
|
|
1248
|
-
|
|
1249
|
-
|
|
1321
|
+
const validator = createTxValidatorForAcceptingTxsOverRPC(
|
|
1322
|
+
db,
|
|
1323
|
+
this.contractDataSource,
|
|
1324
|
+
verifier,
|
|
1325
|
+
{
|
|
1326
|
+
timestamp: nextSlotTimestamp,
|
|
1327
|
+
blockNumber,
|
|
1328
|
+
l1ChainId: this.l1ChainId,
|
|
1329
|
+
rollupVersion: this.version,
|
|
1330
|
+
setupAllowList: [
|
|
1331
|
+
...(await getDefaultAllowedSetupFunctions()),
|
|
1332
|
+
...(this.config.txPublicSetupAllowListExtend ?? []),
|
|
1333
|
+
],
|
|
1334
|
+
gasFees: await this.getCurrentMinFees(),
|
|
1335
|
+
skipFeeEnforcement,
|
|
1336
|
+
txsPermitted: !this.config.disableTransactions,
|
|
1337
|
+
},
|
|
1338
|
+
this.log.getBindings(),
|
|
1339
|
+
);
|
|
1250
1340
|
|
|
1251
1341
|
return await validator.validateTx(tx);
|
|
1252
1342
|
}
|
|
@@ -1315,7 +1405,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1315
1405
|
}
|
|
1316
1406
|
|
|
1317
1407
|
// And it has an L2 block hash
|
|
1318
|
-
const l2BlockHash = await archiver.getL2Tips().then(tips => tips.
|
|
1408
|
+
const l2BlockHash = await archiver.getL2Tips().then(tips => tips.proposed.hash);
|
|
1319
1409
|
if (!l2BlockHash) {
|
|
1320
1410
|
this.metrics.recordSnapshotError();
|
|
1321
1411
|
throw new Error(`Archiver has no latest L2 block hash downloaded. Cannot start snapshot.`);
|
|
@@ -1349,7 +1439,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1349
1439
|
throw new Error('Archiver implementation does not support rollbacks.');
|
|
1350
1440
|
}
|
|
1351
1441
|
|
|
1352
|
-
const finalizedBlock = await archiver.getL2Tips().then(tips => tips.finalized.number);
|
|
1442
|
+
const finalizedBlock = await archiver.getL2Tips().then(tips => tips.finalized.block.number);
|
|
1353
1443
|
if (targetBlock < finalizedBlock) {
|
|
1354
1444
|
if (force) {
|
|
1355
1445
|
this.log.warn(`Clearing world state database to allow rolling back behind finalized block ${finalizedBlock}`);
|
|
@@ -1410,16 +1500,107 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1410
1500
|
}
|
|
1411
1501
|
}
|
|
1412
1502
|
|
|
1503
|
+
public async reloadKeystore(): Promise<void> {
|
|
1504
|
+
if (!this.config.keyStoreDirectory?.length) {
|
|
1505
|
+
throw new BadRequestError(
|
|
1506
|
+
'Cannot reload keystore: node is not using a file-based keystore. ' +
|
|
1507
|
+
'Set KEY_STORE_DIRECTORY to use file-based keystores.',
|
|
1508
|
+
);
|
|
1509
|
+
}
|
|
1510
|
+
if (!this.validatorClient) {
|
|
1511
|
+
throw new BadRequestError('Cannot reload keystore: validator is not enabled.');
|
|
1512
|
+
}
|
|
1513
|
+
|
|
1514
|
+
this.log.info('Reloading keystore from disk');
|
|
1515
|
+
|
|
1516
|
+
// Re-read and validate keystore files
|
|
1517
|
+
const keyStores = loadKeystores(this.config.keyStoreDirectory);
|
|
1518
|
+
const newManager = new KeystoreManager(mergeKeystores(keyStores));
|
|
1519
|
+
await newManager.validateSigners();
|
|
1520
|
+
ValidatorClient.validateKeyStoreConfiguration(newManager, this.log);
|
|
1521
|
+
|
|
1522
|
+
// Validate that every validator's publisher keys overlap with the L1 signers
|
|
1523
|
+
// that were initialized at startup. Publishers cannot be hot-reloaded, so a
|
|
1524
|
+
// validator with a publisher key that doesn't match any existing L1 signer
|
|
1525
|
+
// would silently fail on every proposer slot.
|
|
1526
|
+
if (this.keyStoreManager && this.sequencer) {
|
|
1527
|
+
const oldAdapter = NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager);
|
|
1528
|
+
const availablePublishers = new Set(
|
|
1529
|
+
oldAdapter
|
|
1530
|
+
.getAttesterAddresses()
|
|
1531
|
+
.flatMap(a => oldAdapter.getPublisherAddresses(a).map(p => p.toString().toLowerCase())),
|
|
1532
|
+
);
|
|
1533
|
+
|
|
1534
|
+
const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
|
|
1535
|
+
for (const attester of newAdapter.getAttesterAddresses()) {
|
|
1536
|
+
const pubs = newAdapter.getPublisherAddresses(attester);
|
|
1537
|
+
if (pubs.length > 0 && !pubs.some(p => availablePublishers.has(p.toString().toLowerCase()))) {
|
|
1538
|
+
throw new BadRequestError(
|
|
1539
|
+
`Cannot reload keystore: validator ${attester} has publisher keys ` +
|
|
1540
|
+
`[${pubs.map(p => p.toString()).join(', ')}] but none match the L1 signers initialized at startup ` +
|
|
1541
|
+
`[${[...availablePublishers].join(', ')}]. Publishers cannot be hot-reloaded — ` +
|
|
1542
|
+
`use an existing publisher key or restart the node.`,
|
|
1543
|
+
);
|
|
1544
|
+
}
|
|
1545
|
+
}
|
|
1546
|
+
}
|
|
1547
|
+
|
|
1548
|
+
// Build adapters for old and new keystores to compute diff
|
|
1549
|
+
const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
|
|
1550
|
+
const newAddresses = newAdapter.getAttesterAddresses();
|
|
1551
|
+
const oldAddresses = this.keyStoreManager
|
|
1552
|
+
? NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager).getAttesterAddresses()
|
|
1553
|
+
: [];
|
|
1554
|
+
|
|
1555
|
+
const oldSet = new Set(oldAddresses.map(a => a.toString()));
|
|
1556
|
+
const newSet = new Set(newAddresses.map(a => a.toString()));
|
|
1557
|
+
const added = newAddresses.filter(a => !oldSet.has(a.toString()));
|
|
1558
|
+
const removed = oldAddresses.filter(a => !newSet.has(a.toString()));
|
|
1559
|
+
|
|
1560
|
+
if (added.length > 0) {
|
|
1561
|
+
this.log.info(`Keystore reload: adding attester keys: ${added.map(a => a.toString()).join(', ')}`);
|
|
1562
|
+
}
|
|
1563
|
+
if (removed.length > 0) {
|
|
1564
|
+
this.log.info(`Keystore reload: removing attester keys: ${removed.map(a => a.toString()).join(', ')}`);
|
|
1565
|
+
}
|
|
1566
|
+
if (added.length === 0 && removed.length === 0) {
|
|
1567
|
+
this.log.info('Keystore reload: attester keys unchanged');
|
|
1568
|
+
}
|
|
1569
|
+
|
|
1570
|
+
// Update the validator client (coinbase, feeRecipient, attester keys)
|
|
1571
|
+
this.validatorClient.reloadKeystore(newManager);
|
|
1572
|
+
|
|
1573
|
+
// Update the publisher factory's keystore so newly-added validators
|
|
1574
|
+
// can be matched to existing publisher keys when proposing blocks.
|
|
1575
|
+
if (this.sequencer) {
|
|
1576
|
+
this.sequencer.updatePublisherNodeKeyStore(newAdapter);
|
|
1577
|
+
}
|
|
1578
|
+
|
|
1579
|
+
// Update slasher's "don't-slash-self" list with new validator addresses
|
|
1580
|
+
if (this.slasherClient && !this.config.slashSelfAllowed) {
|
|
1581
|
+
const slashValidatorsNever = unique(
|
|
1582
|
+
[...(this.config.slashValidatorsNever ?? []), ...newAddresses].map(a => a.toString()),
|
|
1583
|
+
).map(EthAddress.fromString);
|
|
1584
|
+
this.slasherClient.updateConfig({ slashValidatorsNever });
|
|
1585
|
+
}
|
|
1586
|
+
|
|
1587
|
+
this.keyStoreManager = newManager;
|
|
1588
|
+
this.log.info('Keystore reloaded: coinbase, feeRecipient, and attester keys updated');
|
|
1589
|
+
}
|
|
1590
|
+
|
|
1591
|
+
#getInitialHeaderHash(): Promise<BlockHash> {
|
|
1592
|
+
if (!this.initialHeaderHashPromise) {
|
|
1593
|
+
this.initialHeaderHashPromise = this.worldStateSynchronizer.getCommitted().getInitialHeader().hash();
|
|
1594
|
+
}
|
|
1595
|
+
return this.initialHeaderHashPromise;
|
|
1596
|
+
}
|
|
1597
|
+
|
|
1413
1598
|
/**
|
|
1414
1599
|
* Returns an instance of MerkleTreeOperations having first ensured the world state is fully synched
|
|
1415
|
-
* @param
|
|
1600
|
+
* @param block - The block parameter (block number, block hash, or 'latest') at which to get the data.
|
|
1416
1601
|
* @returns An instance of a committed MerkleTreeOperations
|
|
1417
1602
|
*/
|
|
1418
|
-
async #getWorldState(
|
|
1419
|
-
if (typeof blockNumber === 'number' && blockNumber < INITIAL_L2_BLOCK_NUM - 1) {
|
|
1420
|
-
throw new Error('Invalid block number to get world state for: ' + blockNumber);
|
|
1421
|
-
}
|
|
1422
|
-
|
|
1603
|
+
async #getWorldState(block: BlockParameter) {
|
|
1423
1604
|
let blockSyncedTo: BlockNumber = BlockNumber.ZERO;
|
|
1424
1605
|
try {
|
|
1425
1606
|
// Attempt to sync the world state if necessary
|
|
@@ -1428,15 +1609,39 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1428
1609
|
this.log.error(`Error getting world state: ${err}`);
|
|
1429
1610
|
}
|
|
1430
1611
|
|
|
1431
|
-
|
|
1432
|
-
|
|
1433
|
-
this.log.debug(`Using committed db for block ${blockNumber}, world state synced upto ${blockSyncedTo}`);
|
|
1612
|
+
if (block === 'latest') {
|
|
1613
|
+
this.log.debug(`Using committed db for block 'latest', world state synced upto ${blockSyncedTo}`);
|
|
1434
1614
|
return this.worldStateSynchronizer.getCommitted();
|
|
1435
|
-
}
|
|
1615
|
+
}
|
|
1616
|
+
|
|
1617
|
+
if (BlockHash.isBlockHash(block)) {
|
|
1618
|
+
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
1619
|
+
if (block.equals(initialBlockHash)) {
|
|
1620
|
+
// Block source doesn't handle initial header so we need to handle the case separately.
|
|
1621
|
+
return this.worldStateSynchronizer.getSnapshot(BlockNumber.ZERO);
|
|
1622
|
+
}
|
|
1623
|
+
|
|
1624
|
+
const header = await this.blockSource.getBlockHeaderByHash(block);
|
|
1625
|
+
if (!header) {
|
|
1626
|
+
throw new Error(
|
|
1627
|
+
`Block hash ${block.toString()} not found when querying world state. If the node API has been queried with anchor block hash possibly a reorg has occurred.`,
|
|
1628
|
+
);
|
|
1629
|
+
}
|
|
1630
|
+
const blockNumber = header.getBlockNumber();
|
|
1436
1631
|
this.log.debug(`Using snapshot for block ${blockNumber}, world state synced upto ${blockSyncedTo}`);
|
|
1437
|
-
return this.worldStateSynchronizer.getSnapshot(blockNumber
|
|
1438
|
-
}
|
|
1439
|
-
|
|
1632
|
+
return this.worldStateSynchronizer.getSnapshot(blockNumber);
|
|
1633
|
+
}
|
|
1634
|
+
|
|
1635
|
+
// Block number provided
|
|
1636
|
+
{
|
|
1637
|
+
const blockNumber = block as BlockNumber;
|
|
1638
|
+
|
|
1639
|
+
if (blockNumber > blockSyncedTo) {
|
|
1640
|
+
throw new Error(`Queried block ${block} not yet synced by the node (node is synced upto ${blockSyncedTo}).`);
|
|
1641
|
+
}
|
|
1642
|
+
|
|
1643
|
+
this.log.debug(`Using snapshot for block ${blockNumber}, world state synced upto ${blockSyncedTo}`);
|
|
1644
|
+
return this.worldStateSynchronizer.getSnapshot(blockNumber);
|
|
1440
1645
|
}
|
|
1441
1646
|
}
|
|
1442
1647
|
|