@aztec/aztec-node 0.0.1-commit.0b941701 → 0.0.1-commit.10bd49492
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/aztec-node/config.d.ts +8 -5
- package/dest/aztec-node/config.d.ts.map +1 -1
- package/dest/aztec-node/config.js +11 -3
- package/dest/aztec-node/node_metrics.d.ts +1 -1
- package/dest/aztec-node/node_metrics.d.ts.map +1 -1
- package/dest/aztec-node/node_metrics.js +8 -4
- package/dest/aztec-node/server.d.ts +36 -24
- package/dest/aztec-node/server.d.ts.map +1 -1
- package/dest/aztec-node/server.js +312 -155
- package/dest/sentinel/factory.d.ts +2 -2
- package/dest/sentinel/factory.d.ts.map +1 -1
- package/dest/sentinel/factory.js +1 -1
- package/dest/sentinel/sentinel.d.ts +2 -2
- package/dest/sentinel/sentinel.d.ts.map +1 -1
- package/dest/sentinel/sentinel.js +56 -30
- package/dest/sentinel/store.d.ts +2 -2
- package/dest/sentinel/store.d.ts.map +1 -1
- package/dest/sentinel/store.js +11 -7
- package/package.json +27 -25
- package/src/aztec-node/config.ts +25 -9
- package/src/aztec-node/node_metrics.ts +12 -5
- package/src/aztec-node/server.ts +397 -205
- package/src/sentinel/factory.ts +2 -7
- package/src/sentinel/sentinel.ts +59 -26
- package/src/sentinel/store.ts +12 -12
package/src/aztec-node/server.ts
CHANGED
|
@@ -1,20 +1,15 @@
|
|
|
1
1
|
import { Archiver, createArchiver } from '@aztec/archiver';
|
|
2
2
|
import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
|
|
3
3
|
import { type BlobClientInterface, createBlobClientWithFileStores } from '@aztec/blob-client/client';
|
|
4
|
-
import {
|
|
5
|
-
|
|
6
|
-
type L1_TO_L2_MSG_TREE_HEIGHT,
|
|
7
|
-
type NOTE_HASH_TREE_HEIGHT,
|
|
8
|
-
type NULLIFIER_TREE_HEIGHT,
|
|
9
|
-
type PUBLIC_DATA_TREE_HEIGHT,
|
|
10
|
-
} from '@aztec/constants';
|
|
4
|
+
import { Blob } from '@aztec/blob-lib';
|
|
5
|
+
import { ARCHIVE_HEIGHT, type L1_TO_L2_MSG_TREE_HEIGHT, type NOTE_HASH_TREE_HEIGHT } from '@aztec/constants';
|
|
11
6
|
import { EpochCache, type EpochCacheInterface } from '@aztec/epoch-cache';
|
|
12
7
|
import { createEthereumChain } from '@aztec/ethereum/chain';
|
|
13
8
|
import { getPublicClient } from '@aztec/ethereum/client';
|
|
14
9
|
import { RegistryContract, RollupContract } from '@aztec/ethereum/contracts';
|
|
15
10
|
import type { L1ContractAddresses } from '@aztec/ethereum/l1-contract-addresses';
|
|
16
11
|
import { BlockNumber, CheckpointNumber, EpochNumber, SlotNumber } from '@aztec/foundation/branded-types';
|
|
17
|
-
import { compactArray, pick } from '@aztec/foundation/collection';
|
|
12
|
+
import { compactArray, pick, unique } from '@aztec/foundation/collection';
|
|
18
13
|
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
19
14
|
import { EthAddress } from '@aztec/foundation/eth-address';
|
|
20
15
|
import { BadRequestError } from '@aztec/foundation/json-rpc';
|
|
@@ -22,14 +17,19 @@ import { type Logger, createLogger } from '@aztec/foundation/log';
|
|
|
22
17
|
import { count } from '@aztec/foundation/string';
|
|
23
18
|
import { DateProvider, Timer } from '@aztec/foundation/timer';
|
|
24
19
|
import { MembershipWitness, SiblingPath } from '@aztec/foundation/trees';
|
|
25
|
-
import { KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
|
|
20
|
+
import { type KeyStore, KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
|
|
26
21
|
import { trySnapshotSync, uploadSnapshot } from '@aztec/node-lib/actions';
|
|
22
|
+
import { createForwarderL1TxUtilsFromSigners, createL1TxUtilsFromSigners } from '@aztec/node-lib/factories';
|
|
27
23
|
import {
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
24
|
+
type P2P,
|
|
25
|
+
type P2PClientDeps,
|
|
26
|
+
createP2PClient,
|
|
27
|
+
createTxValidatorForAcceptingTxsOverRPC,
|
|
28
|
+
getDefaultAllowedSetupFunctions,
|
|
29
|
+
} from '@aztec/p2p';
|
|
32
30
|
import { ProtocolContractAddress } from '@aztec/protocol-contracts';
|
|
31
|
+
import { type ProverNode, type ProverNodeDeps, createProverNode } from '@aztec/prover-node';
|
|
32
|
+
import { createKeyStoreForProver } from '@aztec/prover-node/config';
|
|
33
33
|
import { GlobalVariableBuilder, SequencerClient, type SequencerPublisher } from '@aztec/sequencer-client';
|
|
34
34
|
import { PublicProcessorFactory } from '@aztec/simulator/server';
|
|
35
35
|
import {
|
|
@@ -41,7 +41,14 @@ import {
|
|
|
41
41
|
} from '@aztec/slasher';
|
|
42
42
|
import { CollectionLimitsConfig, PublicSimulatorConfig } from '@aztec/stdlib/avm';
|
|
43
43
|
import { AztecAddress } from '@aztec/stdlib/aztec-address';
|
|
44
|
-
import {
|
|
44
|
+
import {
|
|
45
|
+
type BlockData,
|
|
46
|
+
BlockHash,
|
|
47
|
+
type BlockParameter,
|
|
48
|
+
type DataInBlock,
|
|
49
|
+
L2Block,
|
|
50
|
+
type L2BlockSource,
|
|
51
|
+
} from '@aztec/stdlib/block';
|
|
45
52
|
import type { PublishedCheckpoint } from '@aztec/stdlib/checkpoint';
|
|
46
53
|
import type {
|
|
47
54
|
ContractClassPublic,
|
|
@@ -69,9 +76,9 @@ import {
|
|
|
69
76
|
type WorldStateSynchronizer,
|
|
70
77
|
tryStop,
|
|
71
78
|
} from '@aztec/stdlib/interfaces/server';
|
|
72
|
-
import type { LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
|
|
79
|
+
import type { DebugLogStore, LogFilter, SiloedTag, Tag, TxScopedL2Log } from '@aztec/stdlib/logs';
|
|
80
|
+
import { InMemoryDebugLogStore, NullDebugLogStore } from '@aztec/stdlib/logs';
|
|
73
81
|
import { InboxLeaf, type L1ToL2MessageSource } from '@aztec/stdlib/messaging';
|
|
74
|
-
import { P2PClientType } from '@aztec/stdlib/p2p';
|
|
75
82
|
import type { Offense, SlashPayloadRound } from '@aztec/stdlib/slashing';
|
|
76
83
|
import type { NullifierLeafPreimage, PublicDataTreeLeaf, PublicDataTreeLeafPreimage } from '@aztec/stdlib/trees';
|
|
77
84
|
import { MerkleTreeId, NullifierMembershipWitness, PublicDataWitness } from '@aztec/stdlib/trees';
|
|
@@ -103,7 +110,6 @@ import {
|
|
|
103
110
|
ValidatorClient,
|
|
104
111
|
createBlockProposalHandler,
|
|
105
112
|
createValidatorClient,
|
|
106
|
-
createValidatorForAcceptingTxs,
|
|
107
113
|
} from '@aztec/validator-client';
|
|
108
114
|
import { createWorldStateSynchronizer } from '@aztec/world-state';
|
|
109
115
|
|
|
@@ -119,7 +125,7 @@ import { NodeMetrics } from './node_metrics.js';
|
|
|
119
125
|
*/
|
|
120
126
|
export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
121
127
|
private metrics: NodeMetrics;
|
|
122
|
-
private initialHeaderHashPromise: Promise<
|
|
128
|
+
private initialHeaderHashPromise: Promise<BlockHash> | undefined = undefined;
|
|
123
129
|
|
|
124
130
|
// Prevent two snapshot operations to happen simultaneously
|
|
125
131
|
private isUploadingSnapshot = false;
|
|
@@ -135,6 +141,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
135
141
|
protected readonly l1ToL2MessageSource: L1ToL2MessageSource,
|
|
136
142
|
protected readonly worldStateSynchronizer: WorldStateSynchronizer,
|
|
137
143
|
protected readonly sequencer: SequencerClient | undefined,
|
|
144
|
+
protected readonly proverNode: ProverNode | undefined,
|
|
138
145
|
protected readonly slasherClient: SlasherClientInterface | undefined,
|
|
139
146
|
protected readonly validatorsSentinel: Sentinel | undefined,
|
|
140
147
|
protected readonly epochPruneWatcher: EpochPruneWatcher | undefined,
|
|
@@ -147,12 +154,22 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
147
154
|
private telemetry: TelemetryClient = getTelemetryClient(),
|
|
148
155
|
private log = createLogger('node'),
|
|
149
156
|
private blobClient?: BlobClientInterface,
|
|
157
|
+
private validatorClient?: ValidatorClient,
|
|
158
|
+
private keyStoreManager?: KeystoreManager,
|
|
159
|
+
private debugLogStore: DebugLogStore = new NullDebugLogStore(),
|
|
150
160
|
) {
|
|
151
161
|
this.metrics = new NodeMetrics(telemetry, 'AztecNodeService');
|
|
152
162
|
this.tracer = telemetry.getTracer('AztecNodeService');
|
|
153
163
|
|
|
154
164
|
this.log.info(`Aztec Node version: ${this.packageVersion}`);
|
|
155
165
|
this.log.info(`Aztec Node started on chain 0x${l1ChainId.toString(16)}`, config.l1Contracts);
|
|
166
|
+
|
|
167
|
+
// A defensive check that protects us against introducing a bug in the complex `createAndSync` function. We must
|
|
168
|
+
// never have debugLogStore enabled when not in test mode because then we would be accumulating debug logs in
|
|
169
|
+
// memory which could be a DoS vector on the sequencer (since no fees are paid for debug logs).
|
|
170
|
+
if (debugLogStore.isEnabled && config.realProofs) {
|
|
171
|
+
throw new Error('debugLogStore should never be enabled when realProofs are set');
|
|
172
|
+
}
|
|
156
173
|
}
|
|
157
174
|
|
|
158
175
|
public async getWorldStateSyncStatus(): Promise<WorldStateSyncStatus> {
|
|
@@ -176,11 +193,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
176
193
|
logger?: Logger;
|
|
177
194
|
publisher?: SequencerPublisher;
|
|
178
195
|
dateProvider?: DateProvider;
|
|
179
|
-
p2pClientDeps?: P2PClientDeps
|
|
196
|
+
p2pClientDeps?: P2PClientDeps;
|
|
197
|
+
proverNodeDeps?: Partial<ProverNodeDeps>;
|
|
180
198
|
} = {},
|
|
181
199
|
options: {
|
|
182
200
|
prefilledPublicData?: PublicDataTreeLeaf[];
|
|
183
201
|
dontStartSequencer?: boolean;
|
|
202
|
+
dontStartProverNode?: boolean;
|
|
184
203
|
} = {},
|
|
185
204
|
): Promise<AztecNodeService> {
|
|
186
205
|
const config = { ...inputConfig }; // Copy the config so we dont mutate the input object
|
|
@@ -190,16 +209,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
190
209
|
const dateProvider = deps.dateProvider ?? new DateProvider();
|
|
191
210
|
const ethereumChain = createEthereumChain(config.l1RpcUrls, config.l1ChainId);
|
|
192
211
|
|
|
193
|
-
// Build a key store from file if given or from environment otherwise
|
|
212
|
+
// Build a key store from file if given or from environment otherwise.
|
|
213
|
+
// We keep the raw KeyStore available so we can merge with prover keys if enableProverNode is set.
|
|
194
214
|
let keyStoreManager: KeystoreManager | undefined;
|
|
195
215
|
const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
|
|
196
216
|
if (keyStoreProvided) {
|
|
197
217
|
const keyStores = loadKeystores(config.keyStoreDirectory!);
|
|
198
218
|
keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
|
|
199
219
|
} else {
|
|
200
|
-
const
|
|
201
|
-
|
|
202
|
-
|
|
220
|
+
const rawKeyStores: KeyStore[] = [];
|
|
221
|
+
const validatorKeyStore = createKeyStoreForValidator(config);
|
|
222
|
+
if (validatorKeyStore) {
|
|
223
|
+
rawKeyStores.push(validatorKeyStore);
|
|
224
|
+
}
|
|
225
|
+
if (config.enableProverNode) {
|
|
226
|
+
const proverKeyStore = createKeyStoreForProver(config);
|
|
227
|
+
if (proverKeyStore) {
|
|
228
|
+
rawKeyStores.push(proverKeyStore);
|
|
229
|
+
}
|
|
230
|
+
}
|
|
231
|
+
if (rawKeyStores.length > 0) {
|
|
232
|
+
keyStoreManager = new KeystoreManager(
|
|
233
|
+
rawKeyStores.length === 1 ? rawKeyStores[0] : mergeKeystores(rawKeyStores),
|
|
234
|
+
);
|
|
203
235
|
}
|
|
204
236
|
}
|
|
205
237
|
|
|
@@ -210,10 +242,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
210
242
|
if (keyStoreManager === undefined) {
|
|
211
243
|
throw new Error('Failed to create key store, a requirement for running a validator');
|
|
212
244
|
}
|
|
213
|
-
if (!keyStoreProvided) {
|
|
214
|
-
log.warn(
|
|
215
|
-
'KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS',
|
|
216
|
-
);
|
|
245
|
+
if (!keyStoreProvided && process.env.NODE_ENV !== 'test') {
|
|
246
|
+
log.warn("Keystore created from env: it's recommended to use a file-based key store for production");
|
|
217
247
|
}
|
|
218
248
|
ValidatorClient.validateKeyStoreConfiguration(keyStoreManager, log);
|
|
219
249
|
}
|
|
@@ -241,10 +271,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
241
271
|
config.l1Contracts = { ...config.l1Contracts, ...l1ContractsAddresses };
|
|
242
272
|
|
|
243
273
|
const rollupContract = new RollupContract(publicClient, config.l1Contracts.rollupAddress.toString());
|
|
244
|
-
const [l1GenesisTime, slotDuration, rollupVersionFromRollup] = await Promise.all([
|
|
274
|
+
const [l1GenesisTime, slotDuration, rollupVersionFromRollup, rollupManaLimit] = await Promise.all([
|
|
245
275
|
rollupContract.getL1GenesisTime(),
|
|
246
276
|
rollupContract.getSlotDuration(),
|
|
247
277
|
rollupContract.getVersion(),
|
|
278
|
+
rollupContract.getManaLimit().then(Number),
|
|
248
279
|
] as const);
|
|
249
280
|
|
|
250
281
|
config.rollupVersion ??= Number(rollupVersionFromRollup);
|
|
@@ -255,7 +286,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
255
286
|
);
|
|
256
287
|
}
|
|
257
288
|
|
|
258
|
-
const blobClient = await createBlobClientWithFileStores(config,
|
|
289
|
+
const blobClient = await createBlobClientWithFileStores(config, log.createChild('blob-client'));
|
|
259
290
|
|
|
260
291
|
// attempt snapshot sync if possible
|
|
261
292
|
await trySnapshotSync(config, log);
|
|
@@ -279,14 +310,28 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
279
310
|
config.realProofs || config.debugForceTxProofVerification
|
|
280
311
|
? await BBCircuitVerifier.new(config)
|
|
281
312
|
: new TestCircuitVerifier(config.proverTestVerificationDelayMs);
|
|
313
|
+
|
|
314
|
+
let debugLogStore: DebugLogStore;
|
|
282
315
|
if (!config.realProofs) {
|
|
283
316
|
log.warn(`Aztec node is accepting fake proofs`);
|
|
317
|
+
|
|
318
|
+
debugLogStore = new InMemoryDebugLogStore();
|
|
319
|
+
log.info(
|
|
320
|
+
'Aztec node started in test mode (realProofs set to false) hence debug logs from public functions will be collected and served',
|
|
321
|
+
);
|
|
322
|
+
} else {
|
|
323
|
+
debugLogStore = new NullDebugLogStore();
|
|
284
324
|
}
|
|
325
|
+
|
|
285
326
|
const proofVerifier = new QueuedIVCVerifier(config, circuitVerifier);
|
|
286
327
|
|
|
328
|
+
const proverOnly = config.enableProverNode && config.disableValidator;
|
|
329
|
+
if (proverOnly) {
|
|
330
|
+
log.info('Starting in prover-only mode: skipping validator, sequencer, sentinel, and slasher subsystems');
|
|
331
|
+
}
|
|
332
|
+
|
|
287
333
|
// create the tx pool and the p2p client, which will need the l2 block source
|
|
288
334
|
const p2pClient = await createP2PClient(
|
|
289
|
-
P2PClientType.Full,
|
|
290
335
|
config,
|
|
291
336
|
archiver,
|
|
292
337
|
proofVerifier,
|
|
@@ -298,49 +343,59 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
298
343
|
deps.p2pClientDeps,
|
|
299
344
|
);
|
|
300
345
|
|
|
301
|
-
// We
|
|
302
|
-
|
|
346
|
+
// We'll accumulate sentinel watchers here
|
|
347
|
+
const watchers: Watcher[] = [];
|
|
303
348
|
|
|
304
|
-
// Create FullNodeCheckpointsBuilder for
|
|
349
|
+
// Create FullNodeCheckpointsBuilder for block proposal handling and tx validation.
|
|
350
|
+
// Override maxTxsPerCheckpoint with the validator-specific limit if set.
|
|
305
351
|
const validatorCheckpointsBuilder = new FullNodeCheckpointsBuilder(
|
|
306
|
-
{
|
|
352
|
+
{
|
|
353
|
+
...config,
|
|
354
|
+
l1GenesisTime,
|
|
355
|
+
slotDuration: Number(slotDuration),
|
|
356
|
+
rollupManaLimit,
|
|
357
|
+
maxTxsPerCheckpoint: config.validateMaxTxsPerCheckpoint,
|
|
358
|
+
},
|
|
307
359
|
worldStateSynchronizer,
|
|
308
360
|
archiver,
|
|
309
361
|
dateProvider,
|
|
310
362
|
telemetry,
|
|
311
363
|
);
|
|
312
364
|
|
|
313
|
-
|
|
314
|
-
const watchers: Watcher[] = [];
|
|
365
|
+
let validatorClient: ValidatorClient | undefined;
|
|
315
366
|
|
|
316
|
-
|
|
317
|
-
|
|
318
|
-
|
|
319
|
-
|
|
320
|
-
|
|
321
|
-
|
|
322
|
-
|
|
323
|
-
|
|
324
|
-
|
|
325
|
-
|
|
326
|
-
|
|
327
|
-
|
|
328
|
-
|
|
367
|
+
if (!proverOnly) {
|
|
368
|
+
// Create validator client if required
|
|
369
|
+
validatorClient = await createValidatorClient(config, {
|
|
370
|
+
checkpointsBuilder: validatorCheckpointsBuilder,
|
|
371
|
+
worldState: worldStateSynchronizer,
|
|
372
|
+
p2pClient,
|
|
373
|
+
telemetry,
|
|
374
|
+
dateProvider,
|
|
375
|
+
epochCache,
|
|
376
|
+
blockSource: archiver,
|
|
377
|
+
l1ToL2MessageSource: archiver,
|
|
378
|
+
keyStoreManager,
|
|
379
|
+
blobClient,
|
|
380
|
+
});
|
|
329
381
|
|
|
330
|
-
|
|
331
|
-
|
|
332
|
-
|
|
333
|
-
|
|
334
|
-
|
|
335
|
-
|
|
336
|
-
|
|
382
|
+
// If we have a validator client, register it as a source of offenses for the slasher,
|
|
383
|
+
// and have it register callbacks on the p2p client *before* we start it, otherwise messages
|
|
384
|
+
// like attestations or auths will fail.
|
|
385
|
+
if (validatorClient) {
|
|
386
|
+
watchers.push(validatorClient);
|
|
387
|
+
if (!options.dontStartSequencer) {
|
|
388
|
+
await validatorClient.registerHandlers();
|
|
389
|
+
}
|
|
337
390
|
}
|
|
338
391
|
}
|
|
339
392
|
|
|
340
|
-
// If there's no validator client
|
|
341
|
-
//
|
|
342
|
-
|
|
343
|
-
|
|
393
|
+
// If there's no validator client, create a BlockProposalHandler to handle block proposals
|
|
394
|
+
// for monitoring or reexecution. Reexecution (default) allows us to follow the pending chain,
|
|
395
|
+
// while non-reexecution is used for validating the proposals and collecting their txs.
|
|
396
|
+
if (!validatorClient) {
|
|
397
|
+
const reexecute = !!config.alwaysReexecuteBlockProposals;
|
|
398
|
+
log.info(`Setting up block proposal handler` + (reexecute ? ' with reexecution of proposals' : ''));
|
|
344
399
|
createBlockProposalHandler(config, {
|
|
345
400
|
checkpointsBuilder: validatorCheckpointsBuilder,
|
|
346
401
|
worldState: worldStateSynchronizer,
|
|
@@ -350,7 +405,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
350
405
|
p2pClient,
|
|
351
406
|
dateProvider,
|
|
352
407
|
telemetry,
|
|
353
|
-
}).
|
|
408
|
+
}).register(p2pClient, reexecute);
|
|
354
409
|
}
|
|
355
410
|
|
|
356
411
|
// Start world state and wait for it to sync to the archiver.
|
|
@@ -359,29 +414,33 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
359
414
|
// Start p2p. Note that it depends on world state to be running.
|
|
360
415
|
await p2pClient.start();
|
|
361
416
|
|
|
362
|
-
|
|
363
|
-
if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
|
|
364
|
-
watchers.push(validatorsSentinel);
|
|
365
|
-
}
|
|
366
|
-
|
|
417
|
+
let validatorsSentinel: Awaited<ReturnType<typeof createSentinel>> | undefined;
|
|
367
418
|
let epochPruneWatcher: EpochPruneWatcher | undefined;
|
|
368
|
-
if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
|
|
369
|
-
epochPruneWatcher = new EpochPruneWatcher(
|
|
370
|
-
archiver,
|
|
371
|
-
archiver,
|
|
372
|
-
epochCache,
|
|
373
|
-
p2pClient.getTxProvider(),
|
|
374
|
-
validatorCheckpointsBuilder,
|
|
375
|
-
config,
|
|
376
|
-
);
|
|
377
|
-
watchers.push(epochPruneWatcher);
|
|
378
|
-
}
|
|
379
|
-
|
|
380
|
-
// We assume we want to slash for invalid attestations unless all max penalties are set to 0
|
|
381
419
|
let attestationsBlockWatcher: AttestationsBlockWatcher | undefined;
|
|
382
|
-
|
|
383
|
-
|
|
384
|
-
|
|
420
|
+
|
|
421
|
+
if (!proverOnly) {
|
|
422
|
+
validatorsSentinel = await createSentinel(epochCache, archiver, p2pClient, config);
|
|
423
|
+
if (validatorsSentinel && config.slashInactivityPenalty > 0n) {
|
|
424
|
+
watchers.push(validatorsSentinel);
|
|
425
|
+
}
|
|
426
|
+
|
|
427
|
+
if (config.slashPrunePenalty > 0n || config.slashDataWithholdingPenalty > 0n) {
|
|
428
|
+
epochPruneWatcher = new EpochPruneWatcher(
|
|
429
|
+
archiver,
|
|
430
|
+
archiver,
|
|
431
|
+
epochCache,
|
|
432
|
+
p2pClient.getTxProvider(),
|
|
433
|
+
validatorCheckpointsBuilder,
|
|
434
|
+
config,
|
|
435
|
+
);
|
|
436
|
+
watchers.push(epochPruneWatcher);
|
|
437
|
+
}
|
|
438
|
+
|
|
439
|
+
// We assume we want to slash for invalid attestations unless all max penalties are set to 0
|
|
440
|
+
if (config.slashProposeInvalidAttestationsPenalty > 0n || config.slashAttestDescendantOfInvalidPenalty > 0n) {
|
|
441
|
+
attestationsBlockWatcher = new AttestationsBlockWatcher(archiver, epochCache, config);
|
|
442
|
+
watchers.push(attestationsBlockWatcher);
|
|
443
|
+
}
|
|
385
444
|
}
|
|
386
445
|
|
|
387
446
|
// Start p2p-related services once the archiver has completed sync
|
|
@@ -418,28 +477,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
418
477
|
);
|
|
419
478
|
await slasherClient.start();
|
|
420
479
|
|
|
421
|
-
const l1TxUtils = config.
|
|
422
|
-
? await
|
|
480
|
+
const l1TxUtils = config.sequencerPublisherForwarderAddress
|
|
481
|
+
? await createForwarderL1TxUtilsFromSigners(
|
|
423
482
|
publicClient,
|
|
424
483
|
keyStoreManager!.createAllValidatorPublisherSigners(),
|
|
425
|
-
config.
|
|
484
|
+
config.sequencerPublisherForwarderAddress,
|
|
426
485
|
{ ...config, scope: 'sequencer' },
|
|
427
|
-
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
|
|
486
|
+
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
|
|
428
487
|
)
|
|
429
|
-
: await
|
|
488
|
+
: await createL1TxUtilsFromSigners(
|
|
430
489
|
publicClient,
|
|
431
490
|
keyStoreManager!.createAllValidatorPublisherSigners(),
|
|
432
491
|
{ ...config, scope: 'sequencer' },
|
|
433
|
-
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
|
|
492
|
+
{ telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
|
|
434
493
|
);
|
|
435
494
|
|
|
436
495
|
// Create and start the sequencer client
|
|
437
496
|
const checkpointsBuilder = new CheckpointsBuilder(
|
|
438
|
-
{ ...config, l1GenesisTime, slotDuration: Number(slotDuration) },
|
|
497
|
+
{ ...config, l1GenesisTime, slotDuration: Number(slotDuration), rollupManaLimit },
|
|
439
498
|
worldStateSynchronizer,
|
|
440
499
|
archiver,
|
|
441
500
|
dateProvider,
|
|
442
501
|
telemetry,
|
|
502
|
+
debugLogStore,
|
|
443
503
|
);
|
|
444
504
|
|
|
445
505
|
sequencer = await SequencerClient.new(config, {
|
|
@@ -467,6 +527,29 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
467
527
|
log.warn(`Sequencer created but not started`);
|
|
468
528
|
}
|
|
469
529
|
|
|
530
|
+
// Create prover node subsystem if enabled
|
|
531
|
+
let proverNode: ProverNode | undefined;
|
|
532
|
+
if (config.enableProverNode) {
|
|
533
|
+
proverNode = await createProverNode(config, {
|
|
534
|
+
...deps.proverNodeDeps,
|
|
535
|
+
telemetry,
|
|
536
|
+
dateProvider,
|
|
537
|
+
archiver,
|
|
538
|
+
worldStateSynchronizer,
|
|
539
|
+
p2pClient,
|
|
540
|
+
epochCache,
|
|
541
|
+
blobClient,
|
|
542
|
+
keyStoreManager,
|
|
543
|
+
});
|
|
544
|
+
|
|
545
|
+
if (!options.dontStartProverNode) {
|
|
546
|
+
await proverNode.start();
|
|
547
|
+
log.info(`Prover node subsystem started`);
|
|
548
|
+
} else {
|
|
549
|
+
log.info(`Prover node subsystem created but not started`);
|
|
550
|
+
}
|
|
551
|
+
}
|
|
552
|
+
|
|
470
553
|
const globalVariableBuilder = new GlobalVariableBuilder({
|
|
471
554
|
...config,
|
|
472
555
|
rollupVersion: BigInt(config.rollupVersion),
|
|
@@ -474,7 +557,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
474
557
|
slotDuration: Number(slotDuration),
|
|
475
558
|
});
|
|
476
559
|
|
|
477
|
-
|
|
560
|
+
const node = new AztecNodeService(
|
|
478
561
|
config,
|
|
479
562
|
p2pClient,
|
|
480
563
|
archiver,
|
|
@@ -483,6 +566,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
483
566
|
archiver,
|
|
484
567
|
worldStateSynchronizer,
|
|
485
568
|
sequencer,
|
|
569
|
+
proverNode,
|
|
486
570
|
slasherClient,
|
|
487
571
|
validatorsSentinel,
|
|
488
572
|
epochPruneWatcher,
|
|
@@ -495,7 +579,12 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
495
579
|
telemetry,
|
|
496
580
|
log,
|
|
497
581
|
blobClient,
|
|
582
|
+
validatorClient,
|
|
583
|
+
keyStoreManager,
|
|
584
|
+
debugLogStore,
|
|
498
585
|
);
|
|
586
|
+
|
|
587
|
+
return node;
|
|
499
588
|
}
|
|
500
589
|
|
|
501
590
|
/**
|
|
@@ -506,6 +595,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
506
595
|
return this.sequencer;
|
|
507
596
|
}
|
|
508
597
|
|
|
598
|
+
/** Returns the prover node subsystem, if enabled. */
|
|
599
|
+
public getProverNode(): ProverNode | undefined {
|
|
600
|
+
return this.proverNode;
|
|
601
|
+
}
|
|
602
|
+
|
|
509
603
|
public getBlockSource(): L2BlockSource {
|
|
510
604
|
return this.blockSource;
|
|
511
605
|
}
|
|
@@ -531,7 +625,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
531
625
|
}
|
|
532
626
|
|
|
533
627
|
public async getAllowedPublicSetup(): Promise<AllowedElement[]> {
|
|
534
|
-
return this.config.
|
|
628
|
+
return [...(await getDefaultAllowedSetupFunctions()), ...(this.config.txPublicSetupAllowListExtend ?? [])];
|
|
535
629
|
}
|
|
536
630
|
|
|
537
631
|
/**
|
|
@@ -559,6 +653,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
559
653
|
enr,
|
|
560
654
|
l1ContractAddresses: contractAddresses,
|
|
561
655
|
protocolContractAddresses: protocolContractAddresses,
|
|
656
|
+
realProofs: !!this.config.realProofs,
|
|
562
657
|
};
|
|
563
658
|
|
|
564
659
|
return nodeInfo;
|
|
@@ -570,8 +665,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
570
665
|
* @returns The requested block.
|
|
571
666
|
*/
|
|
572
667
|
public async getBlock(block: BlockParameter): Promise<L2Block | undefined> {
|
|
573
|
-
if (
|
|
574
|
-
return this.getBlockByHash(
|
|
668
|
+
if (BlockHash.isBlockHash(block)) {
|
|
669
|
+
return this.getBlockByHash(block);
|
|
575
670
|
}
|
|
576
671
|
const blockNumber = block === 'latest' ? await this.getBlockNumber() : (block as BlockNumber);
|
|
577
672
|
if (blockNumber === BlockNumber.ZERO) {
|
|
@@ -585,9 +680,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
585
680
|
* @param blockHash - The block hash being requested.
|
|
586
681
|
* @returns The requested block.
|
|
587
682
|
*/
|
|
588
|
-
public async getBlockByHash(blockHash:
|
|
683
|
+
public async getBlockByHash(blockHash: BlockHash): Promise<L2Block | undefined> {
|
|
589
684
|
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
590
|
-
if (blockHash.equals(
|
|
685
|
+
if (blockHash.equals(initialBlockHash)) {
|
|
591
686
|
return this.buildInitialBlock();
|
|
592
687
|
}
|
|
593
688
|
return await this.blockSource.getL2BlockByHash(blockHash);
|
|
@@ -625,6 +720,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
625
720
|
return (await this.blockSource.getCheckpointedBlocks(from, limit)) ?? [];
|
|
626
721
|
}
|
|
627
722
|
|
|
723
|
+
public getCheckpointsDataForEpoch(epochNumber: EpochNumber) {
|
|
724
|
+
return this.blockSource.getCheckpointsDataForEpoch(epochNumber);
|
|
725
|
+
}
|
|
726
|
+
|
|
628
727
|
/**
|
|
629
728
|
* Method to fetch the current min L2 fees.
|
|
630
729
|
* @returns The current min L2 fees.
|
|
@@ -657,6 +756,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
657
756
|
return await this.blockSource.getCheckpointedL2BlockNumber();
|
|
658
757
|
}
|
|
659
758
|
|
|
759
|
+
public getCheckpointNumber(): Promise<CheckpointNumber> {
|
|
760
|
+
return this.blockSource.getCheckpointNumber();
|
|
761
|
+
}
|
|
762
|
+
|
|
660
763
|
/**
|
|
661
764
|
* Method to fetch the version of the package.
|
|
662
765
|
* @returns The node package version
|
|
@@ -692,13 +795,12 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
692
795
|
public async getPrivateLogsByTags(
|
|
693
796
|
tags: SiloedTag[],
|
|
694
797
|
page?: number,
|
|
695
|
-
referenceBlock?:
|
|
798
|
+
referenceBlock?: BlockHash,
|
|
696
799
|
): Promise<TxScopedL2Log[][]> {
|
|
697
800
|
if (referenceBlock) {
|
|
698
801
|
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
699
802
|
if (!referenceBlock.equals(initialBlockHash)) {
|
|
700
|
-
const
|
|
701
|
-
const header = await this.blockSource.getBlockHeaderByHash(blockHashFr);
|
|
803
|
+
const header = await this.blockSource.getBlockHeaderByHash(referenceBlock);
|
|
702
804
|
if (!header) {
|
|
703
805
|
throw new Error(
|
|
704
806
|
`Block ${referenceBlock.toString()} not found in the node. This might indicate a reorg has occurred.`,
|
|
@@ -713,13 +815,12 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
713
815
|
contractAddress: AztecAddress,
|
|
714
816
|
tags: Tag[],
|
|
715
817
|
page?: number,
|
|
716
|
-
referenceBlock?:
|
|
818
|
+
referenceBlock?: BlockHash,
|
|
717
819
|
): Promise<TxScopedL2Log[][]> {
|
|
718
820
|
if (referenceBlock) {
|
|
719
821
|
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
720
822
|
if (!referenceBlock.equals(initialBlockHash)) {
|
|
721
|
-
const
|
|
722
|
-
const header = await this.blockSource.getBlockHeaderByHash(blockHashFr);
|
|
823
|
+
const header = await this.blockSource.getBlockHeaderByHash(referenceBlock);
|
|
723
824
|
if (!header) {
|
|
724
825
|
throw new Error(
|
|
725
826
|
`Block ${referenceBlock.toString()} not found in the node. This might indicate a reorg has occurred.`,
|
|
@@ -769,8 +870,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
769
870
|
}
|
|
770
871
|
|
|
771
872
|
await this.p2pClient!.sendTx(tx);
|
|
772
|
-
|
|
773
|
-
this.
|
|
873
|
+
const duration = timer.ms();
|
|
874
|
+
this.metrics.receivedTx(duration, true);
|
|
875
|
+
this.log.info(`Received tx ${txHash} in ${duration}ms`, { txHash });
|
|
774
876
|
}
|
|
775
877
|
|
|
776
878
|
public async getTxReceipt(txHash: TxHash): Promise<TxReceipt> {
|
|
@@ -782,18 +884,22 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
782
884
|
// Then get the actual tx from the archiver, which tracks every tx in a mined block.
|
|
783
885
|
const settledTxReceipt = await this.blockSource.getSettledTxReceipt(txHash);
|
|
784
886
|
|
|
887
|
+
let receipt: TxReceipt;
|
|
785
888
|
if (settledTxReceipt) {
|
|
786
|
-
|
|
787
|
-
return settledTxReceipt;
|
|
889
|
+
receipt = settledTxReceipt;
|
|
788
890
|
} else if (isKnownToPool) {
|
|
789
891
|
// If the tx is in the pool but not in the archiver, it's pending.
|
|
790
892
|
// This handles race conditions between archiver and p2p, where the archiver
|
|
791
893
|
// has pruned the block in which a tx was mined, but p2p has not caught up yet.
|
|
792
|
-
|
|
894
|
+
receipt = new TxReceipt(txHash, TxStatus.PENDING, undefined, undefined);
|
|
793
895
|
} else {
|
|
794
896
|
// Otherwise, if we don't know the tx, we consider it dropped.
|
|
795
|
-
|
|
897
|
+
receipt = new TxReceipt(txHash, TxStatus.DROPPED, undefined, 'Tx dropped by P2P node');
|
|
796
898
|
}
|
|
899
|
+
|
|
900
|
+
this.debugLogStore.decorateReceiptWithLogs(txHash.toString(), receipt);
|
|
901
|
+
|
|
902
|
+
return receipt;
|
|
797
903
|
}
|
|
798
904
|
|
|
799
905
|
public getTxEffect(txHash: TxHash): Promise<IndexedTxEffect | undefined> {
|
|
@@ -810,6 +916,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
810
916
|
await tryStop(this.slasherClient);
|
|
811
917
|
await tryStop(this.proofVerifier);
|
|
812
918
|
await tryStop(this.sequencer);
|
|
919
|
+
await tryStop(this.proverNode);
|
|
813
920
|
await tryStop(this.p2pClient);
|
|
814
921
|
await tryStop(this.worldStateSynchronizer);
|
|
815
922
|
await tryStop(this.blockSource);
|
|
@@ -859,100 +966,90 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
859
966
|
}
|
|
860
967
|
|
|
861
968
|
public async findLeavesIndexes(
|
|
862
|
-
|
|
969
|
+
referenceBlock: BlockParameter,
|
|
863
970
|
treeId: MerkleTreeId,
|
|
864
971
|
leafValues: Fr[],
|
|
865
972
|
): Promise<(DataInBlock<bigint> | undefined)[]> {
|
|
866
|
-
const committedDb = await this.#getWorldState(
|
|
973
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
867
974
|
const maybeIndices = await committedDb.findLeafIndices(
|
|
868
975
|
treeId,
|
|
869
976
|
leafValues.map(x => x.toBuffer()),
|
|
870
977
|
);
|
|
871
|
-
//
|
|
872
|
-
const
|
|
978
|
+
// Filter out undefined values to query block numbers only for found leaves
|
|
979
|
+
const definedIndices = maybeIndices.filter(x => x !== undefined);
|
|
873
980
|
|
|
874
|
-
// Now we find the block numbers for the indices
|
|
875
|
-
const blockNumbers = await committedDb.getBlockNumbersForLeafIndices(treeId,
|
|
981
|
+
// Now we find the block numbers for the defined indices
|
|
982
|
+
const blockNumbers = await committedDb.getBlockNumbersForLeafIndices(treeId, definedIndices);
|
|
876
983
|
|
|
877
|
-
//
|
|
878
|
-
|
|
879
|
-
|
|
880
|
-
|
|
984
|
+
// Build a map from leaf index to block number
|
|
985
|
+
const indexToBlockNumber = new Map<bigint, BlockNumber>();
|
|
986
|
+
for (let i = 0; i < definedIndices.length; i++) {
|
|
987
|
+
const blockNumber = blockNumbers[i];
|
|
988
|
+
if (blockNumber === undefined) {
|
|
989
|
+
throw new Error(
|
|
990
|
+
`Block number is undefined for leaf index ${definedIndices[i]} in tree ${MerkleTreeId[treeId]}`,
|
|
991
|
+
);
|
|
881
992
|
}
|
|
993
|
+
indexToBlockNumber.set(definedIndices[i], blockNumber);
|
|
882
994
|
}
|
|
883
995
|
|
|
884
996
|
// Get unique block numbers in order to optimize num calls to getLeafValue function.
|
|
885
|
-
const uniqueBlockNumbers = [...new Set(
|
|
997
|
+
const uniqueBlockNumbers = [...new Set(indexToBlockNumber.values())];
|
|
886
998
|
|
|
887
|
-
// Now we obtain the block hashes from the archive tree
|
|
888
|
-
// (note that block number corresponds to the leaf index in the archive tree).
|
|
999
|
+
// Now we obtain the block hashes from the archive tree (block number = leaf index in archive tree).
|
|
889
1000
|
const blockHashes = await Promise.all(
|
|
890
1001
|
uniqueBlockNumbers.map(blockNumber => {
|
|
891
1002
|
return committedDb.getLeafValue(MerkleTreeId.ARCHIVE, BigInt(blockNumber));
|
|
892
1003
|
}),
|
|
893
1004
|
);
|
|
894
1005
|
|
|
895
|
-
//
|
|
1006
|
+
// Build a map from block number to block hash
|
|
1007
|
+
const blockNumberToHash = new Map<BlockNumber, Fr>();
|
|
896
1008
|
for (let i = 0; i < uniqueBlockNumbers.length; i++) {
|
|
897
|
-
|
|
1009
|
+
const blockHash = blockHashes[i];
|
|
1010
|
+
if (blockHash === undefined) {
|
|
898
1011
|
throw new Error(`Block hash is undefined for block number ${uniqueBlockNumbers[i]}`);
|
|
899
1012
|
}
|
|
1013
|
+
blockNumberToHash.set(uniqueBlockNumbers[i], blockHash);
|
|
900
1014
|
}
|
|
901
1015
|
|
|
902
1016
|
// Create DataInBlock objects by combining indices, blockNumbers and blockHashes and return them.
|
|
903
|
-
return maybeIndices.map(
|
|
1017
|
+
return maybeIndices.map(index => {
|
|
904
1018
|
if (index === undefined) {
|
|
905
1019
|
return undefined;
|
|
906
1020
|
}
|
|
907
|
-
const blockNumber =
|
|
1021
|
+
const blockNumber = indexToBlockNumber.get(index);
|
|
908
1022
|
if (blockNumber === undefined) {
|
|
909
|
-
|
|
1023
|
+
throw new Error(`Block number not found for leaf index ${index} in tree ${MerkleTreeId[treeId]}`);
|
|
910
1024
|
}
|
|
911
|
-
const
|
|
912
|
-
|
|
913
|
-
|
|
914
|
-
return undefined;
|
|
1025
|
+
const blockHash = blockNumberToHash.get(blockNumber);
|
|
1026
|
+
if (blockHash === undefined) {
|
|
1027
|
+
throw new Error(`Block hash not found for block number ${blockNumber}`);
|
|
915
1028
|
}
|
|
916
1029
|
return {
|
|
917
|
-
l2BlockNumber:
|
|
918
|
-
l2BlockHash:
|
|
1030
|
+
l2BlockNumber: blockNumber,
|
|
1031
|
+
l2BlockHash: new BlockHash(blockHash),
|
|
919
1032
|
data: index,
|
|
920
1033
|
};
|
|
921
1034
|
});
|
|
922
1035
|
}
|
|
923
1036
|
|
|
924
|
-
public async
|
|
925
|
-
|
|
926
|
-
|
|
927
|
-
): Promise<SiblingPath<typeof NULLIFIER_TREE_HEIGHT>> {
|
|
928
|
-
const committedDb = await this.#getWorldState(block);
|
|
929
|
-
return committedDb.getSiblingPath(MerkleTreeId.NULLIFIER_TREE, leafIndex);
|
|
930
|
-
}
|
|
931
|
-
|
|
932
|
-
public async getNoteHashSiblingPath(
|
|
933
|
-
block: BlockParameter,
|
|
934
|
-
leafIndex: bigint,
|
|
935
|
-
): Promise<SiblingPath<typeof NOTE_HASH_TREE_HEIGHT>> {
|
|
936
|
-
const committedDb = await this.#getWorldState(block);
|
|
937
|
-
return committedDb.getSiblingPath(MerkleTreeId.NOTE_HASH_TREE, leafIndex);
|
|
938
|
-
}
|
|
939
|
-
|
|
940
|
-
public async getArchiveMembershipWitness(
|
|
941
|
-
block: BlockParameter,
|
|
942
|
-
archive: Fr,
|
|
1037
|
+
public async getBlockHashMembershipWitness(
|
|
1038
|
+
referenceBlock: BlockParameter,
|
|
1039
|
+
blockHash: BlockHash,
|
|
943
1040
|
): Promise<MembershipWitness<typeof ARCHIVE_HEIGHT> | undefined> {
|
|
944
|
-
const committedDb = await this.#getWorldState(
|
|
945
|
-
const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.ARCHIVE>(MerkleTreeId.ARCHIVE, [
|
|
1041
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1042
|
+
const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.ARCHIVE>(MerkleTreeId.ARCHIVE, [blockHash]);
|
|
946
1043
|
return pathAndIndex === undefined
|
|
947
1044
|
? undefined
|
|
948
1045
|
: MembershipWitness.fromSiblingPath(pathAndIndex.index, pathAndIndex.path);
|
|
949
1046
|
}
|
|
950
1047
|
|
|
951
1048
|
public async getNoteHashMembershipWitness(
|
|
952
|
-
|
|
1049
|
+
referenceBlock: BlockParameter,
|
|
953
1050
|
noteHash: Fr,
|
|
954
1051
|
): Promise<MembershipWitness<typeof NOTE_HASH_TREE_HEIGHT> | undefined> {
|
|
955
|
-
const committedDb = await this.#getWorldState(
|
|
1052
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
956
1053
|
const [pathAndIndex] = await committedDb.findSiblingPaths<MerkleTreeId.NOTE_HASH_TREE>(
|
|
957
1054
|
MerkleTreeId.NOTE_HASH_TREE,
|
|
958
1055
|
[noteHash],
|
|
@@ -963,10 +1060,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
963
1060
|
}
|
|
964
1061
|
|
|
965
1062
|
public async getL1ToL2MessageMembershipWitness(
|
|
966
|
-
|
|
1063
|
+
referenceBlock: BlockParameter,
|
|
967
1064
|
l1ToL2Message: Fr,
|
|
968
1065
|
): Promise<[bigint, SiblingPath<typeof L1_TO_L2_MSG_TREE_HEIGHT>] | undefined> {
|
|
969
|
-
const db = await this.#getWorldState(
|
|
1066
|
+
const db = await this.#getWorldState(referenceBlock);
|
|
970
1067
|
const [witness] = await db.findSiblingPaths(MerkleTreeId.L1_TO_L2_MESSAGE_TREE, [l1ToL2Message]);
|
|
971
1068
|
if (!witness) {
|
|
972
1069
|
return undefined;
|
|
@@ -976,11 +1073,9 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
976
1073
|
return [witness.index, witness.path];
|
|
977
1074
|
}
|
|
978
1075
|
|
|
979
|
-
public async
|
|
1076
|
+
public async getL1ToL2MessageCheckpoint(l1ToL2Message: Fr): Promise<CheckpointNumber | undefined> {
|
|
980
1077
|
const messageIndex = await this.l1ToL2MessageSource.getL1ToL2MessageIndex(l1ToL2Message);
|
|
981
|
-
return messageIndex
|
|
982
|
-
? BlockNumber.fromCheckpointNumber(InboxLeaf.checkpointNumberFromIndex(messageIndex))
|
|
983
|
-
: undefined;
|
|
1078
|
+
return messageIndex ? InboxLeaf.checkpointNumberFromIndex(messageIndex) : undefined;
|
|
984
1079
|
}
|
|
985
1080
|
|
|
986
1081
|
/**
|
|
@@ -1019,27 +1114,11 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1019
1114
|
);
|
|
1020
1115
|
}
|
|
1021
1116
|
|
|
1022
|
-
public async getArchiveSiblingPath(
|
|
1023
|
-
block: BlockParameter,
|
|
1024
|
-
leafIndex: bigint,
|
|
1025
|
-
): Promise<SiblingPath<typeof ARCHIVE_HEIGHT>> {
|
|
1026
|
-
const committedDb = await this.#getWorldState(block);
|
|
1027
|
-
return committedDb.getSiblingPath(MerkleTreeId.ARCHIVE, leafIndex);
|
|
1028
|
-
}
|
|
1029
|
-
|
|
1030
|
-
public async getPublicDataSiblingPath(
|
|
1031
|
-
block: BlockParameter,
|
|
1032
|
-
leafIndex: bigint,
|
|
1033
|
-
): Promise<SiblingPath<typeof PUBLIC_DATA_TREE_HEIGHT>> {
|
|
1034
|
-
const committedDb = await this.#getWorldState(block);
|
|
1035
|
-
return committedDb.getSiblingPath(MerkleTreeId.PUBLIC_DATA_TREE, leafIndex);
|
|
1036
|
-
}
|
|
1037
|
-
|
|
1038
1117
|
public async getNullifierMembershipWitness(
|
|
1039
|
-
|
|
1118
|
+
referenceBlock: BlockParameter,
|
|
1040
1119
|
nullifier: Fr,
|
|
1041
1120
|
): Promise<NullifierMembershipWitness | undefined> {
|
|
1042
|
-
const db = await this.#getWorldState(
|
|
1121
|
+
const db = await this.#getWorldState(referenceBlock);
|
|
1043
1122
|
const [witness] = await db.findSiblingPaths(MerkleTreeId.NULLIFIER_TREE, [nullifier.toBuffer()]);
|
|
1044
1123
|
if (!witness) {
|
|
1045
1124
|
return undefined;
|
|
@@ -1056,7 +1135,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1056
1135
|
|
|
1057
1136
|
/**
|
|
1058
1137
|
* Returns a low nullifier membership witness for a given nullifier at a given block.
|
|
1059
|
-
* @param
|
|
1138
|
+
* @param referenceBlock - The block parameter (block number, block hash, or 'latest') at which to get the data
|
|
1139
|
+
* (which contains the root of the nullifier tree in which we are searching for the nullifier).
|
|
1060
1140
|
* @param nullifier - Nullifier we try to find the low nullifier witness for.
|
|
1061
1141
|
* @returns The low nullifier membership witness (if found).
|
|
1062
1142
|
* @remarks Low nullifier witness can be used to perform a nullifier non-inclusion proof by leveraging the "linked
|
|
@@ -1069,10 +1149,10 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1069
1149
|
* TODO: This is a confusing behavior and we should eventually address that.
|
|
1070
1150
|
*/
|
|
1071
1151
|
public async getLowNullifierMembershipWitness(
|
|
1072
|
-
|
|
1152
|
+
referenceBlock: BlockParameter,
|
|
1073
1153
|
nullifier: Fr,
|
|
1074
1154
|
): Promise<NullifierMembershipWitness | undefined> {
|
|
1075
|
-
const committedDb = await this.#getWorldState(
|
|
1155
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1076
1156
|
const findResult = await committedDb.getPreviousValueIndex(MerkleTreeId.NULLIFIER_TREE, nullifier.toBigInt());
|
|
1077
1157
|
if (!findResult) {
|
|
1078
1158
|
return undefined;
|
|
@@ -1087,8 +1167,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1087
1167
|
return new NullifierMembershipWitness(BigInt(index), preimageData as NullifierLeafPreimage, siblingPath);
|
|
1088
1168
|
}
|
|
1089
1169
|
|
|
1090
|
-
async getPublicDataWitness(
|
|
1091
|
-
const committedDb = await this.#getWorldState(
|
|
1170
|
+
async getPublicDataWitness(referenceBlock: BlockParameter, leafSlot: Fr): Promise<PublicDataWitness | undefined> {
|
|
1171
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1092
1172
|
const lowLeafResult = await committedDb.getPreviousValueIndex(MerkleTreeId.PUBLIC_DATA_TREE, leafSlot.toBigInt());
|
|
1093
1173
|
if (!lowLeafResult) {
|
|
1094
1174
|
return undefined;
|
|
@@ -1102,8 +1182,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1102
1182
|
}
|
|
1103
1183
|
}
|
|
1104
1184
|
|
|
1105
|
-
public async getPublicStorageAt(
|
|
1106
|
-
const committedDb = await this.#getWorldState(
|
|
1185
|
+
public async getPublicStorageAt(referenceBlock: BlockParameter, contract: AztecAddress, slot: Fr): Promise<Fr> {
|
|
1186
|
+
const committedDb = await this.#getWorldState(referenceBlock);
|
|
1107
1187
|
const leafSlot = await computePublicDataTreeLeafSlot(contract, slot);
|
|
1108
1188
|
|
|
1109
1189
|
const lowLeafResult = await committedDb.getPreviousValueIndex(MerkleTreeId.PUBLIC_DATA_TREE, leafSlot.toBigInt());
|
|
@@ -1118,14 +1198,13 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1118
1198
|
}
|
|
1119
1199
|
|
|
1120
1200
|
public async getBlockHeader(block: BlockParameter = 'latest'): Promise<BlockHeader | undefined> {
|
|
1121
|
-
if (
|
|
1201
|
+
if (BlockHash.isBlockHash(block)) {
|
|
1122
1202
|
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
1123
1203
|
if (block.equals(initialBlockHash)) {
|
|
1124
1204
|
// Block source doesn't handle initial header so we need to handle the case separately.
|
|
1125
1205
|
return this.worldStateSynchronizer.getCommitted().getInitialHeader();
|
|
1126
1206
|
}
|
|
1127
|
-
|
|
1128
|
-
return this.blockSource.getBlockHeaderByHash(blockHashFr);
|
|
1207
|
+
return this.blockSource.getBlockHeaderByHash(block);
|
|
1129
1208
|
} else {
|
|
1130
1209
|
// Block source doesn't handle initial header so we need to handle the case separately.
|
|
1131
1210
|
const blockNumber = block === 'latest' ? await this.getBlockNumber() : (block as BlockNumber);
|
|
@@ -1145,6 +1224,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1145
1224
|
return await this.blockSource.getBlockHeaderByArchive(archive);
|
|
1146
1225
|
}
|
|
1147
1226
|
|
|
1227
|
+
public getBlockData(number: BlockNumber): Promise<BlockData | undefined> {
|
|
1228
|
+
return this.blockSource.getBlockData(number);
|
|
1229
|
+
}
|
|
1230
|
+
|
|
1231
|
+
public getBlockDataByArchive(archive: Fr): Promise<BlockData | undefined> {
|
|
1232
|
+
return this.blockSource.getBlockDataByArchive(archive);
|
|
1233
|
+
}
|
|
1234
|
+
|
|
1148
1235
|
/**
|
|
1149
1236
|
* Simulates the public part of a transaction with the current state.
|
|
1150
1237
|
* @param tx - The transaction to simulate.
|
|
@@ -1168,7 +1255,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1168
1255
|
}
|
|
1169
1256
|
|
|
1170
1257
|
const txHash = tx.getTxHash();
|
|
1171
|
-
const
|
|
1258
|
+
const latestBlockNumber = await this.blockSource.getBlockNumber();
|
|
1259
|
+
const blockNumber = BlockNumber.add(latestBlockNumber, 1);
|
|
1172
1260
|
|
|
1173
1261
|
// If sequencer is not initialized, we just set these values to zero for simulation.
|
|
1174
1262
|
const coinbase = EthAddress.ZERO;
|
|
@@ -1183,6 +1271,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1183
1271
|
this.contractDataSource,
|
|
1184
1272
|
new DateProvider(),
|
|
1185
1273
|
this.telemetry,
|
|
1274
|
+
this.log.getBindings(),
|
|
1186
1275
|
);
|
|
1187
1276
|
|
|
1188
1277
|
this.log.verbose(`Simulating public calls for tx ${txHash}`, {
|
|
@@ -1191,6 +1280,8 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1191
1280
|
blockNumber,
|
|
1192
1281
|
});
|
|
1193
1282
|
|
|
1283
|
+
// Ensure world-state has caught up with the latest block we loaded from the archiver
|
|
1284
|
+
await this.worldStateSynchronizer.syncImmediate(latestBlockNumber);
|
|
1194
1285
|
const merkleTreeFork = await this.worldStateSynchronizer.fork();
|
|
1195
1286
|
try {
|
|
1196
1287
|
const config = PublicSimulatorConfig.from({
|
|
@@ -1206,7 +1297,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1206
1297
|
const processor = publicProcessorFactory.create(merkleTreeFork, newGlobalVariables, config);
|
|
1207
1298
|
|
|
1208
1299
|
// REFACTOR: Consider merging ProcessReturnValues into ProcessedTx
|
|
1209
|
-
const [processedTxs, failedTxs, _usedTxs, returns] = await processor.process([tx]);
|
|
1300
|
+
const [processedTxs, failedTxs, _usedTxs, returns, debugLogs] = await processor.process([tx]);
|
|
1210
1301
|
// REFACTOR: Consider returning the error rather than throwing
|
|
1211
1302
|
if (failedTxs.length) {
|
|
1212
1303
|
this.log.warn(`Simulated tx ${txHash} fails: ${failedTxs[0].error}`, { txHash });
|
|
@@ -1220,6 +1311,7 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1220
1311
|
processedTx.txEffect,
|
|
1221
1312
|
returns,
|
|
1222
1313
|
processedTx.gasUsed,
|
|
1314
|
+
debugLogs,
|
|
1223
1315
|
);
|
|
1224
1316
|
} finally {
|
|
1225
1317
|
await merkleTreeFork.close();
|
|
@@ -1233,19 +1325,32 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1233
1325
|
const db = this.worldStateSynchronizer.getCommitted();
|
|
1234
1326
|
const verifier = isSimulation ? undefined : this.proofVerifier;
|
|
1235
1327
|
|
|
1236
|
-
// We accept transactions if they are not expired by the next slot (checked based on the
|
|
1328
|
+
// We accept transactions if they are not expired by the next slot (checked based on the ExpirationTimestamp field)
|
|
1237
1329
|
const { ts: nextSlotTimestamp } = this.epochCache.getEpochAndSlotInNextL1Slot();
|
|
1238
1330
|
const blockNumber = BlockNumber((await this.blockSource.getBlockNumber()) + 1);
|
|
1239
|
-
const
|
|
1240
|
-
|
|
1241
|
-
|
|
1242
|
-
|
|
1243
|
-
|
|
1244
|
-
|
|
1245
|
-
|
|
1246
|
-
|
|
1247
|
-
|
|
1248
|
-
|
|
1331
|
+
const l1Constants = await this.blockSource.getL1Constants();
|
|
1332
|
+
const validator = createTxValidatorForAcceptingTxsOverRPC(
|
|
1333
|
+
db,
|
|
1334
|
+
this.contractDataSource,
|
|
1335
|
+
verifier,
|
|
1336
|
+
{
|
|
1337
|
+
timestamp: nextSlotTimestamp,
|
|
1338
|
+
blockNumber,
|
|
1339
|
+
l1ChainId: this.l1ChainId,
|
|
1340
|
+
rollupVersion: this.version,
|
|
1341
|
+
setupAllowList: [
|
|
1342
|
+
...(await getDefaultAllowedSetupFunctions()),
|
|
1343
|
+
...(this.config.txPublicSetupAllowListExtend ?? []),
|
|
1344
|
+
],
|
|
1345
|
+
gasFees: await this.getCurrentMinFees(),
|
|
1346
|
+
skipFeeEnforcement,
|
|
1347
|
+
txsPermitted: !this.config.disableTransactions,
|
|
1348
|
+
rollupManaLimit: l1Constants.rollupManaLimit,
|
|
1349
|
+
maxBlockL2Gas: this.config.validateMaxL2BlockGas,
|
|
1350
|
+
maxBlockDAGas: this.config.validateMaxDABlockGas,
|
|
1351
|
+
},
|
|
1352
|
+
this.log.getBindings(),
|
|
1353
|
+
);
|
|
1249
1354
|
|
|
1250
1355
|
return await validator.validateTx(tx);
|
|
1251
1356
|
}
|
|
@@ -1409,7 +1514,95 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1409
1514
|
}
|
|
1410
1515
|
}
|
|
1411
1516
|
|
|
1412
|
-
|
|
1517
|
+
public async reloadKeystore(): Promise<void> {
|
|
1518
|
+
if (!this.config.keyStoreDirectory?.length) {
|
|
1519
|
+
throw new BadRequestError(
|
|
1520
|
+
'Cannot reload keystore: node is not using a file-based keystore. ' +
|
|
1521
|
+
'Set KEY_STORE_DIRECTORY to use file-based keystores.',
|
|
1522
|
+
);
|
|
1523
|
+
}
|
|
1524
|
+
if (!this.validatorClient) {
|
|
1525
|
+
throw new BadRequestError('Cannot reload keystore: validator is not enabled.');
|
|
1526
|
+
}
|
|
1527
|
+
|
|
1528
|
+
this.log.info('Reloading keystore from disk');
|
|
1529
|
+
|
|
1530
|
+
// Re-read and validate keystore files
|
|
1531
|
+
const keyStores = loadKeystores(this.config.keyStoreDirectory);
|
|
1532
|
+
const newManager = new KeystoreManager(mergeKeystores(keyStores));
|
|
1533
|
+
await newManager.validateSigners();
|
|
1534
|
+
ValidatorClient.validateKeyStoreConfiguration(newManager, this.log);
|
|
1535
|
+
|
|
1536
|
+
// Validate that every validator's publisher keys overlap with the L1 signers
|
|
1537
|
+
// that were initialized at startup. Publishers cannot be hot-reloaded, so a
|
|
1538
|
+
// validator with a publisher key that doesn't match any existing L1 signer
|
|
1539
|
+
// would silently fail on every proposer slot.
|
|
1540
|
+
if (this.keyStoreManager && this.sequencer) {
|
|
1541
|
+
const oldAdapter = NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager);
|
|
1542
|
+
const availablePublishers = new Set(
|
|
1543
|
+
oldAdapter
|
|
1544
|
+
.getAttesterAddresses()
|
|
1545
|
+
.flatMap(a => oldAdapter.getPublisherAddresses(a).map(p => p.toString().toLowerCase())),
|
|
1546
|
+
);
|
|
1547
|
+
|
|
1548
|
+
const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
|
|
1549
|
+
for (const attester of newAdapter.getAttesterAddresses()) {
|
|
1550
|
+
const pubs = newAdapter.getPublisherAddresses(attester);
|
|
1551
|
+
if (pubs.length > 0 && !pubs.some(p => availablePublishers.has(p.toString().toLowerCase()))) {
|
|
1552
|
+
throw new BadRequestError(
|
|
1553
|
+
`Cannot reload keystore: validator ${attester} has publisher keys ` +
|
|
1554
|
+
`[${pubs.map(p => p.toString()).join(', ')}] but none match the L1 signers initialized at startup ` +
|
|
1555
|
+
`[${[...availablePublishers].join(', ')}]. Publishers cannot be hot-reloaded — ` +
|
|
1556
|
+
`use an existing publisher key or restart the node.`,
|
|
1557
|
+
);
|
|
1558
|
+
}
|
|
1559
|
+
}
|
|
1560
|
+
}
|
|
1561
|
+
|
|
1562
|
+
// Build adapters for old and new keystores to compute diff
|
|
1563
|
+
const newAdapter = NodeKeystoreAdapter.fromKeyStoreManager(newManager);
|
|
1564
|
+
const newAddresses = newAdapter.getAttesterAddresses();
|
|
1565
|
+
const oldAddresses = this.keyStoreManager
|
|
1566
|
+
? NodeKeystoreAdapter.fromKeyStoreManager(this.keyStoreManager).getAttesterAddresses()
|
|
1567
|
+
: [];
|
|
1568
|
+
|
|
1569
|
+
const oldSet = new Set(oldAddresses.map(a => a.toString()));
|
|
1570
|
+
const newSet = new Set(newAddresses.map(a => a.toString()));
|
|
1571
|
+
const added = newAddresses.filter(a => !oldSet.has(a.toString()));
|
|
1572
|
+
const removed = oldAddresses.filter(a => !newSet.has(a.toString()));
|
|
1573
|
+
|
|
1574
|
+
if (added.length > 0) {
|
|
1575
|
+
this.log.info(`Keystore reload: adding attester keys: ${added.map(a => a.toString()).join(', ')}`);
|
|
1576
|
+
}
|
|
1577
|
+
if (removed.length > 0) {
|
|
1578
|
+
this.log.info(`Keystore reload: removing attester keys: ${removed.map(a => a.toString()).join(', ')}`);
|
|
1579
|
+
}
|
|
1580
|
+
if (added.length === 0 && removed.length === 0) {
|
|
1581
|
+
this.log.info('Keystore reload: attester keys unchanged');
|
|
1582
|
+
}
|
|
1583
|
+
|
|
1584
|
+
// Update the validator client (coinbase, feeRecipient, attester keys)
|
|
1585
|
+
this.validatorClient.reloadKeystore(newManager);
|
|
1586
|
+
|
|
1587
|
+
// Update the publisher factory's keystore so newly-added validators
|
|
1588
|
+
// can be matched to existing publisher keys when proposing blocks.
|
|
1589
|
+
if (this.sequencer) {
|
|
1590
|
+
this.sequencer.updatePublisherNodeKeyStore(newAdapter);
|
|
1591
|
+
}
|
|
1592
|
+
|
|
1593
|
+
// Update slasher's "don't-slash-self" list with new validator addresses
|
|
1594
|
+
if (this.slasherClient && !this.config.slashSelfAllowed) {
|
|
1595
|
+
const slashValidatorsNever = unique(
|
|
1596
|
+
[...(this.config.slashValidatorsNever ?? []), ...newAddresses].map(a => a.toString()),
|
|
1597
|
+
).map(EthAddress.fromString);
|
|
1598
|
+
this.slasherClient.updateConfig({ slashValidatorsNever });
|
|
1599
|
+
}
|
|
1600
|
+
|
|
1601
|
+
this.keyStoreManager = newManager;
|
|
1602
|
+
this.log.info('Keystore reloaded: coinbase, feeRecipient, and attester keys updated');
|
|
1603
|
+
}
|
|
1604
|
+
|
|
1605
|
+
#getInitialHeaderHash(): Promise<BlockHash> {
|
|
1413
1606
|
if (!this.initialHeaderHashPromise) {
|
|
1414
1607
|
this.initialHeaderHashPromise = this.worldStateSynchronizer.getCommitted().getInitialHeader().hash();
|
|
1415
1608
|
}
|
|
@@ -1435,15 +1628,14 @@ export class AztecNodeService implements AztecNode, AztecNodeAdmin, Traceable {
|
|
|
1435
1628
|
return this.worldStateSynchronizer.getCommitted();
|
|
1436
1629
|
}
|
|
1437
1630
|
|
|
1438
|
-
if (
|
|
1631
|
+
if (BlockHash.isBlockHash(block)) {
|
|
1439
1632
|
const initialBlockHash = await this.#getInitialHeaderHash();
|
|
1440
1633
|
if (block.equals(initialBlockHash)) {
|
|
1441
1634
|
// Block source doesn't handle initial header so we need to handle the case separately.
|
|
1442
1635
|
return this.worldStateSynchronizer.getSnapshot(BlockNumber.ZERO);
|
|
1443
1636
|
}
|
|
1444
1637
|
|
|
1445
|
-
const
|
|
1446
|
-
const header = await this.blockSource.getBlockHeaderByHash(blockHashFr);
|
|
1638
|
+
const header = await this.blockSource.getBlockHeaderByHash(block);
|
|
1447
1639
|
if (!header) {
|
|
1448
1640
|
throw new Error(
|
|
1449
1641
|
`Block hash ${block.toString()} not found when querying world state. If the node API has been queried with anchor block hash possibly a reorg has occurred.`,
|