@aztec/archiver 0.0.1-commit.d431d1c → 0.0.1-commit.dbf9cec
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +9 -0
- package/dest/archiver.d.ts +10 -6
- package/dest/archiver.d.ts.map +1 -1
- package/dest/archiver.js +37 -107
- package/dest/errors.d.ts +6 -1
- package/dest/errors.d.ts.map +1 -1
- package/dest/errors.js +8 -0
- package/dest/factory.d.ts +5 -2
- package/dest/factory.d.ts.map +1 -1
- package/dest/factory.js +16 -13
- package/dest/index.d.ts +2 -1
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +1 -0
- package/dest/l1/bin/retrieve-calldata.js +35 -32
- package/dest/l1/calldata_retriever.d.ts +73 -50
- package/dest/l1/calldata_retriever.d.ts.map +1 -1
- package/dest/l1/calldata_retriever.js +190 -259
- package/dest/l1/data_retrieval.d.ts +9 -9
- package/dest/l1/data_retrieval.d.ts.map +1 -1
- package/dest/l1/data_retrieval.js +24 -22
- package/dest/l1/spire_proposer.d.ts +5 -5
- package/dest/l1/spire_proposer.d.ts.map +1 -1
- package/dest/l1/spire_proposer.js +9 -17
- package/dest/l1/validate_trace.d.ts +6 -3
- package/dest/l1/validate_trace.d.ts.map +1 -1
- package/dest/l1/validate_trace.js +13 -9
- package/dest/modules/data_source_base.d.ts +23 -19
- package/dest/modules/data_source_base.d.ts.map +1 -1
- package/dest/modules/data_source_base.js +44 -119
- package/dest/modules/data_store_updater.d.ts +31 -20
- package/dest/modules/data_store_updater.d.ts.map +1 -1
- package/dest/modules/data_store_updater.js +79 -60
- package/dest/modules/instrumentation.d.ts +17 -4
- package/dest/modules/instrumentation.d.ts.map +1 -1
- package/dest/modules/instrumentation.js +36 -12
- package/dest/modules/l1_synchronizer.d.ts +4 -8
- package/dest/modules/l1_synchronizer.d.ts.map +1 -1
- package/dest/modules/l1_synchronizer.js +23 -19
- package/dest/store/block_store.d.ts +50 -32
- package/dest/store/block_store.d.ts.map +1 -1
- package/dest/store/block_store.js +147 -54
- package/dest/store/contract_class_store.d.ts +1 -1
- package/dest/store/contract_class_store.d.ts.map +1 -1
- package/dest/store/contract_class_store.js +11 -7
- package/dest/store/kv_archiver_store.d.ts +43 -25
- package/dest/store/kv_archiver_store.d.ts.map +1 -1
- package/dest/store/kv_archiver_store.js +38 -17
- package/dest/store/l2_tips_cache.d.ts +19 -0
- package/dest/store/l2_tips_cache.d.ts.map +1 -0
- package/dest/store/l2_tips_cache.js +89 -0
- package/dest/store/log_store.d.ts +4 -4
- package/dest/store/log_store.d.ts.map +1 -1
- package/dest/store/log_store.js +57 -37
- package/dest/test/fake_l1_state.d.ts +9 -4
- package/dest/test/fake_l1_state.d.ts.map +1 -1
- package/dest/test/fake_l1_state.js +56 -18
- package/dest/test/index.js +3 -1
- package/dest/test/mock_archiver.d.ts +1 -1
- package/dest/test/mock_archiver.d.ts.map +1 -1
- package/dest/test/mock_archiver.js +3 -2
- package/dest/test/mock_l2_block_source.d.ts +36 -21
- package/dest/test/mock_l2_block_source.d.ts.map +1 -1
- package/dest/test/mock_l2_block_source.js +151 -109
- package/dest/test/mock_structs.d.ts +3 -2
- package/dest/test/mock_structs.d.ts.map +1 -1
- package/dest/test/mock_structs.js +11 -9
- package/dest/test/noop_l1_archiver.d.ts +23 -0
- package/dest/test/noop_l1_archiver.d.ts.map +1 -0
- package/dest/test/noop_l1_archiver.js +68 -0
- package/package.json +14 -13
- package/src/archiver.ts +50 -132
- package/src/errors.ts +12 -0
- package/src/factory.ts +30 -14
- package/src/index.ts +1 -0
- package/src/l1/README.md +25 -68
- package/src/l1/bin/retrieve-calldata.ts +45 -33
- package/src/l1/calldata_retriever.ts +249 -379
- package/src/l1/data_retrieval.ts +27 -29
- package/src/l1/spire_proposer.ts +7 -15
- package/src/l1/validate_trace.ts +24 -6
- package/src/modules/data_source_base.ts +73 -163
- package/src/modules/data_store_updater.ts +92 -63
- package/src/modules/instrumentation.ts +46 -14
- package/src/modules/l1_synchronizer.ts +26 -24
- package/src/store/block_store.ts +188 -92
- package/src/store/contract_class_store.ts +11 -7
- package/src/store/kv_archiver_store.ts +69 -29
- package/src/store/l2_tips_cache.ts +89 -0
- package/src/store/log_store.ts +105 -43
- package/src/test/fake_l1_state.ts +77 -19
- package/src/test/index.ts +3 -0
- package/src/test/mock_archiver.ts +3 -2
- package/src/test/mock_l2_block_source.ts +196 -126
- package/src/test/mock_structs.ts +26 -10
- package/src/test/noop_l1_archiver.ts +109 -0
package/src/archiver.ts
CHANGED
|
@@ -1,5 +1,4 @@
|
|
|
1
1
|
import type { BlobClientInterface } from '@aztec/blob-client/client';
|
|
2
|
-
import { GENESIS_BLOCK_HEADER_HASH, INITIAL_L2_BLOCK_NUM } from '@aztec/constants';
|
|
3
2
|
import { EpochCache } from '@aztec/epoch-cache';
|
|
4
3
|
import { BlockTagTooOldError, RollupContract } from '@aztec/ethereum/contracts';
|
|
5
4
|
import type { L1ContractAddresses } from '@aztec/ethereum/l1-contract-addresses';
|
|
@@ -15,9 +14,7 @@ import { RunningPromise, makeLoggingErrorHandler } from '@aztec/foundation/runni
|
|
|
15
14
|
import { DateProvider } from '@aztec/foundation/timer';
|
|
16
15
|
import {
|
|
17
16
|
type ArchiverEmitter,
|
|
18
|
-
|
|
19
|
-
GENESIS_CHECKPOINT_HEADER_HASH,
|
|
20
|
-
L2BlockNew,
|
|
17
|
+
L2Block,
|
|
21
18
|
type L2BlockSink,
|
|
22
19
|
type L2Tips,
|
|
23
20
|
type ValidateCheckpointResult,
|
|
@@ -26,6 +23,7 @@ import { PublishedCheckpoint } from '@aztec/stdlib/checkpoint';
|
|
|
26
23
|
import {
|
|
27
24
|
type L1RollupConstants,
|
|
28
25
|
getEpochNumberAtTimestamp,
|
|
26
|
+
getSlotAtNextL1Block,
|
|
29
27
|
getSlotAtTimestamp,
|
|
30
28
|
getSlotRangeForEpoch,
|
|
31
29
|
getTimestampRangeForEpoch,
|
|
@@ -40,13 +38,14 @@ import { ArchiverDataStoreUpdater } from './modules/data_store_updater.js';
|
|
|
40
38
|
import type { ArchiverInstrumentation } from './modules/instrumentation.js';
|
|
41
39
|
import type { ArchiverL1Synchronizer } from './modules/l1_synchronizer.js';
|
|
42
40
|
import type { KVArchiverDataStore } from './store/kv_archiver_store.js';
|
|
41
|
+
import { L2TipsCache } from './store/l2_tips_cache.js';
|
|
43
42
|
|
|
44
43
|
/** Export ArchiverEmitter for use in factory and tests. */
|
|
45
44
|
export type { ArchiverEmitter };
|
|
46
45
|
|
|
47
46
|
/** Request to add a block to the archiver, queued for processing by the sync loop. */
|
|
48
47
|
type AddBlockRequest = {
|
|
49
|
-
block:
|
|
48
|
+
block: L2Block;
|
|
50
49
|
resolve: () => void;
|
|
51
50
|
reject: (err: Error) => void;
|
|
52
51
|
};
|
|
@@ -68,7 +67,7 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
68
67
|
public readonly events: ArchiverEmitter;
|
|
69
68
|
|
|
70
69
|
/** A loop in which we will be continually fetching new checkpoints. */
|
|
71
|
-
|
|
70
|
+
protected runningPromise: RunningPromise;
|
|
72
71
|
|
|
73
72
|
/** L1 synchronizer that handles fetching checkpoints and messages from L1. */
|
|
74
73
|
private readonly synchronizer: ArchiverL1Synchronizer;
|
|
@@ -82,6 +81,9 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
82
81
|
/** Helper to handle updates to the store */
|
|
83
82
|
private readonly updater: ArchiverDataStoreUpdater;
|
|
84
83
|
|
|
84
|
+
/** In-memory cache for L2 chain tips. */
|
|
85
|
+
private readonly l2TipsCache: L2TipsCache;
|
|
86
|
+
|
|
85
87
|
public readonly tracer: Tracer;
|
|
86
88
|
|
|
87
89
|
/**
|
|
@@ -121,6 +123,7 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
121
123
|
protected override readonly l1Constants: L1RollupConstants & { l1StartBlockHash: Buffer32; genesisArchiveRoot: Fr },
|
|
122
124
|
synchronizer: ArchiverL1Synchronizer,
|
|
123
125
|
events: ArchiverEmitter,
|
|
126
|
+
l2TipsCache?: L2TipsCache,
|
|
124
127
|
private readonly log: Logger = createLogger('archiver'),
|
|
125
128
|
) {
|
|
126
129
|
super(dataStore, l1Constants);
|
|
@@ -129,7 +132,8 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
129
132
|
this.initialSyncPromise = promiseWithResolvers();
|
|
130
133
|
this.synchronizer = synchronizer;
|
|
131
134
|
this.events = events;
|
|
132
|
-
this.
|
|
135
|
+
this.l2TipsCache = l2TipsCache ?? new L2TipsCache(this.dataStore.blockStore);
|
|
136
|
+
this.updater = new ArchiverDataStoreUpdater(this.dataStore, this.l2TipsCache);
|
|
133
137
|
|
|
134
138
|
// Running promise starts with a small interval inbetween runs, so all iterations needed for the initial sync
|
|
135
139
|
// are done as fast as possible. This then gets updated once the initial sync completes.
|
|
@@ -158,7 +162,11 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
158
162
|
|
|
159
163
|
await this.blobClient.testSources();
|
|
160
164
|
await this.synchronizer.testEthereumNodeSynced();
|
|
161
|
-
await validateAndLogTraceAvailability(
|
|
165
|
+
await validateAndLogTraceAvailability(
|
|
166
|
+
this.debugClient,
|
|
167
|
+
this.config.ethereumAllowNoDebugHosts ?? false,
|
|
168
|
+
this.log.getBindings(),
|
|
169
|
+
);
|
|
162
170
|
|
|
163
171
|
// Log initial state for the archiver
|
|
164
172
|
const { l1StartBlock } = this.l1Constants;
|
|
@@ -187,7 +195,7 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
187
195
|
* @param block - The L2 block to add.
|
|
188
196
|
* @returns A promise that resolves when the block has been added to the store, or rejects on error.
|
|
189
197
|
*/
|
|
190
|
-
public addBlock(block:
|
|
198
|
+
public addBlock(block: L2Block): Promise<void> {
|
|
191
199
|
return new Promise<void>((resolve, reject) => {
|
|
192
200
|
this.blockQueue.push({ block, resolve, reject });
|
|
193
201
|
this.log.debug(`Queued block ${block.number} for processing`);
|
|
@@ -212,10 +220,25 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
212
220
|
const queuedItems = this.blockQueue.splice(0, this.blockQueue.length);
|
|
213
221
|
this.log.debug(`Processing ${queuedItems.length} queued block(s)`);
|
|
214
222
|
|
|
223
|
+
// Calculate slot threshold for validation
|
|
224
|
+
const l1Timestamp = this.synchronizer.getL1Timestamp();
|
|
225
|
+
const slotAtNextL1Block =
|
|
226
|
+
l1Timestamp === undefined ? undefined : getSlotAtNextL1Block(l1Timestamp, this.l1Constants);
|
|
227
|
+
|
|
215
228
|
// Process each block individually to properly resolve/reject each promise
|
|
216
229
|
for (const { block, resolve, reject } of queuedItems) {
|
|
230
|
+
const blockSlot = block.header.globalVariables.slotNumber;
|
|
231
|
+
if (slotAtNextL1Block !== undefined && blockSlot < slotAtNextL1Block) {
|
|
232
|
+
this.log.warn(
|
|
233
|
+
`Rejecting proposed block ${block.number} for past slot ${blockSlot} (current is ${slotAtNextL1Block})`,
|
|
234
|
+
{ block: block.toBlockInfo(), l1Timestamp, slotAtNextL1Block },
|
|
235
|
+
);
|
|
236
|
+
reject(new Error(`Block ${block.number} is for past slot ${blockSlot} (current is ${slotAtNextL1Block})`));
|
|
237
|
+
continue;
|
|
238
|
+
}
|
|
239
|
+
|
|
217
240
|
try {
|
|
218
|
-
await this.updater.
|
|
241
|
+
await this.updater.addProposedBlocks([block]);
|
|
219
242
|
this.log.debug(`Added block ${block.number} to store`);
|
|
220
243
|
resolve();
|
|
221
244
|
} catch (err: any) {
|
|
@@ -323,8 +346,11 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
323
346
|
}
|
|
324
347
|
|
|
325
348
|
public async isEpochComplete(epochNumber: EpochNumber): Promise<boolean> {
|
|
326
|
-
// The epoch is complete if the current L2 block is the last one in the epoch (or later)
|
|
327
|
-
|
|
349
|
+
// The epoch is complete if the current checkpointed L2 block is the last one in the epoch (or later).
|
|
350
|
+
// We use the checkpointed block number (synced from L1) instead of 'latest' to avoid returning true
|
|
351
|
+
// prematurely when proposed blocks have been pushed to the archiver but not yet checkpointed on L1.
|
|
352
|
+
const checkpointedBlockNumber = await this.getCheckpointedL2BlockNumber();
|
|
353
|
+
const header = checkpointedBlockNumber > 0 ? await this.getBlockHeader(checkpointedBlockNumber) : undefined;
|
|
328
354
|
const slot = header ? header.globalVariables.slotNumber : undefined;
|
|
329
355
|
const [_startSlot, endSlot] = getSlotRangeForEpoch(epochNumber, this.l1Constants);
|
|
330
356
|
if (slot && slot >= endSlot) {
|
|
@@ -355,8 +381,8 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
355
381
|
return this.initialSyncComplete;
|
|
356
382
|
}
|
|
357
383
|
|
|
358
|
-
public
|
|
359
|
-
return this.updater.
|
|
384
|
+
public removeCheckpointsAfter(checkpointNumber: CheckpointNumber): Promise<boolean> {
|
|
385
|
+
return this.updater.removeCheckpointsAfter(checkpointNumber);
|
|
360
386
|
}
|
|
361
387
|
|
|
362
388
|
/** Used by TXE to add checkpoints directly without syncing from L1. */
|
|
@@ -364,121 +390,12 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
364
390
|
checkpoints: PublishedCheckpoint[],
|
|
365
391
|
pendingChainValidationStatus?: ValidateCheckpointResult,
|
|
366
392
|
): Promise<boolean> {
|
|
367
|
-
await this.updater.
|
|
393
|
+
await this.updater.addCheckpoints(checkpoints, pendingChainValidationStatus);
|
|
368
394
|
return true;
|
|
369
395
|
}
|
|
370
396
|
|
|
371
|
-
public
|
|
372
|
-
|
|
373
|
-
this.getBlockNumber(),
|
|
374
|
-
this.getProvenBlockNumber(),
|
|
375
|
-
this.getCheckpointedBlockNumber(),
|
|
376
|
-
] as const);
|
|
377
|
-
|
|
378
|
-
// TODO(#13569): Compute proper finalized block number based on L1 finalized block.
|
|
379
|
-
// We just force it 2 epochs worth of proven data for now.
|
|
380
|
-
// NOTE: update end-to-end/src/e2e_epochs/epochs_empty_blocks.test.ts as that uses finalized blocks in computations
|
|
381
|
-
const finalizedBlockNumber = BlockNumber(Math.max(provenBlockNumber - this.l1Constants.epochDuration * 2, 0));
|
|
382
|
-
|
|
383
|
-
const beforeInitialblockNumber = BlockNumber(INITIAL_L2_BLOCK_NUM - 1);
|
|
384
|
-
|
|
385
|
-
// Get the latest block header and checkpointed blocks for proven, finalised and checkpointed blocks
|
|
386
|
-
const [latestBlockHeader, provenCheckpointedBlock, finalizedCheckpointedBlock, checkpointedBlock] =
|
|
387
|
-
await Promise.all([
|
|
388
|
-
latestBlockNumber > beforeInitialblockNumber ? this.getBlockHeader(latestBlockNumber) : undefined,
|
|
389
|
-
provenBlockNumber > beforeInitialblockNumber ? this.getCheckpointedBlock(provenBlockNumber) : undefined,
|
|
390
|
-
finalizedBlockNumber > beforeInitialblockNumber ? this.getCheckpointedBlock(finalizedBlockNumber) : undefined,
|
|
391
|
-
checkpointedBlockNumber > beforeInitialblockNumber
|
|
392
|
-
? this.getCheckpointedBlock(checkpointedBlockNumber)
|
|
393
|
-
: undefined,
|
|
394
|
-
] as const);
|
|
395
|
-
|
|
396
|
-
if (latestBlockNumber > beforeInitialblockNumber && !latestBlockHeader) {
|
|
397
|
-
throw new Error(`Failed to retrieve latest block header for block ${latestBlockNumber}`);
|
|
398
|
-
}
|
|
399
|
-
|
|
400
|
-
// Checkpointed blocks must exist for proven, finalized and checkpointed tips if they are beyond the initial block number.
|
|
401
|
-
if (checkpointedBlockNumber > beforeInitialblockNumber && !checkpointedBlock?.block.header) {
|
|
402
|
-
throw new Error(
|
|
403
|
-
`Failed to retrieve checkpointed block header for block ${checkpointedBlockNumber} (latest block is ${latestBlockNumber})`,
|
|
404
|
-
);
|
|
405
|
-
}
|
|
406
|
-
|
|
407
|
-
if (provenBlockNumber > beforeInitialblockNumber && !provenCheckpointedBlock?.block.header) {
|
|
408
|
-
throw new Error(
|
|
409
|
-
`Failed to retrieve proven checkpointed for block ${provenBlockNumber} (latest block is ${latestBlockNumber})`,
|
|
410
|
-
);
|
|
411
|
-
}
|
|
412
|
-
|
|
413
|
-
if (finalizedBlockNumber > beforeInitialblockNumber && !finalizedCheckpointedBlock?.block.header) {
|
|
414
|
-
throw new Error(
|
|
415
|
-
`Failed to retrieve finalized block header for block ${finalizedBlockNumber} (latest block is ${latestBlockNumber})`,
|
|
416
|
-
);
|
|
417
|
-
}
|
|
418
|
-
|
|
419
|
-
const latestBlockHeaderHash = (await latestBlockHeader?.hash()) ?? GENESIS_BLOCK_HEADER_HASH;
|
|
420
|
-
const provenBlockHeaderHash = (await provenCheckpointedBlock?.block.header?.hash()) ?? GENESIS_BLOCK_HEADER_HASH;
|
|
421
|
-
const finalizedBlockHeaderHash =
|
|
422
|
-
(await finalizedCheckpointedBlock?.block.header?.hash()) ?? GENESIS_BLOCK_HEADER_HASH;
|
|
423
|
-
const checkpointedBlockHeaderHash = (await checkpointedBlock?.block.header?.hash()) ?? GENESIS_BLOCK_HEADER_HASH;
|
|
424
|
-
|
|
425
|
-
// Now attempt to retrieve checkpoints for proven, finalised and checkpointed blocks
|
|
426
|
-
const [[provenBlockCheckpoint], [finalizedBlockCheckpoint], [checkpointedBlockCheckpoint]] = await Promise.all([
|
|
427
|
-
provenCheckpointedBlock !== undefined
|
|
428
|
-
? await this.getPublishedCheckpoints(provenCheckpointedBlock?.checkpointNumber, 1)
|
|
429
|
-
: [undefined],
|
|
430
|
-
finalizedCheckpointedBlock !== undefined
|
|
431
|
-
? await this.getPublishedCheckpoints(finalizedCheckpointedBlock?.checkpointNumber, 1)
|
|
432
|
-
: [undefined],
|
|
433
|
-
checkpointedBlock !== undefined
|
|
434
|
-
? await this.getPublishedCheckpoints(checkpointedBlock?.checkpointNumber, 1)
|
|
435
|
-
: [undefined],
|
|
436
|
-
]);
|
|
437
|
-
|
|
438
|
-
const initialcheckpointId: CheckpointId = {
|
|
439
|
-
number: CheckpointNumber.ZERO,
|
|
440
|
-
hash: GENESIS_CHECKPOINT_HEADER_HASH.toString(),
|
|
441
|
-
};
|
|
442
|
-
|
|
443
|
-
const makeCheckpointId = (checkpoint: PublishedCheckpoint | undefined) => {
|
|
444
|
-
if (checkpoint === undefined) {
|
|
445
|
-
return initialcheckpointId;
|
|
446
|
-
}
|
|
447
|
-
return {
|
|
448
|
-
number: checkpoint.checkpoint.number,
|
|
449
|
-
hash: checkpoint.checkpoint.hash().toString(),
|
|
450
|
-
};
|
|
451
|
-
};
|
|
452
|
-
|
|
453
|
-
const l2Tips: L2Tips = {
|
|
454
|
-
proposed: {
|
|
455
|
-
number: latestBlockNumber,
|
|
456
|
-
hash: latestBlockHeaderHash.toString(),
|
|
457
|
-
},
|
|
458
|
-
proven: {
|
|
459
|
-
block: {
|
|
460
|
-
number: provenBlockNumber,
|
|
461
|
-
hash: provenBlockHeaderHash.toString(),
|
|
462
|
-
},
|
|
463
|
-
checkpoint: makeCheckpointId(provenBlockCheckpoint),
|
|
464
|
-
},
|
|
465
|
-
finalized: {
|
|
466
|
-
block: {
|
|
467
|
-
number: finalizedBlockNumber,
|
|
468
|
-
hash: finalizedBlockHeaderHash.toString(),
|
|
469
|
-
},
|
|
470
|
-
checkpoint: makeCheckpointId(finalizedBlockCheckpoint),
|
|
471
|
-
},
|
|
472
|
-
checkpointed: {
|
|
473
|
-
block: {
|
|
474
|
-
number: checkpointedBlockNumber,
|
|
475
|
-
hash: checkpointedBlockHeaderHash.toString(),
|
|
476
|
-
},
|
|
477
|
-
checkpoint: makeCheckpointId(checkpointedBlockCheckpoint),
|
|
478
|
-
},
|
|
479
|
-
};
|
|
480
|
-
|
|
481
|
-
return l2Tips;
|
|
397
|
+
public getL2Tips(): Promise<L2Tips> {
|
|
398
|
+
return this.l2TipsCache.getL2Tips();
|
|
482
399
|
}
|
|
483
400
|
|
|
484
401
|
public async rollbackTo(targetL2BlockNumber: BlockNumber): Promise<void> {
|
|
@@ -490,13 +407,12 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
490
407
|
if (targetL2BlockNumber >= currentL2Block) {
|
|
491
408
|
throw new Error(`Target L2 block ${targetL2BlockNumber} must be less than current L2 block ${currentL2Block}`);
|
|
492
409
|
}
|
|
493
|
-
const blocksToUnwind = currentL2Block - targetL2BlockNumber;
|
|
494
410
|
const targetL2Block = await this.store.getCheckpointedBlock(targetL2BlockNumber);
|
|
495
411
|
if (!targetL2Block) {
|
|
496
412
|
throw new Error(`Target L2 block ${targetL2BlockNumber} not found`);
|
|
497
413
|
}
|
|
498
414
|
const targetL1BlockNumber = targetL2Block.l1.blockNumber;
|
|
499
|
-
const targetCheckpointNumber =
|
|
415
|
+
const targetCheckpointNumber = targetL2Block.checkpointNumber;
|
|
500
416
|
const targetL1Block = await this.publicClient.getBlock({
|
|
501
417
|
blockNumber: targetL1BlockNumber,
|
|
502
418
|
includeTransactions: false,
|
|
@@ -505,16 +421,18 @@ export class Archiver extends ArchiverDataSourceBase implements L2BlockSink, Tra
|
|
|
505
421
|
throw new Error(`Missing L1 block ${targetL1BlockNumber}`);
|
|
506
422
|
}
|
|
507
423
|
const targetL1BlockHash = Buffer32.fromString(targetL1Block.hash);
|
|
508
|
-
this.log.info(
|
|
509
|
-
|
|
510
|
-
|
|
424
|
+
this.log.info(
|
|
425
|
+
`Removing checkpoints after checkpoint ${targetCheckpointNumber} (target block ${targetL2BlockNumber})`,
|
|
426
|
+
);
|
|
427
|
+
await this.updater.removeCheckpointsAfter(targetCheckpointNumber);
|
|
428
|
+
this.log.info(`Rolling back L1 to L2 messages to checkpoint ${targetCheckpointNumber}`);
|
|
511
429
|
await this.store.rollbackL1ToL2MessagesToCheckpoint(targetCheckpointNumber);
|
|
512
430
|
this.log.info(`Setting L1 syncpoints to ${targetL1BlockNumber}`);
|
|
513
431
|
await this.store.setCheckpointSynchedL1BlockNumber(targetL1BlockNumber);
|
|
514
432
|
await this.store.setMessageSynchedL1Block({ l1BlockNumber: targetL1BlockNumber, l1BlockHash: targetL1BlockHash });
|
|
515
433
|
if (targetL2BlockNumber < currentProvenBlock) {
|
|
516
434
|
this.log.info(`Clearing proven L2 block number`);
|
|
517
|
-
await this.
|
|
435
|
+
await this.updater.setProvenCheckpointNumber(CheckpointNumber.ZERO);
|
|
518
436
|
}
|
|
519
437
|
// TODO(palla/reorg): Set the finalized block when we add support for it.
|
|
520
438
|
// if (targetL2BlockNumber < currentFinalizedBlock) {
|
package/src/errors.ts
CHANGED
|
@@ -88,3 +88,15 @@ export class BlockNotFoundError extends Error {
|
|
|
88
88
|
super(`Failed to find expected block number ${blockNumber}`);
|
|
89
89
|
}
|
|
90
90
|
}
|
|
91
|
+
|
|
92
|
+
export class CannotOverwriteCheckpointedBlockError extends Error {
|
|
93
|
+
constructor(
|
|
94
|
+
public readonly blockNumber: number,
|
|
95
|
+
public readonly lastCheckpointedBlock: number,
|
|
96
|
+
) {
|
|
97
|
+
super(
|
|
98
|
+
`Cannot add block ${blockNumber}: would overwrite checkpointed data (checkpointed up to block ${lastCheckpointedBlock})`,
|
|
99
|
+
);
|
|
100
|
+
this.name = 'CannotOverwriteCheckpointedBlockError';
|
|
101
|
+
}
|
|
102
|
+
}
|
package/src/factory.ts
CHANGED
|
@@ -6,7 +6,6 @@ import { BlockNumber } from '@aztec/foundation/branded-types';
|
|
|
6
6
|
import { Buffer32 } from '@aztec/foundation/buffer';
|
|
7
7
|
import { merge } from '@aztec/foundation/collection';
|
|
8
8
|
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
9
|
-
import { createLogger } from '@aztec/foundation/log';
|
|
10
9
|
import { DateProvider } from '@aztec/foundation/timer';
|
|
11
10
|
import type { DataStoreConfig } from '@aztec/kv-store/config';
|
|
12
11
|
import { createStore } from '@aztec/kv-store/lmdb-v2';
|
|
@@ -15,6 +14,7 @@ import { BundledProtocolContractsProvider } from '@aztec/protocol-contracts/prov
|
|
|
15
14
|
import { FunctionType, decodeFunctionSignature } from '@aztec/stdlib/abi';
|
|
16
15
|
import type { ArchiverEmitter } from '@aztec/stdlib/block';
|
|
17
16
|
import { type ContractClassPublic, computePublicBytecodeCommitment } from '@aztec/stdlib/contract';
|
|
17
|
+
import type { L1RollupConstants } from '@aztec/stdlib/epoch-helpers';
|
|
18
18
|
import { getTelemetryClient } from '@aztec/telemetry-client';
|
|
19
19
|
|
|
20
20
|
import { EventEmitter } from 'events';
|
|
@@ -25,19 +25,21 @@ import { type ArchiverConfig, mapArchiverConfig } from './config.js';
|
|
|
25
25
|
import { ArchiverInstrumentation } from './modules/instrumentation.js';
|
|
26
26
|
import { ArchiverL1Synchronizer } from './modules/l1_synchronizer.js';
|
|
27
27
|
import { ARCHIVER_DB_VERSION, KVArchiverDataStore } from './store/kv_archiver_store.js';
|
|
28
|
+
import { L2TipsCache } from './store/l2_tips_cache.js';
|
|
28
29
|
|
|
29
30
|
export const ARCHIVER_STORE_NAME = 'archiver';
|
|
30
31
|
|
|
31
32
|
/** Creates an archiver store. */
|
|
32
33
|
export async function createArchiverStore(
|
|
33
34
|
userConfig: Pick<ArchiverConfig, 'archiverStoreMapSizeKb' | 'maxLogs'> & DataStoreConfig,
|
|
35
|
+
l1Constants: Pick<L1RollupConstants, 'epochDuration'>,
|
|
34
36
|
) {
|
|
35
37
|
const config = {
|
|
36
38
|
...userConfig,
|
|
37
39
|
dataStoreMapSizeKb: userConfig.archiverStoreMapSizeKb ?? userConfig.dataStoreMapSizeKb,
|
|
38
40
|
};
|
|
39
|
-
const store = await createStore(ARCHIVER_STORE_NAME, ARCHIVER_DB_VERSION, config
|
|
40
|
-
return new KVArchiverDataStore(store, config.maxLogs);
|
|
41
|
+
const store = await createStore(ARCHIVER_STORE_NAME, ARCHIVER_DB_VERSION, config);
|
|
42
|
+
return new KVArchiverDataStore(store, config.maxLogs, l1Constants);
|
|
41
43
|
}
|
|
42
44
|
|
|
43
45
|
/**
|
|
@@ -52,7 +54,7 @@ export async function createArchiver(
|
|
|
52
54
|
deps: ArchiverDeps,
|
|
53
55
|
opts: { blockUntilSync: boolean } = { blockUntilSync: true },
|
|
54
56
|
): Promise<Archiver> {
|
|
55
|
-
const archiverStore = await createArchiverStore(config);
|
|
57
|
+
const archiverStore = await createArchiverStore(config, { epochDuration: config.aztecEpochDuration });
|
|
56
58
|
await registerProtocolContracts(archiverStore);
|
|
57
59
|
|
|
58
60
|
// Create Ethereum clients
|
|
@@ -76,14 +78,21 @@ export async function createArchiver(
|
|
|
76
78
|
const inbox = new InboxContract(publicClient, config.l1Contracts.inboxAddress);
|
|
77
79
|
|
|
78
80
|
// Fetch L1 constants from rollup contract
|
|
79
|
-
const [
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
81
|
+
const [
|
|
82
|
+
l1StartBlock,
|
|
83
|
+
l1GenesisTime,
|
|
84
|
+
proofSubmissionEpochs,
|
|
85
|
+
genesisArchiveRoot,
|
|
86
|
+
slashingProposerAddress,
|
|
87
|
+
targetCommitteeSize,
|
|
88
|
+
] = await Promise.all([
|
|
89
|
+
rollup.getL1StartBlock(),
|
|
90
|
+
rollup.getL1GenesisTime(),
|
|
91
|
+
rollup.getProofSubmissionEpochs(),
|
|
92
|
+
rollup.getGenesisArchiveTreeRoot(),
|
|
93
|
+
rollup.getSlashingProposerAddress(),
|
|
94
|
+
rollup.getTargetCommitteeSize(),
|
|
95
|
+
] as const);
|
|
87
96
|
|
|
88
97
|
const l1StartBlockHash = await publicClient
|
|
89
98
|
.getBlock({ blockNumber: l1StartBlock, includeTransactions: false })
|
|
@@ -99,6 +108,7 @@ export async function createArchiver(
|
|
|
99
108
|
slotDuration,
|
|
100
109
|
ethereumSlotDuration,
|
|
101
110
|
proofSubmissionEpochs: Number(proofSubmissionEpochs),
|
|
111
|
+
targetCommitteeSize,
|
|
102
112
|
genesisArchiveRoot: Fr.fromString(genesisArchiveRoot.toString()),
|
|
103
113
|
};
|
|
104
114
|
|
|
@@ -119,13 +129,15 @@ export async function createArchiver(
|
|
|
119
129
|
// Create the event emitter that will be shared by archiver and synchronizer
|
|
120
130
|
const events = new EventEmitter() as ArchiverEmitter;
|
|
121
131
|
|
|
132
|
+
// Create L2 tips cache shared by archiver and synchronizer
|
|
133
|
+
const l2TipsCache = new L2TipsCache(archiverStore.blockStore);
|
|
134
|
+
|
|
122
135
|
// Create the L1 synchronizer
|
|
123
136
|
const synchronizer = new ArchiverL1Synchronizer(
|
|
124
137
|
publicClient,
|
|
125
138
|
debugClient,
|
|
126
139
|
rollup,
|
|
127
140
|
inbox,
|
|
128
|
-
{ ...config.l1Contracts, slashingProposerAddress },
|
|
129
141
|
archiverStore,
|
|
130
142
|
archiverConfig,
|
|
131
143
|
deps.blobClient,
|
|
@@ -135,6 +147,8 @@ export async function createArchiver(
|
|
|
135
147
|
l1Constants,
|
|
136
148
|
events,
|
|
137
149
|
instrumentation.tracer,
|
|
150
|
+
l2TipsCache,
|
|
151
|
+
undefined, // log (use default)
|
|
138
152
|
);
|
|
139
153
|
|
|
140
154
|
const archiver = new Archiver(
|
|
@@ -149,13 +163,15 @@ export async function createArchiver(
|
|
|
149
163
|
l1Constants,
|
|
150
164
|
synchronizer,
|
|
151
165
|
events,
|
|
166
|
+
l2TipsCache,
|
|
152
167
|
);
|
|
153
168
|
|
|
154
169
|
await archiver.start(opts.blockUntilSync);
|
|
155
170
|
return archiver;
|
|
156
171
|
}
|
|
157
172
|
|
|
158
|
-
|
|
173
|
+
/** Registers protocol contracts in the archiver store. */
|
|
174
|
+
export async function registerProtocolContracts(store: KVArchiverDataStore) {
|
|
159
175
|
const blockNumber = 0;
|
|
160
176
|
for (const name of protocolContractNames) {
|
|
161
177
|
const provider = new BundledProtocolContractsProvider();
|
package/src/index.ts
CHANGED
|
@@ -8,5 +8,6 @@ export * from './config.js';
|
|
|
8
8
|
export { type L1PublishedData } from './structs/published.js';
|
|
9
9
|
export { KVArchiverDataStore, ARCHIVER_DB_VERSION } from './store/kv_archiver_store.js';
|
|
10
10
|
export { ContractInstanceStore } from './store/contract_instance_store.js';
|
|
11
|
+
export { L2TipsCache } from './store/l2_tips_cache.js';
|
|
11
12
|
|
|
12
13
|
export { retrieveCheckpointsFromRollup, retrieveL2ProofVerifiedEvents } from './l1/data_retrieval.js';
|
package/src/l1/README.md
CHANGED
|
@@ -5,29 +5,27 @@ Modules and classes to handle data retrieval from L1 for the archiver.
|
|
|
5
5
|
## Calldata Retriever
|
|
6
6
|
|
|
7
7
|
The sequencer publisher bundles multiple operations into a single multicall3 transaction for gas
|
|
8
|
-
efficiency.
|
|
8
|
+
efficiency. The archiver needs to extract the `propose` calldata from these bundled transactions
|
|
9
|
+
to reconstruct L2 blocks.
|
|
9
10
|
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
11
|
+
The retriever uses hash matching against `attestationsHash` and `payloadDigest` from the
|
|
12
|
+
`CheckpointProposed` L1 event to verify it has found the correct propose calldata. These hashes
|
|
13
|
+
are always required.
|
|
13
14
|
|
|
14
|
-
|
|
15
|
-
L2 blocks. This class needs to handle scenarios where the transaction was submitted via multicall3,
|
|
16
|
-
as well as alternative ways for submitting the `propose` call that other clients might use.
|
|
15
|
+
### Multicall3 Decoding with Hash Matching
|
|
17
16
|
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
First attempt to decode the transaction as a multicall3 `aggregate3` call with validation:
|
|
17
|
+
First attempt to decode the transaction as a multicall3 `aggregate3` call:
|
|
21
18
|
|
|
22
19
|
- Check if transaction is to multicall3 address (`0xcA11bde05977b3631167028862bE2a173976CA11`)
|
|
23
20
|
- Decode as `aggregate3(Call3[] calldata calls)`
|
|
24
|
-
-
|
|
25
|
-
-
|
|
26
|
-
|
|
27
|
-
|
|
21
|
+
- Find all calls matching the rollup contract address and the `propose` function selector
|
|
22
|
+
- Verify each candidate by computing `attestationsHash` (keccak256 of ABI-encoded attestations)
|
|
23
|
+
and `payloadDigest` (keccak256 of the consensus payload signing hash) and comparing against
|
|
24
|
+
expected values from the `CheckpointProposed` event
|
|
25
|
+
- Return the verified candidate (if multiple verify, return the first with a warning)
|
|
28
26
|
|
|
29
|
-
This
|
|
30
|
-
|
|
27
|
+
This approach works regardless of what other calls are in the multicall3 bundle, because hash
|
|
28
|
+
matching identifies the correct propose call without needing an allowlist.
|
|
31
29
|
|
|
32
30
|
### Direct Propose Call
|
|
33
31
|
|
|
@@ -35,64 +33,23 @@ Second attempt to decode the transaction as a direct `propose` call to the rollu
|
|
|
35
33
|
|
|
36
34
|
- Check if transaction is to the rollup address
|
|
37
35
|
- Decode as `propose` function call
|
|
38
|
-
- Verify
|
|
36
|
+
- Verify against expected hashes
|
|
39
37
|
- Return the transaction input as the propose calldata
|
|
40
38
|
|
|
41
|
-
This handles scenarios where clients submit transactions directly to the rollup contract without
|
|
42
|
-
using multicall3 for bundling. Any validation failure triggers fallback to the next step.
|
|
43
|
-
|
|
44
39
|
### Spire Proposer Call
|
|
45
40
|
|
|
46
|
-
Given existing attempts to route the call via the Spire proposer, we also check if the tx is
|
|
47
|
-
proposer known address
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
Similar as with the multicall3 check, we check that there are no other calls in the Spire proposer, so
|
|
51
|
-
we are absolutely sure that the only call is the successful one to the rollup. Any extraneous call would
|
|
52
|
-
imply an unexpected path to calling `propose` in the rollup contract, and since we cannot verify if the
|
|
53
|
-
calldata arguments we extracted are the correct ones (see the section below), we cannot know for sure which
|
|
54
|
-
one is the call that succeeded, so we don't know which calldata to process.
|
|
55
|
-
|
|
56
|
-
Furthermore, since the Spire proposer is upgradeable, we check if the implementation has not changed in
|
|
57
|
-
order to decode. As usual, any validation failure triggers fallback to the next step.
|
|
58
|
-
|
|
59
|
-
### Verifying Multicall3 Arguments
|
|
60
|
-
|
|
61
|
-
**This is NOT implemented for simplicity's sake**
|
|
62
|
-
|
|
63
|
-
If the checks above don't hold, such as when there are multiple calls to `propose`, then we cannot
|
|
64
|
-
reliably extract the `propose` calldata from the multicall3 arguments alone. We can try a best-effort
|
|
65
|
-
where we try all `propose` calls we see and validate them against on-chain data. Note that we can use these
|
|
66
|
-
same strategies if we were to obtain the calldata from another source.
|
|
67
|
-
|
|
68
|
-
#### TempBlockLog Verification
|
|
69
|
-
|
|
70
|
-
Read the stored `TempBlockLog` for the L2 block number from L1 and verify it matches our decoded header hash,
|
|
71
|
-
since the `TempBlockLog` stores the hash of the proposed block header, the payload commitment, and the attestations.
|
|
72
|
-
|
|
73
|
-
However, `TempBlockLog` is only stored temporarily and deleted after proven, so this method only works for recent
|
|
74
|
-
blocks, not for historical data syncing.
|
|
75
|
-
|
|
76
|
-
#### Archive Verification
|
|
77
|
-
|
|
78
|
-
Verify that the archive root in the decoded propose is correct with regard to the block header. This requires
|
|
79
|
-
hashing the block header we have retrieved, inserting it into the archive tree, and checking the resulting root
|
|
80
|
-
against the one we got from L1.
|
|
81
|
-
|
|
82
|
-
However, this requires that the archive keeps a reference to world-state, which is not the case in the current
|
|
83
|
-
system.
|
|
84
|
-
|
|
85
|
-
#### Emit Commitments in Rollup Contract
|
|
86
|
-
|
|
87
|
-
Modify rollup contract to emit commitments to the block header in the `L2BlockProposed` event, allowing us to easily
|
|
88
|
-
verify the calldata we obtained vs the emitted event.
|
|
41
|
+
Given existing attempts to route the call via the Spire proposer, we also check if the tx is
|
|
42
|
+
`to` the proposer known address. If so, we extract all wrapped calls and try each as either
|
|
43
|
+
a multicall3 or direct propose call, using hash matching to find and verify the correct one.
|
|
89
44
|
|
|
90
|
-
|
|
45
|
+
Since the Spire proposer is upgradeable, we check that the implementation has not changed in
|
|
46
|
+
order to decode. Any validation failure triggers fallback to the next step.
|
|
91
47
|
|
|
92
48
|
### Debug and Trace Transaction Fallback
|
|
93
49
|
|
|
94
|
-
Last, we use L1 node's trace/debug RPC methods to definitively identify the one successful
|
|
95
|
-
We can then extract the exact calldata that hit the `propose`
|
|
50
|
+
Last, we use L1 node's trace/debug RPC methods to definitively identify the one successful
|
|
51
|
+
`propose` call within the tx. We can then extract the exact calldata that hit the `propose`
|
|
52
|
+
function in the rollup contract.
|
|
96
53
|
|
|
97
|
-
This approach requires access to a debug-enabled L1 node, which may be more resource-intensive,
|
|
98
|
-
use it as a fallback when
|
|
54
|
+
This approach requires access to a debug-enabled L1 node, which may be more resource-intensive,
|
|
55
|
+
so we only use it as a fallback when earlier steps fail, which should be rare in practice.
|