@aztec/archiver 0.0.1-commit.aada20e3 → 0.0.1-commit.b33fc05d0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (86) hide show
  1. package/dest/archiver.d.ts +7 -4
  2. package/dest/archiver.d.ts.map +1 -1
  3. package/dest/archiver.js +62 -110
  4. package/dest/errors.d.ts +7 -9
  5. package/dest/errors.d.ts.map +1 -1
  6. package/dest/errors.js +9 -14
  7. package/dest/factory.d.ts +3 -4
  8. package/dest/factory.d.ts.map +1 -1
  9. package/dest/factory.js +15 -13
  10. package/dest/index.d.ts +2 -1
  11. package/dest/index.d.ts.map +1 -1
  12. package/dest/index.js +1 -0
  13. package/dest/l1/bin/retrieve-calldata.js +36 -33
  14. package/dest/l1/calldata_retriever.d.ts +73 -50
  15. package/dest/l1/calldata_retriever.d.ts.map +1 -1
  16. package/dest/l1/calldata_retriever.js +190 -259
  17. package/dest/l1/data_retrieval.d.ts +9 -9
  18. package/dest/l1/data_retrieval.d.ts.map +1 -1
  19. package/dest/l1/data_retrieval.js +21 -19
  20. package/dest/l1/spire_proposer.d.ts +5 -5
  21. package/dest/l1/spire_proposer.d.ts.map +1 -1
  22. package/dest/l1/spire_proposer.js +9 -17
  23. package/dest/modules/data_source_base.d.ts +10 -5
  24. package/dest/modules/data_source_base.d.ts.map +1 -1
  25. package/dest/modules/data_source_base.js +29 -73
  26. package/dest/modules/data_store_updater.d.ts +22 -7
  27. package/dest/modules/data_store_updater.d.ts.map +1 -1
  28. package/dest/modules/data_store_updater.js +69 -29
  29. package/dest/modules/instrumentation.d.ts +15 -2
  30. package/dest/modules/instrumentation.d.ts.map +1 -1
  31. package/dest/modules/instrumentation.js +19 -2
  32. package/dest/modules/l1_synchronizer.d.ts +5 -8
  33. package/dest/modules/l1_synchronizer.d.ts.map +1 -1
  34. package/dest/modules/l1_synchronizer.js +41 -10
  35. package/dest/store/block_store.d.ts +29 -26
  36. package/dest/store/block_store.d.ts.map +1 -1
  37. package/dest/store/block_store.js +130 -78
  38. package/dest/store/kv_archiver_store.d.ts +34 -11
  39. package/dest/store/kv_archiver_store.d.ts.map +1 -1
  40. package/dest/store/kv_archiver_store.js +39 -9
  41. package/dest/store/l2_tips_cache.d.ts +19 -0
  42. package/dest/store/l2_tips_cache.d.ts.map +1 -0
  43. package/dest/store/l2_tips_cache.js +89 -0
  44. package/dest/store/log_store.d.ts +1 -1
  45. package/dest/store/log_store.d.ts.map +1 -1
  46. package/dest/store/log_store.js +55 -35
  47. package/dest/store/message_store.js +1 -1
  48. package/dest/test/fake_l1_state.d.ts +13 -1
  49. package/dest/test/fake_l1_state.d.ts.map +1 -1
  50. package/dest/test/fake_l1_state.js +84 -20
  51. package/dest/test/mock_archiver.d.ts +1 -1
  52. package/dest/test/mock_archiver.d.ts.map +1 -1
  53. package/dest/test/mock_archiver.js +3 -2
  54. package/dest/test/mock_l2_block_source.d.ts +21 -5
  55. package/dest/test/mock_l2_block_source.d.ts.map +1 -1
  56. package/dest/test/mock_l2_block_source.js +132 -86
  57. package/dest/test/mock_structs.d.ts +4 -1
  58. package/dest/test/mock_structs.d.ts.map +1 -1
  59. package/dest/test/mock_structs.js +13 -1
  60. package/dest/test/noop_l1_archiver.d.ts +4 -1
  61. package/dest/test/noop_l1_archiver.d.ts.map +1 -1
  62. package/dest/test/noop_l1_archiver.js +5 -1
  63. package/package.json +13 -13
  64. package/src/archiver.ts +74 -130
  65. package/src/errors.ts +10 -24
  66. package/src/factory.ts +29 -14
  67. package/src/index.ts +1 -0
  68. package/src/l1/README.md +25 -68
  69. package/src/l1/bin/retrieve-calldata.ts +46 -39
  70. package/src/l1/calldata_retriever.ts +249 -379
  71. package/src/l1/data_retrieval.ts +23 -25
  72. package/src/l1/spire_proposer.ts +7 -15
  73. package/src/modules/data_source_base.ts +56 -95
  74. package/src/modules/data_store_updater.ts +71 -30
  75. package/src/modules/instrumentation.ts +29 -2
  76. package/src/modules/l1_synchronizer.ts +46 -14
  77. package/src/store/block_store.ts +157 -105
  78. package/src/store/kv_archiver_store.ts +62 -12
  79. package/src/store/l2_tips_cache.ts +89 -0
  80. package/src/store/log_store.ts +93 -31
  81. package/src/store/message_store.ts +1 -1
  82. package/src/test/fake_l1_state.ts +110 -21
  83. package/src/test/mock_archiver.ts +3 -2
  84. package/src/test/mock_l2_block_source.ts +173 -81
  85. package/src/test/mock_structs.ts +20 -6
  86. package/src/test/noop_l1_archiver.ts +7 -1
@@ -0,0 +1,89 @@
1
+ import { GENESIS_BLOCK_HEADER_HASH, INITIAL_L2_BLOCK_NUM } from '@aztec/constants';
2
+ import { BlockNumber, CheckpointNumber } from '@aztec/foundation/branded-types';
3
+ import { type BlockData, type CheckpointId, GENESIS_CHECKPOINT_HEADER_HASH, type L2Tips } from '@aztec/stdlib/block';
4
+
5
+ import type { BlockStore } from './block_store.js';
6
+
7
+ /**
8
+ * In-memory cache for L2 chain tips (proposed, checkpointed, proven, finalized).
9
+ * Populated from the BlockStore on first access, then kept up-to-date by the ArchiverDataStoreUpdater.
10
+ * Refresh calls should happen within the store transaction that mutates block data to ensure consistency.
11
+ */
12
+ export class L2TipsCache {
13
+ #tipsPromise: Promise<L2Tips> | undefined;
14
+
15
+ constructor(private blockStore: BlockStore) {}
16
+
17
+ /** Returns the cached L2 tips. Loads from the block store on first call. */
18
+ public getL2Tips(): Promise<L2Tips> {
19
+ return (this.#tipsPromise ??= this.loadFromStore());
20
+ }
21
+
22
+ /** Reloads the L2 tips from the block store. Should be called within the store transaction that mutates data. */
23
+ public async refresh(): Promise<void> {
24
+ this.#tipsPromise = this.loadFromStore();
25
+ await this.#tipsPromise;
26
+ }
27
+
28
+ private async loadFromStore(): Promise<L2Tips> {
29
+ const [latestBlockNumber, provenBlockNumber, checkpointedBlockNumber, finalizedBlockNumber] = await Promise.all([
30
+ this.blockStore.getLatestBlockNumber(),
31
+ this.blockStore.getProvenBlockNumber(),
32
+ this.blockStore.getCheckpointedL2BlockNumber(),
33
+ this.blockStore.getFinalizedL2BlockNumber(),
34
+ ]);
35
+
36
+ const genesisBlockHeader = {
37
+ blockHash: GENESIS_BLOCK_HEADER_HASH,
38
+ checkpointNumber: CheckpointNumber.ZERO,
39
+ } as const;
40
+ const beforeInitialBlockNumber = BlockNumber(INITIAL_L2_BLOCK_NUM - 1);
41
+
42
+ const getBlockData = (blockNumber: BlockNumber) =>
43
+ blockNumber > beforeInitialBlockNumber ? this.blockStore.getBlockData(blockNumber) : genesisBlockHeader;
44
+
45
+ const [latestBlockData, provenBlockData, checkpointedBlockData, finalizedBlockData] = await Promise.all(
46
+ [latestBlockNumber, provenBlockNumber, checkpointedBlockNumber, finalizedBlockNumber].map(getBlockData),
47
+ );
48
+
49
+ if (!latestBlockData || !provenBlockData || !finalizedBlockData || !checkpointedBlockData) {
50
+ throw new Error('Failed to load block data for L2 tips');
51
+ }
52
+
53
+ const [provenCheckpointId, finalizedCheckpointId, checkpointedCheckpointId] = await Promise.all([
54
+ this.getCheckpointIdForBlock(provenBlockData),
55
+ this.getCheckpointIdForBlock(finalizedBlockData),
56
+ this.getCheckpointIdForBlock(checkpointedBlockData),
57
+ ]);
58
+
59
+ return {
60
+ proposed: { number: latestBlockNumber, hash: latestBlockData.blockHash.toString() },
61
+ proven: {
62
+ block: { number: provenBlockNumber, hash: provenBlockData.blockHash.toString() },
63
+ checkpoint: provenCheckpointId,
64
+ },
65
+ finalized: {
66
+ block: { number: finalizedBlockNumber, hash: finalizedBlockData.blockHash.toString() },
67
+ checkpoint: finalizedCheckpointId,
68
+ },
69
+ checkpointed: {
70
+ block: { number: checkpointedBlockNumber, hash: checkpointedBlockData.blockHash.toString() },
71
+ checkpoint: checkpointedCheckpointId,
72
+ },
73
+ };
74
+ }
75
+
76
+ private async getCheckpointIdForBlock(blockData: Pick<BlockData, 'checkpointNumber'>): Promise<CheckpointId> {
77
+ const checkpointData = await this.blockStore.getCheckpointData(blockData.checkpointNumber);
78
+ if (!checkpointData) {
79
+ return {
80
+ number: CheckpointNumber.ZERO,
81
+ hash: GENESIS_CHECKPOINT_HEADER_HASH.toString(),
82
+ };
83
+ }
84
+ return {
85
+ number: checkpointData.checkpointNumber,
86
+ hash: checkpointData.header.hash().toString(),
87
+ };
88
+ }
89
+ }
@@ -20,6 +20,7 @@ import {
20
20
  Tag,
21
21
  TxScopedL2Log,
22
22
  } from '@aztec/stdlib/logs';
23
+ import { TxHash } from '@aztec/stdlib/tx';
23
24
 
24
25
  import type { BlockStore } from './block_store.js';
25
26
 
@@ -219,6 +220,7 @@ export class LogStore {
219
220
  .map((txEffect, txIndex) =>
220
221
  [
221
222
  numToUInt32BE(txIndex),
223
+ txEffect.txHash.toBuffer(),
222
224
  numToUInt32BE(txEffect.publicLogs.length),
223
225
  txEffect.publicLogs.map(log => log.toBuffer()),
224
226
  ].flat(),
@@ -242,6 +244,7 @@ export class LogStore {
242
244
  .map((txEffect, txIndex) =>
243
245
  [
244
246
  numToUInt32BE(txIndex),
247
+ txEffect.txHash.toBuffer(),
245
248
  numToUInt32BE(txEffect.contractClassLogs.length),
246
249
  txEffect.contractClassLogs.map(log => log.toBuffer()),
247
250
  ].flat(),
@@ -386,24 +389,33 @@ export class LogStore {
386
389
  }
387
390
 
388
391
  const buffer = (await this.#publicLogsByBlock.getAsync(blockNumber)) ?? Buffer.alloc(0);
389
- const publicLogsInBlock: [PublicLog[]] = [[]];
392
+ const publicLogsInBlock: { txHash: TxHash; logs: PublicLog[] }[] = [];
390
393
  const reader = new BufferReader(buffer);
391
394
 
392
395
  const blockHash = this.#unpackBlockHash(reader);
393
396
 
394
397
  while (reader.remainingBytes() > 0) {
395
398
  const indexOfTx = reader.readNumber();
399
+ const txHash = reader.readObject(TxHash);
396
400
  const numLogsInTx = reader.readNumber();
397
- publicLogsInBlock[indexOfTx] = [];
401
+ publicLogsInBlock[indexOfTx] = { txHash, logs: [] };
398
402
  for (let i = 0; i < numLogsInTx; i++) {
399
- publicLogsInBlock[indexOfTx].push(reader.readObject(PublicLog));
403
+ publicLogsInBlock[indexOfTx].logs.push(reader.readObject(PublicLog));
400
404
  }
401
405
  }
402
406
 
403
- const txLogs = publicLogsInBlock[txIndex];
407
+ const txData = publicLogsInBlock[txIndex];
404
408
 
405
409
  const logs: ExtendedPublicLog[] = [];
406
- const maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
410
+ const maxLogsHit = this.#accumulatePublicLogs(
411
+ logs,
412
+ blockNumber,
413
+ blockHash,
414
+ txIndex,
415
+ txData.txHash,
416
+ txData.logs,
417
+ filter,
418
+ );
407
419
 
408
420
  return { logs, maxLogsHit };
409
421
  }
@@ -424,22 +436,31 @@ export class LogStore {
424
436
 
425
437
  let maxLogsHit = false;
426
438
  loopOverBlocks: for await (const [blockNumber, logBuffer] of this.#publicLogsByBlock.entriesAsync({ start, end })) {
427
- const publicLogsInBlock: [PublicLog[]] = [[]];
439
+ const publicLogsInBlock: { txHash: TxHash; logs: PublicLog[] }[] = [];
428
440
  const reader = new BufferReader(logBuffer);
429
441
 
430
442
  const blockHash = this.#unpackBlockHash(reader);
431
443
 
432
444
  while (reader.remainingBytes() > 0) {
433
445
  const indexOfTx = reader.readNumber();
446
+ const txHash = reader.readObject(TxHash);
434
447
  const numLogsInTx = reader.readNumber();
435
- publicLogsInBlock[indexOfTx] = [];
448
+ publicLogsInBlock[indexOfTx] = { txHash, logs: [] };
436
449
  for (let i = 0; i < numLogsInTx; i++) {
437
- publicLogsInBlock[indexOfTx].push(reader.readObject(PublicLog));
450
+ publicLogsInBlock[indexOfTx].logs.push(reader.readObject(PublicLog));
438
451
  }
439
452
  }
440
453
  for (let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < publicLogsInBlock.length; txIndex++) {
441
- const txLogs = publicLogsInBlock[txIndex];
442
- maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
454
+ const txData = publicLogsInBlock[txIndex];
455
+ maxLogsHit = this.#accumulatePublicLogs(
456
+ logs,
457
+ blockNumber,
458
+ blockHash,
459
+ txIndex,
460
+ txData.txHash,
461
+ txData.logs,
462
+ filter,
463
+ );
443
464
  if (maxLogsHit) {
444
465
  this.#log.debug(`Max logs hit at block ${blockNumber}`);
445
466
  break loopOverBlocks;
@@ -475,24 +496,33 @@ export class LogStore {
475
496
  return { logs: [], maxLogsHit: false };
476
497
  }
477
498
  const contractClassLogsBuffer = (await this.#contractClassLogsByBlock.getAsync(blockNumber)) ?? Buffer.alloc(0);
478
- const contractClassLogsInBlock: [ContractClassLog[]] = [[]];
499
+ const contractClassLogsInBlock: { txHash: TxHash; logs: ContractClassLog[] }[] = [];
479
500
 
480
501
  const reader = new BufferReader(contractClassLogsBuffer);
481
502
  const blockHash = this.#unpackBlockHash(reader);
482
503
 
483
504
  while (reader.remainingBytes() > 0) {
484
505
  const indexOfTx = reader.readNumber();
506
+ const txHash = reader.readObject(TxHash);
485
507
  const numLogsInTx = reader.readNumber();
486
- contractClassLogsInBlock[indexOfTx] = [];
508
+ contractClassLogsInBlock[indexOfTx] = { txHash, logs: [] };
487
509
  for (let i = 0; i < numLogsInTx; i++) {
488
- contractClassLogsInBlock[indexOfTx].push(reader.readObject(ContractClassLog));
510
+ contractClassLogsInBlock[indexOfTx].logs.push(reader.readObject(ContractClassLog));
489
511
  }
490
512
  }
491
513
 
492
- const txLogs = contractClassLogsInBlock[txIndex];
514
+ const txData = contractClassLogsInBlock[txIndex];
493
515
 
494
516
  const logs: ExtendedContractClassLog[] = [];
495
- const maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
517
+ const maxLogsHit = this.#accumulateContractClassLogs(
518
+ logs,
519
+ blockNumber,
520
+ blockHash,
521
+ txIndex,
522
+ txData.txHash,
523
+ txData.logs,
524
+ filter,
525
+ );
496
526
 
497
527
  return { logs, maxLogsHit };
498
528
  }
@@ -516,20 +546,29 @@ export class LogStore {
516
546
  start,
517
547
  end,
518
548
  })) {
519
- const contractClassLogsInBlock: [ContractClassLog[]] = [[]];
549
+ const contractClassLogsInBlock: { txHash: TxHash; logs: ContractClassLog[] }[] = [];
520
550
  const reader = new BufferReader(logBuffer);
521
551
  const blockHash = this.#unpackBlockHash(reader);
522
552
  while (reader.remainingBytes() > 0) {
523
553
  const indexOfTx = reader.readNumber();
554
+ const txHash = reader.readObject(TxHash);
524
555
  const numLogsInTx = reader.readNumber();
525
- contractClassLogsInBlock[indexOfTx] = [];
556
+ contractClassLogsInBlock[indexOfTx] = { txHash, logs: [] };
526
557
  for (let i = 0; i < numLogsInTx; i++) {
527
- contractClassLogsInBlock[indexOfTx].push(reader.readObject(ContractClassLog));
558
+ contractClassLogsInBlock[indexOfTx].logs.push(reader.readObject(ContractClassLog));
528
559
  }
529
560
  }
530
561
  for (let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < contractClassLogsInBlock.length; txIndex++) {
531
- const txLogs = contractClassLogsInBlock[txIndex];
532
- maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
562
+ const txData = contractClassLogsInBlock[txIndex];
563
+ maxLogsHit = this.#accumulateContractClassLogs(
564
+ logs,
565
+ blockNumber,
566
+ blockHash,
567
+ txIndex,
568
+ txData.txHash,
569
+ txData.logs,
570
+ filter,
571
+ );
533
572
  if (maxLogsHit) {
534
573
  this.#log.debug(`Max logs hit at block ${blockNumber}`);
535
574
  break loopOverBlocks;
@@ -540,12 +579,13 @@ export class LogStore {
540
579
  return { logs, maxLogsHit };
541
580
  }
542
581
 
543
- #accumulateLogs(
544
- results: (ExtendedContractClassLog | ExtendedPublicLog)[],
582
+ #accumulatePublicLogs(
583
+ results: ExtendedPublicLog[],
545
584
  blockNumber: number,
546
585
  blockHash: BlockHash,
547
586
  txIndex: number,
548
- txLogs: (ContractClassLog | PublicLog)[],
587
+ txHash: TxHash,
588
+ txLogs: PublicLog[],
549
589
  filter: LogFilter = {},
550
590
  ): boolean {
551
591
  let maxLogsHit = false;
@@ -553,15 +593,37 @@ export class LogStore {
553
593
  for (; logIndex < txLogs.length; logIndex++) {
554
594
  const log = txLogs[logIndex];
555
595
  if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
556
- if (log instanceof ContractClassLog) {
557
- results.push(
558
- new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), blockHash, txIndex, logIndex), log),
559
- );
560
- } else if (log instanceof PublicLog) {
561
- results.push(new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txIndex, logIndex), log));
562
- } else {
563
- throw new Error('Unknown log type');
596
+ results.push(
597
+ new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
598
+ );
599
+
600
+ if (results.length >= this.#logsMaxPageSize) {
601
+ maxLogsHit = true;
602
+ break;
564
603
  }
604
+ }
605
+ }
606
+
607
+ return maxLogsHit;
608
+ }
609
+
610
+ #accumulateContractClassLogs(
611
+ results: ExtendedContractClassLog[],
612
+ blockNumber: number,
613
+ blockHash: BlockHash,
614
+ txIndex: number,
615
+ txHash: TxHash,
616
+ txLogs: ContractClassLog[],
617
+ filter: LogFilter = {},
618
+ ): boolean {
619
+ let maxLogsHit = false;
620
+ let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
621
+ for (; logIndex < txLogs.length; logIndex++) {
622
+ const log = txLogs[logIndex];
623
+ if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
624
+ results.push(
625
+ new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
626
+ );
565
627
 
566
628
  if (results.length >= this.#logsMaxPageSize) {
567
629
  maxLogsHit = true;
@@ -137,7 +137,7 @@ export class MessageStore {
137
137
  );
138
138
  }
139
139
 
140
- // Check the first message in a block has the correct index.
140
+ // Check the first message in a checkpoint has the correct index.
141
141
  if (
142
142
  (!lastMessage || message.checkpointNumber > lastMessage.checkpointNumber) &&
143
143
  message.index !== expectedStart
@@ -14,6 +14,7 @@ import { CommitteeAttestation, CommitteeAttestationsAndSigners, L2Block } from '
14
14
  import { Checkpoint } from '@aztec/stdlib/checkpoint';
15
15
  import { getSlotAtTimestamp } from '@aztec/stdlib/epoch-helpers';
16
16
  import { InboxLeaf } from '@aztec/stdlib/messaging';
17
+ import { ConsensusPayload, SignatureDomainSeparator } from '@aztec/stdlib/p2p';
17
18
  import {
18
19
  makeAndSignCommitteeAttestationsAndSigners,
19
20
  makeCheckpointAttestationFromCheckpoint,
@@ -22,7 +23,16 @@ import {
22
23
  import { AppendOnlyTreeSnapshot } from '@aztec/stdlib/trees';
23
24
 
24
25
  import { type MockProxy, mock } from 'jest-mock-extended';
25
- import { type FormattedBlock, type Transaction, encodeFunctionData, multicall3Abi, toHex } from 'viem';
26
+ import {
27
+ type AbiParameter,
28
+ type FormattedBlock,
29
+ type Transaction,
30
+ encodeAbiParameters,
31
+ encodeFunctionData,
32
+ keccak256,
33
+ multicall3Abi,
34
+ toHex,
35
+ } from 'viem';
26
36
 
27
37
  import { updateRollingHash } from '../structs/inbox_message.js';
28
38
 
@@ -87,6 +97,10 @@ type CheckpointData = {
87
97
  blobHashes: `0x${string}`[];
88
98
  blobs: Blob[];
89
99
  signers: Secp256k1Signer[];
100
+ /** Hash of the packed attestations, matching what the L1 event emits. */
101
+ attestationsHash: Buffer32;
102
+ /** Payload digest, matching what the L1 event emits. */
103
+ payloadDigest: Buffer32;
90
104
  /** If true, archiveAt will ignore it */
91
105
  pruned?: boolean;
92
106
  };
@@ -131,12 +145,17 @@ export class FakeL1State {
131
145
  private provenCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
132
146
  private targetCommitteeSize: number = 0;
133
147
  private version: bigint = 1n;
148
+ private canPruneResult: boolean = false;
134
149
 
135
150
  // Computed from checkpoints based on L1 block visibility
136
151
  private pendingCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
137
152
 
153
+ // The L1 block number reported as "finalized" (defaults to the start block)
154
+ private finalizedL1BlockNumber: bigint;
155
+
138
156
  constructor(private readonly config: FakeL1StateConfig) {
139
157
  this.l1BlockNumber = config.l1StartBlock;
158
+ this.finalizedL1BlockNumber = config.l1StartBlock;
140
159
  this.lastArchive = new AppendOnlyTreeSnapshot(config.genesisArchiveRoot, 1);
141
160
  }
142
161
 
@@ -193,10 +212,10 @@ export class FakeL1State {
193
212
  // Store the messages internally so they match the checkpoint's inHash
194
213
  this.addMessages(checkpointNumber, messagesL1BlockNumber, messages);
195
214
 
196
- // Create the transaction and blobs
197
- const tx = this.makeRollupTx(checkpoint, signers);
198
- const blobHashes = this.makeVersionedBlobHashes(checkpoint);
199
- const blobs = this.makeBlobsFromCheckpoint(checkpoint);
215
+ // Create the transaction, blobs, and event hashes
216
+ const { tx, attestationsHash, payloadDigest } = await this.makeRollupTx(checkpoint, signers);
217
+ const blobHashes = await this.makeVersionedBlobHashes(checkpoint);
218
+ const blobs = await this.makeBlobsFromCheckpoint(checkpoint);
200
219
 
201
220
  // Store the checkpoint data
202
221
  this.checkpoints.push({
@@ -207,6 +226,8 @@ export class FakeL1State {
207
226
  blobHashes,
208
227
  blobs,
209
228
  signers,
229
+ attestationsHash,
230
+ payloadDigest,
210
231
  });
211
232
 
212
233
  // Update last archive for auto-chaining
@@ -266,16 +287,40 @@ export class FakeL1State {
266
287
  this.updatePendingCheckpointNumber();
267
288
  }
268
289
 
290
+ /** Sets the L1 block number that will be reported as "finalized". */
291
+ setFinalizedL1BlockNumber(blockNumber: bigint): void {
292
+ this.finalizedL1BlockNumber = blockNumber;
293
+ }
294
+
269
295
  /** Marks a checkpoint as proven. Updates provenCheckpointNumber. */
270
296
  markCheckpointAsProven(checkpointNumber: CheckpointNumber): void {
271
297
  this.provenCheckpointNumber = checkpointNumber;
272
298
  }
273
299
 
300
+ /**
301
+ * Simulates what `rollup.getProvenCheckpointNumber({ blockNumber: atL1Block })` would return.
302
+ */
303
+ getProvenCheckpointNumberAtL1Block(atL1Block: bigint): CheckpointNumber {
304
+ if (this.provenCheckpointNumber === 0) {
305
+ return CheckpointNumber(0);
306
+ }
307
+ const checkpoint = this.checkpoints.find(cp => cp.checkpointNumber === this.provenCheckpointNumber);
308
+ if (checkpoint && checkpoint.l1BlockNumber <= atL1Block) {
309
+ return this.provenCheckpointNumber;
310
+ }
311
+ return CheckpointNumber(0);
312
+ }
313
+
274
314
  /** Sets the target committee size for attestation validation. */
275
315
  setTargetCommitteeSize(size: number): void {
276
316
  this.targetCommitteeSize = size;
277
317
  }
278
318
 
319
+ /** Sets whether the rollup contract would allow pruning at the next block. */
320
+ setCanPrune(value: boolean): void {
321
+ this.canPruneResult = value;
322
+ }
323
+
279
324
  /**
280
325
  * Removes all entries for a checkpoint number (simulates L1 reorg or prune).
281
326
  * Note: Does NOT remove messages for this checkpoint (use numL1ToL2Messages: 0 when re-adding).
@@ -384,6 +429,13 @@ export class FakeL1State {
384
429
  });
385
430
  });
386
431
 
432
+ mockRollup.getProvenCheckpointNumber.mockImplementation((options?: { blockNumber?: bigint }) => {
433
+ const atBlock = options?.blockNumber ?? this.l1BlockNumber;
434
+ return Promise.resolve(this.getProvenCheckpointNumberAtL1Block(atBlock));
435
+ });
436
+
437
+ mockRollup.canPruneAtTime.mockImplementation(() => Promise.resolve(this.canPruneResult));
438
+
387
439
  // Mock the wrapper method for fetching checkpoint events
388
440
  mockRollup.getCheckpointProposedEvents.mockImplementation((fromBlock: bigint, toBlock: bigint) =>
389
441
  Promise.resolve(this.getCheckpointProposedLogs(fromBlock, toBlock)),
@@ -425,10 +477,13 @@ export class FakeL1State {
425
477
  publicClient.getChainId.mockResolvedValue(1);
426
478
  publicClient.getBlockNumber.mockImplementation(() => Promise.resolve(this.l1BlockNumber));
427
479
 
428
- // Use async function pattern that existing test uses for getBlock
429
-
430
- publicClient.getBlock.mockImplementation((async (args: { blockNumber?: bigint } = {}) => {
431
- const blockNum = args.blockNumber ?? (await publicClient.getBlockNumber());
480
+ publicClient.getBlock.mockImplementation((async (args: { blockNumber?: bigint; blockTag?: string } = {}) => {
481
+ let blockNum: bigint;
482
+ if (args.blockTag === 'finalized') {
483
+ blockNum = this.finalizedL1BlockNumber;
484
+ } else {
485
+ blockNum = args.blockNumber ?? (await publicClient.getBlockNumber());
486
+ }
432
487
  return {
433
488
  number: blockNum,
434
489
  timestamp: BigInt(blockNum) * BigInt(this.config.ethereumSlotDuration) + this.config.l1GenesisTime,
@@ -502,10 +557,8 @@ export class FakeL1State {
502
557
  checkpointNumber: cpData.checkpointNumber,
503
558
  archive: cpData.checkpoint.archive.root,
504
559
  versionedBlobHashes: cpData.blobHashes.map(h => Buffer.from(h.slice(2), 'hex')),
505
- // These are intentionally undefined to skip hash validation in the archiver
506
- // (validation is skipped when these fields are falsy)
507
- payloadDigest: undefined,
508
- attestationsHash: undefined,
560
+ attestationsHash: cpData.attestationsHash,
561
+ payloadDigest: cpData.payloadDigest,
509
562
  },
510
563
  }));
511
564
  }
@@ -531,14 +584,17 @@ export class FakeL1State {
531
584
  }));
532
585
  }
533
586
 
534
- private makeRollupTx(checkpoint: Checkpoint, signers: Secp256k1Signer[]): Transaction {
587
+ private async makeRollupTx(
588
+ checkpoint: Checkpoint,
589
+ signers: Secp256k1Signer[],
590
+ ): Promise<{ tx: Transaction; attestationsHash: Buffer32; payloadDigest: Buffer32 }> {
535
591
  const attestations = signers
536
592
  .map(signer => makeCheckpointAttestationFromCheckpoint(checkpoint, signer))
537
593
  .map(attestation => CommitteeAttestation.fromSignature(attestation.signature))
538
594
  .map(committeeAttestation => committeeAttestation.toViem());
539
595
 
540
596
  const header = checkpoint.header.toViem();
541
- const blobInput = getPrefixedEthBlobCommitments(getBlobsPerL1Block(checkpoint.toBlobFields()));
597
+ const blobInput = getPrefixedEthBlobCommitments(await getBlobsPerL1Block(checkpoint.toBlobFields()));
542
598
  const archive = toHex(checkpoint.archive.root.toBuffer());
543
599
  const attestationsAndSigners = new CommitteeAttestationsAndSigners(
544
600
  attestations.map(attestation => CommitteeAttestation.fromViem(attestation)),
@@ -549,6 +605,8 @@ export class FakeL1State {
549
605
  signers[0],
550
606
  );
551
607
 
608
+ const packedAttestations = attestationsAndSigners.getPackedAttestations();
609
+
552
610
  const rollupInput = encodeFunctionData({
553
611
  abi: RollupAbi,
554
612
  functionName: 'propose',
@@ -558,7 +616,7 @@ export class FakeL1State {
558
616
  archive,
559
617
  oracleInput: { feeAssetPriceModifier: 0n },
560
618
  },
561
- attestationsAndSigners.getPackedAttestations(),
619
+ packedAttestations,
562
620
  attestationsAndSigners.getSigners().map(signer => signer.toString()),
563
621
  attestationsAndSignersSignature.toViemSignature(),
564
622
  blobInput,
@@ -579,21 +637,52 @@ export class FakeL1State {
579
637
  ],
580
638
  });
581
639
 
582
- return {
640
+ // Compute attestationsHash (same logic as CalldataRetriever)
641
+ const attestationsHash = Buffer32.fromString(
642
+ keccak256(encodeAbiParameters([this.getCommitteeAttestationsStructDef()], [packedAttestations])),
643
+ );
644
+
645
+ // Compute payloadDigest (same logic as CalldataRetriever)
646
+ const consensusPayload = ConsensusPayload.fromCheckpoint(checkpoint);
647
+ const payloadToSign = consensusPayload.getPayloadToSign(SignatureDomainSeparator.checkpointAttestation);
648
+ const payloadDigest = Buffer32.fromString(keccak256(payloadToSign));
649
+
650
+ const tx = {
583
651
  input: multiCallInput,
584
652
  hash: archive,
585
653
  blockHash: archive,
586
654
  to: MULTI_CALL_3_ADDRESS as `0x${string}`,
587
655
  } as Transaction<bigint, number>;
656
+
657
+ return { tx, attestationsHash, payloadDigest };
658
+ }
659
+
660
+ /** Extracts the CommitteeAttestations struct definition from RollupAbi for hash computation. */
661
+ private getCommitteeAttestationsStructDef(): AbiParameter {
662
+ const proposeFunction = RollupAbi.find(item => item.type === 'function' && item.name === 'propose') as
663
+ | { type: 'function'; name: string; inputs: readonly AbiParameter[] }
664
+ | undefined;
665
+
666
+ if (!proposeFunction) {
667
+ throw new Error('propose function not found in RollupAbi');
668
+ }
669
+
670
+ const attestationsParam = proposeFunction.inputs.find(param => param.name === '_attestations');
671
+ if (!attestationsParam) {
672
+ throw new Error('_attestations parameter not found in propose function');
673
+ }
674
+
675
+ const tupleParam = attestationsParam as unknown as { type: 'tuple'; components?: readonly AbiParameter[] };
676
+ return { type: 'tuple', components: tupleParam.components || [] } as AbiParameter;
588
677
  }
589
678
 
590
- private makeVersionedBlobHashes(checkpoint: Checkpoint): `0x${string}`[] {
591
- return getBlobsPerL1Block(checkpoint.toBlobFields()).map(
679
+ private async makeVersionedBlobHashes(checkpoint: Checkpoint): Promise<`0x${string}`[]> {
680
+ return (await getBlobsPerL1Block(checkpoint.toBlobFields())).map(
592
681
  b => `0x${b.getEthVersionedBlobHash().toString('hex')}` as `0x${string}`,
593
682
  );
594
683
  }
595
684
 
596
- private makeBlobsFromCheckpoint(checkpoint: Checkpoint): Blob[] {
597
- return getBlobsPerL1Block(checkpoint.toBlobFields());
685
+ private async makeBlobsFromCheckpoint(checkpoint: Checkpoint): Promise<Blob[]> {
686
+ return await getBlobsPerL1Block(checkpoint.toBlobFields());
598
687
  }
599
688
  }
@@ -56,8 +56,9 @@ export class MockPrefilledArchiver extends MockArchiver {
56
56
  }
57
57
 
58
58
  const fromBlock = this.l2Blocks.length;
59
- // TODO: Add L2 blocks and checkpoints separately once archiver has the apis for that.
60
- this.addProposedBlocks(this.prefilled.slice(fromBlock, fromBlock + numBlocks).flatMap(c => c.blocks));
59
+ const checkpointsToAdd = this.prefilled.slice(fromBlock, fromBlock + numBlocks);
60
+ this.addProposedBlocks(checkpointsToAdd.flatMap(c => c.blocks));
61
+ this.checkpointList.push(...checkpointsToAdd);
61
62
  return Promise.resolve();
62
63
  }
63
64
  }