@aztec/archiver 0.0.1-commit.ee80a48 → 0.0.1-commit.ef17749e1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (78) hide show
  1. package/dest/archiver.d.ts +7 -2
  2. package/dest/archiver.d.ts.map +1 -1
  3. package/dest/archiver.js +25 -96
  4. package/dest/factory.d.ts +2 -2
  5. package/dest/factory.d.ts.map +1 -1
  6. package/dest/factory.js +12 -8
  7. package/dest/index.d.ts +2 -1
  8. package/dest/index.d.ts.map +1 -1
  9. package/dest/index.js +1 -0
  10. package/dest/l1/bin/retrieve-calldata.js +36 -33
  11. package/dest/l1/calldata_retriever.d.ts +73 -50
  12. package/dest/l1/calldata_retriever.d.ts.map +1 -1
  13. package/dest/l1/calldata_retriever.js +190 -259
  14. package/dest/l1/data_retrieval.d.ts +9 -9
  15. package/dest/l1/data_retrieval.d.ts.map +1 -1
  16. package/dest/l1/data_retrieval.js +21 -19
  17. package/dest/l1/spire_proposer.d.ts +5 -5
  18. package/dest/l1/spire_proposer.d.ts.map +1 -1
  19. package/dest/l1/spire_proposer.js +9 -17
  20. package/dest/modules/data_source_base.d.ts +8 -3
  21. package/dest/modules/data_source_base.d.ts.map +1 -1
  22. package/dest/modules/data_source_base.js +28 -72
  23. package/dest/modules/data_store_updater.d.ts +13 -3
  24. package/dest/modules/data_store_updater.d.ts.map +1 -1
  25. package/dest/modules/data_store_updater.js +48 -19
  26. package/dest/modules/instrumentation.d.ts +15 -2
  27. package/dest/modules/instrumentation.d.ts.map +1 -1
  28. package/dest/modules/instrumentation.js +19 -2
  29. package/dest/modules/l1_synchronizer.d.ts +5 -8
  30. package/dest/modules/l1_synchronizer.d.ts.map +1 -1
  31. package/dest/modules/l1_synchronizer.js +16 -9
  32. package/dest/store/block_store.d.ts +18 -14
  33. package/dest/store/block_store.d.ts.map +1 -1
  34. package/dest/store/block_store.js +69 -17
  35. package/dest/store/kv_archiver_store.d.ts +18 -4
  36. package/dest/store/kv_archiver_store.d.ts.map +1 -1
  37. package/dest/store/kv_archiver_store.js +18 -0
  38. package/dest/store/l2_tips_cache.d.ts +19 -0
  39. package/dest/store/l2_tips_cache.d.ts.map +1 -0
  40. package/dest/store/l2_tips_cache.js +89 -0
  41. package/dest/store/log_store.d.ts +1 -1
  42. package/dest/store/log_store.d.ts.map +1 -1
  43. package/dest/store/log_store.js +55 -35
  44. package/dest/store/message_store.js +1 -1
  45. package/dest/test/fake_l1_state.d.ts +6 -1
  46. package/dest/test/fake_l1_state.d.ts.map +1 -1
  47. package/dest/test/fake_l1_state.js +56 -18
  48. package/dest/test/mock_archiver.d.ts +1 -1
  49. package/dest/test/mock_archiver.d.ts.map +1 -1
  50. package/dest/test/mock_archiver.js +3 -2
  51. package/dest/test/mock_l2_block_source.d.ts +19 -3
  52. package/dest/test/mock_l2_block_source.d.ts.map +1 -1
  53. package/dest/test/mock_l2_block_source.js +128 -82
  54. package/dest/test/mock_structs.d.ts +4 -1
  55. package/dest/test/mock_structs.d.ts.map +1 -1
  56. package/dest/test/mock_structs.js +13 -1
  57. package/package.json +13 -13
  58. package/src/archiver.ts +39 -116
  59. package/src/factory.ts +27 -10
  60. package/src/index.ts +1 -0
  61. package/src/l1/README.md +25 -68
  62. package/src/l1/bin/retrieve-calldata.ts +46 -39
  63. package/src/l1/calldata_retriever.ts +249 -379
  64. package/src/l1/data_retrieval.ts +23 -25
  65. package/src/l1/spire_proposer.ts +7 -15
  66. package/src/modules/data_source_base.ts +53 -92
  67. package/src/modules/data_store_updater.ts +49 -19
  68. package/src/modules/instrumentation.ts +29 -2
  69. package/src/modules/l1_synchronizer.ts +22 -13
  70. package/src/store/block_store.ts +85 -36
  71. package/src/store/kv_archiver_store.ts +35 -3
  72. package/src/store/l2_tips_cache.ts +89 -0
  73. package/src/store/log_store.ts +93 -31
  74. package/src/store/message_store.ts +1 -1
  75. package/src/test/fake_l1_state.ts +75 -17
  76. package/src/test/mock_archiver.ts +3 -2
  77. package/src/test/mock_l2_block_source.ts +164 -78
  78. package/src/test/mock_structs.ts +20 -6
@@ -20,6 +20,7 @@ import {
20
20
  Tag,
21
21
  TxScopedL2Log,
22
22
  } from '@aztec/stdlib/logs';
23
+ import { TxHash } from '@aztec/stdlib/tx';
23
24
 
24
25
  import type { BlockStore } from './block_store.js';
25
26
 
@@ -219,6 +220,7 @@ export class LogStore {
219
220
  .map((txEffect, txIndex) =>
220
221
  [
221
222
  numToUInt32BE(txIndex),
223
+ txEffect.txHash.toBuffer(),
222
224
  numToUInt32BE(txEffect.publicLogs.length),
223
225
  txEffect.publicLogs.map(log => log.toBuffer()),
224
226
  ].flat(),
@@ -242,6 +244,7 @@ export class LogStore {
242
244
  .map((txEffect, txIndex) =>
243
245
  [
244
246
  numToUInt32BE(txIndex),
247
+ txEffect.txHash.toBuffer(),
245
248
  numToUInt32BE(txEffect.contractClassLogs.length),
246
249
  txEffect.contractClassLogs.map(log => log.toBuffer()),
247
250
  ].flat(),
@@ -386,24 +389,33 @@ export class LogStore {
386
389
  }
387
390
 
388
391
  const buffer = (await this.#publicLogsByBlock.getAsync(blockNumber)) ?? Buffer.alloc(0);
389
- const publicLogsInBlock: [PublicLog[]] = [[]];
392
+ const publicLogsInBlock: { txHash: TxHash; logs: PublicLog[] }[] = [];
390
393
  const reader = new BufferReader(buffer);
391
394
 
392
395
  const blockHash = this.#unpackBlockHash(reader);
393
396
 
394
397
  while (reader.remainingBytes() > 0) {
395
398
  const indexOfTx = reader.readNumber();
399
+ const txHash = reader.readObject(TxHash);
396
400
  const numLogsInTx = reader.readNumber();
397
- publicLogsInBlock[indexOfTx] = [];
401
+ publicLogsInBlock[indexOfTx] = { txHash, logs: [] };
398
402
  for (let i = 0; i < numLogsInTx; i++) {
399
- publicLogsInBlock[indexOfTx].push(reader.readObject(PublicLog));
403
+ publicLogsInBlock[indexOfTx].logs.push(reader.readObject(PublicLog));
400
404
  }
401
405
  }
402
406
 
403
- const txLogs = publicLogsInBlock[txIndex];
407
+ const txData = publicLogsInBlock[txIndex];
404
408
 
405
409
  const logs: ExtendedPublicLog[] = [];
406
- const maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
410
+ const maxLogsHit = this.#accumulatePublicLogs(
411
+ logs,
412
+ blockNumber,
413
+ blockHash,
414
+ txIndex,
415
+ txData.txHash,
416
+ txData.logs,
417
+ filter,
418
+ );
407
419
 
408
420
  return { logs, maxLogsHit };
409
421
  }
@@ -424,22 +436,31 @@ export class LogStore {
424
436
 
425
437
  let maxLogsHit = false;
426
438
  loopOverBlocks: for await (const [blockNumber, logBuffer] of this.#publicLogsByBlock.entriesAsync({ start, end })) {
427
- const publicLogsInBlock: [PublicLog[]] = [[]];
439
+ const publicLogsInBlock: { txHash: TxHash; logs: PublicLog[] }[] = [];
428
440
  const reader = new BufferReader(logBuffer);
429
441
 
430
442
  const blockHash = this.#unpackBlockHash(reader);
431
443
 
432
444
  while (reader.remainingBytes() > 0) {
433
445
  const indexOfTx = reader.readNumber();
446
+ const txHash = reader.readObject(TxHash);
434
447
  const numLogsInTx = reader.readNumber();
435
- publicLogsInBlock[indexOfTx] = [];
448
+ publicLogsInBlock[indexOfTx] = { txHash, logs: [] };
436
449
  for (let i = 0; i < numLogsInTx; i++) {
437
- publicLogsInBlock[indexOfTx].push(reader.readObject(PublicLog));
450
+ publicLogsInBlock[indexOfTx].logs.push(reader.readObject(PublicLog));
438
451
  }
439
452
  }
440
453
  for (let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < publicLogsInBlock.length; txIndex++) {
441
- const txLogs = publicLogsInBlock[txIndex];
442
- maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
454
+ const txData = publicLogsInBlock[txIndex];
455
+ maxLogsHit = this.#accumulatePublicLogs(
456
+ logs,
457
+ blockNumber,
458
+ blockHash,
459
+ txIndex,
460
+ txData.txHash,
461
+ txData.logs,
462
+ filter,
463
+ );
443
464
  if (maxLogsHit) {
444
465
  this.#log.debug(`Max logs hit at block ${blockNumber}`);
445
466
  break loopOverBlocks;
@@ -475,24 +496,33 @@ export class LogStore {
475
496
  return { logs: [], maxLogsHit: false };
476
497
  }
477
498
  const contractClassLogsBuffer = (await this.#contractClassLogsByBlock.getAsync(blockNumber)) ?? Buffer.alloc(0);
478
- const contractClassLogsInBlock: [ContractClassLog[]] = [[]];
499
+ const contractClassLogsInBlock: { txHash: TxHash; logs: ContractClassLog[] }[] = [];
479
500
 
480
501
  const reader = new BufferReader(contractClassLogsBuffer);
481
502
  const blockHash = this.#unpackBlockHash(reader);
482
503
 
483
504
  while (reader.remainingBytes() > 0) {
484
505
  const indexOfTx = reader.readNumber();
506
+ const txHash = reader.readObject(TxHash);
485
507
  const numLogsInTx = reader.readNumber();
486
- contractClassLogsInBlock[indexOfTx] = [];
508
+ contractClassLogsInBlock[indexOfTx] = { txHash, logs: [] };
487
509
  for (let i = 0; i < numLogsInTx; i++) {
488
- contractClassLogsInBlock[indexOfTx].push(reader.readObject(ContractClassLog));
510
+ contractClassLogsInBlock[indexOfTx].logs.push(reader.readObject(ContractClassLog));
489
511
  }
490
512
  }
491
513
 
492
- const txLogs = contractClassLogsInBlock[txIndex];
514
+ const txData = contractClassLogsInBlock[txIndex];
493
515
 
494
516
  const logs: ExtendedContractClassLog[] = [];
495
- const maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
517
+ const maxLogsHit = this.#accumulateContractClassLogs(
518
+ logs,
519
+ blockNumber,
520
+ blockHash,
521
+ txIndex,
522
+ txData.txHash,
523
+ txData.logs,
524
+ filter,
525
+ );
496
526
 
497
527
  return { logs, maxLogsHit };
498
528
  }
@@ -516,20 +546,29 @@ export class LogStore {
516
546
  start,
517
547
  end,
518
548
  })) {
519
- const contractClassLogsInBlock: [ContractClassLog[]] = [[]];
549
+ const contractClassLogsInBlock: { txHash: TxHash; logs: ContractClassLog[] }[] = [];
520
550
  const reader = new BufferReader(logBuffer);
521
551
  const blockHash = this.#unpackBlockHash(reader);
522
552
  while (reader.remainingBytes() > 0) {
523
553
  const indexOfTx = reader.readNumber();
554
+ const txHash = reader.readObject(TxHash);
524
555
  const numLogsInTx = reader.readNumber();
525
- contractClassLogsInBlock[indexOfTx] = [];
556
+ contractClassLogsInBlock[indexOfTx] = { txHash, logs: [] };
526
557
  for (let i = 0; i < numLogsInTx; i++) {
527
- contractClassLogsInBlock[indexOfTx].push(reader.readObject(ContractClassLog));
558
+ contractClassLogsInBlock[indexOfTx].logs.push(reader.readObject(ContractClassLog));
528
559
  }
529
560
  }
530
561
  for (let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < contractClassLogsInBlock.length; txIndex++) {
531
- const txLogs = contractClassLogsInBlock[txIndex];
532
- maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
562
+ const txData = contractClassLogsInBlock[txIndex];
563
+ maxLogsHit = this.#accumulateContractClassLogs(
564
+ logs,
565
+ blockNumber,
566
+ blockHash,
567
+ txIndex,
568
+ txData.txHash,
569
+ txData.logs,
570
+ filter,
571
+ );
533
572
  if (maxLogsHit) {
534
573
  this.#log.debug(`Max logs hit at block ${blockNumber}`);
535
574
  break loopOverBlocks;
@@ -540,12 +579,13 @@ export class LogStore {
540
579
  return { logs, maxLogsHit };
541
580
  }
542
581
 
543
- #accumulateLogs(
544
- results: (ExtendedContractClassLog | ExtendedPublicLog)[],
582
+ #accumulatePublicLogs(
583
+ results: ExtendedPublicLog[],
545
584
  blockNumber: number,
546
585
  blockHash: BlockHash,
547
586
  txIndex: number,
548
- txLogs: (ContractClassLog | PublicLog)[],
587
+ txHash: TxHash,
588
+ txLogs: PublicLog[],
549
589
  filter: LogFilter = {},
550
590
  ): boolean {
551
591
  let maxLogsHit = false;
@@ -553,15 +593,37 @@ export class LogStore {
553
593
  for (; logIndex < txLogs.length; logIndex++) {
554
594
  const log = txLogs[logIndex];
555
595
  if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
556
- if (log instanceof ContractClassLog) {
557
- results.push(
558
- new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), blockHash, txIndex, logIndex), log),
559
- );
560
- } else if (log instanceof PublicLog) {
561
- results.push(new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txIndex, logIndex), log));
562
- } else {
563
- throw new Error('Unknown log type');
596
+ results.push(
597
+ new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
598
+ );
599
+
600
+ if (results.length >= this.#logsMaxPageSize) {
601
+ maxLogsHit = true;
602
+ break;
564
603
  }
604
+ }
605
+ }
606
+
607
+ return maxLogsHit;
608
+ }
609
+
610
+ #accumulateContractClassLogs(
611
+ results: ExtendedContractClassLog[],
612
+ blockNumber: number,
613
+ blockHash: BlockHash,
614
+ txIndex: number,
615
+ txHash: TxHash,
616
+ txLogs: ContractClassLog[],
617
+ filter: LogFilter = {},
618
+ ): boolean {
619
+ let maxLogsHit = false;
620
+ let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
621
+ for (; logIndex < txLogs.length; logIndex++) {
622
+ const log = txLogs[logIndex];
623
+ if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
624
+ results.push(
625
+ new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
626
+ );
565
627
 
566
628
  if (results.length >= this.#logsMaxPageSize) {
567
629
  maxLogsHit = true;
@@ -137,7 +137,7 @@ export class MessageStore {
137
137
  );
138
138
  }
139
139
 
140
- // Check the first message in a block has the correct index.
140
+ // Check the first message in a checkpoint has the correct index.
141
141
  if (
142
142
  (!lastMessage || message.checkpointNumber > lastMessage.checkpointNumber) &&
143
143
  message.index !== expectedStart
@@ -14,6 +14,7 @@ import { CommitteeAttestation, CommitteeAttestationsAndSigners, L2Block } from '
14
14
  import { Checkpoint } from '@aztec/stdlib/checkpoint';
15
15
  import { getSlotAtTimestamp } from '@aztec/stdlib/epoch-helpers';
16
16
  import { InboxLeaf } from '@aztec/stdlib/messaging';
17
+ import { ConsensusPayload, SignatureDomainSeparator } from '@aztec/stdlib/p2p';
17
18
  import {
18
19
  makeAndSignCommitteeAttestationsAndSigners,
19
20
  makeCheckpointAttestationFromCheckpoint,
@@ -22,7 +23,16 @@ import {
22
23
  import { AppendOnlyTreeSnapshot } from '@aztec/stdlib/trees';
23
24
 
24
25
  import { type MockProxy, mock } from 'jest-mock-extended';
25
- import { type FormattedBlock, type Transaction, encodeFunctionData, multicall3Abi, toHex } from 'viem';
26
+ import {
27
+ type AbiParameter,
28
+ type FormattedBlock,
29
+ type Transaction,
30
+ encodeAbiParameters,
31
+ encodeFunctionData,
32
+ keccak256,
33
+ multicall3Abi,
34
+ toHex,
35
+ } from 'viem';
26
36
 
27
37
  import { updateRollingHash } from '../structs/inbox_message.js';
28
38
 
@@ -87,6 +97,10 @@ type CheckpointData = {
87
97
  blobHashes: `0x${string}`[];
88
98
  blobs: Blob[];
89
99
  signers: Secp256k1Signer[];
100
+ /** Hash of the packed attestations, matching what the L1 event emits. */
101
+ attestationsHash: Buffer32;
102
+ /** Payload digest, matching what the L1 event emits. */
103
+ payloadDigest: Buffer32;
90
104
  /** If true, archiveAt will ignore it */
91
105
  pruned?: boolean;
92
106
  };
@@ -131,6 +145,7 @@ export class FakeL1State {
131
145
  private provenCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
132
146
  private targetCommitteeSize: number = 0;
133
147
  private version: bigint = 1n;
148
+ private canPruneResult: boolean = false;
134
149
 
135
150
  // Computed from checkpoints based on L1 block visibility
136
151
  private pendingCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
@@ -193,10 +208,10 @@ export class FakeL1State {
193
208
  // Store the messages internally so they match the checkpoint's inHash
194
209
  this.addMessages(checkpointNumber, messagesL1BlockNumber, messages);
195
210
 
196
- // Create the transaction and blobs
197
- const tx = this.makeRollupTx(checkpoint, signers);
198
- const blobHashes = this.makeVersionedBlobHashes(checkpoint);
199
- const blobs = this.makeBlobsFromCheckpoint(checkpoint);
211
+ // Create the transaction, blobs, and event hashes
212
+ const { tx, attestationsHash, payloadDigest } = await this.makeRollupTx(checkpoint, signers);
213
+ const blobHashes = await this.makeVersionedBlobHashes(checkpoint);
214
+ const blobs = await this.makeBlobsFromCheckpoint(checkpoint);
200
215
 
201
216
  // Store the checkpoint data
202
217
  this.checkpoints.push({
@@ -207,6 +222,8 @@ export class FakeL1State {
207
222
  blobHashes,
208
223
  blobs,
209
224
  signers,
225
+ attestationsHash,
226
+ payloadDigest,
210
227
  });
211
228
 
212
229
  // Update last archive for auto-chaining
@@ -276,6 +293,11 @@ export class FakeL1State {
276
293
  this.targetCommitteeSize = size;
277
294
  }
278
295
 
296
+ /** Sets whether the rollup contract would allow pruning at the next block. */
297
+ setCanPrune(value: boolean): void {
298
+ this.canPruneResult = value;
299
+ }
300
+
279
301
  /**
280
302
  * Removes all entries for a checkpoint number (simulates L1 reorg or prune).
281
303
  * Note: Does NOT remove messages for this checkpoint (use numL1ToL2Messages: 0 when re-adding).
@@ -384,6 +406,8 @@ export class FakeL1State {
384
406
  });
385
407
  });
386
408
 
409
+ mockRollup.canPruneAtTime.mockImplementation(() => Promise.resolve(this.canPruneResult));
410
+
387
411
  // Mock the wrapper method for fetching checkpoint events
388
412
  mockRollup.getCheckpointProposedEvents.mockImplementation((fromBlock: bigint, toBlock: bigint) =>
389
413
  Promise.resolve(this.getCheckpointProposedLogs(fromBlock, toBlock)),
@@ -502,10 +526,8 @@ export class FakeL1State {
502
526
  checkpointNumber: cpData.checkpointNumber,
503
527
  archive: cpData.checkpoint.archive.root,
504
528
  versionedBlobHashes: cpData.blobHashes.map(h => Buffer.from(h.slice(2), 'hex')),
505
- // These are intentionally undefined to skip hash validation in the archiver
506
- // (validation is skipped when these fields are falsy)
507
- payloadDigest: undefined,
508
- attestationsHash: undefined,
529
+ attestationsHash: cpData.attestationsHash,
530
+ payloadDigest: cpData.payloadDigest,
509
531
  },
510
532
  }));
511
533
  }
@@ -531,14 +553,17 @@ export class FakeL1State {
531
553
  }));
532
554
  }
533
555
 
534
- private makeRollupTx(checkpoint: Checkpoint, signers: Secp256k1Signer[]): Transaction {
556
+ private async makeRollupTx(
557
+ checkpoint: Checkpoint,
558
+ signers: Secp256k1Signer[],
559
+ ): Promise<{ tx: Transaction; attestationsHash: Buffer32; payloadDigest: Buffer32 }> {
535
560
  const attestations = signers
536
561
  .map(signer => makeCheckpointAttestationFromCheckpoint(checkpoint, signer))
537
562
  .map(attestation => CommitteeAttestation.fromSignature(attestation.signature))
538
563
  .map(committeeAttestation => committeeAttestation.toViem());
539
564
 
540
565
  const header = checkpoint.header.toViem();
541
- const blobInput = getPrefixedEthBlobCommitments(getBlobsPerL1Block(checkpoint.toBlobFields()));
566
+ const blobInput = getPrefixedEthBlobCommitments(await getBlobsPerL1Block(checkpoint.toBlobFields()));
542
567
  const archive = toHex(checkpoint.archive.root.toBuffer());
543
568
  const attestationsAndSigners = new CommitteeAttestationsAndSigners(
544
569
  attestations.map(attestation => CommitteeAttestation.fromViem(attestation)),
@@ -549,6 +574,8 @@ export class FakeL1State {
549
574
  signers[0],
550
575
  );
551
576
 
577
+ const packedAttestations = attestationsAndSigners.getPackedAttestations();
578
+
552
579
  const rollupInput = encodeFunctionData({
553
580
  abi: RollupAbi,
554
581
  functionName: 'propose',
@@ -558,7 +585,7 @@ export class FakeL1State {
558
585
  archive,
559
586
  oracleInput: { feeAssetPriceModifier: 0n },
560
587
  },
561
- attestationsAndSigners.getPackedAttestations(),
588
+ packedAttestations,
562
589
  attestationsAndSigners.getSigners().map(signer => signer.toString()),
563
590
  attestationsAndSignersSignature.toViemSignature(),
564
591
  blobInput,
@@ -579,21 +606,52 @@ export class FakeL1State {
579
606
  ],
580
607
  });
581
608
 
582
- return {
609
+ // Compute attestationsHash (same logic as CalldataRetriever)
610
+ const attestationsHash = Buffer32.fromString(
611
+ keccak256(encodeAbiParameters([this.getCommitteeAttestationsStructDef()], [packedAttestations])),
612
+ );
613
+
614
+ // Compute payloadDigest (same logic as CalldataRetriever)
615
+ const consensusPayload = ConsensusPayload.fromCheckpoint(checkpoint);
616
+ const payloadToSign = consensusPayload.getPayloadToSign(SignatureDomainSeparator.checkpointAttestation);
617
+ const payloadDigest = Buffer32.fromString(keccak256(payloadToSign));
618
+
619
+ const tx = {
583
620
  input: multiCallInput,
584
621
  hash: archive,
585
622
  blockHash: archive,
586
623
  to: MULTI_CALL_3_ADDRESS as `0x${string}`,
587
624
  } as Transaction<bigint, number>;
625
+
626
+ return { tx, attestationsHash, payloadDigest };
627
+ }
628
+
629
+ /** Extracts the CommitteeAttestations struct definition from RollupAbi for hash computation. */
630
+ private getCommitteeAttestationsStructDef(): AbiParameter {
631
+ const proposeFunction = RollupAbi.find(item => item.type === 'function' && item.name === 'propose') as
632
+ | { type: 'function'; name: string; inputs: readonly AbiParameter[] }
633
+ | undefined;
634
+
635
+ if (!proposeFunction) {
636
+ throw new Error('propose function not found in RollupAbi');
637
+ }
638
+
639
+ const attestationsParam = proposeFunction.inputs.find(param => param.name === '_attestations');
640
+ if (!attestationsParam) {
641
+ throw new Error('_attestations parameter not found in propose function');
642
+ }
643
+
644
+ const tupleParam = attestationsParam as unknown as { type: 'tuple'; components?: readonly AbiParameter[] };
645
+ return { type: 'tuple', components: tupleParam.components || [] } as AbiParameter;
588
646
  }
589
647
 
590
- private makeVersionedBlobHashes(checkpoint: Checkpoint): `0x${string}`[] {
591
- return getBlobsPerL1Block(checkpoint.toBlobFields()).map(
648
+ private async makeVersionedBlobHashes(checkpoint: Checkpoint): Promise<`0x${string}`[]> {
649
+ return (await getBlobsPerL1Block(checkpoint.toBlobFields())).map(
592
650
  b => `0x${b.getEthVersionedBlobHash().toString('hex')}` as `0x${string}`,
593
651
  );
594
652
  }
595
653
 
596
- private makeBlobsFromCheckpoint(checkpoint: Checkpoint): Blob[] {
597
- return getBlobsPerL1Block(checkpoint.toBlobFields());
654
+ private async makeBlobsFromCheckpoint(checkpoint: Checkpoint): Promise<Blob[]> {
655
+ return await getBlobsPerL1Block(checkpoint.toBlobFields());
598
656
  }
599
657
  }
@@ -56,8 +56,9 @@ export class MockPrefilledArchiver extends MockArchiver {
56
56
  }
57
57
 
58
58
  const fromBlock = this.l2Blocks.length;
59
- // TODO: Add L2 blocks and checkpoints separately once archiver has the apis for that.
60
- this.addProposedBlocks(this.prefilled.slice(fromBlock, fromBlock + numBlocks).flatMap(c => c.blocks));
59
+ const checkpointsToAdd = this.prefilled.slice(fromBlock, fromBlock + numBlocks);
60
+ this.addProposedBlocks(checkpointsToAdd.flatMap(c => c.blocks));
61
+ this.checkpointList.push(...checkpointsToAdd);
61
62
  return Promise.resolve();
62
63
  }
63
64
  }