@aztec/archiver 0.0.1-commit.ee80a48 → 0.0.1-commit.ef17749e1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/archiver.d.ts +7 -2
- package/dest/archiver.d.ts.map +1 -1
- package/dest/archiver.js +25 -96
- package/dest/factory.d.ts +2 -2
- package/dest/factory.d.ts.map +1 -1
- package/dest/factory.js +12 -8
- package/dest/index.d.ts +2 -1
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +1 -0
- package/dest/l1/bin/retrieve-calldata.js +36 -33
- package/dest/l1/calldata_retriever.d.ts +73 -50
- package/dest/l1/calldata_retriever.d.ts.map +1 -1
- package/dest/l1/calldata_retriever.js +190 -259
- package/dest/l1/data_retrieval.d.ts +9 -9
- package/dest/l1/data_retrieval.d.ts.map +1 -1
- package/dest/l1/data_retrieval.js +21 -19
- package/dest/l1/spire_proposer.d.ts +5 -5
- package/dest/l1/spire_proposer.d.ts.map +1 -1
- package/dest/l1/spire_proposer.js +9 -17
- package/dest/modules/data_source_base.d.ts +8 -3
- package/dest/modules/data_source_base.d.ts.map +1 -1
- package/dest/modules/data_source_base.js +28 -72
- package/dest/modules/data_store_updater.d.ts +13 -3
- package/dest/modules/data_store_updater.d.ts.map +1 -1
- package/dest/modules/data_store_updater.js +48 -19
- package/dest/modules/instrumentation.d.ts +15 -2
- package/dest/modules/instrumentation.d.ts.map +1 -1
- package/dest/modules/instrumentation.js +19 -2
- package/dest/modules/l1_synchronizer.d.ts +5 -8
- package/dest/modules/l1_synchronizer.d.ts.map +1 -1
- package/dest/modules/l1_synchronizer.js +16 -9
- package/dest/store/block_store.d.ts +18 -14
- package/dest/store/block_store.d.ts.map +1 -1
- package/dest/store/block_store.js +69 -17
- package/dest/store/kv_archiver_store.d.ts +18 -4
- package/dest/store/kv_archiver_store.d.ts.map +1 -1
- package/dest/store/kv_archiver_store.js +18 -0
- package/dest/store/l2_tips_cache.d.ts +19 -0
- package/dest/store/l2_tips_cache.d.ts.map +1 -0
- package/dest/store/l2_tips_cache.js +89 -0
- package/dest/store/log_store.d.ts +1 -1
- package/dest/store/log_store.d.ts.map +1 -1
- package/dest/store/log_store.js +55 -35
- package/dest/store/message_store.js +1 -1
- package/dest/test/fake_l1_state.d.ts +6 -1
- package/dest/test/fake_l1_state.d.ts.map +1 -1
- package/dest/test/fake_l1_state.js +56 -18
- package/dest/test/mock_archiver.d.ts +1 -1
- package/dest/test/mock_archiver.d.ts.map +1 -1
- package/dest/test/mock_archiver.js +3 -2
- package/dest/test/mock_l2_block_source.d.ts +19 -3
- package/dest/test/mock_l2_block_source.d.ts.map +1 -1
- package/dest/test/mock_l2_block_source.js +128 -82
- package/dest/test/mock_structs.d.ts +4 -1
- package/dest/test/mock_structs.d.ts.map +1 -1
- package/dest/test/mock_structs.js +13 -1
- package/package.json +13 -13
- package/src/archiver.ts +39 -116
- package/src/factory.ts +27 -10
- package/src/index.ts +1 -0
- package/src/l1/README.md +25 -68
- package/src/l1/bin/retrieve-calldata.ts +46 -39
- package/src/l1/calldata_retriever.ts +249 -379
- package/src/l1/data_retrieval.ts +23 -25
- package/src/l1/spire_proposer.ts +7 -15
- package/src/modules/data_source_base.ts +53 -92
- package/src/modules/data_store_updater.ts +49 -19
- package/src/modules/instrumentation.ts +29 -2
- package/src/modules/l1_synchronizer.ts +22 -13
- package/src/store/block_store.ts +85 -36
- package/src/store/kv_archiver_store.ts +35 -3
- package/src/store/l2_tips_cache.ts +89 -0
- package/src/store/log_store.ts +93 -31
- package/src/store/message_store.ts +1 -1
- package/src/test/fake_l1_state.ts +75 -17
- package/src/test/mock_archiver.ts +3 -2
- package/src/test/mock_l2_block_source.ts +164 -78
- package/src/test/mock_structs.ts +20 -6
package/src/store/log_store.ts
CHANGED
|
@@ -20,6 +20,7 @@ import {
|
|
|
20
20
|
Tag,
|
|
21
21
|
TxScopedL2Log,
|
|
22
22
|
} from '@aztec/stdlib/logs';
|
|
23
|
+
import { TxHash } from '@aztec/stdlib/tx';
|
|
23
24
|
|
|
24
25
|
import type { BlockStore } from './block_store.js';
|
|
25
26
|
|
|
@@ -219,6 +220,7 @@ export class LogStore {
|
|
|
219
220
|
.map((txEffect, txIndex) =>
|
|
220
221
|
[
|
|
221
222
|
numToUInt32BE(txIndex),
|
|
223
|
+
txEffect.txHash.toBuffer(),
|
|
222
224
|
numToUInt32BE(txEffect.publicLogs.length),
|
|
223
225
|
txEffect.publicLogs.map(log => log.toBuffer()),
|
|
224
226
|
].flat(),
|
|
@@ -242,6 +244,7 @@ export class LogStore {
|
|
|
242
244
|
.map((txEffect, txIndex) =>
|
|
243
245
|
[
|
|
244
246
|
numToUInt32BE(txIndex),
|
|
247
|
+
txEffect.txHash.toBuffer(),
|
|
245
248
|
numToUInt32BE(txEffect.contractClassLogs.length),
|
|
246
249
|
txEffect.contractClassLogs.map(log => log.toBuffer()),
|
|
247
250
|
].flat(),
|
|
@@ -386,24 +389,33 @@ export class LogStore {
|
|
|
386
389
|
}
|
|
387
390
|
|
|
388
391
|
const buffer = (await this.#publicLogsByBlock.getAsync(blockNumber)) ?? Buffer.alloc(0);
|
|
389
|
-
const publicLogsInBlock:
|
|
392
|
+
const publicLogsInBlock: { txHash: TxHash; logs: PublicLog[] }[] = [];
|
|
390
393
|
const reader = new BufferReader(buffer);
|
|
391
394
|
|
|
392
395
|
const blockHash = this.#unpackBlockHash(reader);
|
|
393
396
|
|
|
394
397
|
while (reader.remainingBytes() > 0) {
|
|
395
398
|
const indexOfTx = reader.readNumber();
|
|
399
|
+
const txHash = reader.readObject(TxHash);
|
|
396
400
|
const numLogsInTx = reader.readNumber();
|
|
397
|
-
publicLogsInBlock[indexOfTx] = [];
|
|
401
|
+
publicLogsInBlock[indexOfTx] = { txHash, logs: [] };
|
|
398
402
|
for (let i = 0; i < numLogsInTx; i++) {
|
|
399
|
-
publicLogsInBlock[indexOfTx].push(reader.readObject(PublicLog));
|
|
403
|
+
publicLogsInBlock[indexOfTx].logs.push(reader.readObject(PublicLog));
|
|
400
404
|
}
|
|
401
405
|
}
|
|
402
406
|
|
|
403
|
-
const
|
|
407
|
+
const txData = publicLogsInBlock[txIndex];
|
|
404
408
|
|
|
405
409
|
const logs: ExtendedPublicLog[] = [];
|
|
406
|
-
const maxLogsHit = this.#
|
|
410
|
+
const maxLogsHit = this.#accumulatePublicLogs(
|
|
411
|
+
logs,
|
|
412
|
+
blockNumber,
|
|
413
|
+
blockHash,
|
|
414
|
+
txIndex,
|
|
415
|
+
txData.txHash,
|
|
416
|
+
txData.logs,
|
|
417
|
+
filter,
|
|
418
|
+
);
|
|
407
419
|
|
|
408
420
|
return { logs, maxLogsHit };
|
|
409
421
|
}
|
|
@@ -424,22 +436,31 @@ export class LogStore {
|
|
|
424
436
|
|
|
425
437
|
let maxLogsHit = false;
|
|
426
438
|
loopOverBlocks: for await (const [blockNumber, logBuffer] of this.#publicLogsByBlock.entriesAsync({ start, end })) {
|
|
427
|
-
const publicLogsInBlock:
|
|
439
|
+
const publicLogsInBlock: { txHash: TxHash; logs: PublicLog[] }[] = [];
|
|
428
440
|
const reader = new BufferReader(logBuffer);
|
|
429
441
|
|
|
430
442
|
const blockHash = this.#unpackBlockHash(reader);
|
|
431
443
|
|
|
432
444
|
while (reader.remainingBytes() > 0) {
|
|
433
445
|
const indexOfTx = reader.readNumber();
|
|
446
|
+
const txHash = reader.readObject(TxHash);
|
|
434
447
|
const numLogsInTx = reader.readNumber();
|
|
435
|
-
publicLogsInBlock[indexOfTx] = [];
|
|
448
|
+
publicLogsInBlock[indexOfTx] = { txHash, logs: [] };
|
|
436
449
|
for (let i = 0; i < numLogsInTx; i++) {
|
|
437
|
-
publicLogsInBlock[indexOfTx].push(reader.readObject(PublicLog));
|
|
450
|
+
publicLogsInBlock[indexOfTx].logs.push(reader.readObject(PublicLog));
|
|
438
451
|
}
|
|
439
452
|
}
|
|
440
453
|
for (let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < publicLogsInBlock.length; txIndex++) {
|
|
441
|
-
const
|
|
442
|
-
maxLogsHit = this.#
|
|
454
|
+
const txData = publicLogsInBlock[txIndex];
|
|
455
|
+
maxLogsHit = this.#accumulatePublicLogs(
|
|
456
|
+
logs,
|
|
457
|
+
blockNumber,
|
|
458
|
+
blockHash,
|
|
459
|
+
txIndex,
|
|
460
|
+
txData.txHash,
|
|
461
|
+
txData.logs,
|
|
462
|
+
filter,
|
|
463
|
+
);
|
|
443
464
|
if (maxLogsHit) {
|
|
444
465
|
this.#log.debug(`Max logs hit at block ${blockNumber}`);
|
|
445
466
|
break loopOverBlocks;
|
|
@@ -475,24 +496,33 @@ export class LogStore {
|
|
|
475
496
|
return { logs: [], maxLogsHit: false };
|
|
476
497
|
}
|
|
477
498
|
const contractClassLogsBuffer = (await this.#contractClassLogsByBlock.getAsync(blockNumber)) ?? Buffer.alloc(0);
|
|
478
|
-
const contractClassLogsInBlock:
|
|
499
|
+
const contractClassLogsInBlock: { txHash: TxHash; logs: ContractClassLog[] }[] = [];
|
|
479
500
|
|
|
480
501
|
const reader = new BufferReader(contractClassLogsBuffer);
|
|
481
502
|
const blockHash = this.#unpackBlockHash(reader);
|
|
482
503
|
|
|
483
504
|
while (reader.remainingBytes() > 0) {
|
|
484
505
|
const indexOfTx = reader.readNumber();
|
|
506
|
+
const txHash = reader.readObject(TxHash);
|
|
485
507
|
const numLogsInTx = reader.readNumber();
|
|
486
|
-
contractClassLogsInBlock[indexOfTx] = [];
|
|
508
|
+
contractClassLogsInBlock[indexOfTx] = { txHash, logs: [] };
|
|
487
509
|
for (let i = 0; i < numLogsInTx; i++) {
|
|
488
|
-
contractClassLogsInBlock[indexOfTx].push(reader.readObject(ContractClassLog));
|
|
510
|
+
contractClassLogsInBlock[indexOfTx].logs.push(reader.readObject(ContractClassLog));
|
|
489
511
|
}
|
|
490
512
|
}
|
|
491
513
|
|
|
492
|
-
const
|
|
514
|
+
const txData = contractClassLogsInBlock[txIndex];
|
|
493
515
|
|
|
494
516
|
const logs: ExtendedContractClassLog[] = [];
|
|
495
|
-
const maxLogsHit = this.#
|
|
517
|
+
const maxLogsHit = this.#accumulateContractClassLogs(
|
|
518
|
+
logs,
|
|
519
|
+
blockNumber,
|
|
520
|
+
blockHash,
|
|
521
|
+
txIndex,
|
|
522
|
+
txData.txHash,
|
|
523
|
+
txData.logs,
|
|
524
|
+
filter,
|
|
525
|
+
);
|
|
496
526
|
|
|
497
527
|
return { logs, maxLogsHit };
|
|
498
528
|
}
|
|
@@ -516,20 +546,29 @@ export class LogStore {
|
|
|
516
546
|
start,
|
|
517
547
|
end,
|
|
518
548
|
})) {
|
|
519
|
-
const contractClassLogsInBlock:
|
|
549
|
+
const contractClassLogsInBlock: { txHash: TxHash; logs: ContractClassLog[] }[] = [];
|
|
520
550
|
const reader = new BufferReader(logBuffer);
|
|
521
551
|
const blockHash = this.#unpackBlockHash(reader);
|
|
522
552
|
while (reader.remainingBytes() > 0) {
|
|
523
553
|
const indexOfTx = reader.readNumber();
|
|
554
|
+
const txHash = reader.readObject(TxHash);
|
|
524
555
|
const numLogsInTx = reader.readNumber();
|
|
525
|
-
contractClassLogsInBlock[indexOfTx] = [];
|
|
556
|
+
contractClassLogsInBlock[indexOfTx] = { txHash, logs: [] };
|
|
526
557
|
for (let i = 0; i < numLogsInTx; i++) {
|
|
527
|
-
contractClassLogsInBlock[indexOfTx].push(reader.readObject(ContractClassLog));
|
|
558
|
+
contractClassLogsInBlock[indexOfTx].logs.push(reader.readObject(ContractClassLog));
|
|
528
559
|
}
|
|
529
560
|
}
|
|
530
561
|
for (let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < contractClassLogsInBlock.length; txIndex++) {
|
|
531
|
-
const
|
|
532
|
-
maxLogsHit = this.#
|
|
562
|
+
const txData = contractClassLogsInBlock[txIndex];
|
|
563
|
+
maxLogsHit = this.#accumulateContractClassLogs(
|
|
564
|
+
logs,
|
|
565
|
+
blockNumber,
|
|
566
|
+
blockHash,
|
|
567
|
+
txIndex,
|
|
568
|
+
txData.txHash,
|
|
569
|
+
txData.logs,
|
|
570
|
+
filter,
|
|
571
|
+
);
|
|
533
572
|
if (maxLogsHit) {
|
|
534
573
|
this.#log.debug(`Max logs hit at block ${blockNumber}`);
|
|
535
574
|
break loopOverBlocks;
|
|
@@ -540,12 +579,13 @@ export class LogStore {
|
|
|
540
579
|
return { logs, maxLogsHit };
|
|
541
580
|
}
|
|
542
581
|
|
|
543
|
-
#
|
|
544
|
-
results:
|
|
582
|
+
#accumulatePublicLogs(
|
|
583
|
+
results: ExtendedPublicLog[],
|
|
545
584
|
blockNumber: number,
|
|
546
585
|
blockHash: BlockHash,
|
|
547
586
|
txIndex: number,
|
|
548
|
-
|
|
587
|
+
txHash: TxHash,
|
|
588
|
+
txLogs: PublicLog[],
|
|
549
589
|
filter: LogFilter = {},
|
|
550
590
|
): boolean {
|
|
551
591
|
let maxLogsHit = false;
|
|
@@ -553,15 +593,37 @@ export class LogStore {
|
|
|
553
593
|
for (; logIndex < txLogs.length; logIndex++) {
|
|
554
594
|
const log = txLogs[logIndex];
|
|
555
595
|
if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
|
|
556
|
-
|
|
557
|
-
|
|
558
|
-
|
|
559
|
-
|
|
560
|
-
|
|
561
|
-
|
|
562
|
-
|
|
563
|
-
throw new Error('Unknown log type');
|
|
596
|
+
results.push(
|
|
597
|
+
new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
|
|
598
|
+
);
|
|
599
|
+
|
|
600
|
+
if (results.length >= this.#logsMaxPageSize) {
|
|
601
|
+
maxLogsHit = true;
|
|
602
|
+
break;
|
|
564
603
|
}
|
|
604
|
+
}
|
|
605
|
+
}
|
|
606
|
+
|
|
607
|
+
return maxLogsHit;
|
|
608
|
+
}
|
|
609
|
+
|
|
610
|
+
#accumulateContractClassLogs(
|
|
611
|
+
results: ExtendedContractClassLog[],
|
|
612
|
+
blockNumber: number,
|
|
613
|
+
blockHash: BlockHash,
|
|
614
|
+
txIndex: number,
|
|
615
|
+
txHash: TxHash,
|
|
616
|
+
txLogs: ContractClassLog[],
|
|
617
|
+
filter: LogFilter = {},
|
|
618
|
+
): boolean {
|
|
619
|
+
let maxLogsHit = false;
|
|
620
|
+
let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
|
|
621
|
+
for (; logIndex < txLogs.length; logIndex++) {
|
|
622
|
+
const log = txLogs[logIndex];
|
|
623
|
+
if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
|
|
624
|
+
results.push(
|
|
625
|
+
new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
|
|
626
|
+
);
|
|
565
627
|
|
|
566
628
|
if (results.length >= this.#logsMaxPageSize) {
|
|
567
629
|
maxLogsHit = true;
|
|
@@ -137,7 +137,7 @@ export class MessageStore {
|
|
|
137
137
|
);
|
|
138
138
|
}
|
|
139
139
|
|
|
140
|
-
// Check the first message in a
|
|
140
|
+
// Check the first message in a checkpoint has the correct index.
|
|
141
141
|
if (
|
|
142
142
|
(!lastMessage || message.checkpointNumber > lastMessage.checkpointNumber) &&
|
|
143
143
|
message.index !== expectedStart
|
|
@@ -14,6 +14,7 @@ import { CommitteeAttestation, CommitteeAttestationsAndSigners, L2Block } from '
|
|
|
14
14
|
import { Checkpoint } from '@aztec/stdlib/checkpoint';
|
|
15
15
|
import { getSlotAtTimestamp } from '@aztec/stdlib/epoch-helpers';
|
|
16
16
|
import { InboxLeaf } from '@aztec/stdlib/messaging';
|
|
17
|
+
import { ConsensusPayload, SignatureDomainSeparator } from '@aztec/stdlib/p2p';
|
|
17
18
|
import {
|
|
18
19
|
makeAndSignCommitteeAttestationsAndSigners,
|
|
19
20
|
makeCheckpointAttestationFromCheckpoint,
|
|
@@ -22,7 +23,16 @@ import {
|
|
|
22
23
|
import { AppendOnlyTreeSnapshot } from '@aztec/stdlib/trees';
|
|
23
24
|
|
|
24
25
|
import { type MockProxy, mock } from 'jest-mock-extended';
|
|
25
|
-
import {
|
|
26
|
+
import {
|
|
27
|
+
type AbiParameter,
|
|
28
|
+
type FormattedBlock,
|
|
29
|
+
type Transaction,
|
|
30
|
+
encodeAbiParameters,
|
|
31
|
+
encodeFunctionData,
|
|
32
|
+
keccak256,
|
|
33
|
+
multicall3Abi,
|
|
34
|
+
toHex,
|
|
35
|
+
} from 'viem';
|
|
26
36
|
|
|
27
37
|
import { updateRollingHash } from '../structs/inbox_message.js';
|
|
28
38
|
|
|
@@ -87,6 +97,10 @@ type CheckpointData = {
|
|
|
87
97
|
blobHashes: `0x${string}`[];
|
|
88
98
|
blobs: Blob[];
|
|
89
99
|
signers: Secp256k1Signer[];
|
|
100
|
+
/** Hash of the packed attestations, matching what the L1 event emits. */
|
|
101
|
+
attestationsHash: Buffer32;
|
|
102
|
+
/** Payload digest, matching what the L1 event emits. */
|
|
103
|
+
payloadDigest: Buffer32;
|
|
90
104
|
/** If true, archiveAt will ignore it */
|
|
91
105
|
pruned?: boolean;
|
|
92
106
|
};
|
|
@@ -131,6 +145,7 @@ export class FakeL1State {
|
|
|
131
145
|
private provenCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
|
|
132
146
|
private targetCommitteeSize: number = 0;
|
|
133
147
|
private version: bigint = 1n;
|
|
148
|
+
private canPruneResult: boolean = false;
|
|
134
149
|
|
|
135
150
|
// Computed from checkpoints based on L1 block visibility
|
|
136
151
|
private pendingCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
|
|
@@ -193,10 +208,10 @@ export class FakeL1State {
|
|
|
193
208
|
// Store the messages internally so they match the checkpoint's inHash
|
|
194
209
|
this.addMessages(checkpointNumber, messagesL1BlockNumber, messages);
|
|
195
210
|
|
|
196
|
-
// Create the transaction and
|
|
197
|
-
const tx = this.makeRollupTx(checkpoint, signers);
|
|
198
|
-
const blobHashes = this.makeVersionedBlobHashes(checkpoint);
|
|
199
|
-
const blobs = this.makeBlobsFromCheckpoint(checkpoint);
|
|
211
|
+
// Create the transaction, blobs, and event hashes
|
|
212
|
+
const { tx, attestationsHash, payloadDigest } = await this.makeRollupTx(checkpoint, signers);
|
|
213
|
+
const blobHashes = await this.makeVersionedBlobHashes(checkpoint);
|
|
214
|
+
const blobs = await this.makeBlobsFromCheckpoint(checkpoint);
|
|
200
215
|
|
|
201
216
|
// Store the checkpoint data
|
|
202
217
|
this.checkpoints.push({
|
|
@@ -207,6 +222,8 @@ export class FakeL1State {
|
|
|
207
222
|
blobHashes,
|
|
208
223
|
blobs,
|
|
209
224
|
signers,
|
|
225
|
+
attestationsHash,
|
|
226
|
+
payloadDigest,
|
|
210
227
|
});
|
|
211
228
|
|
|
212
229
|
// Update last archive for auto-chaining
|
|
@@ -276,6 +293,11 @@ export class FakeL1State {
|
|
|
276
293
|
this.targetCommitteeSize = size;
|
|
277
294
|
}
|
|
278
295
|
|
|
296
|
+
/** Sets whether the rollup contract would allow pruning at the next block. */
|
|
297
|
+
setCanPrune(value: boolean): void {
|
|
298
|
+
this.canPruneResult = value;
|
|
299
|
+
}
|
|
300
|
+
|
|
279
301
|
/**
|
|
280
302
|
* Removes all entries for a checkpoint number (simulates L1 reorg or prune).
|
|
281
303
|
* Note: Does NOT remove messages for this checkpoint (use numL1ToL2Messages: 0 when re-adding).
|
|
@@ -384,6 +406,8 @@ export class FakeL1State {
|
|
|
384
406
|
});
|
|
385
407
|
});
|
|
386
408
|
|
|
409
|
+
mockRollup.canPruneAtTime.mockImplementation(() => Promise.resolve(this.canPruneResult));
|
|
410
|
+
|
|
387
411
|
// Mock the wrapper method for fetching checkpoint events
|
|
388
412
|
mockRollup.getCheckpointProposedEvents.mockImplementation((fromBlock: bigint, toBlock: bigint) =>
|
|
389
413
|
Promise.resolve(this.getCheckpointProposedLogs(fromBlock, toBlock)),
|
|
@@ -502,10 +526,8 @@ export class FakeL1State {
|
|
|
502
526
|
checkpointNumber: cpData.checkpointNumber,
|
|
503
527
|
archive: cpData.checkpoint.archive.root,
|
|
504
528
|
versionedBlobHashes: cpData.blobHashes.map(h => Buffer.from(h.slice(2), 'hex')),
|
|
505
|
-
|
|
506
|
-
|
|
507
|
-
payloadDigest: undefined,
|
|
508
|
-
attestationsHash: undefined,
|
|
529
|
+
attestationsHash: cpData.attestationsHash,
|
|
530
|
+
payloadDigest: cpData.payloadDigest,
|
|
509
531
|
},
|
|
510
532
|
}));
|
|
511
533
|
}
|
|
@@ -531,14 +553,17 @@ export class FakeL1State {
|
|
|
531
553
|
}));
|
|
532
554
|
}
|
|
533
555
|
|
|
534
|
-
private makeRollupTx(
|
|
556
|
+
private async makeRollupTx(
|
|
557
|
+
checkpoint: Checkpoint,
|
|
558
|
+
signers: Secp256k1Signer[],
|
|
559
|
+
): Promise<{ tx: Transaction; attestationsHash: Buffer32; payloadDigest: Buffer32 }> {
|
|
535
560
|
const attestations = signers
|
|
536
561
|
.map(signer => makeCheckpointAttestationFromCheckpoint(checkpoint, signer))
|
|
537
562
|
.map(attestation => CommitteeAttestation.fromSignature(attestation.signature))
|
|
538
563
|
.map(committeeAttestation => committeeAttestation.toViem());
|
|
539
564
|
|
|
540
565
|
const header = checkpoint.header.toViem();
|
|
541
|
-
const blobInput = getPrefixedEthBlobCommitments(getBlobsPerL1Block(checkpoint.toBlobFields()));
|
|
566
|
+
const blobInput = getPrefixedEthBlobCommitments(await getBlobsPerL1Block(checkpoint.toBlobFields()));
|
|
542
567
|
const archive = toHex(checkpoint.archive.root.toBuffer());
|
|
543
568
|
const attestationsAndSigners = new CommitteeAttestationsAndSigners(
|
|
544
569
|
attestations.map(attestation => CommitteeAttestation.fromViem(attestation)),
|
|
@@ -549,6 +574,8 @@ export class FakeL1State {
|
|
|
549
574
|
signers[0],
|
|
550
575
|
);
|
|
551
576
|
|
|
577
|
+
const packedAttestations = attestationsAndSigners.getPackedAttestations();
|
|
578
|
+
|
|
552
579
|
const rollupInput = encodeFunctionData({
|
|
553
580
|
abi: RollupAbi,
|
|
554
581
|
functionName: 'propose',
|
|
@@ -558,7 +585,7 @@ export class FakeL1State {
|
|
|
558
585
|
archive,
|
|
559
586
|
oracleInput: { feeAssetPriceModifier: 0n },
|
|
560
587
|
},
|
|
561
|
-
|
|
588
|
+
packedAttestations,
|
|
562
589
|
attestationsAndSigners.getSigners().map(signer => signer.toString()),
|
|
563
590
|
attestationsAndSignersSignature.toViemSignature(),
|
|
564
591
|
blobInput,
|
|
@@ -579,21 +606,52 @@ export class FakeL1State {
|
|
|
579
606
|
],
|
|
580
607
|
});
|
|
581
608
|
|
|
582
|
-
|
|
609
|
+
// Compute attestationsHash (same logic as CalldataRetriever)
|
|
610
|
+
const attestationsHash = Buffer32.fromString(
|
|
611
|
+
keccak256(encodeAbiParameters([this.getCommitteeAttestationsStructDef()], [packedAttestations])),
|
|
612
|
+
);
|
|
613
|
+
|
|
614
|
+
// Compute payloadDigest (same logic as CalldataRetriever)
|
|
615
|
+
const consensusPayload = ConsensusPayload.fromCheckpoint(checkpoint);
|
|
616
|
+
const payloadToSign = consensusPayload.getPayloadToSign(SignatureDomainSeparator.checkpointAttestation);
|
|
617
|
+
const payloadDigest = Buffer32.fromString(keccak256(payloadToSign));
|
|
618
|
+
|
|
619
|
+
const tx = {
|
|
583
620
|
input: multiCallInput,
|
|
584
621
|
hash: archive,
|
|
585
622
|
blockHash: archive,
|
|
586
623
|
to: MULTI_CALL_3_ADDRESS as `0x${string}`,
|
|
587
624
|
} as Transaction<bigint, number>;
|
|
625
|
+
|
|
626
|
+
return { tx, attestationsHash, payloadDigest };
|
|
627
|
+
}
|
|
628
|
+
|
|
629
|
+
/** Extracts the CommitteeAttestations struct definition from RollupAbi for hash computation. */
|
|
630
|
+
private getCommitteeAttestationsStructDef(): AbiParameter {
|
|
631
|
+
const proposeFunction = RollupAbi.find(item => item.type === 'function' && item.name === 'propose') as
|
|
632
|
+
| { type: 'function'; name: string; inputs: readonly AbiParameter[] }
|
|
633
|
+
| undefined;
|
|
634
|
+
|
|
635
|
+
if (!proposeFunction) {
|
|
636
|
+
throw new Error('propose function not found in RollupAbi');
|
|
637
|
+
}
|
|
638
|
+
|
|
639
|
+
const attestationsParam = proposeFunction.inputs.find(param => param.name === '_attestations');
|
|
640
|
+
if (!attestationsParam) {
|
|
641
|
+
throw new Error('_attestations parameter not found in propose function');
|
|
642
|
+
}
|
|
643
|
+
|
|
644
|
+
const tupleParam = attestationsParam as unknown as { type: 'tuple'; components?: readonly AbiParameter[] };
|
|
645
|
+
return { type: 'tuple', components: tupleParam.components || [] } as AbiParameter;
|
|
588
646
|
}
|
|
589
647
|
|
|
590
|
-
private makeVersionedBlobHashes(checkpoint: Checkpoint):
|
|
591
|
-
return getBlobsPerL1Block(checkpoint.toBlobFields()).map(
|
|
648
|
+
private async makeVersionedBlobHashes(checkpoint: Checkpoint): Promise<`0x${string}`[]> {
|
|
649
|
+
return (await getBlobsPerL1Block(checkpoint.toBlobFields())).map(
|
|
592
650
|
b => `0x${b.getEthVersionedBlobHash().toString('hex')}` as `0x${string}`,
|
|
593
651
|
);
|
|
594
652
|
}
|
|
595
653
|
|
|
596
|
-
private makeBlobsFromCheckpoint(checkpoint: Checkpoint): Blob[] {
|
|
597
|
-
return getBlobsPerL1Block(checkpoint.toBlobFields());
|
|
654
|
+
private async makeBlobsFromCheckpoint(checkpoint: Checkpoint): Promise<Blob[]> {
|
|
655
|
+
return await getBlobsPerL1Block(checkpoint.toBlobFields());
|
|
598
656
|
}
|
|
599
657
|
}
|
|
@@ -56,8 +56,9 @@ export class MockPrefilledArchiver extends MockArchiver {
|
|
|
56
56
|
}
|
|
57
57
|
|
|
58
58
|
const fromBlock = this.l2Blocks.length;
|
|
59
|
-
|
|
60
|
-
this.addProposedBlocks(
|
|
59
|
+
const checkpointsToAdd = this.prefilled.slice(fromBlock, fromBlock + numBlocks);
|
|
60
|
+
this.addProposedBlocks(checkpointsToAdd.flatMap(c => c.blocks));
|
|
61
|
+
this.checkpointList.push(...checkpointsToAdd);
|
|
61
62
|
return Promise.resolve();
|
|
62
63
|
}
|
|
63
64
|
}
|