@aztec/archiver 0.0.1-commit.c949de6bc → 0.0.1-commit.cbf2c2d5d
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/archiver.d.ts +3 -4
- package/dest/archiver.d.ts.map +1 -1
- package/dest/archiver.js +67 -23
- package/dest/config.d.ts +3 -3
- package/dest/config.d.ts.map +1 -1
- package/dest/config.js +2 -1
- package/dest/errors.d.ts +21 -9
- package/dest/errors.d.ts.map +1 -1
- package/dest/errors.js +27 -14
- package/dest/factory.d.ts +4 -5
- package/dest/factory.d.ts.map +1 -1
- package/dest/factory.js +24 -21
- package/dest/modules/data_source_base.d.ts +5 -5
- package/dest/modules/data_source_base.d.ts.map +1 -1
- package/dest/modules/data_source_base.js +5 -5
- package/dest/modules/data_store_updater.d.ts +17 -12
- package/dest/modules/data_store_updater.d.ts.map +1 -1
- package/dest/modules/data_store_updater.js +78 -77
- package/dest/modules/l1_synchronizer.d.ts +2 -1
- package/dest/modules/l1_synchronizer.d.ts.map +1 -1
- package/dest/modules/l1_synchronizer.js +37 -7
- package/dest/store/block_store.d.ts +12 -13
- package/dest/store/block_store.d.ts.map +1 -1
- package/dest/store/block_store.js +61 -61
- package/dest/store/contract_class_store.d.ts +2 -3
- package/dest/store/contract_class_store.d.ts.map +1 -1
- package/dest/store/contract_class_store.js +7 -67
- package/dest/store/contract_instance_store.d.ts +1 -1
- package/dest/store/contract_instance_store.d.ts.map +1 -1
- package/dest/store/contract_instance_store.js +6 -2
- package/dest/store/kv_archiver_store.d.ts +28 -18
- package/dest/store/kv_archiver_store.d.ts.map +1 -1
- package/dest/store/kv_archiver_store.js +34 -21
- package/dest/store/log_store.d.ts +6 -3
- package/dest/store/log_store.d.ts.map +1 -1
- package/dest/store/log_store.js +93 -16
- package/dest/store/message_store.d.ts +5 -1
- package/dest/store/message_store.d.ts.map +1 -1
- package/dest/store/message_store.js +14 -1
- package/dest/test/fake_l1_state.d.ts +8 -1
- package/dest/test/fake_l1_state.d.ts.map +1 -1
- package/dest/test/fake_l1_state.js +39 -5
- package/dest/test/mock_l2_block_source.d.ts +4 -3
- package/dest/test/mock_l2_block_source.d.ts.map +1 -1
- package/dest/test/mock_l2_block_source.js +7 -4
- package/dest/test/mock_structs.d.ts +4 -1
- package/dest/test/mock_structs.d.ts.map +1 -1
- package/dest/test/mock_structs.js +13 -1
- package/dest/test/noop_l1_archiver.d.ts +4 -1
- package/dest/test/noop_l1_archiver.d.ts.map +1 -1
- package/dest/test/noop_l1_archiver.js +5 -1
- package/package.json +13 -13
- package/src/archiver.ts +80 -23
- package/src/config.ts +8 -1
- package/src/errors.ts +40 -24
- package/src/factory.ts +23 -15
- package/src/modules/data_source_base.ts +11 -6
- package/src/modules/data_store_updater.ts +83 -107
- package/src/modules/l1_synchronizer.ts +47 -13
- package/src/store/block_store.ts +72 -69
- package/src/store/contract_class_store.ts +8 -106
- package/src/store/contract_instance_store.ts +8 -5
- package/src/store/kv_archiver_store.ts +43 -32
- package/src/store/log_store.ts +126 -27
- package/src/store/message_store.ts +20 -1
- package/src/test/fake_l1_state.ts +50 -9
- package/src/test/mock_l2_block_source.ts +15 -3
- package/src/test/mock_structs.ts +20 -6
- package/src/test/noop_l1_archiver.ts +7 -1
package/src/store/log_store.ts
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { INITIAL_L2_BLOCK_NUM } from '@aztec/constants';
|
|
2
2
|
import { BlockNumber } from '@aztec/foundation/branded-types';
|
|
3
|
-
import { filterAsync } from '@aztec/foundation/collection';
|
|
3
|
+
import { compactArray, filterAsync } from '@aztec/foundation/collection';
|
|
4
4
|
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
5
5
|
import { createLogger } from '@aztec/foundation/log';
|
|
6
6
|
import { BufferReader, numToUInt32BE } from '@aztec/foundation/serialize';
|
|
@@ -22,6 +22,7 @@ import {
|
|
|
22
22
|
} from '@aztec/stdlib/logs';
|
|
23
23
|
import { TxHash } from '@aztec/stdlib/tx';
|
|
24
24
|
|
|
25
|
+
import { OutOfOrderLogInsertionError } from '../errors.js';
|
|
25
26
|
import type { BlockStore } from './block_store.js';
|
|
26
27
|
|
|
27
28
|
/**
|
|
@@ -165,10 +166,21 @@ export class LogStore {
|
|
|
165
166
|
|
|
166
167
|
for (const taggedLogBuffer of currentPrivateTaggedLogs) {
|
|
167
168
|
if (taggedLogBuffer.logBuffers && taggedLogBuffer.logBuffers.length > 0) {
|
|
168
|
-
privateTaggedLogs.
|
|
169
|
-
|
|
170
|
-
|
|
171
|
-
|
|
169
|
+
const newLogs = privateTaggedLogs.get(taggedLogBuffer.tag)!;
|
|
170
|
+
if (newLogs.length === 0) {
|
|
171
|
+
continue;
|
|
172
|
+
}
|
|
173
|
+
const lastExisting = TxScopedL2Log.fromBuffer(taggedLogBuffer.logBuffers.at(-1)!);
|
|
174
|
+
const firstNew = TxScopedL2Log.fromBuffer(newLogs[0]);
|
|
175
|
+
if (lastExisting.blockNumber > firstNew.blockNumber) {
|
|
176
|
+
throw new OutOfOrderLogInsertionError(
|
|
177
|
+
'private',
|
|
178
|
+
taggedLogBuffer.tag,
|
|
179
|
+
lastExisting.blockNumber,
|
|
180
|
+
firstNew.blockNumber,
|
|
181
|
+
);
|
|
182
|
+
}
|
|
183
|
+
privateTaggedLogs.set(taggedLogBuffer.tag, taggedLogBuffer.logBuffers.concat(newLogs));
|
|
172
184
|
}
|
|
173
185
|
}
|
|
174
186
|
|
|
@@ -200,10 +212,21 @@ export class LogStore {
|
|
|
200
212
|
|
|
201
213
|
for (const taggedLogBuffer of currentPublicTaggedLogs) {
|
|
202
214
|
if (taggedLogBuffer.logBuffers && taggedLogBuffer.logBuffers.length > 0) {
|
|
203
|
-
publicTaggedLogs.
|
|
204
|
-
|
|
205
|
-
|
|
206
|
-
|
|
215
|
+
const newLogs = publicTaggedLogs.get(taggedLogBuffer.tag)!;
|
|
216
|
+
if (newLogs.length === 0) {
|
|
217
|
+
continue;
|
|
218
|
+
}
|
|
219
|
+
const lastExisting = TxScopedL2Log.fromBuffer(taggedLogBuffer.logBuffers.at(-1)!);
|
|
220
|
+
const firstNew = TxScopedL2Log.fromBuffer(newLogs[0]);
|
|
221
|
+
if (lastExisting.blockNumber > firstNew.blockNumber) {
|
|
222
|
+
throw new OutOfOrderLogInsertionError(
|
|
223
|
+
'public',
|
|
224
|
+
taggedLogBuffer.tag,
|
|
225
|
+
lastExisting.blockNumber,
|
|
226
|
+
firstNew.blockNumber,
|
|
227
|
+
);
|
|
228
|
+
}
|
|
229
|
+
publicTaggedLogs.set(taggedLogBuffer.tag, taggedLogBuffer.logBuffers.concat(newLogs));
|
|
207
230
|
}
|
|
208
231
|
}
|
|
209
232
|
|
|
@@ -290,18 +313,49 @@ export class LogStore {
|
|
|
290
313
|
|
|
291
314
|
deleteLogs(blocks: L2Block[]): Promise<boolean> {
|
|
292
315
|
return this.db.transactionAsync(async () => {
|
|
293
|
-
|
|
294
|
-
|
|
295
|
-
|
|
296
|
-
|
|
297
|
-
|
|
298
|
-
|
|
299
|
-
// Delete public logs
|
|
300
|
-
const publicKeys = (await this.#publicLogKeysByBlock.getAsync(block.number)) ?? [];
|
|
301
|
-
await Promise.all(publicKeys.map(key => this.#publicLogsByContractAndTag.delete(key)));
|
|
302
|
-
}),
|
|
316
|
+
const blockNumbers = new Set(blocks.map(block => block.number));
|
|
317
|
+
const firstBlockToDelete = Math.min(...blockNumbers);
|
|
318
|
+
|
|
319
|
+
// Collect all unique private tags across all blocks being deleted
|
|
320
|
+
const allPrivateTags = new Set(
|
|
321
|
+
compactArray(await Promise.all(blocks.map(block => this.#privateLogKeysByBlock.getAsync(block.number)))).flat(),
|
|
303
322
|
);
|
|
304
323
|
|
|
324
|
+
// Trim private logs: for each tag, delete all instances including and after the first block being deleted.
|
|
325
|
+
// This hinges on the invariant that logs for a given tag are always inserted in order of block number, which is enforced in #addPrivateLogs.
|
|
326
|
+
for (const tag of allPrivateTags) {
|
|
327
|
+
const existing = await this.#privateLogsByTag.getAsync(tag);
|
|
328
|
+
if (existing === undefined || existing.length === 0) {
|
|
329
|
+
continue;
|
|
330
|
+
}
|
|
331
|
+
const lastIndexToKeep = existing.findLastIndex(
|
|
332
|
+
buf => TxScopedL2Log.getBlockNumberFromBuffer(buf) < firstBlockToDelete,
|
|
333
|
+
);
|
|
334
|
+
const remaining = existing.slice(0, lastIndexToKeep + 1);
|
|
335
|
+
await (remaining.length > 0 ? this.#privateLogsByTag.set(tag, remaining) : this.#privateLogsByTag.delete(tag));
|
|
336
|
+
}
|
|
337
|
+
|
|
338
|
+
// Collect all unique public keys across all blocks being deleted
|
|
339
|
+
const allPublicKeys = new Set(
|
|
340
|
+
compactArray(await Promise.all(blocks.map(block => this.#publicLogKeysByBlock.getAsync(block.number)))).flat(),
|
|
341
|
+
);
|
|
342
|
+
|
|
343
|
+
// And do the same as we did with private logs
|
|
344
|
+
for (const key of allPublicKeys) {
|
|
345
|
+
const existing = await this.#publicLogsByContractAndTag.getAsync(key);
|
|
346
|
+
if (existing === undefined || existing.length === 0) {
|
|
347
|
+
continue;
|
|
348
|
+
}
|
|
349
|
+
const lastIndexToKeep = existing.findLastIndex(
|
|
350
|
+
buf => TxScopedL2Log.getBlockNumberFromBuffer(buf) < firstBlockToDelete,
|
|
351
|
+
);
|
|
352
|
+
const remaining = existing.slice(0, lastIndexToKeep + 1);
|
|
353
|
+
await (remaining.length > 0
|
|
354
|
+
? this.#publicLogsByContractAndTag.set(key, remaining)
|
|
355
|
+
: this.#publicLogsByContractAndTag.delete(key));
|
|
356
|
+
}
|
|
357
|
+
|
|
358
|
+
// After trimming the tagged logs, we can delete the block-level keys that track which tags are in which blocks.
|
|
305
359
|
await Promise.all(
|
|
306
360
|
blocks.map(block =>
|
|
307
361
|
Promise.all([
|
|
@@ -322,17 +376,30 @@ export class LogStore {
|
|
|
322
376
|
* array implies no logs match that tag.
|
|
323
377
|
* @param tags - The tags to search for.
|
|
324
378
|
* @param page - The page number (0-indexed) for pagination.
|
|
379
|
+
* @param upToBlockNumber - If set, only return logs from blocks up to and including this block number.
|
|
325
380
|
* @returns An array of log arrays, one per tag. Returns at most MAX_LOGS_PER_TAG logs per tag per page. If
|
|
326
381
|
* MAX_LOGS_PER_TAG logs are returned for a tag, the caller should fetch the next page to check for more logs.
|
|
327
382
|
*/
|
|
328
|
-
async getPrivateLogsByTags(
|
|
383
|
+
async getPrivateLogsByTags(
|
|
384
|
+
tags: SiloedTag[],
|
|
385
|
+
page: number = 0,
|
|
386
|
+
upToBlockNumber?: BlockNumber,
|
|
387
|
+
): Promise<TxScopedL2Log[][]> {
|
|
329
388
|
const logs = await Promise.all(tags.map(tag => this.#privateLogsByTag.getAsync(tag.toString())));
|
|
389
|
+
|
|
330
390
|
const start = page * MAX_LOGS_PER_TAG;
|
|
331
391
|
const end = start + MAX_LOGS_PER_TAG;
|
|
332
392
|
|
|
333
|
-
return logs.map(
|
|
334
|
-
|
|
335
|
-
|
|
393
|
+
return logs.map(logBuffers => {
|
|
394
|
+
const deserialized = logBuffers?.slice(start, end).map(buf => TxScopedL2Log.fromBuffer(buf)) ?? [];
|
|
395
|
+
if (upToBlockNumber !== undefined) {
|
|
396
|
+
const cutoff = deserialized.findIndex(log => log.blockNumber > upToBlockNumber);
|
|
397
|
+
if (cutoff !== -1) {
|
|
398
|
+
return deserialized.slice(0, cutoff);
|
|
399
|
+
}
|
|
400
|
+
}
|
|
401
|
+
return deserialized;
|
|
402
|
+
});
|
|
336
403
|
}
|
|
337
404
|
|
|
338
405
|
/**
|
|
@@ -341,6 +408,7 @@ export class LogStore {
|
|
|
341
408
|
* @param contractAddress - The contract address to search logs for.
|
|
342
409
|
* @param tags - The tags to search for.
|
|
343
410
|
* @param page - The page number (0-indexed) for pagination.
|
|
411
|
+
* @param upToBlockNumber - If set, only return logs from blocks up to and including this block number.
|
|
344
412
|
* @returns An array of log arrays, one per tag. Returns at most MAX_LOGS_PER_TAG logs per tag per page. If
|
|
345
413
|
* MAX_LOGS_PER_TAG logs are returned for a tag, the caller should fetch the next page to check for more logs.
|
|
346
414
|
*/
|
|
@@ -348,6 +416,7 @@ export class LogStore {
|
|
|
348
416
|
contractAddress: AztecAddress,
|
|
349
417
|
tags: Tag[],
|
|
350
418
|
page: number = 0,
|
|
419
|
+
upToBlockNumber?: BlockNumber,
|
|
351
420
|
): Promise<TxScopedL2Log[][]> {
|
|
352
421
|
const logs = await Promise.all(
|
|
353
422
|
tags.map(tag => {
|
|
@@ -358,9 +427,16 @@ export class LogStore {
|
|
|
358
427
|
const start = page * MAX_LOGS_PER_TAG;
|
|
359
428
|
const end = start + MAX_LOGS_PER_TAG;
|
|
360
429
|
|
|
361
|
-
return logs.map(
|
|
362
|
-
|
|
363
|
-
|
|
430
|
+
return logs.map(logBuffers => {
|
|
431
|
+
const deserialized = logBuffers?.slice(start, end).map(buf => TxScopedL2Log.fromBuffer(buf)) ?? [];
|
|
432
|
+
if (upToBlockNumber !== undefined) {
|
|
433
|
+
const cutoff = deserialized.findIndex(log => log.blockNumber > upToBlockNumber);
|
|
434
|
+
if (cutoff !== -1) {
|
|
435
|
+
return deserialized.slice(0, cutoff);
|
|
436
|
+
}
|
|
437
|
+
}
|
|
438
|
+
return deserialized;
|
|
439
|
+
});
|
|
364
440
|
}
|
|
365
441
|
|
|
366
442
|
/**
|
|
@@ -588,11 +664,24 @@ export class LogStore {
|
|
|
588
664
|
txLogs: PublicLog[],
|
|
589
665
|
filter: LogFilter = {},
|
|
590
666
|
): boolean {
|
|
667
|
+
if (filter.fromBlock && blockNumber < filter.fromBlock) {
|
|
668
|
+
return false;
|
|
669
|
+
}
|
|
670
|
+
if (filter.toBlock && blockNumber >= filter.toBlock) {
|
|
671
|
+
return false;
|
|
672
|
+
}
|
|
673
|
+
if (filter.txHash && !txHash.equals(filter.txHash)) {
|
|
674
|
+
return false;
|
|
675
|
+
}
|
|
676
|
+
|
|
591
677
|
let maxLogsHit = false;
|
|
592
678
|
let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
|
|
593
679
|
for (; logIndex < txLogs.length; logIndex++) {
|
|
594
680
|
const log = txLogs[logIndex];
|
|
595
|
-
if (
|
|
681
|
+
if (
|
|
682
|
+
(!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) &&
|
|
683
|
+
(!filter.tag || log.fields[0]?.equals(filter.tag))
|
|
684
|
+
) {
|
|
596
685
|
results.push(
|
|
597
686
|
new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
|
|
598
687
|
);
|
|
@@ -616,6 +705,16 @@ export class LogStore {
|
|
|
616
705
|
txLogs: ContractClassLog[],
|
|
617
706
|
filter: LogFilter = {},
|
|
618
707
|
): boolean {
|
|
708
|
+
if (filter.fromBlock && blockNumber < filter.fromBlock) {
|
|
709
|
+
return false;
|
|
710
|
+
}
|
|
711
|
+
if (filter.toBlock && blockNumber >= filter.toBlock) {
|
|
712
|
+
return false;
|
|
713
|
+
}
|
|
714
|
+
if (filter.txHash && !txHash.equals(filter.txHash)) {
|
|
715
|
+
return false;
|
|
716
|
+
}
|
|
717
|
+
|
|
619
718
|
let maxLogsHit = false;
|
|
620
719
|
let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
|
|
621
720
|
for (; logIndex < txLogs.length; logIndex++) {
|
|
@@ -14,6 +14,7 @@ import {
|
|
|
14
14
|
} from '@aztec/kv-store';
|
|
15
15
|
import { InboxLeaf } from '@aztec/stdlib/messaging';
|
|
16
16
|
|
|
17
|
+
import { L1ToL2MessagesNotReadyError } from '../errors.js';
|
|
17
18
|
import {
|
|
18
19
|
type InboxMessage,
|
|
19
20
|
deserializeInboxMessage,
|
|
@@ -40,6 +41,8 @@ export class MessageStore {
|
|
|
40
41
|
#lastSynchedL1Block: AztecAsyncSingleton<Buffer>;
|
|
41
42
|
/** Stores total messages stored */
|
|
42
43
|
#totalMessageCount: AztecAsyncSingleton<bigint>;
|
|
44
|
+
/** Stores the checkpoint number whose message tree is currently being filled on L1. */
|
|
45
|
+
#inboxTreeInProgress: AztecAsyncSingleton<bigint>;
|
|
43
46
|
|
|
44
47
|
#log = createLogger('archiver:message_store');
|
|
45
48
|
|
|
@@ -48,6 +51,7 @@ export class MessageStore {
|
|
|
48
51
|
this.#l1ToL2MessageIndices = db.openMap('archiver_l1_to_l2_message_indices');
|
|
49
52
|
this.#lastSynchedL1Block = db.openSingleton('archiver_last_l1_block_id');
|
|
50
53
|
this.#totalMessageCount = db.openSingleton('archiver_l1_to_l2_message_count');
|
|
54
|
+
this.#inboxTreeInProgress = db.openSingleton('archiver_inbox_tree_in_progress');
|
|
51
55
|
}
|
|
52
56
|
|
|
53
57
|
public async getTotalL1ToL2MessageCount(): Promise<bigint> {
|
|
@@ -137,7 +141,7 @@ export class MessageStore {
|
|
|
137
141
|
);
|
|
138
142
|
}
|
|
139
143
|
|
|
140
|
-
// Check the first message in a
|
|
144
|
+
// Check the first message in a checkpoint has the correct index.
|
|
141
145
|
if (
|
|
142
146
|
(!lastMessage || message.checkpointNumber > lastMessage.checkpointNumber) &&
|
|
143
147
|
message.index !== expectedStart
|
|
@@ -185,7 +189,22 @@ export class MessageStore {
|
|
|
185
189
|
return msg ? deserializeInboxMessage(msg) : undefined;
|
|
186
190
|
}
|
|
187
191
|
|
|
192
|
+
/** Returns the inbox tree-in-progress checkpoint number from L1, or undefined if not yet set. */
|
|
193
|
+
public getInboxTreeInProgress(): Promise<bigint | undefined> {
|
|
194
|
+
return this.#inboxTreeInProgress.getAsync();
|
|
195
|
+
}
|
|
196
|
+
|
|
197
|
+
/** Persists the inbox tree-in-progress checkpoint number from L1 state. */
|
|
198
|
+
public async setInboxTreeInProgress(value: bigint): Promise<void> {
|
|
199
|
+
await this.#inboxTreeInProgress.set(value);
|
|
200
|
+
}
|
|
201
|
+
|
|
188
202
|
public async getL1ToL2Messages(checkpointNumber: CheckpointNumber): Promise<Fr[]> {
|
|
203
|
+
const treeInProgress = await this.#inboxTreeInProgress.getAsync();
|
|
204
|
+
if (treeInProgress !== undefined && BigInt(checkpointNumber) >= treeInProgress) {
|
|
205
|
+
throw new L1ToL2MessagesNotReadyError(checkpointNumber, treeInProgress);
|
|
206
|
+
}
|
|
207
|
+
|
|
189
208
|
const messages: Fr[] = [];
|
|
190
209
|
|
|
191
210
|
const [startIndex, endIndex] = InboxLeaf.indexRangeForCheckpoint(checkpointNumber);
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
import type { BlobClientInterface } from '@aztec/blob-client/client';
|
|
2
2
|
import { type Blob, getBlobsPerL1Block, getPrefixedEthBlobCommitments } from '@aztec/blob-lib';
|
|
3
|
+
import { INITIAL_CHECKPOINT_NUMBER } from '@aztec/constants';
|
|
3
4
|
import type { CheckpointProposedLog, InboxContract, MessageSentLog, RollupContract } from '@aztec/ethereum/contracts';
|
|
4
5
|
import { MULTI_CALL_3_ADDRESS } from '@aztec/ethereum/contracts';
|
|
5
6
|
import type { ViemPublicClient } from '@aztec/ethereum/types';
|
|
@@ -150,8 +151,12 @@ export class FakeL1State {
|
|
|
150
151
|
// Computed from checkpoints based on L1 block visibility
|
|
151
152
|
private pendingCheckpointNumber: CheckpointNumber = CheckpointNumber(0);
|
|
152
153
|
|
|
154
|
+
// The L1 block number reported as "finalized" (defaults to the start block)
|
|
155
|
+
private finalizedL1BlockNumber: bigint;
|
|
156
|
+
|
|
153
157
|
constructor(private readonly config: FakeL1StateConfig) {
|
|
154
158
|
this.l1BlockNumber = config.l1StartBlock;
|
|
159
|
+
this.finalizedL1BlockNumber = config.l1StartBlock;
|
|
155
160
|
this.lastArchive = new AppendOnlyTreeSnapshot(config.genesisArchiveRoot, 1);
|
|
156
161
|
}
|
|
157
162
|
|
|
@@ -283,11 +288,30 @@ export class FakeL1State {
|
|
|
283
288
|
this.updatePendingCheckpointNumber();
|
|
284
289
|
}
|
|
285
290
|
|
|
291
|
+
/** Sets the L1 block number that will be reported as "finalized". */
|
|
292
|
+
setFinalizedL1BlockNumber(blockNumber: bigint): void {
|
|
293
|
+
this.finalizedL1BlockNumber = blockNumber;
|
|
294
|
+
}
|
|
295
|
+
|
|
286
296
|
/** Marks a checkpoint as proven. Updates provenCheckpointNumber. */
|
|
287
297
|
markCheckpointAsProven(checkpointNumber: CheckpointNumber): void {
|
|
288
298
|
this.provenCheckpointNumber = checkpointNumber;
|
|
289
299
|
}
|
|
290
300
|
|
|
301
|
+
/**
|
|
302
|
+
* Simulates what `rollup.getProvenCheckpointNumber({ blockNumber: atL1Block })` would return.
|
|
303
|
+
*/
|
|
304
|
+
getProvenCheckpointNumberAtL1Block(atL1Block: bigint): CheckpointNumber {
|
|
305
|
+
if (this.provenCheckpointNumber === 0) {
|
|
306
|
+
return CheckpointNumber(0);
|
|
307
|
+
}
|
|
308
|
+
const checkpoint = this.checkpoints.find(cp => cp.checkpointNumber === this.provenCheckpointNumber);
|
|
309
|
+
if (checkpoint && checkpoint.l1BlockNumber <= atL1Block) {
|
|
310
|
+
return this.provenCheckpointNumber;
|
|
311
|
+
}
|
|
312
|
+
return CheckpointNumber(0);
|
|
313
|
+
}
|
|
314
|
+
|
|
291
315
|
/** Sets the target committee size for attestation validation. */
|
|
292
316
|
setTargetCommitteeSize(size: number): void {
|
|
293
317
|
this.targetCommitteeSize = size;
|
|
@@ -406,6 +430,11 @@ export class FakeL1State {
|
|
|
406
430
|
});
|
|
407
431
|
});
|
|
408
432
|
|
|
433
|
+
mockRollup.getProvenCheckpointNumber.mockImplementation((options?: { blockNumber?: bigint }) => {
|
|
434
|
+
const atBlock = options?.blockNumber ?? this.l1BlockNumber;
|
|
435
|
+
return Promise.resolve(this.getProvenCheckpointNumberAtL1Block(atBlock));
|
|
436
|
+
});
|
|
437
|
+
|
|
409
438
|
mockRollup.canPruneAtTime.mockImplementation(() => Promise.resolve(this.canPruneResult));
|
|
410
439
|
|
|
411
440
|
// Mock the wrapper method for fetching checkpoint events
|
|
@@ -422,13 +451,22 @@ export class FakeL1State {
|
|
|
422
451
|
createMockInboxContract(_publicClient: MockProxy<ViemPublicClient>): MockProxy<InboxContract> {
|
|
423
452
|
const mockInbox = mock<InboxContract>();
|
|
424
453
|
|
|
425
|
-
mockInbox.getState.mockImplementation(() =>
|
|
426
|
-
|
|
454
|
+
mockInbox.getState.mockImplementation(() => {
|
|
455
|
+
// treeInProgress must be > any sealed checkpoint. On L1, a checkpoint can only be proposed
|
|
456
|
+
// after its messages are sealed, so treeInProgress > checkpointNumber for all published checkpoints.
|
|
457
|
+
const maxFromMessages =
|
|
458
|
+
this.messages.length > 0 ? Math.max(...this.messages.map(m => Number(m.checkpointNumber))) + 1 : 0;
|
|
459
|
+
const maxFromCheckpoints =
|
|
460
|
+
this.checkpoints.length > 0
|
|
461
|
+
? Math.max(...this.checkpoints.filter(cp => !cp.pruned).map(cp => Number(cp.checkpointNumber))) + 1
|
|
462
|
+
: 0;
|
|
463
|
+
const treeInProgress = Math.max(maxFromMessages, maxFromCheckpoints, INITIAL_CHECKPOINT_NUMBER);
|
|
464
|
+
return Promise.resolve({
|
|
427
465
|
messagesRollingHash: this.messagesRollingHash,
|
|
428
466
|
totalMessagesInserted: BigInt(this.messages.length),
|
|
429
|
-
treeInProgress:
|
|
430
|
-
})
|
|
431
|
-
);
|
|
467
|
+
treeInProgress: BigInt(treeInProgress),
|
|
468
|
+
});
|
|
469
|
+
});
|
|
432
470
|
|
|
433
471
|
// Mock the wrapper methods for fetching message events
|
|
434
472
|
mockInbox.getMessageSentEvents.mockImplementation((fromBlock: bigint, toBlock: bigint) =>
|
|
@@ -449,10 +487,13 @@ export class FakeL1State {
|
|
|
449
487
|
publicClient.getChainId.mockResolvedValue(1);
|
|
450
488
|
publicClient.getBlockNumber.mockImplementation(() => Promise.resolve(this.l1BlockNumber));
|
|
451
489
|
|
|
452
|
-
|
|
453
|
-
|
|
454
|
-
|
|
455
|
-
|
|
490
|
+
publicClient.getBlock.mockImplementation((async (args: { blockNumber?: bigint; blockTag?: string } = {}) => {
|
|
491
|
+
let blockNum: bigint;
|
|
492
|
+
if (args.blockTag === 'finalized') {
|
|
493
|
+
blockNum = this.finalizedL1BlockNumber;
|
|
494
|
+
} else {
|
|
495
|
+
blockNum = args.blockNumber ?? (await publicClient.getBlockNumber());
|
|
496
|
+
}
|
|
456
497
|
return {
|
|
457
498
|
number: blockNum,
|
|
458
499
|
timestamp: BigInt(blockNum) * BigInt(this.config.ethereumSlotDuration) + this.config.l1GenesisTime,
|
|
@@ -18,7 +18,12 @@ import {
|
|
|
18
18
|
} from '@aztec/stdlib/block';
|
|
19
19
|
import { Checkpoint, type CheckpointData, L1PublishedData, PublishedCheckpoint } from '@aztec/stdlib/checkpoint';
|
|
20
20
|
import type { ContractClassPublic, ContractDataSource, ContractInstanceWithAddress } from '@aztec/stdlib/contract';
|
|
21
|
-
import {
|
|
21
|
+
import {
|
|
22
|
+
EmptyL1RollupConstants,
|
|
23
|
+
type L1RollupConstants,
|
|
24
|
+
getEpochAtSlot,
|
|
25
|
+
getSlotRangeForEpoch,
|
|
26
|
+
} from '@aztec/stdlib/epoch-helpers';
|
|
22
27
|
import { computeCheckpointOutHash } from '@aztec/stdlib/messaging';
|
|
23
28
|
import { CheckpointHeader } from '@aztec/stdlib/rollup';
|
|
24
29
|
import { type BlockHeader, TxExecutionResult, TxHash, TxReceipt, TxStatus } from '@aztec/stdlib/tx';
|
|
@@ -42,6 +47,12 @@ export class MockL2BlockSource implements L2BlockSource, ContractDataSource {
|
|
|
42
47
|
await this.createCheckpoints(numBlocks, 1);
|
|
43
48
|
}
|
|
44
49
|
|
|
50
|
+
public getCheckpointNumber(): Promise<CheckpointNumber> {
|
|
51
|
+
return Promise.resolve(
|
|
52
|
+
this.checkpointList.length === 0 ? CheckpointNumber.ZERO : CheckpointNumber(this.checkpointList.length),
|
|
53
|
+
);
|
|
54
|
+
}
|
|
55
|
+
|
|
45
56
|
/** Creates checkpoints, each containing `blocksPerCheckpoint` blocks. */
|
|
46
57
|
public async createCheckpoints(numCheckpoints: number, blocksPerCheckpoint: number = 1) {
|
|
47
58
|
for (let c = 0; c < numCheckpoints; c++) {
|
|
@@ -388,6 +399,7 @@ export class MockL2BlockSource implements L2BlockSource, ContractDataSource {
|
|
|
388
399
|
txEffect.transactionFee.toBigInt(),
|
|
389
400
|
await block.hash(),
|
|
390
401
|
block.number,
|
|
402
|
+
getEpochAtSlot(block.slot, EmptyL1RollupConstants),
|
|
391
403
|
);
|
|
392
404
|
}
|
|
393
405
|
}
|
|
@@ -441,11 +453,11 @@ export class MockL2BlockSource implements L2BlockSource, ContractDataSource {
|
|
|
441
453
|
};
|
|
442
454
|
}
|
|
443
455
|
|
|
444
|
-
|
|
456
|
+
getSyncedL2EpochNumber(): Promise<EpochNumber> {
|
|
445
457
|
throw new Error('Method not implemented.');
|
|
446
458
|
}
|
|
447
459
|
|
|
448
|
-
|
|
460
|
+
getSyncedL2SlotNumber(): Promise<SlotNumber> {
|
|
449
461
|
throw new Error('Method not implemented.');
|
|
450
462
|
}
|
|
451
463
|
|
package/src/test/mock_structs.ts
CHANGED
|
@@ -127,6 +127,25 @@ export function makeL1PublishedData(l1BlockNumber: number): L1PublishedData {
|
|
|
127
127
|
return new L1PublishedData(BigInt(l1BlockNumber), BigInt(l1BlockNumber * 1000), makeBlockHash(l1BlockNumber));
|
|
128
128
|
}
|
|
129
129
|
|
|
130
|
+
/** Creates a Checkpoint from a list of blocks with a header that matches the blocks' structure. */
|
|
131
|
+
export function makeCheckpoint(blocks: L2Block[], checkpointNumber = CheckpointNumber(1)): Checkpoint {
|
|
132
|
+
const firstBlock = blocks[0];
|
|
133
|
+
const { slotNumber, timestamp, coinbase, feeRecipient, gasFees } = firstBlock.header.globalVariables;
|
|
134
|
+
return new Checkpoint(
|
|
135
|
+
blocks.at(-1)!.archive,
|
|
136
|
+
CheckpointHeader.random({
|
|
137
|
+
lastArchiveRoot: firstBlock.header.lastArchive.root,
|
|
138
|
+
slotNumber,
|
|
139
|
+
timestamp,
|
|
140
|
+
coinbase,
|
|
141
|
+
feeRecipient,
|
|
142
|
+
gasFees,
|
|
143
|
+
}),
|
|
144
|
+
blocks,
|
|
145
|
+
checkpointNumber,
|
|
146
|
+
);
|
|
147
|
+
}
|
|
148
|
+
|
|
130
149
|
/** Wraps a Checkpoint with L1 published data and random attestations. */
|
|
131
150
|
export function makePublishedCheckpoint(
|
|
132
151
|
checkpoint: Checkpoint,
|
|
@@ -301,11 +320,6 @@ export async function makeCheckpointWithLogs(
|
|
|
301
320
|
return txEffect;
|
|
302
321
|
});
|
|
303
322
|
|
|
304
|
-
const checkpoint =
|
|
305
|
-
AppendOnlyTreeSnapshot.random(),
|
|
306
|
-
CheckpointHeader.random(),
|
|
307
|
-
[block],
|
|
308
|
-
CheckpointNumber.fromBlockNumber(BlockNumber(blockNumber)),
|
|
309
|
-
);
|
|
323
|
+
const checkpoint = makeCheckpoint([block], CheckpointNumber.fromBlockNumber(BlockNumber(blockNumber)));
|
|
310
324
|
return makePublishedCheckpoint(checkpoint, blockNumber);
|
|
311
325
|
}
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import type { BlobClientInterface } from '@aztec/blob-client/client';
|
|
2
2
|
import type { RollupContract } from '@aztec/ethereum/contracts';
|
|
3
3
|
import type { ViemPublicClient, ViemPublicDebugClient } from '@aztec/ethereum/types';
|
|
4
|
+
import { SlotNumber } from '@aztec/foundation/branded-types';
|
|
4
5
|
import { Buffer32 } from '@aztec/foundation/buffer';
|
|
5
6
|
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
6
7
|
import { EthAddress } from '@aztec/foundation/eth-address';
|
|
@@ -30,7 +31,7 @@ class NoopL1Synchronizer implements FunctionsOf<ArchiverL1Synchronizer> {
|
|
|
30
31
|
return 0n;
|
|
31
32
|
}
|
|
32
33
|
getL1Timestamp(): bigint | undefined {
|
|
33
|
-
return
|
|
34
|
+
return undefined;
|
|
34
35
|
}
|
|
35
36
|
testEthereumNodeSynced(): Promise<void> {
|
|
36
37
|
return Promise.resolve();
|
|
@@ -96,6 +97,11 @@ export class NoopL1Archiver extends Archiver {
|
|
|
96
97
|
this.runningPromise.start();
|
|
97
98
|
return Promise.resolve();
|
|
98
99
|
}
|
|
100
|
+
|
|
101
|
+
/** Always reports as fully synced since there is no real L1 to sync from. */
|
|
102
|
+
public override getSyncedL2SlotNumber(): Promise<SlotNumber | undefined> {
|
|
103
|
+
return Promise.resolve(SlotNumber(Number.MAX_SAFE_INTEGER));
|
|
104
|
+
}
|
|
99
105
|
}
|
|
100
106
|
|
|
101
107
|
/** Creates an archiver with mocked L1 connectivity for testing. */
|