@aztec/archiver 0.0.1-commit.b64cb54f6 → 0.0.1-commit.b6e433891
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/archiver.d.ts +1 -2
- package/dest/archiver.d.ts.map +1 -1
- package/dest/archiver.js +0 -1
- package/dest/config.d.ts +3 -3
- package/dest/config.d.ts.map +1 -1
- package/dest/config.js +2 -1
- package/dest/errors.d.ts +15 -1
- package/dest/errors.d.ts.map +1 -1
- package/dest/errors.js +18 -0
- package/dest/factory.d.ts +1 -1
- package/dest/factory.d.ts.map +1 -1
- package/dest/factory.js +9 -7
- package/dest/modules/data_source_base.d.ts +3 -3
- package/dest/modules/data_source_base.d.ts.map +1 -1
- package/dest/modules/data_source_base.js +4 -4
- package/dest/modules/data_store_updater.js +1 -1
- package/dest/modules/l1_synchronizer.d.ts +1 -1
- package/dest/modules/l1_synchronizer.d.ts.map +1 -1
- package/dest/modules/l1_synchronizer.js +9 -5
- package/dest/store/kv_archiver_store.d.ts +7 -3
- package/dest/store/kv_archiver_store.d.ts.map +1 -1
- package/dest/store/kv_archiver_store.js +9 -4
- package/dest/store/log_store.d.ts +6 -3
- package/dest/store/log_store.d.ts.map +1 -1
- package/dest/store/log_store.js +93 -16
- package/dest/store/message_store.d.ts +5 -1
- package/dest/store/message_store.d.ts.map +1 -1
- package/dest/store/message_store.js +13 -0
- package/dest/test/fake_l1_state.d.ts +1 -1
- package/dest/test/fake_l1_state.d.ts.map +1 -1
- package/dest/test/fake_l1_state.js +11 -3
- package/package.json +13 -13
- package/src/archiver.ts +0 -1
- package/src/config.ts +8 -1
- package/src/errors.ts +30 -0
- package/src/factory.ts +5 -3
- package/src/modules/data_source_base.ts +8 -3
- package/src/modules/data_store_updater.ts +1 -1
- package/src/modules/l1_synchronizer.ts +15 -10
- package/src/store/kv_archiver_store.ts +11 -3
- package/src/store/log_store.ts +126 -27
- package/src/store/message_store.ts +19 -0
- package/src/test/fake_l1_state.ts +15 -5
|
@@ -3,6 +3,7 @@ import { EpochCache } from '@aztec/epoch-cache';
|
|
|
3
3
|
import { InboxContract, RollupContract } from '@aztec/ethereum/contracts';
|
|
4
4
|
import type { L1BlockId } from '@aztec/ethereum/l1-types';
|
|
5
5
|
import type { ViemPublicClient, ViemPublicDebugClient } from '@aztec/ethereum/types';
|
|
6
|
+
import { asyncPool } from '@aztec/foundation/async-pool';
|
|
6
7
|
import { maxBigint } from '@aztec/foundation/bigint';
|
|
7
8
|
import { BlockNumber, CheckpointNumber, EpochNumber } from '@aztec/foundation/branded-types';
|
|
8
9
|
import { Buffer32 } from '@aztec/foundation/buffer';
|
|
@@ -333,17 +334,20 @@ export class ArchiverL1Synchronizer implements Traceable {
|
|
|
333
334
|
|
|
334
335
|
const checkpointsToUnwind = localPendingCheckpointNumber - provenCheckpointNumber;
|
|
335
336
|
|
|
336
|
-
|
|
337
|
-
|
|
338
|
-
|
|
339
|
-
const
|
|
340
|
-
|
|
341
|
-
|
|
342
|
-
checkpoints
|
|
343
|
-
.filter(isDefined)
|
|
344
|
-
.map(cp => this.store.getBlocksForCheckpoint(CheckpointNumber(cp.checkpointNumber))),
|
|
337
|
+
// Fetch checkpoints and blocks in bounded batches to avoid unbounded concurrent
|
|
338
|
+
// promises when the gap between local pending and proven checkpoint numbers is large.
|
|
339
|
+
const BATCH_SIZE = 10;
|
|
340
|
+
const indices = Array.from({ length: checkpointsToUnwind }, (_, i) => CheckpointNumber(i + pruneFrom));
|
|
341
|
+
const checkpoints = (await asyncPool(BATCH_SIZE, indices, idx => this.store.getCheckpointData(idx))).filter(
|
|
342
|
+
isDefined,
|
|
345
343
|
);
|
|
346
|
-
const newBlocks =
|
|
344
|
+
const newBlocks = (
|
|
345
|
+
await asyncPool(BATCH_SIZE, checkpoints, cp =>
|
|
346
|
+
this.store.getBlocksForCheckpoint(CheckpointNumber(cp.checkpointNumber)),
|
|
347
|
+
)
|
|
348
|
+
)
|
|
349
|
+
.filter(isDefined)
|
|
350
|
+
.flat();
|
|
347
351
|
|
|
348
352
|
// Emit an event for listening services to react to the chain prune
|
|
349
353
|
this.events.emit(L2BlockSourceEvents.L2PruneUnproven, {
|
|
@@ -391,6 +395,7 @@ export class ArchiverL1Synchronizer implements Traceable {
|
|
|
391
395
|
const localMessagesInserted = await this.store.getTotalL1ToL2MessageCount();
|
|
392
396
|
const localLastMessage = await this.store.getLastL1ToL2Message();
|
|
393
397
|
const remoteMessagesState = await this.inbox.getState({ blockNumber: currentL1BlockNumber });
|
|
398
|
+
await this.store.setInboxTreeInProgress(remoteMessagesState.treeInProgress);
|
|
394
399
|
|
|
395
400
|
this.log.trace(`Retrieved remote inbox state at L1 block ${currentL1BlockNumber}.`, {
|
|
396
401
|
localMessagesInserted,
|
|
@@ -474,10 +474,11 @@ export class KVArchiverDataStore implements ContractDataSource {
|
|
|
474
474
|
* array implies no logs match that tag.
|
|
475
475
|
* @param tags - The tags to search for.
|
|
476
476
|
* @param page - The page number (0-indexed) for pagination. Returns at most 10 logs per tag per page.
|
|
477
|
+
* @param upToBlockNumber - If set, only return logs from blocks up to and including this block number.
|
|
477
478
|
*/
|
|
478
|
-
getPrivateLogsByTags(tags: SiloedTag[], page?: number): Promise<TxScopedL2Log[][]> {
|
|
479
|
+
getPrivateLogsByTags(tags: SiloedTag[], page?: number, upToBlockNumber?: BlockNumber): Promise<TxScopedL2Log[][]> {
|
|
479
480
|
try {
|
|
480
|
-
return this.#logStore.getPrivateLogsByTags(tags, page);
|
|
481
|
+
return this.#logStore.getPrivateLogsByTags(tags, page, upToBlockNumber);
|
|
481
482
|
} catch (err) {
|
|
482
483
|
return Promise.reject(err);
|
|
483
484
|
}
|
|
@@ -489,14 +490,16 @@ export class KVArchiverDataStore implements ContractDataSource {
|
|
|
489
490
|
* @param contractAddress - The contract address to search logs for.
|
|
490
491
|
* @param tags - The tags to search for.
|
|
491
492
|
* @param page - The page number (0-indexed) for pagination. Returns at most 10 logs per tag per page.
|
|
493
|
+
* @param upToBlockNumber - If set, only return logs from blocks up to and including this block number.
|
|
492
494
|
*/
|
|
493
495
|
getPublicLogsByTagsFromContract(
|
|
494
496
|
contractAddress: AztecAddress,
|
|
495
497
|
tags: Tag[],
|
|
496
498
|
page?: number,
|
|
499
|
+
upToBlockNumber?: BlockNumber,
|
|
497
500
|
): Promise<TxScopedL2Log[][]> {
|
|
498
501
|
try {
|
|
499
|
-
return this.#logStore.getPublicLogsByTagsFromContract(contractAddress, tags, page);
|
|
502
|
+
return this.#logStore.getPublicLogsByTagsFromContract(contractAddress, tags, page, upToBlockNumber);
|
|
500
503
|
} catch (err) {
|
|
501
504
|
return Promise.reject(err);
|
|
502
505
|
}
|
|
@@ -603,6 +606,11 @@ export class KVArchiverDataStore implements ContractDataSource {
|
|
|
603
606
|
return this.#messageStore.rollbackL1ToL2MessagesToCheckpoint(targetCheckpointNumber);
|
|
604
607
|
}
|
|
605
608
|
|
|
609
|
+
/** Persists the inbox tree-in-progress checkpoint number from L1 state. */
|
|
610
|
+
public setInboxTreeInProgress(value: bigint): Promise<void> {
|
|
611
|
+
return this.#messageStore.setInboxTreeInProgress(value);
|
|
612
|
+
}
|
|
613
|
+
|
|
606
614
|
/** Returns an async iterator to all L1 to L2 messages on the range. */
|
|
607
615
|
public iterateL1ToL2Messages(range: CustomRange<bigint> = {}): AsyncIterableIterator<InboxMessage> {
|
|
608
616
|
return this.#messageStore.iterateL1ToL2Messages(range);
|
package/src/store/log_store.ts
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { INITIAL_L2_BLOCK_NUM } from '@aztec/constants';
|
|
2
2
|
import { BlockNumber } from '@aztec/foundation/branded-types';
|
|
3
|
-
import { filterAsync } from '@aztec/foundation/collection';
|
|
3
|
+
import { compactArray, filterAsync } from '@aztec/foundation/collection';
|
|
4
4
|
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
5
5
|
import { createLogger } from '@aztec/foundation/log';
|
|
6
6
|
import { BufferReader, numToUInt32BE } from '@aztec/foundation/serialize';
|
|
@@ -22,6 +22,7 @@ import {
|
|
|
22
22
|
} from '@aztec/stdlib/logs';
|
|
23
23
|
import { TxHash } from '@aztec/stdlib/tx';
|
|
24
24
|
|
|
25
|
+
import { OutOfOrderLogInsertionError } from '../errors.js';
|
|
25
26
|
import type { BlockStore } from './block_store.js';
|
|
26
27
|
|
|
27
28
|
/**
|
|
@@ -165,10 +166,21 @@ export class LogStore {
|
|
|
165
166
|
|
|
166
167
|
for (const taggedLogBuffer of currentPrivateTaggedLogs) {
|
|
167
168
|
if (taggedLogBuffer.logBuffers && taggedLogBuffer.logBuffers.length > 0) {
|
|
168
|
-
privateTaggedLogs.
|
|
169
|
-
|
|
170
|
-
|
|
171
|
-
|
|
169
|
+
const newLogs = privateTaggedLogs.get(taggedLogBuffer.tag)!;
|
|
170
|
+
if (newLogs.length === 0) {
|
|
171
|
+
continue;
|
|
172
|
+
}
|
|
173
|
+
const lastExisting = TxScopedL2Log.fromBuffer(taggedLogBuffer.logBuffers.at(-1)!);
|
|
174
|
+
const firstNew = TxScopedL2Log.fromBuffer(newLogs[0]);
|
|
175
|
+
if (lastExisting.blockNumber > firstNew.blockNumber) {
|
|
176
|
+
throw new OutOfOrderLogInsertionError(
|
|
177
|
+
'private',
|
|
178
|
+
taggedLogBuffer.tag,
|
|
179
|
+
lastExisting.blockNumber,
|
|
180
|
+
firstNew.blockNumber,
|
|
181
|
+
);
|
|
182
|
+
}
|
|
183
|
+
privateTaggedLogs.set(taggedLogBuffer.tag, taggedLogBuffer.logBuffers.concat(newLogs));
|
|
172
184
|
}
|
|
173
185
|
}
|
|
174
186
|
|
|
@@ -200,10 +212,21 @@ export class LogStore {
|
|
|
200
212
|
|
|
201
213
|
for (const taggedLogBuffer of currentPublicTaggedLogs) {
|
|
202
214
|
if (taggedLogBuffer.logBuffers && taggedLogBuffer.logBuffers.length > 0) {
|
|
203
|
-
publicTaggedLogs.
|
|
204
|
-
|
|
205
|
-
|
|
206
|
-
|
|
215
|
+
const newLogs = publicTaggedLogs.get(taggedLogBuffer.tag)!;
|
|
216
|
+
if (newLogs.length === 0) {
|
|
217
|
+
continue;
|
|
218
|
+
}
|
|
219
|
+
const lastExisting = TxScopedL2Log.fromBuffer(taggedLogBuffer.logBuffers.at(-1)!);
|
|
220
|
+
const firstNew = TxScopedL2Log.fromBuffer(newLogs[0]);
|
|
221
|
+
if (lastExisting.blockNumber > firstNew.blockNumber) {
|
|
222
|
+
throw new OutOfOrderLogInsertionError(
|
|
223
|
+
'public',
|
|
224
|
+
taggedLogBuffer.tag,
|
|
225
|
+
lastExisting.blockNumber,
|
|
226
|
+
firstNew.blockNumber,
|
|
227
|
+
);
|
|
228
|
+
}
|
|
229
|
+
publicTaggedLogs.set(taggedLogBuffer.tag, taggedLogBuffer.logBuffers.concat(newLogs));
|
|
207
230
|
}
|
|
208
231
|
}
|
|
209
232
|
|
|
@@ -290,18 +313,49 @@ export class LogStore {
|
|
|
290
313
|
|
|
291
314
|
deleteLogs(blocks: L2Block[]): Promise<boolean> {
|
|
292
315
|
return this.db.transactionAsync(async () => {
|
|
293
|
-
|
|
294
|
-
|
|
295
|
-
|
|
296
|
-
|
|
297
|
-
|
|
298
|
-
|
|
299
|
-
// Delete public logs
|
|
300
|
-
const publicKeys = (await this.#publicLogKeysByBlock.getAsync(block.number)) ?? [];
|
|
301
|
-
await Promise.all(publicKeys.map(key => this.#publicLogsByContractAndTag.delete(key)));
|
|
302
|
-
}),
|
|
316
|
+
const blockNumbers = new Set(blocks.map(block => block.number));
|
|
317
|
+
const firstBlockToDelete = Math.min(...blockNumbers);
|
|
318
|
+
|
|
319
|
+
// Collect all unique private tags across all blocks being deleted
|
|
320
|
+
const allPrivateTags = new Set(
|
|
321
|
+
compactArray(await Promise.all(blocks.map(block => this.#privateLogKeysByBlock.getAsync(block.number)))).flat(),
|
|
303
322
|
);
|
|
304
323
|
|
|
324
|
+
// Trim private logs: for each tag, delete all instances including and after the first block being deleted.
|
|
325
|
+
// This hinges on the invariant that logs for a given tag are always inserted in order of block number, which is enforced in #addPrivateLogs.
|
|
326
|
+
for (const tag of allPrivateTags) {
|
|
327
|
+
const existing = await this.#privateLogsByTag.getAsync(tag);
|
|
328
|
+
if (existing === undefined || existing.length === 0) {
|
|
329
|
+
continue;
|
|
330
|
+
}
|
|
331
|
+
const lastIndexToKeep = existing.findLastIndex(
|
|
332
|
+
buf => TxScopedL2Log.getBlockNumberFromBuffer(buf) < firstBlockToDelete,
|
|
333
|
+
);
|
|
334
|
+
const remaining = existing.slice(0, lastIndexToKeep + 1);
|
|
335
|
+
await (remaining.length > 0 ? this.#privateLogsByTag.set(tag, remaining) : this.#privateLogsByTag.delete(tag));
|
|
336
|
+
}
|
|
337
|
+
|
|
338
|
+
// Collect all unique public keys across all blocks being deleted
|
|
339
|
+
const allPublicKeys = new Set(
|
|
340
|
+
compactArray(await Promise.all(blocks.map(block => this.#publicLogKeysByBlock.getAsync(block.number)))).flat(),
|
|
341
|
+
);
|
|
342
|
+
|
|
343
|
+
// And do the same as we did with private logs
|
|
344
|
+
for (const key of allPublicKeys) {
|
|
345
|
+
const existing = await this.#publicLogsByContractAndTag.getAsync(key);
|
|
346
|
+
if (existing === undefined || existing.length === 0) {
|
|
347
|
+
continue;
|
|
348
|
+
}
|
|
349
|
+
const lastIndexToKeep = existing.findLastIndex(
|
|
350
|
+
buf => TxScopedL2Log.getBlockNumberFromBuffer(buf) < firstBlockToDelete,
|
|
351
|
+
);
|
|
352
|
+
const remaining = existing.slice(0, lastIndexToKeep + 1);
|
|
353
|
+
await (remaining.length > 0
|
|
354
|
+
? this.#publicLogsByContractAndTag.set(key, remaining)
|
|
355
|
+
: this.#publicLogsByContractAndTag.delete(key));
|
|
356
|
+
}
|
|
357
|
+
|
|
358
|
+
// After trimming the tagged logs, we can delete the block-level keys that track which tags are in which blocks.
|
|
305
359
|
await Promise.all(
|
|
306
360
|
blocks.map(block =>
|
|
307
361
|
Promise.all([
|
|
@@ -322,17 +376,30 @@ export class LogStore {
|
|
|
322
376
|
* array implies no logs match that tag.
|
|
323
377
|
* @param tags - The tags to search for.
|
|
324
378
|
* @param page - The page number (0-indexed) for pagination.
|
|
379
|
+
* @param upToBlockNumber - If set, only return logs from blocks up to and including this block number.
|
|
325
380
|
* @returns An array of log arrays, one per tag. Returns at most MAX_LOGS_PER_TAG logs per tag per page. If
|
|
326
381
|
* MAX_LOGS_PER_TAG logs are returned for a tag, the caller should fetch the next page to check for more logs.
|
|
327
382
|
*/
|
|
328
|
-
async getPrivateLogsByTags(
|
|
383
|
+
async getPrivateLogsByTags(
|
|
384
|
+
tags: SiloedTag[],
|
|
385
|
+
page: number = 0,
|
|
386
|
+
upToBlockNumber?: BlockNumber,
|
|
387
|
+
): Promise<TxScopedL2Log[][]> {
|
|
329
388
|
const logs = await Promise.all(tags.map(tag => this.#privateLogsByTag.getAsync(tag.toString())));
|
|
389
|
+
|
|
330
390
|
const start = page * MAX_LOGS_PER_TAG;
|
|
331
391
|
const end = start + MAX_LOGS_PER_TAG;
|
|
332
392
|
|
|
333
|
-
return logs.map(
|
|
334
|
-
|
|
335
|
-
|
|
393
|
+
return logs.map(logBuffers => {
|
|
394
|
+
const deserialized = logBuffers?.slice(start, end).map(buf => TxScopedL2Log.fromBuffer(buf)) ?? [];
|
|
395
|
+
if (upToBlockNumber !== undefined) {
|
|
396
|
+
const cutoff = deserialized.findIndex(log => log.blockNumber > upToBlockNumber);
|
|
397
|
+
if (cutoff !== -1) {
|
|
398
|
+
return deserialized.slice(0, cutoff);
|
|
399
|
+
}
|
|
400
|
+
}
|
|
401
|
+
return deserialized;
|
|
402
|
+
});
|
|
336
403
|
}
|
|
337
404
|
|
|
338
405
|
/**
|
|
@@ -341,6 +408,7 @@ export class LogStore {
|
|
|
341
408
|
* @param contractAddress - The contract address to search logs for.
|
|
342
409
|
* @param tags - The tags to search for.
|
|
343
410
|
* @param page - The page number (0-indexed) for pagination.
|
|
411
|
+
* @param upToBlockNumber - If set, only return logs from blocks up to and including this block number.
|
|
344
412
|
* @returns An array of log arrays, one per tag. Returns at most MAX_LOGS_PER_TAG logs per tag per page. If
|
|
345
413
|
* MAX_LOGS_PER_TAG logs are returned for a tag, the caller should fetch the next page to check for more logs.
|
|
346
414
|
*/
|
|
@@ -348,6 +416,7 @@ export class LogStore {
|
|
|
348
416
|
contractAddress: AztecAddress,
|
|
349
417
|
tags: Tag[],
|
|
350
418
|
page: number = 0,
|
|
419
|
+
upToBlockNumber?: BlockNumber,
|
|
351
420
|
): Promise<TxScopedL2Log[][]> {
|
|
352
421
|
const logs = await Promise.all(
|
|
353
422
|
tags.map(tag => {
|
|
@@ -358,9 +427,16 @@ export class LogStore {
|
|
|
358
427
|
const start = page * MAX_LOGS_PER_TAG;
|
|
359
428
|
const end = start + MAX_LOGS_PER_TAG;
|
|
360
429
|
|
|
361
|
-
return logs.map(
|
|
362
|
-
|
|
363
|
-
|
|
430
|
+
return logs.map(logBuffers => {
|
|
431
|
+
const deserialized = logBuffers?.slice(start, end).map(buf => TxScopedL2Log.fromBuffer(buf)) ?? [];
|
|
432
|
+
if (upToBlockNumber !== undefined) {
|
|
433
|
+
const cutoff = deserialized.findIndex(log => log.blockNumber > upToBlockNumber);
|
|
434
|
+
if (cutoff !== -1) {
|
|
435
|
+
return deserialized.slice(0, cutoff);
|
|
436
|
+
}
|
|
437
|
+
}
|
|
438
|
+
return deserialized;
|
|
439
|
+
});
|
|
364
440
|
}
|
|
365
441
|
|
|
366
442
|
/**
|
|
@@ -588,11 +664,24 @@ export class LogStore {
|
|
|
588
664
|
txLogs: PublicLog[],
|
|
589
665
|
filter: LogFilter = {},
|
|
590
666
|
): boolean {
|
|
667
|
+
if (filter.fromBlock && blockNumber < filter.fromBlock) {
|
|
668
|
+
return false;
|
|
669
|
+
}
|
|
670
|
+
if (filter.toBlock && blockNumber >= filter.toBlock) {
|
|
671
|
+
return false;
|
|
672
|
+
}
|
|
673
|
+
if (filter.txHash && !txHash.equals(filter.txHash)) {
|
|
674
|
+
return false;
|
|
675
|
+
}
|
|
676
|
+
|
|
591
677
|
let maxLogsHit = false;
|
|
592
678
|
let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
|
|
593
679
|
for (; logIndex < txLogs.length; logIndex++) {
|
|
594
680
|
const log = txLogs[logIndex];
|
|
595
|
-
if (
|
|
681
|
+
if (
|
|
682
|
+
(!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) &&
|
|
683
|
+
(!filter.tag || log.fields[0]?.equals(filter.tag))
|
|
684
|
+
) {
|
|
596
685
|
results.push(
|
|
597
686
|
new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txHash, txIndex, logIndex), log),
|
|
598
687
|
);
|
|
@@ -616,6 +705,16 @@ export class LogStore {
|
|
|
616
705
|
txLogs: ContractClassLog[],
|
|
617
706
|
filter: LogFilter = {},
|
|
618
707
|
): boolean {
|
|
708
|
+
if (filter.fromBlock && blockNumber < filter.fromBlock) {
|
|
709
|
+
return false;
|
|
710
|
+
}
|
|
711
|
+
if (filter.toBlock && blockNumber >= filter.toBlock) {
|
|
712
|
+
return false;
|
|
713
|
+
}
|
|
714
|
+
if (filter.txHash && !txHash.equals(filter.txHash)) {
|
|
715
|
+
return false;
|
|
716
|
+
}
|
|
717
|
+
|
|
619
718
|
let maxLogsHit = false;
|
|
620
719
|
let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
|
|
621
720
|
for (; logIndex < txLogs.length; logIndex++) {
|
|
@@ -14,6 +14,7 @@ import {
|
|
|
14
14
|
} from '@aztec/kv-store';
|
|
15
15
|
import { InboxLeaf } from '@aztec/stdlib/messaging';
|
|
16
16
|
|
|
17
|
+
import { L1ToL2MessagesNotReadyError } from '../errors.js';
|
|
17
18
|
import {
|
|
18
19
|
type InboxMessage,
|
|
19
20
|
deserializeInboxMessage,
|
|
@@ -40,6 +41,8 @@ export class MessageStore {
|
|
|
40
41
|
#lastSynchedL1Block: AztecAsyncSingleton<Buffer>;
|
|
41
42
|
/** Stores total messages stored */
|
|
42
43
|
#totalMessageCount: AztecAsyncSingleton<bigint>;
|
|
44
|
+
/** Stores the checkpoint number whose message tree is currently being filled on L1. */
|
|
45
|
+
#inboxTreeInProgress: AztecAsyncSingleton<bigint>;
|
|
43
46
|
|
|
44
47
|
#log = createLogger('archiver:message_store');
|
|
45
48
|
|
|
@@ -48,6 +51,7 @@ export class MessageStore {
|
|
|
48
51
|
this.#l1ToL2MessageIndices = db.openMap('archiver_l1_to_l2_message_indices');
|
|
49
52
|
this.#lastSynchedL1Block = db.openSingleton('archiver_last_l1_block_id');
|
|
50
53
|
this.#totalMessageCount = db.openSingleton('archiver_l1_to_l2_message_count');
|
|
54
|
+
this.#inboxTreeInProgress = db.openSingleton('archiver_inbox_tree_in_progress');
|
|
51
55
|
}
|
|
52
56
|
|
|
53
57
|
public async getTotalL1ToL2MessageCount(): Promise<bigint> {
|
|
@@ -185,7 +189,22 @@ export class MessageStore {
|
|
|
185
189
|
return msg ? deserializeInboxMessage(msg) : undefined;
|
|
186
190
|
}
|
|
187
191
|
|
|
192
|
+
/** Returns the inbox tree-in-progress checkpoint number from L1, or undefined if not yet set. */
|
|
193
|
+
public getInboxTreeInProgress(): Promise<bigint | undefined> {
|
|
194
|
+
return this.#inboxTreeInProgress.getAsync();
|
|
195
|
+
}
|
|
196
|
+
|
|
197
|
+
/** Persists the inbox tree-in-progress checkpoint number from L1 state. */
|
|
198
|
+
public async setInboxTreeInProgress(value: bigint): Promise<void> {
|
|
199
|
+
await this.#inboxTreeInProgress.set(value);
|
|
200
|
+
}
|
|
201
|
+
|
|
188
202
|
public async getL1ToL2Messages(checkpointNumber: CheckpointNumber): Promise<Fr[]> {
|
|
203
|
+
const treeInProgress = await this.#inboxTreeInProgress.getAsync();
|
|
204
|
+
if (treeInProgress !== undefined && BigInt(checkpointNumber) >= treeInProgress) {
|
|
205
|
+
throw new L1ToL2MessagesNotReadyError(checkpointNumber, treeInProgress);
|
|
206
|
+
}
|
|
207
|
+
|
|
189
208
|
const messages: Fr[] = [];
|
|
190
209
|
|
|
191
210
|
const [startIndex, endIndex] = InboxLeaf.indexRangeForCheckpoint(checkpointNumber);
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
import type { BlobClientInterface } from '@aztec/blob-client/client';
|
|
2
2
|
import { type Blob, getBlobsPerL1Block, getPrefixedEthBlobCommitments } from '@aztec/blob-lib';
|
|
3
|
+
import { INITIAL_CHECKPOINT_NUMBER } from '@aztec/constants';
|
|
3
4
|
import type { CheckpointProposedLog, InboxContract, MessageSentLog, RollupContract } from '@aztec/ethereum/contracts';
|
|
4
5
|
import { MULTI_CALL_3_ADDRESS } from '@aztec/ethereum/contracts';
|
|
5
6
|
import type { ViemPublicClient } from '@aztec/ethereum/types';
|
|
@@ -450,13 +451,22 @@ export class FakeL1State {
|
|
|
450
451
|
createMockInboxContract(_publicClient: MockProxy<ViemPublicClient>): MockProxy<InboxContract> {
|
|
451
452
|
const mockInbox = mock<InboxContract>();
|
|
452
453
|
|
|
453
|
-
mockInbox.getState.mockImplementation(() =>
|
|
454
|
-
|
|
454
|
+
mockInbox.getState.mockImplementation(() => {
|
|
455
|
+
// treeInProgress must be > any sealed checkpoint. On L1, a checkpoint can only be proposed
|
|
456
|
+
// after its messages are sealed, so treeInProgress > checkpointNumber for all published checkpoints.
|
|
457
|
+
const maxFromMessages =
|
|
458
|
+
this.messages.length > 0 ? Math.max(...this.messages.map(m => Number(m.checkpointNumber))) + 1 : 0;
|
|
459
|
+
const maxFromCheckpoints =
|
|
460
|
+
this.checkpoints.length > 0
|
|
461
|
+
? Math.max(...this.checkpoints.filter(cp => !cp.pruned).map(cp => Number(cp.checkpointNumber))) + 1
|
|
462
|
+
: 0;
|
|
463
|
+
const treeInProgress = Math.max(maxFromMessages, maxFromCheckpoints, INITIAL_CHECKPOINT_NUMBER);
|
|
464
|
+
return Promise.resolve({
|
|
455
465
|
messagesRollingHash: this.messagesRollingHash,
|
|
456
466
|
totalMessagesInserted: BigInt(this.messages.length),
|
|
457
|
-
treeInProgress:
|
|
458
|
-
})
|
|
459
|
-
);
|
|
467
|
+
treeInProgress: BigInt(treeInProgress),
|
|
468
|
+
});
|
|
469
|
+
});
|
|
460
470
|
|
|
461
471
|
// Mock the wrapper methods for fetching message events
|
|
462
472
|
mockInbox.getMessageSentEvents.mockImplementation((fromBlock: bigint, toBlock: bigint) =>
|