@aztec/archiver 0.0.1-commit.d3ec352c → 0.0.1-commit.fcb71a6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (122) hide show
  1. package/dest/archiver/archiver.d.ts +84 -70
  2. package/dest/archiver/archiver.d.ts.map +1 -1
  3. package/dest/archiver/archiver.js +439 -228
  4. package/dest/archiver/archiver_store.d.ts +95 -43
  5. package/dest/archiver/archiver_store.d.ts.map +1 -1
  6. package/dest/archiver/archiver_store_test_suite.d.ts +1 -1
  7. package/dest/archiver/archiver_store_test_suite.d.ts.map +1 -1
  8. package/dest/archiver/archiver_store_test_suite.js +1847 -366
  9. package/dest/archiver/config.d.ts +5 -4
  10. package/dest/archiver/config.d.ts.map +1 -1
  11. package/dest/archiver/config.js +10 -3
  12. package/dest/archiver/errors.d.ts +25 -1
  13. package/dest/archiver/errors.d.ts.map +1 -1
  14. package/dest/archiver/errors.js +37 -0
  15. package/dest/archiver/index.d.ts +2 -2
  16. package/dest/archiver/index.d.ts.map +1 -1
  17. package/dest/archiver/instrumentation.d.ts +3 -1
  18. package/dest/archiver/instrumentation.d.ts.map +1 -1
  19. package/dest/archiver/instrumentation.js +11 -0
  20. package/dest/archiver/kv_archiver_store/block_store.d.ts +50 -18
  21. package/dest/archiver/kv_archiver_store/block_store.d.ts.map +1 -1
  22. package/dest/archiver/kv_archiver_store/block_store.js +320 -84
  23. package/dest/archiver/kv_archiver_store/contract_class_store.d.ts +2 -2
  24. package/dest/archiver/kv_archiver_store/contract_class_store.d.ts.map +1 -1
  25. package/dest/archiver/kv_archiver_store/contract_class_store.js +1 -1
  26. package/dest/archiver/kv_archiver_store/contract_instance_store.d.ts +2 -2
  27. package/dest/archiver/kv_archiver_store/contract_instance_store.d.ts.map +1 -1
  28. package/dest/archiver/kv_archiver_store/kv_archiver_store.d.ts +40 -51
  29. package/dest/archiver/kv_archiver_store/kv_archiver_store.d.ts.map +1 -1
  30. package/dest/archiver/kv_archiver_store/kv_archiver_store.js +65 -48
  31. package/dest/archiver/kv_archiver_store/log_store.d.ts +12 -16
  32. package/dest/archiver/kv_archiver_store/log_store.d.ts.map +1 -1
  33. package/dest/archiver/kv_archiver_store/log_store.js +148 -84
  34. package/dest/archiver/kv_archiver_store/message_store.d.ts +6 -5
  35. package/dest/archiver/kv_archiver_store/message_store.d.ts.map +1 -1
  36. package/dest/archiver/kv_archiver_store/message_store.js +15 -14
  37. package/dest/archiver/l1/bin/retrieve-calldata.d.ts +3 -0
  38. package/dest/archiver/l1/bin/retrieve-calldata.d.ts.map +1 -0
  39. package/dest/archiver/l1/bin/retrieve-calldata.js +149 -0
  40. package/dest/archiver/l1/calldata_retriever.d.ts +112 -0
  41. package/dest/archiver/l1/calldata_retriever.d.ts.map +1 -0
  42. package/dest/archiver/l1/calldata_retriever.js +471 -0
  43. package/dest/archiver/l1/data_retrieval.d.ts +90 -0
  44. package/dest/archiver/l1/data_retrieval.d.ts.map +1 -0
  45. package/dest/archiver/{data_retrieval.js → l1/data_retrieval.js} +50 -106
  46. package/dest/archiver/l1/debug_tx.d.ts +19 -0
  47. package/dest/archiver/l1/debug_tx.d.ts.map +1 -0
  48. package/dest/archiver/l1/debug_tx.js +73 -0
  49. package/dest/archiver/l1/spire_proposer.d.ts +70 -0
  50. package/dest/archiver/l1/spire_proposer.d.ts.map +1 -0
  51. package/dest/archiver/l1/spire_proposer.js +157 -0
  52. package/dest/archiver/l1/trace_tx.d.ts +97 -0
  53. package/dest/archiver/l1/trace_tx.d.ts.map +1 -0
  54. package/dest/archiver/l1/trace_tx.js +91 -0
  55. package/dest/archiver/l1/types.d.ts +12 -0
  56. package/dest/archiver/l1/types.d.ts.map +1 -0
  57. package/dest/archiver/l1/types.js +3 -0
  58. package/dest/archiver/l1/validate_trace.d.ts +29 -0
  59. package/dest/archiver/l1/validate_trace.d.ts.map +1 -0
  60. package/dest/archiver/l1/validate_trace.js +150 -0
  61. package/dest/archiver/structs/inbox_message.d.ts +4 -4
  62. package/dest/archiver/structs/inbox_message.d.ts.map +1 -1
  63. package/dest/archiver/structs/inbox_message.js +6 -6
  64. package/dest/archiver/structs/published.d.ts +1 -2
  65. package/dest/archiver/structs/published.d.ts.map +1 -1
  66. package/dest/factory.d.ts +1 -1
  67. package/dest/factory.js +1 -1
  68. package/dest/index.d.ts +2 -2
  69. package/dest/index.d.ts.map +1 -1
  70. package/dest/index.js +1 -1
  71. package/dest/test/mock_archiver.d.ts +4 -5
  72. package/dest/test/mock_archiver.d.ts.map +1 -1
  73. package/dest/test/mock_archiver.js +5 -9
  74. package/dest/test/mock_l1_to_l2_message_source.d.ts +5 -6
  75. package/dest/test/mock_l1_to_l2_message_source.d.ts.map +1 -1
  76. package/dest/test/mock_l1_to_l2_message_source.js +7 -11
  77. package/dest/test/mock_l2_block_source.d.ts +11 -4
  78. package/dest/test/mock_l2_block_source.d.ts.map +1 -1
  79. package/dest/test/mock_l2_block_source.js +18 -17
  80. package/dest/test/mock_structs.d.ts +3 -2
  81. package/dest/test/mock_structs.d.ts.map +1 -1
  82. package/dest/test/mock_structs.js +9 -9
  83. package/package.json +15 -14
  84. package/src/archiver/archiver.ts +567 -290
  85. package/src/archiver/archiver_store.ts +104 -42
  86. package/src/archiver/archiver_store_test_suite.ts +1895 -347
  87. package/src/archiver/config.ts +15 -10
  88. package/src/archiver/errors.ts +64 -0
  89. package/src/archiver/index.ts +1 -1
  90. package/src/archiver/instrumentation.ts +14 -0
  91. package/src/archiver/kv_archiver_store/block_store.ts +435 -95
  92. package/src/archiver/kv_archiver_store/contract_class_store.ts +1 -1
  93. package/src/archiver/kv_archiver_store/contract_instance_store.ts +1 -1
  94. package/src/archiver/kv_archiver_store/kv_archiver_store.ts +81 -66
  95. package/src/archiver/kv_archiver_store/log_store.ts +208 -99
  96. package/src/archiver/kv_archiver_store/message_store.ts +21 -18
  97. package/src/archiver/l1/README.md +98 -0
  98. package/src/archiver/l1/bin/retrieve-calldata.ts +182 -0
  99. package/src/archiver/l1/calldata_retriever.ts +641 -0
  100. package/src/archiver/{data_retrieval.ts → l1/data_retrieval.ts} +96 -161
  101. package/src/archiver/l1/debug_tx.ts +99 -0
  102. package/src/archiver/l1/spire_proposer.ts +160 -0
  103. package/src/archiver/l1/trace_tx.ts +128 -0
  104. package/src/archiver/l1/types.ts +13 -0
  105. package/src/archiver/l1/validate_trace.ts +211 -0
  106. package/src/archiver/structs/inbox_message.ts +7 -8
  107. package/src/archiver/structs/published.ts +0 -1
  108. package/src/factory.ts +1 -1
  109. package/src/index.ts +1 -1
  110. package/src/test/fixtures/debug_traceTransaction-multicall3.json +88 -0
  111. package/src/test/fixtures/debug_traceTransaction-multiplePropose.json +153 -0
  112. package/src/test/fixtures/debug_traceTransaction-proxied.json +122 -0
  113. package/src/test/fixtures/trace_transaction-multicall3.json +65 -0
  114. package/src/test/fixtures/trace_transaction-multiplePropose.json +319 -0
  115. package/src/test/fixtures/trace_transaction-proxied.json +128 -0
  116. package/src/test/fixtures/trace_transaction-randomRevert.json +216 -0
  117. package/src/test/mock_archiver.ts +6 -11
  118. package/src/test/mock_l1_to_l2_message_source.ts +6 -11
  119. package/src/test/mock_l2_block_source.ts +22 -18
  120. package/src/test/mock_structs.ts +10 -10
  121. package/dest/archiver/data_retrieval.d.ts +0 -80
  122. package/dest/archiver/data_retrieval.d.ts.map +0 -1
@@ -1,16 +1,21 @@
1
- import { INITIAL_L2_BLOCK_NUM, MAX_NOTE_HASHES_PER_TX } from '@aztec/constants';
1
+ import { INITIAL_L2_BLOCK_NUM } from '@aztec/constants';
2
2
  import { BlockNumber } from '@aztec/foundation/branded-types';
3
+ import { Fr } from '@aztec/foundation/curves/bn254';
3
4
  import { createLogger } from '@aztec/foundation/log';
4
5
  import { BufferReader, numToUInt32BE } from '@aztec/foundation/serialize';
5
- import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, PrivateLog, PublicLog, TxScopedL2Log } from '@aztec/stdlib/logs';
6
+ import { L2BlockHash } from '@aztec/stdlib/block';
7
+ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, PublicLog, TxScopedL2Log } from '@aztec/stdlib/logs';
6
8
  /**
7
9
  * A store for logs
8
10
  */ export class LogStore {
9
11
  db;
10
12
  blockStore;
11
- #logsByTag;
12
- #logTagsByBlock;
13
- #privateLogsByBlock;
13
+ // `tag` --> private logs
14
+ #privateLogsByTag;
15
+ // `{contractAddress}_${tag}` --> public logs
16
+ #publicLogsByContractAndTag;
17
+ #privateLogKeysByBlock;
18
+ #publicLogKeysByBlock;
14
19
  #publicLogsByBlock;
15
20
  #contractClassLogsByBlock;
16
21
  #logsMaxPageSize;
@@ -19,68 +24,118 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
19
24
  this.db = db;
20
25
  this.blockStore = blockStore;
21
26
  this.#log = createLogger('archiver:log_store');
22
- this.#logsByTag = db.openMap('archiver_tagged_logs_by_tag');
23
- this.#logTagsByBlock = db.openMap('archiver_log_tags_by_block');
24
- this.#privateLogsByBlock = db.openMap('archiver_private_logs_by_block');
27
+ this.#privateLogsByTag = db.openMap('archiver_private_tagged_logs_by_tag');
28
+ this.#publicLogsByContractAndTag = db.openMap('archiver_public_tagged_logs_by_tag');
29
+ this.#privateLogKeysByBlock = db.openMap('archiver_private_log_keys_by_block');
30
+ this.#publicLogKeysByBlock = db.openMap('archiver_public_log_keys_by_block');
25
31
  this.#publicLogsByBlock = db.openMap('archiver_public_logs_by_block');
26
32
  this.#contractClassLogsByBlock = db.openMap('archiver_contract_class_logs_by_block');
27
33
  this.#logsMaxPageSize = logsMaxPageSize;
28
34
  }
29
- #extractTaggedLogs(block) {
30
- const taggedLogs = new Map();
31
- const dataStartIndexForBlock = block.header.state.partial.noteHashTree.nextAvailableLeafIndex - block.body.txEffects.length * MAX_NOTE_HASHES_PER_TX;
32
- block.body.txEffects.forEach((txEffect, txIndex)=>{
35
+ /**
36
+ * Extracts tagged logs from a single block, grouping them into private and public maps.
37
+ *
38
+ * @param block - The L2 block to extract logs from.
39
+ * @returns An object containing the private and public tagged logs for the block.
40
+ */ #extractTaggedLogsFromBlock(block) {
41
+ // SiloedTag (as string) -> array of log buffers.
42
+ const privateTaggedLogs = new Map();
43
+ // "{contractAddress}_{tag}" (as string) -> array of log buffers.
44
+ const publicTaggedLogs = new Map();
45
+ block.body.txEffects.forEach((txEffect)=>{
33
46
  const txHash = txEffect.txHash;
34
- const dataStartIndexForTx = dataStartIndexForBlock + txIndex * MAX_NOTE_HASHES_PER_TX;
35
- txEffect.privateLogs.forEach((log, logIndex)=>{
47
+ txEffect.privateLogs.forEach((log)=>{
48
+ // Private logs use SiloedTag (already siloed by kernel)
36
49
  const tag = log.fields[0];
37
50
  this.#log.debug(`Found private log with tag ${tag.toString()} in block ${block.number}`);
38
- const currentLogs = taggedLogs.get(tag.toString()) ?? [];
39
- currentLogs.push(new TxScopedL2Log(txHash, dataStartIndexForTx, logIndex, block.number, log).toBuffer());
40
- taggedLogs.set(tag.toString(), currentLogs);
51
+ const currentLogs = privateTaggedLogs.get(tag.toString()) ?? [];
52
+ currentLogs.push(new TxScopedL2Log(txHash, block.number, block.timestamp, log.getEmittedFields(), txEffect.noteHashes, txEffect.nullifiers[0]).toBuffer());
53
+ privateTaggedLogs.set(tag.toString(), currentLogs);
41
54
  });
42
- txEffect.publicLogs.forEach((log, logIndex)=>{
55
+ txEffect.publicLogs.forEach((log)=>{
56
+ // Public logs use Tag directly (not siloed) and are stored with contract address
43
57
  const tag = log.fields[0];
44
- this.#log.debug(`Found public log with tag ${tag.toString()} in block ${block.number}`);
45
- const currentLogs = taggedLogs.get(tag.toString()) ?? [];
46
- currentLogs.push(new TxScopedL2Log(txHash, dataStartIndexForTx, logIndex, block.number, log).toBuffer());
47
- taggedLogs.set(tag.toString(), currentLogs);
58
+ const contractAddress = log.contractAddress;
59
+ const key = `${contractAddress.toString()}_${tag.toString()}`;
60
+ this.#log.debug(`Found public log with tag ${tag.toString()} from contract ${contractAddress.toString()} in block ${block.number}`);
61
+ const currentLogs = publicTaggedLogs.get(key) ?? [];
62
+ currentLogs.push(new TxScopedL2Log(txHash, block.number, block.timestamp, log.getEmittedFields(), txEffect.noteHashes, txEffect.nullifiers[0]).toBuffer());
63
+ publicTaggedLogs.set(key, currentLogs);
48
64
  });
49
65
  });
50
- return taggedLogs;
66
+ return {
67
+ privateTaggedLogs,
68
+ publicTaggedLogs
69
+ };
51
70
  }
52
71
  /**
53
- * Append new logs to the store's list.
54
- * @param blocks - The blocks for which to add the logs.
55
- * @returns True if the operation is successful.
56
- */ addLogs(blocks) {
57
- const taggedLogsToAdd = blocks.map((block)=>this.#extractTaggedLogs(block)).reduce((acc, val)=>{
58
- for (const [tag, logs] of val.entries()){
72
+ * Extracts and aggregates tagged logs from a list of blocks.
73
+ * @param blocks - The blocks to extract logs from.
74
+ * @returns A map from tag (as string) to an array of serialized private logs belonging to that tag, and a map from
75
+ * "{contractAddress}_{tag}" (as string) to an array of serialized public logs belonging to that key.
76
+ */ #extractTaggedLogs(blocks) {
77
+ const taggedLogsInBlocks = blocks.map((block)=>this.#extractTaggedLogsFromBlock(block));
78
+ // Now we merge the maps from each block into a single map.
79
+ const privateTaggedLogs = taggedLogsInBlocks.reduce((acc, { privateTaggedLogs })=>{
80
+ for (const [tag, logs] of privateTaggedLogs.entries()){
59
81
  const currentLogs = acc.get(tag) ?? [];
60
82
  acc.set(tag, currentLogs.concat(logs));
61
83
  }
62
84
  return acc;
63
85
  }, new Map());
64
- const tagsToUpdate = Array.from(taggedLogsToAdd.keys());
86
+ const publicTaggedLogs = taggedLogsInBlocks.reduce((acc, { publicTaggedLogs })=>{
87
+ for (const [key, logs] of publicTaggedLogs.entries()){
88
+ const currentLogs = acc.get(key) ?? [];
89
+ acc.set(key, currentLogs.concat(logs));
90
+ }
91
+ return acc;
92
+ }, new Map());
93
+ return {
94
+ privateTaggedLogs,
95
+ publicTaggedLogs
96
+ };
97
+ }
98
+ /**
99
+ * Append new logs to the store's list.
100
+ * @param blocks - The blocks for which to add the logs.
101
+ * @returns True if the operation is successful.
102
+ */ addLogs(blocks) {
103
+ const { privateTaggedLogs, publicTaggedLogs } = this.#extractTaggedLogs(blocks);
104
+ const keysOfPrivateLogsToUpdate = Array.from(privateTaggedLogs.keys());
105
+ const keysOfPublicLogsToUpdate = Array.from(publicTaggedLogs.keys());
65
106
  return this.db.transactionAsync(async ()=>{
66
- const currentTaggedLogs = await Promise.all(tagsToUpdate.map(async (tag)=>({
67
- tag,
68
- logBuffers: await this.#logsByTag.getAsync(tag)
107
+ const currentPrivateTaggedLogs = await Promise.all(keysOfPrivateLogsToUpdate.map(async (key)=>({
108
+ tag: key,
109
+ logBuffers: await this.#privateLogsByTag.getAsync(key)
110
+ })));
111
+ currentPrivateTaggedLogs.forEach((taggedLogBuffer)=>{
112
+ if (taggedLogBuffer.logBuffers && taggedLogBuffer.logBuffers.length > 0) {
113
+ privateTaggedLogs.set(taggedLogBuffer.tag, taggedLogBuffer.logBuffers.concat(privateTaggedLogs.get(taggedLogBuffer.tag)));
114
+ }
115
+ });
116
+ const currentPublicTaggedLogs = await Promise.all(keysOfPublicLogsToUpdate.map(async (key)=>({
117
+ key,
118
+ logBuffers: await this.#publicLogsByContractAndTag.getAsync(key)
69
119
  })));
70
- currentTaggedLogs.forEach((taggedLogBuffer)=>{
120
+ currentPublicTaggedLogs.forEach((taggedLogBuffer)=>{
71
121
  if (taggedLogBuffer.logBuffers && taggedLogBuffer.logBuffers.length > 0) {
72
- taggedLogsToAdd.set(taggedLogBuffer.tag, taggedLogBuffer.logBuffers.concat(taggedLogsToAdd.get(taggedLogBuffer.tag)));
122
+ publicTaggedLogs.set(taggedLogBuffer.key, taggedLogBuffer.logBuffers.concat(publicTaggedLogs.get(taggedLogBuffer.key)));
73
123
  }
74
124
  });
75
125
  for (const block of blocks){
76
- const tagsInBlock = [];
77
- for (const [tag, logs] of taggedLogsToAdd.entries()){
78
- await this.#logsByTag.set(tag, logs);
79
- tagsInBlock.push(tag);
126
+ const blockHash = await block.hash();
127
+ const privateTagsInBlock = [];
128
+ for (const [tag, logs] of privateTaggedLogs.entries()){
129
+ await this.#privateLogsByTag.set(tag, logs);
130
+ privateTagsInBlock.push(tag);
131
+ }
132
+ await this.#privateLogKeysByBlock.set(block.number, privateTagsInBlock);
133
+ const publicKeysInBlock = [];
134
+ for (const [key, logs] of publicTaggedLogs.entries()){
135
+ await this.#publicLogsByContractAndTag.set(key, logs);
136
+ publicKeysInBlock.push(key);
80
137
  }
81
- await this.#logTagsByBlock.set(block.number, tagsInBlock);
82
- const privateLogsInBlock = block.body.txEffects.map((txEffect)=>txEffect.privateLogs).flat().map((log)=>log.toBuffer());
83
- await this.#privateLogsByBlock.set(block.number, Buffer.concat(privateLogsInBlock));
138
+ await this.#publicLogKeysByBlock.set(block.number, publicKeysInBlock);
84
139
  const publicLogsInBlock = block.body.txEffects.map((txEffect, txIndex)=>[
85
140
  numToUInt32BE(txIndex),
86
141
  numToUInt32BE(txEffect.publicLogs.length),
@@ -91,57 +146,60 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
91
146
  numToUInt32BE(txEffect.contractClassLogs.length),
92
147
  txEffect.contractClassLogs.map((log)=>log.toBuffer())
93
148
  ].flat()).flat();
94
- await this.#publicLogsByBlock.set(block.number, Buffer.concat(publicLogsInBlock));
95
- await this.#contractClassLogsByBlock.set(block.number, Buffer.concat(contractClassLogsInBlock));
149
+ await this.#publicLogsByBlock.set(block.number, this.#packWithBlockHash(blockHash, publicLogsInBlock));
150
+ await this.#contractClassLogsByBlock.set(block.number, this.#packWithBlockHash(blockHash, contractClassLogsInBlock));
96
151
  }
97
152
  return true;
98
153
  });
99
154
  }
155
+ #packWithBlockHash(blockHash, data) {
156
+ return Buffer.concat([
157
+ blockHash.toBuffer(),
158
+ ...data
159
+ ]);
160
+ }
161
+ #unpackBlockHash(reader) {
162
+ const blockHash = reader.remainingBytes() > 0 ? reader.readObject(Fr) : undefined;
163
+ if (!blockHash) {
164
+ throw new Error('Failed to read block hash from log entry buffer');
165
+ }
166
+ return L2BlockHash.fromField(blockHash);
167
+ }
100
168
  deleteLogs(blocks) {
101
169
  return this.db.transactionAsync(async ()=>{
102
- const tagsToDelete = (await Promise.all(blocks.map(async (block)=>{
103
- const tags = await this.#logTagsByBlock.getAsync(block.number);
104
- return tags ?? [];
105
- }))).flat();
170
+ await Promise.all(blocks.map(async (block)=>{
171
+ // Delete private logs
172
+ const privateKeys = await this.#privateLogKeysByBlock.getAsync(block.number) ?? [];
173
+ await Promise.all(privateKeys.map((tag)=>this.#privateLogsByTag.delete(tag)));
174
+ // Delete public logs
175
+ const publicKeys = await this.#publicLogKeysByBlock.getAsync(block.number) ?? [];
176
+ await Promise.all(publicKeys.map((key)=>this.#publicLogsByContractAndTag.delete(key)));
177
+ }));
106
178
  await Promise.all(blocks.map((block)=>Promise.all([
107
- this.#privateLogsByBlock.delete(block.number),
108
179
  this.#publicLogsByBlock.delete(block.number),
109
- this.#logTagsByBlock.delete(block.number),
180
+ this.#privateLogKeysByBlock.delete(block.number),
181
+ this.#publicLogKeysByBlock.delete(block.number),
110
182
  this.#contractClassLogsByBlock.delete(block.number)
111
183
  ])));
112
- await Promise.all(tagsToDelete.map((tag)=>this.#logsByTag.delete(tag.toString())));
113
184
  return true;
114
185
  });
115
186
  }
116
187
  /**
117
- * Retrieves all private logs from up to `limit` blocks, starting from the block number `start`.
118
- * @param start - The block number from which to begin retrieving logs.
119
- * @param limit - The maximum number of blocks to retrieve logs from.
120
- * @returns An array of private logs from the specified range of blocks.
121
- */ async getPrivateLogs(start, limit) {
122
- const logs = [];
123
- for await (const buffer of this.#privateLogsByBlock.valuesAsync({
124
- start,
125
- limit
126
- })){
127
- const reader = new BufferReader(buffer);
128
- while(reader.remainingBytes() > 0){
129
- logs.push(reader.readObject(PrivateLog));
130
- }
131
- }
132
- return logs;
188
+ * Gets all private logs that match any of the `tags`. For each tag, an array of matching logs is returned. An empty
189
+ * array implies no logs match that tag.
190
+ */ async getPrivateLogsByTags(tags) {
191
+ const logs = await Promise.all(tags.map((tag)=>this.#privateLogsByTag.getAsync(tag.toString())));
192
+ return logs.map((logBuffers)=>logBuffers?.map((logBuffer)=>TxScopedL2Log.fromBuffer(logBuffer)) ?? []);
133
193
  }
134
194
  /**
135
- * Gets all logs that match any of the received tags (i.e. logs with their first field equal to a tag).
136
- * @param tags - The tags to filter the logs by.
137
- * @returns For each received tag, an array of matching logs is returned. An empty array implies no logs match
138
- * that tag.
139
- */ async getLogsByTags(tags, limitPerTag) {
140
- if (limitPerTag !== undefined && limitPerTag <= 0) {
141
- throw new TypeError('limitPerTag needs to be greater than 0');
142
- }
143
- const logs = await Promise.all(tags.map((tag)=>this.#logsByTag.getAsync(tag.toString())));
144
- return logs.map((logBuffers)=>logBuffers?.slice(0, limitPerTag).map((logBuffer)=>TxScopedL2Log.fromBuffer(logBuffer)) ?? []);
195
+ * Gets all public logs that match any of the `tags` from the specified contract. For each tag, an array of matching
196
+ * logs is returned. An empty array implies no logs match that tag.
197
+ */ async getPublicLogsByTagsFromContract(contractAddress, tags) {
198
+ const logs = await Promise.all(tags.map((tag)=>{
199
+ const key = `${contractAddress.toString()}_${tag.value.toString()}`;
200
+ return this.#publicLogsByContractAndTag.getAsync(key);
201
+ }));
202
+ return logs.map((logBuffers)=>logBuffers?.map((logBuffer)=>TxScopedL2Log.fromBuffer(logBuffer)) ?? []);
145
203
  }
146
204
  /**
147
205
  * Gets public logs based on the provided filter.
@@ -172,6 +230,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
172
230
  []
173
231
  ];
174
232
  const reader = new BufferReader(buffer);
233
+ const blockHash = this.#unpackBlockHash(reader);
175
234
  while(reader.remainingBytes() > 0){
176
235
  const indexOfTx = reader.readNumber();
177
236
  const numLogsInTx = reader.readNumber();
@@ -182,7 +241,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
182
241
  }
183
242
  const txLogs = publicLogsInBlock[txIndex];
184
243
  const logs = [];
185
- const maxLogsHit = this.#accumulateLogs(logs, blockNumber, txIndex, txLogs, filter);
244
+ const maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
186
245
  return {
187
246
  logs,
188
247
  maxLogsHit
@@ -207,6 +266,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
207
266
  []
208
267
  ];
209
268
  const reader = new BufferReader(logBuffer);
269
+ const blockHash = this.#unpackBlockHash(reader);
210
270
  while(reader.remainingBytes() > 0){
211
271
  const indexOfTx = reader.readNumber();
212
272
  const numLogsInTx = reader.readNumber();
@@ -217,7 +277,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
217
277
  }
218
278
  for(let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < publicLogsInBlock.length; txIndex++){
219
279
  const txLogs = publicLogsInBlock[txIndex];
220
- maxLogsHit = this.#accumulateLogs(logs, blockNumber, txIndex, txLogs, filter);
280
+ maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
221
281
  if (maxLogsHit) {
222
282
  this.#log.debug(`Max logs hit at block ${blockNumber}`);
223
283
  break loopOverBlocks;
@@ -258,6 +318,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
258
318
  []
259
319
  ];
260
320
  const reader = new BufferReader(contractClassLogsBuffer);
321
+ const blockHash = this.#unpackBlockHash(reader);
261
322
  while(reader.remainingBytes() > 0){
262
323
  const indexOfTx = reader.readNumber();
263
324
  const numLogsInTx = reader.readNumber();
@@ -268,7 +329,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
268
329
  }
269
330
  const txLogs = contractClassLogsInBlock[txIndex];
270
331
  const logs = [];
271
- const maxLogsHit = this.#accumulateLogs(logs, blockNumber, txIndex, txLogs, filter);
332
+ const maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
272
333
  return {
273
334
  logs,
274
335
  maxLogsHit
@@ -293,6 +354,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
293
354
  []
294
355
  ];
295
356
  const reader = new BufferReader(logBuffer);
357
+ const blockHash = this.#unpackBlockHash(reader);
296
358
  while(reader.remainingBytes() > 0){
297
359
  const indexOfTx = reader.readNumber();
298
360
  const numLogsInTx = reader.readNumber();
@@ -303,7 +365,7 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
303
365
  }
304
366
  for(let txIndex = filter.afterLog?.txIndex ?? 0; txIndex < contractClassLogsInBlock.length; txIndex++){
305
367
  const txLogs = contractClassLogsInBlock[txIndex];
306
- maxLogsHit = this.#accumulateLogs(logs, blockNumber, txIndex, txLogs, filter);
368
+ maxLogsHit = this.#accumulateLogs(logs, blockNumber, blockHash, txIndex, txLogs, filter);
307
369
  if (maxLogsHit) {
308
370
  this.#log.debug(`Max logs hit at block ${blockNumber}`);
309
371
  break loopOverBlocks;
@@ -315,16 +377,18 @@ import { ContractClassLog, ExtendedContractClassLog, ExtendedPublicLog, LogId, P
315
377
  maxLogsHit
316
378
  };
317
379
  }
318
- #accumulateLogs(results, blockNumber, txIndex, txLogs, filter) {
380
+ #accumulateLogs(results, blockNumber, blockHash, txIndex, txLogs, filter = {}) {
319
381
  let maxLogsHit = false;
320
382
  let logIndex = typeof filter.afterLog?.logIndex === 'number' ? filter.afterLog.logIndex + 1 : 0;
321
383
  for(; logIndex < txLogs.length; logIndex++){
322
384
  const log = txLogs[logIndex];
323
385
  if (!filter.contractAddress || log.contractAddress.equals(filter.contractAddress)) {
324
386
  if (log instanceof ContractClassLog) {
325
- results.push(new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), txIndex, logIndex), log));
387
+ results.push(new ExtendedContractClassLog(new LogId(BlockNumber(blockNumber), blockHash, txIndex, logIndex), log));
388
+ } else if (log instanceof PublicLog) {
389
+ results.push(new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), blockHash, txIndex, logIndex), log));
326
390
  } else {
327
- results.push(new ExtendedPublicLog(new LogId(BlockNumber(blockNumber), txIndex, logIndex), log));
391
+ throw new Error('Unknown log type');
328
392
  }
329
393
  if (results.length >= this.#logsMaxPageSize) {
330
394
  maxLogsHit = true;
@@ -1,5 +1,6 @@
1
- import type { L1BlockId } from '@aztec/ethereum';
2
- import { Fr } from '@aztec/foundation/fields';
1
+ import type { L1BlockId } from '@aztec/ethereum/l1-types';
2
+ import { CheckpointNumber } from '@aztec/foundation/branded-types';
3
+ import { Fr } from '@aztec/foundation/curves/bn254';
3
4
  import { type AztecAsyncKVStore, type CustomRange } from '@aztec/kv-store';
4
5
  import { type InboxMessage } from '../structs/inbox_message.js';
5
6
  export declare class MessageStoreError extends Error {
@@ -28,12 +29,12 @@ export declare class MessageStore {
28
29
  */
29
30
  getL1ToL2MessageIndex(l1ToL2Message: Fr): Promise<bigint | undefined>;
30
31
  getLastMessage(): Promise<InboxMessage | undefined>;
31
- getL1ToL2Messages(blockNumber: number): Promise<Fr[]>;
32
+ getL1ToL2Messages(checkpointNumber: CheckpointNumber): Promise<Fr[]>;
32
33
  iterateL1ToL2Messages(range?: CustomRange<bigint>): AsyncIterableIterator<InboxMessage>;
33
34
  removeL1ToL2Messages(startIndex: bigint): Promise<void>;
34
- rollbackL1ToL2MessagesToL2Block(targetBlockNumber: number): Promise<void>;
35
+ rollbackL1ToL2MessagesToCheckpoint(targetCheckpointNumber: CheckpointNumber): Promise<void>;
35
36
  private indexToKey;
36
37
  private leafToIndexKey;
37
38
  private increaseTotalMessageCount;
38
39
  }
39
- //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoibWVzc2FnZV9zdG9yZS5kLnRzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vLi4vLi4vc3JjL2FyY2hpdmVyL2t2X2FyY2hpdmVyX3N0b3JlL21lc3NhZ2Vfc3RvcmUudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQUEsT0FBTyxLQUFLLEVBQUUsU0FBUyxFQUFFLE1BQU0saUJBQWlCLENBQUM7QUFFakQsT0FBTyxFQUFFLEVBQUUsRUFBRSxNQUFNLDBCQUEwQixDQUFDO0FBSTlDLE9BQU8sRUFDTCxLQUFLLGlCQUFpQixFQUd0QixLQUFLLFdBQVcsRUFFakIsTUFBTSxpQkFBaUIsQ0FBQztBQUd6QixPQUFPLEVBQ0wsS0FBSyxZQUFZLEVBSWxCLE1BQU0sNkJBQTZCLENBQUM7QUFFckMscUJBQWEsaUJBQWtCLFNBQVEsS0FBSzthQUd4QixZQUFZLEVBQUUsWUFBWTtJQUY1QyxZQUNFLE9BQU8sRUFBRSxNQUFNLEVBQ0MsWUFBWSxFQUFFLFlBQVksRUFJM0M7Q0FDRjtBQUVELHFCQUFhLFlBQVk7O0lBWVgsT0FBTyxDQUFDLEVBQUU7SUFBdEIsWUFBb0IsRUFBRSxFQUFFLGlCQUFpQixFQUt4QztJQUVZLDBCQUEwQixJQUFJLE9BQU8sQ0FBQyxNQUFNLENBQUMsQ0FFekQ7SUFFRCxxQ0FBcUM7SUFDeEIsaUJBQWlCLElBQUksT0FBTyxDQUFDLFNBQVMsR0FBRyxTQUFTLENBQUMsQ0FRL0Q7SUFFRCxvQ0FBb0M7SUFDdkIsaUJBQWlCLENBQUMsT0FBTyxFQUFFLFNBQVMsR0FBRyxPQUFPLENBQUMsSUFBSSxDQUFDLENBR2hFO0lBRUQ7Ozs7T0FJRztJQUNJLGlCQUFpQixDQUFDLFFBQVEsRUFBRSxZQUFZLEVBQUUsR0FBRyxPQUFPLENBQUMsSUFBSSxDQUFDLENBNkZoRTtJQUVEOzs7O09BSUc7SUFDSSxxQkFBcUIsQ0FBQyxhQUFhLEVBQUUsRUFBRSxHQUFHLE9BQU8sQ0FBQyxNQUFNLEdBQUcsU0FBUyxDQUFDLENBRTNFO0lBRVksY0FBYyxJQUFJLE9BQU8sQ0FBQyxZQUFZLEdBQUcsU0FBUyxDQUFDLENBRy9EO0lBRVksaUJBQWlCLENBQUMsV0FBVyxFQUFFLE1BQU0sR0FBRyxPQUFPLENBQUMsRUFBRSxFQUFFLENBQUMsQ0FxQmpFO0lBRWEscUJBQXFCLENBQUMsS0FBSyxHQUFFLFdBQVcsQ0FBQyxNQUFNLENBQU0sR0FBRyxxQkFBcUIsQ0FBQyxZQUFZLENBQUMsQ0FLeEc7SUFFTSxvQkFBb0IsQ0FBQyxVQUFVLEVBQUUsTUFBTSxHQUFHLE9BQU8sQ0FBQyxJQUFJLENBQUMsQ0FnQjdEO0lBRU0sK0JBQStCLENBQUMsaUJBQWlCLEVBQUUsTUFBTSxHQUFHLE9BQU8sQ0FBQyxJQUFJLENBQUMsQ0FJL0U7SUFFRCxPQUFPLENBQUMsVUFBVTtJQUlsQixPQUFPLENBQUMsY0FBYztZQUlSLHlCQUF5QjtDQVN4QyJ9
40
+ //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoibWVzc2FnZV9zdG9yZS5kLnRzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vLi4vLi4vc3JjL2FyY2hpdmVyL2t2X2FyY2hpdmVyX3N0b3JlL21lc3NhZ2Vfc3RvcmUudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQUEsT0FBTyxLQUFLLEVBQUUsU0FBUyxFQUFFLE1BQU0sMEJBQTBCLENBQUM7QUFDMUQsT0FBTyxFQUFFLGdCQUFnQixFQUFFLE1BQU0saUNBQWlDLENBQUM7QUFFbkUsT0FBTyxFQUFFLEVBQUUsRUFBRSxNQUFNLGdDQUFnQyxDQUFDO0FBSXBELE9BQU8sRUFDTCxLQUFLLGlCQUFpQixFQUd0QixLQUFLLFdBQVcsRUFFakIsTUFBTSxpQkFBaUIsQ0FBQztBQUd6QixPQUFPLEVBQ0wsS0FBSyxZQUFZLEVBSWxCLE1BQU0sNkJBQTZCLENBQUM7QUFFckMscUJBQWEsaUJBQWtCLFNBQVEsS0FBSzthQUd4QixZQUFZLEVBQUUsWUFBWTtJQUY1QyxZQUNFLE9BQU8sRUFBRSxNQUFNLEVBQ0MsWUFBWSxFQUFFLFlBQVksRUFJM0M7Q0FDRjtBQUVELHFCQUFhLFlBQVk7O0lBWVgsT0FBTyxDQUFDLEVBQUU7SUFBdEIsWUFBb0IsRUFBRSxFQUFFLGlCQUFpQixFQUt4QztJQUVZLDBCQUEwQixJQUFJLE9BQU8sQ0FBQyxNQUFNLENBQUMsQ0FFekQ7SUFFRCxxQ0FBcUM7SUFDeEIsaUJBQWlCLElBQUksT0FBTyxDQUFDLFNBQVMsR0FBRyxTQUFTLENBQUMsQ0FRL0Q7SUFFRCxvQ0FBb0M7SUFDdkIsaUJBQWlCLENBQUMsT0FBTyxFQUFFLFNBQVMsR0FBRyxPQUFPLENBQUMsSUFBSSxDQUFDLENBR2hFO0lBRUQ7Ozs7T0FJRztJQUNJLGlCQUFpQixDQUFDLFFBQVEsRUFBRSxZQUFZLEVBQUUsR0FBRyxPQUFPLENBQUMsSUFBSSxDQUFDLENBNkZoRTtJQUVEOzs7O09BSUc7SUFDSSxxQkFBcUIsQ0FBQyxhQUFhLEVBQUUsRUFBRSxHQUFHLE9BQU8sQ0FBQyxNQUFNLEdBQUcsU0FBUyxDQUFDLENBRTNFO0lBRVksY0FBYyxJQUFJLE9BQU8sQ0FBQyxZQUFZLEdBQUcsU0FBUyxDQUFDLENBRy9EO0lBRVksaUJBQWlCLENBQUMsZ0JBQWdCLEVBQUUsZ0JBQWdCLEdBQUcsT0FBTyxDQUFDLEVBQUUsRUFBRSxDQUFDLENBdUJoRjtJQUVhLHFCQUFxQixDQUFDLEtBQUssR0FBRSxXQUFXLENBQUMsTUFBTSxDQUFNLEdBQUcscUJBQXFCLENBQUMsWUFBWSxDQUFDLENBS3hHO0lBRU0sb0JBQW9CLENBQUMsVUFBVSxFQUFFLE1BQU0sR0FBRyxPQUFPLENBQUMsSUFBSSxDQUFDLENBZ0I3RDtJQUVNLGtDQUFrQyxDQUFDLHNCQUFzQixFQUFFLGdCQUFnQixHQUFHLE9BQU8sQ0FBQyxJQUFJLENBQUMsQ0FJakc7SUFFRCxPQUFPLENBQUMsVUFBVTtJQUlsQixPQUFPLENBQUMsY0FBYztZQUlSLHlCQUF5QjtDQVN4QyJ9
@@ -1 +1 @@
1
- {"version":3,"file":"message_store.d.ts","sourceRoot":"","sources":["../../../src/archiver/kv_archiver_store/message_store.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAC;AAEjD,OAAO,EAAE,EAAE,EAAE,MAAM,0BAA0B,CAAC;AAI9C,OAAO,EACL,KAAK,iBAAiB,EAGtB,KAAK,WAAW,EAEjB,MAAM,iBAAiB,CAAC;AAGzB,OAAO,EACL,KAAK,YAAY,EAIlB,MAAM,6BAA6B,CAAC;AAErC,qBAAa,iBAAkB,SAAQ,KAAK;aAGxB,YAAY,EAAE,YAAY;IAF5C,YACE,OAAO,EAAE,MAAM,EACC,YAAY,EAAE,YAAY,EAI3C;CACF;AAED,qBAAa,YAAY;;IAYX,OAAO,CAAC,EAAE;IAAtB,YAAoB,EAAE,EAAE,iBAAiB,EAKxC;IAEY,0BAA0B,IAAI,OAAO,CAAC,MAAM,CAAC,CAEzD;IAED,qCAAqC;IACxB,iBAAiB,IAAI,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC,CAQ/D;IAED,oCAAoC;IACvB,iBAAiB,CAAC,OAAO,EAAE,SAAS,GAAG,OAAO,CAAC,IAAI,CAAC,CAGhE;IAED;;;;OAIG;IACI,iBAAiB,CAAC,QAAQ,EAAE,YAAY,EAAE,GAAG,OAAO,CAAC,IAAI,CAAC,CA6FhE;IAED;;;;OAIG;IACI,qBAAqB,CAAC,aAAa,EAAE,EAAE,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC,CAE3E;IAEY,cAAc,IAAI,OAAO,CAAC,YAAY,GAAG,SAAS,CAAC,CAG/D;IAEY,iBAAiB,CAAC,WAAW,EAAE,MAAM,GAAG,OAAO,CAAC,EAAE,EAAE,CAAC,CAqBjE;IAEa,qBAAqB,CAAC,KAAK,GAAE,WAAW,CAAC,MAAM,CAAM,GAAG,qBAAqB,CAAC,YAAY,CAAC,CAKxG;IAEM,oBAAoB,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAgB7D;IAEM,+BAA+B,CAAC,iBAAiB,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAI/E;IAED,OAAO,CAAC,UAAU;IAIlB,OAAO,CAAC,cAAc;YAIR,yBAAyB;CASxC"}
1
+ {"version":3,"file":"message_store.d.ts","sourceRoot":"","sources":["../../../src/archiver/kv_archiver_store/message_store.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,0BAA0B,CAAC;AAC1D,OAAO,EAAE,gBAAgB,EAAE,MAAM,iCAAiC,CAAC;AAEnE,OAAO,EAAE,EAAE,EAAE,MAAM,gCAAgC,CAAC;AAIpD,OAAO,EACL,KAAK,iBAAiB,EAGtB,KAAK,WAAW,EAEjB,MAAM,iBAAiB,CAAC;AAGzB,OAAO,EACL,KAAK,YAAY,EAIlB,MAAM,6BAA6B,CAAC;AAErC,qBAAa,iBAAkB,SAAQ,KAAK;aAGxB,YAAY,EAAE,YAAY;IAF5C,YACE,OAAO,EAAE,MAAM,EACC,YAAY,EAAE,YAAY,EAI3C;CACF;AAED,qBAAa,YAAY;;IAYX,OAAO,CAAC,EAAE;IAAtB,YAAoB,EAAE,EAAE,iBAAiB,EAKxC;IAEY,0BAA0B,IAAI,OAAO,CAAC,MAAM,CAAC,CAEzD;IAED,qCAAqC;IACxB,iBAAiB,IAAI,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC,CAQ/D;IAED,oCAAoC;IACvB,iBAAiB,CAAC,OAAO,EAAE,SAAS,GAAG,OAAO,CAAC,IAAI,CAAC,CAGhE;IAED;;;;OAIG;IACI,iBAAiB,CAAC,QAAQ,EAAE,YAAY,EAAE,GAAG,OAAO,CAAC,IAAI,CAAC,CA6FhE;IAED;;;;OAIG;IACI,qBAAqB,CAAC,aAAa,EAAE,EAAE,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC,CAE3E;IAEY,cAAc,IAAI,OAAO,CAAC,YAAY,GAAG,SAAS,CAAC,CAG/D;IAEY,iBAAiB,CAAC,gBAAgB,EAAE,gBAAgB,GAAG,OAAO,CAAC,EAAE,EAAE,CAAC,CAuBhF;IAEa,qBAAqB,CAAC,KAAK,GAAE,WAAW,CAAC,MAAM,CAAM,GAAG,qBAAqB,CAAC,YAAY,CAAC,CAKxG;IAEM,oBAAoB,CAAC,UAAU,EAAE,MAAM,GAAG,OAAO,CAAC,IAAI,CAAC,CAgB7D;IAEM,kCAAkC,CAAC,sBAAsB,EAAE,gBAAgB,GAAG,OAAO,CAAC,IAAI,CAAC,CAIjG;IAED,OAAO,CAAC,UAAU;IAIlB,OAAO,CAAC,cAAc;YAIR,yBAAyB;CASxC"}
@@ -1,3 +1,4 @@
1
+ import { CheckpointNumber } from '@aztec/foundation/branded-types';
1
2
  import { Buffer16, Buffer32 } from '@aztec/foundation/buffer';
2
3
  import { toArray } from '@aztec/foundation/iterable';
3
4
  import { createLogger } from '@aztec/foundation/log';
@@ -77,18 +78,18 @@ export class MessageStore {
77
78
  if (!expectedRollingHash.equals(message.rollingHash)) {
78
79
  throw new MessageStoreError(`Invalid rolling hash for incoming L1 to L2 message ${message.leaf.toString()} ` + `with index ${message.index} ` + `(expected ${expectedRollingHash.toString()} from previous hash ${previousRollingHash} but got ${message.rollingHash.toString()})`, message);
79
80
  }
80
- // Check index corresponds to the L2 block number.
81
- const [expectedStart, expectedEnd] = InboxLeaf.indexRangeFromL2Block(message.l2BlockNumber);
81
+ // Check index corresponds to the checkpoint number.
82
+ const [expectedStart, expectedEnd] = InboxLeaf.indexRangeForCheckpoint(message.checkpointNumber);
82
83
  if (message.index < expectedStart || message.index >= expectedEnd) {
83
- throw new MessageStoreError(`Invalid index ${message.index} for incoming L1 to L2 message ${message.leaf.toString()} ` + `at block ${message.l2BlockNumber} (expected value in range [${expectedStart}, ${expectedEnd}))`, message);
84
+ throw new MessageStoreError(`Invalid index ${message.index} for incoming L1 to L2 message ${message.leaf.toString()} ` + `at checkpoint ${message.checkpointNumber} (expected value in range [${expectedStart}, ${expectedEnd}))`, message);
84
85
  }
85
- // Check there are no gaps in the indices within the same block.
86
- if (lastMessage && message.l2BlockNumber === lastMessage.l2BlockNumber && message.index !== lastMessage.index + 1n) {
86
+ // Check there are no gaps in the indices within the same checkpoint.
87
+ if (lastMessage && message.checkpointNumber === lastMessage.checkpointNumber && message.index !== lastMessage.index + 1n) {
87
88
  throw new MessageStoreError(`Missing prior message for incoming L1 to L2 message ${message.leaf.toString()} ` + `with index ${message.index}`, message);
88
89
  }
89
90
  // Check the first message in a block has the correct index.
90
- if ((!lastMessage || message.l2BlockNumber > lastMessage.l2BlockNumber) && message.index !== InboxLeaf.smallestIndexFromL2Block(message.l2BlockNumber)) {
91
- throw new MessageStoreError(`Message ${message.leaf.toString()} for L2 block ${message.l2BlockNumber} has wrong index ` + `${message.index} (expected ${InboxLeaf.smallestIndexFromL2Block(message.l2BlockNumber)})`, message);
91
+ if ((!lastMessage || message.checkpointNumber > lastMessage.checkpointNumber) && message.index !== expectedStart) {
92
+ throw new MessageStoreError(`Message ${message.leaf.toString()} for checkpoint ${message.checkpointNumber} has wrong index ` + `${message.index} (expected ${expectedStart})`, message);
92
93
  }
93
94
  // Perform the insertions.
94
95
  await this.#l1ToL2Messages.set(this.indexToKey(message.index), serializeInboxMessage(message));
@@ -123,17 +124,17 @@ export class MessageStore {
123
124
  }));
124
125
  return msg ? deserializeInboxMessage(msg) : undefined;
125
126
  }
126
- async getL1ToL2Messages(blockNumber) {
127
+ async getL1ToL2Messages(checkpointNumber) {
127
128
  const messages = [];
128
- const [startIndex, endIndex] = InboxLeaf.indexRangeFromL2Block(blockNumber);
129
+ const [startIndex, endIndex] = InboxLeaf.indexRangeForCheckpoint(checkpointNumber);
129
130
  let lastIndex = startIndex - 1n;
130
131
  for await (const msgBuffer of this.#l1ToL2Messages.valuesAsync({
131
132
  start: this.indexToKey(startIndex),
132
133
  end: this.indexToKey(endIndex)
133
134
  })){
134
135
  const msg = deserializeInboxMessage(msgBuffer);
135
- if (msg.l2BlockNumber !== blockNumber) {
136
- throw new Error(`L1 to L2 message with index ${msg.index} has invalid block number ${msg.l2BlockNumber}`);
136
+ if (msg.checkpointNumber !== checkpointNumber) {
137
+ throw new Error(`L1 to L2 message with index ${msg.index} has invalid checkpoint number ${msg.checkpointNumber}`);
137
138
  } else if (msg.index !== lastIndex + 1n) {
138
139
  throw new Error(`Expected L1 to L2 message with index ${lastIndex + 1n} but got ${msg.index}`);
139
140
  }
@@ -164,9 +165,9 @@ export class MessageStore {
164
165
  this.#log.warn(`Deleted ${deleteCount} L1 to L2 messages from index ${startIndex} from the store`);
165
166
  });
166
167
  }
167
- rollbackL1ToL2MessagesToL2Block(targetBlockNumber) {
168
- this.#log.debug(`Deleting L1 to L2 messages up to target L2 block ${targetBlockNumber}`);
169
- const startIndex = InboxLeaf.smallestIndexFromL2Block(targetBlockNumber + 1);
168
+ rollbackL1ToL2MessagesToCheckpoint(targetCheckpointNumber) {
169
+ this.#log.debug(`Deleting L1 to L2 messages up to target checkpoint ${targetCheckpointNumber}`);
170
+ const startIndex = InboxLeaf.smallestIndexForCheckpoint(CheckpointNumber(targetCheckpointNumber + 1));
170
171
  return this.removeL1ToL2Messages(startIndex);
171
172
  }
172
173
  indexToKey(index) {
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ export {};
3
+ //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicmV0cmlldmUtY2FsbGRhdGEuZC50cyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uLy4uL3NyYy9hcmNoaXZlci9sMS9iaW4vcmV0cmlldmUtY2FsbGRhdGEudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IiJ9
@@ -0,0 +1 @@
1
+ {"version":3,"file":"retrieve-calldata.d.ts","sourceRoot":"","sources":["../../../../src/archiver/l1/bin/retrieve-calldata.ts"],"names":[],"mappings":""}
@@ -0,0 +1,149 @@
1
+ #!/usr/bin/env node
2
+ import { CheckpointNumber } from '@aztec/foundation/branded-types';
3
+ import { EthAddress } from '@aztec/foundation/eth-address';
4
+ import { createLogger } from '@aztec/foundation/log';
5
+ import { createPublicClient, http } from 'viem';
6
+ import { mainnet } from 'viem/chains';
7
+ import { CalldataRetriever } from '../calldata_retriever.js';
8
+ const logger = createLogger('archiver:calldata-test');
9
+ function parseArgs() {
10
+ const args = process.argv.slice(2);
11
+ if (args.length < 2) {
12
+ // eslint-disable-next-line no-console
13
+ console.error('Usage: node index.js <rollup-address> <tx-hash> [target-committee-size]');
14
+ // eslint-disable-next-line no-console
15
+ console.error('');
16
+ // eslint-disable-next-line no-console
17
+ console.error('Environment variables:');
18
+ // eslint-disable-next-line no-console
19
+ console.error(' ETHEREUM_HOST or RPC_URL - Ethereum RPC endpoint');
20
+ // eslint-disable-next-line no-console
21
+ console.error('');
22
+ // eslint-disable-next-line no-console
23
+ console.error('Example:');
24
+ // eslint-disable-next-line no-console
25
+ console.error(' RPC_URL=https://eth-mainnet.g.alchemy.com/v2/YOUR-API-KEY \\');
26
+ // eslint-disable-next-line no-console
27
+ console.error(' node index.js 0x1234... 0xabcd... 32');
28
+ process.exit(1);
29
+ }
30
+ const rollupAddress = EthAddress.fromString(args[0]);
31
+ const txHash = args[1];
32
+ const targetCommitteeSize = args[2] ? parseInt(args[2], 10) : 24;
33
+ const rpcUrl = process.env.ETHEREUM_HOST || process.env.RPC_URL;
34
+ if (!rpcUrl) {
35
+ // eslint-disable-next-line no-console
36
+ console.error('Error: ETHEREUM_HOST or RPC_URL environment variable must be set');
37
+ process.exit(1);
38
+ }
39
+ if (targetCommitteeSize <= 0 || targetCommitteeSize > 256) {
40
+ // eslint-disable-next-line no-console
41
+ console.error('Error: target-committee-size must be between 1 and 256');
42
+ process.exit(1);
43
+ }
44
+ return {
45
+ rollupAddress,
46
+ txHash,
47
+ rpcUrl,
48
+ targetCommitteeSize
49
+ };
50
+ }
51
+ async function main() {
52
+ const { rollupAddress, txHash, rpcUrl, targetCommitteeSize } = parseArgs();
53
+ logger.info('Calldata Retriever Test Script');
54
+ logger.info('===============================');
55
+ logger.info(`Rollup Address: ${rollupAddress.toString()}`);
56
+ logger.info(`Transaction Hash: ${txHash}`);
57
+ logger.info(`RPC URL: ${rpcUrl}`);
58
+ logger.info(`Target Committee Size: ${targetCommitteeSize}`);
59
+ logger.info('');
60
+ try {
61
+ // Create viem public client
62
+ const publicClient = createPublicClient({
63
+ chain: mainnet,
64
+ transport: http(rpcUrl, {
65
+ batch: false
66
+ })
67
+ });
68
+ logger.info('Fetching transaction...');
69
+ const tx = await publicClient.getTransaction({
70
+ hash: txHash
71
+ });
72
+ if (!tx) {
73
+ throw new Error(`Transaction ${txHash} not found`);
74
+ }
75
+ logger.info(`Transaction found in block ${tx.blockNumber}`);
76
+ // For simplicity, use zero addresses for optional contract addresses
77
+ // In production, these would be fetched from the rollup contract or configuration
78
+ const slashingProposerAddress = EthAddress.ZERO;
79
+ const governanceProposerAddress = EthAddress.ZERO;
80
+ const slashFactoryAddress = undefined;
81
+ logger.info('Using zero addresses for governance/slashing (can be configured if needed)');
82
+ // Create CalldataRetriever
83
+ const retriever = new CalldataRetriever(publicClient, publicClient, targetCommitteeSize, undefined, logger, {
84
+ rollupAddress,
85
+ governanceProposerAddress,
86
+ slashingProposerAddress,
87
+ slashFactoryAddress
88
+ });
89
+ // Extract L2 block number from transaction logs
90
+ logger.info('Decoding transaction to extract L2 block number...');
91
+ const receipt = await publicClient.getTransactionReceipt({
92
+ hash: txHash
93
+ });
94
+ const l2BlockProposedEvent = receipt.logs.find((log)=>{
95
+ try {
96
+ // Try to match the L2BlockProposed event
97
+ return log.address.toLowerCase() === rollupAddress.toString().toLowerCase() && log.topics[0] === '0x2f1d0e696fa5186494a2f2f89a0e0bcbb15d607f6c5eac4637e07e1e5e7d3c00' // L2BlockProposed event signature
98
+ ;
99
+ } catch {
100
+ return false;
101
+ }
102
+ });
103
+ let l2BlockNumber;
104
+ if (l2BlockProposedEvent && l2BlockProposedEvent.topics[1]) {
105
+ // L2 block number is typically the first indexed parameter
106
+ l2BlockNumber = Number(BigInt(l2BlockProposedEvent.topics[1]));
107
+ logger.info(`L2 Block Number (from event): ${l2BlockNumber}`);
108
+ } else {
109
+ // Fallback: try to extract from transaction data or use a default
110
+ logger.warn('Could not extract L2 block number from event, using block number as fallback');
111
+ l2BlockNumber = Number(tx.blockNumber);
112
+ }
113
+ logger.info('');
114
+ logger.info('Retrieving block header from rollup transaction...');
115
+ logger.info('');
116
+ // For this script, we don't have blob hashes or expected hashes, so pass empty arrays/objects
117
+ const result = await retriever.getCheckpointFromRollupTx(txHash, [], CheckpointNumber(l2BlockNumber), {});
118
+ logger.info(' Successfully retrieved block header!');
119
+ logger.info('');
120
+ logger.info('Block Header Details:');
121
+ logger.info('====================');
122
+ logger.info(`Checkpoint Number: ${result.checkpointNumber}`);
123
+ logger.info(`Block Hash: ${result.blockHash}`);
124
+ logger.info(`Archive Root: ${result.archiveRoot.toString()}`);
125
+ logger.info('');
126
+ logger.info('Header:');
127
+ logger.info(` Slot Number: ${result.header.slotNumber.toString()}`);
128
+ logger.info(` Timestamp: ${result.header.timestamp.toString()}`);
129
+ logger.info(` Coinbase: ${result.header.coinbase.toString()}`);
130
+ logger.info(` Fee Recipient: ${result.header.feeRecipient.toString()}`);
131
+ logger.info(` Total Mana Used: ${result.header.totalManaUsed.toString()}`);
132
+ logger.info('');
133
+ logger.info('Attestations:');
134
+ logger.info(` Count: ${result.attestations.length}`);
135
+ logger.info(` Non-empty attestations: ${result.attestations.filter((a)=>!a.signature.isEmpty()).length}`);
136
+ process.exit(0);
137
+ } catch (error) {
138
+ logger.error('Error retrieving block header:');
139
+ logger.error(error instanceof Error ? error.message : String(error));
140
+ if (error instanceof Error && error.stack) {
141
+ logger.debug(error.stack);
142
+ }
143
+ process.exit(1);
144
+ }
145
+ }
146
+ // Only run if this is the main module
147
+ if (import.meta.url === `file://${process.argv[1]}`) {
148
+ void main();
149
+ }