@aztec/p2p 0.0.1-commit.f2ce05ee → 0.0.1-commit.f8ca9b2f3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/client/factory.d.ts +3 -3
- package/dest/client/factory.d.ts.map +1 -1
- package/dest/client/factory.js +38 -7
- package/dest/client/interface.d.ts +26 -15
- package/dest/client/interface.d.ts.map +1 -1
- package/dest/client/p2p_client.d.ts +31 -35
- package/dest/client/p2p_client.d.ts.map +1 -1
- package/dest/client/p2p_client.js +97 -138
- package/dest/config.d.ts +10 -2
- package/dest/config.d.ts.map +1 -1
- package/dest/config.js +9 -0
- package/dest/index.d.ts +2 -1
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +1 -0
- package/dest/mem_pools/attestation_pool/attestation_pool.d.ts +18 -11
- package/dest/mem_pools/attestation_pool/attestation_pool.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/attestation_pool.js +62 -37
- package/dest/mem_pools/attestation_pool/attestation_pool_test_suite.d.ts +1 -1
- package/dest/mem_pools/attestation_pool/attestation_pool_test_suite.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/attestation_pool_test_suite.js +53 -53
- package/dest/mem_pools/attestation_pool/index.d.ts +2 -2
- package/dest/mem_pools/attestation_pool/index.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/index.js +1 -1
- package/dest/mem_pools/index.d.ts +2 -1
- package/dest/mem_pools/index.d.ts.map +1 -1
- package/dest/mem_pools/interface.d.ts +3 -3
- package/dest/mem_pools/interface.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool_v2/deleted_pool.d.ts +87 -0
- package/dest/mem_pools/tx_pool_v2/deleted_pool.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool_v2/deleted_pool.js +180 -0
- package/dest/mem_pools/tx_pool_v2/index.d.ts +2 -1
- package/dest/mem_pools/tx_pool_v2/index.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool_v2/index.js +1 -0
- package/dest/mem_pools/tx_pool_v2/interfaces.d.ts +3 -3
- package/dest/mem_pools/tx_pool_v2/interfaces.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool_v2/tx_metadata.d.ts +25 -3
- package/dest/mem_pools/tx_pool_v2/tx_metadata.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool_v2/tx_metadata.js +36 -4
- package/dest/mem_pools/tx_pool_v2/tx_pool_v2_impl.d.ts +1 -1
- package/dest/mem_pools/tx_pool_v2/tx_pool_v2_impl.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool_v2/tx_pool_v2_impl.js +77 -74
- package/dest/msg_validators/tx_validator/aggregate_tx_validator.d.ts +3 -3
- package/dest/msg_validators/tx_validator/aggregate_tx_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/block_header_validator.d.ts +16 -3
- package/dest/msg_validators/tx_validator/block_header_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/block_header_validator.js +1 -1
- package/dest/msg_validators/tx_validator/double_spend_validator.d.ts +13 -3
- package/dest/msg_validators/tx_validator/double_spend_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/double_spend_validator.js +4 -4
- package/dest/msg_validators/tx_validator/timestamp_validator.d.ts +20 -4
- package/dest/msg_validators/tx_validator/timestamp_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/timestamp_validator.js +2 -2
- package/dest/services/dummy_service.d.ts +6 -2
- package/dest/services/dummy_service.d.ts.map +1 -1
- package/dest/services/dummy_service.js +3 -0
- package/dest/services/encoding.d.ts +2 -2
- package/dest/services/encoding.d.ts.map +1 -1
- package/dest/services/encoding.js +2 -2
- package/dest/services/libp2p/libp2p_service.d.ts +11 -3
- package/dest/services/libp2p/libp2p_service.d.ts.map +1 -1
- package/dest/services/libp2p/libp2p_service.js +71 -35
- package/dest/services/reqresp/interface.d.ts +10 -1
- package/dest/services/reqresp/interface.d.ts.map +1 -1
- package/dest/services/reqresp/interface.js +15 -1
- package/dest/services/reqresp/protocols/block_txs/block_txs_handler.d.ts +3 -3
- package/dest/services/reqresp/protocols/block_txs/block_txs_handler.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/block_txs/block_txs_reqresp.d.ts +7 -1
- package/dest/services/reqresp/protocols/block_txs/block_txs_reqresp.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/block_txs/block_txs_reqresp.js +15 -0
- package/dest/services/reqresp/protocols/tx.d.ts +7 -1
- package/dest/services/reqresp/protocols/tx.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/tx.js +20 -0
- package/dest/services/reqresp/reqresp.d.ts +1 -1
- package/dest/services/reqresp/reqresp.d.ts.map +1 -1
- package/dest/services/reqresp/reqresp.js +11 -4
- package/dest/services/service.d.ts +18 -1
- package/dest/services/service.d.ts.map +1 -1
- package/dest/services/tx_collection/config.d.ts +7 -1
- package/dest/services/tx_collection/config.d.ts.map +1 -1
- package/dest/services/tx_collection/config.js +16 -0
- package/dest/services/tx_collection/fast_tx_collection.d.ts +3 -1
- package/dest/services/tx_collection/fast_tx_collection.d.ts.map +1 -1
- package/dest/services/tx_collection/fast_tx_collection.js +17 -3
- package/dest/services/tx_collection/file_store_tx_collection.d.ts +44 -0
- package/dest/services/tx_collection/file_store_tx_collection.d.ts.map +1 -0
- package/dest/services/tx_collection/file_store_tx_collection.js +118 -0
- package/dest/services/tx_collection/file_store_tx_source.d.ts +27 -0
- package/dest/services/tx_collection/file_store_tx_source.d.ts.map +1 -0
- package/dest/services/tx_collection/file_store_tx_source.js +57 -0
- package/dest/services/tx_collection/index.d.ts +2 -1
- package/dest/services/tx_collection/index.d.ts.map +1 -1
- package/dest/services/tx_collection/index.js +1 -0
- package/dest/services/tx_collection/slow_tx_collection.d.ts +3 -1
- package/dest/services/tx_collection/slow_tx_collection.d.ts.map +1 -1
- package/dest/services/tx_collection/slow_tx_collection.js +48 -19
- package/dest/services/tx_collection/tx_collection.d.ts +17 -7
- package/dest/services/tx_collection/tx_collection.d.ts.map +1 -1
- package/dest/services/tx_collection/tx_collection.js +58 -2
- package/dest/services/tx_collection/tx_collection_sink.d.ts +15 -6
- package/dest/services/tx_collection/tx_collection_sink.d.ts.map +1 -1
- package/dest/services/tx_collection/tx_collection_sink.js +13 -7
- package/dest/services/tx_file_store/config.d.ts +1 -3
- package/dest/services/tx_file_store/config.d.ts.map +1 -1
- package/dest/services/tx_file_store/config.js +0 -4
- package/dest/services/tx_file_store/tx_file_store.d.ts +3 -3
- package/dest/services/tx_file_store/tx_file_store.d.ts.map +1 -1
- package/dest/services/tx_provider.d.ts +3 -3
- package/dest/services/tx_provider.d.ts.map +1 -1
- package/dest/services/tx_provider.js +5 -4
- package/dest/test-helpers/make-test-p2p-clients.d.ts +3 -3
- package/dest/test-helpers/make-test-p2p-clients.d.ts.map +1 -1
- package/dest/test-helpers/mock-pubsub.d.ts +27 -1
- package/dest/test-helpers/mock-pubsub.d.ts.map +1 -1
- package/dest/test-helpers/mock-pubsub.js +97 -2
- package/dest/test-helpers/testbench-utils.d.ts +30 -24
- package/dest/test-helpers/testbench-utils.d.ts.map +1 -1
- package/dest/test-helpers/testbench-utils.js +87 -35
- package/dest/testbench/p2p_client_testbench_worker.js +1 -1
- package/package.json +14 -14
- package/src/client/factory.ts +65 -8
- package/src/client/interface.ts +26 -13
- package/src/client/p2p_client.ts +123 -158
- package/src/config.ts +16 -0
- package/src/index.ts +1 -0
- package/src/mem_pools/attestation_pool/attestation_pool.ts +87 -44
- package/src/mem_pools/attestation_pool/attestation_pool_test_suite.ts +57 -53
- package/src/mem_pools/attestation_pool/index.ts +3 -3
- package/src/mem_pools/index.ts +3 -0
- package/src/mem_pools/interface.ts +2 -2
- package/src/mem_pools/tx_pool_v2/README.md +59 -9
- package/src/mem_pools/tx_pool_v2/deleted_pool.ts +234 -0
- package/src/mem_pools/tx_pool_v2/index.ts +1 -0
- package/src/mem_pools/tx_pool_v2/interfaces.ts +2 -2
- package/src/mem_pools/tx_pool_v2/tx_metadata.ts +51 -5
- package/src/mem_pools/tx_pool_v2/tx_pool_v2_impl.ts +90 -77
- package/src/msg_validators/tx_validator/aggregate_tx_validator.ts +2 -2
- package/src/msg_validators/tx_validator/block_header_validator.ts +15 -3
- package/src/msg_validators/tx_validator/double_spend_validator.ts +11 -6
- package/src/msg_validators/tx_validator/timestamp_validator.ts +19 -14
- package/src/services/dummy_service.ts +6 -0
- package/src/services/encoding.ts +2 -2
- package/src/services/libp2p/libp2p_service.ts +70 -37
- package/src/services/reqresp/interface.ts +26 -1
- package/src/services/reqresp/protocols/block_txs/block_txs_handler.ts +2 -2
- package/src/services/reqresp/protocols/block_txs/block_txs_reqresp.ts +17 -0
- package/src/services/reqresp/protocols/tx.ts +22 -0
- package/src/services/reqresp/reqresp.ts +13 -3
- package/src/services/service.ts +20 -0
- package/src/services/tx_collection/config.ts +26 -0
- package/src/services/tx_collection/fast_tx_collection.ts +14 -2
- package/src/services/tx_collection/file_store_tx_collection.ts +152 -0
- package/src/services/tx_collection/file_store_tx_source.ts +70 -0
- package/src/services/tx_collection/index.ts +1 -0
- package/src/services/tx_collection/slow_tx_collection.ts +55 -26
- package/src/services/tx_collection/tx_collection.ts +78 -12
- package/src/services/tx_collection/tx_collection_sink.ts +17 -7
- package/src/services/tx_file_store/config.ts +0 -6
- package/src/services/tx_file_store/tx_file_store.ts +4 -4
- package/src/services/tx_provider.ts +8 -7
- package/src/test-helpers/make-test-p2p-clients.ts +3 -3
- package/src/test-helpers/mock-pubsub.ts +133 -3
- package/src/test-helpers/testbench-utils.ts +100 -40
- package/src/testbench/p2p_client_testbench_worker.ts +1 -1
|
@@ -0,0 +1,152 @@
|
|
|
1
|
+
import { type Logger, createLogger } from '@aztec/foundation/log';
|
|
2
|
+
import { FifoMemoryQueue } from '@aztec/foundation/queue';
|
|
3
|
+
import { Tx, TxHash } from '@aztec/stdlib/tx';
|
|
4
|
+
|
|
5
|
+
import type { FileStoreTxSource } from './file_store_tx_source.js';
|
|
6
|
+
import type { TxAddContext, TxCollectionSink } from './tx_collection_sink.js';
|
|
7
|
+
|
|
8
|
+
// Internal constants (not configurable by node operators)
|
|
9
|
+
const FILE_STORE_DOWNLOAD_CONCURRENCY = 5; // Max concurrent downloads
|
|
10
|
+
|
|
11
|
+
/**
|
|
12
|
+
* Collects txs from file stores as a fallback after P2P methods have been tried.
|
|
13
|
+
* Runs in parallel to slow/fast collection. The delay before starting file store
|
|
14
|
+
* collection is managed by the TxCollection orchestrator, not this class.
|
|
15
|
+
*/
|
|
16
|
+
export class FileStoreTxCollection {
|
|
17
|
+
/** Map from tx hash to add context for txs queued for download. */
|
|
18
|
+
private pendingTxs = new Map<string, TxAddContext>();
|
|
19
|
+
|
|
20
|
+
/**
|
|
21
|
+
* Tracks tx hashes found elsewhere, even before startCollecting is called.
|
|
22
|
+
* Needed because the orchestrator delays startCollecting via a real sleep, but foundTxs
|
|
23
|
+
* may arrive during that delay — before the hashes are added to pendingTxs.
|
|
24
|
+
*/
|
|
25
|
+
private foundTxHashes = new Set<string>();
|
|
26
|
+
|
|
27
|
+
/** Queue of tx hashes to be downloaded. */
|
|
28
|
+
private downloadQueue = new FifoMemoryQueue<TxHash>();
|
|
29
|
+
|
|
30
|
+
/** Worker promises for concurrent downloads. */
|
|
31
|
+
private workers: Promise<void>[] = [];
|
|
32
|
+
|
|
33
|
+
/** Whether the collection has been started. */
|
|
34
|
+
private started = false;
|
|
35
|
+
|
|
36
|
+
constructor(
|
|
37
|
+
private readonly fileStoreSources: FileStoreTxSource[],
|
|
38
|
+
private readonly txCollectionSink: TxCollectionSink,
|
|
39
|
+
private readonly log: Logger = createLogger('p2p:file_store_tx_collection'),
|
|
40
|
+
) {}
|
|
41
|
+
|
|
42
|
+
/** Starts the file store collection workers. */
|
|
43
|
+
public start() {
|
|
44
|
+
if (this.fileStoreSources.length === 0) {
|
|
45
|
+
this.log.debug('No file store sources configured, skipping file store collection');
|
|
46
|
+
return;
|
|
47
|
+
}
|
|
48
|
+
|
|
49
|
+
this.started = true;
|
|
50
|
+
this.downloadQueue = new FifoMemoryQueue<TxHash>();
|
|
51
|
+
|
|
52
|
+
// Start concurrent download workers
|
|
53
|
+
for (let i = 0; i < FILE_STORE_DOWNLOAD_CONCURRENCY; i++) {
|
|
54
|
+
this.workers.push(this.downloadQueue.process(txHash => this.processDownload(txHash)));
|
|
55
|
+
}
|
|
56
|
+
|
|
57
|
+
this.log.info(`Started file store tx collection with ${this.fileStoreSources.length} sources`, {
|
|
58
|
+
sources: this.fileStoreSources.map(s => s.getInfo()),
|
|
59
|
+
concurrency: FILE_STORE_DOWNLOAD_CONCURRENCY,
|
|
60
|
+
});
|
|
61
|
+
}
|
|
62
|
+
|
|
63
|
+
/** Stops all collection activity. */
|
|
64
|
+
public async stop() {
|
|
65
|
+
if (!this.started) {
|
|
66
|
+
return;
|
|
67
|
+
}
|
|
68
|
+
this.started = false;
|
|
69
|
+
this.downloadQueue.end();
|
|
70
|
+
await Promise.all(this.workers);
|
|
71
|
+
this.workers = [];
|
|
72
|
+
this.pendingTxs.clear();
|
|
73
|
+
this.foundTxHashes.clear();
|
|
74
|
+
}
|
|
75
|
+
|
|
76
|
+
/** Remove the given tx hashes from pending. */
|
|
77
|
+
public stopCollecting(txHashes: TxHash[]) {
|
|
78
|
+
for (const txHash of txHashes) {
|
|
79
|
+
const hashStr = txHash.toString();
|
|
80
|
+
this.pendingTxs.delete(hashStr);
|
|
81
|
+
}
|
|
82
|
+
}
|
|
83
|
+
|
|
84
|
+
/** Clears all pending state. Items already in the download queue will still be processed but won't be re-queued. */
|
|
85
|
+
public clearPending() {
|
|
86
|
+
this.pendingTxs.clear();
|
|
87
|
+
this.foundTxHashes.clear();
|
|
88
|
+
}
|
|
89
|
+
|
|
90
|
+
/** Queue the given tx hashes for file store collection. */
|
|
91
|
+
public startCollecting(txHashes: TxHash[], context: TxAddContext) {
|
|
92
|
+
for (const txHash of txHashes) {
|
|
93
|
+
const hashStr = txHash.toString();
|
|
94
|
+
if (!this.pendingTxs.has(hashStr) && !this.foundTxHashes.has(hashStr)) {
|
|
95
|
+
this.pendingTxs.set(hashStr, context);
|
|
96
|
+
this.downloadQueue.put(txHash);
|
|
97
|
+
}
|
|
98
|
+
}
|
|
99
|
+
}
|
|
100
|
+
|
|
101
|
+
/** Stop tracking txs that were found elsewhere. */
|
|
102
|
+
public foundTxs(txs: Tx[]) {
|
|
103
|
+
for (const tx of txs) {
|
|
104
|
+
const hashStr = tx.getTxHash().toString();
|
|
105
|
+
this.pendingTxs.delete(hashStr);
|
|
106
|
+
this.foundTxHashes.add(hashStr);
|
|
107
|
+
}
|
|
108
|
+
}
|
|
109
|
+
|
|
110
|
+
/** Processes a single tx hash from the download queue. */
|
|
111
|
+
private async processDownload(txHash: TxHash) {
|
|
112
|
+
const hashStr = txHash.toString();
|
|
113
|
+
const context = this.pendingTxs.get(hashStr);
|
|
114
|
+
|
|
115
|
+
// Skip if already found by another method
|
|
116
|
+
if (!context) {
|
|
117
|
+
return;
|
|
118
|
+
}
|
|
119
|
+
|
|
120
|
+
await this.downloadTx(txHash, context);
|
|
121
|
+
this.pendingTxs.delete(hashStr);
|
|
122
|
+
}
|
|
123
|
+
|
|
124
|
+
/** Attempt to download a tx from file stores (round-robin). */
|
|
125
|
+
private async downloadTx(txHash: TxHash, context: TxAddContext) {
|
|
126
|
+
const startIndex = Math.floor(Math.random() * this.fileStoreSources.length);
|
|
127
|
+
for (let i = startIndex; i < startIndex + this.fileStoreSources.length; i++) {
|
|
128
|
+
const source = this.fileStoreSources[i % this.fileStoreSources.length];
|
|
129
|
+
|
|
130
|
+
try {
|
|
131
|
+
const result = await this.txCollectionSink.collect(
|
|
132
|
+
hashes => source.getTxsByHash(hashes),
|
|
133
|
+
[txHash],
|
|
134
|
+
{
|
|
135
|
+
description: `file-store ${source.getInfo()}`,
|
|
136
|
+
method: 'file-store',
|
|
137
|
+
fileStore: source.getInfo(),
|
|
138
|
+
},
|
|
139
|
+
context,
|
|
140
|
+
);
|
|
141
|
+
|
|
142
|
+
if (result.txs.length > 0) {
|
|
143
|
+
return;
|
|
144
|
+
}
|
|
145
|
+
} catch (err) {
|
|
146
|
+
this.log.trace(`Failed to download tx ${txHash} from ${source.getInfo()}`, { err });
|
|
147
|
+
}
|
|
148
|
+
}
|
|
149
|
+
|
|
150
|
+
this.log.trace(`Tx ${txHash} not found in any file store`);
|
|
151
|
+
}
|
|
152
|
+
}
|
|
@@ -0,0 +1,70 @@
|
|
|
1
|
+
import { type Logger, createLogger } from '@aztec/foundation/log';
|
|
2
|
+
import { type ReadOnlyFileStore, createReadOnlyFileStore } from '@aztec/stdlib/file-store';
|
|
3
|
+
import { Tx, type TxHash } from '@aztec/stdlib/tx';
|
|
4
|
+
|
|
5
|
+
import type { TxSource } from './tx_source.js';
|
|
6
|
+
|
|
7
|
+
/** TxSource implementation that downloads txs from a file store. */
|
|
8
|
+
export class FileStoreTxSource implements TxSource {
|
|
9
|
+
private constructor(
|
|
10
|
+
private readonly fileStore: ReadOnlyFileStore,
|
|
11
|
+
private readonly baseUrl: string,
|
|
12
|
+
private readonly log: Logger,
|
|
13
|
+
) {}
|
|
14
|
+
|
|
15
|
+
/**
|
|
16
|
+
* Creates a FileStoreTxSource from a URL.
|
|
17
|
+
* @param url - The file store URL (s3://, gs://, file://, http://, https://).
|
|
18
|
+
* @param log - Optional logger.
|
|
19
|
+
* @returns The FileStoreTxSource instance, or undefined if creation fails.
|
|
20
|
+
*/
|
|
21
|
+
public static async create(
|
|
22
|
+
url: string,
|
|
23
|
+
log: Logger = createLogger('p2p:file_store_tx_source'),
|
|
24
|
+
): Promise<FileStoreTxSource | undefined> {
|
|
25
|
+
try {
|
|
26
|
+
const fileStore = await createReadOnlyFileStore(url, log);
|
|
27
|
+
if (!fileStore) {
|
|
28
|
+
log.warn(`Failed to create file store for URL: ${url}`);
|
|
29
|
+
return undefined;
|
|
30
|
+
}
|
|
31
|
+
return new FileStoreTxSource(fileStore, url, log);
|
|
32
|
+
} catch (err) {
|
|
33
|
+
log.warn(`Error creating file store for URL: ${url}`, { error: err });
|
|
34
|
+
return undefined;
|
|
35
|
+
}
|
|
36
|
+
}
|
|
37
|
+
|
|
38
|
+
public getInfo(): string {
|
|
39
|
+
return `file-store:${this.baseUrl}`;
|
|
40
|
+
}
|
|
41
|
+
|
|
42
|
+
public getTxsByHash(txHashes: TxHash[]): Promise<(Tx | undefined)[]> {
|
|
43
|
+
return Promise.all(
|
|
44
|
+
txHashes.map(async txHash => {
|
|
45
|
+
const path = `txs/${txHash.toString()}.bin`;
|
|
46
|
+
try {
|
|
47
|
+
const buffer = await this.fileStore.read(path);
|
|
48
|
+
return Tx.fromBuffer(buffer);
|
|
49
|
+
} catch {
|
|
50
|
+
// Tx not found or error reading - return undefined
|
|
51
|
+
return undefined;
|
|
52
|
+
}
|
|
53
|
+
}),
|
|
54
|
+
);
|
|
55
|
+
}
|
|
56
|
+
}
|
|
57
|
+
|
|
58
|
+
/**
|
|
59
|
+
* Creates FileStoreTxSource instances from URLs.
|
|
60
|
+
* @param urls - Array of file store URLs.
|
|
61
|
+
* @param log - Optional logger.
|
|
62
|
+
* @returns Array of successfully created FileStoreTxSource instances.
|
|
63
|
+
*/
|
|
64
|
+
export async function createFileStoreTxSources(
|
|
65
|
+
urls: string[],
|
|
66
|
+
log: Logger = createLogger('p2p:file_store_tx_source'),
|
|
67
|
+
): Promise<FileStoreTxSource[]> {
|
|
68
|
+
const sources = await Promise.all(urls.map(url => FileStoreTxSource.create(url, log)));
|
|
69
|
+
return sources.filter((s): s is FileStoreTxSource => s !== undefined);
|
|
70
|
+
}
|
|
@@ -91,6 +91,7 @@ export class SlowTxCollection {
|
|
|
91
91
|
|
|
92
92
|
for (const txHash of txHashes) {
|
|
93
93
|
this.missingTxs.set(txHash.toString(), {
|
|
94
|
+
block,
|
|
94
95
|
blockNumber: block.number,
|
|
95
96
|
deadline: this.getDeadlineForSlot(block.header.getSlot()),
|
|
96
97
|
readyForReqResp: this.nodes.length === 0, // If we have no nodes, we can start reqresp immediately
|
|
@@ -109,18 +110,26 @@ export class SlowTxCollection {
|
|
|
109
110
|
|
|
110
111
|
// Gather all missing txs that are not in fast collection and request them from the node
|
|
111
112
|
const missingTxs = this.getMissingTxsForSlowCollection();
|
|
112
|
-
|
|
113
|
-
if (missingTxHashes.length === 0) {
|
|
113
|
+
if (missingTxs.length === 0) {
|
|
114
114
|
return;
|
|
115
115
|
}
|
|
116
116
|
|
|
117
|
-
//
|
|
118
|
-
for (const
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
117
|
+
// Group by block so we pass the correct mined context to the sink
|
|
118
|
+
for (const entries of this.groupByBlock(missingTxs)) {
|
|
119
|
+
const block = entries[0][1].block;
|
|
120
|
+
const txHashes = entries.map(([txHash]) => TxHash.fromString(txHash));
|
|
121
|
+
for (const batch of chunk(txHashes, this.config.txCollectionNodeRpcMaxBatchSize)) {
|
|
122
|
+
await this.txCollectionSink.collect(
|
|
123
|
+
hashes => node.getTxsByHash(hashes),
|
|
124
|
+
batch,
|
|
125
|
+
{
|
|
126
|
+
description: `node ${node.getInfo()}`,
|
|
127
|
+
node: node.getInfo(),
|
|
128
|
+
method: 'slow-node-rpc',
|
|
129
|
+
},
|
|
130
|
+
{ type: 'mined', block },
|
|
131
|
+
);
|
|
132
|
+
}
|
|
124
133
|
}
|
|
125
134
|
|
|
126
135
|
// Mark every tx that is still missing as ready for reqresp.
|
|
@@ -149,25 +158,30 @@ export class SlowTxCollection {
|
|
|
149
158
|
|
|
150
159
|
const pinnedPeer = undefined;
|
|
151
160
|
const timeoutMs = this.config.txCollectionSlowReqRespTimeoutMs;
|
|
152
|
-
const maxPeers = boundInclusive(Math.ceil(missingTxs.length / 3), 4, 16);
|
|
153
161
|
const maxRetryAttempts = 3;
|
|
154
|
-
// Send a batch request via reqresp for the missing txs
|
|
155
|
-
await this.txCollectionSink.collect(
|
|
156
|
-
async txHashes => {
|
|
157
|
-
const txs = await this.reqResp.sendBatchRequest<ReqRespSubProtocol.TX>(
|
|
158
|
-
ReqRespSubProtocol.TX,
|
|
159
|
-
chunkTxHashesRequest(txHashes),
|
|
160
|
-
pinnedPeer,
|
|
161
|
-
timeoutMs,
|
|
162
|
-
maxPeers,
|
|
163
|
-
maxRetryAttempts,
|
|
164
|
-
);
|
|
165
162
|
|
|
166
|
-
|
|
167
|
-
|
|
168
|
-
|
|
169
|
-
|
|
170
|
-
|
|
163
|
+
// Group by block so we pass the correct mined context to the sink
|
|
164
|
+
for (const entries of this.groupByBlock(missingTxs)) {
|
|
165
|
+
const block = entries[0][1].block;
|
|
166
|
+
const txHashes = entries.map(([txHash]) => TxHash.fromString(txHash));
|
|
167
|
+
const maxPeers = boundInclusive(Math.ceil(txHashes.length / 3), 4, 16);
|
|
168
|
+
await this.txCollectionSink.collect(
|
|
169
|
+
async hashes => {
|
|
170
|
+
const txs = await this.reqResp.sendBatchRequest<ReqRespSubProtocol.TX>(
|
|
171
|
+
ReqRespSubProtocol.TX,
|
|
172
|
+
chunkTxHashesRequest(hashes),
|
|
173
|
+
pinnedPeer,
|
|
174
|
+
timeoutMs,
|
|
175
|
+
maxPeers,
|
|
176
|
+
maxRetryAttempts,
|
|
177
|
+
);
|
|
178
|
+
return txs.flat();
|
|
179
|
+
},
|
|
180
|
+
txHashes,
|
|
181
|
+
{ description: 'slow reqresp', timeoutMs, method: 'slow-req-resp' },
|
|
182
|
+
{ type: 'mined', block },
|
|
183
|
+
);
|
|
184
|
+
}
|
|
171
185
|
}
|
|
172
186
|
|
|
173
187
|
/** Retrieves all missing txs for the slow collection process. This is, all missing txs that are not part of a fast request. */
|
|
@@ -223,6 +237,21 @@ export class SlowTxCollection {
|
|
|
223
237
|
}
|
|
224
238
|
}
|
|
225
239
|
|
|
240
|
+
/** Groups missing tx entries by block number. */
|
|
241
|
+
private groupByBlock(entries: [string, MissingTxInfo][]): [string, MissingTxInfo][][] {
|
|
242
|
+
const groups = new Map<number, [string, MissingTxInfo][]>();
|
|
243
|
+
for (const entry of entries) {
|
|
244
|
+
const bn = +entry[1].blockNumber;
|
|
245
|
+
let group = groups.get(bn);
|
|
246
|
+
if (!group) {
|
|
247
|
+
group = [];
|
|
248
|
+
groups.set(bn, group);
|
|
249
|
+
}
|
|
250
|
+
group.push(entry);
|
|
251
|
+
}
|
|
252
|
+
return [...groups.values()];
|
|
253
|
+
}
|
|
254
|
+
|
|
226
255
|
/** Computes the proof submission deadline for a given slot, a tx mined in this slot is no longer interesting after this deadline */
|
|
227
256
|
private getDeadlineForSlot(slotNumber: SlotNumber): Date {
|
|
228
257
|
const epoch = getEpochAtSlot(slotNumber, this.constants);
|
|
@@ -2,6 +2,7 @@ import { BlockNumber } from '@aztec/foundation/branded-types';
|
|
|
2
2
|
import { compactArray } from '@aztec/foundation/collection';
|
|
3
3
|
import { type Logger, createLogger } from '@aztec/foundation/log';
|
|
4
4
|
import { type PromiseWithResolvers, RunningPromise } from '@aztec/foundation/promise';
|
|
5
|
+
import { sleep } from '@aztec/foundation/sleep';
|
|
5
6
|
import { DateProvider } from '@aztec/foundation/timer';
|
|
6
7
|
import type { L2Block, L2BlockInfo } from '@aztec/stdlib/block';
|
|
7
8
|
import type { L1RollupConstants } from '@aztec/stdlib/epoch-helpers';
|
|
@@ -11,18 +12,19 @@ import { type TelemetryClient, getTelemetryClient } from '@aztec/telemetry-clien
|
|
|
11
12
|
|
|
12
13
|
import type { PeerId } from '@libp2p/interface';
|
|
13
14
|
|
|
14
|
-
import type {
|
|
15
|
-
import type { TxPoolEvents } from '../../mem_pools/tx_pool/tx_pool.js';
|
|
15
|
+
import type { TxPoolV2, TxPoolV2Events } from '../../mem_pools/tx_pool_v2/interfaces.js';
|
|
16
16
|
import type { BatchTxRequesterLibP2PService } from '../reqresp/batch-tx-requester/interface.js';
|
|
17
17
|
import type { TxCollectionConfig } from './config.js';
|
|
18
18
|
import { FastTxCollection } from './fast_tx_collection.js';
|
|
19
|
+
import { FileStoreTxCollection } from './file_store_tx_collection.js';
|
|
20
|
+
import type { FileStoreTxSource } from './file_store_tx_source.js';
|
|
19
21
|
import { SlowTxCollection } from './slow_tx_collection.js';
|
|
20
|
-
import { TxCollectionSink } from './tx_collection_sink.js';
|
|
22
|
+
import { type TxAddContext, TxCollectionSink } from './tx_collection_sink.js';
|
|
21
23
|
import type { TxSource } from './tx_source.js';
|
|
22
24
|
|
|
23
|
-
export type CollectionMethod = 'fast-req-resp' | 'fast-node-rpc' | 'slow-req-resp' | 'slow-node-rpc';
|
|
25
|
+
export type CollectionMethod = 'fast-req-resp' | 'fast-node-rpc' | 'slow-req-resp' | 'slow-node-rpc' | 'file-store';
|
|
24
26
|
|
|
25
|
-
export type MissingTxInfo = { blockNumber: BlockNumber; deadline: Date; readyForReqResp: boolean };
|
|
27
|
+
export type MissingTxInfo = { block: L2Block; blockNumber: BlockNumber; deadline: Date; readyForReqResp: boolean };
|
|
26
28
|
|
|
27
29
|
export type FastCollectionRequestInput =
|
|
28
30
|
| { type: 'block'; block: L2Block }
|
|
@@ -54,6 +56,9 @@ export class TxCollection {
|
|
|
54
56
|
/** Fast collection methods */
|
|
55
57
|
protected readonly fastCollection: FastTxCollection;
|
|
56
58
|
|
|
59
|
+
/** File store collection */
|
|
60
|
+
protected readonly fileStoreCollection: FileStoreTxCollection;
|
|
61
|
+
|
|
57
62
|
/** Loop for periodically reconciling found transactions from the tx pool in case we missed some */
|
|
58
63
|
private readonly reconcileFoundTxsLoop: RunningPromise;
|
|
59
64
|
|
|
@@ -61,17 +66,24 @@ export class TxCollection {
|
|
|
61
66
|
private readonly txCollectionSink: TxCollectionSink;
|
|
62
67
|
|
|
63
68
|
/** Handler for the txs-added event from the tx pool */
|
|
64
|
-
protected readonly handleTxsAddedToPool:
|
|
69
|
+
protected readonly handleTxsAddedToPool: TxPoolV2Events['txs-added'];
|
|
65
70
|
|
|
66
71
|
/** Handler for the txs-added event from the tx collection sink */
|
|
67
|
-
protected readonly handleTxsFound:
|
|
72
|
+
protected readonly handleTxsFound: TxPoolV2Events['txs-added'];
|
|
73
|
+
|
|
74
|
+
/** Whether the service has been started. */
|
|
75
|
+
private started = false;
|
|
76
|
+
|
|
77
|
+
/** Whether file store sources are configured. */
|
|
78
|
+
private readonly hasFileStoreSources: boolean;
|
|
68
79
|
|
|
69
80
|
constructor(
|
|
70
81
|
private readonly p2pService: BatchTxRequesterLibP2PService,
|
|
71
82
|
private readonly nodes: TxSource[],
|
|
72
83
|
private readonly constants: L1RollupConstants,
|
|
73
|
-
private readonly txPool:
|
|
84
|
+
private readonly txPool: TxPoolV2,
|
|
74
85
|
private readonly config: TxCollectionConfig,
|
|
86
|
+
fileStoreSources: FileStoreTxSource[] = [],
|
|
75
87
|
private readonly dateProvider: DateProvider = new DateProvider(),
|
|
76
88
|
telemetryClient: TelemetryClient = getTelemetryClient(),
|
|
77
89
|
private readonly log: Logger = createLogger('p2p:tx_collection_service'),
|
|
@@ -98,18 +110,21 @@ export class TxCollection {
|
|
|
98
110
|
this.log,
|
|
99
111
|
);
|
|
100
112
|
|
|
113
|
+
this.hasFileStoreSources = fileStoreSources.length > 0;
|
|
114
|
+
this.fileStoreCollection = new FileStoreTxCollection(fileStoreSources, this.txCollectionSink, this.log);
|
|
115
|
+
|
|
101
116
|
this.reconcileFoundTxsLoop = new RunningPromise(
|
|
102
117
|
() => this.reconcileFoundTxsWithPool(),
|
|
103
118
|
this.log,
|
|
104
119
|
this.config.txCollectionReconcileIntervalMs,
|
|
105
120
|
);
|
|
106
121
|
|
|
107
|
-
this.handleTxsFound = (args: Parameters<
|
|
122
|
+
this.handleTxsFound = (args: Parameters<TxPoolV2Events['txs-added']>[0]) => {
|
|
108
123
|
this.foundTxs(args.txs);
|
|
109
124
|
};
|
|
110
125
|
this.txCollectionSink.on('txs-added', this.handleTxsFound);
|
|
111
126
|
|
|
112
|
-
this.handleTxsAddedToPool = (args: Parameters<
|
|
127
|
+
this.handleTxsAddedToPool = (args: Parameters<TxPoolV2Events['txs-added']>[0]) => {
|
|
113
128
|
const { txs, source } = args;
|
|
114
129
|
if (source !== 'tx-collection') {
|
|
115
130
|
this.foundTxs(txs);
|
|
@@ -120,7 +135,9 @@ export class TxCollection {
|
|
|
120
135
|
|
|
121
136
|
/** Starts all collection loops. */
|
|
122
137
|
public start(): Promise<void> {
|
|
138
|
+
this.started = true;
|
|
123
139
|
this.slowCollection.start();
|
|
140
|
+
this.fileStoreCollection.start();
|
|
124
141
|
this.reconcileFoundTxsLoop.start();
|
|
125
142
|
|
|
126
143
|
// TODO(palla/txs): Collect mined unproven tx hashes for txs we dont have in the pool and populate missingTxs on startup
|
|
@@ -129,7 +146,13 @@ export class TxCollection {
|
|
|
129
146
|
|
|
130
147
|
/** Stops all activity. */
|
|
131
148
|
public async stop() {
|
|
132
|
-
|
|
149
|
+
this.started = false;
|
|
150
|
+
await Promise.all([
|
|
151
|
+
this.slowCollection.stop(),
|
|
152
|
+
this.fastCollection.stop(),
|
|
153
|
+
this.fileStoreCollection.stop(),
|
|
154
|
+
this.reconcileFoundTxsLoop.stop(),
|
|
155
|
+
]);
|
|
133
156
|
|
|
134
157
|
this.txPool.removeListener('txs-added', this.handleTxsAddedToPool);
|
|
135
158
|
this.txCollectionSink.removeListener('txs-added', this.handleTxsFound);
|
|
@@ -147,7 +170,24 @@ export class TxCollection {
|
|
|
147
170
|
|
|
148
171
|
/** Starts collecting the given tx hashes for the given L2Block in the slow loop */
|
|
149
172
|
public startCollecting(block: L2Block, txHashes: TxHash[]) {
|
|
150
|
-
|
|
173
|
+
this.slowCollection.startCollecting(block, txHashes);
|
|
174
|
+
|
|
175
|
+
// Delay file store collection to give P2P methods time to find txs first
|
|
176
|
+
if (this.hasFileStoreSources) {
|
|
177
|
+
const context: TxAddContext = { type: 'mined', block };
|
|
178
|
+
sleep(this.config.txCollectionFileStoreSlowDelayMs)
|
|
179
|
+
.then(() => {
|
|
180
|
+
if (this.started) {
|
|
181
|
+
// Only queue txs that are still missing after the delay
|
|
182
|
+
const stillMissing = new Set(this.slowCollection.getMissingTxHashes().map(h => h.toString()));
|
|
183
|
+
const remaining = txHashes.filter(h => stillMissing.has(h.toString()));
|
|
184
|
+
if (remaining.length > 0) {
|
|
185
|
+
this.fileStoreCollection.startCollecting(remaining, context);
|
|
186
|
+
}
|
|
187
|
+
}
|
|
188
|
+
})
|
|
189
|
+
.catch(err => this.log.error('Error in file store slow delay', err));
|
|
190
|
+
}
|
|
151
191
|
}
|
|
152
192
|
|
|
153
193
|
/** Collects the set of txs for the given block proposal as fast as possible */
|
|
@@ -175,13 +215,37 @@ export class TxCollection {
|
|
|
175
215
|
txHashes: TxHash[] | string[],
|
|
176
216
|
opts: { deadline: Date; pinnedPeer?: PeerId },
|
|
177
217
|
) {
|
|
218
|
+
const hashes = txHashes.map(h => (typeof h === 'string' ? TxHash.fromString(h) : h));
|
|
219
|
+
|
|
220
|
+
// Delay file store collection to give P2P methods time to find txs first
|
|
221
|
+
if (this.hasFileStoreSources) {
|
|
222
|
+
const context = this.getAddContextForInput(input);
|
|
223
|
+
sleep(this.config.txCollectionFileStoreFastDelayMs)
|
|
224
|
+
.then(() => {
|
|
225
|
+
if (this.started) {
|
|
226
|
+
this.fileStoreCollection.startCollecting(hashes, context);
|
|
227
|
+
}
|
|
228
|
+
})
|
|
229
|
+
.catch(err => this.log.error('Error in file store fast delay', err));
|
|
230
|
+
}
|
|
231
|
+
|
|
178
232
|
return this.fastCollection.collectFastFor(input, txHashes, opts);
|
|
179
233
|
}
|
|
180
234
|
|
|
235
|
+
/** Returns the TxAddContext for the given fast collection request input */
|
|
236
|
+
private getAddContextForInput(input: FastCollectionRequestInput): TxAddContext {
|
|
237
|
+
if (input.type === 'proposal') {
|
|
238
|
+
return { type: 'proposal', blockHeader: input.blockProposal.blockHeader };
|
|
239
|
+
} else {
|
|
240
|
+
return { type: 'mined', block: input.block };
|
|
241
|
+
}
|
|
242
|
+
}
|
|
243
|
+
|
|
181
244
|
/** Mark the given txs as found. Stops collecting them. */
|
|
182
245
|
private foundTxs(txs: Tx[]) {
|
|
183
246
|
this.slowCollection.foundTxs(txs);
|
|
184
247
|
this.fastCollection.foundTxs(txs);
|
|
248
|
+
this.fileStoreCollection.foundTxs(txs);
|
|
185
249
|
}
|
|
186
250
|
|
|
187
251
|
/**
|
|
@@ -191,6 +255,7 @@ export class TxCollection {
|
|
|
191
255
|
public stopCollectingForBlocksUpTo(blockNumber: BlockNumber): void {
|
|
192
256
|
this.slowCollection.stopCollectingForBlocksUpTo(blockNumber);
|
|
193
257
|
this.fastCollection.stopCollectingForBlocksUpTo(blockNumber);
|
|
258
|
+
this.fileStoreCollection.clearPending();
|
|
194
259
|
}
|
|
195
260
|
|
|
196
261
|
/**
|
|
@@ -200,6 +265,7 @@ export class TxCollection {
|
|
|
200
265
|
public stopCollectingForBlocksAfter(blockNumber: BlockNumber): void {
|
|
201
266
|
this.slowCollection.stopCollectingForBlocksAfter(blockNumber);
|
|
202
267
|
this.fastCollection.stopCollectingForBlocksAfter(blockNumber);
|
|
268
|
+
this.fileStoreCollection.clearPending();
|
|
203
269
|
}
|
|
204
270
|
|
|
205
271
|
/** Every now and then, check if the pool has received one of the txs we are looking for, just to catch any race conditions */
|
|
@@ -1,24 +1,28 @@
|
|
|
1
1
|
import type { Logger } from '@aztec/foundation/log';
|
|
2
2
|
import { elapsed } from '@aztec/foundation/timer';
|
|
3
3
|
import type { TypedEventEmitter } from '@aztec/foundation/types';
|
|
4
|
-
import {
|
|
4
|
+
import type { L2Block } from '@aztec/stdlib/block';
|
|
5
|
+
import type { BlockHeader, Tx, TxHash } from '@aztec/stdlib/tx';
|
|
5
6
|
import type { TelemetryClient } from '@aztec/telemetry-client';
|
|
6
7
|
|
|
7
8
|
import EventEmitter from 'node:events';
|
|
8
9
|
|
|
9
|
-
import type {
|
|
10
|
+
import type { TxPoolV2, TxPoolV2Events } from '../../mem_pools/tx_pool_v2/interfaces.js';
|
|
10
11
|
import { TxCollectionInstrumentation } from './instrumentation.js';
|
|
11
12
|
import type { CollectionMethod } from './tx_collection.js';
|
|
12
13
|
|
|
14
|
+
/** Context determining how collected txs should be added to the pool. */
|
|
15
|
+
export type TxAddContext = { type: 'proposal'; blockHeader: BlockHeader } | { type: 'mined'; block: L2Block };
|
|
16
|
+
|
|
13
17
|
/**
|
|
14
18
|
* Executes collection requests from the fast and slow collection loops, and handles collected txs
|
|
15
19
|
* by adding them to the tx pool and emitting events, as well as handling logging and metrics.
|
|
16
20
|
*/
|
|
17
|
-
export class TxCollectionSink extends (EventEmitter as new () => TypedEventEmitter<
|
|
21
|
+
export class TxCollectionSink extends (EventEmitter as new () => TypedEventEmitter<TxPoolV2Events>) {
|
|
18
22
|
private readonly instrumentation: TxCollectionInstrumentation;
|
|
19
23
|
|
|
20
24
|
constructor(
|
|
21
|
-
private readonly txPool:
|
|
25
|
+
private readonly txPool: TxPoolV2,
|
|
22
26
|
telemetryClient: TelemetryClient,
|
|
23
27
|
private readonly log: Logger,
|
|
24
28
|
) {
|
|
@@ -30,6 +34,7 @@ export class TxCollectionSink extends (EventEmitter as new () => TypedEventEmitt
|
|
|
30
34
|
collectValidTxsFn: (txHashes: TxHash[]) => Promise<(Tx | undefined)[]>,
|
|
31
35
|
requested: TxHash[],
|
|
32
36
|
info: Record<string, any> & { description: string; method: CollectionMethod },
|
|
37
|
+
context: TxAddContext,
|
|
33
38
|
) {
|
|
34
39
|
this.log.trace(`Requesting ${requested.length} txs via ${info.description}`, {
|
|
35
40
|
...info,
|
|
@@ -99,12 +104,13 @@ export class TxCollectionSink extends (EventEmitter as new () => TypedEventEmitt
|
|
|
99
104
|
},
|
|
100
105
|
);
|
|
101
106
|
|
|
102
|
-
return await this.foundTxs(validTxs, { ...info, duration });
|
|
107
|
+
return await this.foundTxs(validTxs, { ...info, duration }, context);
|
|
103
108
|
}
|
|
104
109
|
|
|
105
110
|
private async foundTxs(
|
|
106
111
|
txs: Tx[],
|
|
107
112
|
info: Record<string, any> & { description: string; method: CollectionMethod; duration: number },
|
|
113
|
+
context: TxAddContext,
|
|
108
114
|
) {
|
|
109
115
|
// Report metrics for the collection
|
|
110
116
|
this.instrumentation.increaseTxsFor(info.method, txs.length, info.duration);
|
|
@@ -112,9 +118,13 @@ export class TxCollectionSink extends (EventEmitter as new () => TypedEventEmitt
|
|
|
112
118
|
// Mark txs as found in the slow missing txs set and all fast requests
|
|
113
119
|
this.emit('txs-added', { txs });
|
|
114
120
|
|
|
115
|
-
// Add the txs to the tx pool
|
|
121
|
+
// Add the txs to the tx pool using the appropriate method based on context
|
|
116
122
|
try {
|
|
117
|
-
|
|
123
|
+
if (context.type === 'mined') {
|
|
124
|
+
await this.txPool.addMinedTxs(txs, context.block.header, { source: 'tx-collection' });
|
|
125
|
+
} else {
|
|
126
|
+
await this.txPool.addProtectedTxs(txs, context.blockHeader, { source: 'tx-collection' });
|
|
127
|
+
}
|
|
118
128
|
} catch (err) {
|
|
119
129
|
this.log.error(`Error adding txs to the pool via ${info.description}`, err, {
|
|
120
130
|
...info,
|
|
@@ -6,8 +6,6 @@ import { type ConfigMappingsType, booleanConfigHelper, numberConfigHelper } from
|
|
|
6
6
|
export type TxFileStoreConfig = {
|
|
7
7
|
/** URL for uploading txs to file storage (s3://, gs://, file://) */
|
|
8
8
|
txFileStoreUrl?: string;
|
|
9
|
-
/** URL for downloading txs from file storage */
|
|
10
|
-
txFileStoreDownloadUrl?: string;
|
|
11
9
|
/** Max concurrent uploads */
|
|
12
10
|
txFileStoreUploadConcurrency: number;
|
|
13
11
|
/** Max queue size to prevent unbounded memory growth */
|
|
@@ -21,10 +19,6 @@ export const txFileStoreConfigMappings: ConfigMappingsType<TxFileStoreConfig> =
|
|
|
21
19
|
env: 'TX_FILE_STORE_URL',
|
|
22
20
|
description: 'URL for uploading txs to file storage (s3://, gs://, file://)',
|
|
23
21
|
},
|
|
24
|
-
txFileStoreDownloadUrl: {
|
|
25
|
-
env: 'TX_FILE_STORE_DOWNLOAD_URL',
|
|
26
|
-
description: 'URL for downloading txs from file storage',
|
|
27
|
-
},
|
|
28
22
|
txFileStoreUploadConcurrency: {
|
|
29
23
|
env: 'TX_FILE_STORE_UPLOAD_CONCURRENCY',
|
|
30
24
|
description: 'Maximum number of concurrent tx uploads',
|
|
@@ -6,7 +6,7 @@ import { type FileStore, createFileStore } from '@aztec/stdlib/file-store';
|
|
|
6
6
|
import type { Tx } from '@aztec/stdlib/tx';
|
|
7
7
|
import { type TelemetryClient, getTelemetryClient } from '@aztec/telemetry-client';
|
|
8
8
|
|
|
9
|
-
import type {
|
|
9
|
+
import type { TxPoolV2 } from '../../mem_pools/index.js';
|
|
10
10
|
import type { TxFileStoreConfig } from './config.js';
|
|
11
11
|
import { TxFileStoreInstrumentation } from './instrumentation.js';
|
|
12
12
|
|
|
@@ -18,7 +18,7 @@ export class TxFileStore {
|
|
|
18
18
|
private uploadQueue: Tx[] = [];
|
|
19
19
|
private activeUploads = 0;
|
|
20
20
|
private readonly queueProcessor: RunningPromise;
|
|
21
|
-
private readonly handleTxsAdded:
|
|
21
|
+
private readonly handleTxsAdded: (args: { txs: Tx[]; source?: string }) => void;
|
|
22
22
|
|
|
23
23
|
/** Recently uploaded tx hashes for deduplication. */
|
|
24
24
|
private recentUploads: Set<string> = new Set();
|
|
@@ -27,7 +27,7 @@ export class TxFileStore {
|
|
|
27
27
|
|
|
28
28
|
private constructor(
|
|
29
29
|
private readonly fileStore: FileStore,
|
|
30
|
-
private readonly txPool:
|
|
30
|
+
private readonly txPool: TxPoolV2,
|
|
31
31
|
private readonly config: TxFileStoreConfig,
|
|
32
32
|
private readonly instrumentation: TxFileStoreInstrumentation,
|
|
33
33
|
private readonly log: Logger,
|
|
@@ -48,7 +48,7 @@ export class TxFileStore {
|
|
|
48
48
|
* @returns The file store instance, or undefined if not configured/enabled.
|
|
49
49
|
*/
|
|
50
50
|
static async create(
|
|
51
|
-
txPool:
|
|
51
|
+
txPool: TxPoolV2,
|
|
52
52
|
config: TxFileStoreConfig,
|
|
53
53
|
log: Logger = createLogger('p2p:tx_file_store'),
|
|
54
54
|
telemetry: TelemetryClient = getTelemetryClient(),
|