@aztec/prover-node 0.0.0-test.1 → 0.0.1-commit.b655e406
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/actions/download-epoch-proving-job.d.ts +18 -0
- package/dest/actions/download-epoch-proving-job.d.ts.map +1 -0
- package/dest/actions/download-epoch-proving-job.js +37 -0
- package/dest/actions/index.d.ts +3 -0
- package/dest/actions/index.d.ts.map +1 -0
- package/dest/actions/index.js +2 -0
- package/dest/actions/rerun-epoch-proving-job.d.ts +11 -0
- package/dest/actions/rerun-epoch-proving-job.d.ts.map +1 -0
- package/dest/actions/rerun-epoch-proving-job.js +40 -0
- package/dest/actions/upload-epoch-proof-failure.d.ts +15 -0
- package/dest/actions/upload-epoch-proof-failure.d.ts.map +1 -0
- package/dest/actions/upload-epoch-proof-failure.js +78 -0
- package/dest/bin/run-failed-epoch.d.ts +2 -0
- package/dest/bin/run-failed-epoch.d.ts.map +1 -0
- package/dest/bin/run-failed-epoch.js +67 -0
- package/dest/config.d.ts +12 -9
- package/dest/config.d.ts.map +1 -1
- package/dest/config.js +81 -14
- package/dest/factory.d.ts +12 -8
- package/dest/factory.d.ts.map +1 -1
- package/dest/factory.js +95 -31
- package/dest/index.d.ts +1 -1
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +1 -1
- package/dest/job/epoch-proving-job-data.d.ts +16 -0
- package/dest/job/epoch-proving-job-data.d.ts.map +1 -0
- package/dest/job/epoch-proving-job-data.js +52 -0
- package/dest/job/epoch-proving-job.d.ts +30 -15
- package/dest/job/epoch-proving-job.d.ts.map +1 -1
- package/dest/job/epoch-proving-job.js +149 -50
- package/dest/metrics.d.ts +28 -4
- package/dest/metrics.d.ts.map +1 -1
- package/dest/metrics.js +141 -35
- package/dest/monitors/epoch-monitor.d.ts +3 -1
- package/dest/monitors/epoch-monitor.d.ts.map +1 -1
- package/dest/monitors/epoch-monitor.js +15 -2
- package/dest/prover-node-publisher.d.ts +7 -10
- package/dest/prover-node-publisher.d.ts.map +1 -1
- package/dest/prover-node-publisher.js +59 -60
- package/dest/prover-node.d.ts +43 -39
- package/dest/prover-node.d.ts.map +1 -1
- package/dest/prover-node.js +171 -100
- package/dest/prover-publisher-factory.d.ts +21 -0
- package/dest/prover-publisher-factory.d.ts.map +1 -0
- package/dest/prover-publisher-factory.js +26 -0
- package/dest/test/index.d.ts +4 -2
- package/dest/test/index.d.ts.map +1 -1
- package/dest/test/index.js +1 -3
- package/package.json +36 -31
- package/src/actions/download-epoch-proving-job.ts +44 -0
- package/src/actions/index.ts +2 -0
- package/src/actions/rerun-epoch-proving-job.ts +61 -0
- package/src/actions/upload-epoch-proof-failure.ts +88 -0
- package/src/bin/run-failed-epoch.ts +77 -0
- package/src/config.ts +108 -24
- package/src/factory.ts +161 -43
- package/src/index.ts +1 -1
- package/src/job/epoch-proving-job-data.ts +76 -0
- package/src/job/epoch-proving-job.ts +215 -50
- package/src/metrics.ts +135 -37
- package/src/monitors/epoch-monitor.ts +16 -5
- package/src/prover-node-publisher.ts +93 -86
- package/src/prover-node.ts +203 -126
- package/src/prover-publisher-factory.ts +37 -0
- package/src/test/index.ts +7 -4
- package/dest/http.d.ts +0 -8
- package/dest/http.d.ts.map +0 -1
- package/dest/http.js +0 -9
- package/dest/prover-coordination/config.d.ts +0 -7
- package/dest/prover-coordination/config.d.ts.map +0 -1
- package/dest/prover-coordination/config.js +0 -11
- package/dest/prover-coordination/factory.d.ts +0 -22
- package/dest/prover-coordination/factory.d.ts.map +0 -1
- package/dest/prover-coordination/factory.js +0 -42
- package/dest/prover-coordination/index.d.ts +0 -3
- package/dest/prover-coordination/index.d.ts.map +0 -1
- package/dest/prover-coordination/index.js +0 -2
- package/src/http.ts +0 -13
- package/src/prover-coordination/config.ts +0 -17
- package/src/prover-coordination/factory.ts +0 -72
- package/src/prover-coordination/index.ts +0 -2
package/dest/factory.js
CHANGED
|
@@ -1,23 +1,78 @@
|
|
|
1
1
|
import { createArchiver } from '@aztec/archiver';
|
|
2
|
+
import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
|
|
2
3
|
import { createBlobSinkClient } from '@aztec/blob-sink/client';
|
|
3
4
|
import { EpochCache } from '@aztec/epoch-cache';
|
|
4
|
-
import {
|
|
5
|
+
import { PublisherManager, RollupContract, createEthereumChain } from '@aztec/ethereum';
|
|
6
|
+
import { pick } from '@aztec/foundation/collection';
|
|
5
7
|
import { createLogger } from '@aztec/foundation/log';
|
|
8
|
+
import { DateProvider } from '@aztec/foundation/timer';
|
|
9
|
+
import { KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
|
|
10
|
+
import { trySnapshotSync } from '@aztec/node-lib/actions';
|
|
11
|
+
import { createL1TxUtilsFromEthSignerWithStore } from '@aztec/node-lib/factories';
|
|
12
|
+
import { NodeRpcTxSource, createP2PClient } from '@aztec/p2p';
|
|
6
13
|
import { createProverClient } from '@aztec/prover-client';
|
|
7
14
|
import { createAndStartProvingBroker } from '@aztec/prover-client/broker';
|
|
8
|
-
import {
|
|
15
|
+
import { P2PClientType } from '@aztec/stdlib/p2p';
|
|
16
|
+
import { getPackageVersion } from '@aztec/stdlib/update-checker';
|
|
17
|
+
import { L1Metrics, getTelemetryClient } from '@aztec/telemetry-client';
|
|
9
18
|
import { createWorldStateSynchronizer } from '@aztec/world-state';
|
|
19
|
+
import { createPublicClient, fallback, http } from 'viem';
|
|
20
|
+
import { createKeyStoreForProver } from './config.js';
|
|
10
21
|
import { EpochMonitor } from './monitors/epoch-monitor.js';
|
|
11
|
-
import { createProverCoordination } from './prover-coordination/factory.js';
|
|
12
|
-
import { ProverNodePublisher } from './prover-node-publisher.js';
|
|
13
22
|
import { ProverNode } from './prover-node.js';
|
|
14
|
-
|
|
23
|
+
import { ProverPublisherFactory } from './prover-publisher-factory.js';
|
|
24
|
+
/** Creates a new prover node given a config. */ export async function createProverNode(userConfig, deps = {}, options = {}) {
|
|
25
|
+
const config = {
|
|
26
|
+
...userConfig
|
|
27
|
+
};
|
|
15
28
|
const telemetry = deps.telemetry ?? getTelemetryClient();
|
|
16
|
-
const
|
|
29
|
+
const dateProvider = deps.dateProvider ?? new DateProvider();
|
|
30
|
+
const blobSinkClient = deps.blobSinkClient ?? createBlobSinkClient(config, {
|
|
31
|
+
logger: createLogger('prover-node:blob-sink:client')
|
|
32
|
+
});
|
|
17
33
|
const log = deps.log ?? createLogger('prover-node');
|
|
18
|
-
|
|
34
|
+
// Build a key store from file if given or from environment otherwise
|
|
35
|
+
let keyStoreManager;
|
|
36
|
+
const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
|
|
37
|
+
if (keyStoreProvided) {
|
|
38
|
+
const keyStores = loadKeystores(config.keyStoreDirectory);
|
|
39
|
+
keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
|
|
40
|
+
} else {
|
|
41
|
+
const keyStore = createKeyStoreForProver(config);
|
|
42
|
+
if (keyStore) {
|
|
43
|
+
keyStoreManager = new KeystoreManager(keyStore);
|
|
44
|
+
}
|
|
45
|
+
}
|
|
46
|
+
await keyStoreManager?.validateSigners();
|
|
47
|
+
// Extract the prover signers from the key store and verify that we have one.
|
|
48
|
+
const proverSigners = keyStoreManager?.createProverSigners();
|
|
49
|
+
if (proverSigners === undefined) {
|
|
50
|
+
throw new Error('Failed to create prover key store configuration');
|
|
51
|
+
} else if (proverSigners.signers.length === 0) {
|
|
52
|
+
throw new Error('No prover signers found in the key store');
|
|
53
|
+
} else if (!keyStoreProvided) {
|
|
54
|
+
log.warn('KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS');
|
|
55
|
+
}
|
|
56
|
+
log.info(`Creating prover with publishers ${proverSigners.signers.map((signer)=>signer.address.toString()).join()}`);
|
|
57
|
+
// Only consider user provided config if it is valid
|
|
58
|
+
const proverIdInUserConfig = config.proverId === undefined || config.proverId.isZero() ? undefined : config.proverId;
|
|
59
|
+
// ProverId: Take from key store if provided, otherwise from user config if valid, otherwise address of first signer
|
|
60
|
+
const proverId = proverSigners.id ?? proverIdInUserConfig ?? proverSigners.signers[0].address;
|
|
61
|
+
// Now create the prover client configuration from this.
|
|
62
|
+
const proverClientConfig = {
|
|
63
|
+
...config,
|
|
64
|
+
proverId
|
|
65
|
+
};
|
|
66
|
+
await trySnapshotSync(config, log);
|
|
67
|
+
const epochCache = await EpochCache.create(config.l1Contracts.rollupAddress, config);
|
|
68
|
+
const archiver = deps.archiver ?? await createArchiver(config, {
|
|
69
|
+
blobSinkClient,
|
|
70
|
+
epochCache,
|
|
71
|
+
telemetry,
|
|
72
|
+
dateProvider
|
|
73
|
+
}, {
|
|
19
74
|
blockUntilSync: true
|
|
20
|
-
}
|
|
75
|
+
});
|
|
21
76
|
log.verbose(`Created archiver and synced to block ${await archiver.getBlockNumber()}`);
|
|
22
77
|
const worldStateConfig = {
|
|
23
78
|
...config,
|
|
@@ -26,35 +81,44 @@ import { ProverNode } from './prover-node.js';
|
|
|
26
81
|
const worldStateSynchronizer = await createWorldStateSynchronizer(worldStateConfig, archiver, options.prefilledPublicData, telemetry);
|
|
27
82
|
await worldStateSynchronizer.start();
|
|
28
83
|
const broker = deps.broker ?? await createAndStartProvingBroker(config, telemetry);
|
|
29
|
-
const prover = await createProverClient(
|
|
30
|
-
const { l1RpcUrls: rpcUrls, l1ChainId: chainId
|
|
84
|
+
const prover = await createProverClient(proverClientConfig, worldStateSynchronizer, broker, telemetry);
|
|
85
|
+
const { l1RpcUrls: rpcUrls, l1ChainId: chainId } = config;
|
|
31
86
|
const chain = createEthereumChain(rpcUrls, chainId);
|
|
32
|
-
const
|
|
87
|
+
const publicClient = createPublicClient({
|
|
88
|
+
chain: chain.chainInfo,
|
|
89
|
+
transport: fallback(config.l1RpcUrls.map((url)=>http(url))),
|
|
90
|
+
pollingInterval: config.viemPollingIntervalMS
|
|
91
|
+
});
|
|
33
92
|
const rollupContract = new RollupContract(publicClient, config.l1Contracts.rollupAddress.toString());
|
|
34
|
-
const l1TxUtils = deps.l1TxUtils
|
|
35
|
-
|
|
93
|
+
const l1TxUtils = deps.l1TxUtils ? [
|
|
94
|
+
deps.l1TxUtils
|
|
95
|
+
] : await createL1TxUtilsFromEthSignerWithStore(publicClient, proverSigners.signers, {
|
|
96
|
+
...config,
|
|
97
|
+
scope: 'prover'
|
|
98
|
+
}, {
|
|
36
99
|
telemetry,
|
|
37
|
-
|
|
38
|
-
|
|
100
|
+
logger: log.createChild('l1-tx-utils'),
|
|
101
|
+
dateProvider
|
|
39
102
|
});
|
|
40
|
-
const
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
const proverCoordination = await createProverCoordination(config, {
|
|
44
|
-
aztecNodeTxProvider: deps.aztecNodeTxProvider,
|
|
45
|
-
worldStateSynchronizer,
|
|
46
|
-
archiver,
|
|
47
|
-
epochCache,
|
|
103
|
+
const publisherFactory = deps.publisherFactory ?? new ProverPublisherFactory(config, {
|
|
104
|
+
rollupContract,
|
|
105
|
+
publisherManager: new PublisherManager(l1TxUtils, config),
|
|
48
106
|
telemetry
|
|
49
107
|
});
|
|
108
|
+
const proofVerifier = new QueuedIVCVerifier(config, config.realProofs ? await BBCircuitVerifier.new(config) : new TestCircuitVerifier());
|
|
109
|
+
const p2pClient = await createP2PClient(P2PClientType.Prover, config, archiver, proofVerifier, worldStateSynchronizer, epochCache, getPackageVersion() ?? '', dateProvider, telemetry, {
|
|
110
|
+
txCollectionNodeSources: deps.aztecNodeTxProvider ? [
|
|
111
|
+
new NodeRpcTxSource(deps.aztecNodeTxProvider, 'TestNode')
|
|
112
|
+
] : []
|
|
113
|
+
});
|
|
114
|
+
await p2pClient.start();
|
|
50
115
|
const proverNodeConfig = {
|
|
51
|
-
|
|
52
|
-
pollingIntervalMs: config.proverNodePollingIntervalMs,
|
|
53
|
-
maxParallelBlocksPerEpoch: config.proverNodeMaxParallelBlocksPerEpoch,
|
|
54
|
-
txGatheringMaxParallelRequests: config.txGatheringMaxParallelRequests,
|
|
55
|
-
txGatheringIntervalMs: config.txGatheringIntervalMs,
|
|
56
|
-
txGatheringTimeoutMs: config.txGatheringTimeoutMs
|
|
116
|
+
...pick(config, 'proverNodeMaxPendingJobs', 'proverNodeMaxParallelBlocksPerEpoch', 'proverNodePollingIntervalMs', 'proverNodeEpochProvingDelayMs', 'txGatheringMaxParallelRequests', 'txGatheringIntervalMs', 'txGatheringTimeoutMs', 'proverNodeFailedEpochStore', 'proverNodeDisableProofPublish', 'dataDirectory', 'l1ChainId', 'rollupVersion')
|
|
57
117
|
};
|
|
58
|
-
const epochMonitor = await EpochMonitor.create(archiver,
|
|
59
|
-
|
|
118
|
+
const epochMonitor = await EpochMonitor.create(archiver, {
|
|
119
|
+
pollingIntervalMs: config.proverNodePollingIntervalMs,
|
|
120
|
+
provingDelayMs: config.proverNodeEpochProvingDelayMs
|
|
121
|
+
}, telemetry);
|
|
122
|
+
const l1Metrics = new L1Metrics(telemetry.getMeter('ProverNodeL1Metrics'), publicClient, l1TxUtils.map((utils)=>utils.getSenderAddress()));
|
|
123
|
+
return new ProverNode(prover, publisherFactory, archiver, archiver, archiver, worldStateSynchronizer, p2pClient, epochMonitor, rollupContract, l1Metrics, proverNodeConfig, telemetry);
|
|
60
124
|
}
|
package/dest/index.d.ts
CHANGED
package/dest/index.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,cAAc,
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,cAAc,oBAAoB,CAAC;AACnC,cAAc,aAAa,CAAC;AAC5B,cAAc,cAAc,CAAC;AAC7B,cAAc,4BAA4B,CAAC;AAC3C,cAAc,kBAAkB,CAAC"}
|
package/dest/index.js
CHANGED
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
import { Fr } from '@aztec/foundation/fields';
|
|
2
|
+
import { CommitteeAttestation, L2Block } from '@aztec/stdlib/block';
|
|
3
|
+
import { BlockHeader, Tx } from '@aztec/stdlib/tx';
|
|
4
|
+
/** All data from an epoch used in proving. */
|
|
5
|
+
export type EpochProvingJobData = {
|
|
6
|
+
epochNumber: bigint;
|
|
7
|
+
blocks: L2Block[];
|
|
8
|
+
txs: Map<string, Tx>;
|
|
9
|
+
l1ToL2Messages: Record<number, Fr[]>;
|
|
10
|
+
previousBlockHeader: BlockHeader;
|
|
11
|
+
attestations: CommitteeAttestation[];
|
|
12
|
+
};
|
|
13
|
+
export declare function validateEpochProvingJobData(data: EpochProvingJobData): void;
|
|
14
|
+
export declare function serializeEpochProvingJobData(data: EpochProvingJobData): Buffer;
|
|
15
|
+
export declare function deserializeEpochProvingJobData(buf: Buffer): EpochProvingJobData;
|
|
16
|
+
//# sourceMappingURL=epoch-proving-job-data.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"epoch-proving-job-data.d.ts","sourceRoot":"","sources":["../../src/job/epoch-proving-job-data.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,EAAE,EAAE,MAAM,0BAA0B,CAAC;AAE9C,OAAO,EAAE,oBAAoB,EAAE,OAAO,EAAE,MAAM,qBAAqB,CAAC;AACpE,OAAO,EAAE,WAAW,EAAE,EAAE,EAAE,MAAM,kBAAkB,CAAC;AAEnD,8CAA8C;AAC9C,MAAM,MAAM,mBAAmB,GAAG;IAChC,WAAW,EAAE,MAAM,CAAC;IACpB,MAAM,EAAE,OAAO,EAAE,CAAC;IAClB,GAAG,EAAE,GAAG,CAAC,MAAM,EAAE,EAAE,CAAC,CAAC;IACrB,cAAc,EAAE,MAAM,CAAC,MAAM,EAAE,EAAE,EAAE,CAAC,CAAC;IACrC,mBAAmB,EAAE,WAAW,CAAC;IACjC,YAAY,EAAE,oBAAoB,EAAE,CAAC;CACtC,CAAC;AAEF,wBAAgB,2BAA2B,CAAC,IAAI,EAAE,mBAAmB,QAcpE;AAED,wBAAgB,4BAA4B,CAAC,IAAI,EAAE,mBAAmB,GAAG,MAAM,CAsB9E;AAED,wBAAgB,8BAA8B,CAAC,GAAG,EAAE,MAAM,GAAG,mBAAmB,CAoB/E"}
|
|
@@ -0,0 +1,52 @@
|
|
|
1
|
+
import { Fr } from '@aztec/foundation/fields';
|
|
2
|
+
import { BufferReader, serializeToBuffer } from '@aztec/foundation/serialize';
|
|
3
|
+
import { CommitteeAttestation, L2Block } from '@aztec/stdlib/block';
|
|
4
|
+
import { BlockHeader, Tx } from '@aztec/stdlib/tx';
|
|
5
|
+
export function validateEpochProvingJobData(data) {
|
|
6
|
+
if (data.blocks.length > 0 && data.previousBlockHeader.getBlockNumber() + 1 !== data.blocks[0].number) {
|
|
7
|
+
throw new Error(`Initial block number ${data.blocks[0].number} does not match previous block header ${data.previousBlockHeader.getBlockNumber()}`);
|
|
8
|
+
}
|
|
9
|
+
for (const blockNumber of data.blocks.map((block)=>block.number)){
|
|
10
|
+
if (!(blockNumber in data.l1ToL2Messages)) {
|
|
11
|
+
throw new Error(`Missing L1 to L2 messages for block number ${blockNumber}`);
|
|
12
|
+
}
|
|
13
|
+
}
|
|
14
|
+
}
|
|
15
|
+
export function serializeEpochProvingJobData(data) {
|
|
16
|
+
const blocks = data.blocks.map((block)=>block.toBuffer());
|
|
17
|
+
const txs = Array.from(data.txs.values()).map((tx)=>tx.toBuffer());
|
|
18
|
+
const l1ToL2Messages = Object.entries(data.l1ToL2Messages).map(([blockNumber, messages])=>[
|
|
19
|
+
Number(blockNumber),
|
|
20
|
+
messages.length,
|
|
21
|
+
...messages
|
|
22
|
+
]);
|
|
23
|
+
const attestations = data.attestations.map((attestation)=>attestation.toBuffer());
|
|
24
|
+
return serializeToBuffer(Number(data.epochNumber), data.previousBlockHeader, blocks.length, ...blocks, txs.length, ...txs, l1ToL2Messages.length, ...l1ToL2Messages, attestations.length, ...attestations);
|
|
25
|
+
}
|
|
26
|
+
export function deserializeEpochProvingJobData(buf) {
|
|
27
|
+
const reader = BufferReader.asReader(buf);
|
|
28
|
+
const epochNumber = BigInt(reader.readNumber());
|
|
29
|
+
const previousBlockHeader = reader.readObject(BlockHeader);
|
|
30
|
+
const blocks = reader.readVector(L2Block);
|
|
31
|
+
const txArray = reader.readVector(Tx);
|
|
32
|
+
const l1ToL2MessageBlockCount = reader.readNumber();
|
|
33
|
+
const l1ToL2Messages = {};
|
|
34
|
+
for(let i = 0; i < l1ToL2MessageBlockCount; i++){
|
|
35
|
+
const blockNumber = reader.readNumber();
|
|
36
|
+
const messages = reader.readVector(Fr);
|
|
37
|
+
l1ToL2Messages[blockNumber] = messages;
|
|
38
|
+
}
|
|
39
|
+
const attestations = reader.readVector(CommitteeAttestation);
|
|
40
|
+
const txs = new Map(txArray.map((tx)=>[
|
|
41
|
+
tx.getTxHash().toString(),
|
|
42
|
+
tx
|
|
43
|
+
]));
|
|
44
|
+
return {
|
|
45
|
+
epochNumber,
|
|
46
|
+
previousBlockHeader,
|
|
47
|
+
blocks,
|
|
48
|
+
txs,
|
|
49
|
+
l1ToL2Messages,
|
|
50
|
+
attestations
|
|
51
|
+
};
|
|
52
|
+
}
|
|
@@ -1,50 +1,65 @@
|
|
|
1
1
|
import type { PublicProcessorFactory } from '@aztec/simulator/server';
|
|
2
|
-
import type {
|
|
3
|
-
import { type EpochProver, type EpochProvingJobState, type ForkMerkleTreeOperations } from '@aztec/stdlib/interfaces/server';
|
|
4
|
-
import type { L1ToL2MessageSource } from '@aztec/stdlib/messaging';
|
|
5
|
-
import type { Tx } from '@aztec/stdlib/tx';
|
|
2
|
+
import type { L2BlockSource } from '@aztec/stdlib/block';
|
|
3
|
+
import { type EpochProver, type EpochProvingJobState, EpochProvingJobTerminalState, type ForkMerkleTreeOperations } from '@aztec/stdlib/interfaces/server';
|
|
6
4
|
import { type Traceable, type Tracer } from '@aztec/telemetry-client';
|
|
7
|
-
import type {
|
|
5
|
+
import type { ProverNodeJobMetrics } from '../metrics.js';
|
|
8
6
|
import type { ProverNodePublisher } from '../prover-node-publisher.js';
|
|
7
|
+
import { type EpochProvingJobData } from './epoch-proving-job-data.js';
|
|
8
|
+
export type EpochProvingJobOptions = {
|
|
9
|
+
parallelBlockLimit?: number;
|
|
10
|
+
skipEpochCheck?: boolean;
|
|
11
|
+
skipSubmitProof?: boolean;
|
|
12
|
+
};
|
|
9
13
|
/**
|
|
10
|
-
* Job that grabs a range of blocks from the
|
|
14
|
+
* Job that grabs a range of blocks from the unfinalized chain from L1, gets their txs given their hashes,
|
|
11
15
|
* re-executes their public calls, generates a rollup proof, and submits it to L1. This job will update the
|
|
12
16
|
* world state as part of public call execution via the public processor.
|
|
13
17
|
*/
|
|
14
18
|
export declare class EpochProvingJob implements Traceable {
|
|
19
|
+
private data;
|
|
15
20
|
private dbProvider;
|
|
16
|
-
private epochNumber;
|
|
17
|
-
private blocks;
|
|
18
|
-
private txs;
|
|
19
21
|
private prover;
|
|
20
22
|
private publicProcessorFactory;
|
|
21
23
|
private publisher;
|
|
22
24
|
private l2BlockSource;
|
|
23
|
-
private l1ToL2MessageSource;
|
|
24
25
|
private metrics;
|
|
25
26
|
private deadline;
|
|
26
27
|
private config;
|
|
27
|
-
private cleanUp;
|
|
28
28
|
private state;
|
|
29
29
|
private log;
|
|
30
30
|
private uuid;
|
|
31
31
|
private runPromise;
|
|
32
|
+
private epochCheckPromise;
|
|
32
33
|
private deadlineTimeoutHandler;
|
|
33
34
|
readonly tracer: Tracer;
|
|
34
|
-
constructor(
|
|
35
|
-
parallelBlockLimit: number;
|
|
36
|
-
}, cleanUp?: (job: EpochProvingJob) => Promise<void>);
|
|
35
|
+
constructor(data: EpochProvingJobData, dbProvider: Pick<ForkMerkleTreeOperations, 'fork'>, prover: EpochProver, publicProcessorFactory: PublicProcessorFactory, publisher: Pick<ProverNodePublisher, 'submitEpochProof'>, l2BlockSource: L2BlockSource | undefined, metrics: ProverNodeJobMetrics, deadline: Date | undefined, config: EpochProvingJobOptions);
|
|
37
36
|
getId(): string;
|
|
38
37
|
getState(): EpochProvingJobState;
|
|
39
38
|
getEpochNumber(): bigint;
|
|
39
|
+
getDeadline(): Date | undefined;
|
|
40
|
+
getProvingData(): EpochProvingJobData;
|
|
41
|
+
private get epochNumber();
|
|
42
|
+
private get blocks();
|
|
43
|
+
private get txs();
|
|
44
|
+
private get attestations();
|
|
40
45
|
/**
|
|
41
46
|
* Proves the given epoch and submits the proof to L1.
|
|
42
47
|
*/
|
|
43
48
|
run(): Promise<void>;
|
|
49
|
+
/**
|
|
50
|
+
* Create a new db fork for tx processing, inserting all L1 to L2.
|
|
51
|
+
* REFACTOR: The prover already spawns a db fork of its own for each block, so we may be able to do away with just one fork.
|
|
52
|
+
*/
|
|
53
|
+
private createFork;
|
|
44
54
|
private progressState;
|
|
45
55
|
private checkState;
|
|
46
|
-
stop(state?:
|
|
56
|
+
stop(state?: EpochProvingJobTerminalState): Promise<void>;
|
|
47
57
|
private scheduleDeadlineStop;
|
|
58
|
+
/**
|
|
59
|
+
* Kicks off a running promise that queries the archiver for the set of L2 blocks of the current epoch.
|
|
60
|
+
* If those change, stops the proving job with a `rerun` state, so the node re-enqueues it.
|
|
61
|
+
*/
|
|
62
|
+
private scheduleEpochCheck;
|
|
48
63
|
private getBlockHeader;
|
|
49
64
|
private getTxs;
|
|
50
65
|
private getL1ToL2Messages;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"epoch-proving-job.d.ts","sourceRoot":"","sources":["../../src/job/epoch-proving-job.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"epoch-proving-job.d.ts","sourceRoot":"","sources":["../../src/job/epoch-proving-job.ts"],"names":[],"mappings":"AAUA,OAAO,KAAK,EAAmB,sBAAsB,EAAE,MAAM,yBAAyB,CAAC;AACvF,OAAO,KAAK,EAAW,aAAa,EAAE,MAAM,qBAAqB,CAAC;AAClE,OAAO,EACL,KAAK,WAAW,EAChB,KAAK,oBAAoB,EACzB,4BAA4B,EAC5B,KAAK,wBAAwB,EAC9B,MAAM,iCAAiC,CAAC;AAIzC,OAAO,EAAc,KAAK,SAAS,EAAE,KAAK,MAAM,EAAa,MAAM,yBAAyB,CAAC;AAI7F,OAAO,KAAK,EAAE,oBAAoB,EAAE,MAAM,eAAe,CAAC;AAC1D,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,6BAA6B,CAAC;AACvE,OAAO,EAAE,KAAK,mBAAmB,EAA+B,MAAM,6BAA6B,CAAC;AAEpG,MAAM,MAAM,sBAAsB,GAAG;IACnC,kBAAkB,CAAC,EAAE,MAAM,CAAC;IAC5B,cAAc,CAAC,EAAE,OAAO,CAAC;IACzB,eAAe,CAAC,EAAE,OAAO,CAAC;CAC3B,CAAC;AAEF;;;;GAIG;AACH,qBAAa,eAAgB,YAAW,SAAS;IAY7C,OAAO,CAAC,IAAI;IACZ,OAAO,CAAC,UAAU;IAClB,OAAO,CAAC,MAAM;IACd,OAAO,CAAC,sBAAsB;IAC9B,OAAO,CAAC,SAAS;IACjB,OAAO,CAAC,aAAa;IACrB,OAAO,CAAC,OAAO;IACf,OAAO,CAAC,QAAQ;IAChB,OAAO,CAAC,MAAM;IAnBhB,OAAO,CAAC,KAAK,CAAuC;IACpD,OAAO,CAAC,GAAG,CAAiD;IAC5D,OAAO,CAAC,IAAI,CAAS;IAErB,OAAO,CAAC,UAAU,CAA4B;IAC9C,OAAO,CAAC,iBAAiB,CAA6B;IACtD,OAAO,CAAC,sBAAsB,CAA6B;IAE3D,SAAgB,MAAM,EAAE,MAAM,CAAC;gBAGrB,IAAI,EAAE,mBAAmB,EACzB,UAAU,EAAE,IAAI,CAAC,wBAAwB,EAAE,MAAM,CAAC,EAClD,MAAM,EAAE,WAAW,EACnB,sBAAsB,EAAE,sBAAsB,EAC9C,SAAS,EAAE,IAAI,CAAC,mBAAmB,EAAE,kBAAkB,CAAC,EACxD,aAAa,EAAE,aAAa,GAAG,SAAS,EACxC,OAAO,EAAE,oBAAoB,EAC7B,QAAQ,EAAE,IAAI,GAAG,SAAS,EAC1B,MAAM,EAAE,sBAAsB;IAOjC,KAAK,IAAI,MAAM;IAIf,QAAQ,IAAI,oBAAoB;IAIhC,cAAc,IAAI,MAAM;IAIxB,WAAW,IAAI,IAAI,GAAG,SAAS;IAI/B,cAAc,IAAI,mBAAmB;IAI5C,OAAO,KAAK,WAAW,GAEtB;IAED,OAAO,KAAK,MAAM,GAEjB;IAED,OAAO,KAAK,GAAG,GAEd;IAED,OAAO,KAAK,YAAY,GAEvB;IAED;;OAEG;IAIU,GAAG;IAiKhB;;;OAGG;YACW,UAAU;IAgBxB,OAAO,CAAC,aAAa;IAKrB,OAAO,CAAC,UAAU;IAML,IAAI,CAAC,KAAK,GAAE,4BAAwC;IAQjE,OAAO,CAAC,oBAAoB;IAoB5B;;;OAGG;YACW,kBAAkB;IAiChC,OAAO,CAAC,cAAc;IAiBtB,OAAO,CAAC,MAAM;IAId,OAAO,CAAC,iBAAiB;YAIX,UAAU;CAmBzB;AASD,OAAO,EAAE,KAAK,oBAAoB,EAAE,CAAC"}
|