@aztec/prover-node 0.0.0-test.1 → 0.0.1-commit.b655e406

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (81) hide show
  1. package/dest/actions/download-epoch-proving-job.d.ts +18 -0
  2. package/dest/actions/download-epoch-proving-job.d.ts.map +1 -0
  3. package/dest/actions/download-epoch-proving-job.js +37 -0
  4. package/dest/actions/index.d.ts +3 -0
  5. package/dest/actions/index.d.ts.map +1 -0
  6. package/dest/actions/index.js +2 -0
  7. package/dest/actions/rerun-epoch-proving-job.d.ts +11 -0
  8. package/dest/actions/rerun-epoch-proving-job.d.ts.map +1 -0
  9. package/dest/actions/rerun-epoch-proving-job.js +40 -0
  10. package/dest/actions/upload-epoch-proof-failure.d.ts +15 -0
  11. package/dest/actions/upload-epoch-proof-failure.d.ts.map +1 -0
  12. package/dest/actions/upload-epoch-proof-failure.js +78 -0
  13. package/dest/bin/run-failed-epoch.d.ts +2 -0
  14. package/dest/bin/run-failed-epoch.d.ts.map +1 -0
  15. package/dest/bin/run-failed-epoch.js +67 -0
  16. package/dest/config.d.ts +12 -9
  17. package/dest/config.d.ts.map +1 -1
  18. package/dest/config.js +81 -14
  19. package/dest/factory.d.ts +12 -8
  20. package/dest/factory.d.ts.map +1 -1
  21. package/dest/factory.js +95 -31
  22. package/dest/index.d.ts +1 -1
  23. package/dest/index.d.ts.map +1 -1
  24. package/dest/index.js +1 -1
  25. package/dest/job/epoch-proving-job-data.d.ts +16 -0
  26. package/dest/job/epoch-proving-job-data.d.ts.map +1 -0
  27. package/dest/job/epoch-proving-job-data.js +52 -0
  28. package/dest/job/epoch-proving-job.d.ts +30 -15
  29. package/dest/job/epoch-proving-job.d.ts.map +1 -1
  30. package/dest/job/epoch-proving-job.js +149 -50
  31. package/dest/metrics.d.ts +28 -4
  32. package/dest/metrics.d.ts.map +1 -1
  33. package/dest/metrics.js +141 -35
  34. package/dest/monitors/epoch-monitor.d.ts +3 -1
  35. package/dest/monitors/epoch-monitor.d.ts.map +1 -1
  36. package/dest/monitors/epoch-monitor.js +15 -2
  37. package/dest/prover-node-publisher.d.ts +7 -10
  38. package/dest/prover-node-publisher.d.ts.map +1 -1
  39. package/dest/prover-node-publisher.js +59 -60
  40. package/dest/prover-node.d.ts +43 -39
  41. package/dest/prover-node.d.ts.map +1 -1
  42. package/dest/prover-node.js +171 -100
  43. package/dest/prover-publisher-factory.d.ts +21 -0
  44. package/dest/prover-publisher-factory.d.ts.map +1 -0
  45. package/dest/prover-publisher-factory.js +26 -0
  46. package/dest/test/index.d.ts +4 -2
  47. package/dest/test/index.d.ts.map +1 -1
  48. package/dest/test/index.js +1 -3
  49. package/package.json +36 -31
  50. package/src/actions/download-epoch-proving-job.ts +44 -0
  51. package/src/actions/index.ts +2 -0
  52. package/src/actions/rerun-epoch-proving-job.ts +61 -0
  53. package/src/actions/upload-epoch-proof-failure.ts +88 -0
  54. package/src/bin/run-failed-epoch.ts +77 -0
  55. package/src/config.ts +108 -24
  56. package/src/factory.ts +161 -43
  57. package/src/index.ts +1 -1
  58. package/src/job/epoch-proving-job-data.ts +76 -0
  59. package/src/job/epoch-proving-job.ts +215 -50
  60. package/src/metrics.ts +135 -37
  61. package/src/monitors/epoch-monitor.ts +16 -5
  62. package/src/prover-node-publisher.ts +93 -86
  63. package/src/prover-node.ts +203 -126
  64. package/src/prover-publisher-factory.ts +37 -0
  65. package/src/test/index.ts +7 -4
  66. package/dest/http.d.ts +0 -8
  67. package/dest/http.d.ts.map +0 -1
  68. package/dest/http.js +0 -9
  69. package/dest/prover-coordination/config.d.ts +0 -7
  70. package/dest/prover-coordination/config.d.ts.map +0 -1
  71. package/dest/prover-coordination/config.js +0 -11
  72. package/dest/prover-coordination/factory.d.ts +0 -22
  73. package/dest/prover-coordination/factory.d.ts.map +0 -1
  74. package/dest/prover-coordination/factory.js +0 -42
  75. package/dest/prover-coordination/index.d.ts +0 -3
  76. package/dest/prover-coordination/index.d.ts.map +0 -1
  77. package/dest/prover-coordination/index.js +0 -2
  78. package/src/http.ts +0 -13
  79. package/src/prover-coordination/config.ts +0 -17
  80. package/src/prover-coordination/factory.ts +0 -72
  81. package/src/prover-coordination/index.ts +0 -2
@@ -0,0 +1,18 @@
1
+ import type { Logger } from '@aztec/foundation/log';
2
+ /**
3
+ * Given a location returned by `uploadEpochProofFailure`, downloads the world state and archiver snapshots
4
+ * and the proving job data, so we can re-run the job later using `rerunEpochProvingJob`. This is decoupled
5
+ * from actually proving so we can download once and run multiple times.
6
+ */
7
+ export declare function downloadEpochProvingJob(location: string, log: Logger, config: {
8
+ dataDirectory: string;
9
+ jobDataDownloadPath: string;
10
+ }): Promise<{
11
+ rollupVersion: number;
12
+ l2BlockNumber: number;
13
+ l2BlockHash: string;
14
+ rollupAddress: import("@aztec/foundation/schemas").EthAddress;
15
+ l1ChainId: number;
16
+ l1BlockNumber: number;
17
+ }>;
18
+ //# sourceMappingURL=download-epoch-proving-job.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"download-epoch-proving-job.d.ts","sourceRoot":"","sources":["../../src/actions/download-epoch-proving-job.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,MAAM,EAAE,MAAM,uBAAuB,CAAC;AAUpD;;;;GAIG;AACH,wBAAsB,uBAAuB,CAC3C,QAAQ,EAAE,MAAM,EAChB,GAAG,EAAE,MAAM,EACX,MAAM,EAAE;IACN,aAAa,EAAE,MAAM,CAAC;IACtB,mBAAmB,EAAE,MAAM,CAAC;CAC7B;;;;;;;GAqBF"}
@@ -0,0 +1,37 @@
1
+ import { jsonParseWithSchema } from '@aztec/foundation/json-rpc';
2
+ import { urlJoin } from '@aztec/foundation/string';
3
+ import { snapshotSync } from '@aztec/node-lib/actions';
4
+ import { createReadOnlyFileStore } from '@aztec/stdlib/file-store';
5
+ import { UploadSnapshotMetadataSchema, makeSnapshotPaths } from '@aztec/stdlib/snapshots';
6
+ import { readFileSync } from 'fs';
7
+ import { deserializeEpochProvingJobData } from '../job/epoch-proving-job-data.js';
8
+ /**
9
+ * Given a location returned by `uploadEpochProofFailure`, downloads the world state and archiver snapshots
10
+ * and the proving job data, so we can re-run the job later using `rerunEpochProvingJob`. This is decoupled
11
+ * from actually proving so we can download once and run multiple times.
12
+ */ export async function downloadEpochProvingJob(location, log, config) {
13
+ log.info(`Downloading epoch proving job data from ${location}`);
14
+ const fileStore = await createReadOnlyFileStore(location);
15
+ const metadataUrl = urlJoin(location, 'metadata.json');
16
+ const metadataRaw = await fileStore.read(metadataUrl);
17
+ const metadata = jsonParseWithSchema(metadataRaw.toString(), UploadSnapshotMetadataSchema);
18
+ const dataUrls = makeSnapshotPaths(location);
19
+ log.info(`Downloading state snapshot from ${location} to local data directory`, {
20
+ metadata,
21
+ dataUrls
22
+ });
23
+ await snapshotSync({
24
+ dataUrls
25
+ }, log, {
26
+ ...config,
27
+ ...metadata,
28
+ snapshotsUrl: location
29
+ });
30
+ const dataPath = urlJoin(location, 'data.bin');
31
+ const localPath = config.jobDataDownloadPath;
32
+ log.info(`Downloading epoch proving job data from ${dataPath} to ${localPath}`);
33
+ await fileStore.download(dataPath, localPath);
34
+ const jobData = deserializeEpochProvingJobData(readFileSync(localPath));
35
+ log.info(`Epoch proving job data for epoch ${jobData.epochNumber} downloaded successfully`);
36
+ return metadata;
37
+ }
@@ -0,0 +1,3 @@
1
+ export * from './download-epoch-proving-job.js';
2
+ export * from './rerun-epoch-proving-job.js';
3
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/actions/index.ts"],"names":[],"mappings":"AAAA,cAAc,iCAAiC,CAAC;AAChD,cAAc,8BAA8B,CAAC"}
@@ -0,0 +1,2 @@
1
+ export * from './download-epoch-proving-job.js';
2
+ export * from './rerun-epoch-proving-job.js';
@@ -0,0 +1,11 @@
1
+ import type { Logger } from '@aztec/foundation/log';
2
+ import type { DataStoreConfig } from '@aztec/kv-store/config';
3
+ import { type ProverClientConfig } from '@aztec/prover-client';
4
+ import { ProverBrokerConfig } from '@aztec/prover-client/broker';
5
+ /**
6
+ * Given a local folder where `downloadEpochProvingJob` was called, creates a new archiver and world state
7
+ * using the state snapshots, and creates a new epoch proving job to prove the downloaded proving job.
8
+ * Proving is done with a local proving broker and agents as specified by the config.
9
+ */
10
+ export declare function rerunEpochProvingJob(localPath: string, log: Logger, config: DataStoreConfig & ProverBrokerConfig & ProverClientConfig): Promise<"initialized" | "processing" | "awaiting-prover" | "publishing-proof" | "completed" | "failed" | "stopped" | "timed-out" | "reorg">;
11
+ //# sourceMappingURL=rerun-epoch-proving-job.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"rerun-epoch-proving-job.d.ts","sourceRoot":"","sources":["../../src/actions/rerun-epoch-proving-job.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,MAAM,EAAE,MAAM,uBAAuB,CAAC;AACpD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,wBAAwB,CAAC;AAC9D,OAAO,EAAE,KAAK,kBAAkB,EAAsB,MAAM,sBAAsB,CAAC;AACnF,OAAO,EAAE,kBAAkB,EAA+B,MAAM,6BAA6B,CAAC;AAW9F;;;;GAIG;AACH,wBAAsB,oBAAoB,CACxC,SAAS,EAAE,MAAM,EACjB,GAAG,EAAE,MAAM,EACX,MAAM,EAAE,eAAe,GAAG,kBAAkB,GAAG,kBAAkB,+IAqClE"}
@@ -0,0 +1,40 @@
1
+ import { createArchiverStore } from '@aztec/archiver';
2
+ import { createProverClient } from '@aztec/prover-client';
3
+ import { createAndStartProvingBroker } from '@aztec/prover-client/broker';
4
+ import { PublicProcessorFactory } from '@aztec/simulator/server';
5
+ import { getTelemetryClient } from '@aztec/telemetry-client';
6
+ import { createWorldState } from '@aztec/world-state';
7
+ import { readFileSync } from 'fs';
8
+ import { deserializeEpochProvingJobData } from '../job/epoch-proving-job-data.js';
9
+ import { EpochProvingJob } from '../job/epoch-proving-job.js';
10
+ import { ProverNodeJobMetrics } from '../metrics.js';
11
+ /**
12
+ * Given a local folder where `downloadEpochProvingJob` was called, creates a new archiver and world state
13
+ * using the state snapshots, and creates a new epoch proving job to prove the downloaded proving job.
14
+ * Proving is done with a local proving broker and agents as specified by the config.
15
+ */ export async function rerunEpochProvingJob(localPath, log, config) {
16
+ const jobData = deserializeEpochProvingJobData(readFileSync(localPath));
17
+ log.info(`Loaded proving job data for epoch ${jobData.epochNumber}`);
18
+ const telemetry = getTelemetryClient();
19
+ const metrics = new ProverNodeJobMetrics(telemetry.getMeter('prover-job'), telemetry.getTracer('prover-job'));
20
+ const worldState = await createWorldState(config);
21
+ const archiver = await createArchiverStore(config);
22
+ const publicProcessorFactory = new PublicProcessorFactory(archiver);
23
+ const publisher = {
24
+ submitEpochProof: ()=>Promise.resolve(true)
25
+ };
26
+ const l2BlockSourceForReorgDetection = undefined;
27
+ const deadline = undefined;
28
+ // This starts a local proving broker that does not get exposed as a service. This should be good enough for
29
+ // smallish epochs to be proven if we run on a large machine, but as epochs grow larger, we may want to switch
30
+ // this out for a live proving broker with multiple agents that we can connect to.
31
+ const broker = await createAndStartProvingBroker(config, telemetry);
32
+ const prover = await createProverClient(config, worldState, broker, telemetry);
33
+ const provingJob = new EpochProvingJob(jobData, worldState, prover.createEpochProver(), publicProcessorFactory, publisher, l2BlockSourceForReorgDetection, metrics, deadline, {
34
+ skipEpochCheck: true
35
+ });
36
+ log.info(`Rerunning epoch proving job for epoch ${jobData.epochNumber}`);
37
+ await provingJob.run();
38
+ log.info(`Completed job for epoch ${jobData.epochNumber} with status ${provingJob.getState()}`);
39
+ return provingJob.getState();
40
+ }
@@ -0,0 +1,15 @@
1
+ import { type Archiver } from '@aztec/archiver';
2
+ import type { Logger } from '@aztec/foundation/log';
3
+ import type { DataStoreConfig } from '@aztec/kv-store/config';
4
+ import type { ChainConfig } from '@aztec/stdlib/config';
5
+ import type { WorldStateSynchronizer } from '@aztec/stdlib/interfaces/server';
6
+ import { type EpochProvingJobData } from '../job/epoch-proving-job-data.js';
7
+ type UploadEpochProofConfig = Pick<ChainConfig, 'l1ChainId' | 'rollupVersion'> & Pick<DataStoreConfig, 'dataDirectory'>;
8
+ /**
9
+ * Uploads a snapshot of world state and archiver (requires pausing them) along with the proving job data,
10
+ * so we can download and re-run the job later under the same conditions.
11
+ * @param location The location to upload the data to (used to create the `FileStore`).
12
+ */
13
+ export declare function uploadEpochProofFailure(location: string, jobId: string, jobData: EpochProvingJobData, archiver: Archiver, worldState: WorldStateSynchronizer, config: UploadEpochProofConfig, log: Logger): Promise<string>;
14
+ export {};
15
+ //# sourceMappingURL=upload-epoch-proof-failure.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"upload-epoch-proof-failure.d.ts","sourceRoot":"","sources":["../../src/actions/upload-epoch-proof-failure.ts"],"names":[],"mappings":"AAAA,OAAO,EAAuB,KAAK,QAAQ,EAAE,MAAM,iBAAiB,CAAC;AAGrE,OAAO,KAAK,EAAE,MAAM,EAAE,MAAM,uBAAuB,CAAC;AAEpD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,wBAAwB,CAAC;AAE9D,OAAO,KAAK,EAAE,WAAW,EAAE,MAAM,sBAAsB,CAAC;AAExD,OAAO,KAAK,EAAE,sBAAsB,EAAE,MAAM,iCAAiC,CAAC;AAQ9E,OAAO,EAAE,KAAK,mBAAmB,EAAgC,MAAM,kCAAkC,CAAC;AAE1G,KAAK,sBAAsB,GAAG,IAAI,CAAC,WAAW,EAAE,WAAW,GAAG,eAAe,CAAC,GAAG,IAAI,CAAC,eAAe,EAAE,eAAe,CAAC,CAAC;AAKxH;;;;GAIG;AACH,wBAAsB,uBAAuB,CAC3C,QAAQ,EAAE,MAAM,EAChB,KAAK,EAAE,MAAM,EACb,OAAO,EAAE,mBAAmB,EAC5B,QAAQ,EAAE,QAAQ,EAClB,UAAU,EAAE,sBAAsB,EAClC,MAAM,EAAE,sBAAsB,EAC9B,GAAG,EAAE,MAAM,mBAuCZ"}
@@ -0,0 +1,78 @@
1
+ import { ARCHIVER_DB_VERSION } from '@aztec/archiver';
2
+ import { tryRmDir } from '@aztec/foundation/fs';
3
+ import { jsonStringify } from '@aztec/foundation/json-rpc';
4
+ import { isoDate } from '@aztec/foundation/string';
5
+ import { buildSnapshotMetadata, createBackups } from '@aztec/node-lib/actions';
6
+ import { createFileStore } from '@aztec/stdlib/file-store';
7
+ import { getBasePath, uploadSnapshotData } from '@aztec/stdlib/snapshots';
8
+ import { WORLD_STATE_DB_VERSION } from '@aztec/world-state';
9
+ import { mkdtemp } from 'fs/promises';
10
+ import { tmpdir } from 'os';
11
+ import { dirname, join } from 'path';
12
+ import { serializeEpochProvingJobData } from '../job/epoch-proving-job-data.js';
13
+ /** Whether uploaded data to the file store should be of public access. */ const PUBLIC_UPLOADS = true;
14
+ /**
15
+ * Uploads a snapshot of world state and archiver (requires pausing them) along with the proving job data,
16
+ * so we can download and re-run the job later under the same conditions.
17
+ * @param location The location to upload the data to (used to create the `FileStore`).
18
+ */ export async function uploadEpochProofFailure(location, jobId, jobData, archiver, worldState, config, log) {
19
+ const epochNumber = jobData.epochNumber;
20
+ log.warn(`Uploading epoch proof failure for ${epochNumber} to ${location}`, {
21
+ epochNumber,
22
+ jobId,
23
+ location
24
+ });
25
+ const backupDir = await mkdtemp(join(config.dataDirectory ?? tmpdir(), 'epoch-proof-data-'));
26
+ const store = await createFileStore(location);
27
+ if (!store) {
28
+ throw new Error(`Failed to create file store for epoch proof failure upload for location ${location}.`);
29
+ }
30
+ try {
31
+ const versions = {
32
+ archiver: ARCHIVER_DB_VERSION,
33
+ worldState: WORLD_STATE_DB_VERSION
34
+ };
35
+ const uploadMetadata = await buildSnapshotMetadata(archiver, config);
36
+ const paths = await createBackups(backupDir, archiver, worldState, log);
37
+ const basePath = `${getBasePath(uploadMetadata)}/${epochNumber}-${isoDate()}-${jobId}`;
38
+ const pathFor = (key)=>`${basePath}/${key}.db`;
39
+ const [metadata, dataUrl, metadataUrl] = await Promise.all([
40
+ uploadSnapshotData(paths, versions, uploadMetadata, store, {
41
+ pathFor,
42
+ private: !PUBLIC_UPLOADS
43
+ }),
44
+ uploadJobData(jobData, store, basePath),
45
+ uploadSnapshotMetadata(uploadMetadata, store, basePath)
46
+ ]);
47
+ const baseUrl = dirname(metadataUrl);
48
+ log.warn(`Uploaded epoch ${epochNumber} proof failure data to ${baseUrl}`, {
49
+ epochNumber,
50
+ location,
51
+ basePath,
52
+ metadataUrl,
53
+ dataUrl,
54
+ metadata,
55
+ jobId
56
+ });
57
+ return baseUrl;
58
+ } finally{
59
+ log.info(`Cleaning up backup dir ${backupDir}`);
60
+ await tryRmDir(backupDir, log);
61
+ }
62
+ }
63
+ async function uploadJobData(jobData, store, basePath) {
64
+ const data = serializeEpochProvingJobData(jobData);
65
+ const path = `${basePath}/data.bin`;
66
+ return await store.save(path, data, {
67
+ compress: true,
68
+ public: PUBLIC_UPLOADS
69
+ });
70
+ }
71
+ async function uploadSnapshotMetadata(metadata, store, basePath) {
72
+ const data = Buffer.from(jsonStringify(metadata), 'utf-8');
73
+ const path = `${basePath}/metadata.json`;
74
+ return await store.save(path, data, {
75
+ compress: false,
76
+ public: PUBLIC_UPLOADS
77
+ });
78
+ }
@@ -0,0 +1,2 @@
1
+ export {};
2
+ //# sourceMappingURL=run-failed-epoch.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"run-failed-epoch.d.ts","sourceRoot":"","sources":["../../src/bin/run-failed-epoch.ts"],"names":[],"mappings":""}
@@ -0,0 +1,67 @@
1
+ /* eslint-disable no-console */ import { EthAddress } from '@aztec/foundation/eth-address';
2
+ import { jsonParseWithSchema, jsonStringify } from '@aztec/foundation/json-rpc';
3
+ import { createLogger } from '@aztec/foundation/log';
4
+ import { downloadEpochProvingJob, getProverNodeConfigFromEnv, rerunEpochProvingJob } from '@aztec/prover-node';
5
+ import { UploadSnapshotMetadataSchema } from '@aztec/stdlib/snapshots';
6
+ import { existsSync, mkdirSync } from 'fs';
7
+ import { readFile, writeFile } from 'fs/promises';
8
+ import { basename, join } from 'path';
9
+ const logger = createLogger('prover-node:run-failed-epoch');
10
+ function printUsage() {
11
+ console.error('Usage: run-failed-epoch <proof-uri> [out-dir=./data]');
12
+ }
13
+ async function rerunFailedEpoch(provingJobUrl, baseLocalDir) {
14
+ const localDir = join(baseLocalDir, basename(provingJobUrl));
15
+ const jobPath = join(localDir, 'data.bin');
16
+ const dataDir = join(localDir, 'state');
17
+ const env = getProverNodeConfigFromEnv();
18
+ const config = {
19
+ ...getProverNodeConfigFromEnv(),
20
+ dataDirectory: dataDir,
21
+ dataStoreMapSizeKb: env.dataStoreMapSizeKb ?? 1024 * 1024,
22
+ proverId: env.proverId ?? EthAddress.random()
23
+ };
24
+ let metadata;
25
+ const metadataPath = join(localDir, 'metadata.json');
26
+ if (existsSync(metadataPath)) {
27
+ logger.info(`Using downloaded data`);
28
+ metadata = jsonParseWithSchema(await readFile(metadataPath, 'utf-8'), UploadSnapshotMetadataSchema);
29
+ } else {
30
+ logger.info(`Downloading epoch proving job data and state from ${provingJobUrl} to ${localDir}`);
31
+ metadata = await downloadEpochProvingJob(provingJobUrl, logger, {
32
+ jobDataDownloadPath: jobPath,
33
+ dataDirectory: dataDir
34
+ });
35
+ await writeFile(metadataPath, jsonStringify(metadata, true));
36
+ logger.info(`Download to ${localDir} complete`);
37
+ }
38
+ logger.info(`Rerunning proving job from ${jobPath} with state from ${dataDir}`, metadata);
39
+ const result = await rerunEpochProvingJob(jobPath, logger, {
40
+ ...config,
41
+ l1Contracts: {
42
+ rollupAddress: metadata.rollupAddress
43
+ },
44
+ rollupVersion: metadata.rollupVersion
45
+ });
46
+ console.error(`Epoch proving job complete with result ${result}`);
47
+ }
48
+ async function main() {
49
+ if (process.argv[2] === '--help') {
50
+ printUsage();
51
+ return;
52
+ }
53
+ const uri = process.argv[2];
54
+ const outDir = process.argv[3] || './data';
55
+ if (!uri) {
56
+ printUsage();
57
+ throw new Error('Missing URL to epoch proving job');
58
+ }
59
+ mkdirSync(outDir, {
60
+ recursive: true
61
+ });
62
+ await rerunFailedEpoch(uri, outDir);
63
+ }
64
+ main().catch((err)=>{
65
+ console.error(err);
66
+ process.exit(1);
67
+ });
package/dest/config.d.ts CHANGED
@@ -1,28 +1,31 @@
1
1
  import { type ArchiverConfig } from '@aztec/archiver/config';
2
2
  import type { ACVMConfig, BBConfig } from '@aztec/bb-prover/config';
3
+ import { type GenesisStateConfig } from '@aztec/ethereum';
3
4
  import { type ConfigMappingsType } from '@aztec/foundation/config';
4
5
  import { type DataStoreConfig } from '@aztec/kv-store/config';
6
+ import { type KeyStore, type KeyStoreConfig } from '@aztec/node-keystore';
7
+ import { type SharedNodeConfig } from '@aztec/node-lib/config';
5
8
  import { type P2PConfig } from '@aztec/p2p/config';
6
9
  import { type ProverAgentConfig, type ProverBrokerConfig } from '@aztec/prover-client/broker';
7
- import { type ProverClientConfig } from '@aztec/prover-client/config';
10
+ import { type ProverClientUserConfig } from '@aztec/prover-client/config';
8
11
  import { type PublisherConfig, type TxSenderConfig } from '@aztec/sequencer-client/config';
9
12
  import { type WorldStateConfig } from '@aztec/world-state/config';
10
- import { type ProverCoordinationConfig } from './prover-coordination/config.js';
11
- export type ProverNodeConfig = ArchiverConfig & ProverClientConfig & P2PConfig & WorldStateConfig & PublisherConfig & TxSenderConfig & DataStoreConfig & ProverCoordinationConfig & SpecificProverNodeConfig & {
12
- /** Whether to populate the genesis state with initial fee juice for the test accounts */
13
- testAccounts: boolean;
14
- };
15
- type SpecificProverNodeConfig = {
13
+ export type ProverNodeConfig = ArchiverConfig & ProverClientUserConfig & P2PConfig & WorldStateConfig & PublisherConfig & TxSenderConfig & DataStoreConfig & KeyStoreConfig & SharedNodeConfig & SpecificProverNodeConfig & GenesisStateConfig;
14
+ export type SpecificProverNodeConfig = {
16
15
  proverNodeMaxPendingJobs: number;
17
16
  proverNodePollingIntervalMs: number;
18
17
  proverNodeMaxParallelBlocksPerEpoch: number;
18
+ proverNodeFailedEpochStore: string | undefined;
19
+ proverNodeEpochProvingDelayMs: number | undefined;
20
+ proverNodeDisableProofPublish?: boolean;
19
21
  txGatheringTimeoutMs: number;
20
22
  txGatheringIntervalMs: number;
21
- txGatheringMaxParallelRequests: number;
23
+ txGatheringBatchSize: number;
24
+ txGatheringMaxParallelRequestsPerNode: number;
22
25
  };
23
26
  export declare const proverNodeConfigMappings: ConfigMappingsType<ProverNodeConfig>;
24
27
  export declare function getProverNodeConfigFromEnv(): ProverNodeConfig;
25
28
  export declare function getProverNodeBrokerConfigFromEnv(): ProverBrokerConfig;
26
29
  export declare function getProverNodeAgentConfigFromEnv(): ProverAgentConfig & BBConfig & ACVMConfig;
27
- export {};
30
+ export declare function createKeyStoreForProver(config: ProverNodeConfig): KeyStore | undefined;
28
31
  //# sourceMappingURL=config.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"config.d.ts","sourceRoot":"","sources":["../src/config.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,KAAK,cAAc,EAA0B,MAAM,wBAAwB,CAAC;AACrF,OAAO,KAAK,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,yBAAyB,CAAC;AACpE,OAAO,EACL,KAAK,kBAAkB,EAIxB,MAAM,0BAA0B,CAAC;AAClC,OAAO,EAAE,KAAK,eAAe,EAAsB,MAAM,wBAAwB,CAAC;AAClF,OAAO,EAAE,KAAK,SAAS,EAAqB,MAAM,mBAAmB,CAAC;AACtE,OAAO,EACL,KAAK,iBAAiB,EACtB,KAAK,kBAAkB,EAGxB,MAAM,6BAA6B,CAAC;AACrC,OAAO,EAAE,KAAK,kBAAkB,EAAgD,MAAM,6BAA6B,CAAC;AACpH,OAAO,EACL,KAAK,eAAe,EACpB,KAAK,cAAc,EAGpB,MAAM,gCAAgC,CAAC;AACxC,OAAO,EAAE,KAAK,gBAAgB,EAA4B,MAAM,2BAA2B,CAAC;AAE5F,OAAO,EAAE,KAAK,wBAAwB,EAAoC,MAAM,iCAAiC,CAAC;AAElH,MAAM,MAAM,gBAAgB,GAAG,cAAc,GAC3C,kBAAkB,GAClB,SAAS,GACT,gBAAgB,GAChB,eAAe,GACf,cAAc,GACd,eAAe,GACf,wBAAwB,GACxB,wBAAwB,GAAG;IACzB,yFAAyF;IACzF,YAAY,EAAE,OAAO,CAAC;CACvB,CAAC;AAEJ,KAAK,wBAAwB,GAAG;IAC9B,wBAAwB,EAAE,MAAM,CAAC;IACjC,2BAA2B,EAAE,MAAM,CAAC;IACpC,mCAAmC,EAAE,MAAM,CAAC;IAC5C,oBAAoB,EAAE,MAAM,CAAC;IAC7B,qBAAqB,EAAE,MAAM,CAAC;IAC9B,8BAA8B,EAAE,MAAM,CAAC;CACxC,CAAC;AAmCF,eAAO,MAAM,wBAAwB,EAAE,kBAAkB,CAAC,gBAAgB,CAezE,CAAC;AAEF,wBAAgB,0BAA0B,IAAI,gBAAgB,CAE7D;AAED,wBAAgB,gCAAgC,IAAI,kBAAkB,CAIrE;AAED,wBAAgB,+BAA+B,IAAI,iBAAiB,GAAG,QAAQ,GAAG,UAAU,CAK3F"}
1
+ {"version":3,"file":"config.d.ts","sourceRoot":"","sources":["../src/config.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,KAAK,cAAc,EAA0B,MAAM,wBAAwB,CAAC;AACrF,OAAO,KAAK,EAAE,UAAU,EAAE,QAAQ,EAAE,MAAM,yBAAyB,CAAC;AACpE,OAAO,EAAE,KAAK,kBAAkB,EAA8B,MAAM,iBAAiB,CAAC;AACtF,OAAO,EACL,KAAK,kBAAkB,EAIxB,MAAM,0BAA0B,CAAC;AAClC,OAAO,EAAE,KAAK,eAAe,EAAsB,MAAM,wBAAwB,CAAC;AAClF,OAAO,EAAE,KAAK,QAAQ,EAAE,KAAK,cAAc,EAA+C,MAAM,sBAAsB,CAAC;AACvH,OAAO,EAAE,KAAK,gBAAgB,EAA4B,MAAM,wBAAwB,CAAC;AACzF,OAAO,EAAE,KAAK,SAAS,EAAqB,MAAM,mBAAmB,CAAC;AACtE,OAAO,EACL,KAAK,iBAAiB,EACtB,KAAK,kBAAkB,EAGxB,MAAM,6BAA6B,CAAC;AACrC,OAAO,EAAE,KAAK,sBAAsB,EAAgD,MAAM,6BAA6B,CAAC;AACxH,OAAO,EACL,KAAK,eAAe,EACpB,KAAK,cAAc,EAGpB,MAAM,gCAAgC,CAAC;AACxC,OAAO,EAAE,KAAK,gBAAgB,EAA4B,MAAM,2BAA2B,CAAC;AAE5F,MAAM,MAAM,gBAAgB,GAAG,cAAc,GAC3C,sBAAsB,GACtB,SAAS,GACT,gBAAgB,GAChB,eAAe,GACf,cAAc,GACd,eAAe,GACf,cAAc,GACd,gBAAgB,GAChB,wBAAwB,GACxB,kBAAkB,CAAC;AAErB,MAAM,MAAM,wBAAwB,GAAG;IACrC,wBAAwB,EAAE,MAAM,CAAC;IACjC,2BAA2B,EAAE,MAAM,CAAC;IACpC,mCAAmC,EAAE,MAAM,CAAC;IAC5C,0BAA0B,EAAE,MAAM,GAAG,SAAS,CAAC;IAC/C,6BAA6B,EAAE,MAAM,GAAG,SAAS,CAAC;IAClD,6BAA6B,CAAC,EAAE,OAAO,CAAC;IACxC,oBAAoB,EAAE,MAAM,CAAC;IAC7B,qBAAqB,EAAE,MAAM,CAAC;IAC9B,oBAAoB,EAAE,MAAM,CAAC;IAC7B,qCAAqC,EAAE,MAAM,CAAC;CAC/C,CAAC;AAsDF,eAAO,MAAM,wBAAwB,EAAE,kBAAkB,CAAC,gBAAgB,CAYzE,CAAC;AAEF,wBAAgB,0BAA0B,IAAI,gBAAgB,CAE7D;AAED,wBAAgB,gCAAgC,IAAI,kBAAkB,CAIrE;AAED,wBAAgB,+BAA+B,IAAI,iBAAiB,GAAG,QAAQ,GAAG,UAAU,CAK3F;AA0DD,wBAAgB,uBAAuB,CAAC,MAAM,EAAE,gBAAgB,GAAG,QAAQ,GAAG,SAAS,CAMtF"}
package/dest/config.js CHANGED
@@ -1,12 +1,14 @@
1
1
  import { archiverConfigMappings } from '@aztec/archiver/config';
2
+ import { genesisStateConfigMappings } from '@aztec/ethereum';
2
3
  import { booleanConfigHelper, getConfigFromMappings, numberConfigHelper } from '@aztec/foundation/config';
3
4
  import { dataConfigMappings } from '@aztec/kv-store/config';
5
+ import { ethPrivateKeySchema, keyStoreConfigMappings } from '@aztec/node-keystore';
6
+ import { sharedNodeConfigMappings } from '@aztec/node-lib/config';
4
7
  import { p2pConfigMappings } from '@aztec/p2p/config';
5
8
  import { proverAgentConfigMappings, proverBrokerConfigMappings } from '@aztec/prover-client/broker';
6
9
  import { bbConfigMappings, proverClientConfigMappings } from '@aztec/prover-client/config';
7
10
  import { getPublisherConfigMappings, getTxSenderConfigMappings } from '@aztec/sequencer-client/config';
8
11
  import { worldStateConfigMappings } from '@aztec/world-state/config';
9
- import { proverCoordinationConfigMappings } from './prover-coordination/config.js';
10
12
  const specificProverNodeConfigMappings = {
11
13
  proverNodeMaxPendingJobs: {
12
14
  env: 'PROVER_NODE_MAX_PENDING_JOBS',
@@ -23,37 +25,53 @@ const specificProverNodeConfigMappings = {
23
25
  description: 'The Maximum number of blocks to process in parallel while proving an epoch',
24
26
  ...numberConfigHelper(32)
25
27
  },
26
- txGatheringTimeoutMs: {
27
- env: 'PROVER_NODE_TX_GATHERING_TIMEOUT_MS',
28
- description: 'The maximum amount of time to wait for tx data to be available',
29
- ...numberConfigHelper(60_000)
28
+ proverNodeFailedEpochStore: {
29
+ env: 'PROVER_NODE_FAILED_EPOCH_STORE',
30
+ description: 'File store where to upload node state when an epoch fails to be proven',
31
+ defaultValue: undefined
32
+ },
33
+ proverNodeEpochProvingDelayMs: {
34
+ description: 'Optional delay in milliseconds to wait before proving a new epoch',
35
+ defaultValue: undefined
30
36
  },
31
37
  txGatheringIntervalMs: {
32
38
  env: 'PROVER_NODE_TX_GATHERING_INTERVAL_MS',
33
39
  description: 'How often to check that tx data is available',
34
40
  ...numberConfigHelper(1_000)
35
41
  },
36
- txGatheringMaxParallelRequests: {
37
- env: 'PROVER_NODE_TX_GATHERING_MAX_PARALLEL_REQUESTS',
38
- description: 'How many txs to load up a time',
42
+ txGatheringBatchSize: {
43
+ env: 'PROVER_NODE_TX_GATHERING_BATCH_SIZE',
44
+ description: 'How many transactions to gather from a node in a single request',
45
+ ...numberConfigHelper(10)
46
+ },
47
+ txGatheringMaxParallelRequestsPerNode: {
48
+ env: 'PROVER_NODE_TX_GATHERING_MAX_PARALLEL_REQUESTS_PER_NODE',
49
+ description: 'How many tx requests to make in parallel to each node',
39
50
  ...numberConfigHelper(100)
51
+ },
52
+ txGatheringTimeoutMs: {
53
+ env: 'PROVER_NODE_TX_GATHERING_TIMEOUT_MS',
54
+ description: 'How long to wait for tx data to be available before giving up',
55
+ ...numberConfigHelper(120_000)
56
+ },
57
+ proverNodeDisableProofPublish: {
58
+ env: 'PROVER_NODE_DISABLE_PROOF_PUBLISH',
59
+ description: 'Whether the prover node skips publishing proofs to L1',
60
+ ...booleanConfigHelper(false)
40
61
  }
41
62
  };
42
63
  export const proverNodeConfigMappings = {
43
64
  ...dataConfigMappings,
65
+ ...keyStoreConfigMappings,
44
66
  ...archiverConfigMappings,
45
67
  ...proverClientConfigMappings,
46
68
  ...p2pConfigMappings,
47
69
  ...worldStateConfigMappings,
48
70
  ...getPublisherConfigMappings('PROVER'),
49
71
  ...getTxSenderConfigMappings('PROVER'),
50
- ...proverCoordinationConfigMappings,
51
72
  ...specificProverNodeConfigMappings,
52
- testAccounts: {
53
- env: 'TEST_ACCOUNTS',
54
- description: 'Whether to populate the genesis state with initial fee juice for the test accounts.',
55
- ...booleanConfigHelper(false)
56
- }
73
+ ...genesisStateConfigMappings,
74
+ ...sharedNodeConfigMappings
57
75
  };
58
76
  export function getProverNodeConfigFromEnv() {
59
77
  return getConfigFromMappings(proverNodeConfigMappings);
@@ -69,3 +87,52 @@ export function getProverNodeAgentConfigFromEnv() {
69
87
  ...getConfigFromMappings(bbConfigMappings)
70
88
  };
71
89
  }
90
+ function createKeyStoreFromWeb3Signer(config) {
91
+ // If we don't have a valid prover Id then we can't build a valid key store with remote signers
92
+ if (config.proverId === undefined) {
93
+ return undefined;
94
+ }
95
+ // Also, we need at least one publisher address.
96
+ const publishers = config.publisherAddresses ?? [];
97
+ if (publishers.length === 0) {
98
+ return undefined;
99
+ }
100
+ const keyStore = {
101
+ schemaVersion: 1,
102
+ slasher: undefined,
103
+ prover: {
104
+ id: config.proverId,
105
+ publisher: publishers
106
+ },
107
+ remoteSigner: config.web3SignerUrl,
108
+ validators: undefined
109
+ };
110
+ return keyStore;
111
+ }
112
+ function createKeyStoreFromPublisherKeys(config) {
113
+ // Extract the publisher keys from the provided config.
114
+ const publisherKeys = config.publisherPrivateKeys ? config.publisherPrivateKeys.map((k)=>ethPrivateKeySchema.parse(k.getValue())) : [];
115
+ // There must be at least 1.
116
+ if (publisherKeys.length === 0) {
117
+ return undefined;
118
+ }
119
+ // If we have a valid proverId then create a prover key store of the form { id, publisher: [publisherKeys] }
120
+ // Otherwise create one of the form ("0x12345678....." as EthAccount).
121
+ const keyStore = {
122
+ schemaVersion: 1,
123
+ slasher: undefined,
124
+ prover: config.proverId === undefined ? publisherKeys[0] : {
125
+ id: config.proverId,
126
+ publisher: publisherKeys
127
+ },
128
+ remoteSigner: undefined,
129
+ validators: undefined
130
+ };
131
+ return keyStore;
132
+ }
133
+ export function createKeyStoreForProver(config) {
134
+ if (config.web3SignerUrl !== undefined && config.web3SignerUrl.length > 0) {
135
+ return createKeyStoreFromWeb3Signer(config);
136
+ }
137
+ return createKeyStoreFromPublisherKeys(config);
138
+ }
package/dest/factory.d.ts CHANGED
@@ -2,24 +2,28 @@ import { type Archiver } from '@aztec/archiver';
2
2
  import { type BlobSinkClientInterface } from '@aztec/blob-sink/client';
3
3
  import { L1TxUtils } from '@aztec/ethereum';
4
4
  import { type Logger } from '@aztec/foundation/log';
5
+ import { DateProvider } from '@aztec/foundation/timer';
5
6
  import type { DataStoreConfig } from '@aztec/kv-store/config';
6
- import type { ProverCoordination, ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
7
+ import { type KeyStoreConfig } from '@aztec/node-keystore';
8
+ import type { AztecNode, ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
7
9
  import type { PublicDataTreeLeaf } from '@aztec/stdlib/trees';
8
10
  import { type TelemetryClient } from '@aztec/telemetry-client';
9
- import type { ProverNodeConfig } from './config.js';
10
- import { ProverNodePublisher } from './prover-node-publisher.js';
11
+ import { type ProverNodeConfig } from './config.js';
11
12
  import { ProverNode } from './prover-node.js';
12
- /** Creates a new prover node given a config. */
13
- export declare function createProverNode(config: ProverNodeConfig & DataStoreConfig, deps?: {
13
+ import { ProverPublisherFactory } from './prover-publisher-factory.js';
14
+ export type ProverNodeDeps = {
14
15
  telemetry?: TelemetryClient;
15
16
  log?: Logger;
16
- aztecNodeTxProvider?: ProverCoordination;
17
+ aztecNodeTxProvider?: Pick<AztecNode, 'getTxsByHash'>;
17
18
  archiver?: Archiver;
18
- publisher?: ProverNodePublisher;
19
+ publisherFactory?: ProverPublisherFactory;
19
20
  blobSinkClient?: BlobSinkClientInterface;
20
21
  broker?: ProvingJobBroker;
21
22
  l1TxUtils?: L1TxUtils;
22
- }, options?: {
23
+ dateProvider?: DateProvider;
24
+ };
25
+ /** Creates a new prover node given a config. */
26
+ export declare function createProverNode(userConfig: ProverNodeConfig & DataStoreConfig & KeyStoreConfig, deps?: ProverNodeDeps, options?: {
23
27
  prefilledPublicData?: PublicDataTreeLeaf[];
24
28
  }): Promise<ProverNode>;
25
29
  //# sourceMappingURL=factory.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"factory.d.ts","sourceRoot":"","sources":["../src/factory.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,KAAK,QAAQ,EAAkB,MAAM,iBAAiB,CAAC;AAChE,OAAO,EAAE,KAAK,uBAAuB,EAAwB,MAAM,yBAAyB,CAAC;AAE7F,OAAO,EAAE,SAAS,EAAwD,MAAM,iBAAiB,CAAC;AAClG,OAAO,EAAE,KAAK,MAAM,EAAgB,MAAM,uBAAuB,CAAC;AAClE,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,wBAAwB,CAAC;AAG9D,OAAO,KAAK,EAAE,kBAAkB,EAAE,gBAAgB,EAAE,MAAM,iCAAiC,CAAC;AAC5F,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,qBAAqB,CAAC;AAC9D,OAAO,EAAE,KAAK,eAAe,EAAsB,MAAM,yBAAyB,CAAC;AAGnF,OAAO,KAAK,EAAE,gBAAgB,EAAE,MAAM,aAAa,CAAC;AAGpD,OAAO,EAAE,mBAAmB,EAAE,MAAM,4BAA4B,CAAC;AACjE,OAAO,EAAE,UAAU,EAA0B,MAAM,kBAAkB,CAAC;AAEtE,gDAAgD;AAChD,wBAAsB,gBAAgB,CACpC,MAAM,EAAE,gBAAgB,GAAG,eAAe,EAC1C,IAAI,GAAE;IACJ,SAAS,CAAC,EAAE,eAAe,CAAC;IAC5B,GAAG,CAAC,EAAE,MAAM,CAAC;IACb,mBAAmB,CAAC,EAAE,kBAAkB,CAAC;IACzC,QAAQ,CAAC,EAAE,QAAQ,CAAC;IACpB,SAAS,CAAC,EAAE,mBAAmB,CAAC;IAChC,cAAc,CAAC,EAAE,uBAAuB,CAAC;IACzC,MAAM,CAAC,EAAE,gBAAgB,CAAC;IAC1B,SAAS,CAAC,EAAE,SAAS,CAAC;CAClB,EACN,OAAO,GAAE;IACP,mBAAmB,CAAC,EAAE,kBAAkB,EAAE,CAAC;CACvC,uBAgEP"}
1
+ {"version":3,"file":"factory.d.ts","sourceRoot":"","sources":["../src/factory.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,KAAK,QAAQ,EAAkB,MAAM,iBAAiB,CAAC;AAEhE,OAAO,EAAE,KAAK,uBAAuB,EAAwB,MAAM,yBAAyB,CAAC;AAE7F,OAAO,EAAE,SAAS,EAAyD,MAAM,iBAAiB,CAAC;AAEnG,OAAO,EAAE,KAAK,MAAM,EAAgB,MAAM,uBAAuB,CAAC;AAClE,OAAO,EAAE,YAAY,EAAE,MAAM,yBAAyB,CAAC;AACvD,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,wBAAwB,CAAC;AAC9D,OAAO,EAAE,KAAK,cAAc,EAAkD,MAAM,sBAAsB,CAAC;AAM3G,OAAO,KAAK,EAAE,SAAS,EAAE,gBAAgB,EAAE,MAAM,iCAAiC,CAAC;AAEnF,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,qBAAqB,CAAC;AAE9D,OAAO,EAAa,KAAK,eAAe,EAAsB,MAAM,yBAAyB,CAAC;AAK9F,OAAO,EAAE,KAAK,gBAAgB,EAA2B,MAAM,aAAa,CAAC;AAE7E,OAAO,EAAE,UAAU,EAAE,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAE,sBAAsB,EAAE,MAAM,+BAA+B,CAAC;AAEvE,MAAM,MAAM,cAAc,GAAG;IAC3B,SAAS,CAAC,EAAE,eAAe,CAAC;IAC5B,GAAG,CAAC,EAAE,MAAM,CAAC;IACb,mBAAmB,CAAC,EAAE,IAAI,CAAC,SAAS,EAAE,cAAc,CAAC,CAAC;IACtD,QAAQ,CAAC,EAAE,QAAQ,CAAC;IACpB,gBAAgB,CAAC,EAAE,sBAAsB,CAAC;IAC1C,cAAc,CAAC,EAAE,uBAAuB,CAAC;IACzC,MAAM,CAAC,EAAE,gBAAgB,CAAC;IAC1B,SAAS,CAAC,EAAE,SAAS,CAAC;IACtB,YAAY,CAAC,EAAE,YAAY,CAAC;CAC7B,CAAC;AAEF,gDAAgD;AAChD,wBAAsB,gBAAgB,CACpC,UAAU,EAAE,gBAAgB,GAAG,eAAe,GAAG,cAAc,EAC/D,IAAI,GAAE,cAAmB,EACzB,OAAO,GAAE;IACP,mBAAmB,CAAC,EAAE,kBAAkB,EAAE,CAAC;CACvC,uBAyKP"}