@aztec/prover-node 0.0.1-commit.b655e406 → 0.0.1-commit.c0b82b2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (54) hide show
  1. package/dest/actions/download-epoch-proving-job.d.ts +4 -4
  2. package/dest/actions/index.d.ts +1 -1
  3. package/dest/actions/rerun-epoch-proving-job.d.ts +3 -2
  4. package/dest/actions/rerun-epoch-proving-job.d.ts.map +1 -1
  5. package/dest/actions/rerun-epoch-proving-job.js +5 -3
  6. package/dest/actions/upload-epoch-proof-failure.d.ts +1 -1
  7. package/dest/bin/run-failed-epoch.d.ts +1 -1
  8. package/dest/bin/run-failed-epoch.js +5 -2
  9. package/dest/config.d.ts +7 -9
  10. package/dest/config.d.ts.map +1 -1
  11. package/dest/config.js +17 -19
  12. package/dest/factory.d.ts +20 -15
  13. package/dest/factory.d.ts.map +1 -1
  14. package/dest/factory.js +28 -61
  15. package/dest/index.d.ts +2 -1
  16. package/dest/index.d.ts.map +1 -1
  17. package/dest/index.js +1 -0
  18. package/dest/job/epoch-proving-job-data.d.ts +8 -6
  19. package/dest/job/epoch-proving-job-data.d.ts.map +1 -1
  20. package/dest/job/epoch-proving-job-data.js +25 -18
  21. package/dest/job/epoch-proving-job.d.ts +7 -13
  22. package/dest/job/epoch-proving-job.d.ts.map +1 -1
  23. package/dest/job/epoch-proving-job.js +503 -105
  24. package/dest/metrics.d.ts +4 -3
  25. package/dest/metrics.d.ts.map +1 -1
  26. package/dest/metrics.js +30 -98
  27. package/dest/monitors/epoch-monitor.d.ts +3 -2
  28. package/dest/monitors/epoch-monitor.d.ts.map +1 -1
  29. package/dest/monitors/epoch-monitor.js +3 -11
  30. package/dest/monitors/index.d.ts +1 -1
  31. package/dest/prover-node-publisher.d.ts +14 -11
  32. package/dest/prover-node-publisher.d.ts.map +1 -1
  33. package/dest/prover-node-publisher.js +49 -42
  34. package/dest/prover-node.d.ts +26 -15
  35. package/dest/prover-node.d.ts.map +1 -1
  36. package/dest/prover-node.js +441 -57
  37. package/dest/prover-publisher-factory.d.ts +9 -5
  38. package/dest/prover-publisher-factory.d.ts.map +1 -1
  39. package/dest/prover-publisher-factory.js +4 -2
  40. package/dest/test/index.d.ts +1 -1
  41. package/dest/test/index.d.ts.map +1 -1
  42. package/package.json +26 -25
  43. package/src/actions/rerun-epoch-proving-job.ts +5 -3
  44. package/src/bin/run-failed-epoch.ts +5 -2
  45. package/src/config.ts +25 -31
  46. package/src/factory.ts +65 -99
  47. package/src/index.ts +1 -0
  48. package/src/job/epoch-proving-job-data.ts +31 -25
  49. package/src/job/epoch-proving-job.ts +139 -110
  50. package/src/metrics.ts +31 -82
  51. package/src/monitors/epoch-monitor.ts +5 -11
  52. package/src/prover-node-publisher.ts +71 -58
  53. package/src/prover-node.ts +60 -50
  54. package/src/prover-publisher-factory.ts +16 -8
@@ -2,9 +2,11 @@ import { ProverNodePublisher } from './prover-node-publisher.js';
2
2
  export class ProverPublisherFactory {
3
3
  config;
4
4
  deps;
5
- constructor(config, deps){
5
+ bindings;
6
+ constructor(config, deps, bindings){
6
7
  this.config = config;
7
8
  this.deps = deps;
9
+ this.bindings = bindings;
8
10
  }
9
11
  async start() {
10
12
  await this.deps.publisherManager.loadState();
@@ -21,6 +23,6 @@ export class ProverPublisherFactory {
21
23
  rollupContract: this.deps.rollupContract,
22
24
  l1TxUtils: l1Publisher,
23
25
  telemetry: this.deps.telemetry
24
- });
26
+ }, this.bindings);
25
27
  }
26
28
  }
@@ -9,4 +9,4 @@ declare abstract class TestProverNodeClass extends ProverNode {
9
9
  }
10
10
  export type TestProverNode = TestProverNodeClass;
11
11
  export {};
12
- //# sourceMappingURL=index.d.ts.map
12
+ //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiaW5kZXguZC50cyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uL3NyYy90ZXN0L2luZGV4LnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBLE9BQU8sS0FBSyxFQUFFLGtCQUFrQixFQUFFLE1BQU0saUNBQWlDLENBQUM7QUFFMUUsT0FBTyxLQUFLLEVBQUUsZUFBZSxFQUFFLE1BQU0sNkJBQTZCLENBQUM7QUFDbkUsT0FBTyxLQUFLLEVBQUUsbUJBQW1CLEVBQUUsTUFBTSw2QkFBNkIsQ0FBQztBQUN2RSxPQUFPLEVBQUUsVUFBVSxFQUFFLE1BQU0sbUJBQW1CLENBQUM7QUFFL0MsdUJBQWUsbUJBQW9CLFNBQVEsVUFBVTtJQUNwQyxNQUFNLEVBQUUsa0JBQWtCLENBQUM7SUFDM0IsU0FBUyxFQUFFLG1CQUFtQixDQUFDO0lBRTlDLFNBQXlCLHFCQUFxQixDQUFDLEdBQUcsRUFBRSxlQUFlLEdBQUcsT0FBTyxDQUFDLE1BQU0sR0FBRyxTQUFTLENBQUMsQ0FBQztDQUNuRztBQUVELE1BQU0sTUFBTSxjQUFjLEdBQUcsbUJBQW1CLENBQUMifQ==
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/test/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,iCAAiC,CAAC;AAE1E,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,6BAA6B,CAAC;AACnE,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,6BAA6B,CAAC;AACvE,OAAO,EAAE,UAAU,EAAE,MAAM,mBAAmB,CAAC;AAE/C,uBAAe,mBAAoB,SAAQ,UAAU;IACpC,MAAM,EAAE,kBAAkB,CAAC;IAC3B,SAAS,EAAE,mBAAmB,CAAC;aAErB,qBAAqB,CAAC,GAAG,EAAE,eAAe,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC;CAClG;AAED,MAAM,MAAM,cAAc,GAAG,mBAAmB,CAAC"}
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/test/index.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,kBAAkB,EAAE,MAAM,iCAAiC,CAAC;AAE1E,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,6BAA6B,CAAC;AACnE,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,6BAA6B,CAAC;AACvE,OAAO,EAAE,UAAU,EAAE,MAAM,mBAAmB,CAAC;AAE/C,uBAAe,mBAAoB,SAAQ,UAAU;IACpC,MAAM,EAAE,kBAAkB,CAAC;IAC3B,SAAS,EAAE,mBAAmB,CAAC;IAE9C,SAAyB,qBAAqB,CAAC,GAAG,EAAE,eAAe,GAAG,OAAO,CAAC,MAAM,GAAG,SAAS,CAAC,CAAC;CACnG;AAED,MAAM,MAAM,cAAc,GAAG,mBAAmB,CAAC"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@aztec/prover-node",
3
- "version": "0.0.1-commit.b655e406",
3
+ "version": "0.0.1-commit.c0b82b2",
4
4
  "type": "module",
5
5
  "exports": {
6
6
  ".": "./dest/index.js",
@@ -11,8 +11,8 @@
11
11
  "../package.common.json"
12
12
  ],
13
13
  "scripts": {
14
- "build": "yarn clean && tsc -b",
15
- "build:dev": "tsc -b --watch",
14
+ "build": "yarn clean && ../scripts/tsc.sh",
15
+ "build:dev": "../scripts/tsc.sh --watch",
16
16
  "clean": "rm -rf ./dest .tsbuildinfo",
17
17
  "bb": "node --no-warnings ./dest/bb/index.js",
18
18
  "test": "NODE_NO_WARNINGS=1 node --experimental-vm-modules ../node_modules/.bin/jest --passWithNoTests --maxWorkers=${JEST_MAX_WORKERS:-8}",
@@ -56,36 +56,37 @@
56
56
  ]
57
57
  },
58
58
  "dependencies": {
59
- "@aztec/archiver": "0.0.1-commit.b655e406",
60
- "@aztec/bb-prover": "0.0.1-commit.b655e406",
61
- "@aztec/blob-lib": "0.0.1-commit.b655e406",
62
- "@aztec/blob-sink": "0.0.1-commit.b655e406",
63
- "@aztec/constants": "0.0.1-commit.b655e406",
64
- "@aztec/epoch-cache": "0.0.1-commit.b655e406",
65
- "@aztec/ethereum": "0.0.1-commit.b655e406",
66
- "@aztec/foundation": "0.0.1-commit.b655e406",
67
- "@aztec/kv-store": "0.0.1-commit.b655e406",
68
- "@aztec/l1-artifacts": "0.0.1-commit.b655e406",
69
- "@aztec/node-keystore": "0.0.1-commit.b655e406",
70
- "@aztec/node-lib": "0.0.1-commit.b655e406",
71
- "@aztec/noir-protocol-circuits-types": "0.0.1-commit.b655e406",
72
- "@aztec/p2p": "0.0.1-commit.b655e406",
73
- "@aztec/protocol-contracts": "0.0.1-commit.b655e406",
74
- "@aztec/prover-client": "0.0.1-commit.b655e406",
75
- "@aztec/sequencer-client": "0.0.1-commit.b655e406",
76
- "@aztec/simulator": "0.0.1-commit.b655e406",
77
- "@aztec/stdlib": "0.0.1-commit.b655e406",
78
- "@aztec/telemetry-client": "0.0.1-commit.b655e406",
79
- "@aztec/world-state": "0.0.1-commit.b655e406",
59
+ "@aztec/archiver": "0.0.1-commit.c0b82b2",
60
+ "@aztec/bb-prover": "0.0.1-commit.c0b82b2",
61
+ "@aztec/blob-client": "0.0.1-commit.c0b82b2",
62
+ "@aztec/blob-lib": "0.0.1-commit.c0b82b2",
63
+ "@aztec/constants": "0.0.1-commit.c0b82b2",
64
+ "@aztec/epoch-cache": "0.0.1-commit.c0b82b2",
65
+ "@aztec/ethereum": "0.0.1-commit.c0b82b2",
66
+ "@aztec/foundation": "0.0.1-commit.c0b82b2",
67
+ "@aztec/kv-store": "0.0.1-commit.c0b82b2",
68
+ "@aztec/l1-artifacts": "0.0.1-commit.c0b82b2",
69
+ "@aztec/node-keystore": "0.0.1-commit.c0b82b2",
70
+ "@aztec/node-lib": "0.0.1-commit.c0b82b2",
71
+ "@aztec/noir-protocol-circuits-types": "0.0.1-commit.c0b82b2",
72
+ "@aztec/p2p": "0.0.1-commit.c0b82b2",
73
+ "@aztec/protocol-contracts": "0.0.1-commit.c0b82b2",
74
+ "@aztec/prover-client": "0.0.1-commit.c0b82b2",
75
+ "@aztec/sequencer-client": "0.0.1-commit.c0b82b2",
76
+ "@aztec/simulator": "0.0.1-commit.c0b82b2",
77
+ "@aztec/stdlib": "0.0.1-commit.c0b82b2",
78
+ "@aztec/telemetry-client": "0.0.1-commit.c0b82b2",
79
+ "@aztec/world-state": "0.0.1-commit.c0b82b2",
80
80
  "source-map-support": "^0.5.21",
81
81
  "tslib": "^2.4.0",
82
- "viem": "npm:@spalladino/viem@2.38.2-eip7594.0"
82
+ "viem": "npm:@aztec/viem@2.38.2"
83
83
  },
84
84
  "devDependencies": {
85
85
  "@jest/globals": "^30.0.0",
86
86
  "@types/jest": "^30.0.0",
87
87
  "@types/node": "^22.15.17",
88
88
  "@types/source-map-support": "^0.5.10",
89
+ "@typescript/native-preview": "7.0.0-dev.20260113.1",
89
90
  "jest": "^30.0.0",
90
91
  "jest-mock-extended": "^4.0.0",
91
92
  "ts-node": "^10.9.1",
@@ -1,4 +1,5 @@
1
1
  import { createArchiverStore } from '@aztec/archiver';
2
+ import type { L1ContractsConfig } from '@aztec/ethereum/config';
2
3
  import type { Logger } from '@aztec/foundation/log';
3
4
  import type { DataStoreConfig } from '@aztec/kv-store/config';
4
5
  import { type ProverClientConfig, createProverClient } from '@aztec/prover-client';
@@ -21,7 +22,7 @@ import { ProverNodeJobMetrics } from '../metrics.js';
21
22
  export async function rerunEpochProvingJob(
22
23
  localPath: string,
23
24
  log: Logger,
24
- config: DataStoreConfig & ProverBrokerConfig & ProverClientConfig,
25
+ config: DataStoreConfig & ProverBrokerConfig & ProverClientConfig & Pick<L1ContractsConfig, 'aztecEpochDuration'>,
25
26
  ) {
26
27
  const jobData = deserializeEpochProvingJobData(readFileSync(localPath));
27
28
  log.info(`Loaded proving job data for epoch ${jobData.epochNumber}`);
@@ -29,8 +30,8 @@ export async function rerunEpochProvingJob(
29
30
  const telemetry = getTelemetryClient();
30
31
  const metrics = new ProverNodeJobMetrics(telemetry.getMeter('prover-job'), telemetry.getTracer('prover-job'));
31
32
  const worldState = await createWorldState(config);
32
- const archiver = await createArchiverStore(config);
33
- const publicProcessorFactory = new PublicProcessorFactory(archiver);
33
+ const archiver = await createArchiverStore(config, { epochDuration: config.aztecEpochDuration });
34
+ const publicProcessorFactory = new PublicProcessorFactory(archiver, undefined, undefined, log.getBindings());
34
35
 
35
36
  const publisher = { submitEpochProof: () => Promise.resolve(true) };
36
37
  const l2BlockSourceForReorgDetection = undefined;
@@ -52,6 +53,7 @@ export async function rerunEpochProvingJob(
52
53
  metrics,
53
54
  deadline,
54
55
  { skipEpochCheck: true },
56
+ log.getBindings(),
55
57
  );
56
58
 
57
59
  log.info(`Rerunning epoch proving job for epoch ${jobData.epochNumber}`);
@@ -1,5 +1,6 @@
1
1
  /* eslint-disable no-console */
2
- import type { L1ContractAddresses } from '@aztec/ethereum';
2
+ import { getL1ContractsConfigEnvVars } from '@aztec/ethereum/config';
3
+ import type { L1ContractAddresses } from '@aztec/ethereum/l1-contract-addresses';
3
4
  import { EthAddress } from '@aztec/foundation/eth-address';
4
5
  import { jsonParseWithSchema, jsonStringify } from '@aztec/foundation/json-rpc';
5
6
  import { createLogger } from '@aztec/foundation/log';
@@ -22,8 +23,10 @@ async function rerunFailedEpoch(provingJobUrl: string, baseLocalDir: string) {
22
23
  const dataDir = join(localDir, 'state');
23
24
 
24
25
  const env = getProverNodeConfigFromEnv();
26
+ const l1Config = getL1ContractsConfigEnvVars();
25
27
  const config = {
26
- ...getProverNodeConfigFromEnv(),
28
+ ...env,
29
+ ...l1Config,
27
30
  dataDirectory: dataDir,
28
31
  dataStoreMapSizeKb: env.dataStoreMapSizeKb ?? 1024 * 1024,
29
32
  proverId: env.proverId ?? EthAddress.random(),
package/src/config.ts CHANGED
@@ -1,42 +1,37 @@
1
- import { type ArchiverConfig, archiverConfigMappings } from '@aztec/archiver/config';
2
1
  import type { ACVMConfig, BBConfig } from '@aztec/bb-prover/config';
3
- import { type GenesisStateConfig, genesisStateConfigMappings } from '@aztec/ethereum';
4
2
  import {
5
3
  type ConfigMappingsType,
6
4
  booleanConfigHelper,
7
5
  getConfigFromMappings,
8
6
  numberConfigHelper,
7
+ pickConfigMappings,
9
8
  } from '@aztec/foundation/config';
10
9
  import { type DataStoreConfig, dataConfigMappings } from '@aztec/kv-store/config';
11
- import { type KeyStore, type KeyStoreConfig, ethPrivateKeySchema, keyStoreConfigMappings } from '@aztec/node-keystore';
10
+ import { type KeyStoreConfig, keyStoreConfigMappings } from '@aztec/node-keystore/config';
11
+ import { ethPrivateKeySchema } from '@aztec/node-keystore/schemas';
12
+ import type { KeyStore } from '@aztec/node-keystore/types';
12
13
  import { type SharedNodeConfig, sharedNodeConfigMappings } from '@aztec/node-lib/config';
13
- import { type P2PConfig, p2pConfigMappings } from '@aztec/p2p/config';
14
14
  import {
15
15
  type ProverAgentConfig,
16
16
  type ProverBrokerConfig,
17
17
  proverAgentConfigMappings,
18
18
  proverBrokerConfigMappings,
19
- } from '@aztec/prover-client/broker';
19
+ } from '@aztec/prover-client/broker/config';
20
20
  import { type ProverClientUserConfig, bbConfigMappings, proverClientConfigMappings } from '@aztec/prover-client/config';
21
21
  import {
22
- type PublisherConfig,
23
- type TxSenderConfig,
24
- getPublisherConfigMappings,
25
- getTxSenderConfigMappings,
22
+ type ProverPublisherConfig,
23
+ type ProverTxSenderConfig,
24
+ proverPublisherConfigMappings,
25
+ proverTxSenderConfigMappings,
26
26
  } from '@aztec/sequencer-client/config';
27
- import { type WorldStateConfig, worldStateConfigMappings } from '@aztec/world-state/config';
28
-
29
- export type ProverNodeConfig = ArchiverConfig &
30
- ProverClientUserConfig &
31
- P2PConfig &
32
- WorldStateConfig &
33
- PublisherConfig &
34
- TxSenderConfig &
27
+
28
+ export type ProverNodeConfig = ProverClientUserConfig &
29
+ ProverPublisherConfig &
30
+ ProverTxSenderConfig &
35
31
  DataStoreConfig &
36
32
  KeyStoreConfig &
37
- SharedNodeConfig &
38
33
  SpecificProverNodeConfig &
39
- GenesisStateConfig;
34
+ Pick<SharedNodeConfig, 'web3SignerUrl'>;
40
35
 
41
36
  export type SpecificProverNodeConfig = {
42
37
  proverNodeMaxPendingJobs: number;
@@ -51,7 +46,7 @@ export type SpecificProverNodeConfig = {
51
46
  txGatheringMaxParallelRequestsPerNode: number;
52
47
  };
53
48
 
54
- const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeConfig> = {
49
+ export const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeConfig> = {
55
50
  proverNodeMaxPendingJobs: {
56
51
  env: 'PROVER_NODE_MAX_PENDING_JOBS',
57
52
  description: 'The maximum number of pending jobs for the prover node',
@@ -106,15 +101,11 @@ const specificProverNodeConfigMappings: ConfigMappingsType<SpecificProverNodeCon
106
101
  export const proverNodeConfigMappings: ConfigMappingsType<ProverNodeConfig> = {
107
102
  ...dataConfigMappings,
108
103
  ...keyStoreConfigMappings,
109
- ...archiverConfigMappings,
110
104
  ...proverClientConfigMappings,
111
- ...p2pConfigMappings,
112
- ...worldStateConfigMappings,
113
- ...getPublisherConfigMappings('PROVER'),
114
- ...getTxSenderConfigMappings('PROVER'),
105
+ ...proverPublisherConfigMappings,
106
+ ...proverTxSenderConfigMappings,
115
107
  ...specificProverNodeConfigMappings,
116
- ...genesisStateConfigMappings,
117
- ...sharedNodeConfigMappings,
108
+ ...pickConfigMappings(sharedNodeConfigMappings, ['web3SignerUrl']),
118
109
  };
119
110
 
120
111
  export function getProverNodeConfigFromEnv(): ProverNodeConfig {
@@ -141,7 +132,7 @@ function createKeyStoreFromWeb3Signer(config: ProverNodeConfig): KeyStore | unde
141
132
  }
142
133
 
143
134
  // Also, we need at least one publisher address.
144
- const publishers = config.publisherAddresses ?? [];
135
+ const publishers = config.proverPublisherAddresses ?? [];
145
136
 
146
137
  if (publishers.length === 0) {
147
138
  return undefined;
@@ -162,8 +153,8 @@ function createKeyStoreFromWeb3Signer(config: ProverNodeConfig): KeyStore | unde
162
153
 
163
154
  function createKeyStoreFromPublisherKeys(config: ProverNodeConfig): KeyStore | undefined {
164
155
  // Extract the publisher keys from the provided config.
165
- const publisherKeys = config.publisherPrivateKeys
166
- ? config.publisherPrivateKeys.map(k => ethPrivateKeySchema.parse(k.getValue()))
156
+ const publisherKeys = config.proverPublisherPrivateKeys
157
+ ? config.proverPublisherPrivateKeys.map((k: { getValue: () => string }) => ethPrivateKeySchema.parse(k.getValue()))
167
158
  : [];
168
159
 
169
160
  // There must be at least 1.
@@ -192,7 +183,10 @@ function createKeyStoreFromPublisherKeys(config: ProverNodeConfig): KeyStore | u
192
183
 
193
184
  export function createKeyStoreForProver(config: ProverNodeConfig): KeyStore | undefined {
194
185
  if (config.web3SignerUrl !== undefined && config.web3SignerUrl.length > 0) {
195
- return createKeyStoreFromWeb3Signer(config);
186
+ const keyStore = createKeyStoreFromWeb3Signer(config);
187
+ if (keyStore) {
188
+ return keyStore;
189
+ }
196
190
  }
197
191
 
198
192
  return createKeyStoreFromPublisherKeys(config);
package/src/factory.ts CHANGED
@@ -1,28 +1,36 @@
1
- import { type Archiver, createArchiver } from '@aztec/archiver';
2
- import { BBCircuitVerifier, QueuedIVCVerifier, TestCircuitVerifier } from '@aztec/bb-prover';
3
- import { type BlobSinkClientInterface, createBlobSinkClient } from '@aztec/blob-sink/client';
4
- import { EpochCache } from '@aztec/epoch-cache';
5
- import { L1TxUtils, PublisherManager, RollupContract, createEthereumChain } from '@aztec/ethereum';
1
+ import type { Archiver } from '@aztec/archiver';
2
+ import type { BlobClientInterface } from '@aztec/blob-client/client';
3
+ import { Blob } from '@aztec/blob-lib';
4
+ import type { EpochCacheInterface } from '@aztec/epoch-cache';
5
+ import { createEthereumChain } from '@aztec/ethereum/chain';
6
+ import { RollupContract } from '@aztec/ethereum/contracts';
7
+ import { L1TxUtils } from '@aztec/ethereum/l1-tx-utils';
8
+ import { PublisherManager } from '@aztec/ethereum/publisher-manager';
6
9
  import { pick } from '@aztec/foundation/collection';
7
10
  import { type Logger, createLogger } from '@aztec/foundation/log';
8
11
  import { DateProvider } from '@aztec/foundation/timer';
9
- import type { DataStoreConfig } from '@aztec/kv-store/config';
10
- import { type KeyStoreConfig, KeystoreManager, loadKeystores, mergeKeystores } from '@aztec/node-keystore';
11
- import { trySnapshotSync } from '@aztec/node-lib/actions';
12
- import { createL1TxUtilsFromEthSignerWithStore } from '@aztec/node-lib/factories';
13
- import { NodeRpcTxSource, createP2PClient } from '@aztec/p2p';
14
- import { type ProverClientConfig, createProverClient } from '@aztec/prover-client';
12
+ import { KeystoreManager } from '@aztec/node-keystore';
13
+ import { createForwarderL1TxUtilsFromSigners, createL1TxUtilsFromSigners } from '@aztec/node-lib/factories';
14
+ import { type ProverClientConfig, type ProverClientUserConfig, createProverClient } from '@aztec/prover-client';
15
15
  import { createAndStartProvingBroker } from '@aztec/prover-client/broker';
16
- import type { AztecNode, ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
17
- import { P2PClientType } from '@aztec/stdlib/p2p';
18
- import type { PublicDataTreeLeaf } from '@aztec/stdlib/trees';
19
- import { getPackageVersion } from '@aztec/stdlib/update-checker';
16
+ import {
17
+ type ProverPublisherConfig,
18
+ type ProverTxSenderConfig,
19
+ getPublisherConfigFromProverConfig,
20
+ } from '@aztec/sequencer-client';
21
+ import type {
22
+ AztecNode,
23
+ ITxProvider,
24
+ ProverConfig,
25
+ ProvingJobBroker,
26
+ Service,
27
+ WorldStateSynchronizer,
28
+ } from '@aztec/stdlib/interfaces/server';
20
29
  import { L1Metrics, type TelemetryClient, getTelemetryClient } from '@aztec/telemetry-client';
21
- import { createWorldStateSynchronizer } from '@aztec/world-state';
22
30
 
23
31
  import { createPublicClient, fallback, http } from 'viem';
24
32
 
25
- import { type ProverNodeConfig, createKeyStoreForProver } from './config.js';
33
+ import type { SpecificProverNodeConfig } from './config.js';
26
34
  import { EpochMonitor } from './monitors/epoch-monitor.js';
27
35
  import { ProverNode } from './prover-node.js';
28
36
  import { ProverPublisherFactory } from './prover-publisher-factory.js';
@@ -31,55 +39,42 @@ export type ProverNodeDeps = {
31
39
  telemetry?: TelemetryClient;
32
40
  log?: Logger;
33
41
  aztecNodeTxProvider?: Pick<AztecNode, 'getTxsByHash'>;
34
- archiver?: Archiver;
42
+ archiver: Archiver;
35
43
  publisherFactory?: ProverPublisherFactory;
36
- blobSinkClient?: BlobSinkClientInterface;
37
44
  broker?: ProvingJobBroker;
38
45
  l1TxUtils?: L1TxUtils;
39
46
  dateProvider?: DateProvider;
47
+ worldStateSynchronizer: WorldStateSynchronizer;
48
+ p2pClient: { getTxProvider(): ITxProvider } & Partial<Service>;
49
+ epochCache: EpochCacheInterface;
50
+ blobClient: BlobClientInterface;
51
+ keyStoreManager?: KeystoreManager;
40
52
  };
41
53
 
42
- /** Creates a new prover node given a config. */
54
+ /** Creates a new prover node subsystem given a config and dependencies */
43
55
  export async function createProverNode(
44
- userConfig: ProverNodeConfig & DataStoreConfig & KeyStoreConfig,
45
- deps: ProverNodeDeps = {},
46
- options: {
47
- prefilledPublicData?: PublicDataTreeLeaf[];
48
- } = {},
56
+ userConfig: SpecificProverNodeConfig &
57
+ ProverConfig &
58
+ ProverClientUserConfig &
59
+ ProverPublisherConfig &
60
+ ProverTxSenderConfig,
61
+ deps: ProverNodeDeps,
49
62
  ) {
50
63
  const config = { ...userConfig };
51
64
  const telemetry = deps.telemetry ?? getTelemetryClient();
52
65
  const dateProvider = deps.dateProvider ?? new DateProvider();
53
- const blobSinkClient =
54
- deps.blobSinkClient ?? createBlobSinkClient(config, { logger: createLogger('prover-node:blob-sink:client') });
55
- const log = deps.log ?? createLogger('prover-node');
56
-
57
- // Build a key store from file if given or from environment otherwise
58
- let keyStoreManager: KeystoreManager | undefined;
59
- const keyStoreProvided = config.keyStoreDirectory !== undefined && config.keyStoreDirectory.length > 0;
60
- if (keyStoreProvided) {
61
- const keyStores = loadKeystores(config.keyStoreDirectory!);
62
- keyStoreManager = new KeystoreManager(mergeKeystores(keyStores));
63
- } else {
64
- const keyStore = createKeyStoreForProver(config);
65
- if (keyStore) {
66
- keyStoreManager = new KeystoreManager(keyStore);
67
- }
68
- }
66
+ const log = deps.log ?? createLogger('prover');
69
67
 
70
- await keyStoreManager?.validateSigners();
68
+ const { p2pClient, archiver, keyStoreManager, worldStateSynchronizer } = deps;
71
69
 
72
70
  // Extract the prover signers from the key store and verify that we have one.
71
+ await keyStoreManager?.validateSigners();
73
72
  const proverSigners = keyStoreManager?.createProverSigners();
74
73
 
75
74
  if (proverSigners === undefined) {
76
75
  throw new Error('Failed to create prover key store configuration');
77
76
  } else if (proverSigners.signers.length === 0) {
78
77
  throw new Error('No prover signers found in the key store');
79
- } else if (!keyStoreProvided) {
80
- log.warn(
81
- 'KEY STORE CREATED FROM ENVIRONMENT, IT IS RECOMMENDED TO USE A FILE-BASED KEY STORE IN PRODUCTION ENVIRONMENTS',
82
- );
83
78
  }
84
79
 
85
80
  log.info(`Creating prover with publishers ${proverSigners.signers.map(signer => signer.address.toString()).join()}`);
@@ -91,28 +86,7 @@ export async function createProverNode(
91
86
  const proverId = proverSigners.id ?? proverIdInUserConfig ?? proverSigners.signers[0].address;
92
87
 
93
88
  // Now create the prover client configuration from this.
94
- const proverClientConfig: ProverClientConfig = {
95
- ...config,
96
- proverId,
97
- };
98
-
99
- await trySnapshotSync(config, log);
100
-
101
- const epochCache = await EpochCache.create(config.l1Contracts.rollupAddress, config);
102
-
103
- const archiver =
104
- deps.archiver ??
105
- (await createArchiver(config, { blobSinkClient, epochCache, telemetry, dateProvider }, { blockUntilSync: true }));
106
- log.verbose(`Created archiver and synced to block ${await archiver.getBlockNumber()}`);
107
-
108
- const worldStateConfig = { ...config, worldStateProvenBlocksOnly: false };
109
- const worldStateSynchronizer = await createWorldStateSynchronizer(
110
- worldStateConfig,
111
- archiver,
112
- options.prefilledPublicData,
113
- telemetry,
114
- );
115
- await worldStateSynchronizer.start();
89
+ const proverClientConfig: ProverClientConfig = { ...config, proverId };
116
90
 
117
91
  const broker = deps.broker ?? (await createAndStartProvingBroker(config, telemetry));
118
92
 
@@ -123,7 +97,7 @@ export async function createProverNode(
123
97
 
124
98
  const publicClient = createPublicClient({
125
99
  chain: chain.chainInfo,
126
- transport: fallback(config.l1RpcUrls.map((url: string) => http(url))),
100
+ transport: fallback(config.l1RpcUrls.map((url: string) => http(url, { batch: false }))),
127
101
  pollingInterval: config.viemPollingIntervalMS,
128
102
  });
129
103
 
@@ -131,44 +105,31 @@ export async function createProverNode(
131
105
 
132
106
  const l1TxUtils = deps.l1TxUtils
133
107
  ? [deps.l1TxUtils]
134
- : await createL1TxUtilsFromEthSignerWithStore(
135
- publicClient,
136
- proverSigners.signers,
137
- { ...config, scope: 'prover' },
138
- { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
139
- );
108
+ : config.proverPublisherForwarderAddress
109
+ ? await createForwarderL1TxUtilsFromSigners(
110
+ publicClient,
111
+ proverSigners.signers,
112
+ config.proverPublisherForwarderAddress,
113
+ { ...config, scope: 'prover' },
114
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider, kzg: Blob.getViemKzgInstance() },
115
+ )
116
+ : await createL1TxUtilsFromSigners(
117
+ publicClient,
118
+ proverSigners.signers,
119
+ { ...config, scope: 'prover' },
120
+ { telemetry, logger: log.createChild('l1-tx-utils'), dateProvider },
121
+ );
140
122
 
141
123
  const publisherFactory =
142
124
  deps.publisherFactory ??
143
125
  new ProverPublisherFactory(config, {
144
126
  rollupContract,
145
- publisherManager: new PublisherManager(l1TxUtils, config),
127
+ publisherManager: new PublisherManager(l1TxUtils, getPublisherConfigFromProverConfig(config), log.getBindings()),
146
128
  telemetry,
147
129
  });
148
130
 
149
- const proofVerifier = new QueuedIVCVerifier(
150
- config,
151
- config.realProofs ? await BBCircuitVerifier.new(config) : new TestCircuitVerifier(),
152
- );
153
-
154
- const p2pClient = await createP2PClient(
155
- P2PClientType.Prover,
156
- config,
157
- archiver,
158
- proofVerifier,
159
- worldStateSynchronizer,
160
- epochCache,
161
- getPackageVersion() ?? '',
162
- dateProvider,
163
- telemetry,
164
- {
165
- txCollectionNodeSources: deps.aztecNodeTxProvider
166
- ? [new NodeRpcTxSource(deps.aztecNodeTxProvider, 'TestNode')]
167
- : [],
168
- },
169
- );
170
-
171
- await p2pClient.start();
131
+ // TODO(#20393): Check that the tx collection node sources are properly injected
132
+ // See aztecNodeTxProvider
172
133
 
173
134
  const proverNodeConfig = {
174
135
  ...pick(
@@ -200,6 +161,9 @@ export async function createProverNode(
200
161
  l1TxUtils.map(utils => utils.getSenderAddress()),
201
162
  );
202
163
 
164
+ // Extract the shared delayer from the first L1TxUtils instance (all instances share the same delayer)
165
+ const delayer = l1TxUtils[0]?.delayer;
166
+
203
167
  return new ProverNode(
204
168
  prover,
205
169
  publisherFactory,
@@ -213,5 +177,7 @@ export async function createProverNode(
213
177
  l1Metrics,
214
178
  proverNodeConfig,
215
179
  telemetry,
180
+ delayer,
181
+ dateProvider,
216
182
  );
217
183
  }
package/src/index.ts CHANGED
@@ -1,5 +1,6 @@
1
1
  export * from './actions/index.js';
2
2
  export * from './config.js';
3
3
  export * from './factory.js';
4
+ export * from './monitors/index.js';
4
5
  export * from './prover-node-publisher.js';
5
6
  export * from './prover-node.js';
@@ -1,49 +1,55 @@
1
- import { Fr } from '@aztec/foundation/fields';
1
+ import { CheckpointNumber, EpochNumber } from '@aztec/foundation/branded-types';
2
+ import { Fr } from '@aztec/foundation/curves/bn254';
2
3
  import { BufferReader, serializeToBuffer } from '@aztec/foundation/serialize';
3
- import { CommitteeAttestation, L2Block } from '@aztec/stdlib/block';
4
+ import { CommitteeAttestation } from '@aztec/stdlib/block';
5
+ import { Checkpoint } from '@aztec/stdlib/checkpoint';
4
6
  import { BlockHeader, Tx } from '@aztec/stdlib/tx';
5
7
 
6
8
  /** All data from an epoch used in proving. */
7
9
  export type EpochProvingJobData = {
8
- epochNumber: bigint;
9
- blocks: L2Block[];
10
+ epochNumber: EpochNumber;
11
+ checkpoints: Checkpoint[];
10
12
  txs: Map<string, Tx>;
11
- l1ToL2Messages: Record<number, Fr[]>;
13
+ l1ToL2Messages: Record<CheckpointNumber, Fr[]>;
12
14
  previousBlockHeader: BlockHeader;
13
15
  attestations: CommitteeAttestation[];
14
16
  };
15
17
 
16
18
  export function validateEpochProvingJobData(data: EpochProvingJobData) {
17
- if (data.blocks.length > 0 && data.previousBlockHeader.getBlockNumber() + 1 !== data.blocks[0].number) {
19
+ if (data.checkpoints.length === 0) {
20
+ throw new Error('No checkpoints to prove');
21
+ }
22
+
23
+ const firstBlockNumber = data.checkpoints[0].blocks[0].number;
24
+ const previousBlockNumber = data.previousBlockHeader.getBlockNumber();
25
+ if (previousBlockNumber + 1 !== firstBlockNumber) {
18
26
  throw new Error(
19
- `Initial block number ${
20
- data.blocks[0].number
21
- } does not match previous block header ${data.previousBlockHeader.getBlockNumber()}`,
27
+ `Initial block number ${firstBlockNumber} does not match previous block header ${previousBlockNumber}`,
22
28
  );
23
29
  }
24
30
 
25
- for (const blockNumber of data.blocks.map(block => block.number)) {
26
- if (!(blockNumber in data.l1ToL2Messages)) {
27
- throw new Error(`Missing L1 to L2 messages for block number ${blockNumber}`);
31
+ for (const checkpoint of data.checkpoints) {
32
+ if (!(checkpoint.number in data.l1ToL2Messages)) {
33
+ throw new Error(`Missing L1 to L2 messages for checkpoint number ${checkpoint.number}`);
28
34
  }
29
35
  }
30
36
  }
31
37
 
32
38
  export function serializeEpochProvingJobData(data: EpochProvingJobData): Buffer {
33
- const blocks = data.blocks.map(block => block.toBuffer());
39
+ const checkpoints = data.checkpoints.map(checkpoint => checkpoint.toBuffer());
34
40
  const txs = Array.from(data.txs.values()).map(tx => tx.toBuffer());
35
- const l1ToL2Messages = Object.entries(data.l1ToL2Messages).map(([blockNumber, messages]) => [
36
- Number(blockNumber),
41
+ const l1ToL2Messages = Object.entries(data.l1ToL2Messages).map(([checkpointNumber, messages]) => [
42
+ Number(checkpointNumber),
37
43
  messages.length,
38
44
  ...messages,
39
45
  ]);
40
46
  const attestations = data.attestations.map(attestation => attestation.toBuffer());
41
47
 
42
48
  return serializeToBuffer(
43
- Number(data.epochNumber),
49
+ data.epochNumber,
44
50
  data.previousBlockHeader,
45
- blocks.length,
46
- ...blocks,
51
+ checkpoints.length,
52
+ ...checkpoints,
47
53
  txs.length,
48
54
  ...txs,
49
55
  l1ToL2Messages.length,
@@ -55,22 +61,22 @@ export function serializeEpochProvingJobData(data: EpochProvingJobData): Buffer
55
61
 
56
62
  export function deserializeEpochProvingJobData(buf: Buffer): EpochProvingJobData {
57
63
  const reader = BufferReader.asReader(buf);
58
- const epochNumber = BigInt(reader.readNumber());
64
+ const epochNumber = EpochNumber(reader.readNumber());
59
65
  const previousBlockHeader = reader.readObject(BlockHeader);
60
- const blocks = reader.readVector(L2Block);
66
+ const checkpoints = reader.readVector(Checkpoint);
61
67
  const txArray = reader.readVector(Tx);
62
68
 
63
- const l1ToL2MessageBlockCount = reader.readNumber();
69
+ const l1ToL2MessageCheckpointCount = reader.readNumber();
64
70
  const l1ToL2Messages: Record<number, Fr[]> = {};
65
- for (let i = 0; i < l1ToL2MessageBlockCount; i++) {
66
- const blockNumber = reader.readNumber();
71
+ for (let i = 0; i < l1ToL2MessageCheckpointCount; i++) {
72
+ const checkpointNumber = CheckpointNumber(reader.readNumber());
67
73
  const messages = reader.readVector(Fr);
68
- l1ToL2Messages[blockNumber] = messages;
74
+ l1ToL2Messages[checkpointNumber] = messages;
69
75
  }
70
76
 
71
77
  const attestations = reader.readVector(CommitteeAttestation);
72
78
 
73
79
  const txs = new Map<string, Tx>(txArray.map(tx => [tx.getTxHash().toString(), tx]));
74
80
 
75
- return { epochNumber, previousBlockHeader, blocks, txs, l1ToL2Messages, attestations };
81
+ return { epochNumber, previousBlockHeader, checkpoints, txs, l1ToL2Messages, attestations };
76
82
  }