@aztec/aztec 0.0.1-commit.6d3c34e → 0.0.1-commit.7035c9bd6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (126) hide show
  1. package/dest/bin/index.js +7 -3
  2. package/dest/cli/admin_api_key_store.d.ts +45 -0
  3. package/dest/cli/admin_api_key_store.d.ts.map +1 -0
  4. package/dest/cli/admin_api_key_store.js +98 -0
  5. package/dest/cli/aztec_start_action.d.ts +1 -1
  6. package/dest/cli/aztec_start_action.d.ts.map +1 -1
  7. package/dest/cli/aztec_start_action.js +61 -16
  8. package/dest/cli/aztec_start_options.d.ts +1 -1
  9. package/dest/cli/aztec_start_options.d.ts.map +1 -1
  10. package/dest/cli/aztec_start_options.js +29 -11
  11. package/dest/cli/cli.d.ts +1 -1
  12. package/dest/cli/cli.d.ts.map +1 -1
  13. package/dest/cli/cli.js +3 -4
  14. package/dest/cli/cmds/compile.d.ts +4 -0
  15. package/dest/cli/cmds/compile.d.ts.map +1 -0
  16. package/dest/cli/cmds/compile.js +166 -0
  17. package/dest/cli/cmds/profile.d.ts +4 -0
  18. package/dest/cli/cmds/profile.d.ts.map +1 -0
  19. package/dest/cli/cmds/profile.js +8 -0
  20. package/dest/cli/cmds/profile_flamegraph.d.ts +4 -0
  21. package/dest/cli/cmds/profile_flamegraph.d.ts.map +1 -0
  22. package/dest/cli/cmds/profile_flamegraph.js +52 -0
  23. package/dest/cli/cmds/profile_gates.d.ts +4 -0
  24. package/dest/cli/cmds/profile_gates.d.ts.map +1 -0
  25. package/dest/cli/cmds/profile_gates.js +58 -0
  26. package/dest/cli/cmds/profile_utils.d.ts +18 -0
  27. package/dest/cli/cmds/profile_utils.d.ts.map +1 -0
  28. package/dest/cli/cmds/profile_utils.js +50 -0
  29. package/dest/cli/cmds/standby.d.ts +32 -0
  30. package/dest/cli/cmds/standby.d.ts.map +1 -0
  31. package/dest/cli/cmds/standby.js +97 -0
  32. package/dest/cli/cmds/start_archiver.d.ts +2 -2
  33. package/dest/cli/cmds/start_archiver.d.ts.map +1 -1
  34. package/dest/cli/cmds/start_archiver.js +6 -8
  35. package/dest/cli/cmds/start_bot.d.ts +3 -3
  36. package/dest/cli/cmds/start_bot.d.ts.map +1 -1
  37. package/dest/cli/cmds/start_bot.js +9 -5
  38. package/dest/cli/cmds/start_node.d.ts +3 -2
  39. package/dest/cli/cmds/start_node.d.ts.map +1 -1
  40. package/dest/cli/cmds/start_node.js +83 -27
  41. package/dest/cli/cmds/start_p2p_bootstrap.d.ts +2 -2
  42. package/dest/cli/cmds/start_p2p_bootstrap.d.ts.map +1 -1
  43. package/dest/cli/cmds/start_p2p_bootstrap.js +1 -2
  44. package/dest/cli/cmds/start_prover_agent.js +2 -2
  45. package/dest/cli/cmds/start_prover_broker.d.ts +1 -1
  46. package/dest/cli/cmds/start_prover_broker.d.ts.map +1 -1
  47. package/dest/cli/cmds/start_prover_broker.js +14 -8
  48. package/dest/cli/cmds/utils/artifacts.d.ts +21 -0
  49. package/dest/cli/cmds/utils/artifacts.d.ts.map +1 -0
  50. package/dest/cli/cmds/utils/artifacts.js +24 -0
  51. package/dest/cli/cmds/utils/needs_recompile.d.ts +10 -0
  52. package/dest/cli/cmds/utils/needs_recompile.d.ts.map +1 -0
  53. package/dest/cli/cmds/utils/needs_recompile.js +134 -0
  54. package/dest/cli/cmds/utils/spawn.d.ts +3 -0
  55. package/dest/cli/cmds/utils/spawn.d.ts.map +1 -0
  56. package/dest/cli/cmds/utils/spawn.js +16 -0
  57. package/dest/cli/util.d.ts +7 -18
  58. package/dest/cli/util.d.ts.map +1 -1
  59. package/dest/cli/util.js +51 -87
  60. package/dest/examples/token.js +9 -9
  61. package/dest/local-network/banana_fpc.d.ts +1 -1
  62. package/dest/local-network/banana_fpc.d.ts.map +1 -1
  63. package/dest/local-network/banana_fpc.js +3 -3
  64. package/dest/local-network/local-network.d.ts +4 -3
  65. package/dest/local-network/local-network.d.ts.map +1 -1
  66. package/dest/local-network/local-network.js +52 -21
  67. package/dest/testing/anvil_test_watcher.d.ts +9 -1
  68. package/dest/testing/anvil_test_watcher.d.ts.map +1 -1
  69. package/dest/testing/anvil_test_watcher.js +52 -15
  70. package/dest/testing/cheat_codes.d.ts +3 -1
  71. package/dest/testing/cheat_codes.d.ts.map +1 -1
  72. package/dest/testing/epoch_test_settler.d.ts +4 -2
  73. package/dest/testing/epoch_test_settler.d.ts.map +1 -1
  74. package/dest/testing/epoch_test_settler.js +18 -9
  75. package/dest/testing/index.d.ts +2 -1
  76. package/dest/testing/index.d.ts.map +1 -1
  77. package/dest/testing/index.js +1 -0
  78. package/dest/testing/token_allowed_setup.d.ts +7 -0
  79. package/dest/testing/token_allowed_setup.d.ts.map +1 -0
  80. package/dest/testing/token_allowed_setup.js +20 -0
  81. package/package.json +35 -35
  82. package/scripts/add_crate.sh +102 -0
  83. package/scripts/aztec.sh +11 -5
  84. package/scripts/init.sh +23 -19
  85. package/scripts/new.sh +48 -24
  86. package/scripts/setup_workspace.sh +68 -0
  87. package/src/bin/index.ts +7 -3
  88. package/src/cli/admin_api_key_store.ts +128 -0
  89. package/src/cli/aztec_start_action.ts +60 -12
  90. package/src/cli/aztec_start_options.ts +30 -9
  91. package/src/cli/cli.ts +3 -4
  92. package/src/cli/cmds/compile.ts +191 -0
  93. package/src/cli/cmds/profile.ts +25 -0
  94. package/src/cli/cmds/profile_flamegraph.ts +64 -0
  95. package/src/cli/cmds/profile_gates.ts +68 -0
  96. package/src/cli/cmds/profile_utils.ts +58 -0
  97. package/src/cli/cmds/standby.ts +132 -0
  98. package/src/cli/cmds/start_archiver.ts +3 -14
  99. package/src/cli/cmds/start_bot.ts +8 -5
  100. package/src/cli/cmds/start_node.ts +83 -33
  101. package/src/cli/cmds/start_p2p_bootstrap.ts +2 -2
  102. package/src/cli/cmds/start_prover_agent.ts +2 -2
  103. package/src/cli/cmds/start_prover_broker.ts +19 -15
  104. package/src/cli/cmds/utils/artifacts.ts +44 -0
  105. package/src/cli/cmds/utils/needs_recompile.ts +151 -0
  106. package/src/cli/cmds/utils/spawn.ts +16 -0
  107. package/src/cli/util.ts +58 -96
  108. package/src/examples/token.ts +10 -10
  109. package/src/local-network/banana_fpc.ts +11 -7
  110. package/src/local-network/local-network.ts +62 -23
  111. package/src/testing/anvil_test_watcher.ts +59 -15
  112. package/src/testing/epoch_test_settler.ts +19 -8
  113. package/src/testing/index.ts +1 -0
  114. package/src/testing/token_allowed_setup.ts +19 -0
  115. package/dest/cli/cmds/start_prover_node.d.ts +0 -7
  116. package/dest/cli/cmds/start_prover_node.d.ts.map +0 -1
  117. package/dest/cli/cmds/start_prover_node.js +0 -108
  118. package/dest/cli/release_version.d.ts +0 -2
  119. package/dest/cli/release_version.d.ts.map +0 -1
  120. package/dest/cli/release_version.js +0 -14
  121. package/scripts/compile.sh +0 -44
  122. package/scripts/extract_function.js +0 -47
  123. package/scripts/flamegraph.sh +0 -59
  124. package/scripts/setup_project.sh +0 -31
  125. package/src/cli/cmds/start_prover_node.ts +0 -124
  126. package/src/cli/release_version.ts +0 -21
@@ -18,13 +18,17 @@ import type { LogFn } from '@aztec/foundation/log';
18
18
  import { DateProvider, TestDateProvider } from '@aztec/foundation/timer';
19
19
  import { getVKTreeRoot } from '@aztec/noir-protocol-circuits-types/vk-tree';
20
20
  import { protocolContractsHash } from '@aztec/protocol-contracts';
21
+ import { SequencerState } from '@aztec/sequencer-client';
22
+ import { AztecAddress } from '@aztec/stdlib/aztec-address';
23
+ import type { ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
21
24
  import type { PublicDataTreeLeaf } from '@aztec/stdlib/trees';
22
25
  import {
23
26
  type TelemetryClient,
24
27
  getConfigEnvVars as getTelemetryClientConfig,
25
28
  initTelemetryClient,
26
29
  } from '@aztec/telemetry-client';
27
- import { TestWallet, deployFundedSchnorrAccounts } from '@aztec/test-wallet/server';
30
+ import { EmbeddedWallet } from '@aztec/wallets/embedded';
31
+ import { deployFundedSchnorrAccounts } from '@aztec/wallets/testing';
28
32
  import { getGenesisValues } from '@aztec/world-state/testing';
29
33
 
30
34
  import { type Hex, createPublicClient, fallback, http as httpViemTransport } from 'viem';
@@ -35,6 +39,7 @@ import { createAccountLogs } from '../cli/util.js';
35
39
  import { DefaultMnemonic } from '../mnemonic.js';
36
40
  import { AnvilTestWatcher } from '../testing/anvil_test_watcher.js';
37
41
  import { EpochTestSettler } from '../testing/epoch_test_settler.js';
42
+ import { getTokenAllowedSetupFunctions } from '../testing/token_allowed_setup.js';
38
43
  import { getBananaFPCAddress, setupBananaFPC } from './banana_fpc.js';
39
44
  import { getSponsoredFPCAddress } from './sponsored_fpc.js';
40
45
 
@@ -51,7 +56,6 @@ export async function deployContractsToL1(
51
56
  aztecNodeConfig: AztecNodeConfig,
52
57
  privateKey: Hex,
53
58
  opts: {
54
- assumeProvenThroughBlockNumber?: number;
55
59
  genesisArchiveRoot?: Fr;
56
60
  feeJuicePortalInitialBalance?: bigint;
57
61
  } = {},
@@ -99,18 +103,25 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
99
103
  logger.warn(`Multiple L1 RPC URLs provided. Local networks will only use the first one: ${l1RpcUrl}`);
100
104
  }
101
105
 
106
+ // The local network deploys a banana FPC with Token contracts, so include Token entries
107
+ // in the setup allowlist so FPC-based fee payments work out of the box.
108
+ const tokenAllowList = await getTokenAllowedSetupFunctions();
109
+
102
110
  const aztecNodeConfig: AztecNodeConfig = {
103
111
  ...getConfigEnvVars(),
104
112
  ...config,
113
+ txPublicSetupAllowListExtend: [...tokenAllowList, ...(config.txPublicSetupAllowListExtend ?? [])],
105
114
  };
106
115
  const hdAccount = mnemonicToAccount(config.l1Mnemonic || DefaultMnemonic);
107
116
  if (
108
- aztecNodeConfig.publisherPrivateKeys == undefined ||
109
- !aztecNodeConfig.publisherPrivateKeys.length ||
110
- aztecNodeConfig.publisherPrivateKeys[0].getValue() === NULL_KEY
117
+ aztecNodeConfig.sequencerPublisherPrivateKeys == undefined ||
118
+ !aztecNodeConfig.sequencerPublisherPrivateKeys.length ||
119
+ aztecNodeConfig.sequencerPublisherPrivateKeys[0].getValue() === NULL_KEY
111
120
  ) {
112
121
  const privKey = hdAccount.getHdKey().privateKey;
113
- aztecNodeConfig.publisherPrivateKeys = [new SecretValue(`0x${Buffer.from(privKey!).toString('hex')}` as const)];
122
+ aztecNodeConfig.sequencerPublisherPrivateKeys = [
123
+ new SecretValue(`0x${Buffer.from(privKey!).toString('hex')}` as const),
124
+ ];
114
125
  }
115
126
  if (!aztecNodeConfig.validatorPrivateKeys?.getValue().length) {
116
127
  const privKey = hdAccount.getHdKey().privateKey;
@@ -134,9 +145,12 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
134
145
 
135
146
  const bananaFPC = await getBananaFPCAddress(initialAccounts);
136
147
  const sponsoredFPC = await getSponsoredFPCAddress();
137
- const fundedAddresses = initialAccounts.length
138
- ? [...initialAccounts.map(a => a.address), bananaFPC, sponsoredFPC]
139
- : [];
148
+ const prefundAddresses = (aztecNodeConfig.prefundAddresses ?? []).map(a => AztecAddress.fromString(a));
149
+ const fundedAddresses = [
150
+ ...initialAccounts.map(a => a.address),
151
+ ...(initialAccounts.length ? [bananaFPC, sponsoredFPC] : []),
152
+ ...prefundAddresses,
153
+ ];
140
154
  const { genesisArchiveRoot, prefilledPublicData, fundingNeeded } = await getGenesisValues(fundedAddresses);
141
155
 
142
156
  const dateProvider = new TestDateProvider();
@@ -149,7 +163,6 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
149
163
  aztecNodeConfig,
150
164
  aztecNodeConfig.validatorPrivateKeys.getValue()[0],
151
165
  {
152
- assumeProvenThroughBlockNumber: Number.MAX_SAFE_INTEGER,
153
166
  genesisArchiveRoot,
154
167
  feeJuicePortalInitialBalance: fundingNeeded,
155
168
  },
@@ -179,25 +192,42 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
179
192
  const blobClient = createBlobClient();
180
193
  const node = await createAztecNode(aztecNodeConfig, { telemetry, blobClient, dateProvider }, { prefilledPublicData });
181
194
 
195
+ // Now that the node is up, let the watcher check for pending txs so it can skip unfilled slots faster when
196
+ // transactions are waiting in the mempool. Also let it check if the sequencer is actively building, to avoid
197
+ // warping time out from under an in-progress block.
198
+ watcher?.setGetPendingTxCount(() => node.getPendingTxCount());
199
+ const sequencer = node.getSequencer()?.getSequencer();
200
+ if (sequencer) {
201
+ const idleStates: Set<string> = new Set([
202
+ SequencerState.STOPPED,
203
+ SequencerState.STOPPING,
204
+ SequencerState.IDLE,
205
+ SequencerState.SYNCHRONIZING,
206
+ ]);
207
+ watcher?.setIsSequencerBuilding(() => !idleStates.has(sequencer.getState()));
208
+ }
209
+
182
210
  let epochTestSettler: EpochTestSettler | undefined;
183
211
  if (!aztecNodeConfig.p2pEnabled) {
184
- epochTestSettler = new EpochTestSettler(cheatcodes!, rollupAddress!, node.getBlockSource(), {
185
- pollingIntervalMs: 200,
186
- });
212
+ epochTestSettler = new EpochTestSettler(
213
+ cheatcodes!,
214
+ rollupAddress!,
215
+ node.getBlockSource(),
216
+ logger.createChild('epoch-settler'),
217
+ { pollingIntervalMs: 200 },
218
+ );
187
219
  await epochTestSettler.start();
188
220
  }
189
221
 
190
222
  if (initialAccounts.length) {
191
- const PXEConfig = { proverEnabled: aztecNodeConfig.realProofs };
192
- const wallet = await TestWallet.create(node, PXEConfig);
223
+ const wallet = await EmbeddedWallet.create(node, {
224
+ pxeConfig: { proverEnabled: aztecNodeConfig.realProofs },
225
+ ephemeral: true,
226
+ });
193
227
 
194
228
  userLog('Setting up funded test accounts...');
195
- const accountManagers = await deployFundedSchnorrAccounts(wallet, node, initialAccounts);
196
- const accountsWithSecrets = accountManagers.map((manager, i) => ({
197
- account: manager,
198
- secretKey: initialAccounts[i].secret,
199
- }));
200
- const accLogs = await createAccountLogs(accountsWithSecrets, wallet);
229
+ const accountManagers = await deployFundedSchnorrAccounts(wallet, initialAccounts);
230
+ const accLogs = await createAccountLogs(accountManagers, wallet);
201
231
  userLog(accLogs.join(''));
202
232
 
203
233
  await setupBananaFPC(initialAccounts, wallet, userLog);
@@ -223,7 +253,12 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
223
253
  */
224
254
  export async function createAztecNode(
225
255
  config: Partial<AztecNodeConfig> = {},
226
- deps: { telemetry?: TelemetryClient; blobClient?: BlobClientInterface; dateProvider?: DateProvider } = {},
256
+ deps: {
257
+ telemetry?: TelemetryClient;
258
+ blobClient?: BlobClientInterface;
259
+ dateProvider?: DateProvider;
260
+ proverBroker?: ProvingJobBroker;
261
+ } = {},
227
262
  options: { prefilledPublicData?: PublicDataTreeLeaf[] } = {},
228
263
  ) {
229
264
  // TODO(#12272): will clean this up. This is criminal.
@@ -233,6 +268,10 @@ export async function createAztecNode(
233
268
  ...config,
234
269
  l1Contracts: { ...l1Contracts, ...config.l1Contracts },
235
270
  };
236
- const node = await AztecNodeService.createAndSync(aztecNodeConfig, deps, options);
271
+ const node = await AztecNodeService.createAndSync(
272
+ aztecNodeConfig,
273
+ { ...deps, proverNodeDeps: { broker: deps.proverBroker } },
274
+ options,
275
+ );
237
276
  return node;
238
277
  }
@@ -31,6 +31,15 @@ export class AnvilTestWatcher {
31
31
 
32
32
  private isMarkingAsProven = true;
33
33
 
34
+ // Optional callback to check if there are pending txs in the mempool.
35
+ private getPendingTxCount?: () => Promise<number>;
36
+
37
+ // Optional callback to check if the sequencer is actively building a block.
38
+ private isSequencerBuilding?: () => boolean;
39
+
40
+ // Tracks when we first observed the current unfilled slot with pending txs (real wall time).
41
+ private unfilledSlotFirstSeen?: { slot: number; realTime: number };
42
+
34
43
  constructor(
35
44
  private cheatcodes: EthCheatCodes,
36
45
  rollupAddress: EthAddress,
@@ -59,6 +68,16 @@ export class AnvilTestWatcher {
59
68
  this.isLocalNetwork = isLocalNetwork;
60
69
  }
61
70
 
71
+ /** Sets a callback to check for pending txs, used to skip unfilled slots faster when txs are waiting. */
72
+ setGetPendingTxCount(fn: () => Promise<number>) {
73
+ this.getPendingTxCount = fn;
74
+ }
75
+
76
+ /** Sets a callback to check if the sequencer is actively building, to avoid warping while it works. */
77
+ setIsSequencerBuilding(fn: () => boolean) {
78
+ this.isSequencerBuilding = fn;
79
+ }
80
+
62
81
  async start() {
63
82
  if (this.filledRunningPromise) {
64
83
  throw new Error('Watcher already watching for filled slot');
@@ -131,15 +150,8 @@ export class AnvilTestWatcher {
131
150
  const nextSlotTimestamp = Number(await this.rollup.read.getTimestampForSlot([BigInt(nextSlot)]));
132
151
 
133
152
  if (BigInt(currentSlot) === checkpointLog.slotNumber) {
134
- // We should jump to the next slot
135
- try {
136
- await this.cheatcodes.warp(nextSlotTimestamp, {
137
- resetBlockInterval: true,
138
- });
139
- } catch (e) {
140
- this.logger.error(`Failed to warp to timestamp ${nextSlotTimestamp}: ${e}`);
141
- }
142
-
153
+ // The current slot has been filled, we should jump to the next slot.
154
+ await this.warpToTimestamp(nextSlotTimestamp);
143
155
  this.logger.info(`Slot ${currentSlot} was filled, jumped to next slot`);
144
156
  return;
145
157
  }
@@ -149,18 +161,50 @@ export class AnvilTestWatcher {
149
161
  return;
150
162
  }
151
163
 
152
- const currentTimestamp = this.dateProvider?.now() ?? Date.now();
153
- if (currentTimestamp > nextSlotTimestamp * 1000) {
154
- try {
155
- await this.cheatcodes.warp(nextSlotTimestamp, { resetBlockInterval: true });
156
- } catch (e) {
157
- this.logger.error(`Failed to warp to timestamp ${nextSlotTimestamp}: ${e}`);
164
+ // If there are pending txs and the sequencer missed them, warp quickly (after a 2s real-time debounce) so the
165
+ // sequencer can retry in the next slot. Without this, we'd have to wait a full real-time slot duration (~36s) for
166
+ // the dateProvider to catch up to the next slot timestamp. We skip the warp if the sequencer is actively building
167
+ // to avoid invalidating its in-progress work.
168
+ if (this.getPendingTxCount) {
169
+ const pendingTxs = await this.getPendingTxCount();
170
+ if (pendingTxs > 0) {
171
+ if (this.isSequencerBuilding?.()) {
172
+ this.unfilledSlotFirstSeen = undefined;
173
+ return;
174
+ }
175
+
176
+ const realNow = Date.now();
177
+ if (!this.unfilledSlotFirstSeen || this.unfilledSlotFirstSeen.slot !== currentSlot) {
178
+ this.unfilledSlotFirstSeen = { slot: currentSlot, realTime: realNow };
179
+ return;
180
+ }
181
+
182
+ if (realNow - this.unfilledSlotFirstSeen.realTime > 2000) {
183
+ await this.warpToTimestamp(nextSlotTimestamp);
184
+ this.unfilledSlotFirstSeen = undefined;
185
+ this.logger.info(`Slot ${currentSlot} was missed with pending txs, jumped to next slot`);
186
+ }
187
+
188
+ return;
158
189
  }
190
+ }
159
191
 
192
+ // Fallback: warp when the dateProvider time has passed the next slot timestamp.
193
+ const currentTimestamp = this.dateProvider?.now() ?? Date.now();
194
+ if (currentTimestamp > nextSlotTimestamp * 1000) {
195
+ await this.warpToTimestamp(nextSlotTimestamp);
160
196
  this.logger.info(`Slot ${currentSlot} was missed, jumped to next slot`);
161
197
  }
162
198
  } catch {
163
199
  this.logger.error('mineIfSlotFilled failed');
164
200
  }
165
201
  }
202
+
203
+ private async warpToTimestamp(timestamp: number) {
204
+ try {
205
+ await this.cheatcodes.warp(timestamp, { resetBlockInterval: true });
206
+ } catch (e) {
207
+ this.logger.error(`Failed to warp to timestamp ${timestamp}: ${e}`);
208
+ }
209
+ }
166
210
  }
@@ -1,9 +1,10 @@
1
1
  import { Fr } from '@aztec/aztec.js/fields';
2
2
  import { type EthCheatCodes, RollupCheatCodes } from '@aztec/ethereum/test';
3
3
  import { type EpochNumber, SlotNumber } from '@aztec/foundation/branded-types';
4
+ import type { Logger } from '@aztec/foundation/log';
4
5
  import { EpochMonitor } from '@aztec/prover-node';
5
6
  import type { EthAddress, L2BlockSource } from '@aztec/stdlib/block';
6
- import { computeL2ToL1MembershipWitnessFromMessagesInEpoch } from '@aztec/stdlib/messaging';
7
+ import { computeEpochOutHash } from '@aztec/stdlib/messaging';
7
8
 
8
9
  export class EpochTestSettler {
9
10
  private rollupCheatCodes: RollupCheatCodes;
@@ -13,6 +14,7 @@ export class EpochTestSettler {
13
14
  cheatcodes: EthCheatCodes,
14
15
  rollupAddress: EthAddress,
15
16
  private l2BlockSource: L2BlockSource,
17
+ private log: Logger,
16
18
  private options: { pollingIntervalMs: number; provingDelayMs?: number },
17
19
  ) {
18
20
  this.rollupCheatCodes = new RollupCheatCodes(cheatcodes, { rollupAddress });
@@ -29,10 +31,16 @@ export class EpochTestSettler {
29
31
  }
30
32
 
31
33
  async handleEpochReadyToProve(epoch: EpochNumber): Promise<boolean> {
32
- const blocks = await this.l2BlockSource.getBlocksForEpoch(epoch);
34
+ const checkpointedBlocks = await this.l2BlockSource.getCheckpointedBlocksForEpoch(epoch);
35
+ const blocks = checkpointedBlocks.map(b => b.block);
36
+ this.log.info(
37
+ `Settling epoch ${epoch} with blocks ${blocks[0]?.header.getBlockNumber()} to ${blocks.at(-1)?.header.getBlockNumber()}`,
38
+ { blocks: blocks.map(b => b.toBlockInfo()) },
39
+ );
33
40
  const messagesInEpoch: Fr[][][][] = [];
34
41
  let previousSlotNumber = SlotNumber.ZERO;
35
42
  let checkpointIndex = -1;
43
+
36
44
  for (const block of blocks) {
37
45
  const slotNumber = block.header.globalVariables.slotNumber;
38
46
  if (slotNumber !== previousSlotNumber) {
@@ -43,15 +51,18 @@ export class EpochTestSettler {
43
51
  messagesInEpoch[checkpointIndex].push(block.body.txEffects.map(txEffect => txEffect.l2ToL1Msgs));
44
52
  }
45
53
 
46
- const [firstMessage] = messagesInEpoch.flat(3);
47
- if (firstMessage) {
48
- const { root: outHash } = computeL2ToL1MembershipWitnessFromMessagesInEpoch(messagesInEpoch, firstMessage);
54
+ const outHash = computeEpochOutHash(messagesInEpoch);
55
+ if (!outHash.isZero()) {
49
56
  await this.rollupCheatCodes.insertOutbox(epoch, outHash.toBigInt());
57
+ } else {
58
+ this.log.info(`No L2 to L1 messages in epoch ${epoch}`);
50
59
  }
51
60
 
52
- // Mark the blocks as proven.
53
- for (const block of blocks) {
54
- await this.rollupCheatCodes.markAsProven(block.number);
61
+ const lastCheckpoint = checkpointedBlocks.at(-1)?.checkpointNumber;
62
+ if (lastCheckpoint !== undefined) {
63
+ await this.rollupCheatCodes.markAsProven(lastCheckpoint);
64
+ } else {
65
+ this.log.warn(`No checkpoint found for epoch ${epoch}`);
55
66
  }
56
67
 
57
68
  return true;
@@ -2,3 +2,4 @@ export { AnvilTestWatcher } from './anvil_test_watcher.js';
2
2
  export { EthCheatCodes, RollupCheatCodes } from '@aztec/ethereum/test';
3
3
  export { CheatCodes } from './cheat_codes.js';
4
4
  export { EpochTestSettler } from './epoch_test_settler.js';
5
+ export { getTokenAllowedSetupFunctions } from './token_allowed_setup.js';
@@ -0,0 +1,19 @@
1
+ import { TokenContractArtifact } from '@aztec/noir-contracts.js/Token';
2
+ import { buildAllowedElement } from '@aztec/p2p/msg_validators';
3
+ import { getContractClassFromArtifact } from '@aztec/stdlib/contract';
4
+ import type { AllowedElement } from '@aztec/stdlib/interfaces/server';
5
+
6
+ /**
7
+ * Returns Token-specific allowlist entries needed for FPC-based fee payments.
8
+ * These are test-only: FPC-based fee payment with custom tokens won't work on mainnet alpha.
9
+ */
10
+ export async function getTokenAllowedSetupFunctions(): Promise<AllowedElement[]> {
11
+ const tokenClassId = (await getContractClassFromArtifact(TokenContractArtifact)).id;
12
+ const target = { classId: tokenClassId };
13
+ return Promise.all([
14
+ // Token: needed for private transfers via FPC (transfer_to_public enqueues this)
15
+ buildAllowedElement(TokenContractArtifact, target, '_increase_public_balance', { onlySelf: true }),
16
+ // Token: needed for public transfers via FPC (fee_entrypoint_public enqueues this)
17
+ buildAllowedElement(TokenContractArtifact, target, 'transfer_in_public'),
18
+ ]);
19
+ }
@@ -1,7 +0,0 @@
1
- import type { NamespacedApiHandlers } from '@aztec/foundation/json-rpc/server';
2
- import type { LogFn } from '@aztec/foundation/log';
3
- import { type ProverNodeConfig } from '@aztec/prover-node';
4
- export declare function startProverNode(options: any, signalHandlers: (() => Promise<void>)[], services: NamespacedApiHandlers, userLog: LogFn): Promise<{
5
- config: ProverNodeConfig;
6
- }>;
7
- //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoic3RhcnRfcHJvdmVyX25vZGUuZC50cyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9jbGkvY21kcy9zdGFydF9wcm92ZXJfbm9kZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFLQSxPQUFPLEtBQUssRUFBRSxxQkFBcUIsRUFBRSxNQUFNLG1DQUFtQyxDQUFDO0FBRS9FLE9BQU8sS0FBSyxFQUFFLEtBQUssRUFBRSxNQUFNLHVCQUF1QixDQUFDO0FBRW5ELE9BQU8sRUFDTCxLQUFLLGdCQUFnQixFQUl0QixNQUFNLG9CQUFvQixDQUFDO0FBUzVCLHdCQUFzQixlQUFlLENBQ25DLE9BQU8sRUFBRSxHQUFHLEVBQ1osY0FBYyxFQUFFLENBQUMsTUFBTSxPQUFPLENBQUMsSUFBSSxDQUFDLENBQUMsRUFBRSxFQUN2QyxRQUFRLEVBQUUscUJBQXFCLEVBQy9CLE9BQU8sRUFBRSxLQUFLLEdBQ2IsT0FBTyxDQUFDO0lBQUUsTUFBTSxFQUFFLGdCQUFnQixDQUFBO0NBQUUsQ0FBQyxDQStGdkMifQ==
@@ -1 +0,0 @@
1
- {"version":3,"file":"start_prover_node.d.ts","sourceRoot":"","sources":["../../../src/cli/cmds/start_prover_node.ts"],"names":[],"mappings":"AAKA,OAAO,KAAK,EAAE,qBAAqB,EAAE,MAAM,mCAAmC,CAAC;AAE/E,OAAO,KAAK,EAAE,KAAK,EAAE,MAAM,uBAAuB,CAAC;AAEnD,OAAO,EACL,KAAK,gBAAgB,EAItB,MAAM,oBAAoB,CAAC;AAS5B,wBAAsB,eAAe,CACnC,OAAO,EAAE,GAAG,EACZ,cAAc,EAAE,CAAC,MAAM,OAAO,CAAC,IAAI,CAAC,CAAC,EAAE,EACvC,QAAQ,EAAE,qBAAqB,EAC/B,OAAO,EAAE,KAAK,GACb,OAAO,CAAC;IAAE,MAAM,EAAE,gBAAgB,CAAA;CAAE,CAAC,CA+FvC"}
@@ -1,108 +0,0 @@
1
- import { getInitialTestAccountsData } from '@aztec/accounts/testing';
2
- import { Fr } from '@aztec/aztec.js/fields';
3
- import { getSponsoredFPCAddress } from '@aztec/cli/cli-utils';
4
- import { getL1Config } from '@aztec/cli/config';
5
- import { getPublicClient } from '@aztec/ethereum/client';
6
- import { Agent, makeUndiciFetch } from '@aztec/foundation/json-rpc/undici';
7
- import { ProvingJobConsumerSchema, createProvingJobBrokerClient } from '@aztec/prover-client/broker';
8
- import { createProverNode, getProverNodeConfigFromEnv, proverNodeConfigMappings } from '@aztec/prover-node';
9
- import { P2PApiSchema, ProverNodeApiSchema } from '@aztec/stdlib/interfaces/server';
10
- import { initTelemetryClient, makeTracedFetch, telemetryClientConfigMappings } from '@aztec/telemetry-client';
11
- import { getGenesisValues } from '@aztec/world-state/testing';
12
- import { extractRelevantOptions, preloadCrsDataForVerifying, setupUpdateMonitor } from '../util.js';
13
- import { getVersions } from '../versioning.js';
14
- import { startProverBroker } from './start_prover_broker.js';
15
- export async function startProverNode(options, signalHandlers, services, userLog) {
16
- if (options.node || options.sequencer || options.pxe || options.p2pBootstrap || options.txe) {
17
- userLog(`Starting a prover-node with --node, --sequencer, --pxe, --p2p-bootstrap, or --txe is not supported.`);
18
- process.exit(1);
19
- }
20
- let proverConfig = {
21
- ...getProverNodeConfigFromEnv(),
22
- ...extractRelevantOptions(options, proverNodeConfigMappings, 'proverNode')
23
- };
24
- if (!proverConfig.l1Contracts.registryAddress || proverConfig.l1Contracts.registryAddress.isZero()) {
25
- throw new Error('L1 registry address is required to start a Prover Node');
26
- }
27
- const followsCanonicalRollup = typeof proverConfig.rollupVersion !== 'number';
28
- const { addresses, config } = await getL1Config(proverConfig.l1Contracts.registryAddress, proverConfig.l1RpcUrls, proverConfig.l1ChainId, proverConfig.rollupVersion);
29
- process.env.ROLLUP_CONTRACT_ADDRESS ??= addresses.rollupAddress.toString();
30
- proverConfig.l1Contracts = addresses;
31
- proverConfig = {
32
- ...proverConfig,
33
- ...config
34
- };
35
- const testAccounts = proverConfig.testAccounts ? (await getInitialTestAccountsData()).map((a)=>a.address) : [];
36
- const sponsoredFPCAccounts = proverConfig.sponsoredFPC ? [
37
- await getSponsoredFPCAddress()
38
- ] : [];
39
- const initialFundedAccounts = testAccounts.concat(sponsoredFPCAccounts);
40
- userLog(`Initial funded accounts: ${initialFundedAccounts.map((a)=>a.toString()).join(', ')}`);
41
- const { genesisArchiveRoot, prefilledPublicData } = await getGenesisValues(initialFundedAccounts);
42
- userLog(`Genesis archive root: ${genesisArchiveRoot.toString()}`);
43
- if (!Fr.fromHexString(config.genesisArchiveTreeRoot).equals(genesisArchiveRoot)) {
44
- throw new Error(`The computed genesis archive tree root ${genesisArchiveRoot} does not match the expected genesis archive tree root ${config.genesisArchiveTreeRoot} for the rollup deployed at ${addresses.rollupAddress}`);
45
- }
46
- const telemetry = await initTelemetryClient(extractRelevantOptions(options, telemetryClientConfigMappings, 'tel'));
47
- let broker;
48
- if (proverConfig.proverBrokerUrl) {
49
- // at 1TPS we'd enqueue ~1k chonk verifier proofs and ~1k AVM proofs immediately
50
- // set a lower connection limit such that we don't overload the server
51
- // Keep retrying up to 30s
52
- const fetch = makeTracedFetch([
53
- 1,
54
- 2,
55
- 3,
56
- 3,
57
- 3,
58
- 3,
59
- 3,
60
- 3,
61
- 3,
62
- 3,
63
- 3
64
- ], false, makeUndiciFetch(new Agent({
65
- connections: 100
66
- })));
67
- broker = createProvingJobBrokerClient(proverConfig.proverBrokerUrl, getVersions(proverConfig), fetch);
68
- } else if (options.proverBroker) {
69
- ({ broker } = await startProverBroker(options, signalHandlers, services, userLog));
70
- } else {
71
- userLog(`--prover-broker-url or --prover-broker is required to start a Prover Node`);
72
- process.exit(1);
73
- }
74
- if (proverConfig.proverAgentCount === 0) {
75
- userLog(`Running prover node without local prover agent. Connect one or more prover agents to this node or pass --proverAgent.proverAgentCount`);
76
- }
77
- await preloadCrsDataForVerifying(proverConfig, userLog);
78
- const proverNode = await createProverNode(proverConfig, {
79
- telemetry,
80
- broker
81
- }, {
82
- prefilledPublicData
83
- });
84
- services.proverNode = [
85
- proverNode,
86
- ProverNodeApiSchema
87
- ];
88
- if (proverNode.getP2P()) {
89
- services.p2p = [
90
- proverNode.getP2P(),
91
- P2PApiSchema
92
- ];
93
- }
94
- if (!proverConfig.proverBrokerUrl) {
95
- services.provingJobSource = [
96
- proverNode.getProver().getProvingJobSource(),
97
- ProvingJobConsumerSchema
98
- ];
99
- }
100
- signalHandlers.push(proverNode.stop.bind(proverNode));
101
- await proverNode.start();
102
- if (proverConfig.autoUpdate !== 'disabled' && proverConfig.autoUpdateUrl) {
103
- await setupUpdateMonitor(proverConfig.autoUpdate, new URL(proverConfig.autoUpdateUrl), followsCanonicalRollup, getPublicClient(proverConfig), proverConfig.l1Contracts.registryAddress, signalHandlers);
104
- }
105
- return {
106
- config: proverConfig
107
- };
108
- }
@@ -1,2 +0,0 @@
1
- export declare const getCliVersion: () => any;
2
- //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicmVsZWFzZV92ZXJzaW9uLmQudHMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY2xpL3JlbGVhc2VfdmVyc2lvbi50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFLQSxlQUFPLE1BQU0sYUFBYSxXQWV6QixDQUFDIn0=
@@ -1 +0,0 @@
1
- {"version":3,"file":"release_version.d.ts","sourceRoot":"","sources":["../../src/cli/release_version.ts"],"names":[],"mappings":"AAKA,eAAO,MAAM,aAAa,WAezB,CAAC"}
@@ -1,14 +0,0 @@
1
- import { fileURLToPath } from '@aztec/foundation/url';
2
- import { readFileSync } from 'fs';
3
- import { dirname, resolve } from 'path';
4
- export const getCliVersion = ()=>{
5
- const packageJsonPath = resolve(dirname(fileURLToPath(import.meta.url)), '../../package.json');
6
- const cliVersion = JSON.parse(readFileSync(packageJsonPath).toString()).version;
7
- // If the version is 0.1.0, this is a placeholder version and we are in a docker container; query release please for the latest version
8
- if (cliVersion === '0.1.0') {
9
- const releasePleasePath = resolve(dirname(fileURLToPath(import.meta.url)), '../../../../.release-please-manifest.json');
10
- const releaseVersion = JSON.parse(readFileSync(releasePleasePath).toString())['.'];
11
- return releaseVersion;
12
- }
13
- return cliVersion;
14
- };
@@ -1,44 +0,0 @@
1
- #!/usr/bin/env bash
2
- set -euo pipefail
3
-
4
- NARGO=${NARGO:-nargo}
5
- BB=${BB:-bb}
6
-
7
- # If help is requested, show Aztec-specific info then run nargo compile help and then exit in order to not trigger
8
- # transpilation
9
- for arg in "$@"; do
10
- if [ "$arg" == "--help" ] || [ "$arg" == "-h" ]; then
11
- cat << 'EOF'
12
- Aztec Compile - Compile Aztec Noir contracts
13
-
14
- This command compiles Aztec Noir contracts using nargo and then automatically
15
- postprocesses them to generate Aztec specific artifacts including:
16
- - Transpiled contract artifacts
17
- - Verification keys
18
-
19
- The compiled contracts will be placed in the target/ directory by default.
20
-
21
- ---
22
- Underlying nargo compile options:
23
-
24
- EOF
25
- nargo compile --help
26
- exit 0
27
- fi
28
- done
29
-
30
- # Run nargo compile.
31
- $NARGO compile "$@"
32
-
33
- echo "Postprocessing contract..."
34
- $BB aztec_process
35
-
36
- # Strip internal prefixes from all compiled contract JSONs in target directory
37
- # TODO: This should be part of bb aztec_process!
38
- for json in target/*.json; do
39
- temp_file="${json}.tmp"
40
- jq '.functions |= map(.name |= sub("^__aztec_nr_internals__"; ""))' "$json" > "$temp_file"
41
- mv "$temp_file" "$json"
42
- done
43
-
44
- echo "Compilation complete!"
@@ -1,47 +0,0 @@
1
- #!/usr/bin/env node
2
- import fs from 'fs/promises';
3
- import path from 'path';
4
-
5
- // Simple script to extract a contract function as a separate Noir artifact.
6
- // We need to use this since the transpiling that we do on public functions make the contract artifacts
7
- // unreadable by noir tooling, since they are no longer following the noir artifact format.
8
- async function main() {
9
- let [contractArtifactPath, functionName] = process.argv.slice(2);
10
- if (!contractArtifactPath || !functionName) {
11
- console.log('Usage: node extractFunctionAsNoirArtifact.js <contractArtifactPath> <functionName>');
12
- return;
13
- }
14
-
15
- const contractArtifact = JSON.parse(await fs.readFile(contractArtifactPath, 'utf8'));
16
- const func = contractArtifact.functions.find(f => f.name === functionName);
17
- if (!func) {
18
- console.error(`Function ${functionName} not found in ${contractArtifactPath}`);
19
- return;
20
- }
21
-
22
- const artifact = {
23
- noir_version: contractArtifact.noir_version,
24
- hash: 0,
25
- abi: func.abi,
26
- bytecode: func.bytecode,
27
- debug_symbols: func.debug_symbols,
28
- file_map: contractArtifact.file_map,
29
- expression_width: {
30
- Bounded: {
31
- width: 4,
32
- },
33
- },
34
- };
35
-
36
- const outputDir = path.dirname(contractArtifactPath);
37
- const outputName = path.basename(contractArtifactPath, '.json') + `-${functionName}.json`;
38
-
39
- const outPath = path.join(outputDir, outputName);
40
-
41
- await fs.writeFile(outPath, JSON.stringify(artifact, null, 2));
42
- }
43
-
44
- main().catch(err => {
45
- console.error(err);
46
- process.exit(1);
47
- });
@@ -1,59 +0,0 @@
1
- #!/usr/bin/env bash
2
- set -eu
3
-
4
- # If first arg is -h or --help, print usage.
5
- if [ $# -lt 2 ] || [ "$1" == "-h" ] || [ "$1" == "--help" ]; then
6
- cat << 'EOF'
7
- Aztec Flamegraph - Generate a gate count flamegraph for an aztec contract function.
8
-
9
- Usage: aztec flamegraph <contract_artifact> <function>
10
-
11
- Options:
12
- -h, --help Print help
13
-
14
- Will output an svg at <artifact_path>/<contract>-<function>-flamegraph.svg.
15
- You can open it in your browser to view it.
16
-
17
- EOF
18
- exit 0
19
- fi
20
-
21
- cleanup() {
22
- set +e
23
- if [ -f "$function_artifact" ]; then
24
- rm -f "$function_artifact"
25
- fi
26
- }
27
-
28
- trap cleanup EXIT
29
-
30
- # Get the directory of the script
31
- script_dir=$(realpath $(dirname $0))
32
-
33
- PROFILER=${PROFILER_PATH:-noir-profiler}
34
- BB=${BB:-bb}
35
-
36
- # first console arg is contract name in camel case or path to contract artifact
37
- contract=$1
38
-
39
- # second console arg is the contract function
40
- function=$2
41
-
42
- if [ ! -f "$contract" ]; then
43
- echo "Error: Contract artifact not found at: $contract"
44
- exit 1
45
- fi
46
- artifact_path=$contract
47
- function_artifact="${artifact_path%%.json}-${function}.json"
48
- output_dir=$(dirname "$artifact_path")
49
-
50
- # Extract artifact for the specific function.
51
- node $script_dir/extract_function.js "$artifact_path" $function
52
-
53
- # Generate the flamegraph
54
- $PROFILER gates --artifact-path "$function_artifact" --backend-path "$BB" --backend-gates-command "gates" --output "$output_dir" --scheme chonk --include_gates_per_opcode
55
-
56
- # Save as $artifact_name-$function-flamegraph.svg
57
- output_file="${function_artifact%%.json}-flamegraph.svg"
58
- mv "$output_dir/__aztec_nr_internals__${function}_gates.svg" "$output_file"
59
- echo "Flamegraph generated at: $output_file"