@aztec/aztec 0.0.1-commit.4eabbdb → 0.0.1-commit.5358163d3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/bin/index.js +5 -1
- package/dest/cli/admin_api_key_store.d.ts +3 -3
- package/dest/cli/admin_api_key_store.d.ts.map +1 -1
- package/dest/cli/admin_api_key_store.js +3 -3
- package/dest/cli/aztec_start_action.js +4 -4
- package/dest/cli/aztec_start_options.d.ts +1 -1
- package/dest/cli/aztec_start_options.d.ts.map +1 -1
- package/dest/cli/aztec_start_options.js +4 -3
- package/dest/cli/cli.d.ts +1 -1
- package/dest/cli/cli.d.ts.map +1 -1
- package/dest/cli/cli.js +0 -1
- package/dest/cli/cmds/compile.d.ts +4 -0
- package/dest/cli/cmds/compile.d.ts.map +1 -0
- package/dest/cli/cmds/compile.js +160 -0
- package/dest/cli/cmds/profile.d.ts +4 -0
- package/dest/cli/cmds/profile.d.ts.map +1 -0
- package/dest/cli/cmds/profile.js +8 -0
- package/dest/cli/cmds/profile_flamegraph.d.ts +4 -0
- package/dest/cli/cmds/profile_flamegraph.d.ts.map +1 -0
- package/dest/cli/cmds/profile_flamegraph.js +51 -0
- package/dest/cli/cmds/profile_gates.d.ts +4 -0
- package/dest/cli/cmds/profile_gates.d.ts.map +1 -0
- package/dest/cli/cmds/profile_gates.js +57 -0
- package/dest/cli/cmds/profile_utils.d.ts +18 -0
- package/dest/cli/cmds/profile_utils.d.ts.map +1 -0
- package/dest/cli/cmds/profile_utils.js +50 -0
- package/dest/cli/cmds/start_node.d.ts +1 -1
- package/dest/cli/cmds/start_node.d.ts.map +1 -1
- package/dest/cli/cmds/start_node.js +59 -1
- package/dest/cli/cmds/utils/artifacts.d.ts +21 -0
- package/dest/cli/cmds/utils/artifacts.d.ts.map +1 -0
- package/dest/cli/cmds/utils/artifacts.js +24 -0
- package/dest/cli/cmds/utils/spawn.d.ts +3 -0
- package/dest/cli/cmds/utils/spawn.d.ts.map +1 -0
- package/dest/cli/cmds/utils/spawn.js +16 -0
- package/dest/cli/util.js +3 -3
- package/dest/local-network/local-network.d.ts +1 -1
- package/dest/local-network/local-network.d.ts.map +1 -1
- package/dest/local-network/local-network.js +24 -4
- package/dest/testing/anvil_test_watcher.d.ts +9 -1
- package/dest/testing/anvil_test_watcher.d.ts.map +1 -1
- package/dest/testing/anvil_test_watcher.js +52 -15
- package/package.json +34 -34
- package/scripts/aztec.sh +7 -4
- package/scripts/init.sh +23 -13
- package/scripts/new.sh +17 -16
- package/scripts/setup_workspace.sh +124 -0
- package/src/bin/index.ts +5 -1
- package/src/cli/admin_api_key_store.ts +4 -4
- package/src/cli/aztec_start_action.ts +4 -4
- package/src/cli/aztec_start_options.ts +4 -3
- package/src/cli/cli.ts +0 -1
- package/src/cli/cmds/compile.ts +184 -0
- package/src/cli/cmds/profile.ts +25 -0
- package/src/cli/cmds/profile_flamegraph.ts +63 -0
- package/src/cli/cmds/profile_gates.ts +67 -0
- package/src/cli/cmds/profile_utils.ts +58 -0
- package/src/cli/cmds/start_node.ts +89 -1
- package/src/cli/cmds/utils/artifacts.ts +44 -0
- package/src/cli/cmds/utils/spawn.ts +16 -0
- package/src/cli/util.ts +2 -2
- package/src/local-network/local-network.ts +23 -3
- package/src/testing/anvil_test_watcher.ts +59 -15
- package/scripts/compile.sh +0 -44
- package/scripts/extract_function.js +0 -47
- package/scripts/flamegraph.sh +0 -59
- package/scripts/setup_project.sh +0 -31
|
@@ -0,0 +1,58 @@
|
|
|
1
|
+
import { mkdtemp, writeFile } from 'fs/promises';
|
|
2
|
+
import { tmpdir } from 'os';
|
|
3
|
+
import { join } from 'path';
|
|
4
|
+
|
|
5
|
+
import type { CompiledArtifact, ContractFunction } from './utils/artifacts.js';
|
|
6
|
+
import { readArtifactFiles } from './utils/artifacts.js';
|
|
7
|
+
|
|
8
|
+
export const MAX_CONCURRENT = 4;
|
|
9
|
+
|
|
10
|
+
export interface DiscoveredArtifact {
|
|
11
|
+
name: string;
|
|
12
|
+
filePath: string;
|
|
13
|
+
type: 'contract-function' | 'program';
|
|
14
|
+
}
|
|
15
|
+
|
|
16
|
+
/**
|
|
17
|
+
* Reads a target directory and returns a list of discovered artifacts with temp files
|
|
18
|
+
* created for contract functions. Caller must clean up tmpDir when done.
|
|
19
|
+
*/
|
|
20
|
+
export async function discoverArtifacts(
|
|
21
|
+
targetDir: string,
|
|
22
|
+
): Promise<{ artifacts: DiscoveredArtifact[]; tmpDir: string }> {
|
|
23
|
+
const files = await readArtifactFiles(targetDir);
|
|
24
|
+
const tmpDir = await mkdtemp(join(tmpdir(), 'aztec-profile-'));
|
|
25
|
+
const artifacts: DiscoveredArtifact[] = [];
|
|
26
|
+
|
|
27
|
+
for (const file of files) {
|
|
28
|
+
if (Array.isArray(file.content.functions)) {
|
|
29
|
+
for (const func of file.content.functions) {
|
|
30
|
+
if (!func.bytecode || func.is_unconstrained) {
|
|
31
|
+
continue;
|
|
32
|
+
}
|
|
33
|
+
const name = `${file.name}::${func.name}`;
|
|
34
|
+
const tmpPath = join(tmpDir, `${file.name}-${func.name}.json`);
|
|
35
|
+
await writeFile(tmpPath, makeFunctionArtifact(file.content, func));
|
|
36
|
+
artifacts.push({ name, filePath: tmpPath, type: 'contract-function' });
|
|
37
|
+
}
|
|
38
|
+
} else if (file.content.bytecode) {
|
|
39
|
+
artifacts.push({ name: file.name, filePath: file.filePath, type: 'program' });
|
|
40
|
+
}
|
|
41
|
+
}
|
|
42
|
+
|
|
43
|
+
return { artifacts, tmpDir };
|
|
44
|
+
}
|
|
45
|
+
|
|
46
|
+
/** Extracts a contract function as a standalone program artifact JSON string. */
|
|
47
|
+
export function makeFunctionArtifact(artifact: CompiledArtifact, func: ContractFunction) {
|
|
48
|
+
/* eslint-disable camelcase */
|
|
49
|
+
return JSON.stringify({
|
|
50
|
+
noir_version: artifact.noir_version,
|
|
51
|
+
hash: 0,
|
|
52
|
+
abi: func.abi,
|
|
53
|
+
bytecode: func.bytecode,
|
|
54
|
+
debug_symbols: func.debug_symbols,
|
|
55
|
+
file_map: artifact.file_map,
|
|
56
|
+
});
|
|
57
|
+
/* eslint-enable camelcase */
|
|
58
|
+
}
|
|
@@ -4,12 +4,17 @@ import { Fr } from '@aztec/aztec.js/fields';
|
|
|
4
4
|
import { getSponsoredFPCAddress } from '@aztec/cli/cli-utils';
|
|
5
5
|
import { getL1Config } from '@aztec/cli/config';
|
|
6
6
|
import { getPublicClient } from '@aztec/ethereum/client';
|
|
7
|
+
import { RegistryContract, RollupContract } from '@aztec/ethereum/contracts';
|
|
7
8
|
import { SecretValue } from '@aztec/foundation/config';
|
|
9
|
+
import { EthAddress } from '@aztec/foundation/eth-address';
|
|
8
10
|
import type { NamespacedApiHandlers } from '@aztec/foundation/json-rpc/server';
|
|
11
|
+
import { startHttpRpcServer } from '@aztec/foundation/json-rpc/server';
|
|
9
12
|
import { Agent, makeUndiciFetch } from '@aztec/foundation/json-rpc/undici';
|
|
10
13
|
import type { LogFn } from '@aztec/foundation/log';
|
|
14
|
+
import { sleep } from '@aztec/foundation/sleep';
|
|
11
15
|
import { ProvingJobConsumerSchema, createProvingJobBrokerClient } from '@aztec/prover-client/broker';
|
|
12
16
|
import { type CliPXEOptions, type PXEConfig, allPxeConfigMappings } from '@aztec/pxe/config';
|
|
17
|
+
import { AztecAddress } from '@aztec/stdlib/aztec-address';
|
|
13
18
|
import { AztecNodeAdminApiSchema, AztecNodeApiSchema } from '@aztec/stdlib/interfaces/client';
|
|
14
19
|
import { P2PApiSchema, ProverNodeApiSchema, type ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
|
|
15
20
|
import {
|
|
@@ -21,6 +26,8 @@ import {
|
|
|
21
26
|
import { EmbeddedWallet } from '@aztec/wallets/embedded';
|
|
22
27
|
import { getGenesisValues } from '@aztec/world-state/testing';
|
|
23
28
|
|
|
29
|
+
import Koa from 'koa';
|
|
30
|
+
|
|
24
31
|
import { createAztecNode } from '../../local-network/index.js';
|
|
25
32
|
import {
|
|
26
33
|
extractNamespacedOptions,
|
|
@@ -31,6 +38,72 @@ import {
|
|
|
31
38
|
import { getVersions } from '../versioning.js';
|
|
32
39
|
import { startProverBroker } from './start_prover_broker.js';
|
|
33
40
|
|
|
41
|
+
const ROLLUP_POLL_INTERVAL_MS = 600_000;
|
|
42
|
+
|
|
43
|
+
/**
|
|
44
|
+
* Waits until the canonical rollup's genesis archive root matches the expected local genesis root.
|
|
45
|
+
* If the rollup is not yet compatible (e.g. during L1 contract upgrades), enters standby mode:
|
|
46
|
+
* starts a lightweight HTTP server for K8s liveness probes and polls until a compatible rollup appears.
|
|
47
|
+
*/
|
|
48
|
+
async function waitForCompatibleRollup(
|
|
49
|
+
publicClient: ReturnType<typeof getPublicClient>,
|
|
50
|
+
registryAddress: EthAddress,
|
|
51
|
+
rollupVersion: number | 'canonical',
|
|
52
|
+
expectedGenesisRoot: Fr,
|
|
53
|
+
port: number | undefined,
|
|
54
|
+
userLog: LogFn,
|
|
55
|
+
): Promise<void> {
|
|
56
|
+
const registry = new RegistryContract(publicClient, registryAddress);
|
|
57
|
+
const rollupAddress = await registry.getRollupAddress(rollupVersion);
|
|
58
|
+
const rollup = new RollupContract(publicClient, rollupAddress.toString());
|
|
59
|
+
|
|
60
|
+
let l1GenesisRoot: Fr;
|
|
61
|
+
try {
|
|
62
|
+
l1GenesisRoot = await rollup.getGenesisArchiveTreeRoot();
|
|
63
|
+
} catch (err: any) {
|
|
64
|
+
throw new Error(
|
|
65
|
+
`Could not retrieve genesis archive root from canonical rollup at ${rollupAddress}: ${err.message}`,
|
|
66
|
+
);
|
|
67
|
+
}
|
|
68
|
+
|
|
69
|
+
if (l1GenesisRoot.equals(expectedGenesisRoot)) {
|
|
70
|
+
return;
|
|
71
|
+
}
|
|
72
|
+
|
|
73
|
+
userLog(
|
|
74
|
+
`Genesis root mismatch: expected ${expectedGenesisRoot}, got ${l1GenesisRoot} from rollup at ${rollupAddress}. ` +
|
|
75
|
+
`Entering standby mode. Will poll every ${ROLLUP_POLL_INTERVAL_MS / 1000}s for a compatible rollup...`,
|
|
76
|
+
);
|
|
77
|
+
|
|
78
|
+
const standbyServer = await startHttpRpcServer({ getApp: () => new Koa(), isHealthy: () => true }, { port });
|
|
79
|
+
userLog(`Standby status server listening on port ${standbyServer.port}`);
|
|
80
|
+
|
|
81
|
+
try {
|
|
82
|
+
while (true) {
|
|
83
|
+
await sleep(ROLLUP_POLL_INTERVAL_MS);
|
|
84
|
+
|
|
85
|
+
const currentRollupAddress = await registry.getRollupAddress(rollupVersion);
|
|
86
|
+
const currentRollup = new RollupContract(publicClient, currentRollupAddress.toString());
|
|
87
|
+
|
|
88
|
+
try {
|
|
89
|
+
l1GenesisRoot = await currentRollup.getGenesisArchiveTreeRoot();
|
|
90
|
+
} catch {
|
|
91
|
+
userLog(`Failed to fetch genesis root from rollup at ${currentRollupAddress}. Retrying...`);
|
|
92
|
+
continue;
|
|
93
|
+
}
|
|
94
|
+
|
|
95
|
+
if (l1GenesisRoot.equals(expectedGenesisRoot)) {
|
|
96
|
+
userLog(`Compatible rollup found at ${currentRollupAddress}. Exiting standby mode.`);
|
|
97
|
+
return;
|
|
98
|
+
}
|
|
99
|
+
|
|
100
|
+
userLog(`Still waiting. Rollup at ${currentRollupAddress} has genesis root ${l1GenesisRoot}.`);
|
|
101
|
+
}
|
|
102
|
+
} finally {
|
|
103
|
+
await new Promise<void>((resolve, reject) => standbyServer.close(err => (err ? reject(err) : resolve())));
|
|
104
|
+
}
|
|
105
|
+
}
|
|
106
|
+
|
|
34
107
|
export async function startNode(
|
|
35
108
|
options: any,
|
|
36
109
|
signalHandlers: (() => Promise<void>)[],
|
|
@@ -82,7 +155,8 @@ export async function startNode(
|
|
|
82
155
|
|
|
83
156
|
const testAccounts = nodeConfig.testAccounts ? (await getInitialTestAccountsData()).map(a => a.address) : [];
|
|
84
157
|
const sponsoredFPCAccounts = nodeConfig.sponsoredFPC ? [await getSponsoredFPCAddress()] : [];
|
|
85
|
-
const
|
|
158
|
+
const prefundAddresses = (nodeConfig.prefundAddresses ?? []).map(a => AztecAddress.fromString(a));
|
|
159
|
+
const initialFundedAccounts = testAccounts.concat(sponsoredFPCAccounts).concat(prefundAddresses);
|
|
86
160
|
|
|
87
161
|
userLog(`Initial funded accounts: ${initialFundedAccounts.map(a => a.toString()).join(', ')}`);
|
|
88
162
|
|
|
@@ -96,6 +170,20 @@ export async function startNode(
|
|
|
96
170
|
if (!nodeConfig.l1Contracts.registryAddress || nodeConfig.l1Contracts.registryAddress.isZero()) {
|
|
97
171
|
throw new Error('L1 registry address is required to start Aztec Node');
|
|
98
172
|
}
|
|
173
|
+
|
|
174
|
+
// Wait for a compatible rollup before proceeding with full L1 config fetch.
|
|
175
|
+
// This prevents crashes when the canonical rollup hasn't been upgraded yet.
|
|
176
|
+
const publicClient = getPublicClient(nodeConfig);
|
|
177
|
+
const rollupVersion: number | 'canonical' = nodeConfig.rollupVersion ?? 'canonical';
|
|
178
|
+
await waitForCompatibleRollup(
|
|
179
|
+
publicClient,
|
|
180
|
+
nodeConfig.l1Contracts.registryAddress,
|
|
181
|
+
rollupVersion,
|
|
182
|
+
genesisArchiveRoot,
|
|
183
|
+
options.port,
|
|
184
|
+
userLog,
|
|
185
|
+
);
|
|
186
|
+
|
|
99
187
|
const { addresses, config } = await getL1Config(
|
|
100
188
|
nodeConfig.l1Contracts.registryAddress,
|
|
101
189
|
nodeConfig.l1RpcUrls,
|
|
@@ -0,0 +1,44 @@
|
|
|
1
|
+
import { readFile, readdir } from 'fs/promises';
|
|
2
|
+
import { join } from 'path';
|
|
3
|
+
|
|
4
|
+
export interface CompiledArtifact {
|
|
5
|
+
noir_version: string;
|
|
6
|
+
file_map: unknown;
|
|
7
|
+
functions: ContractFunction[];
|
|
8
|
+
bytecode?: string;
|
|
9
|
+
}
|
|
10
|
+
|
|
11
|
+
export interface ContractFunction {
|
|
12
|
+
name: string;
|
|
13
|
+
abi: unknown;
|
|
14
|
+
bytecode: string;
|
|
15
|
+
debug_symbols: unknown;
|
|
16
|
+
is_unconstrained?: boolean;
|
|
17
|
+
}
|
|
18
|
+
|
|
19
|
+
export interface ArtifactFile {
|
|
20
|
+
name: string;
|
|
21
|
+
filePath: string;
|
|
22
|
+
content: CompiledArtifact;
|
|
23
|
+
}
|
|
24
|
+
|
|
25
|
+
/** Reads all JSON artifact files from a target directory and returns their parsed contents. */
|
|
26
|
+
export async function readArtifactFiles(targetDir: string): Promise<ArtifactFile[]> {
|
|
27
|
+
let entries: string[];
|
|
28
|
+
try {
|
|
29
|
+
entries = (await readdir(targetDir)).filter(f => f.endsWith('.json'));
|
|
30
|
+
} catch (err: any) {
|
|
31
|
+
if (err?.code === 'ENOENT') {
|
|
32
|
+
throw new Error(`Target directory '${targetDir}' does not exist. Compile first with 'aztec compile'.`);
|
|
33
|
+
}
|
|
34
|
+
throw err;
|
|
35
|
+
}
|
|
36
|
+
|
|
37
|
+
const artifacts: ArtifactFile[] = [];
|
|
38
|
+
for (const file of entries) {
|
|
39
|
+
const filePath = join(targetDir, file);
|
|
40
|
+
const content = JSON.parse(await readFile(filePath, 'utf-8')) as CompiledArtifact;
|
|
41
|
+
artifacts.push({ name: file.replace('.json', ''), filePath, content });
|
|
42
|
+
}
|
|
43
|
+
return artifacts;
|
|
44
|
+
}
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
import { spawn } from 'child_process';
|
|
2
|
+
|
|
3
|
+
/** Spawns a command with inherited stdio and rejects on non-zero exit. */
|
|
4
|
+
export function run(cmd: string, args: string[]): Promise<void> {
|
|
5
|
+
return new Promise((resolve, reject) => {
|
|
6
|
+
const child = spawn(cmd, args, { stdio: 'inherit' });
|
|
7
|
+
child.on('error', reject);
|
|
8
|
+
child.on('close', code => {
|
|
9
|
+
if (code !== 0) {
|
|
10
|
+
reject(new Error(`${cmd} exited with code ${code}`));
|
|
11
|
+
} else {
|
|
12
|
+
resolve();
|
|
13
|
+
}
|
|
14
|
+
});
|
|
15
|
+
});
|
|
16
|
+
}
|
package/src/cli/util.ts
CHANGED
|
@@ -271,7 +271,7 @@ export async function preloadCrsDataForVerifying(
|
|
|
271
271
|
): Promise<void> {
|
|
272
272
|
if (realProofs) {
|
|
273
273
|
const { Crs, GrumpkinCrs } = await import('@aztec/bb.js');
|
|
274
|
-
await Promise.all([Crs.new(2 ** 1, undefined, log), GrumpkinCrs.new(2 ** 16
|
|
274
|
+
await Promise.all([Crs.new(2 ** 1, undefined, log), GrumpkinCrs.new(2 ** 16, undefined, log)]);
|
|
275
275
|
}
|
|
276
276
|
}
|
|
277
277
|
|
|
@@ -286,7 +286,7 @@ export async function preloadCrsDataForServerSideProving(
|
|
|
286
286
|
): Promise<void> {
|
|
287
287
|
if (realProofs) {
|
|
288
288
|
const { Crs, GrumpkinCrs } = await import('@aztec/bb.js');
|
|
289
|
-
await Promise.all([Crs.new(2 ** 25
|
|
289
|
+
await Promise.all([Crs.new(2 ** 25, undefined, log), GrumpkinCrs.new(2 ** 18, undefined, log)]);
|
|
290
290
|
}
|
|
291
291
|
}
|
|
292
292
|
|
|
@@ -18,6 +18,8 @@ import type { LogFn } from '@aztec/foundation/log';
|
|
|
18
18
|
import { DateProvider, TestDateProvider } from '@aztec/foundation/timer';
|
|
19
19
|
import { getVKTreeRoot } from '@aztec/noir-protocol-circuits-types/vk-tree';
|
|
20
20
|
import { protocolContractsHash } from '@aztec/protocol-contracts';
|
|
21
|
+
import { SequencerState } from '@aztec/sequencer-client';
|
|
22
|
+
import { AztecAddress } from '@aztec/stdlib/aztec-address';
|
|
21
23
|
import type { ProvingJobBroker } from '@aztec/stdlib/interfaces/server';
|
|
22
24
|
import type { PublicDataTreeLeaf } from '@aztec/stdlib/trees';
|
|
23
25
|
import {
|
|
@@ -137,9 +139,12 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
|
|
|
137
139
|
|
|
138
140
|
const bananaFPC = await getBananaFPCAddress(initialAccounts);
|
|
139
141
|
const sponsoredFPC = await getSponsoredFPCAddress();
|
|
140
|
-
const
|
|
141
|
-
|
|
142
|
-
|
|
142
|
+
const prefundAddresses = (aztecNodeConfig.prefundAddresses ?? []).map(a => AztecAddress.fromString(a));
|
|
143
|
+
const fundedAddresses = [
|
|
144
|
+
...initialAccounts.map(a => a.address),
|
|
145
|
+
...(initialAccounts.length ? [bananaFPC, sponsoredFPC] : []),
|
|
146
|
+
...prefundAddresses,
|
|
147
|
+
];
|
|
143
148
|
const { genesisArchiveRoot, prefilledPublicData, fundingNeeded } = await getGenesisValues(fundedAddresses);
|
|
144
149
|
|
|
145
150
|
const dateProvider = new TestDateProvider();
|
|
@@ -181,6 +186,21 @@ export async function createLocalNetwork(config: Partial<LocalNetworkConfig> = {
|
|
|
181
186
|
const blobClient = createBlobClient();
|
|
182
187
|
const node = await createAztecNode(aztecNodeConfig, { telemetry, blobClient, dateProvider }, { prefilledPublicData });
|
|
183
188
|
|
|
189
|
+
// Now that the node is up, let the watcher check for pending txs so it can skip unfilled slots faster when
|
|
190
|
+
// transactions are waiting in the mempool. Also let it check if the sequencer is actively building, to avoid
|
|
191
|
+
// warping time out from under an in-progress block.
|
|
192
|
+
watcher?.setGetPendingTxCount(() => node.getPendingTxCount());
|
|
193
|
+
const sequencer = node.getSequencer()?.getSequencer();
|
|
194
|
+
if (sequencer) {
|
|
195
|
+
const idleStates: Set<string> = new Set([
|
|
196
|
+
SequencerState.STOPPED,
|
|
197
|
+
SequencerState.STOPPING,
|
|
198
|
+
SequencerState.IDLE,
|
|
199
|
+
SequencerState.SYNCHRONIZING,
|
|
200
|
+
]);
|
|
201
|
+
watcher?.setIsSequencerBuilding(() => !idleStates.has(sequencer.getState()));
|
|
202
|
+
}
|
|
203
|
+
|
|
184
204
|
let epochTestSettler: EpochTestSettler | undefined;
|
|
185
205
|
if (!aztecNodeConfig.p2pEnabled) {
|
|
186
206
|
epochTestSettler = new EpochTestSettler(
|
|
@@ -31,6 +31,15 @@ export class AnvilTestWatcher {
|
|
|
31
31
|
|
|
32
32
|
private isMarkingAsProven = true;
|
|
33
33
|
|
|
34
|
+
// Optional callback to check if there are pending txs in the mempool.
|
|
35
|
+
private getPendingTxCount?: () => Promise<number>;
|
|
36
|
+
|
|
37
|
+
// Optional callback to check if the sequencer is actively building a block.
|
|
38
|
+
private isSequencerBuilding?: () => boolean;
|
|
39
|
+
|
|
40
|
+
// Tracks when we first observed the current unfilled slot with pending txs (real wall time).
|
|
41
|
+
private unfilledSlotFirstSeen?: { slot: number; realTime: number };
|
|
42
|
+
|
|
34
43
|
constructor(
|
|
35
44
|
private cheatcodes: EthCheatCodes,
|
|
36
45
|
rollupAddress: EthAddress,
|
|
@@ -59,6 +68,16 @@ export class AnvilTestWatcher {
|
|
|
59
68
|
this.isLocalNetwork = isLocalNetwork;
|
|
60
69
|
}
|
|
61
70
|
|
|
71
|
+
/** Sets a callback to check for pending txs, used to skip unfilled slots faster when txs are waiting. */
|
|
72
|
+
setGetPendingTxCount(fn: () => Promise<number>) {
|
|
73
|
+
this.getPendingTxCount = fn;
|
|
74
|
+
}
|
|
75
|
+
|
|
76
|
+
/** Sets a callback to check if the sequencer is actively building, to avoid warping while it works. */
|
|
77
|
+
setIsSequencerBuilding(fn: () => boolean) {
|
|
78
|
+
this.isSequencerBuilding = fn;
|
|
79
|
+
}
|
|
80
|
+
|
|
62
81
|
async start() {
|
|
63
82
|
if (this.filledRunningPromise) {
|
|
64
83
|
throw new Error('Watcher already watching for filled slot');
|
|
@@ -131,15 +150,8 @@ export class AnvilTestWatcher {
|
|
|
131
150
|
const nextSlotTimestamp = Number(await this.rollup.read.getTimestampForSlot([BigInt(nextSlot)]));
|
|
132
151
|
|
|
133
152
|
if (BigInt(currentSlot) === checkpointLog.slotNumber) {
|
|
134
|
-
//
|
|
135
|
-
|
|
136
|
-
await this.cheatcodes.warp(nextSlotTimestamp, {
|
|
137
|
-
resetBlockInterval: true,
|
|
138
|
-
});
|
|
139
|
-
} catch (e) {
|
|
140
|
-
this.logger.error(`Failed to warp to timestamp ${nextSlotTimestamp}: ${e}`);
|
|
141
|
-
}
|
|
142
|
-
|
|
153
|
+
// The current slot has been filled, we should jump to the next slot.
|
|
154
|
+
await this.warpToTimestamp(nextSlotTimestamp);
|
|
143
155
|
this.logger.info(`Slot ${currentSlot} was filled, jumped to next slot`);
|
|
144
156
|
return;
|
|
145
157
|
}
|
|
@@ -149,18 +161,50 @@ export class AnvilTestWatcher {
|
|
|
149
161
|
return;
|
|
150
162
|
}
|
|
151
163
|
|
|
152
|
-
|
|
153
|
-
|
|
154
|
-
|
|
155
|
-
|
|
156
|
-
|
|
157
|
-
|
|
164
|
+
// If there are pending txs and the sequencer missed them, warp quickly (after a 2s real-time debounce) so the
|
|
165
|
+
// sequencer can retry in the next slot. Without this, we'd have to wait a full real-time slot duration (~36s) for
|
|
166
|
+
// the dateProvider to catch up to the next slot timestamp. We skip the warp if the sequencer is actively building
|
|
167
|
+
// to avoid invalidating its in-progress work.
|
|
168
|
+
if (this.getPendingTxCount) {
|
|
169
|
+
const pendingTxs = await this.getPendingTxCount();
|
|
170
|
+
if (pendingTxs > 0) {
|
|
171
|
+
if (this.isSequencerBuilding?.()) {
|
|
172
|
+
this.unfilledSlotFirstSeen = undefined;
|
|
173
|
+
return;
|
|
174
|
+
}
|
|
175
|
+
|
|
176
|
+
const realNow = Date.now();
|
|
177
|
+
if (!this.unfilledSlotFirstSeen || this.unfilledSlotFirstSeen.slot !== currentSlot) {
|
|
178
|
+
this.unfilledSlotFirstSeen = { slot: currentSlot, realTime: realNow };
|
|
179
|
+
return;
|
|
180
|
+
}
|
|
181
|
+
|
|
182
|
+
if (realNow - this.unfilledSlotFirstSeen.realTime > 2000) {
|
|
183
|
+
await this.warpToTimestamp(nextSlotTimestamp);
|
|
184
|
+
this.unfilledSlotFirstSeen = undefined;
|
|
185
|
+
this.logger.info(`Slot ${currentSlot} was missed with pending txs, jumped to next slot`);
|
|
186
|
+
}
|
|
187
|
+
|
|
188
|
+
return;
|
|
158
189
|
}
|
|
190
|
+
}
|
|
159
191
|
|
|
192
|
+
// Fallback: warp when the dateProvider time has passed the next slot timestamp.
|
|
193
|
+
const currentTimestamp = this.dateProvider?.now() ?? Date.now();
|
|
194
|
+
if (currentTimestamp > nextSlotTimestamp * 1000) {
|
|
195
|
+
await this.warpToTimestamp(nextSlotTimestamp);
|
|
160
196
|
this.logger.info(`Slot ${currentSlot} was missed, jumped to next slot`);
|
|
161
197
|
}
|
|
162
198
|
} catch {
|
|
163
199
|
this.logger.error('mineIfSlotFilled failed');
|
|
164
200
|
}
|
|
165
201
|
}
|
|
202
|
+
|
|
203
|
+
private async warpToTimestamp(timestamp: number) {
|
|
204
|
+
try {
|
|
205
|
+
await this.cheatcodes.warp(timestamp, { resetBlockInterval: true });
|
|
206
|
+
} catch (e) {
|
|
207
|
+
this.logger.error(`Failed to warp to timestamp ${timestamp}: ${e}`);
|
|
208
|
+
}
|
|
209
|
+
}
|
|
166
210
|
}
|
package/scripts/compile.sh
DELETED
|
@@ -1,44 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/env bash
|
|
2
|
-
set -euo pipefail
|
|
3
|
-
|
|
4
|
-
NARGO=${NARGO:-nargo}
|
|
5
|
-
BB=${BB:-bb}
|
|
6
|
-
|
|
7
|
-
# If help is requested, show Aztec-specific info then run nargo compile help and then exit in order to not trigger
|
|
8
|
-
# transpilation
|
|
9
|
-
for arg in "$@"; do
|
|
10
|
-
if [ "$arg" == "--help" ] || [ "$arg" == "-h" ]; then
|
|
11
|
-
cat << 'EOF'
|
|
12
|
-
Aztec Compile - Compile Aztec Noir contracts
|
|
13
|
-
|
|
14
|
-
This command compiles Aztec Noir contracts using nargo and then automatically
|
|
15
|
-
postprocesses them to generate Aztec specific artifacts including:
|
|
16
|
-
- Transpiled contract artifacts
|
|
17
|
-
- Verification keys
|
|
18
|
-
|
|
19
|
-
The compiled contracts will be placed in the target/ directory by default.
|
|
20
|
-
|
|
21
|
-
---
|
|
22
|
-
Underlying nargo compile options:
|
|
23
|
-
|
|
24
|
-
EOF
|
|
25
|
-
nargo compile --help
|
|
26
|
-
exit 0
|
|
27
|
-
fi
|
|
28
|
-
done
|
|
29
|
-
|
|
30
|
-
# Run nargo compile.
|
|
31
|
-
$NARGO compile "$@"
|
|
32
|
-
|
|
33
|
-
echo "Postprocessing contract..."
|
|
34
|
-
$BB aztec_process
|
|
35
|
-
|
|
36
|
-
# Strip internal prefixes from all compiled contract JSONs in target directory
|
|
37
|
-
# TODO: This should be part of bb aztec_process!
|
|
38
|
-
for json in target/*.json; do
|
|
39
|
-
temp_file="${json}.tmp"
|
|
40
|
-
jq '.functions |= map(.name |= sub("^__aztec_nr_internals__"; ""))' "$json" > "$temp_file"
|
|
41
|
-
mv "$temp_file" "$json"
|
|
42
|
-
done
|
|
43
|
-
|
|
44
|
-
echo "Compilation complete!"
|
|
@@ -1,47 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/env node
|
|
2
|
-
import fs from 'fs/promises';
|
|
3
|
-
import path from 'path';
|
|
4
|
-
|
|
5
|
-
// Simple script to extract a contract function as a separate Noir artifact.
|
|
6
|
-
// We need to use this since the transpiling that we do on public functions make the contract artifacts
|
|
7
|
-
// unreadable by noir tooling, since they are no longer following the noir artifact format.
|
|
8
|
-
async function main() {
|
|
9
|
-
let [contractArtifactPath, functionName] = process.argv.slice(2);
|
|
10
|
-
if (!contractArtifactPath || !functionName) {
|
|
11
|
-
console.log('Usage: node extractFunctionAsNoirArtifact.js <contractArtifactPath> <functionName>');
|
|
12
|
-
return;
|
|
13
|
-
}
|
|
14
|
-
|
|
15
|
-
const contractArtifact = JSON.parse(await fs.readFile(contractArtifactPath, 'utf8'));
|
|
16
|
-
const func = contractArtifact.functions.find(f => f.name === functionName);
|
|
17
|
-
if (!func) {
|
|
18
|
-
console.error(`Function ${functionName} not found in ${contractArtifactPath}`);
|
|
19
|
-
return;
|
|
20
|
-
}
|
|
21
|
-
|
|
22
|
-
const artifact = {
|
|
23
|
-
noir_version: contractArtifact.noir_version,
|
|
24
|
-
hash: 0,
|
|
25
|
-
abi: func.abi,
|
|
26
|
-
bytecode: func.bytecode,
|
|
27
|
-
debug_symbols: func.debug_symbols,
|
|
28
|
-
file_map: contractArtifact.file_map,
|
|
29
|
-
expression_width: {
|
|
30
|
-
Bounded: {
|
|
31
|
-
width: 4,
|
|
32
|
-
},
|
|
33
|
-
},
|
|
34
|
-
};
|
|
35
|
-
|
|
36
|
-
const outputDir = path.dirname(contractArtifactPath);
|
|
37
|
-
const outputName = path.basename(contractArtifactPath, '.json') + `-${functionName}.json`;
|
|
38
|
-
|
|
39
|
-
const outPath = path.join(outputDir, outputName);
|
|
40
|
-
|
|
41
|
-
await fs.writeFile(outPath, JSON.stringify(artifact, null, 2));
|
|
42
|
-
}
|
|
43
|
-
|
|
44
|
-
main().catch(err => {
|
|
45
|
-
console.error(err);
|
|
46
|
-
process.exit(1);
|
|
47
|
-
});
|
package/scripts/flamegraph.sh
DELETED
|
@@ -1,59 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/env bash
|
|
2
|
-
set -eu
|
|
3
|
-
|
|
4
|
-
# If first arg is -h or --help, print usage.
|
|
5
|
-
if [ $# -lt 2 ] || [ "$1" == "-h" ] || [ "$1" == "--help" ]; then
|
|
6
|
-
cat << 'EOF'
|
|
7
|
-
Aztec Flamegraph - Generate a gate count flamegraph for an aztec contract function.
|
|
8
|
-
|
|
9
|
-
Usage: aztec flamegraph <contract_artifact> <function>
|
|
10
|
-
|
|
11
|
-
Options:
|
|
12
|
-
-h, --help Print help
|
|
13
|
-
|
|
14
|
-
Will output an svg at <artifact_path>/<contract>-<function>-flamegraph.svg.
|
|
15
|
-
You can open it in your browser to view it.
|
|
16
|
-
|
|
17
|
-
EOF
|
|
18
|
-
exit 0
|
|
19
|
-
fi
|
|
20
|
-
|
|
21
|
-
cleanup() {
|
|
22
|
-
set +e
|
|
23
|
-
if [ -f "$function_artifact" ]; then
|
|
24
|
-
rm -f "$function_artifact"
|
|
25
|
-
fi
|
|
26
|
-
}
|
|
27
|
-
|
|
28
|
-
trap cleanup EXIT
|
|
29
|
-
|
|
30
|
-
# Get the directory of the script
|
|
31
|
-
script_dir=$(realpath $(dirname $0))
|
|
32
|
-
|
|
33
|
-
PROFILER=${PROFILER_PATH:-noir-profiler}
|
|
34
|
-
BB=${BB:-bb}
|
|
35
|
-
|
|
36
|
-
# first console arg is contract name in camel case or path to contract artifact
|
|
37
|
-
contract=$1
|
|
38
|
-
|
|
39
|
-
# second console arg is the contract function
|
|
40
|
-
function=$2
|
|
41
|
-
|
|
42
|
-
if [ ! -f "$contract" ]; then
|
|
43
|
-
echo "Error: Contract artifact not found at: $contract"
|
|
44
|
-
exit 1
|
|
45
|
-
fi
|
|
46
|
-
artifact_path=$contract
|
|
47
|
-
function_artifact="${artifact_path%%.json}-${function}.json"
|
|
48
|
-
output_dir=$(dirname "$artifact_path")
|
|
49
|
-
|
|
50
|
-
# Extract artifact for the specific function.
|
|
51
|
-
node $script_dir/extract_function.js "$artifact_path" $function
|
|
52
|
-
|
|
53
|
-
# Generate the flamegraph
|
|
54
|
-
$PROFILER gates --artifact-path "$function_artifact" --backend-path "$BB" --backend-gates-command "gates" --output "$output_dir" --scheme chonk --include_gates_per_opcode
|
|
55
|
-
|
|
56
|
-
# Save as $artifact_name-$function-flamegraph.svg
|
|
57
|
-
output_file="${function_artifact%%.json}-flamegraph.svg"
|
|
58
|
-
mv "$output_dir/__aztec_nr_internals__${function}_gates.svg" "$output_file"
|
|
59
|
-
echo "Flamegraph generated at: $output_file"
|
package/scripts/setup_project.sh
DELETED
|
@@ -1,31 +0,0 @@
|
|
|
1
|
-
#!/usr/bin/env bash
|
|
2
|
-
set -euo pipefail
|
|
3
|
-
|
|
4
|
-
# Get the actual aztec version for the git tag.
|
|
5
|
-
AZTEC_VERSION=$(jq -r '.version' $(dirname $0)/../package.json)
|
|
6
|
-
NARGO_TOML_PATH="Nargo.toml"
|
|
7
|
-
MAIN_NR_PATH="src/main.nr"
|
|
8
|
-
|
|
9
|
-
if [ ! -f "$NARGO_TOML_PATH" ]; then
|
|
10
|
-
>&2 echo "Warning: Could not find Nargo.toml at $NARGO_TOML_PATH to add aztec dependency"
|
|
11
|
-
exit 1
|
|
12
|
-
fi
|
|
13
|
-
|
|
14
|
-
if [ ! -f "$MAIN_NR_PATH" ]; then
|
|
15
|
-
>&2 echo "Warning: Could not find main.nr at $MAIN_NR_PATH"
|
|
16
|
-
exit 1
|
|
17
|
-
fi
|
|
18
|
-
|
|
19
|
-
# Add aztec dependency to Nargo.toml
|
|
20
|
-
echo "" >> "$NARGO_TOML_PATH"
|
|
21
|
-
echo "aztec = { git=\"https://github.com/AztecProtocol/aztec-nr\", tag=\"v${AZTEC_VERSION}\", directory=\"aztec\" }" >> "$NARGO_TOML_PATH"
|
|
22
|
-
echo "Added aztec dependency (v${AZTEC_VERSION}) to Nargo.toml"
|
|
23
|
-
|
|
24
|
-
# Replace the contents of main.nr with the Aztec contract template
|
|
25
|
-
cat > "$MAIN_NR_PATH" << 'EOF'
|
|
26
|
-
use aztec::macros::aztec;
|
|
27
|
-
|
|
28
|
-
#[aztec]
|
|
29
|
-
contract Main {}
|
|
30
|
-
EOF
|
|
31
|
-
echo "Created main.nr with Aztec contract template"
|