@aztec/foundation 0.0.1-commit.6c91f13 → 0.0.1-commit.6d63667d

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (152) hide show
  1. package/dest/branded-types/block_number.d.ts +4 -1
  2. package/dest/branded-types/block_number.d.ts.map +1 -1
  3. package/dest/branded-types/block_number.js +3 -0
  4. package/dest/branded-types/checkpoint_number.d.ts +10 -2
  5. package/dest/branded-types/checkpoint_number.d.ts.map +1 -1
  6. package/dest/branded-types/checkpoint_number.js +17 -8
  7. package/dest/branded-types/index.d.ts +3 -2
  8. package/dest/branded-types/index.d.ts.map +1 -1
  9. package/dest/branded-types/index.js +2 -1
  10. package/dest/branded-types/index_within_checkpoint.d.ts +42 -0
  11. package/dest/branded-types/index_within_checkpoint.d.ts.map +1 -0
  12. package/dest/branded-types/index_within_checkpoint.js +59 -0
  13. package/dest/buffer/buffer16.d.ts +4 -1
  14. package/dest/buffer/buffer16.d.ts.map +1 -1
  15. package/dest/buffer/buffer16.js +3 -1
  16. package/dest/buffer/buffer32.d.ts +4 -1
  17. package/dest/buffer/buffer32.d.ts.map +1 -1
  18. package/dest/buffer/buffer32.js +3 -1
  19. package/dest/collection/array.d.ts +12 -1
  20. package/dest/collection/array.d.ts.map +1 -1
  21. package/dest/collection/array.js +51 -0
  22. package/dest/config/env_var.d.ts +2 -2
  23. package/dest/config/env_var.d.ts.map +1 -1
  24. package/dest/config/index.d.ts +6 -3
  25. package/dest/config/index.d.ts.map +1 -1
  26. package/dest/config/index.js +16 -7
  27. package/dest/config/parse-env.d.ts +3 -0
  28. package/dest/config/parse-env.d.ts.map +1 -0
  29. package/dest/config/parse-env.js +7 -0
  30. package/dest/config/secret_value.js +3 -1
  31. package/dest/crypto/ecdsa/signature.d.ts +10 -1
  32. package/dest/crypto/ecdsa/signature.d.ts.map +1 -1
  33. package/dest/crypto/poseidon/index.d.ts +1 -2
  34. package/dest/crypto/poseidon/index.d.ts.map +1 -1
  35. package/dest/crypto/poseidon/index.js +0 -9
  36. package/dest/crypto/random/randomness_singleton.d.ts +4 -3
  37. package/dest/crypto/random/randomness_singleton.d.ts.map +1 -1
  38. package/dest/crypto/random/randomness_singleton.js +5 -5
  39. package/dest/crypto/sync/poseidon/index.d.ts +1 -2
  40. package/dest/crypto/sync/poseidon/index.d.ts.map +1 -1
  41. package/dest/crypto/sync/poseidon/index.js +0 -8
  42. package/dest/curves/bls12/field.js +6 -3
  43. package/dest/curves/bls12/point.d.ts +10 -1
  44. package/dest/curves/bls12/point.d.ts.map +1 -1
  45. package/dest/curves/bls12/point.js +3 -1
  46. package/dest/curves/bn254/field.js +5 -2
  47. package/dest/curves/grumpkin/point.d.ts +11 -2
  48. package/dest/curves/grumpkin/point.d.ts.map +1 -1
  49. package/dest/error/index.d.ts +4 -4
  50. package/dest/error/index.d.ts.map +1 -1
  51. package/dest/eth-address/index.js +4 -2
  52. package/dest/eth-signature/eth_signature.d.ts +4 -1
  53. package/dest/eth-signature/eth_signature.d.ts.map +1 -1
  54. package/dest/jest/setup.js +4 -1
  55. package/dest/json-rpc/client/undici.d.ts +1 -1
  56. package/dest/json-rpc/client/undici.d.ts.map +1 -1
  57. package/dest/json-rpc/client/undici.js +21 -4
  58. package/dest/json-rpc/fixtures/class_a.d.ts +3 -3
  59. package/dest/json-rpc/fixtures/class_a.d.ts.map +1 -1
  60. package/dest/json-rpc/fixtures/class_b.d.ts +3 -3
  61. package/dest/json-rpc/fixtures/class_b.d.ts.map +1 -1
  62. package/dest/json-rpc/server/safe_json_rpc_server.js +1 -1
  63. package/dest/log/libp2p_logger.d.ts +5 -2
  64. package/dest/log/libp2p_logger.d.ts.map +1 -1
  65. package/dest/log/libp2p_logger.js +14 -4
  66. package/dest/log/pino-logger-server.d.ts +9 -0
  67. package/dest/log/pino-logger-server.d.ts.map +1 -0
  68. package/dest/log/pino-logger-server.js +18 -0
  69. package/dest/log/pino-logger.d.ts +37 -8
  70. package/dest/log/pino-logger.d.ts.map +1 -1
  71. package/dest/log/pino-logger.js +116 -27
  72. package/dest/queue/semaphore.d.ts +5 -1
  73. package/dest/queue/semaphore.d.ts.map +1 -1
  74. package/dest/retry/index.d.ts +11 -1
  75. package/dest/retry/index.d.ts.map +1 -1
  76. package/dest/retry/index.js +11 -0
  77. package/dest/serialize/buffer_reader.d.ts +7 -4
  78. package/dest/serialize/buffer_reader.d.ts.map +1 -1
  79. package/dest/serialize/buffer_reader.js +13 -4
  80. package/dest/timer/date.d.ts +23 -1
  81. package/dest/timer/date.d.ts.map +1 -1
  82. package/dest/timer/date.js +29 -0
  83. package/dest/trees/balanced_merkle_tree_root.d.ts +17 -0
  84. package/dest/trees/balanced_merkle_tree_root.d.ts.map +1 -0
  85. package/dest/trees/{balanced_merkle_tree.js → balanced_merkle_tree_root.js} +2 -17
  86. package/dest/trees/hasher.d.ts +3 -1
  87. package/dest/trees/hasher.d.ts.map +1 -1
  88. package/dest/trees/hasher.js +10 -5
  89. package/dest/trees/index.d.ts +4 -4
  90. package/dest/trees/index.d.ts.map +1 -1
  91. package/dest/trees/index.js +3 -3
  92. package/dest/trees/membership_witness.d.ts +11 -1
  93. package/dest/trees/membership_witness.d.ts.map +1 -1
  94. package/dest/trees/membership_witness.js +9 -0
  95. package/dest/trees/merkle_tree_calculator.d.ts +1 -1
  96. package/dest/trees/merkle_tree_calculator.d.ts.map +1 -1
  97. package/dest/trees/merkle_tree_calculator.js +2 -2
  98. package/dest/trees/sibling_path.d.ts +2 -1
  99. package/dest/trees/sibling_path.d.ts.map +1 -1
  100. package/dest/trees/sibling_path.js +2 -2
  101. package/dest/trees/unbalanced_merkle_tree_calculator.d.ts +2 -3
  102. package/dest/trees/unbalanced_merkle_tree_calculator.d.ts.map +1 -1
  103. package/dest/trees/unbalanced_merkle_tree_calculator.js +1 -5
  104. package/dest/trees/{unbalanced_merkle_tree.d.ts → unbalanced_merkle_tree_root.d.ts} +3 -5
  105. package/dest/trees/unbalanced_merkle_tree_root.d.ts.map +1 -0
  106. package/dest/trees/{unbalanced_merkle_tree.js → unbalanced_merkle_tree_root.js} +11 -49
  107. package/dest/trees/unbalanced_tree_store.d.ts +5 -1
  108. package/dest/trees/unbalanced_tree_store.d.ts.map +1 -1
  109. package/dest/trees/unbalanced_tree_store.js +49 -1
  110. package/dest/types/index.d.ts +4 -2
  111. package/dest/types/index.d.ts.map +1 -1
  112. package/dest/types/index.js +3 -0
  113. package/package.json +25 -4
  114. package/src/branded-types/block_number.ts +5 -0
  115. package/src/branded-types/checkpoint_number.ts +22 -7
  116. package/src/branded-types/index.ts +2 -1
  117. package/src/branded-types/index_within_checkpoint.ts +88 -0
  118. package/src/collection/array.ts +52 -0
  119. package/src/config/env_var.ts +32 -7
  120. package/src/config/index.ts +19 -4
  121. package/src/config/parse-env.ts +4 -0
  122. package/src/crypto/poseidon/index.ts +0 -10
  123. package/src/crypto/random/randomness_singleton.ts +6 -4
  124. package/src/crypto/sync/poseidon/index.ts +0 -9
  125. package/src/eth-address/index.ts +1 -1
  126. package/src/jest/setup.mjs +4 -1
  127. package/src/json-rpc/client/undici.ts +21 -3
  128. package/src/json-rpc/server/safe_json_rpc_server.ts +1 -1
  129. package/src/log/libp2p_logger.ts +12 -5
  130. package/src/log/pino-logger-server.ts +25 -0
  131. package/src/log/pino-logger.ts +138 -37
  132. package/src/queue/semaphore.ts +5 -0
  133. package/src/retry/index.ts +18 -0
  134. package/src/serialize/buffer_reader.ts +21 -9
  135. package/src/timer/date.ts +42 -0
  136. package/src/trees/{balanced_merkle_tree.ts → balanced_merkle_tree_root.ts} +2 -18
  137. package/src/trees/hasher.ts +9 -0
  138. package/src/trees/index.ts +3 -3
  139. package/src/trees/membership_witness.ts +8 -0
  140. package/src/trees/merkle_tree_calculator.ts +2 -2
  141. package/src/trees/sibling_path.ts +2 -2
  142. package/src/trees/unbalanced_merkle_tree_calculator.ts +1 -12
  143. package/src/trees/{unbalanced_merkle_tree.ts → unbalanced_merkle_tree_root.ts} +17 -61
  144. package/src/trees/unbalanced_tree_store.ts +57 -2
  145. package/src/types/index.ts +6 -1
  146. package/dest/json-rpc/server/telemetry.d.ts +0 -2
  147. package/dest/json-rpc/server/telemetry.d.ts.map +0 -1
  148. package/dest/json-rpc/server/telemetry.js +0 -0
  149. package/dest/trees/balanced_merkle_tree.d.ts +0 -22
  150. package/dest/trees/balanced_merkle_tree.d.ts.map +0 -1
  151. package/dest/trees/unbalanced_merkle_tree.d.ts.map +0 -1
  152. package/src/json-rpc/server/telemetry.ts +0 -0
@@ -35,16 +35,6 @@ export async function poseidon2HashWithSeparator(input: Fieldable[], separator:
35
35
  return Fr.fromBuffer(Buffer.from(response.hash));
36
36
  }
37
37
 
38
- export async function poseidon2HashAccumulate(input: Fieldable[]): Promise<Fr> {
39
- const inputFields = serializeToFields(input);
40
- await BarretenbergSync.initSingleton();
41
- const api = BarretenbergSync.getSingleton();
42
- const response = api.poseidon2HashAccumulate({
43
- inputs: inputFields.map(i => i.toBuffer()),
44
- });
45
- return Fr.fromBuffer(Buffer.from(response.hash));
46
- }
47
-
48
38
  /**
49
39
  * Runs a Poseidon2 permutation.
50
40
  * @param input the input state. Expected to be of size 4.
@@ -1,4 +1,4 @@
1
- import { createLogger } from '../../log/pino-logger.js';
1
+ import { type Logger, type LoggerBindings, createLogger } from '../../log/pino-logger.js';
2
2
 
3
3
  /**
4
4
  * A number generator which is used as a source of randomness in the system. If the SEED env variable is set, the
@@ -12,11 +12,13 @@ export class RandomnessSingleton {
12
12
  private static instance: RandomnessSingleton;
13
13
 
14
14
  private counter = 0;
15
+ private log: Logger;
15
16
 
16
17
  private constructor(
17
18
  private readonly seed?: number,
18
- private readonly log = createLogger('foundation:randomness_singleton'),
19
+ bindings?: LoggerBindings,
19
20
  ) {
21
+ this.log = createLogger('foundation:randomness_singleton', bindings);
20
22
  if (seed !== undefined) {
21
23
  this.log.debug(`Using pseudo-randomness with seed: ${seed}`);
22
24
  this.counter = seed;
@@ -25,10 +27,10 @@ export class RandomnessSingleton {
25
27
  }
26
28
  }
27
29
 
28
- public static getInstance(): RandomnessSingleton {
30
+ public static getInstance(bindings?: LoggerBindings): RandomnessSingleton {
29
31
  if (!RandomnessSingleton.instance) {
30
32
  const seed = process.env.SEED ? Number(process.env.SEED) : undefined;
31
- RandomnessSingleton.instance = new RandomnessSingleton(seed);
33
+ RandomnessSingleton.instance = new RandomnessSingleton(seed, bindings);
32
34
  }
33
35
 
34
36
  return RandomnessSingleton.instance;
@@ -34,15 +34,6 @@ export function poseidon2HashWithSeparator(input: Fieldable[], separator: number
34
34
  return Fr.fromBuffer(Buffer.from(response.hash));
35
35
  }
36
36
 
37
- export function poseidon2HashAccumulate(input: Fieldable[]): Fr {
38
- const inputFields = serializeToFields(input);
39
- const api = BarretenbergSync.getSingleton();
40
- const response = api.poseidon2HashAccumulate({
41
- inputs: inputFields.map(i => i.toBuffer()),
42
- });
43
- return Fr.fromBuffer(Buffer.from(response.hash));
44
- }
45
-
46
37
  /**
47
38
  * Runs a Poseidon2 permutation.
48
39
  * @param input the input state. Expected to be of size 4.
@@ -249,7 +249,7 @@ export class EthAddress {
249
249
  /** Converts a number into an address. Useful for testing. */
250
250
  static fromNumber(num: bigint | number): EthAddress {
251
251
  const buffer = Buffer.alloc(EthAddress.SIZE_IN_BYTES);
252
- buffer.writeBigUInt64BE(BigInt(num), 0);
252
+ buffer.writeBigUInt64BE(BigInt(num), EthAddress.SIZE_IN_BYTES - 8);
253
253
  return new EthAddress(buffer);
254
254
  }
255
255
 
@@ -1,3 +1,4 @@
1
+ import { parseBooleanEnv } from '@aztec/foundation/config';
1
2
  import { overwriteLoggingStream, pinoPrettyOpts } from '@aztec/foundation/log';
2
3
 
3
4
  import pretty from 'pino-pretty';
@@ -6,4 +7,6 @@ import pretty from 'pino-pretty';
6
7
  // file so we don't mess up with dependencies in non-testing environments,
7
8
  // since pino-pretty messes up with browser bundles.
8
9
  // See also https://www.npmjs.com/package/pino-pretty?activeTab=readme#user-content-usage-with-jest
9
- overwriteLoggingStream(pretty(pinoPrettyOpts));
10
+ if (!parseBooleanEnv(process.env.LOG_JSON)) {
11
+ overwriteLoggingStream(pretty(pinoPrettyOpts));
12
+ }
@@ -1,3 +1,5 @@
1
+ import { promisify } from 'node:util';
2
+ import { gunzip as gunzipCb, gzip as gzipCb } from 'node:zlib';
1
3
  import { Agent, type Dispatcher } from 'undici';
2
4
 
3
5
  import { createLogger } from '../../log/pino-logger.js';
@@ -5,8 +7,14 @@ import { NoRetryError } from '../../retry/index.js';
5
7
  import { jsonStringify } from '../convert.js';
6
8
  import type { JsonRpcFetch } from './fetch.js';
7
9
 
10
+ const gzip = promisify(gzipCb);
11
+ const gunzip = promisify(gunzipCb);
12
+
8
13
  const log = createLogger('json-rpc:json_rpc_client:undici');
9
14
 
15
+ /** Minimum request size in bytes to trigger compression. */
16
+ const COMPRESSION_THRESHOLD = 1024;
17
+
10
18
  export { Agent };
11
19
 
12
20
  export function makeUndiciFetch(client = new Agent()): JsonRpcFetch {
@@ -14,14 +22,18 @@ export function makeUndiciFetch(client = new Agent()): JsonRpcFetch {
14
22
  log.trace(`JsonRpcClient.fetch: ${host}`, { host, body });
15
23
  let resp: Dispatcher.ResponseData;
16
24
  try {
25
+ const jsonBody = Buffer.from(jsonStringify(body));
26
+ const shouldCompress = jsonBody.length >= COMPRESSION_THRESHOLD;
17
27
  resp = await client.request({
18
28
  method: 'POST',
19
29
  origin: new URL(host),
20
30
  path: '/',
21
- body: jsonStringify(body),
31
+ body: shouldCompress ? await gzip(jsonBody) : jsonBody,
22
32
  headers: {
23
33
  ...extraHeaders,
24
34
  'content-type': 'application/json',
35
+ ...(shouldCompress && { 'content-encoding': 'gzip' }),
36
+ 'accept-encoding': 'gzip',
25
37
  },
26
38
  });
27
39
  } catch (err) {
@@ -31,13 +43,19 @@ export function makeUndiciFetch(client = new Agent()): JsonRpcFetch {
31
43
 
32
44
  let responseJson: any;
33
45
  const responseOk = resp.statusCode >= 200 && resp.statusCode <= 299;
46
+ const contentEncoding = resp.headers['content-encoding'];
34
47
  try {
35
- responseJson = await resp.body.json();
48
+ if (contentEncoding === 'gzip') {
49
+ const jsonBuffer = await gunzip(await resp.body.arrayBuffer());
50
+ responseJson = JSON.parse(jsonBuffer.toString('utf-8'));
51
+ } else {
52
+ responseJson = await resp.body.json();
53
+ }
36
54
  } catch {
37
55
  if (!responseOk) {
38
56
  throw new Error('HTTP ' + resp.statusCode);
39
57
  }
40
- throw new Error(`Failed to parse body as JSON: ${await resp.body.text()}`);
58
+ throw new Error(`Failed to parse body as JSON. encoding: ${contentEncoding}, body: ${await resp.body.text()}`);
41
59
  }
42
60
 
43
61
  if (!responseOk) {
@@ -35,7 +35,7 @@ export type SafeJsonRpcServerConfig = {
35
35
  const defaultServerConfig: SafeJsonRpcServerConfig = {
36
36
  http200OnError: false,
37
37
  maxBatchSize: 100,
38
- maxBodySizeBytes: '50mb',
38
+ maxBodySizeBytes: '1mb',
39
39
  };
40
40
 
41
41
  export class SafeJsonRpcServer {
@@ -2,15 +2,17 @@ import type { ComponentLogger, Logger } from '@libp2p/interface';
2
2
 
3
3
  import { getLogLevelFromFilters } from './log-filters.js';
4
4
  import type { LogLevel } from './log-levels.js';
5
- import { logFilters, logger } from './pino-logger.js';
5
+ import { type LoggerBindings, logFilters, logger } from './pino-logger.js';
6
6
 
7
7
  /**
8
8
  * Creates a libp2p compatible logger that wraps our pino logger.
9
9
  * This adapter implements the ComponentLogger interface required by libp2p.
10
+ * @param namespace - Base namespace for the logger
11
+ * @param bindings - Optional bindings to pass to the logger (actor, instanceId)
10
12
  */
11
- export function createLibp2pComponentLogger(namespace: string): ComponentLogger {
13
+ export function createLibp2pComponentLogger(namespace: string, bindings?: LoggerBindings): ComponentLogger {
12
14
  return {
13
- forComponent: (component: string) => createLibp2pLogger(`${namespace}:${component}`),
15
+ forComponent: (component: string) => createLibp2pLogger(`${namespace}:${component}`, bindings),
14
16
  };
15
17
  }
16
18
 
@@ -24,9 +26,14 @@ function replaceFormatting(message: string) {
24
26
  return message.replace(/(%p|%a)/g, '%s');
25
27
  }
26
28
 
27
- function createLibp2pLogger(component: string): Logger {
29
+ function createLibp2pLogger(component: string, bindings?: LoggerBindings): Logger {
28
30
  // Create a direct pino logger instance for libp2p that supports string interpolation
29
- const log = logger.child({ module: component }, { level: getLogLevelFromFilters(logFilters, component) });
31
+ const actor = bindings?.actor;
32
+ const instanceId = bindings?.instanceId;
33
+ const log = logger.child(
34
+ { module: component, ...(actor && { actor }), ...(instanceId && { instanceId }) },
35
+ { level: getLogLevelFromFilters(logFilters, component) },
36
+ );
30
37
 
31
38
  const logIfEnabled = (level: LogLevel, message: string, ...args: unknown[]) => {
32
39
  if (!log.isLevelEnabled(level)) {
@@ -0,0 +1,25 @@
1
+ import { AsyncLocalStorage } from 'node:async_hooks';
2
+
3
+ import { type LoggerBindings, addLogBindingsHandler, removeLogBindingsHandler } from './pino-logger.js';
4
+
5
+ /** AsyncLocalStorage for logger bindings context propagation (Node.js only). */
6
+ const bindingsStorage = new AsyncLocalStorage<LoggerBindings>();
7
+
8
+ /** Returns the current bindings from AsyncLocalStorage, if any. */
9
+ export function getBindings(): LoggerBindings | undefined {
10
+ return bindingsStorage.getStore();
11
+ }
12
+
13
+ /**
14
+ * Runs a callback within a bindings context. All loggers created within the callback
15
+ * will automatically inherit the bindings (actor, instanceId) via the log bindings handler.
16
+ */
17
+ export async function withLoggerBindings<T>(bindings: LoggerBindings, callback: () => Promise<T>): Promise<T> {
18
+ const handler = () => bindingsStorage.getStore();
19
+ addLogBindingsHandler(handler);
20
+ try {
21
+ return await bindingsStorage.run(bindings, callback);
22
+ } finally {
23
+ removeLogBindingsHandler(handler);
24
+ }
25
+ }
@@ -1,19 +1,62 @@
1
- import { createColors, isColorSupported } from 'colorette';
1
+ import { type Color, createColors, isColorSupported } from 'colorette';
2
2
  import isNode from 'detect-node';
3
3
  import { pino, symbols } from 'pino';
4
4
  import type { Writable } from 'stream';
5
5
  import { inspect } from 'util';
6
6
 
7
7
  import { compactArray } from '../collection/array.js';
8
- import { type EnvVar, parseBooleanEnv } from '../config/index.js';
8
+ import type { EnvVar } from '../config/index.js';
9
+ import { parseBooleanEnv } from '../config/parse-env.js';
9
10
  import { GoogleCloudLoggerConfig } from './gcloud-logger-config.js';
10
11
  import { getLogLevelFromFilters, parseEnv } from './log-filters.js';
11
12
  import type { LogLevel } from './log-levels.js';
12
13
  import type { LogData, LogFn } from './log_fn.js';
13
14
 
14
- export function createLogger(module: string): Logger {
15
- module = logNameHandlers.reduce((moduleName, handler) => handler(moduleName), module.replace(/^aztec:/, ''));
16
- const pinoLogger = logger.child({ module }, { level: getLogLevelFromFilters(logFilters, module) });
15
+ /** Optional bindings to pass to createLogger for additional context. */
16
+ export type LoggerBindings = {
17
+ /** Actor label shown in logs (e.g., 'MAIN', 'prover-node'). */
18
+ actor?: string;
19
+ /** Instance identifier for distinguishing multiple instances of the same component. */
20
+ instanceId?: string;
21
+ };
22
+
23
+ // Allow global hooks for providing default bindings.
24
+ // Used by withLoggerBindings in pino-logger-server to propagate bindings via AsyncLocalStorage.
25
+ type LogBindingsHandler = () => LoggerBindings | undefined;
26
+ const logBindingsHandlers: LogBindingsHandler[] = [];
27
+
28
+ export function addLogBindingsHandler(handler: LogBindingsHandler): void {
29
+ logBindingsHandlers.push(handler);
30
+ }
31
+
32
+ export function removeLogBindingsHandler(handler: LogBindingsHandler) {
33
+ const index = logBindingsHandlers.indexOf(handler);
34
+ if (index !== -1) {
35
+ logBindingsHandlers.splice(index, 1);
36
+ }
37
+ }
38
+
39
+ function getBindingsFromHandlers(): LoggerBindings | undefined {
40
+ for (const handler of logBindingsHandlers) {
41
+ const bindings = handler();
42
+ if (bindings) {
43
+ return bindings;
44
+ }
45
+ }
46
+ return undefined;
47
+ }
48
+
49
+ export function createLogger(module: string, bindings?: LoggerBindings): Logger {
50
+ module = module.replace(/^aztec:/, '');
51
+
52
+ const resolvedBindings = { ...getBindingsFromHandlers(), ...bindings };
53
+ const actor = resolvedBindings?.actor;
54
+ const instanceId = resolvedBindings?.instanceId;
55
+
56
+ const pinoLogger = logger.child(
57
+ { module, ...(actor && { actor }), ...(instanceId && { instanceId }) },
58
+ { level: getLogLevelFromFilters(logFilters, module) },
59
+ );
17
60
 
18
61
  // We check manually for isLevelEnabled to avoid calling processLogData unnecessarily.
19
62
  // Note that isLevelEnabled is missing from the browser version of pino.
@@ -43,11 +86,24 @@ export function createLogger(module: string): Logger {
43
86
  isLevelEnabled: (level: LogLevel) => isLevelEnabled(pinoLogger, level),
44
87
  /** Module name for the logger. */
45
88
  module,
46
- /** Creates another logger by extending this logger module name. */
47
- createChild: (childModule: string) => createLogger(`${module}:${childModule}`),
89
+ /** Creates another logger by extending this logger module name and preserving bindings. */
90
+ createChild: (childModule: string) => createLogger(`${module}:${childModule}`, { actor, instanceId }),
91
+ /** Returns the bindings (actor, instanceId) for this logger. */
92
+ getBindings: () => ({ actor, instanceId }),
48
93
  };
49
94
  }
50
95
 
96
+ /**
97
+ * Returns a logger for the given module. If loggerOrBindings is already a Logger, returns it directly.
98
+ * Otherwise, creates a new logger with the given module name and bindings.
99
+ */
100
+ export function resolveLogger(module: string, loggerOrBindings?: Logger | LoggerBindings): Logger {
101
+ if (loggerOrBindings && 'info' in loggerOrBindings) {
102
+ return loggerOrBindings as Logger;
103
+ }
104
+ return createLogger(module, loggerOrBindings);
105
+ }
106
+
51
107
  // Allow global hooks for processing log data.
52
108
  // Used for injecting OTEL trace_id in telemetry client.
53
109
  type LogDataHandler = (data: LogData) => LogData;
@@ -61,31 +117,6 @@ function processLogData(data: LogData): LogData {
61
117
  return logDataHandlers.reduce((accum, handler) => handler(accum), data);
62
118
  }
63
119
 
64
- // Allow global hooks for tweaking module names.
65
- // Used in tests to add a uid to modules, so we can differentiate multiple nodes in the same process.
66
- type LogNameHandler = (module: string) => string;
67
- const logNameHandlers: LogNameHandler[] = [];
68
-
69
- export function addLogNameHandler(handler: LogNameHandler): void {
70
- logNameHandlers.push(handler);
71
- }
72
-
73
- export function removeLogNameHandler(handler: LogNameHandler) {
74
- const index = logNameHandlers.indexOf(handler);
75
- if (index !== -1) {
76
- logNameHandlers.splice(index, 1);
77
- }
78
- }
79
-
80
- /** Creates all loggers within the given callback with the suffix appended to the module name. */
81
- export async function withLogNameSuffix<T>(suffix: string, callback: () => Promise<T>): Promise<T> {
82
- const logNameHandler = (module: string) => `${module}:${suffix}`;
83
- addLogNameHandler(logNameHandler);
84
- const result = await callback();
85
- removeLogNameHandler(logNameHandler);
86
- return result;
87
- }
88
-
89
120
  // Patch isLevelEnabled missing from pino/browser.
90
121
  function isLevelEnabled(logger: pino.Logger<'verbose', boolean>, level: LogLevel): boolean {
91
122
  return typeof logger.isLevelEnabled === 'function'
@@ -145,22 +176,90 @@ export const levels = {
145
176
  // Transport options for pretty logging to stderr via pino-pretty.
146
177
  const colorEnv = process.env['FORCE_COLOR' satisfies EnvVar];
147
178
  const useColor = colorEnv === undefined ? isColorSupported : parseBooleanEnv(colorEnv);
148
- const { bold, reset } = createColors({ useColor });
149
- export const pinoPrettyOpts = {
179
+ const { bold, reset, cyan, magenta, yellow, blue, green, magentaBright, yellowBright, blueBright, greenBright } =
180
+ createColors({ useColor });
181
+
182
+ // Per-actor coloring: each unique actor gets a different color for easier visual distinction.
183
+ // Disabled when LOG_NO_COLOR_PER_ACTOR is set to a truthy value.
184
+ const useColorPerActor = useColor && !parseBooleanEnv(process.env['LOG_NO_COLOR_PER_ACTOR' satisfies EnvVar]);
185
+ const actorColors: Color[] = [yellow, magenta, blue, green, magentaBright, yellowBright, blueBright, greenBright];
186
+ const actorColorMap = new Map<string, Color>();
187
+ let nextColorIndex = 0;
188
+
189
+ /** Returns the color function assigned to a given actor, assigning a new one if needed. */
190
+ export function getActorColor(actor: string): Color {
191
+ let color = actorColorMap.get(actor);
192
+ if (!color) {
193
+ color = actorColors[nextColorIndex % actorColors.length];
194
+ actorColorMap.set(actor, color);
195
+ nextColorIndex++;
196
+ }
197
+ return color;
198
+ }
199
+
200
+ /** Resets the actor-to-color mapping. Useful for testing. */
201
+ export function resetActorColors(): void {
202
+ actorColorMap.clear();
203
+ nextColorIndex = 0;
204
+ }
205
+
206
+ // String template for messageFormat (used in worker threads and when per-actor coloring is disabled).
207
+ const messageFormatString = `${bold('{module}')}{if actor} ${cyan('{actor}')}{end}{if instanceId} ${reset(cyan('{instanceId}'))}{end} ${reset('{msg}')}`;
208
+
209
+ // Function for messageFormat when per-actor coloring is enabled (can only be used in-process, not worker threads).
210
+ type LogObject = { actor?: string; module?: string; instanceId?: string; msg?: string };
211
+
212
+ /** Formats a log message with per-actor coloring. Actor, module, and instanceId share the same color. */
213
+ export function formatLogMessage(log: LogObject, messageKey: string): string {
214
+ const actor = log.actor;
215
+ const module = log.module ?? '';
216
+ const instanceId = log.instanceId;
217
+ const msg = log[messageKey as keyof LogObject] ?? '';
218
+
219
+ // Use actor color for actor, module, and instanceId when actor is present
220
+ const color = actor ? getActorColor(actor) : cyan;
221
+
222
+ let result = bold(color(module));
223
+ if (actor) {
224
+ result += ' ' + color(actor);
225
+ }
226
+ if (instanceId) {
227
+ result += ' ' + reset(color(instanceId));
228
+ }
229
+ result += ' ' + reset(String(msg));
230
+ return result;
231
+ }
232
+
233
+ // Base options for pino-pretty (shared between transport and direct use).
234
+ const pinoPrettyBaseOpts = {
150
235
  destination: 2,
151
236
  sync: true,
152
237
  colorize: useColor,
153
- ignore: 'module,pid,hostname,trace_id,span_id,trace_flags,severity',
154
- messageFormat: `${bold('{module}')} ${reset('{msg}')}`,
238
+ ignore: 'module,actor,instanceId,pid,hostname,trace_id,span_id,trace_flags,severity',
155
239
  customLevels: 'fatal:60,error:50,warn:40,info:30,verbose:25,debug:20,trace:10',
156
240
  customColors: 'fatal:bgRed,error:red,warn:yellow,info:green,verbose:magenta,debug:blue,trace:gray',
157
241
  minimumLevel: 'trace' as const,
158
242
  singleLine: !parseBooleanEnv(process.env['LOG_MULTILINE' satisfies EnvVar]),
159
243
  };
160
244
 
245
+ /**
246
+ * Pino-pretty options for direct use (e.g., jest/setup.mjs).
247
+ * Includes function-based messageFormat for per-actor coloring when enabled.
248
+ */
249
+ export const pinoPrettyOpts = {
250
+ ...pinoPrettyBaseOpts,
251
+ messageFormat: useColorPerActor ? formatLogMessage : messageFormatString,
252
+ };
253
+
254
+ // Transport options use string template only (functions can't be serialized to worker threads).
255
+ const prettyTransportOpts = {
256
+ ...pinoPrettyBaseOpts,
257
+ messageFormat: messageFormatString,
258
+ };
259
+
161
260
  const prettyTransport: pino.TransportTargetOptions = {
162
261
  target: 'pino-pretty',
163
- options: pinoPrettyOpts,
262
+ options: prettyTransportOpts,
164
263
  level: 'trace',
165
264
  };
166
265
 
@@ -261,6 +360,8 @@ export type Logger = { [K in LogLevel]: LogFn } & { /** Error log function */ er
261
360
  isLevelEnabled: (level: LogLevel) => boolean;
262
361
  module: string;
263
362
  createChild: (childModule: string) => Logger;
363
+ /** Returns the bindings (actor, instanceId) for this logger. */
364
+ getBindings: () => LoggerBindings;
264
365
  };
265
366
 
266
367
  /**
@@ -1,5 +1,10 @@
1
1
  import { FifoMemoryQueue } from './fifo_memory_queue.js';
2
2
 
3
+ export interface ISemaphore {
4
+ acquire(): Promise<void>;
5
+ release(): void;
6
+ }
7
+
3
8
  /**
4
9
  * Allows the acquiring of up to `size` tokens before calls to acquire block, waiting for a call to release().
5
10
  */
@@ -103,3 +103,21 @@ export async function retryUntil<T>(
103
103
  }
104
104
  }
105
105
  }
106
+
107
+ /**
108
+ * Convenience wrapper around retryUntil with fast polling for tests.
109
+ * Uses 10s timeout and 100ms polling interval by default.
110
+ *
111
+ * @param fn - The function to retry until it returns a truthy value.
112
+ * @param name - Description of what we're waiting for (for error messages).
113
+ * @param timeout - Optional timeout in seconds. Defaults to 10s.
114
+ * @param interval - Optional interval in seconds. Defaults to 0.1s (100ms).
115
+ */
116
+ export function retryFastUntil<T>(
117
+ fn: () => (T | undefined) | Promise<T | undefined>,
118
+ name = '',
119
+ timeout = 10,
120
+ interval = 0.1,
121
+ ) {
122
+ return retryUntil(fn, name, timeout, interval);
123
+ }
@@ -224,15 +224,22 @@ export class BufferReader {
224
224
  * deserializing each one using the 'fromBuffer' method of 'itemDeserializer'.
225
225
  *
226
226
  * @param itemDeserializer - Object with 'fromBuffer' method to deserialize vector elements.
227
+ * @param maxSize - Optional maximum allowed size for the vector. If the size exceeds this, an error is thrown.
227
228
  * @returns An array of deserialized elements of type T.
228
229
  */
229
- public readVector<T>(itemDeserializer: {
230
- /**
231
- * A method to deserialize data from a buffer.
232
- */
233
- fromBuffer: (reader: BufferReader) => T;
234
- }): T[] {
230
+ public readVector<T>(
231
+ itemDeserializer: {
232
+ /**
233
+ * A method to deserialize data from a buffer.
234
+ */
235
+ fromBuffer: (reader: BufferReader) => T;
236
+ },
237
+ maxSize?: number,
238
+ ): T[] {
235
239
  const size = this.readNumber();
240
+ if (maxSize !== undefined && size > maxSize) {
241
+ throw new Error(`Vector size ${size} exceeds maximum allowed ${maxSize}`);
242
+ }
236
243
  const result = new Array<T>(size);
237
244
  for (let i = 0; i < size; i++) {
238
245
  result[i] = itemDeserializer.fromBuffer(this);
@@ -344,10 +351,11 @@ export class BufferReader {
344
351
  * The method first reads the size of the string, then reads the corresponding
345
352
  * number of bytes from the buffer and converts them to a string.
346
353
  *
354
+ * @param maxSize - Optional maximum allowed size for the string buffer. If the size exceeds this, an error is thrown.
347
355
  * @returns The read string from the buffer.
348
356
  */
349
- public readString(): string {
350
- return this.readBuffer().toString();
357
+ public readString(maxSize?: number): string {
358
+ return this.readBuffer(maxSize).toString();
351
359
  }
352
360
 
353
361
  /**
@@ -356,10 +364,14 @@ export class BufferReader {
356
364
  * a Buffer with that size containing the bytes. Useful for reading variable-length
357
365
  * binary data encoded as (size, data) format.
358
366
  *
367
+ * @param maxSize - Optional maximum allowed size for the buffer. If the size exceeds this, an error is thrown.
359
368
  * @returns A Buffer containing the read bytes.
360
369
  */
361
- public readBuffer(): Buffer {
370
+ public readBuffer(maxSize?: number): Buffer {
362
371
  const size = this.readNumber();
372
+ if (maxSize !== undefined && size > maxSize) {
373
+ throw new Error(`Buffer size ${size} exceeds maximum allowed ${maxSize}`);
374
+ }
363
375
  this.#rangeCheck(size);
364
376
  return this.readBytes(size);
365
377
  }
package/src/timer/date.ts CHANGED
@@ -31,4 +31,46 @@ export class TestDateProvider extends DateProvider {
31
31
  this.offset = timeMs - Date.now();
32
32
  this.logger.warn(`Time set to ${new Date(timeMs).toISOString()}`, { offset: this.offset, timeMs });
33
33
  }
34
+
35
+ /** Advances the time by the given number of seconds. */
36
+ public advanceTime(seconds: number) {
37
+ this.offset += seconds * 1000;
38
+ }
39
+ }
40
+
41
+ /**
42
+ * A date provider for tests that only advances time via explicit advanceTime() calls.
43
+ * Unlike TestDateProvider, this does NOT track real time progression - time is completely
44
+ * frozen until explicitly advanced. This eliminates flakiness from tests taking
45
+ * varying amounts of real time to execute.
46
+ */
47
+ export class ManualDateProvider extends DateProvider {
48
+ private currentTimeMs: number;
49
+
50
+ /**
51
+ * @param initialTimeMs - Initial time in milliseconds. Defaults to a round timestamp for easy visualization.
52
+ */
53
+ constructor(initialTimeMs: number = Date.UTC(2025, 0, 1, 0, 0, 0)) {
54
+ super();
55
+ this.currentTimeMs = initialTimeMs;
56
+ }
57
+
58
+ public override now(): number {
59
+ return this.currentTimeMs;
60
+ }
61
+
62
+ /** Sets the current time to the given timestamp in milliseconds. */
63
+ public setTime(timeMs: number) {
64
+ this.currentTimeMs = timeMs;
65
+ }
66
+
67
+ /** Advances the time by the given number of seconds. */
68
+ public advanceTime(seconds: number) {
69
+ this.currentTimeMs += seconds * 1000;
70
+ }
71
+
72
+ /** Advances the time by the given number of milliseconds. */
73
+ public advanceTimeMs(ms: number) {
74
+ this.currentTimeMs += ms;
75
+ }
34
76
  }
@@ -1,23 +1,7 @@
1
- import { pedersenHash as pedersenHashArray } from '@aztec/foundation/crypto/pedersen';
2
- import { poseidon2Hash } from '@aztec/foundation/crypto/poseidon';
3
- import { sha256Trunc } from '@aztec/foundation/crypto/sha256';
4
-
5
- import type { AsyncHasher, Hasher } from './hasher.js';
6
-
7
- export const shaMerkleHash: Hasher['hash'] = (left: Buffer, right: Buffer) =>
8
- sha256Trunc(Buffer.concat([left, right])) as Buffer<ArrayBuffer>;
9
-
10
- export const pedersenMerkleHash: AsyncHasher['hash'] = async (left: Buffer, right: Buffer) =>
11
- (await pedersenHashArray([left, right])).toBuffer() as Buffer<ArrayBuffer>;
12
-
13
- export const poseidonMerkleHash: AsyncHasher['hash'] = async (left: Buffer, right: Buffer) =>
14
- (await poseidon2Hash([left, right])).toBuffer() as Buffer<ArrayBuffer>;
1
+ import { poseidonMerkleHash, shaMerkleHash } from './hasher.js';
15
2
 
16
3
  export const computeBalancedShaRoot = (leaves: Buffer[]) => computeBalancedMerkleTreeRoot(leaves);
17
4
 
18
- export const computeBalancedPedersenRoot = async (leaves: Buffer[]) =>
19
- await computeBalancedMerkleTreeRootAsync(leaves, pedersenMerkleHash);
20
-
21
5
  export const computeBalancedPoseidonRoot = async (leaves: Buffer[]) =>
22
6
  await computeBalancedMerkleTreeRootAsync(leaves, poseidonMerkleHash);
23
7
 
@@ -46,7 +30,7 @@ export function computeBalancedMerkleTreeRoot(leaves: Buffer[], hasher = shaMerk
46
30
 
47
31
  /**
48
32
  * Computes the Merkle root with the provided leaves **asynchronously**.
49
- * This method uses an asynchronous hash function (defaults to `pedersenHash`).
33
+ * This method uses an asynchronous hash function (defaults to `poseidon2Hash`).
50
34
  *
51
35
  * @throws If the number of leaves is not a power of two.
52
36
  */
@@ -1,3 +1,6 @@
1
+ import { poseidon2Hash } from '../crypto/poseidon/index.js';
2
+ import { sha256Trunc } from '../crypto/sha256/index.js';
3
+
1
4
  /**
2
5
  * Defines hasher interface used by Merkle trees.
3
6
  */
@@ -37,3 +40,9 @@ export interface AsyncHasher {
37
40
  */
38
41
  hashInputs(inputs: Buffer[]): Promise<Buffer<ArrayBuffer>>;
39
42
  }
43
+
44
+ export const shaMerkleHash: Hasher['hash'] = (left: Buffer, right: Buffer) =>
45
+ sha256Trunc(Buffer.concat([left, right])) as Buffer<ArrayBuffer>;
46
+
47
+ export const poseidonMerkleHash: AsyncHasher['hash'] = async (left: Buffer, right: Buffer) =>
48
+ (await poseidon2Hash([left, right])).toBuffer() as Buffer<ArrayBuffer>;
@@ -1,6 +1,4 @@
1
- export * from './balanced_merkle_tree.js';
2
- export * from './unbalanced_merkle_tree.js';
3
- export * from './unbalanced_tree_store.js';
1
+ export * from './balanced_merkle_tree_root.js';
4
2
  export * from './merkle_tree_calculator.js';
5
3
  export * from './merkle_tree.js';
6
4
  export * from './indexed_merkle_tree_calculator.js';
@@ -10,3 +8,5 @@ export * from './membership_witness.js';
10
8
  export * from './hasher.js';
11
9
  export * from './indexed_tree_leaf.js';
12
10
  export * from './unbalanced_merkle_tree_calculator.js';
11
+ export * from './unbalanced_merkle_tree_root.js';
12
+ export * from './unbalanced_tree_store.js';