@aztec/foundation 0.0.1-commit.bf2612ae → 0.0.1-commit.c0b82b2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (92) hide show
  1. package/dest/array/sorted_array.d.ts +6 -1
  2. package/dest/array/sorted_array.d.ts.map +1 -1
  3. package/dest/array/sorted_array.js +18 -15
  4. package/dest/collection/array.d.ts +7 -1
  5. package/dest/collection/array.d.ts.map +1 -1
  6. package/dest/collection/array.js +24 -0
  7. package/dest/config/env_var.d.ts +2 -2
  8. package/dest/config/env_var.d.ts.map +1 -1
  9. package/dest/config/network_name.d.ts +2 -2
  10. package/dest/config/network_name.d.ts.map +1 -1
  11. package/dest/config/network_name.js +2 -0
  12. package/dest/crypto/poseidon/index.js +13 -13
  13. package/dest/crypto/random/randomness_singleton.d.ts +4 -3
  14. package/dest/crypto/random/randomness_singleton.d.ts.map +1 -1
  15. package/dest/crypto/random/randomness_singleton.js +4 -4
  16. package/dest/curves/bn254/field.d.ts +2 -1
  17. package/dest/curves/bn254/field.d.ts.map +1 -1
  18. package/dest/curves/bn254/field.js +5 -2
  19. package/dest/jest/setup.js +28 -1
  20. package/dest/json-rpc/client/safe_json_rpc_client.d.ts +2 -1
  21. package/dest/json-rpc/client/safe_json_rpc_client.d.ts.map +1 -1
  22. package/dest/json-rpc/client/safe_json_rpc_client.js +1 -1
  23. package/dest/json-rpc/server/api_key_auth.d.ts +19 -0
  24. package/dest/json-rpc/server/api_key_auth.d.ts.map +1 -0
  25. package/dest/json-rpc/server/api_key_auth.js +57 -0
  26. package/dest/json-rpc/server/index.d.ts +2 -1
  27. package/dest/json-rpc/server/index.d.ts.map +1 -1
  28. package/dest/json-rpc/server/index.js +1 -0
  29. package/dest/log/bigint-utils.d.ts +5 -0
  30. package/dest/log/bigint-utils.d.ts.map +1 -0
  31. package/dest/log/bigint-utils.js +18 -0
  32. package/dest/log/gcloud-logger-config.d.ts +1 -1
  33. package/dest/log/gcloud-logger-config.d.ts.map +1 -1
  34. package/dest/log/gcloud-logger-config.js +3 -0
  35. package/dest/log/libp2p_logger.d.ts +5 -2
  36. package/dest/log/libp2p_logger.d.ts.map +1 -1
  37. package/dest/log/libp2p_logger.js +14 -4
  38. package/dest/log/log-filters.d.ts +17 -4
  39. package/dest/log/log-filters.d.ts.map +1 -1
  40. package/dest/log/log-filters.js +26 -12
  41. package/dest/log/pino-logger-server.d.ts +9 -0
  42. package/dest/log/pino-logger-server.d.ts.map +1 -0
  43. package/dest/log/pino-logger-server.js +18 -0
  44. package/dest/log/pino-logger.d.ts +37 -8
  45. package/dest/log/pino-logger.d.ts.map +1 -1
  46. package/dest/log/pino-logger.js +121 -28
  47. package/dest/queue/base_memory_queue.d.ts +2 -2
  48. package/dest/queue/base_memory_queue.d.ts.map +1 -1
  49. package/dest/queue/semaphore.d.ts +5 -1
  50. package/dest/queue/semaphore.d.ts.map +1 -1
  51. package/dest/serialize/buffer_reader.d.ts +8 -1
  52. package/dest/serialize/buffer_reader.d.ts.map +1 -1
  53. package/dest/serialize/buffer_reader.js +13 -0
  54. package/dest/serialize/serialize.d.ts +19 -1
  55. package/dest/serialize/serialize.d.ts.map +1 -1
  56. package/dest/serialize/serialize.js +31 -0
  57. package/dest/sleep/index.d.ts +2 -1
  58. package/dest/sleep/index.d.ts.map +1 -1
  59. package/dest/sleep/index.js +10 -1
  60. package/dest/timer/date.d.ts +25 -1
  61. package/dest/timer/date.d.ts.map +1 -1
  62. package/dest/timer/date.js +33 -0
  63. package/dest/transport/transport_client.js +2 -2
  64. package/dest/types/index.d.ts +3 -1
  65. package/dest/types/index.d.ts.map +1 -1
  66. package/dest/types/index.js +3 -0
  67. package/package.json +23 -2
  68. package/src/array/sorted_array.ts +22 -17
  69. package/src/collection/array.ts +21 -0
  70. package/src/config/env_var.ts +38 -2
  71. package/src/config/network_name.ts +4 -1
  72. package/src/crypto/poseidon/index.ts +13 -13
  73. package/src/crypto/random/randomness_singleton.ts +9 -5
  74. package/src/curves/bn254/field.ts +6 -2
  75. package/src/jest/setup.mjs +31 -1
  76. package/src/json-rpc/client/safe_json_rpc_client.ts +2 -0
  77. package/src/json-rpc/server/api_key_auth.ts +63 -0
  78. package/src/json-rpc/server/index.ts +1 -0
  79. package/src/log/bigint-utils.ts +22 -0
  80. package/src/log/gcloud-logger-config.ts +5 -0
  81. package/src/log/libp2p_logger.ts +12 -5
  82. package/src/log/log-filters.ts +29 -11
  83. package/src/log/pino-logger-server.ts +25 -0
  84. package/src/log/pino-logger.ts +142 -38
  85. package/src/queue/base_memory_queue.ts +1 -1
  86. package/src/queue/semaphore.ts +5 -0
  87. package/src/serialize/buffer_reader.ts +15 -0
  88. package/src/serialize/serialize.ts +32 -0
  89. package/src/sleep/index.ts +10 -1
  90. package/src/timer/date.ts +48 -0
  91. package/src/transport/transport_client.ts +2 -2
  92. package/src/types/index.ts +5 -0
@@ -12,6 +12,9 @@ export type EnvVar =
12
12
  | 'ARCHIVER_VIEM_POLLING_INTERVAL_MS'
13
13
  | 'ARCHIVER_BATCH_SIZE'
14
14
  | 'AZTEC_ADMIN_PORT'
15
+ | 'AZTEC_ADMIN_API_KEY_HASH'
16
+ | 'AZTEC_DISABLE_ADMIN_API_KEY'
17
+ | 'AZTEC_RESET_ADMIN_API_KEY'
15
18
  | 'AZTEC_NODE_ADMIN_URL'
16
19
  | 'AZTEC_NODE_URL'
17
20
  | 'AZTEC_PORT'
@@ -47,7 +50,10 @@ export type EnvVar =
47
50
  | 'BOT_TX_MINED_WAIT_SECONDS'
48
51
  | 'BOT_MAX_CONSECUTIVE_ERRORS'
49
52
  | 'BOT_STOP_WHEN_UNHEALTHY'
50
- | 'BOT_AMM_TXS'
53
+ | 'BOT_MODE'
54
+ | 'BOT_L2_TO_L1_MESSAGES_PER_TX'
55
+ | 'BOT_L1_TO_L2_SEED_COUNT'
56
+ | 'BOT_L1_TO_L2_SEED_INTERVAL'
51
57
  | 'COINBASE'
52
58
  | 'CRS_PATH'
53
59
  | 'DATA_DIRECTORY'
@@ -64,6 +70,7 @@ export type EnvVar =
64
70
  | 'PUBLIC_DATA_TREE_MAP_SIZE_KB'
65
71
  | 'DEBUG'
66
72
  | 'DEBUG_P2P_DISABLE_COLOCATION_PENALTY'
73
+ | 'ENABLE_PROVER_NODE'
67
74
  | 'ETHEREUM_HOSTS'
68
75
  | 'ETHEREUM_DEBUG_HOSTS'
69
76
  | 'ETHEREUM_ALLOW_NO_DEBUG_HOSTS'
@@ -75,8 +82,10 @@ export type EnvVar =
75
82
  | 'L1_CONSENSUS_HOST_URLS'
76
83
  | 'L1_CONSENSUS_HOST_API_KEYS'
77
84
  | 'L1_CONSENSUS_HOST_API_KEY_HEADERS'
85
+ | 'L1_TX_FAILED_STORE'
78
86
  | 'LOG_JSON'
79
87
  | 'LOG_MULTILINE'
88
+ | 'LOG_NO_COLOR_PER_ACTOR'
80
89
  | 'LOG_LEVEL'
81
90
  | 'MNEMONIC'
82
91
  | 'NETWORK'
@@ -94,7 +103,12 @@ export type EnvVar =
94
103
  | 'PUBLIC_OTEL_INCLUDE_METRICS'
95
104
  | 'PUBLIC_OTEL_COLLECT_FROM'
96
105
  | 'PUBLIC_OTEL_OPT_OUT'
106
+ | 'P2P_BATCH_TX_REQUESTER_SMART_PARALLEL_WORKER_COUNT'
107
+ | 'P2P_BATCH_TX_REQUESTER_DUMB_PARALLEL_WORKER_COUNT'
108
+ | 'P2P_BATCH_TX_REQUESTER_TX_BATCH_SIZE'
109
+ | 'P2P_BATCH_TX_REQUESTER_BAD_PEER_THRESHOLD'
97
110
  | 'P2P_BLOCK_CHECK_INTERVAL_MS'
111
+ | 'P2P_SLOT_CHECK_INTERVAL_MS'
98
112
  | 'P2P_BLOCK_REQUEST_BATCH_SIZE'
99
113
  | 'P2P_BOOTSTRAP_NODE_ENR_VERSION_CHECK'
100
114
  | 'P2P_BOOTSTRAP_NODES_AS_FULL_PEERS'
@@ -135,9 +149,9 @@ export type EnvVar =
135
149
  | 'P2P_PREFERRED_PEERS'
136
150
  | 'P2P_MAX_PENDING_TX_COUNT'
137
151
  | 'P2P_SEEN_MSG_CACHE_SIZE'
138
- | 'P2P_DROP_TX'
139
152
  | 'P2P_DROP_TX_CHANCE'
140
153
  | 'P2P_TX_POOL_DELETE_TXS_AFTER_REORG'
154
+ | 'P2P_MIN_TX_POOL_AGE_MS'
141
155
  | 'DEBUG_P2P_INSTRUMENT_MESSAGES'
142
156
  | 'PEER_ID_PRIVATE_KEY'
143
157
  | 'PEER_ID_PRIVATE_KEY_PATH'
@@ -151,8 +165,10 @@ export type EnvVar =
151
165
  | 'PROVER_BROKER_BATCH_INTERVAL_MS'
152
166
  | 'PROVER_BROKER_BATCH_SIZE'
153
167
  | 'PROVER_BROKER_MAX_EPOCHS_TO_KEEP_RESULTS_FOR'
168
+ | 'PROVER_BROKER_DEBUG_REPLAY_ENABLED'
154
169
  | 'PROVER_CANCEL_JOBS_ON_STOP'
155
170
  | 'PROVER_COORDINATION_NODE_URLS'
171
+ | 'PROVER_PROOF_STORE'
156
172
  | 'PROVER_FAILED_PROOF_STORE'
157
173
  | 'PROVER_NODE_FAILED_EPOCH_STORE'
158
174
  | 'PROVER_NODE_DISABLE_PROOF_PUBLISH'
@@ -200,9 +216,11 @@ export type EnvVar =
200
216
  | 'SEQ_L1_PUBLISHING_TIME_ALLOWANCE_IN_SLOT'
201
217
  | 'SEQ_ATTESTATION_PROPAGATION_TIME'
202
218
  | 'SEQ_BLOCK_DURATION_MS'
219
+ | 'SEQ_EXPECTED_BLOCK_PROPOSALS_PER_SLOT'
203
220
  | 'SEQ_BUILD_CHECKPOINT_IF_EMPTY'
204
221
  | 'SEQ_SECONDS_BEFORE_INVALIDATING_BLOCK_AS_COMMITTEE_MEMBER'
205
222
  | 'SEQ_SECONDS_BEFORE_INVALIDATING_BLOCK_AS_NON_COMMITTEE_MEMBER'
223
+ | 'SEQ_SKIP_CHECKPOINT_PUBLISH_PERCENT'
206
224
  | 'SLASH_MIN_PENALTY_PERCENTAGE'
207
225
  | 'SLASH_MAX_PENALTY_PERCENTAGE'
208
226
  | 'SLASH_VALIDATORS_ALWAYS'
@@ -213,6 +231,8 @@ export type EnvVar =
213
231
  | 'SLASH_INACTIVITY_TARGET_PERCENTAGE'
214
232
  | 'SLASH_INACTIVITY_CONSECUTIVE_EPOCH_THRESHOLD'
215
233
  | 'SLASH_INVALID_BLOCK_PENALTY'
234
+ | 'SLASH_DUPLICATE_PROPOSAL_PENALTY'
235
+ | 'SLASH_DUPLICATE_ATTESTATION_PENALTY'
216
236
  | 'SLASH_OVERRIDE_PAYLOAD'
217
237
  | 'SLASH_PROPOSE_INVALID_ATTESTATIONS_PENALTY'
218
238
  | 'SLASH_ATTEST_DESCENDANT_OF_INVALID_PENALTY'
@@ -237,6 +257,20 @@ export type EnvVar =
237
257
  | 'TX_COLLECTION_FAST_MAX_PARALLEL_REQUESTS_PER_NODE'
238
258
  | 'TX_COLLECTION_NODE_RPC_MAX_BATCH_SIZE'
239
259
  | 'TX_COLLECTION_NODE_RPC_URLS'
260
+ | 'TX_COLLECTION_MISSING_TXS_COLLECTOR_TYPE'
261
+ | 'TX_COLLECTION_FILE_STORE_URLS'
262
+ | 'TX_COLLECTION_FILE_STORE_SLOW_DELAY_MS'
263
+ | 'TX_COLLECTION_FILE_STORE_FAST_DELAY_MS'
264
+ | 'TX_COLLECTION_FILE_STORE_FAST_WORKER_COUNT'
265
+ | 'TX_COLLECTION_FILE_STORE_SLOW_WORKER_COUNT'
266
+ | 'TX_COLLECTION_FILE_STORE_FAST_BACKOFF_BASE_MS'
267
+ | 'TX_COLLECTION_FILE_STORE_SLOW_BACKOFF_BASE_MS'
268
+ | 'TX_COLLECTION_FILE_STORE_FAST_BACKOFF_MAX_MS'
269
+ | 'TX_COLLECTION_FILE_STORE_SLOW_BACKOFF_MAX_MS'
270
+ | 'TX_FILE_STORE_URL'
271
+ | 'TX_FILE_STORE_UPLOAD_CONCURRENCY'
272
+ | 'TX_FILE_STORE_MAX_QUEUE_SIZE'
273
+ | 'TX_FILE_STORE_ENABLED'
240
274
  | 'TX_PUBLIC_SETUP_ALLOWLIST'
241
275
  | 'TXE_PORT'
242
276
  | 'TRANSACTIONS_DISABLED'
@@ -251,6 +285,7 @@ export type EnvVar =
251
285
  | 'WS_BLOCK_REQUEST_BATCH_SIZE'
252
286
  | 'L1_READER_VIEM_POLLING_INTERVAL_MS'
253
287
  | 'WS_DATA_DIRECTORY'
288
+ | 'WS_NUM_HISTORIC_CHECKPOINTS'
254
289
  | 'WS_NUM_HISTORIC_BLOCKS'
255
290
  | 'ETHEREUM_SLOT_DURATION'
256
291
  | 'AZTEC_SLOT_DURATION'
@@ -320,6 +355,7 @@ export type EnvVar =
320
355
  | 'VALIDATOR_HA_POLLING_INTERVAL_MS'
321
356
  | 'VALIDATOR_HA_SIGNING_TIMEOUT_MS'
322
357
  | 'VALIDATOR_HA_MAX_STUCK_DUTIES_AGE_MS'
358
+ | 'VALIDATOR_HA_OLD_DUTIES_MAX_AGE_H'
323
359
  | 'VALIDATOR_HA_DATABASE_URL'
324
360
  | 'VALIDATOR_HA_RUN_MIGRATIONS'
325
361
  | 'VALIDATOR_HA_POOL_MAX'
@@ -5,7 +5,8 @@ export type NetworkNames =
5
5
  | 'testnet'
6
6
  | 'mainnet'
7
7
  | 'next-net'
8
- | 'devnet';
8
+ | 'devnet'
9
+ | `v${number}-devnet-${number}`;
9
10
 
10
11
  export function getActiveNetworkName(name?: string): NetworkNames {
11
12
  const network = name || process.env.NETWORK;
@@ -23,6 +24,8 @@ export function getActiveNetworkName(name?: string): NetworkNames {
23
24
  return 'next-net';
24
25
  } else if (network === 'devnet') {
25
26
  return 'devnet';
27
+ } else if (/^v\d+-devnet-\d+$/.test(network)) {
28
+ return network as `v${number}-devnet-${number}`;
26
29
  }
27
30
  throw new Error(`Unknown network: ${network}`);
28
31
  }
@@ -1,4 +1,4 @@
1
- import { BarretenbergSync } from '@aztec/bb.js';
1
+ import { Barretenberg } from '@aztec/bb.js';
2
2
 
3
3
  import { Fr } from '../../curves/bn254/field.js';
4
4
  import { type Fieldable, serializeToFields } from '../../serialize/serialize.js';
@@ -10,9 +10,9 @@ import { type Fieldable, serializeToFields } from '../../serialize/serialize.js'
10
10
  */
11
11
  export async function poseidon2Hash(input: Fieldable[]): Promise<Fr> {
12
12
  const inputFields = serializeToFields(input);
13
- await BarretenbergSync.initSingleton();
14
- const api = BarretenbergSync.getSingleton();
15
- const response = api.poseidon2Hash({
13
+ await Barretenberg.initSingleton();
14
+ const api = Barretenberg.getSingleton();
15
+ const response = await api.poseidon2Hash({
16
16
  inputs: inputFields.map(i => i.toBuffer()),
17
17
  });
18
18
  return Fr.fromBuffer(Buffer.from(response.hash));
@@ -27,9 +27,9 @@ export async function poseidon2Hash(input: Fieldable[]): Promise<Fr> {
27
27
  export async function poseidon2HashWithSeparator(input: Fieldable[], separator: number): Promise<Fr> {
28
28
  const inputFields = serializeToFields(input);
29
29
  inputFields.unshift(new Fr(separator));
30
- await BarretenbergSync.initSingleton();
31
- const api = BarretenbergSync.getSingleton();
32
- const response = api.poseidon2Hash({
30
+ await Barretenberg.initSingleton();
31
+ const api = Barretenberg.getSingleton();
32
+ const response = await api.poseidon2Hash({
33
33
  inputs: inputFields.map(i => i.toBuffer()),
34
34
  });
35
35
  return Fr.fromBuffer(Buffer.from(response.hash));
@@ -44,9 +44,9 @@ export async function poseidon2Permutation(input: Fieldable[]): Promise<Fr[]> {
44
44
  const inputFields = serializeToFields(input);
45
45
  // We'd like this assertion but it's not possible to use it in the browser.
46
46
  // assert(input.length === 4, 'Input state must be of size 4');
47
- await BarretenbergSync.initSingleton();
48
- const api = BarretenbergSync.getSingleton();
49
- const response = api.poseidon2Permutation({
47
+ await Barretenberg.initSingleton();
48
+ const api = Barretenberg.getSingleton();
49
+ const response = await api.poseidon2Permutation({
50
50
  inputs: inputFields.map(i => i.toBuffer()),
51
51
  });
52
52
  // We'd like this assertion but it's not possible to use it in the browser.
@@ -65,9 +65,9 @@ export async function poseidon2HashBytes(input: Buffer): Promise<Fr> {
65
65
  inputFields.push(Fr.fromBuffer(fieldBytes));
66
66
  }
67
67
 
68
- await BarretenbergSync.initSingleton();
69
- const api = BarretenbergSync.getSingleton();
70
- const response = api.poseidon2Hash({
68
+ await Barretenberg.initSingleton();
69
+ const api = Barretenberg.getSingleton();
70
+ const response = await api.poseidon2Hash({
71
71
  inputs: inputFields.map(i => i.toBuffer()),
72
72
  });
73
73
 
@@ -1,4 +1,4 @@
1
- import { createLogger } from '../../log/pino-logger.js';
1
+ import { type Logger, type LoggerBindings, createLogger } from '../../log/pino-logger.js';
2
2
 
3
3
  /**
4
4
  * A number generator which is used as a source of randomness in the system. If the SEED env variable is set, the
@@ -12,9 +12,13 @@ export class RandomnessSingleton {
12
12
  private static instance: RandomnessSingleton;
13
13
 
14
14
  private counter = 0;
15
- private readonly log = createLogger('foundation:randomness_singleton');
15
+ private log: Logger;
16
16
 
17
- private constructor(private readonly seed?: number) {
17
+ private constructor(
18
+ private readonly seed?: number,
19
+ bindings?: LoggerBindings,
20
+ ) {
21
+ this.log = createLogger('foundation:randomness_singleton', bindings);
18
22
  if (seed !== undefined) {
19
23
  this.log.debug(`Using pseudo-randomness with seed: ${seed}`);
20
24
  this.counter = seed;
@@ -23,10 +27,10 @@ export class RandomnessSingleton {
23
27
  }
24
28
  }
25
29
 
26
- public static getInstance(): RandomnessSingleton {
30
+ public static getInstance(bindings?: LoggerBindings): RandomnessSingleton {
27
31
  if (!RandomnessSingleton.instance) {
28
32
  const seed = process.env.SEED ? Number(process.env.SEED) : undefined;
29
- RandomnessSingleton.instance = new RandomnessSingleton(seed);
33
+ RandomnessSingleton.instance = new RandomnessSingleton(seed, bindings);
30
34
  }
31
35
 
32
36
  return RandomnessSingleton.instance;
@@ -118,14 +118,18 @@ abstract class BaseField {
118
118
  }
119
119
 
120
120
  cmp(rhs: BaseField): -1 | 0 | 1 {
121
- const rhsBigInt = rhs.asBigInt;
122
- return this.asBigInt === rhsBigInt ? 0 : this.asBigInt < rhsBigInt ? -1 : 1;
121
+ return BaseField.cmpAsBigInt(this.asBigInt, rhs.asBigInt);
123
122
  }
124
123
 
125
124
  static cmp(lhs: BaseField, rhs: BaseField): -1 | 0 | 1 {
126
125
  return lhs.cmp(rhs);
127
126
  }
128
127
 
128
+ // Actual bigint comparison. Arguments must have been validated previously.
129
+ static cmpAsBigInt(lhs: bigint, rhs: bigint): -1 | 0 | 1 {
130
+ return lhs === rhs ? 0 : lhs < rhs ? -1 : 1;
131
+ }
132
+
129
133
  isZero(): boolean {
130
134
  return this.asBigInt === 0n;
131
135
  }
@@ -1,3 +1,4 @@
1
+ import { parseBooleanEnv } from '@aztec/foundation/config';
1
2
  import { overwriteLoggingStream, pinoPrettyOpts } from '@aztec/foundation/log';
2
3
 
3
4
  import pretty from 'pino-pretty';
@@ -6,4 +7,33 @@ import pretty from 'pino-pretty';
6
7
  // file so we don't mess up with dependencies in non-testing environments,
7
8
  // since pino-pretty messes up with browser bundles.
8
9
  // See also https://www.npmjs.com/package/pino-pretty?activeTab=readme#user-content-usage-with-jest
9
- overwriteLoggingStream(pretty(pinoPrettyOpts));
10
+ if (!parseBooleanEnv(process.env.LOG_JSON)) {
11
+ overwriteLoggingStream(pretty(pinoPrettyOpts));
12
+ }
13
+
14
+ // Prevent timers from keeping the process alive after tests complete.
15
+ // Libraries like viem create internal polling loops (via setTimeout) that
16
+ // reschedule themselves indefinitely. In test environments we never want a
17
+ // timer to be the reason the process can't exit. We also unref stdout/stderr
18
+ // which, when they are pipes (as in Jest workers), remain ref'd by default.
19
+ {
20
+ const origSetTimeout = globalThis.setTimeout;
21
+ const origSetInterval = globalThis.setInterval;
22
+ globalThis.setTimeout = function unrefSetTimeout(...args) {
23
+ const id = origSetTimeout.apply(this, args);
24
+ id?.unref?.();
25
+ return id;
26
+ };
27
+ // Preserve .unref, .__promisify__ etc. that may exist on the original
28
+ Object.setPrototypeOf(globalThis.setTimeout, origSetTimeout);
29
+
30
+ globalThis.setInterval = function unrefSetInterval(...args) {
31
+ const id = origSetInterval.apply(this, args);
32
+ id?.unref?.();
33
+ return id;
34
+ };
35
+ Object.setPrototypeOf(globalThis.setInterval, origSetInterval);
36
+
37
+ if (process.stdout?._handle?.unref) process.stdout._handle.unref();
38
+ if (process.stderr?._handle?.unref) process.stderr._handle.unref();
39
+ }
@@ -24,6 +24,7 @@ export type SafeJsonRpcClientOptions = {
24
24
  batchWindowMS?: number;
25
25
  maxBatchSize?: number;
26
26
  maxRequestBodySize?: number;
27
+ extraHeaders?: Record<string, string>;
27
28
  onResponse?: (res: {
28
29
  response: any;
29
30
  headers: { get: (header: string) => string | null | undefined };
@@ -129,6 +130,7 @@ export function createSafeJsonRpcClient<T extends object>(
129
130
  const { response, headers } = await fetch(
130
131
  host,
131
132
  rpcCalls.map(({ request }) => request),
133
+ config.extraHeaders,
132
134
  );
133
135
 
134
136
  if (config.onResponse) {
@@ -0,0 +1,63 @@
1
+ import { timingSafeEqual } from 'crypto';
2
+ import type Koa from 'koa';
3
+
4
+ import { sha256 } from '../../crypto/sha256/index.js';
5
+ import { createLogger } from '../../log/index.js';
6
+
7
+ const log = createLogger('json-rpc:api-key-auth');
8
+
9
+ /**
10
+ * Computes the SHA-256 hash of a string and returns it as a Buffer.
11
+ * @param input - The input string to hash.
12
+ * @returns The SHA-256 hash as a Buffer.
13
+ */
14
+ export function sha256Hash(input: string): Buffer {
15
+ return sha256(Buffer.from(input));
16
+ }
17
+
18
+ /**
19
+ * Creates a Koa middleware that enforces API key authentication on all requests
20
+ * except the health check endpoint (GET /status).
21
+ *
22
+ * The API key can be provided via the `x-api-key` header or the `Authorization: Bearer <key>` header.
23
+ * Comparison is done by hashing the provided key with SHA-256 and comparing against the stored hash.
24
+ *
25
+ * @param apiKeyHash - The SHA-256 hash of the expected API key as a Buffer.
26
+ * @returns A Koa middleware that rejects requests without a valid API key.
27
+ */
28
+ export function getApiKeyAuthMiddleware(
29
+ apiKeyHash: Buffer,
30
+ ): (ctx: Koa.Context, next: () => Promise<void>) => Promise<void> {
31
+ return async (ctx: Koa.Context, next: () => Promise<void>) => {
32
+ // Allow health check through without auth
33
+ if (ctx.path === '/status' && ctx.method === 'GET') {
34
+ return next();
35
+ }
36
+
37
+ const providedKey = ctx.get('x-api-key') || ctx.get('authorization')?.replace(/^Bearer\s+/i, '');
38
+ if (!providedKey) {
39
+ log.warn(`Rejected admin RPC request from ${ctx.ip}: missing API key`);
40
+ ctx.status = 401;
41
+ ctx.body = {
42
+ jsonrpc: '2.0',
43
+ id: null,
44
+ error: { code: -32000, message: 'Unauthorized: invalid or missing API key' },
45
+ };
46
+ return;
47
+ }
48
+
49
+ const providedHashBuf = sha256Hash(providedKey);
50
+ if (!timingSafeEqual(apiKeyHash, providedHashBuf)) {
51
+ log.warn(`Rejected admin RPC request from ${ctx.ip}: invalid API key`);
52
+ ctx.status = 401;
53
+ ctx.body = {
54
+ jsonrpc: '2.0',
55
+ id: null,
56
+ error: { code: -32000, message: 'Unauthorized: invalid or missing API key' },
57
+ };
58
+ return;
59
+ }
60
+
61
+ await next();
62
+ };
63
+ }
@@ -1 +1,2 @@
1
+ export * from './api_key_auth.js';
1
2
  export * from './safe_json_rpc_server.js';
@@ -0,0 +1,22 @@
1
+ /**
2
+ * Converts bigint values to strings recursively in a log object to avoid serialization issues.
3
+ */
4
+ export function convertBigintsToStrings(obj: unknown): unknown {
5
+ if (typeof obj === 'bigint') {
6
+ return String(obj);
7
+ }
8
+
9
+ if (Array.isArray(obj)) {
10
+ return obj.map(item => convertBigintsToStrings(item));
11
+ }
12
+
13
+ if (obj !== null && typeof obj === 'object') {
14
+ const result: Record<string, unknown> = {};
15
+ for (const key in obj) {
16
+ result[key] = convertBigintsToStrings((obj as Record<string, unknown>)[key]);
17
+ }
18
+ return result;
19
+ }
20
+
21
+ return obj;
22
+ }
@@ -1,5 +1,7 @@
1
1
  import type { pino } from 'pino';
2
2
 
3
+ import { convertBigintsToStrings } from './bigint-utils.js';
4
+
3
5
  /* eslint-disable camelcase */
4
6
 
5
7
  const GOOGLE_CLOUD_TRACE_ID = 'logging.googleapis.com/trace';
@@ -15,6 +17,9 @@ export const GoogleCloudLoggerConfig = {
15
17
  messageKey: 'message',
16
18
  formatters: {
17
19
  log(object: Record<string, unknown>): Record<string, unknown> {
20
+ // Convert bigints to strings recursively to avoid serialization issues
21
+ object = convertBigintsToStrings(object) as Record<string, unknown>;
22
+
18
23
  // Add trace context attributes following Cloud Logging structured log format described
19
24
  // in https://cloud.google.com/logging/docs/structured-logging#special-payload-fields
20
25
  const { trace_id, span_id, trace_flags, ...rest } = object;
@@ -2,15 +2,17 @@ import type { ComponentLogger, Logger } from '@libp2p/interface';
2
2
 
3
3
  import { getLogLevelFromFilters } from './log-filters.js';
4
4
  import type { LogLevel } from './log-levels.js';
5
- import { logFilters, logger } from './pino-logger.js';
5
+ import { type LoggerBindings, logFilters, logger } from './pino-logger.js';
6
6
 
7
7
  /**
8
8
  * Creates a libp2p compatible logger that wraps our pino logger.
9
9
  * This adapter implements the ComponentLogger interface required by libp2p.
10
+ * @param namespace - Base namespace for the logger
11
+ * @param bindings - Optional bindings to pass to the logger (actor, instanceId)
10
12
  */
11
- export function createLibp2pComponentLogger(namespace: string): ComponentLogger {
13
+ export function createLibp2pComponentLogger(namespace: string, bindings?: LoggerBindings): ComponentLogger {
12
14
  return {
13
- forComponent: (component: string) => createLibp2pLogger(`${namespace}:${component}`),
15
+ forComponent: (component: string) => createLibp2pLogger(`${namespace}:${component}`, bindings),
14
16
  };
15
17
  }
16
18
 
@@ -24,9 +26,14 @@ function replaceFormatting(message: string) {
24
26
  return message.replace(/(%p|%a)/g, '%s');
25
27
  }
26
28
 
27
- function createLibp2pLogger(component: string): Logger {
29
+ function createLibp2pLogger(component: string, bindings?: LoggerBindings): Logger {
28
30
  // Create a direct pino logger instance for libp2p that supports string interpolation
29
- const log = logger.child({ module: component }, { level: getLogLevelFromFilters(logFilters, component) });
31
+ const actor = bindings?.actor;
32
+ const instanceId = bindings?.instanceId;
33
+ const log = logger.child(
34
+ { module: component, ...(actor && { actor }), ...(instanceId && { instanceId }) },
35
+ { level: getLogLevelFromFilters(logFilters, component) },
36
+ );
30
37
 
31
38
  const logIfEnabled = (level: LogLevel, message: string, ...args: unknown[]) => {
32
39
  if (!log.isLevelEnabled(level)) {
@@ -19,22 +19,40 @@ export function getLogLevelFromFilters(filters: LogFilters, module: string): Log
19
19
  return undefined;
20
20
  }
21
21
 
22
- export function assertLogLevel(level: string): asserts level is LogLevel {
23
- if (!LogLevels.includes(level as LogLevel)) {
24
- throw new Error(`Invalid log level: ${level}`);
22
+ /**
23
+ * Parses the LOG_LEVEL env string into a default level and per-module filter overrides.
24
+ *
25
+ * Format: `<default_level>;<level>:<module1>,<module2>;<level>:<module3>;...`
26
+ * - First segment (before the first `;`) is the default log level for all modules.
27
+ * - Remaining segments are `level:module` pairs: apply the given level to the listed modules (comma-separated).
28
+ * - Later filters override earlier ones for overlapping module matches.
29
+ * - The `aztec:` prefix is stripped from module names; spaces are trimmed.
30
+ *
31
+ * @example
32
+ * ```ts
33
+ * parseLogLevel('debug;warn:module1,module2;error:module3', 'info')
34
+ * // => ['debug', [['module3', 'error'], ['module2', 'warn'], ['module1', 'warn']]]
35
+ * ```
36
+ */
37
+ export function parseLogLevelEnvVar(
38
+ logLevelEnvVar: string | undefined,
39
+ defaultLevel: LogLevel,
40
+ ): [LogLevel, LogFilters] {
41
+ if (!logLevelEnvVar) {
42
+ return [defaultLevel, []];
25
43
  }
44
+ const [level] = logLevelEnvVar.split(';', 1);
45
+ assertValidLogLevel(level);
46
+ return [level, parseFilters(logLevelEnvVar.slice(level.length + 1))];
26
47
  }
27
48
 
28
- export function parseEnv(env: string | undefined, defaultLevel: LogLevel): [LogLevel, LogFilters] {
29
- if (!env) {
30
- return [defaultLevel, []];
49
+ function assertValidLogLevel(level: string): asserts level is LogLevel {
50
+ if (!LogLevels.includes(level as LogLevel)) {
51
+ throw new Error(`Invalid log level: ${level}`);
31
52
  }
32
- const [level] = env.split(';', 1);
33
- assertLogLevel(level);
34
- return [level, parseFilters(env.slice(level.length + 1))];
35
53
  }
36
54
 
37
- export function parseFilters(definition: string | undefined): LogFilters {
55
+ function parseFilters(definition: string | undefined): LogFilters {
38
56
  if (!definition) {
39
57
  return [];
40
58
  }
@@ -48,7 +66,7 @@ export function parseFilters(definition: string | undefined): LogFilters {
48
66
  throw new Error(`Invalid log filter statement: ${statement}`);
49
67
  }
50
68
  const sanitizedLevel = level.trim().toLowerCase();
51
- assertLogLevel(sanitizedLevel);
69
+ assertValidLogLevel(sanitizedLevel);
52
70
  for (const module of modules.split(',')) {
53
71
  filters.push([
54
72
  module
@@ -0,0 +1,25 @@
1
+ import { AsyncLocalStorage } from 'node:async_hooks';
2
+
3
+ import { type LoggerBindings, addLogBindingsHandler, removeLogBindingsHandler } from './pino-logger.js';
4
+
5
+ /** AsyncLocalStorage for logger bindings context propagation (Node.js only). */
6
+ const bindingsStorage = new AsyncLocalStorage<LoggerBindings>();
7
+
8
+ /** Returns the current bindings from AsyncLocalStorage, if any. */
9
+ export function getBindings(): LoggerBindings | undefined {
10
+ return bindingsStorage.getStore();
11
+ }
12
+
13
+ /**
14
+ * Runs a callback within a bindings context. All loggers created within the callback
15
+ * will automatically inherit the bindings (actor, instanceId) via the log bindings handler.
16
+ */
17
+ export async function withLoggerBindings<T>(bindings: LoggerBindings, callback: () => Promise<T>): Promise<T> {
18
+ const handler = () => bindingsStorage.getStore();
19
+ addLogBindingsHandler(handler);
20
+ try {
21
+ return await bindingsStorage.run(bindings, callback);
22
+ } finally {
23
+ removeLogBindingsHandler(handler);
24
+ }
25
+ }