@devtion/actions 0.0.0-9c50f66 → 0.0.0-a9e4cd4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -67,7 +67,7 @@ yarn add @p0tion/actions
67
67
 
68
68
  ### Local Development
69
69
 
70
- **Prerequisities**
70
+ **Prerequisites**
71
71
 
72
72
  - Node.js version 16.0 or higher.
73
73
  - Yarn version 3.5.0 or higher.
package/dist/index.mjs CHANGED
@@ -1,6 +1,6 @@
1
1
  /**
2
2
  * @module @p0tion/actions
3
- * @version 1.0.5
3
+ * @version 1.1.1
4
4
  * @file A set of actions and helpers for CLI commands
5
5
  * @copyright Ethereum Foundation 2022
6
6
  * @license MIT
@@ -17,8 +17,7 @@ import crypto from 'crypto';
17
17
  import blake from 'blakejs';
18
18
  import { utils } from 'ffjavascript';
19
19
  import winston from 'winston';
20
- import { S3Client, HeadObjectCommand, GetObjectCommand } from '@aws-sdk/client-s3';
21
- import { pipeline, Readable } from 'stream';
20
+ import { pipeline } from 'stream';
22
21
  import { promisify } from 'util';
23
22
  import { initializeApp } from 'firebase/app';
24
23
  import { signInWithCredential, initializeAuth, getAuth } from 'firebase/auth';
@@ -340,7 +339,7 @@ const commonTerms = {
340
339
  finalizeCircuit: "finalizeCircuit",
341
340
  finalizeCeremony: "finalizeCeremony",
342
341
  downloadCircuitArtifacts: "downloadCircuitArtifacts",
343
- transferObject: "transferObject",
342
+ transferObject: "transferObject"
344
343
  }
345
344
  };
346
345
 
@@ -691,11 +690,15 @@ const getChunksAndPreSignedUrls = async (cloudFunctions, bucketName, objectKey,
691
690
  * @param cloudFunctions <Functions> - the Firebase Cloud Functions service instance.
692
691
  * @param ceremonyId <string> - the unique identifier of the ceremony.
693
692
  * @param alreadyUploadedChunks Array<ETagWithPartNumber> - the temporary information about the already uploaded chunks.
693
+ * @param logger <GenericBar> - an optional logger to show progress.
694
694
  * @returns <Promise<Array<ETagWithPartNumber>>> - the completed (uploaded) chunks information.
695
695
  */
696
- const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremonyId, alreadyUploadedChunks) => {
696
+ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremonyId, alreadyUploadedChunks, logger) => {
697
697
  // Keep track of uploaded chunks.
698
698
  const uploadedChunks = alreadyUploadedChunks || [];
699
+ // if we were passed a logger, start it
700
+ if (logger)
701
+ logger.start(chunksWithUrls.length, 0);
699
702
  // Loop through remaining chunks.
700
703
  for (let i = alreadyUploadedChunks ? alreadyUploadedChunks.length : 0; i < chunksWithUrls.length; i += 1) {
701
704
  // Consume the pre-signed url to upload the chunk.
@@ -727,6 +730,9 @@ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremony
727
730
  // nb. this must be done only when contributing (not finalizing).
728
731
  if (!!ceremonyId && !!cloudFunctions)
729
732
  await temporaryStoreCurrentContributionUploadedChunkData(cloudFunctions, ceremonyId, chunk);
733
+ // increment the count on the logger
734
+ if (logger)
735
+ logger.increment();
730
736
  }
731
737
  return uploadedChunks;
732
738
  };
@@ -747,8 +753,9 @@ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremony
747
753
  * @param configStreamChunkSize <number> - size of each chunk into which the artifact is going to be splitted (nb. will be converted in MB).
748
754
  * @param [ceremonyId] <string> - the unique identifier of the ceremony (used as a double-edge sword - as identifier and as a check if current contributor is the coordinator finalizing the ceremony).
749
755
  * @param [temporaryDataToResumeMultiPartUpload] <TemporaryParticipantContributionData> - the temporary information necessary to resume an already started multi-part upload.
756
+ * @param logger <GenericBar> - an optional logger to show progress.
750
757
  */
751
- const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFilePath, configStreamChunkSize, ceremonyId, temporaryDataToResumeMultiPartUpload) => {
758
+ const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFilePath, configStreamChunkSize, ceremonyId, temporaryDataToResumeMultiPartUpload, logger) => {
752
759
  // The unique identifier of the multi-part upload.
753
760
  let multiPartUploadId = "";
754
761
  // The list of already uploaded chunks.
@@ -772,7 +779,7 @@ const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFileP
772
779
  const chunksWithUrlsZkey = await getChunksAndPreSignedUrls(cloudFunctions, bucketName, objectKey, localFilePath, multiPartUploadId, configStreamChunkSize, ceremonyId);
773
780
  // Step (2).
774
781
  const partNumbersAndETagsZkey = await uploadParts(chunksWithUrlsZkey, mime.lookup(localFilePath), // content-type.
775
- cloudFunctions, ceremonyId, alreadyUploadedChunks);
782
+ cloudFunctions, ceremonyId, alreadyUploadedChunks, logger);
776
783
  // Step (3).
777
784
  await completeMultiPartUpload(cloudFunctions, bucketName, objectKey, multiPartUploadId, partNumbersAndETagsZkey, ceremonyId);
778
785
  };
@@ -1044,207 +1051,22 @@ const compareHashes = async (path1, path2) => {
1044
1051
  };
1045
1052
 
1046
1053
  /**
1047
- * Parse and validate that the ceremony configuration is correct
1048
- * @notice this does not upload any files to storage
1049
- * @param path <string> - the path to the configuration file
1050
- * @param cleanup <boolean> - whether to delete the r1cs file after parsing
1051
- * @returns any - the data to pass to the cloud function for setup and the circuit artifacts
1054
+ * Return a string with double digits if the provided input is one digit only.
1055
+ * @param in <number> - the input number to be converted.
1056
+ * @returns <string> - the two digits stringified number derived from the conversion.
1052
1057
  */
1053
- const parseCeremonyFile = async (path, cleanup = false) => {
1054
- // check that the path exists
1055
- if (!fs.existsSync(path))
1056
- throw new Error("The provided path to the configuration file does not exist. Please provide an absolute path and try again.");
1057
- try {
1058
- // read the data
1059
- const data = JSON.parse(fs.readFileSync(path).toString());
1060
- // verify that the data is correct
1061
- if (data['timeoutMechanismType'] !== "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */ && data['timeoutMechanismType'] !== "FIXED" /* CeremonyTimeoutType.FIXED */)
1062
- throw new Error("Invalid timeout type. Please choose between DYNAMIC and FIXED.");
1063
- // validate that we have at least 1 circuit input data
1064
- if (!data.circuits || data.circuits.length === 0)
1065
- throw new Error("You need to provide the data for at least 1 circuit.");
1066
- // validate that the end date is in the future
1067
- let endDate;
1068
- let startDate;
1069
- try {
1070
- endDate = new Date(data.endDate);
1071
- startDate = new Date(data.startDate);
1072
- }
1073
- catch (error) {
1074
- throw new Error("The dates should follow this format: 2023-07-04T00:00:00.");
1075
- }
1076
- if (endDate <= startDate)
1077
- throw new Error("The end date should be greater than the start date.");
1078
- const currentDate = new Date();
1079
- if (endDate <= currentDate || startDate <= currentDate)
1080
- throw new Error("The start and end dates should be in the future.");
1081
- // validate penalty
1082
- if (data.penalty <= 0)
1083
- throw new Error("The penalty should be greater than zero.");
1084
- const circuits = [];
1085
- const urlPattern = /(https?:\/\/[^\s]+)/g;
1086
- const commitHashPattern = /^[a-f0-9]{40}$/i;
1087
- const circuitArtifacts = [];
1088
- for (let i = 0; i < data.circuits.length; i++) {
1089
- const circuitData = data.circuits[i];
1090
- const artifacts = circuitData.artifacts;
1091
- circuitArtifacts.push({
1092
- artifacts: artifacts
1093
- });
1094
- const r1csPath = artifacts.r1csStoragePath;
1095
- const wasmPath = artifacts.wasmStoragePath;
1096
- // where we storing the r1cs downloaded
1097
- const localR1csPath = `./${circuitData.name}.r1cs`;
1098
- // check that the artifacts exist in S3
1099
- // we don't need any privileges to download this
1100
- // just the correct region
1101
- const s3 = new S3Client({ region: artifacts.region });
1102
- try {
1103
- await s3.send(new HeadObjectCommand({
1104
- Bucket: artifacts.bucket,
1105
- Key: r1csPath
1106
- }));
1107
- }
1108
- catch (error) {
1109
- throw new Error(`The r1cs file (${r1csPath}) seems to not exist. Please ensure this is correct and that the object is publicly available.`);
1110
- }
1111
- try {
1112
- await s3.send(new HeadObjectCommand({
1113
- Bucket: artifacts.bucket,
1114
- Key: wasmPath
1115
- }));
1116
- }
1117
- catch (error) {
1118
- throw new Error(`The wasm file (${wasmPath}) seems to not exist. Please ensure this is correct and that the object is publicly available.`);
1119
- }
1120
- // download the r1cs to extract the metadata
1121
- const command = new GetObjectCommand({ Bucket: artifacts.bucket, Key: artifacts.r1csStoragePath });
1122
- const response = await s3.send(command);
1123
- const streamPipeline = promisify(pipeline);
1124
- if (response.$metadata.httpStatusCode !== 200)
1125
- throw new Error("There was an error while trying to download the r1cs file. Please check that the file has the correct permissions (public) set.");
1126
- if (response.Body instanceof Readable)
1127
- await streamPipeline(response.Body, fs.createWriteStream(localR1csPath));
1128
- // extract the metadata from the r1cs
1129
- const metadata = getR1CSInfo(localR1csPath);
1130
- // validate that the circuit hash and template links are valid
1131
- const template = circuitData.template;
1132
- const URLMatch = template.source.match(urlPattern);
1133
- if (!URLMatch || URLMatch.length === 0 || URLMatch.length > 1)
1134
- throw new Error("You should provide the URL to the circuits templates on GitHub.");
1135
- const hashMatch = template.commitHash.match(commitHashPattern);
1136
- if (!hashMatch || hashMatch.length === 0 || hashMatch.length > 1)
1137
- throw new Error("You should provide a valid commit hash of the circuit templates.");
1138
- // calculate the hash of the r1cs file
1139
- const r1csBlake2bHash = await blake512FromPath(localR1csPath);
1140
- const circuitPrefix = extractPrefix(circuitData.name);
1141
- // filenames
1142
- const doubleDigitsPowers = convertToDoubleDigits(metadata.pot);
1143
- const r1csCompleteFilename = `${circuitData.name}.r1cs`;
1144
- const wasmCompleteFilename = `${circuitData.name}.wasm`;
1145
- const smallestPowersOfTauCompleteFilenameForCircuit = `${potFilenameTemplate}${doubleDigitsPowers}.ptau`;
1146
- const firstZkeyCompleteFilename = `${circuitPrefix}_${genesisZkeyIndex}.zkey`;
1147
- // storage paths
1148
- const r1csStorageFilePath = getR1csStorageFilePath(circuitPrefix, r1csCompleteFilename);
1149
- const wasmStorageFilePath = getWasmStorageFilePath(circuitPrefix, wasmCompleteFilename);
1150
- const potStorageFilePath = getPotStorageFilePath(smallestPowersOfTauCompleteFilenameForCircuit);
1151
- const zkeyStorageFilePath = getZkeyStorageFilePath(circuitPrefix, firstZkeyCompleteFilename);
1152
- const files = {
1153
- potFilename: smallestPowersOfTauCompleteFilenameForCircuit,
1154
- r1csFilename: r1csCompleteFilename,
1155
- wasmFilename: wasmCompleteFilename,
1156
- initialZkeyFilename: firstZkeyCompleteFilename,
1157
- potStoragePath: potStorageFilePath,
1158
- r1csStoragePath: r1csStorageFilePath,
1159
- wasmStoragePath: wasmStorageFilePath,
1160
- initialZkeyStoragePath: zkeyStorageFilePath,
1161
- r1csBlake2bHash: r1csBlake2bHash
1162
- };
1163
- // validate that the compiler hash is a valid hash
1164
- const compiler = circuitData.compiler;
1165
- const compilerHashMatch = compiler.commitHash.match(commitHashPattern);
1166
- if (!compilerHashMatch || compilerHashMatch.length === 0 || compilerHashMatch.length > 1)
1167
- throw new Error("You should provide a valid commit hash of the circuit compiler.");
1168
- // validate that the verification options are valid
1169
- const verification = circuitData.verification;
1170
- if (verification.cfOrVm !== "CF" && verification.cfOrVm !== "VM")
1171
- throw new Error("Please enter a valid verification mechanism: either CF or VM");
1172
- // @todo VM parameters verification
1173
- // if (verification['cfOrVM'] === "VM") {}
1174
- // check that the timeout is provided for the correct configuration
1175
- let dynamicThreshold;
1176
- let fixedTimeWindow;
1177
- let circuit = {};
1178
- if (data.timeoutMechanismType === "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */) {
1179
- if (circuitData.dynamicThreshold <= 0)
1180
- throw new Error("The dynamic threshold should be > 0.");
1181
- dynamicThreshold = circuitData.dynamicThreshold;
1182
- // the Circuit data for the ceremony setup
1183
- circuit = {
1184
- name: circuitData.name,
1185
- description: circuitData.description,
1186
- prefix: circuitPrefix,
1187
- sequencePosition: i + 1,
1188
- metadata: metadata,
1189
- files: files,
1190
- template: template,
1191
- compiler: compiler,
1192
- verification: verification,
1193
- dynamicThreshold: dynamicThreshold,
1194
- avgTimings: {
1195
- contributionComputation: 0,
1196
- fullContribution: 0,
1197
- verifyCloudFunction: 0
1198
- },
1199
- };
1200
- }
1201
- if (data.timeoutMechanismType === "FIXED" /* CeremonyTimeoutType.FIXED */) {
1202
- if (circuitData.fixedTimeWindow <= 0)
1203
- throw new Error("The fixed time window threshold should be > 0.");
1204
- fixedTimeWindow = circuitData.fixedTimeWindow;
1205
- // the Circuit data for the ceremony setup
1206
- circuit = {
1207
- name: circuitData.name,
1208
- description: circuitData.description,
1209
- prefix: circuitPrefix,
1210
- sequencePosition: i + 1,
1211
- metadata: metadata,
1212
- files: files,
1213
- template: template,
1214
- compiler: compiler,
1215
- verification: verification,
1216
- fixedTimeWindow: fixedTimeWindow,
1217
- avgTimings: {
1218
- contributionComputation: 0,
1219
- fullContribution: 0,
1220
- verifyCloudFunction: 0
1221
- },
1222
- };
1223
- }
1224
- circuits.push(circuit);
1225
- // remove the local r1cs download (if used for verifying the config only vs setup)
1226
- if (cleanup)
1227
- fs.unlinkSync(localR1csPath);
1228
- }
1229
- const setupData = {
1230
- ceremonyInputData: {
1231
- title: data.title,
1232
- description: data.description,
1233
- startDate: startDate.valueOf(),
1234
- endDate: endDate.valueOf(),
1235
- timeoutMechanismType: data.timeoutMechanismType,
1236
- penalty: data.penalty
1237
- },
1238
- ceremonyPrefix: extractPrefix(data.title),
1239
- circuits: circuits,
1240
- circuitArtifacts: circuitArtifacts
1241
- };
1242
- return setupData;
1243
- }
1244
- catch (error) {
1245
- throw new Error(`Error while parsing up the ceremony setup file. ${error.message}`);
1246
- }
1247
- };
1058
+ const convertToDoubleDigits = (amount) => (amount < 10 ? `0${amount}` : amount.toString());
1059
+ /**
1060
+ * Extract a prefix consisting of alphanumeric and underscore characters from a string with arbitrary characters.
1061
+ * @dev replaces all special symbols and whitespaces with an underscore char ('_'). Convert all uppercase chars to lowercase.
1062
+ * @notice example: str = 'Multiplier-2!2.4.zkey'; output prefix = 'multiplier_2_2_4.zkey'.
1063
+ * NB. Prefix extraction is a key process that conditions the name of the ceremony artifacts, download/upload from/to storage, collections paths.
1064
+ * @param str <string> - the arbitrary string from which to extract the prefix.
1065
+ * @returns <string> - the resulting prefix.
1066
+ */
1067
+ const extractPrefix = (str) =>
1068
+ // eslint-disable-next-line no-useless-escape
1069
+ str.replace(/[`\s~!@#$%^&*()|+\-=?;:'",.<>\{\}\[\]\\\/]/gi, "-").toLowerCase();
1248
1070
  /**
1249
1071
  * Extract data from a R1CS metadata file generated with a custom file-based logger.
1250
1072
  * @notice useful for extracting metadata circuits contained in the generated file using a logger
@@ -1301,17 +1123,6 @@ const formatZkeyIndex = (progress) => {
1301
1123
  * @returns <number> - the amount of powers.
1302
1124
  */
1303
1125
  const extractPoTFromFilename = (potCompleteFilename) => Number(potCompleteFilename.split("_").pop()?.split(".").at(0));
1304
- /**
1305
- * Extract a prefix consisting of alphanumeric and underscore characters from a string with arbitrary characters.
1306
- * @dev replaces all special symbols and whitespaces with an underscore char ('_'). Convert all uppercase chars to lowercase.
1307
- * @notice example: str = 'Multiplier-2!2.4.zkey'; output prefix = 'multiplier_2_2_4.zkey'.
1308
- * NB. Prefix extraction is a key process that conditions the name of the ceremony artifacts, download/upload from/to storage, collections paths.
1309
- * @param str <string> - the arbitrary string from which to extract the prefix.
1310
- * @returns <string> - the resulting prefix.
1311
- */
1312
- const extractPrefix = (str) =>
1313
- // eslint-disable-next-line no-useless-escape
1314
- str.replace(/[`\s~!@#$%^&*()|+\-=?;:'",.<>\{\}\[\]\\\/]/gi, "-").toLowerCase();
1315
1126
  /**
1316
1127
  * Automate the generation of an entropy for a contribution.
1317
1128
  * @dev Took inspiration from here https://github.com/glamperd/setup-mpc-ui/blob/master/client/src/state/Compute.tsx#L112.
@@ -1378,7 +1189,9 @@ const getContributionsValidityForContributor = async (firestoreDatabase, circuit
1378
1189
  * @param isFinalizing <boolean> - true when the coordinator is finalizing the ceremony, otherwise false.
1379
1190
  * @returns <string> - the public attestation preamble.
1380
1191
  */
1381
- const getPublicAttestationPreambleForContributor = (contributorIdentifier, ceremonyName, isFinalizing) => `Hey, I'm ${contributorIdentifier} and I have ${isFinalizing ? "finalized" : "contributed to"} the ${ceremonyName} MPC Phase2 Trusted Setup ceremony.\nThe following are my contribution signatures:`;
1192
+ const getPublicAttestationPreambleForContributor = (contributorIdentifier, ceremonyName, isFinalizing) => `Hey, I'm ${contributorIdentifier} and I have ${isFinalizing ? "finalized" : "contributed to"} the ${ceremonyName}${ceremonyName.toLowerCase().includes("trusted setup") || ceremonyName.toLowerCase().includes("ceremony")
1193
+ ? "."
1194
+ : " MPC Phase2 Trusted Setup ceremony."}\nThe following are my contribution signatures:`;
1382
1195
  /**
1383
1196
  * Check and prepare public attestation for the contributor made only of its valid contributions.
1384
1197
  * @param firestoreDatabase <Firestore> - the Firestore service instance associated to the current Firebase application.
@@ -1512,8 +1325,9 @@ const getR1CSInfo = (localR1CSFilePath) => {
1512
1325
  const numberOfSections = utils.leBuff2int(readBytesFromFile(localR1CSFilePath, 0, 4, 8));
1513
1326
  // Jump to first section.
1514
1327
  pointer = 12;
1328
+ let found = false;
1515
1329
  // For each section
1516
- for (let i = 0; i < numberOfSections; i++) {
1330
+ for (let i = 0; i < numberOfSections && !found; i++) {
1517
1331
  // Read section type.
1518
1332
  const sectionType = utils.leBuff2int(readBytesFromFile(localR1CSFilePath, 0, 4, pointer));
1519
1333
  // Jump to section size.
@@ -1562,6 +1376,7 @@ const getR1CSInfo = (localR1CSFilePath) => {
1562
1376
  labels = Number(utils.leBuff2int(readBytesFromFile(localR1CSFilePath, 0, 8, pointer)));
1563
1377
  pointer += 8;
1564
1378
  constraints = Number(utils.leBuff2int(readBytesFromFile(localR1CSFilePath, 0, 4, pointer)));
1379
+ found = true;
1565
1380
  }
1566
1381
  pointer += 8 + Number(sectionSize);
1567
1382
  }
@@ -1581,11 +1396,194 @@ const getR1CSInfo = (localR1CSFilePath) => {
1581
1396
  }
1582
1397
  };
1583
1398
  /**
1584
- * Return a string with double digits if the provided input is one digit only.
1585
- * @param in <number> - the input number to be converted.
1586
- * @returns <string> - the two digits stringified number derived from the conversion.
1399
+ * Parse and validate that the ceremony configuration is correct
1400
+ * @notice this does not upload any files to storage
1401
+ * @param path <string> - the path to the configuration file
1402
+ * @param cleanup <boolean> - whether to delete the r1cs file after parsing
1403
+ * @returns any - the data to pass to the cloud function for setup and the circuit artifacts
1587
1404
  */
1588
- const convertToDoubleDigits = (amount) => (amount < 10 ? `0${amount}` : amount.toString());
1405
+ const parseCeremonyFile = async (path, cleanup = false) => {
1406
+ // check that the path exists
1407
+ if (!fs.existsSync(path))
1408
+ throw new Error("The provided path to the configuration file does not exist. Please provide an absolute path and try again.");
1409
+ try {
1410
+ // read the data
1411
+ const data = JSON.parse(fs.readFileSync(path).toString());
1412
+ // verify that the data is correct
1413
+ if (data.timeoutMechanismType !== "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */ &&
1414
+ data.timeoutMechanismType !== "FIXED" /* CeremonyTimeoutType.FIXED */)
1415
+ throw new Error("Invalid timeout type. Please choose between DYNAMIC and FIXED.");
1416
+ // validate that we have at least 1 circuit input data
1417
+ if (!data.circuits || data.circuits.length === 0)
1418
+ throw new Error("You need to provide the data for at least 1 circuit.");
1419
+ // validate that the end date is in the future
1420
+ let endDate;
1421
+ let startDate;
1422
+ try {
1423
+ endDate = new Date(data.endDate);
1424
+ startDate = new Date(data.startDate);
1425
+ }
1426
+ catch (error) {
1427
+ throw new Error("The dates should follow this format: 2023-07-04T00:00:00.");
1428
+ }
1429
+ if (endDate <= startDate)
1430
+ throw new Error("The end date should be greater than the start date.");
1431
+ const currentDate = new Date();
1432
+ if (endDate <= currentDate || startDate <= currentDate)
1433
+ throw new Error("The start and end dates should be in the future.");
1434
+ // validate penalty
1435
+ if (data.penalty <= 0)
1436
+ throw new Error("The penalty should be greater than zero.");
1437
+ const circuits = [];
1438
+ const urlPattern = /(https?:\/\/[^\s]+)/g;
1439
+ const commitHashPattern = /^[a-f0-9]{40}$/i;
1440
+ const circuitArtifacts = [];
1441
+ for (let i = 0; i < data.circuits.length; i++) {
1442
+ const circuitData = data.circuits[i];
1443
+ const { artifacts } = circuitData;
1444
+ circuitArtifacts.push({
1445
+ artifacts
1446
+ });
1447
+ // where we storing the r1cs downloaded
1448
+ const localR1csPath = `./${circuitData.name}.r1cs`;
1449
+ // where we storing the wasm downloaded
1450
+ const localWasmPath = `./${circuitData.name}.wasm`;
1451
+ // download the r1cs to extract the metadata
1452
+ const streamPipeline = promisify(pipeline);
1453
+ // Make the call.
1454
+ const responseR1CS = await fetch(artifacts.r1csStoragePath);
1455
+ // Handle errors.
1456
+ if (!responseR1CS.ok && responseR1CS.status !== 200)
1457
+ throw new Error(`There was an error while trying to download the r1cs file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
1458
+ await streamPipeline(responseR1CS.body, createWriteStream(localR1csPath));
1459
+ // Write the file locally
1460
+ // extract the metadata from the r1cs
1461
+ const metadata = getR1CSInfo(localR1csPath);
1462
+ // download wasm too to ensure it's available
1463
+ const responseWASM = await fetch(artifacts.wasmStoragePath);
1464
+ if (!responseWASM.ok && responseWASM.status !== 200)
1465
+ throw new Error(`There was an error while trying to download the WASM file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
1466
+ await streamPipeline(responseWASM.body, createWriteStream(localWasmPath));
1467
+ // validate that the circuit hash and template links are valid
1468
+ const { template } = circuitData;
1469
+ const URLMatch = template.source.match(urlPattern);
1470
+ if (!URLMatch || URLMatch.length === 0 || URLMatch.length > 1)
1471
+ throw new Error("You should provide the URL to the circuits templates on GitHub.");
1472
+ const hashMatch = template.commitHash.match(commitHashPattern);
1473
+ if (!hashMatch || hashMatch.length === 0 || hashMatch.length > 1)
1474
+ throw new Error("You should provide a valid commit hash of the circuit templates.");
1475
+ // calculate the hash of the r1cs file
1476
+ const r1csBlake2bHash = await blake512FromPath(localR1csPath);
1477
+ const circuitPrefix = extractPrefix(circuitData.name);
1478
+ // filenames
1479
+ const doubleDigitsPowers = convertToDoubleDigits(metadata.pot);
1480
+ const r1csCompleteFilename = `${circuitData.name}.r1cs`;
1481
+ const wasmCompleteFilename = `${circuitData.name}.wasm`;
1482
+ const smallestPowersOfTauCompleteFilenameForCircuit = `${potFilenameTemplate}${doubleDigitsPowers}.ptau`;
1483
+ const firstZkeyCompleteFilename = `${circuitPrefix}_${genesisZkeyIndex}.zkey`;
1484
+ // storage paths
1485
+ const r1csStorageFilePath = getR1csStorageFilePath(circuitPrefix, r1csCompleteFilename);
1486
+ const wasmStorageFilePath = getWasmStorageFilePath(circuitPrefix, wasmCompleteFilename);
1487
+ const potStorageFilePath = getPotStorageFilePath(smallestPowersOfTauCompleteFilenameForCircuit);
1488
+ const zkeyStorageFilePath = getZkeyStorageFilePath(circuitPrefix, firstZkeyCompleteFilename);
1489
+ const files = {
1490
+ potFilename: smallestPowersOfTauCompleteFilenameForCircuit,
1491
+ r1csFilename: r1csCompleteFilename,
1492
+ wasmFilename: wasmCompleteFilename,
1493
+ initialZkeyFilename: firstZkeyCompleteFilename,
1494
+ potStoragePath: potStorageFilePath,
1495
+ r1csStoragePath: r1csStorageFilePath,
1496
+ wasmStoragePath: wasmStorageFilePath,
1497
+ initialZkeyStoragePath: zkeyStorageFilePath,
1498
+ r1csBlake2bHash
1499
+ };
1500
+ // validate that the compiler hash is a valid hash
1501
+ const { compiler } = circuitData;
1502
+ const compilerHashMatch = compiler.commitHash.match(commitHashPattern);
1503
+ if (!compilerHashMatch || compilerHashMatch.length === 0 || compilerHashMatch.length > 1)
1504
+ throw new Error("You should provide a valid commit hash of the circuit compiler.");
1505
+ // validate that the verification options are valid
1506
+ const { verification } = circuitData;
1507
+ if (verification.cfOrVm !== "CF" && verification.cfOrVm !== "VM")
1508
+ throw new Error("Please enter a valid verification mechanism: either CF or VM");
1509
+ // @todo VM parameters verification
1510
+ // if (verification['cfOrVM'] === "VM") {}
1511
+ // check that the timeout is provided for the correct configuration
1512
+ let dynamicThreshold;
1513
+ let fixedTimeWindow;
1514
+ let circuit = {};
1515
+ if (data.timeoutMechanismType === "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */) {
1516
+ if (circuitData.dynamicThreshold <= 0)
1517
+ throw new Error("The dynamic threshold should be > 0.");
1518
+ dynamicThreshold = circuitData.dynamicThreshold;
1519
+ // the Circuit data for the ceremony setup
1520
+ circuit = {
1521
+ name: circuitData.name,
1522
+ description: circuitData.description,
1523
+ prefix: circuitPrefix,
1524
+ sequencePosition: i + 1,
1525
+ metadata,
1526
+ files,
1527
+ template,
1528
+ compiler,
1529
+ verification,
1530
+ dynamicThreshold,
1531
+ avgTimings: {
1532
+ contributionComputation: 0,
1533
+ fullContribution: 0,
1534
+ verifyCloudFunction: 0
1535
+ }
1536
+ };
1537
+ }
1538
+ if (data.timeoutMechanismType === "FIXED" /* CeremonyTimeoutType.FIXED */) {
1539
+ if (circuitData.fixedTimeWindow <= 0)
1540
+ throw new Error("The fixed time window threshold should be > 0.");
1541
+ fixedTimeWindow = circuitData.fixedTimeWindow;
1542
+ // the Circuit data for the ceremony setup
1543
+ circuit = {
1544
+ name: circuitData.name,
1545
+ description: circuitData.description,
1546
+ prefix: circuitPrefix,
1547
+ sequencePosition: i + 1,
1548
+ metadata,
1549
+ files,
1550
+ template,
1551
+ compiler,
1552
+ verification,
1553
+ fixedTimeWindow,
1554
+ avgTimings: {
1555
+ contributionComputation: 0,
1556
+ fullContribution: 0,
1557
+ verifyCloudFunction: 0
1558
+ }
1559
+ };
1560
+ }
1561
+ circuits.push(circuit);
1562
+ // remove the local r1cs and wasm downloads (if used for verifying the config only vs setup)
1563
+ if (cleanup) {
1564
+ fs.unlinkSync(localR1csPath);
1565
+ fs.unlinkSync(localWasmPath);
1566
+ }
1567
+ }
1568
+ const setupData = {
1569
+ ceremonyInputData: {
1570
+ title: data.title,
1571
+ description: data.description,
1572
+ startDate: startDate.valueOf(),
1573
+ endDate: endDate.valueOf(),
1574
+ timeoutMechanismType: data.timeoutMechanismType,
1575
+ penalty: data.penalty
1576
+ },
1577
+ ceremonyPrefix: extractPrefix(data.title),
1578
+ circuits,
1579
+ circuitArtifacts
1580
+ };
1581
+ return setupData;
1582
+ }
1583
+ catch (error) {
1584
+ throw new Error(`Error while parsing up the ceremony setup file. ${error.message}`);
1585
+ }
1586
+ };
1589
1587
 
1590
1588
  /**
1591
1589
  * Verify that a zKey is valid
@@ -1834,7 +1832,7 @@ const getFirestoreDatabase = (app) => getFirestore(app);
1834
1832
  * @param app <FirebaseApp> - the Firebase application.
1835
1833
  * @returns <Functions> - the Cloud Functions associated to the application.
1836
1834
  */
1837
- const getFirebaseFunctions = (app) => getFunctions(app, 'europe-west1');
1835
+ const getFirebaseFunctions = (app) => getFunctions(app, "europe-west1");
1838
1836
  /**
1839
1837
  * Retrieve the configuration variables for the AWS services (S3, EC2).
1840
1838
  * @returns <AWSVariables> - the values of the AWS services configuration variables.
@@ -1843,14 +1841,14 @@ const getAWSVariables = () => {
1843
1841
  if (!process.env.AWS_ACCESS_KEY_ID ||
1844
1842
  !process.env.AWS_SECRET_ACCESS_KEY ||
1845
1843
  !process.env.AWS_REGION ||
1846
- !process.env.AWS_ROLE_ARN ||
1844
+ !process.env.AWS_INSTANCE_PROFILE_ARN ||
1847
1845
  !process.env.AWS_AMI_ID)
1848
1846
  throw new Error("Could not retrieve the AWS environment variables. Please, verify your environment configuration and retry");
1849
1847
  return {
1850
1848
  accessKeyId: process.env.AWS_ACCESS_KEY_ID,
1851
1849
  secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY,
1852
1850
  region: process.env.AWS_REGION || "us-east-1",
1853
- roleArn: process.env.AWS_ROLE_ARN,
1851
+ instanceProfileArn: process.env.AWS_INSTANCE_PROFILE_ARN,
1854
1852
  amiId: process.env.AWS_AMI_ID
1855
1853
  };
1856
1854
  };
@@ -2103,7 +2101,8 @@ const getGitHubStats = async (user) => {
2103
2101
  following: jsonData.following,
2104
2102
  followers: jsonData.followers,
2105
2103
  publicRepos: jsonData.public_repos,
2106
- avatarUrl: jsonData.avatar_url
2104
+ avatarUrl: jsonData.avatar_url,
2105
+ age: jsonData.created_at
2107
2106
  };
2108
2107
  return data;
2109
2108
  };
@@ -2115,20 +2114,21 @@ const getGitHubStats = async (user) => {
2115
2114
  * @param minimumAmountOfPublicRepos <number> The minimum amount of public repos the user should have
2116
2115
  * @returns <any> Return the avatar URL of the user if the user is reputable, false otherwise
2117
2116
  */
2118
- const githubReputation = async (userLogin, minimumAmountOfFollowing, minimumAmountOfFollowers, minimumAmountOfPublicRepos) => {
2117
+ const githubReputation = async (userLogin, minimumAmountOfFollowing, minimumAmountOfFollowers, minimumAmountOfPublicRepos, minimumAge) => {
2119
2118
  if (!process.env.GITHUB_ACCESS_TOKEN)
2120
2119
  throw new Error("The GitHub access token is missing. Please insert a valid token to be used for anti-sybil checks on user registation, and then try again.");
2121
- const { following, followers, publicRepos, avatarUrl } = await getGitHubStats(userLogin);
2120
+ const { following, followers, publicRepos, avatarUrl, age } = await getGitHubStats(userLogin);
2122
2121
  if (following < minimumAmountOfFollowing ||
2123
2122
  publicRepos < minimumAmountOfPublicRepos ||
2124
- followers < minimumAmountOfFollowers)
2123
+ followers < minimumAmountOfFollowers ||
2124
+ new Date(age) > new Date(Date.now() - minimumAge))
2125
2125
  return {
2126
2126
  reputable: false,
2127
2127
  avatarUrl: ""
2128
2128
  };
2129
2129
  return {
2130
2130
  reputable: true,
2131
- avatarUrl: avatarUrl
2131
+ avatarUrl
2132
2132
  };
2133
2133
  };
2134
2134
 
@@ -2337,8 +2337,13 @@ const vmDependenciesAndCacheArtifactsCommand = (zKeyPath, potPath, snsTopic, reg
2337
2337
  // eslint-disable-next-line no-template-curly-in-string
2338
2338
  "touch ${MARKER_FILE}",
2339
2339
  "sudo yum update -y",
2340
- "curl -sL https://rpm.nodesource.com/setup_16.x | sudo bash - ",
2341
- "sudo yum install -y nodejs",
2340
+ "curl -O https://nodejs.org/dist/v16.13.0/node-v16.13.0-linux-x64.tar.xz",
2341
+ "tar -xf node-v16.13.0-linux-x64.tar.xz",
2342
+ "mv node-v16.13.0-linux-x64 nodejs",
2343
+ "sudo mv nodejs /opt/",
2344
+ "echo 'export NODEJS_HOME=/opt/nodejs' >> /etc/profile",
2345
+ "echo 'export PATH=$NODEJS_HOME/bin:$PATH' >> /etc/profile",
2346
+ "source /etc/profile",
2342
2347
  "npm install -g snarkjs",
2343
2348
  `aws s3 cp s3://${zKeyPath} /var/tmp/genesisZkey.zkey`,
2344
2349
  `aws s3 cp s3://${potPath} /var/tmp/pot.ptau`,
@@ -2357,6 +2362,7 @@ const vmDependenciesAndCacheArtifactsCommand = (zKeyPath, potPath, snsTopic, reg
2357
2362
  * @returns Array<string> - the list of commands for contribution verification.
2358
2363
  */
2359
2364
  const vmContributionVerificationCommand = (bucketName, lastZkeyStoragePath, verificationTranscriptStoragePathAndFilename) => [
2365
+ `source /etc/profile`,
2360
2366
  `aws s3 cp s3://${bucketName}/${lastZkeyStoragePath} /var/tmp/lastZKey.zkey > /var/tmp/log.txt`,
2361
2367
  `snarkjs zkvi /var/tmp/genesisZkey.zkey /var/tmp/pot.ptau /var/tmp/lastZKey.zkey > /var/tmp/verification_transcript.log`,
2362
2368
  `aws s3 cp /var/tmp/verification_transcript.log s3://${bucketName}/${verificationTranscriptStoragePathAndFilename} &>/dev/null`,
@@ -2383,7 +2389,7 @@ const computeDiskSizeForVM = (zKeySizeInBytes, pot) => Math.ceil(2 * convertByte
2383
2389
  */
2384
2390
  const createEC2Instance = async (ec2, commands, instanceType, volumeSize, diskType) => {
2385
2391
  // Get the AWS variables.
2386
- const { amiId, roleArn } = getAWSVariables();
2392
+ const { amiId, instanceProfileArn } = getAWSVariables();
2387
2393
  // Parametrize the VM EC2 instance.
2388
2394
  const params = {
2389
2395
  ImageId: amiId,
@@ -2392,7 +2398,7 @@ const createEC2Instance = async (ec2, commands, instanceType, volumeSize, diskTy
2392
2398
  MinCount: 1,
2393
2399
  // nb. to find this: iam -> roles -> role_name.
2394
2400
  IamInstanceProfile: {
2395
- Arn: roleArn
2401
+ Arn: instanceProfileArn
2396
2402
  },
2397
2403
  // nb. for running commands at the startup.
2398
2404
  UserData: Buffer.from(commands.join("\n")).toString("base64"),