@devtion/actions 0.0.0-101d43f → 0.0.0-2319823
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/index.mjs +238 -227
- package/dist/index.node.js +238 -227
- package/dist/types/src/helpers/security.d.ts +1 -1
- package/dist/types/src/helpers/security.d.ts.map +1 -1
- package/dist/types/src/helpers/storage.d.ts +5 -2
- package/dist/types/src/helpers/storage.d.ts.map +1 -1
- package/dist/types/src/helpers/utils.d.ts +19 -19
- package/dist/types/src/helpers/utils.d.ts.map +1 -1
- package/dist/types/src/helpers/vm.d.ts.map +1 -1
- package/dist/types/src/types/index.d.ts +2 -2
- package/dist/types/src/types/index.d.ts.map +1 -1
- package/package.json +2 -6
- package/src/helpers/constants.ts +1 -1
- package/src/helpers/functions.ts +1 -1
- package/src/helpers/security.ts +11 -10
- package/src/helpers/services.ts +3 -3
- package/src/helpers/storage.ts +15 -3
- package/src/helpers/utils.ts +262 -263
- package/src/helpers/vm.ts +11 -5
- package/src/index.ts +2 -2
- package/src/types/index.ts +11 -7
package/dist/index.node.js
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* @module @
|
|
3
|
-
* @version 1.0
|
|
2
|
+
* @module @devtion/actions
|
|
3
|
+
* @version 1.1.0
|
|
4
4
|
* @file A set of actions and helpers for CLI commands
|
|
5
5
|
* @copyright Ethereum Foundation 2022
|
|
6
6
|
* @license MIT
|
|
@@ -19,7 +19,6 @@ var crypto = require('crypto');
|
|
|
19
19
|
var blake = require('blakejs');
|
|
20
20
|
var ffjavascript = require('ffjavascript');
|
|
21
21
|
var winston = require('winston');
|
|
22
|
-
var clientS3 = require('@aws-sdk/client-s3');
|
|
23
22
|
var stream = require('stream');
|
|
24
23
|
var util = require('util');
|
|
25
24
|
var app = require('firebase/app');
|
|
@@ -342,7 +341,7 @@ const commonTerms = {
|
|
|
342
341
|
finalizeCircuit: "finalizeCircuit",
|
|
343
342
|
finalizeCeremony: "finalizeCeremony",
|
|
344
343
|
downloadCircuitArtifacts: "downloadCircuitArtifacts",
|
|
345
|
-
transferObject: "transferObject"
|
|
344
|
+
transferObject: "transferObject"
|
|
346
345
|
}
|
|
347
346
|
};
|
|
348
347
|
|
|
@@ -693,11 +692,15 @@ const getChunksAndPreSignedUrls = async (cloudFunctions, bucketName, objectKey,
|
|
|
693
692
|
* @param cloudFunctions <Functions> - the Firebase Cloud Functions service instance.
|
|
694
693
|
* @param ceremonyId <string> - the unique identifier of the ceremony.
|
|
695
694
|
* @param alreadyUploadedChunks Array<ETagWithPartNumber> - the temporary information about the already uploaded chunks.
|
|
695
|
+
* @param logger <GenericBar> - an optional logger to show progress.
|
|
696
696
|
* @returns <Promise<Array<ETagWithPartNumber>>> - the completed (uploaded) chunks information.
|
|
697
697
|
*/
|
|
698
|
-
const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremonyId, alreadyUploadedChunks) => {
|
|
698
|
+
const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremonyId, alreadyUploadedChunks, logger) => {
|
|
699
699
|
// Keep track of uploaded chunks.
|
|
700
700
|
const uploadedChunks = alreadyUploadedChunks || [];
|
|
701
|
+
// if we were passed a logger, start it
|
|
702
|
+
if (logger)
|
|
703
|
+
logger.start(chunksWithUrls.length, 0);
|
|
701
704
|
// Loop through remaining chunks.
|
|
702
705
|
for (let i = alreadyUploadedChunks ? alreadyUploadedChunks.length : 0; i < chunksWithUrls.length; i += 1) {
|
|
703
706
|
// Consume the pre-signed url to upload the chunk.
|
|
@@ -729,6 +732,9 @@ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremony
|
|
|
729
732
|
// nb. this must be done only when contributing (not finalizing).
|
|
730
733
|
if (!!ceremonyId && !!cloudFunctions)
|
|
731
734
|
await temporaryStoreCurrentContributionUploadedChunkData(cloudFunctions, ceremonyId, chunk);
|
|
735
|
+
// increment the count on the logger
|
|
736
|
+
if (logger)
|
|
737
|
+
logger.increment();
|
|
732
738
|
}
|
|
733
739
|
return uploadedChunks;
|
|
734
740
|
};
|
|
@@ -749,8 +755,9 @@ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremony
|
|
|
749
755
|
* @param configStreamChunkSize <number> - size of each chunk into which the artifact is going to be splitted (nb. will be converted in MB).
|
|
750
756
|
* @param [ceremonyId] <string> - the unique identifier of the ceremony (used as a double-edge sword - as identifier and as a check if current contributor is the coordinator finalizing the ceremony).
|
|
751
757
|
* @param [temporaryDataToResumeMultiPartUpload] <TemporaryParticipantContributionData> - the temporary information necessary to resume an already started multi-part upload.
|
|
758
|
+
* @param logger <GenericBar> - an optional logger to show progress.
|
|
752
759
|
*/
|
|
753
|
-
const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFilePath, configStreamChunkSize, ceremonyId, temporaryDataToResumeMultiPartUpload) => {
|
|
760
|
+
const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFilePath, configStreamChunkSize, ceremonyId, temporaryDataToResumeMultiPartUpload, logger) => {
|
|
754
761
|
// The unique identifier of the multi-part upload.
|
|
755
762
|
let multiPartUploadId = "";
|
|
756
763
|
// The list of already uploaded chunks.
|
|
@@ -774,7 +781,7 @@ const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFileP
|
|
|
774
781
|
const chunksWithUrlsZkey = await getChunksAndPreSignedUrls(cloudFunctions, bucketName, objectKey, localFilePath, multiPartUploadId, configStreamChunkSize, ceremonyId);
|
|
775
782
|
// Step (2).
|
|
776
783
|
const partNumbersAndETagsZkey = await uploadParts(chunksWithUrlsZkey, mime.lookup(localFilePath), // content-type.
|
|
777
|
-
cloudFunctions, ceremonyId, alreadyUploadedChunks);
|
|
784
|
+
cloudFunctions, ceremonyId, alreadyUploadedChunks, logger);
|
|
778
785
|
// Step (3).
|
|
779
786
|
await completeMultiPartUpload(cloudFunctions, bucketName, objectKey, multiPartUploadId, partNumbersAndETagsZkey, ceremonyId);
|
|
780
787
|
};
|
|
@@ -1046,199 +1053,22 @@ const compareHashes = async (path1, path2) => {
|
|
|
1046
1053
|
};
|
|
1047
1054
|
|
|
1048
1055
|
/**
|
|
1049
|
-
*
|
|
1050
|
-
* @
|
|
1051
|
-
* @
|
|
1052
|
-
* @param cleanup <boolean> - whether to delete the r1cs file after parsing
|
|
1053
|
-
* @returns any - the data to pass to the cloud function for setup and the circuit artifacts
|
|
1056
|
+
* Return a string with double digits if the provided input is one digit only.
|
|
1057
|
+
* @param in <number> - the input number to be converted.
|
|
1058
|
+
* @returns <string> - the two digits stringified number derived from the conversion.
|
|
1054
1059
|
*/
|
|
1055
|
-
const
|
|
1056
|
-
|
|
1057
|
-
|
|
1058
|
-
|
|
1059
|
-
|
|
1060
|
-
|
|
1061
|
-
|
|
1062
|
-
|
|
1063
|
-
|
|
1064
|
-
|
|
1065
|
-
|
|
1066
|
-
|
|
1067
|
-
throw new Error("You need to provide the data for at least 1 circuit.");
|
|
1068
|
-
// validate that the end date is in the future
|
|
1069
|
-
let endDate;
|
|
1070
|
-
let startDate;
|
|
1071
|
-
try {
|
|
1072
|
-
endDate = new Date(data.endDate);
|
|
1073
|
-
startDate = new Date(data.startDate);
|
|
1074
|
-
}
|
|
1075
|
-
catch (error) {
|
|
1076
|
-
throw new Error("The dates should follow this format: 2023-07-04T00:00:00.");
|
|
1077
|
-
}
|
|
1078
|
-
if (endDate <= startDate)
|
|
1079
|
-
throw new Error("The end date should be greater than the start date.");
|
|
1080
|
-
const currentDate = new Date();
|
|
1081
|
-
if (endDate <= currentDate || startDate <= currentDate)
|
|
1082
|
-
throw new Error("The start and end dates should be in the future.");
|
|
1083
|
-
// validate penalty
|
|
1084
|
-
if (data.penalty <= 0)
|
|
1085
|
-
throw new Error("The penalty should be greater than zero.");
|
|
1086
|
-
const circuits = [];
|
|
1087
|
-
const urlPattern = /(https?:\/\/[^\s]+)/g;
|
|
1088
|
-
const commitHashPattern = /^[a-f0-9]{40}$/i;
|
|
1089
|
-
const circuitArtifacts = [];
|
|
1090
|
-
for (let i = 0; i < data.circuits.length; i++) {
|
|
1091
|
-
const circuitData = data.circuits[i];
|
|
1092
|
-
const artifacts = circuitData.artifacts;
|
|
1093
|
-
circuitArtifacts.push({
|
|
1094
|
-
artifacts: artifacts
|
|
1095
|
-
});
|
|
1096
|
-
// where we storing the r1cs downloaded
|
|
1097
|
-
const localR1csPath = `./${circuitData.name}.r1cs`;
|
|
1098
|
-
// where we storing the wasm downloaded
|
|
1099
|
-
const localWasmPath = `./${circuitData.name}.wasm`;
|
|
1100
|
-
// check that the artifacts exist in S3
|
|
1101
|
-
// we don't need any privileges to download this
|
|
1102
|
-
// just the correct region
|
|
1103
|
-
const s3 = new clientS3.S3Client({
|
|
1104
|
-
region: artifacts.region,
|
|
1105
|
-
credentials: undefined
|
|
1106
|
-
});
|
|
1107
|
-
// download the r1cs to extract the metadata
|
|
1108
|
-
const command = new clientS3.GetObjectCommand({ Bucket: artifacts.bucket, Key: artifacts.r1csStoragePath });
|
|
1109
|
-
const response = await s3.send(command);
|
|
1110
|
-
const streamPipeline = util.promisify(stream.pipeline);
|
|
1111
|
-
if (response.$metadata.httpStatusCode !== 200)
|
|
1112
|
-
throw new Error(`There was an error while trying to download the r1cs file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
|
|
1113
|
-
if (response.Body instanceof stream.Readable)
|
|
1114
|
-
await streamPipeline(response.Body, fs.createWriteStream(localR1csPath));
|
|
1115
|
-
// extract the metadata from the r1cs
|
|
1116
|
-
const metadata = getR1CSInfo(localR1csPath);
|
|
1117
|
-
// download wasm too to ensure it's available
|
|
1118
|
-
const wasmCommand = new clientS3.GetObjectCommand({ Bucket: artifacts.bucket, Key: artifacts.wasmStoragePath });
|
|
1119
|
-
const wasmResponse = await s3.send(wasmCommand);
|
|
1120
|
-
if (wasmResponse.$metadata.httpStatusCode !== 200)
|
|
1121
|
-
throw new Error(`There was an error while trying to download the wasm file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
|
|
1122
|
-
if (wasmResponse.Body instanceof stream.Readable)
|
|
1123
|
-
await streamPipeline(wasmResponse.Body, fs.createWriteStream(localWasmPath));
|
|
1124
|
-
// validate that the circuit hash and template links are valid
|
|
1125
|
-
const template = circuitData.template;
|
|
1126
|
-
const URLMatch = template.source.match(urlPattern);
|
|
1127
|
-
if (!URLMatch || URLMatch.length === 0 || URLMatch.length > 1)
|
|
1128
|
-
throw new Error("You should provide the URL to the circuits templates on GitHub.");
|
|
1129
|
-
const hashMatch = template.commitHash.match(commitHashPattern);
|
|
1130
|
-
if (!hashMatch || hashMatch.length === 0 || hashMatch.length > 1)
|
|
1131
|
-
throw new Error("You should provide a valid commit hash of the circuit templates.");
|
|
1132
|
-
// calculate the hash of the r1cs file
|
|
1133
|
-
const r1csBlake2bHash = await blake512FromPath(localR1csPath);
|
|
1134
|
-
const circuitPrefix = extractPrefix(circuitData.name);
|
|
1135
|
-
// filenames
|
|
1136
|
-
const doubleDigitsPowers = convertToDoubleDigits(metadata.pot);
|
|
1137
|
-
const r1csCompleteFilename = `${circuitData.name}.r1cs`;
|
|
1138
|
-
const wasmCompleteFilename = `${circuitData.name}.wasm`;
|
|
1139
|
-
const smallestPowersOfTauCompleteFilenameForCircuit = `${potFilenameTemplate}${doubleDigitsPowers}.ptau`;
|
|
1140
|
-
const firstZkeyCompleteFilename = `${circuitPrefix}_${genesisZkeyIndex}.zkey`;
|
|
1141
|
-
// storage paths
|
|
1142
|
-
const r1csStorageFilePath = getR1csStorageFilePath(circuitPrefix, r1csCompleteFilename);
|
|
1143
|
-
const wasmStorageFilePath = getWasmStorageFilePath(circuitPrefix, wasmCompleteFilename);
|
|
1144
|
-
const potStorageFilePath = getPotStorageFilePath(smallestPowersOfTauCompleteFilenameForCircuit);
|
|
1145
|
-
const zkeyStorageFilePath = getZkeyStorageFilePath(circuitPrefix, firstZkeyCompleteFilename);
|
|
1146
|
-
const files = {
|
|
1147
|
-
potFilename: smallestPowersOfTauCompleteFilenameForCircuit,
|
|
1148
|
-
r1csFilename: r1csCompleteFilename,
|
|
1149
|
-
wasmFilename: wasmCompleteFilename,
|
|
1150
|
-
initialZkeyFilename: firstZkeyCompleteFilename,
|
|
1151
|
-
potStoragePath: potStorageFilePath,
|
|
1152
|
-
r1csStoragePath: r1csStorageFilePath,
|
|
1153
|
-
wasmStoragePath: wasmStorageFilePath,
|
|
1154
|
-
initialZkeyStoragePath: zkeyStorageFilePath,
|
|
1155
|
-
r1csBlake2bHash: r1csBlake2bHash
|
|
1156
|
-
};
|
|
1157
|
-
// validate that the compiler hash is a valid hash
|
|
1158
|
-
const compiler = circuitData.compiler;
|
|
1159
|
-
const compilerHashMatch = compiler.commitHash.match(commitHashPattern);
|
|
1160
|
-
if (!compilerHashMatch || compilerHashMatch.length === 0 || compilerHashMatch.length > 1)
|
|
1161
|
-
throw new Error("You should provide a valid commit hash of the circuit compiler.");
|
|
1162
|
-
// validate that the verification options are valid
|
|
1163
|
-
const verification = circuitData.verification;
|
|
1164
|
-
if (verification.cfOrVm !== "CF" && verification.cfOrVm !== "VM")
|
|
1165
|
-
throw new Error("Please enter a valid verification mechanism: either CF or VM");
|
|
1166
|
-
// @todo VM parameters verification
|
|
1167
|
-
// if (verification['cfOrVM'] === "VM") {}
|
|
1168
|
-
// check that the timeout is provided for the correct configuration
|
|
1169
|
-
let dynamicThreshold;
|
|
1170
|
-
let fixedTimeWindow;
|
|
1171
|
-
let circuit = {};
|
|
1172
|
-
if (data.timeoutMechanismType === "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */) {
|
|
1173
|
-
if (circuitData.dynamicThreshold <= 0)
|
|
1174
|
-
throw new Error("The dynamic threshold should be > 0.");
|
|
1175
|
-
dynamicThreshold = circuitData.dynamicThreshold;
|
|
1176
|
-
// the Circuit data for the ceremony setup
|
|
1177
|
-
circuit = {
|
|
1178
|
-
name: circuitData.name,
|
|
1179
|
-
description: circuitData.description,
|
|
1180
|
-
prefix: circuitPrefix,
|
|
1181
|
-
sequencePosition: i + 1,
|
|
1182
|
-
metadata: metadata,
|
|
1183
|
-
files: files,
|
|
1184
|
-
template: template,
|
|
1185
|
-
compiler: compiler,
|
|
1186
|
-
verification: verification,
|
|
1187
|
-
dynamicThreshold: dynamicThreshold,
|
|
1188
|
-
avgTimings: {
|
|
1189
|
-
contributionComputation: 0,
|
|
1190
|
-
fullContribution: 0,
|
|
1191
|
-
verifyCloudFunction: 0
|
|
1192
|
-
},
|
|
1193
|
-
};
|
|
1194
|
-
}
|
|
1195
|
-
if (data.timeoutMechanismType === "FIXED" /* CeremonyTimeoutType.FIXED */) {
|
|
1196
|
-
if (circuitData.fixedTimeWindow <= 0)
|
|
1197
|
-
throw new Error("The fixed time window threshold should be > 0.");
|
|
1198
|
-
fixedTimeWindow = circuitData.fixedTimeWindow;
|
|
1199
|
-
// the Circuit data for the ceremony setup
|
|
1200
|
-
circuit = {
|
|
1201
|
-
name: circuitData.name,
|
|
1202
|
-
description: circuitData.description,
|
|
1203
|
-
prefix: circuitPrefix,
|
|
1204
|
-
sequencePosition: i + 1,
|
|
1205
|
-
metadata: metadata,
|
|
1206
|
-
files: files,
|
|
1207
|
-
template: template,
|
|
1208
|
-
compiler: compiler,
|
|
1209
|
-
verification: verification,
|
|
1210
|
-
fixedTimeWindow: fixedTimeWindow,
|
|
1211
|
-
avgTimings: {
|
|
1212
|
-
contributionComputation: 0,
|
|
1213
|
-
fullContribution: 0,
|
|
1214
|
-
verifyCloudFunction: 0
|
|
1215
|
-
},
|
|
1216
|
-
};
|
|
1217
|
-
}
|
|
1218
|
-
circuits.push(circuit);
|
|
1219
|
-
// remove the local r1cs download (if used for verifying the config only vs setup)
|
|
1220
|
-
if (cleanup)
|
|
1221
|
-
fs.unlinkSync(localR1csPath);
|
|
1222
|
-
}
|
|
1223
|
-
const setupData = {
|
|
1224
|
-
ceremonyInputData: {
|
|
1225
|
-
title: data.title,
|
|
1226
|
-
description: data.description,
|
|
1227
|
-
startDate: startDate.valueOf(),
|
|
1228
|
-
endDate: endDate.valueOf(),
|
|
1229
|
-
timeoutMechanismType: data.timeoutMechanismType,
|
|
1230
|
-
penalty: data.penalty
|
|
1231
|
-
},
|
|
1232
|
-
ceremonyPrefix: extractPrefix(data.title),
|
|
1233
|
-
circuits: circuits,
|
|
1234
|
-
circuitArtifacts: circuitArtifacts
|
|
1235
|
-
};
|
|
1236
|
-
return setupData;
|
|
1237
|
-
}
|
|
1238
|
-
catch (error) {
|
|
1239
|
-
throw new Error(`Error while parsing up the ceremony setup file. ${error.message}`);
|
|
1240
|
-
}
|
|
1241
|
-
};
|
|
1060
|
+
const convertToDoubleDigits = (amount) => (amount < 10 ? `0${amount}` : amount.toString());
|
|
1061
|
+
/**
|
|
1062
|
+
* Extract a prefix consisting of alphanumeric and underscore characters from a string with arbitrary characters.
|
|
1063
|
+
* @dev replaces all special symbols and whitespaces with an underscore char ('_'). Convert all uppercase chars to lowercase.
|
|
1064
|
+
* @notice example: str = 'Multiplier-2!2.4.zkey'; output prefix = 'multiplier_2_2_4.zkey'.
|
|
1065
|
+
* NB. Prefix extraction is a key process that conditions the name of the ceremony artifacts, download/upload from/to storage, collections paths.
|
|
1066
|
+
* @param str <string> - the arbitrary string from which to extract the prefix.
|
|
1067
|
+
* @returns <string> - the resulting prefix.
|
|
1068
|
+
*/
|
|
1069
|
+
const extractPrefix = (str) =>
|
|
1070
|
+
// eslint-disable-next-line no-useless-escape
|
|
1071
|
+
str.replace(/[`\s~!@#$%^&*()|+\-=?;:'",.<>\{\}\[\]\\\/]/gi, "-").toLowerCase();
|
|
1242
1072
|
/**
|
|
1243
1073
|
* Extract data from a R1CS metadata file generated with a custom file-based logger.
|
|
1244
1074
|
* @notice useful for extracting metadata circuits contained in the generated file using a logger
|
|
@@ -1295,17 +1125,6 @@ const formatZkeyIndex = (progress) => {
|
|
|
1295
1125
|
* @returns <number> - the amount of powers.
|
|
1296
1126
|
*/
|
|
1297
1127
|
const extractPoTFromFilename = (potCompleteFilename) => Number(potCompleteFilename.split("_").pop()?.split(".").at(0));
|
|
1298
|
-
/**
|
|
1299
|
-
* Extract a prefix consisting of alphanumeric and underscore characters from a string with arbitrary characters.
|
|
1300
|
-
* @dev replaces all special symbols and whitespaces with an underscore char ('_'). Convert all uppercase chars to lowercase.
|
|
1301
|
-
* @notice example: str = 'Multiplier-2!2.4.zkey'; output prefix = 'multiplier_2_2_4.zkey'.
|
|
1302
|
-
* NB. Prefix extraction is a key process that conditions the name of the ceremony artifacts, download/upload from/to storage, collections paths.
|
|
1303
|
-
* @param str <string> - the arbitrary string from which to extract the prefix.
|
|
1304
|
-
* @returns <string> - the resulting prefix.
|
|
1305
|
-
*/
|
|
1306
|
-
const extractPrefix = (str) =>
|
|
1307
|
-
// eslint-disable-next-line no-useless-escape
|
|
1308
|
-
str.replace(/[`\s~!@#$%^&*()|+\-=?;:'",.<>\{\}\[\]\\\/]/gi, "-").toLowerCase();
|
|
1309
1128
|
/**
|
|
1310
1129
|
* Automate the generation of an entropy for a contribution.
|
|
1311
1130
|
* @dev Took inspiration from here https://github.com/glamperd/setup-mpc-ui/blob/master/client/src/state/Compute.tsx#L112.
|
|
@@ -1372,7 +1191,9 @@ const getContributionsValidityForContributor = async (firestoreDatabase, circuit
|
|
|
1372
1191
|
* @param isFinalizing <boolean> - true when the coordinator is finalizing the ceremony, otherwise false.
|
|
1373
1192
|
* @returns <string> - the public attestation preamble.
|
|
1374
1193
|
*/
|
|
1375
|
-
const getPublicAttestationPreambleForContributor = (contributorIdentifier, ceremonyName, isFinalizing) => `Hey, I'm ${contributorIdentifier} and I have ${isFinalizing ? "finalized" : "contributed to"} the ${ceremonyName}
|
|
1194
|
+
const getPublicAttestationPreambleForContributor = (contributorIdentifier, ceremonyName, isFinalizing) => `Hey, I'm ${contributorIdentifier} and I have ${isFinalizing ? "finalized" : "contributed to"} the ${ceremonyName}${ceremonyName.toLowerCase().includes("trusted setup") || ceremonyName.toLowerCase().includes("ceremony")
|
|
1195
|
+
? "."
|
|
1196
|
+
: " MPC Phase2 Trusted Setup ceremony."}\nThe following are my contribution signatures:`;
|
|
1376
1197
|
/**
|
|
1377
1198
|
* Check and prepare public attestation for the contributor made only of its valid contributions.
|
|
1378
1199
|
* @param firestoreDatabase <Firestore> - the Firestore service instance associated to the current Firebase application.
|
|
@@ -1575,11 +1396,193 @@ const getR1CSInfo = (localR1CSFilePath) => {
|
|
|
1575
1396
|
}
|
|
1576
1397
|
};
|
|
1577
1398
|
/**
|
|
1578
|
-
*
|
|
1579
|
-
* @
|
|
1580
|
-
* @
|
|
1399
|
+
* Parse and validate that the ceremony configuration is correct
|
|
1400
|
+
* @notice this does not upload any files to storage
|
|
1401
|
+
* @param path <string> - the path to the configuration file
|
|
1402
|
+
* @param cleanup <boolean> - whether to delete the r1cs file after parsing
|
|
1403
|
+
* @returns any - the data to pass to the cloud function for setup and the circuit artifacts
|
|
1581
1404
|
*/
|
|
1582
|
-
const
|
|
1405
|
+
const parseCeremonyFile = async (path, cleanup = false) => {
|
|
1406
|
+
// check that the path exists
|
|
1407
|
+
if (!fs.existsSync(path))
|
|
1408
|
+
throw new Error("The provided path to the configuration file does not exist. Please provide an absolute path and try again.");
|
|
1409
|
+
try {
|
|
1410
|
+
// read the data
|
|
1411
|
+
const data = JSON.parse(fs.readFileSync(path).toString());
|
|
1412
|
+
// verify that the data is correct
|
|
1413
|
+
if (data.timeoutMechanismType !== "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */ &&
|
|
1414
|
+
data.timeoutMechanismType !== "FIXED" /* CeremonyTimeoutType.FIXED */)
|
|
1415
|
+
throw new Error("Invalid timeout type. Please choose between DYNAMIC and FIXED.");
|
|
1416
|
+
// validate that we have at least 1 circuit input data
|
|
1417
|
+
if (!data.circuits || data.circuits.length === 0)
|
|
1418
|
+
throw new Error("You need to provide the data for at least 1 circuit.");
|
|
1419
|
+
// validate that the end date is in the future
|
|
1420
|
+
let endDate;
|
|
1421
|
+
let startDate;
|
|
1422
|
+
try {
|
|
1423
|
+
endDate = new Date(data.endDate);
|
|
1424
|
+
startDate = new Date(data.startDate);
|
|
1425
|
+
}
|
|
1426
|
+
catch (error) {
|
|
1427
|
+
throw new Error("The dates should follow this format: 2023-07-04T00:00:00.");
|
|
1428
|
+
}
|
|
1429
|
+
if (endDate <= startDate)
|
|
1430
|
+
throw new Error("The end date should be greater than the start date.");
|
|
1431
|
+
const currentDate = new Date();
|
|
1432
|
+
if (endDate <= currentDate || startDate <= currentDate)
|
|
1433
|
+
throw new Error("The start and end dates should be in the future.");
|
|
1434
|
+
// validate penalty
|
|
1435
|
+
if (data.penalty <= 0)
|
|
1436
|
+
throw new Error("The penalty should be greater than zero.");
|
|
1437
|
+
const circuits = [];
|
|
1438
|
+
const urlPattern = /(https?:\/\/[^\s]+)/g;
|
|
1439
|
+
const commitHashPattern = /^[a-f0-9]{40}$/i;
|
|
1440
|
+
const circuitArtifacts = [];
|
|
1441
|
+
for (let i = 0; i < data.circuits.length; i++) {
|
|
1442
|
+
const circuitData = data.circuits[i];
|
|
1443
|
+
const { artifacts } = circuitData;
|
|
1444
|
+
circuitArtifacts.push({
|
|
1445
|
+
artifacts
|
|
1446
|
+
});
|
|
1447
|
+
// where we storing the r1cs downloaded
|
|
1448
|
+
const localR1csPath = `./${circuitData.name}.r1cs`;
|
|
1449
|
+
// where we storing the wasm downloaded
|
|
1450
|
+
const localWasmPath = `./${circuitData.name}.wasm`;
|
|
1451
|
+
// download the r1cs to extract the metadata
|
|
1452
|
+
const streamPipeline = util.promisify(stream.pipeline);
|
|
1453
|
+
// Make the call.
|
|
1454
|
+
const responseR1CS = await fetch(artifacts.r1csStoragePath);
|
|
1455
|
+
// Handle errors.
|
|
1456
|
+
if (!responseR1CS.ok && responseR1CS.status !== 200)
|
|
1457
|
+
throw new Error(`There was an error while trying to download the r1cs file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
|
|
1458
|
+
await streamPipeline(responseR1CS.body, fs.createWriteStream(localR1csPath));
|
|
1459
|
+
// Write the file locally
|
|
1460
|
+
// extract the metadata from the r1cs
|
|
1461
|
+
const metadata = getR1CSInfo(localR1csPath);
|
|
1462
|
+
// download wasm too to ensure it's available
|
|
1463
|
+
const responseWASM = await fetch(artifacts.wasmStoragePath);
|
|
1464
|
+
if (!responseWASM.ok && responseWASM.status !== 200)
|
|
1465
|
+
throw new Error(`There was an error while trying to download the WASM file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
|
|
1466
|
+
await streamPipeline(responseWASM.body, fs.createWriteStream(localWasmPath));
|
|
1467
|
+
// validate that the circuit hash and template links are valid
|
|
1468
|
+
const { template } = circuitData;
|
|
1469
|
+
const URLMatch = template.source.match(urlPattern);
|
|
1470
|
+
if (!URLMatch || URLMatch.length === 0 || URLMatch.length > 1)
|
|
1471
|
+
throw new Error("You should provide the URL to the circuits templates on GitHub.");
|
|
1472
|
+
const hashMatch = template.commitHash.match(commitHashPattern);
|
|
1473
|
+
if (!hashMatch || hashMatch.length === 0 || hashMatch.length > 1)
|
|
1474
|
+
throw new Error("You should provide a valid commit hash of the circuit templates.");
|
|
1475
|
+
// calculate the hash of the r1cs file
|
|
1476
|
+
const r1csBlake2bHash = await blake512FromPath(localR1csPath);
|
|
1477
|
+
const circuitPrefix = extractPrefix(circuitData.name);
|
|
1478
|
+
// filenames
|
|
1479
|
+
const doubleDigitsPowers = convertToDoubleDigits(metadata.pot);
|
|
1480
|
+
const r1csCompleteFilename = `${circuitData.name}.r1cs`;
|
|
1481
|
+
const wasmCompleteFilename = `${circuitData.name}.wasm`;
|
|
1482
|
+
const smallestPowersOfTauCompleteFilenameForCircuit = `${potFilenameTemplate}${doubleDigitsPowers}.ptau`;
|
|
1483
|
+
const firstZkeyCompleteFilename = `${circuitPrefix}_${genesisZkeyIndex}.zkey`;
|
|
1484
|
+
// storage paths
|
|
1485
|
+
const r1csStorageFilePath = getR1csStorageFilePath(circuitPrefix, r1csCompleteFilename);
|
|
1486
|
+
const wasmStorageFilePath = getWasmStorageFilePath(circuitPrefix, wasmCompleteFilename);
|
|
1487
|
+
const potStorageFilePath = getPotStorageFilePath(smallestPowersOfTauCompleteFilenameForCircuit);
|
|
1488
|
+
const zkeyStorageFilePath = getZkeyStorageFilePath(circuitPrefix, firstZkeyCompleteFilename);
|
|
1489
|
+
const files = {
|
|
1490
|
+
potFilename: smallestPowersOfTauCompleteFilenameForCircuit,
|
|
1491
|
+
r1csFilename: r1csCompleteFilename,
|
|
1492
|
+
wasmFilename: wasmCompleteFilename,
|
|
1493
|
+
initialZkeyFilename: firstZkeyCompleteFilename,
|
|
1494
|
+
potStoragePath: potStorageFilePath,
|
|
1495
|
+
r1csStoragePath: r1csStorageFilePath,
|
|
1496
|
+
wasmStoragePath: wasmStorageFilePath,
|
|
1497
|
+
initialZkeyStoragePath: zkeyStorageFilePath,
|
|
1498
|
+
r1csBlake2bHash
|
|
1499
|
+
};
|
|
1500
|
+
// validate that the compiler hash is a valid hash
|
|
1501
|
+
const { compiler } = circuitData;
|
|
1502
|
+
const compilerHashMatch = compiler.commitHash.match(commitHashPattern);
|
|
1503
|
+
if (!compilerHashMatch || compilerHashMatch.length === 0 || compilerHashMatch.length > 1)
|
|
1504
|
+
throw new Error("You should provide a valid commit hash of the circuit compiler.");
|
|
1505
|
+
// validate that the verification options are valid
|
|
1506
|
+
const { verification } = circuitData;
|
|
1507
|
+
if (verification.cfOrVm !== "CF" && verification.cfOrVm !== "VM")
|
|
1508
|
+
throw new Error("Please enter a valid verification mechanism: either CF or VM");
|
|
1509
|
+
// @todo VM parameters verification
|
|
1510
|
+
// if (verification['cfOrVM'] === "VM") {}
|
|
1511
|
+
// check that the timeout is provided for the correct configuration
|
|
1512
|
+
let dynamicThreshold;
|
|
1513
|
+
let fixedTimeWindow;
|
|
1514
|
+
let circuit = {};
|
|
1515
|
+
if (data.timeoutMechanismType === "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */) {
|
|
1516
|
+
if (circuitData.dynamicThreshold <= 0)
|
|
1517
|
+
throw new Error("The dynamic threshold should be > 0.");
|
|
1518
|
+
dynamicThreshold = circuitData.dynamicThreshold;
|
|
1519
|
+
// the Circuit data for the ceremony setup
|
|
1520
|
+
circuit = {
|
|
1521
|
+
name: circuitData.name,
|
|
1522
|
+
description: circuitData.description,
|
|
1523
|
+
prefix: circuitPrefix,
|
|
1524
|
+
sequencePosition: i + 1,
|
|
1525
|
+
metadata,
|
|
1526
|
+
files,
|
|
1527
|
+
template,
|
|
1528
|
+
compiler,
|
|
1529
|
+
verification,
|
|
1530
|
+
dynamicThreshold,
|
|
1531
|
+
avgTimings: {
|
|
1532
|
+
contributionComputation: 0,
|
|
1533
|
+
fullContribution: 0,
|
|
1534
|
+
verifyCloudFunction: 0
|
|
1535
|
+
}
|
|
1536
|
+
};
|
|
1537
|
+
}
|
|
1538
|
+
if (data.timeoutMechanismType === "FIXED" /* CeremonyTimeoutType.FIXED */) {
|
|
1539
|
+
if (circuitData.fixedTimeWindow <= 0)
|
|
1540
|
+
throw new Error("The fixed time window threshold should be > 0.");
|
|
1541
|
+
fixedTimeWindow = circuitData.fixedTimeWindow;
|
|
1542
|
+
// the Circuit data for the ceremony setup
|
|
1543
|
+
circuit = {
|
|
1544
|
+
name: circuitData.name,
|
|
1545
|
+
description: circuitData.description,
|
|
1546
|
+
prefix: circuitPrefix,
|
|
1547
|
+
sequencePosition: i + 1,
|
|
1548
|
+
metadata,
|
|
1549
|
+
files,
|
|
1550
|
+
template,
|
|
1551
|
+
compiler,
|
|
1552
|
+
verification,
|
|
1553
|
+
fixedTimeWindow,
|
|
1554
|
+
avgTimings: {
|
|
1555
|
+
contributionComputation: 0,
|
|
1556
|
+
fullContribution: 0,
|
|
1557
|
+
verifyCloudFunction: 0
|
|
1558
|
+
}
|
|
1559
|
+
};
|
|
1560
|
+
}
|
|
1561
|
+
circuits.push(circuit);
|
|
1562
|
+
// remove the local r1cs and wasm downloads (if used for verifying the config only vs setup)
|
|
1563
|
+
if (cleanup)
|
|
1564
|
+
fs.unlinkSync(localR1csPath);
|
|
1565
|
+
fs.unlinkSync(localWasmPath);
|
|
1566
|
+
}
|
|
1567
|
+
const setupData = {
|
|
1568
|
+
ceremonyInputData: {
|
|
1569
|
+
title: data.title,
|
|
1570
|
+
description: data.description,
|
|
1571
|
+
startDate: startDate.valueOf(),
|
|
1572
|
+
endDate: endDate.valueOf(),
|
|
1573
|
+
timeoutMechanismType: data.timeoutMechanismType,
|
|
1574
|
+
penalty: data.penalty
|
|
1575
|
+
},
|
|
1576
|
+
ceremonyPrefix: extractPrefix(data.title),
|
|
1577
|
+
circuits,
|
|
1578
|
+
circuitArtifacts
|
|
1579
|
+
};
|
|
1580
|
+
return setupData;
|
|
1581
|
+
}
|
|
1582
|
+
catch (error) {
|
|
1583
|
+
throw new Error(`Error while parsing up the ceremony setup file. ${error.message}`);
|
|
1584
|
+
}
|
|
1585
|
+
};
|
|
1583
1586
|
|
|
1584
1587
|
/**
|
|
1585
1588
|
* Verify that a zKey is valid
|
|
@@ -1828,7 +1831,7 @@ const getFirestoreDatabase = (app) => firestore.getFirestore(app);
|
|
|
1828
1831
|
* @param app <FirebaseApp> - the Firebase application.
|
|
1829
1832
|
* @returns <Functions> - the Cloud Functions associated to the application.
|
|
1830
1833
|
*/
|
|
1831
|
-
const getFirebaseFunctions = (app) => functions.getFunctions(app,
|
|
1834
|
+
const getFirebaseFunctions = (app) => functions.getFunctions(app, "europe-west1");
|
|
1832
1835
|
/**
|
|
1833
1836
|
* Retrieve the configuration variables for the AWS services (S3, EC2).
|
|
1834
1837
|
* @returns <AWSVariables> - the values of the AWS services configuration variables.
|
|
@@ -1837,14 +1840,14 @@ const getAWSVariables = () => {
|
|
|
1837
1840
|
if (!process.env.AWS_ACCESS_KEY_ID ||
|
|
1838
1841
|
!process.env.AWS_SECRET_ACCESS_KEY ||
|
|
1839
1842
|
!process.env.AWS_REGION ||
|
|
1840
|
-
!process.env.
|
|
1843
|
+
!process.env.AWS_INSTANCE_PROFILE_ARN ||
|
|
1841
1844
|
!process.env.AWS_AMI_ID)
|
|
1842
1845
|
throw new Error("Could not retrieve the AWS environment variables. Please, verify your environment configuration and retry");
|
|
1843
1846
|
return {
|
|
1844
1847
|
accessKeyId: process.env.AWS_ACCESS_KEY_ID,
|
|
1845
1848
|
secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY,
|
|
1846
1849
|
region: process.env.AWS_REGION || "us-east-1",
|
|
1847
|
-
|
|
1850
|
+
instanceProfileArn: process.env.AWS_INSTANCE_PROFILE_ARN,
|
|
1848
1851
|
amiId: process.env.AWS_AMI_ID
|
|
1849
1852
|
};
|
|
1850
1853
|
};
|
|
@@ -2097,7 +2100,8 @@ const getGitHubStats = async (user) => {
|
|
|
2097
2100
|
following: jsonData.following,
|
|
2098
2101
|
followers: jsonData.followers,
|
|
2099
2102
|
publicRepos: jsonData.public_repos,
|
|
2100
|
-
avatarUrl: jsonData.avatar_url
|
|
2103
|
+
avatarUrl: jsonData.avatar_url,
|
|
2104
|
+
age: jsonData.created_at
|
|
2101
2105
|
};
|
|
2102
2106
|
return data;
|
|
2103
2107
|
};
|
|
@@ -2109,20 +2113,21 @@ const getGitHubStats = async (user) => {
|
|
|
2109
2113
|
* @param minimumAmountOfPublicRepos <number> The minimum amount of public repos the user should have
|
|
2110
2114
|
* @returns <any> Return the avatar URL of the user if the user is reputable, false otherwise
|
|
2111
2115
|
*/
|
|
2112
|
-
const githubReputation = async (userLogin, minimumAmountOfFollowing, minimumAmountOfFollowers, minimumAmountOfPublicRepos) => {
|
|
2116
|
+
const githubReputation = async (userLogin, minimumAmountOfFollowing, minimumAmountOfFollowers, minimumAmountOfPublicRepos, minimumAge) => {
|
|
2113
2117
|
if (!process.env.GITHUB_ACCESS_TOKEN)
|
|
2114
2118
|
throw new Error("The GitHub access token is missing. Please insert a valid token to be used for anti-sybil checks on user registation, and then try again.");
|
|
2115
|
-
const { following, followers, publicRepos, avatarUrl } = await getGitHubStats(userLogin);
|
|
2119
|
+
const { following, followers, publicRepos, avatarUrl, age } = await getGitHubStats(userLogin);
|
|
2116
2120
|
if (following < minimumAmountOfFollowing ||
|
|
2117
2121
|
publicRepos < minimumAmountOfPublicRepos ||
|
|
2118
|
-
followers < minimumAmountOfFollowers
|
|
2122
|
+
followers < minimumAmountOfFollowers ||
|
|
2123
|
+
new Date(age) > new Date(Date.now() - minimumAge))
|
|
2119
2124
|
return {
|
|
2120
2125
|
reputable: false,
|
|
2121
2126
|
avatarUrl: ""
|
|
2122
2127
|
};
|
|
2123
2128
|
return {
|
|
2124
2129
|
reputable: true,
|
|
2125
|
-
avatarUrl
|
|
2130
|
+
avatarUrl
|
|
2126
2131
|
};
|
|
2127
2132
|
};
|
|
2128
2133
|
|
|
@@ -2331,8 +2336,13 @@ const vmDependenciesAndCacheArtifactsCommand = (zKeyPath, potPath, snsTopic, reg
|
|
|
2331
2336
|
// eslint-disable-next-line no-template-curly-in-string
|
|
2332
2337
|
"touch ${MARKER_FILE}",
|
|
2333
2338
|
"sudo yum update -y",
|
|
2334
|
-
"curl -
|
|
2335
|
-
"
|
|
2339
|
+
"curl -O https://nodejs.org/dist/v16.13.0/node-v16.13.0-linux-x64.tar.xz",
|
|
2340
|
+
"tar -xf node-v16.13.0-linux-x64.tar.xz",
|
|
2341
|
+
"mv node-v16.13.0-linux-x64 nodejs",
|
|
2342
|
+
"sudo mv nodejs /opt/",
|
|
2343
|
+
"echo 'export NODEJS_HOME=/opt/nodejs' >> /etc/profile",
|
|
2344
|
+
"echo 'export PATH=$NODEJS_HOME/bin:$PATH' >> /etc/profile",
|
|
2345
|
+
"source /etc/profile",
|
|
2336
2346
|
"npm install -g snarkjs",
|
|
2337
2347
|
`aws s3 cp s3://${zKeyPath} /var/tmp/genesisZkey.zkey`,
|
|
2338
2348
|
`aws s3 cp s3://${potPath} /var/tmp/pot.ptau`,
|
|
@@ -2351,6 +2361,7 @@ const vmDependenciesAndCacheArtifactsCommand = (zKeyPath, potPath, snsTopic, reg
|
|
|
2351
2361
|
* @returns Array<string> - the list of commands for contribution verification.
|
|
2352
2362
|
*/
|
|
2353
2363
|
const vmContributionVerificationCommand = (bucketName, lastZkeyStoragePath, verificationTranscriptStoragePathAndFilename) => [
|
|
2364
|
+
`source /etc/profile`,
|
|
2354
2365
|
`aws s3 cp s3://${bucketName}/${lastZkeyStoragePath} /var/tmp/lastZKey.zkey > /var/tmp/log.txt`,
|
|
2355
2366
|
`snarkjs zkvi /var/tmp/genesisZkey.zkey /var/tmp/pot.ptau /var/tmp/lastZKey.zkey > /var/tmp/verification_transcript.log`,
|
|
2356
2367
|
`aws s3 cp /var/tmp/verification_transcript.log s3://${bucketName}/${verificationTranscriptStoragePathAndFilename} &>/dev/null`,
|
|
@@ -2377,7 +2388,7 @@ const computeDiskSizeForVM = (zKeySizeInBytes, pot) => Math.ceil(2 * convertByte
|
|
|
2377
2388
|
*/
|
|
2378
2389
|
const createEC2Instance = async (ec2, commands, instanceType, volumeSize, diskType) => {
|
|
2379
2390
|
// Get the AWS variables.
|
|
2380
|
-
const { amiId,
|
|
2391
|
+
const { amiId, instanceProfileArn } = getAWSVariables();
|
|
2381
2392
|
// Parametrize the VM EC2 instance.
|
|
2382
2393
|
const params = {
|
|
2383
2394
|
ImageId: amiId,
|
|
@@ -2386,7 +2397,7 @@ const createEC2Instance = async (ec2, commands, instanceType, volumeSize, diskTy
|
|
|
2386
2397
|
MinCount: 1,
|
|
2387
2398
|
// nb. to find this: iam -> roles -> role_name.
|
|
2388
2399
|
IamInstanceProfile: {
|
|
2389
|
-
Arn:
|
|
2400
|
+
Arn: instanceProfileArn
|
|
2390
2401
|
},
|
|
2391
2402
|
// nb. for running commands at the startup.
|
|
2392
2403
|
UserData: Buffer.from(commands.join("\n")).toString("base64"),
|
|
@@ -6,5 +6,5 @@
|
|
|
6
6
|
* @param minimumAmountOfPublicRepos <number> The minimum amount of public repos the user should have
|
|
7
7
|
* @returns <any> Return the avatar URL of the user if the user is reputable, false otherwise
|
|
8
8
|
*/
|
|
9
|
-
export declare const githubReputation: (userLogin: string, minimumAmountOfFollowing: number, minimumAmountOfFollowers: number, minimumAmountOfPublicRepos: number) => Promise<any>;
|
|
9
|
+
export declare const githubReputation: (userLogin: string, minimumAmountOfFollowing: number, minimumAmountOfFollowers: number, minimumAmountOfPublicRepos: number, minimumAge: number) => Promise<any>;
|
|
10
10
|
//# sourceMappingURL=security.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"security.d.ts","sourceRoot":"","sources":["../../../../src/helpers/security.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"security.d.ts","sourceRoot":"","sources":["../../../../src/helpers/security.ts"],"names":[],"mappings":"AA6BA;;;;;;;GAOG;AACH,eAAO,MAAM,gBAAgB,cACd,MAAM,4BACS,MAAM,4BACN,MAAM,8BACJ,MAAM,cACtB,MAAM,KACnB,QAAQ,GAAG,CAuBb,CAAA"}
|