@devtion/actions 0.0.0-0fd4368 → 0.0.0-142ec0a
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/index.mjs +312 -292
- package/dist/index.node.js +312 -292
- package/dist/types/src/helpers/constants.d.ts +6 -0
- package/dist/types/src/helpers/constants.d.ts.map +1 -1
- package/dist/types/src/helpers/contracts.d.ts.map +1 -1
- package/dist/types/src/helpers/security.d.ts +2 -2
- package/dist/types/src/helpers/security.d.ts.map +1 -1
- package/dist/types/src/helpers/storage.d.ts +5 -2
- package/dist/types/src/helpers/storage.d.ts.map +1 -1
- package/dist/types/src/helpers/utils.d.ts +33 -20
- package/dist/types/src/helpers/utils.d.ts.map +1 -1
- package/dist/types/src/helpers/vm.d.ts.map +1 -1
- package/dist/types/src/types/index.d.ts +9 -3
- package/dist/types/src/types/index.d.ts.map +1 -1
- package/package.json +2 -7
- package/src/helpers/constants.ts +7 -1
- package/src/helpers/contracts.ts +3 -3
- package/src/helpers/functions.ts +1 -1
- package/src/helpers/security.ts +33 -52
- package/src/helpers/services.ts +3 -3
- package/src/helpers/storage.ts +15 -3
- package/src/helpers/utils.ts +316 -277
- package/src/helpers/vm.ts +14 -7
- package/src/index.ts +2 -2
- package/src/types/index.ts +32 -8
package/dist/index.node.js
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* @module @
|
|
3
|
-
* @version 1.
|
|
2
|
+
* @module @devtion/actions
|
|
3
|
+
* @version 1.1.1
|
|
4
4
|
* @file A set of actions and helpers for CLI commands
|
|
5
5
|
* @copyright Ethereum Foundation 2022
|
|
6
6
|
* @license MIT
|
|
@@ -17,9 +17,7 @@ var firestore = require('firebase/firestore');
|
|
|
17
17
|
var snarkjs = require('snarkjs');
|
|
18
18
|
var crypto = require('crypto');
|
|
19
19
|
var blake = require('blakejs');
|
|
20
|
-
var ffjavascript = require('ffjavascript');
|
|
21
20
|
var winston = require('winston');
|
|
22
|
-
var clientS3 = require('@aws-sdk/client-s3');
|
|
23
21
|
var stream = require('stream');
|
|
24
22
|
var util = require('util');
|
|
25
23
|
var app = require('firebase/app');
|
|
@@ -246,6 +244,12 @@ const commonTerms = {
|
|
|
246
244
|
verificationStartedAt: "verificationStartedAt"
|
|
247
245
|
}
|
|
248
246
|
},
|
|
247
|
+
avatars: {
|
|
248
|
+
name: "avatars",
|
|
249
|
+
fields: {
|
|
250
|
+
avatarUrl: "avatarUrl"
|
|
251
|
+
}
|
|
252
|
+
},
|
|
249
253
|
ceremonies: {
|
|
250
254
|
name: "ceremonies",
|
|
251
255
|
fields: {
|
|
@@ -336,7 +340,7 @@ const commonTerms = {
|
|
|
336
340
|
finalizeCircuit: "finalizeCircuit",
|
|
337
341
|
finalizeCeremony: "finalizeCeremony",
|
|
338
342
|
downloadCircuitArtifacts: "downloadCircuitArtifacts",
|
|
339
|
-
transferObject: "transferObject"
|
|
343
|
+
transferObject: "transferObject"
|
|
340
344
|
}
|
|
341
345
|
};
|
|
342
346
|
|
|
@@ -687,11 +691,15 @@ const getChunksAndPreSignedUrls = async (cloudFunctions, bucketName, objectKey,
|
|
|
687
691
|
* @param cloudFunctions <Functions> - the Firebase Cloud Functions service instance.
|
|
688
692
|
* @param ceremonyId <string> - the unique identifier of the ceremony.
|
|
689
693
|
* @param alreadyUploadedChunks Array<ETagWithPartNumber> - the temporary information about the already uploaded chunks.
|
|
694
|
+
* @param logger <GenericBar> - an optional logger to show progress.
|
|
690
695
|
* @returns <Promise<Array<ETagWithPartNumber>>> - the completed (uploaded) chunks information.
|
|
691
696
|
*/
|
|
692
|
-
const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremonyId, alreadyUploadedChunks) => {
|
|
697
|
+
const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremonyId, alreadyUploadedChunks, logger) => {
|
|
693
698
|
// Keep track of uploaded chunks.
|
|
694
699
|
const uploadedChunks = alreadyUploadedChunks || [];
|
|
700
|
+
// if we were passed a logger, start it
|
|
701
|
+
if (logger)
|
|
702
|
+
logger.start(chunksWithUrls.length, 0);
|
|
695
703
|
// Loop through remaining chunks.
|
|
696
704
|
for (let i = alreadyUploadedChunks ? alreadyUploadedChunks.length : 0; i < chunksWithUrls.length; i += 1) {
|
|
697
705
|
// Consume the pre-signed url to upload the chunk.
|
|
@@ -723,6 +731,9 @@ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremony
|
|
|
723
731
|
// nb. this must be done only when contributing (not finalizing).
|
|
724
732
|
if (!!ceremonyId && !!cloudFunctions)
|
|
725
733
|
await temporaryStoreCurrentContributionUploadedChunkData(cloudFunctions, ceremonyId, chunk);
|
|
734
|
+
// increment the count on the logger
|
|
735
|
+
if (logger)
|
|
736
|
+
logger.increment();
|
|
726
737
|
}
|
|
727
738
|
return uploadedChunks;
|
|
728
739
|
};
|
|
@@ -743,8 +754,9 @@ const uploadParts = async (chunksWithUrls, contentType, cloudFunctions, ceremony
|
|
|
743
754
|
* @param configStreamChunkSize <number> - size of each chunk into which the artifact is going to be splitted (nb. will be converted in MB).
|
|
744
755
|
* @param [ceremonyId] <string> - the unique identifier of the ceremony (used as a double-edge sword - as identifier and as a check if current contributor is the coordinator finalizing the ceremony).
|
|
745
756
|
* @param [temporaryDataToResumeMultiPartUpload] <TemporaryParticipantContributionData> - the temporary information necessary to resume an already started multi-part upload.
|
|
757
|
+
* @param logger <GenericBar> - an optional logger to show progress.
|
|
746
758
|
*/
|
|
747
|
-
const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFilePath, configStreamChunkSize, ceremonyId, temporaryDataToResumeMultiPartUpload) => {
|
|
759
|
+
const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFilePath, configStreamChunkSize, ceremonyId, temporaryDataToResumeMultiPartUpload, logger) => {
|
|
748
760
|
// The unique identifier of the multi-part upload.
|
|
749
761
|
let multiPartUploadId = "";
|
|
750
762
|
// The list of already uploaded chunks.
|
|
@@ -768,7 +780,7 @@ const multiPartUpload = async (cloudFunctions, bucketName, objectKey, localFileP
|
|
|
768
780
|
const chunksWithUrlsZkey = await getChunksAndPreSignedUrls(cloudFunctions, bucketName, objectKey, localFilePath, multiPartUploadId, configStreamChunkSize, ceremonyId);
|
|
769
781
|
// Step (2).
|
|
770
782
|
const partNumbersAndETagsZkey = await uploadParts(chunksWithUrlsZkey, mime.lookup(localFilePath), // content-type.
|
|
771
|
-
cloudFunctions, ceremonyId, alreadyUploadedChunks);
|
|
783
|
+
cloudFunctions, ceremonyId, alreadyUploadedChunks, logger);
|
|
772
784
|
// Step (3).
|
|
773
785
|
await completeMultiPartUpload(cloudFunctions, bucketName, objectKey, multiPartUploadId, partNumbersAndETagsZkey, ceremonyId);
|
|
774
786
|
};
|
|
@@ -1040,207 +1052,22 @@ const compareHashes = async (path1, path2) => {
|
|
|
1040
1052
|
};
|
|
1041
1053
|
|
|
1042
1054
|
/**
|
|
1043
|
-
*
|
|
1044
|
-
* @
|
|
1045
|
-
* @
|
|
1046
|
-
* @param cleanup <boolean> - whether to delete the r1cs file after parsing
|
|
1047
|
-
* @returns any - the data to pass to the cloud function for setup and the circuit artifacts
|
|
1055
|
+
* Return a string with double digits if the provided input is one digit only.
|
|
1056
|
+
* @param in <number> - the input number to be converted.
|
|
1057
|
+
* @returns <string> - the two digits stringified number derived from the conversion.
|
|
1048
1058
|
*/
|
|
1049
|
-
const
|
|
1050
|
-
|
|
1051
|
-
|
|
1052
|
-
|
|
1053
|
-
|
|
1054
|
-
|
|
1055
|
-
|
|
1056
|
-
|
|
1057
|
-
|
|
1058
|
-
|
|
1059
|
-
|
|
1060
|
-
|
|
1061
|
-
throw new Error("You need to provide the data for at least 1 circuit.");
|
|
1062
|
-
// validate that the end date is in the future
|
|
1063
|
-
let endDate;
|
|
1064
|
-
let startDate;
|
|
1065
|
-
try {
|
|
1066
|
-
endDate = new Date(data.endDate);
|
|
1067
|
-
startDate = new Date(data.startDate);
|
|
1068
|
-
}
|
|
1069
|
-
catch (error) {
|
|
1070
|
-
throw new Error("The dates should follow this format: 2023-07-04T00:00:00.");
|
|
1071
|
-
}
|
|
1072
|
-
if (endDate <= startDate)
|
|
1073
|
-
throw new Error("The end date should be greater than the start date.");
|
|
1074
|
-
const currentDate = new Date();
|
|
1075
|
-
if (endDate <= currentDate || startDate <= currentDate)
|
|
1076
|
-
throw new Error("The start and end dates should be in the future.");
|
|
1077
|
-
// validate penalty
|
|
1078
|
-
if (data.penalty <= 0)
|
|
1079
|
-
throw new Error("The penalty should be greater than zero.");
|
|
1080
|
-
const circuits = [];
|
|
1081
|
-
const urlPattern = /(https?:\/\/[^\s]+)/g;
|
|
1082
|
-
const commitHashPattern = /^[a-f0-9]{40}$/i;
|
|
1083
|
-
const circuitArtifacts = [];
|
|
1084
|
-
for (let i = 0; i < data.circuits.length; i++) {
|
|
1085
|
-
const circuitData = data.circuits[i];
|
|
1086
|
-
const artifacts = circuitData.artifacts;
|
|
1087
|
-
circuitArtifacts.push({
|
|
1088
|
-
artifacts: artifacts
|
|
1089
|
-
});
|
|
1090
|
-
const r1csPath = artifacts.r1csStoragePath;
|
|
1091
|
-
const wasmPath = artifacts.wasmStoragePath;
|
|
1092
|
-
// where we storing the r1cs downloaded
|
|
1093
|
-
const localR1csPath = `./${circuitData.name}.r1cs`;
|
|
1094
|
-
// check that the artifacts exist in S3
|
|
1095
|
-
// we don't need any privileges to download this
|
|
1096
|
-
// just the correct region
|
|
1097
|
-
const s3 = new clientS3.S3Client({ region: artifacts.region });
|
|
1098
|
-
try {
|
|
1099
|
-
await s3.send(new clientS3.HeadObjectCommand({
|
|
1100
|
-
Bucket: artifacts.bucket,
|
|
1101
|
-
Key: r1csPath
|
|
1102
|
-
}));
|
|
1103
|
-
}
|
|
1104
|
-
catch (error) {
|
|
1105
|
-
throw new Error(`The r1cs file (${r1csPath}) seems to not exist. Please ensure this is correct and that the object is publicly available.`);
|
|
1106
|
-
}
|
|
1107
|
-
try {
|
|
1108
|
-
await s3.send(new clientS3.HeadObjectCommand({
|
|
1109
|
-
Bucket: artifacts.bucket,
|
|
1110
|
-
Key: wasmPath
|
|
1111
|
-
}));
|
|
1112
|
-
}
|
|
1113
|
-
catch (error) {
|
|
1114
|
-
throw new Error(`The wasm file (${wasmPath}) seems to not exist. Please ensure this is correct and that the object is publicly available.`);
|
|
1115
|
-
}
|
|
1116
|
-
// download the r1cs to extract the metadata
|
|
1117
|
-
const command = new clientS3.GetObjectCommand({ Bucket: artifacts.bucket, Key: artifacts.r1csStoragePath });
|
|
1118
|
-
const response = await s3.send(command);
|
|
1119
|
-
const streamPipeline = util.promisify(stream.pipeline);
|
|
1120
|
-
if (response.$metadata.httpStatusCode !== 200)
|
|
1121
|
-
throw new Error("There was an error while trying to download the r1cs file. Please check that the file has the correct permissions (public) set.");
|
|
1122
|
-
if (response.Body instanceof stream.Readable)
|
|
1123
|
-
await streamPipeline(response.Body, fs.createWriteStream(localR1csPath));
|
|
1124
|
-
// extract the metadata from the r1cs
|
|
1125
|
-
const metadata = getR1CSInfo(localR1csPath);
|
|
1126
|
-
// validate that the circuit hash and template links are valid
|
|
1127
|
-
const template = circuitData.template;
|
|
1128
|
-
const URLMatch = template.source.match(urlPattern);
|
|
1129
|
-
if (!URLMatch || URLMatch.length === 0 || URLMatch.length > 1)
|
|
1130
|
-
throw new Error("You should provide the URL to the circuits templates on GitHub.");
|
|
1131
|
-
const hashMatch = template.commitHash.match(commitHashPattern);
|
|
1132
|
-
if (!hashMatch || hashMatch.length === 0 || hashMatch.length > 1)
|
|
1133
|
-
throw new Error("You should provide a valid commit hash of the circuit templates.");
|
|
1134
|
-
// calculate the hash of the r1cs file
|
|
1135
|
-
const r1csBlake2bHash = await blake512FromPath(localR1csPath);
|
|
1136
|
-
const circuitPrefix = extractPrefix(circuitData.name);
|
|
1137
|
-
// filenames
|
|
1138
|
-
const doubleDigitsPowers = convertToDoubleDigits(metadata.pot);
|
|
1139
|
-
const r1csCompleteFilename = `${circuitData.name}.r1cs`;
|
|
1140
|
-
const wasmCompleteFilename = `${circuitData.name}.wasm`;
|
|
1141
|
-
const smallestPowersOfTauCompleteFilenameForCircuit = `${potFilenameTemplate}${doubleDigitsPowers}.ptau`;
|
|
1142
|
-
const firstZkeyCompleteFilename = `${circuitPrefix}_${genesisZkeyIndex}.zkey`;
|
|
1143
|
-
// storage paths
|
|
1144
|
-
const r1csStorageFilePath = getR1csStorageFilePath(circuitPrefix, r1csCompleteFilename);
|
|
1145
|
-
const wasmStorageFilePath = getWasmStorageFilePath(circuitPrefix, wasmCompleteFilename);
|
|
1146
|
-
const potStorageFilePath = getPotStorageFilePath(smallestPowersOfTauCompleteFilenameForCircuit);
|
|
1147
|
-
const zkeyStorageFilePath = getZkeyStorageFilePath(circuitPrefix, firstZkeyCompleteFilename);
|
|
1148
|
-
const files = {
|
|
1149
|
-
potFilename: smallestPowersOfTauCompleteFilenameForCircuit,
|
|
1150
|
-
r1csFilename: r1csCompleteFilename,
|
|
1151
|
-
wasmFilename: wasmCompleteFilename,
|
|
1152
|
-
initialZkeyFilename: firstZkeyCompleteFilename,
|
|
1153
|
-
potStoragePath: potStorageFilePath,
|
|
1154
|
-
r1csStoragePath: r1csStorageFilePath,
|
|
1155
|
-
wasmStoragePath: wasmStorageFilePath,
|
|
1156
|
-
initialZkeyStoragePath: zkeyStorageFilePath,
|
|
1157
|
-
r1csBlake2bHash: r1csBlake2bHash
|
|
1158
|
-
};
|
|
1159
|
-
// validate that the compiler hash is a valid hash
|
|
1160
|
-
const compiler = circuitData.compiler;
|
|
1161
|
-
const compilerHashMatch = compiler.commitHash.match(commitHashPattern);
|
|
1162
|
-
if (!compilerHashMatch || compilerHashMatch.length === 0 || compilerHashMatch.length > 1)
|
|
1163
|
-
throw new Error("You should provide a valid commit hash of the circuit compiler.");
|
|
1164
|
-
// validate that the verification options are valid
|
|
1165
|
-
const verification = circuitData.verification;
|
|
1166
|
-
if (verification.cfOrVm !== "CF" && verification.cfOrVm !== "VM")
|
|
1167
|
-
throw new Error("Please enter a valid verification mechanism: either CF or VM");
|
|
1168
|
-
// @todo VM parameters verification
|
|
1169
|
-
// if (verification['cfOrVM'] === "VM") {}
|
|
1170
|
-
// check that the timeout is provided for the correct configuration
|
|
1171
|
-
let dynamicThreshold;
|
|
1172
|
-
let fixedTimeWindow;
|
|
1173
|
-
let circuit = {};
|
|
1174
|
-
if (data.timeoutMechanismType === "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */) {
|
|
1175
|
-
if (circuitData.dynamicThreshold <= 0)
|
|
1176
|
-
throw new Error("The dynamic threshold should be > 0.");
|
|
1177
|
-
dynamicThreshold = circuitData.dynamicThreshold;
|
|
1178
|
-
// the Circuit data for the ceremony setup
|
|
1179
|
-
circuit = {
|
|
1180
|
-
name: circuitData.name,
|
|
1181
|
-
description: circuitData.description,
|
|
1182
|
-
prefix: circuitPrefix,
|
|
1183
|
-
sequencePosition: i + 1,
|
|
1184
|
-
metadata: metadata,
|
|
1185
|
-
files: files,
|
|
1186
|
-
template: template,
|
|
1187
|
-
compiler: compiler,
|
|
1188
|
-
verification: verification,
|
|
1189
|
-
dynamicThreshold: dynamicThreshold,
|
|
1190
|
-
avgTimings: {
|
|
1191
|
-
contributionComputation: 0,
|
|
1192
|
-
fullContribution: 0,
|
|
1193
|
-
verifyCloudFunction: 0
|
|
1194
|
-
},
|
|
1195
|
-
};
|
|
1196
|
-
}
|
|
1197
|
-
if (data.timeoutMechanismType === "FIXED" /* CeremonyTimeoutType.FIXED */) {
|
|
1198
|
-
if (circuitData.fixedTimeWindow <= 0)
|
|
1199
|
-
throw new Error("The fixed time window threshold should be > 0.");
|
|
1200
|
-
fixedTimeWindow = circuitData.fixedTimeWindow;
|
|
1201
|
-
// the Circuit data for the ceremony setup
|
|
1202
|
-
circuit = {
|
|
1203
|
-
name: circuitData.name,
|
|
1204
|
-
description: circuitData.description,
|
|
1205
|
-
prefix: circuitPrefix,
|
|
1206
|
-
sequencePosition: i + 1,
|
|
1207
|
-
metadata: metadata,
|
|
1208
|
-
files: files,
|
|
1209
|
-
template: template,
|
|
1210
|
-
compiler: compiler,
|
|
1211
|
-
verification: verification,
|
|
1212
|
-
fixedTimeWindow: fixedTimeWindow,
|
|
1213
|
-
avgTimings: {
|
|
1214
|
-
contributionComputation: 0,
|
|
1215
|
-
fullContribution: 0,
|
|
1216
|
-
verifyCloudFunction: 0
|
|
1217
|
-
},
|
|
1218
|
-
};
|
|
1219
|
-
}
|
|
1220
|
-
circuits.push(circuit);
|
|
1221
|
-
// remove the local r1cs download (if used for verifying the config only vs setup)
|
|
1222
|
-
if (cleanup)
|
|
1223
|
-
fs.unlinkSync(localR1csPath);
|
|
1224
|
-
}
|
|
1225
|
-
const setupData = {
|
|
1226
|
-
ceremonyInputData: {
|
|
1227
|
-
title: data.title,
|
|
1228
|
-
description: data.description,
|
|
1229
|
-
startDate: startDate.valueOf(),
|
|
1230
|
-
endDate: endDate.valueOf(),
|
|
1231
|
-
timeoutMechanismType: data.timeoutMechanismType,
|
|
1232
|
-
penalty: data.penalty
|
|
1233
|
-
},
|
|
1234
|
-
ceremonyPrefix: extractPrefix(data.title),
|
|
1235
|
-
circuits: circuits,
|
|
1236
|
-
circuitArtifacts: circuitArtifacts
|
|
1237
|
-
};
|
|
1238
|
-
return setupData;
|
|
1239
|
-
}
|
|
1240
|
-
catch (error) {
|
|
1241
|
-
throw new Error(`Error while parsing up the ceremony setup file. ${error.message}`);
|
|
1242
|
-
}
|
|
1243
|
-
};
|
|
1059
|
+
const convertToDoubleDigits = (amount) => (amount < 10 ? `0${amount}` : amount.toString());
|
|
1060
|
+
/**
|
|
1061
|
+
* Extract a prefix consisting of alphanumeric and underscore characters from a string with arbitrary characters.
|
|
1062
|
+
* @dev replaces all special symbols and whitespaces with an underscore char ('_'). Convert all uppercase chars to lowercase.
|
|
1063
|
+
* @notice example: str = 'Multiplier-2!2.4.zkey'; output prefix = 'multiplier_2_2_4.zkey'.
|
|
1064
|
+
* NB. Prefix extraction is a key process that conditions the name of the ceremony artifacts, download/upload from/to storage, collections paths.
|
|
1065
|
+
* @param str <string> - the arbitrary string from which to extract the prefix.
|
|
1066
|
+
* @returns <string> - the resulting prefix.
|
|
1067
|
+
*/
|
|
1068
|
+
const extractPrefix = (str) =>
|
|
1069
|
+
// eslint-disable-next-line no-useless-escape
|
|
1070
|
+
str.replace(/[`\s~!@#$%^&*()|+\-=?;:'",.<>\{\}\[\]\\\/]/gi, "-").toLowerCase();
|
|
1244
1071
|
/**
|
|
1245
1072
|
* Extract data from a R1CS metadata file generated with a custom file-based logger.
|
|
1246
1073
|
* @notice useful for extracting metadata circuits contained in the generated file using a logger
|
|
@@ -1297,17 +1124,6 @@ const formatZkeyIndex = (progress) => {
|
|
|
1297
1124
|
* @returns <number> - the amount of powers.
|
|
1298
1125
|
*/
|
|
1299
1126
|
const extractPoTFromFilename = (potCompleteFilename) => Number(potCompleteFilename.split("_").pop()?.split(".").at(0));
|
|
1300
|
-
/**
|
|
1301
|
-
* Extract a prefix consisting of alphanumeric and underscore characters from a string with arbitrary characters.
|
|
1302
|
-
* @dev replaces all special symbols and whitespaces with an underscore char ('_'). Convert all uppercase chars to lowercase.
|
|
1303
|
-
* @notice example: str = 'Multiplier-2!2.4.zkey'; output prefix = 'multiplier_2_2_4.zkey'.
|
|
1304
|
-
* NB. Prefix extraction is a key process that conditions the name of the ceremony artifacts, download/upload from/to storage, collections paths.
|
|
1305
|
-
* @param str <string> - the arbitrary string from which to extract the prefix.
|
|
1306
|
-
* @returns <string> - the resulting prefix.
|
|
1307
|
-
*/
|
|
1308
|
-
const extractPrefix = (str) =>
|
|
1309
|
-
// eslint-disable-next-line no-useless-escape
|
|
1310
|
-
str.replace(/[`\s~!@#$%^&*()|+\-=?;:'",.<>\{\}\[\]\\\/]/gi, "-").toLowerCase();
|
|
1311
1127
|
/**
|
|
1312
1128
|
* Automate the generation of an entropy for a contribution.
|
|
1313
1129
|
* @dev Took inspiration from here https://github.com/glamperd/setup-mpc-ui/blob/master/client/src/state/Compute.tsx#L112.
|
|
@@ -1374,7 +1190,9 @@ const getContributionsValidityForContributor = async (firestoreDatabase, circuit
|
|
|
1374
1190
|
* @param isFinalizing <boolean> - true when the coordinator is finalizing the ceremony, otherwise false.
|
|
1375
1191
|
* @returns <string> - the public attestation preamble.
|
|
1376
1192
|
*/
|
|
1377
|
-
const getPublicAttestationPreambleForContributor = (contributorIdentifier, ceremonyName, isFinalizing) => `Hey, I'm ${contributorIdentifier} and I have ${isFinalizing ? "finalized" : "contributed to"} the ${ceremonyName}
|
|
1193
|
+
const getPublicAttestationPreambleForContributor = (contributorIdentifier, ceremonyName, isFinalizing) => `Hey, I'm ${contributorIdentifier} and I have ${isFinalizing ? "finalized" : "contributed to"} the ${ceremonyName}${ceremonyName.toLowerCase().includes("trusted setup") || ceremonyName.toLowerCase().includes("ceremony")
|
|
1194
|
+
? "."
|
|
1195
|
+
: " MPC Phase2 Trusted Setup ceremony."}\nThe following are my contribution signatures:`;
|
|
1378
1196
|
/**
|
|
1379
1197
|
* Check and prepare public attestation for the contributor made only of its valid contributions.
|
|
1380
1198
|
* @param firestoreDatabase <Firestore> - the Firestore service instance associated to the current Firebase application.
|
|
@@ -1445,6 +1263,41 @@ const readBytesFromFile = (localFilePath, offset, length, position) => {
|
|
|
1445
1263
|
// Return the read bytes.
|
|
1446
1264
|
return buffer;
|
|
1447
1265
|
};
|
|
1266
|
+
/**
|
|
1267
|
+
* Given a buffer in little endian format, convert it to bigint
|
|
1268
|
+
* @param buffer
|
|
1269
|
+
* @returns
|
|
1270
|
+
*/
|
|
1271
|
+
function leBufferToBigint(buffer) {
|
|
1272
|
+
return BigInt(`0x${buffer.reverse().toString("hex")}`);
|
|
1273
|
+
}
|
|
1274
|
+
/**
|
|
1275
|
+
* Given an input containing string values, convert them
|
|
1276
|
+
* to bigint
|
|
1277
|
+
* @param input - The input to convert
|
|
1278
|
+
* @returns the input with string values converted to bigint
|
|
1279
|
+
*/
|
|
1280
|
+
const unstringifyBigInts = (input) => {
|
|
1281
|
+
if (typeof input === "string" && /^[0-9]+$/.test(input)) {
|
|
1282
|
+
return BigInt(input);
|
|
1283
|
+
}
|
|
1284
|
+
if (typeof input === "string" && /^0x[0-9a-fA-F]+$/.test(input)) {
|
|
1285
|
+
return BigInt(input);
|
|
1286
|
+
}
|
|
1287
|
+
if (Array.isArray(input)) {
|
|
1288
|
+
return input.map(unstringifyBigInts);
|
|
1289
|
+
}
|
|
1290
|
+
if (input === null) {
|
|
1291
|
+
return null;
|
|
1292
|
+
}
|
|
1293
|
+
if (typeof input === "object") {
|
|
1294
|
+
return Object.entries(input).reduce((acc, [key, value]) => {
|
|
1295
|
+
acc[key] = unstringifyBigInts(value);
|
|
1296
|
+
return acc;
|
|
1297
|
+
}, {});
|
|
1298
|
+
}
|
|
1299
|
+
return input;
|
|
1300
|
+
};
|
|
1448
1301
|
/**
|
|
1449
1302
|
* Return the info about the R1CS file.ù
|
|
1450
1303
|
* @dev this method was built taking inspiration from
|
|
@@ -1505,17 +1358,17 @@ const getR1CSInfo = (localR1CSFilePath) => {
|
|
|
1505
1358
|
let constraints = 0;
|
|
1506
1359
|
try {
|
|
1507
1360
|
// Get 'number of section' (jump magic r1cs and version1 data).
|
|
1508
|
-
const numberOfSections =
|
|
1361
|
+
const numberOfSections = leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, 8));
|
|
1509
1362
|
// Jump to first section.
|
|
1510
1363
|
pointer = 12;
|
|
1511
1364
|
// For each section
|
|
1512
1365
|
for (let i = 0; i < numberOfSections; i++) {
|
|
1513
1366
|
// Read section type.
|
|
1514
|
-
const sectionType =
|
|
1367
|
+
const sectionType = leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, pointer));
|
|
1515
1368
|
// Jump to section size.
|
|
1516
1369
|
pointer += 4;
|
|
1517
1370
|
// Read section size
|
|
1518
|
-
const sectionSize = Number(
|
|
1371
|
+
const sectionSize = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 8, pointer)));
|
|
1519
1372
|
// If at header section (0x00000001 : Header Section).
|
|
1520
1373
|
if (sectionType === BigInt(1)) {
|
|
1521
1374
|
// Read info from header section.
|
|
@@ -1547,17 +1400,17 @@ const getR1CSInfo = (localR1CSFilePath) => {
|
|
|
1547
1400
|
*/
|
|
1548
1401
|
pointer += sectionSize - 20;
|
|
1549
1402
|
// Read R1CS info.
|
|
1550
|
-
wires = Number(
|
|
1403
|
+
wires = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, pointer)));
|
|
1551
1404
|
pointer += 4;
|
|
1552
|
-
publicOutputs = Number(
|
|
1405
|
+
publicOutputs = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, pointer)));
|
|
1553
1406
|
pointer += 4;
|
|
1554
|
-
publicInputs = Number(
|
|
1407
|
+
publicInputs = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, pointer)));
|
|
1555
1408
|
pointer += 4;
|
|
1556
|
-
privateInputs = Number(
|
|
1409
|
+
privateInputs = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, pointer)));
|
|
1557
1410
|
pointer += 4;
|
|
1558
|
-
labels = Number(
|
|
1411
|
+
labels = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 8, pointer)));
|
|
1559
1412
|
pointer += 8;
|
|
1560
|
-
constraints = Number(
|
|
1413
|
+
constraints = Number(leBufferToBigint(readBytesFromFile(localR1CSFilePath, 0, 4, pointer)));
|
|
1561
1414
|
}
|
|
1562
1415
|
pointer += 8 + Number(sectionSize);
|
|
1563
1416
|
}
|
|
@@ -1577,11 +1430,194 @@ const getR1CSInfo = (localR1CSFilePath) => {
|
|
|
1577
1430
|
}
|
|
1578
1431
|
};
|
|
1579
1432
|
/**
|
|
1580
|
-
*
|
|
1581
|
-
* @
|
|
1582
|
-
* @
|
|
1433
|
+
* Parse and validate that the ceremony configuration is correct
|
|
1434
|
+
* @notice this does not upload any files to storage
|
|
1435
|
+
* @param path <string> - the path to the configuration file
|
|
1436
|
+
* @param cleanup <boolean> - whether to delete the r1cs file after parsing
|
|
1437
|
+
* @returns any - the data to pass to the cloud function for setup and the circuit artifacts
|
|
1583
1438
|
*/
|
|
1584
|
-
const
|
|
1439
|
+
const parseCeremonyFile = async (path, cleanup = false) => {
|
|
1440
|
+
// check that the path exists
|
|
1441
|
+
if (!fs.existsSync(path))
|
|
1442
|
+
throw new Error("The provided path to the configuration file does not exist. Please provide an absolute path and try again.");
|
|
1443
|
+
try {
|
|
1444
|
+
// read the data
|
|
1445
|
+
const data = JSON.parse(fs.readFileSync(path).toString());
|
|
1446
|
+
// verify that the data is correct
|
|
1447
|
+
if (data.timeoutMechanismType !== "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */ &&
|
|
1448
|
+
data.timeoutMechanismType !== "FIXED" /* CeremonyTimeoutType.FIXED */)
|
|
1449
|
+
throw new Error("Invalid timeout type. Please choose between DYNAMIC and FIXED.");
|
|
1450
|
+
// validate that we have at least 1 circuit input data
|
|
1451
|
+
if (!data.circuits || data.circuits.length === 0)
|
|
1452
|
+
throw new Error("You need to provide the data for at least 1 circuit.");
|
|
1453
|
+
// validate that the end date is in the future
|
|
1454
|
+
let endDate;
|
|
1455
|
+
let startDate;
|
|
1456
|
+
try {
|
|
1457
|
+
endDate = new Date(data.endDate);
|
|
1458
|
+
startDate = new Date(data.startDate);
|
|
1459
|
+
}
|
|
1460
|
+
catch (error) {
|
|
1461
|
+
throw new Error("The dates should follow this format: 2023-07-04T00:00:00.");
|
|
1462
|
+
}
|
|
1463
|
+
if (endDate <= startDate)
|
|
1464
|
+
throw new Error("The end date should be greater than the start date.");
|
|
1465
|
+
const currentDate = new Date();
|
|
1466
|
+
if (endDate <= currentDate || startDate <= currentDate)
|
|
1467
|
+
throw new Error("The start and end dates should be in the future.");
|
|
1468
|
+
// validate penalty
|
|
1469
|
+
if (data.penalty <= 0)
|
|
1470
|
+
throw new Error("The penalty should be greater than zero.");
|
|
1471
|
+
const circuits = [];
|
|
1472
|
+
const urlPattern = /(https?:\/\/[^\s]+)/g;
|
|
1473
|
+
const commitHashPattern = /^[a-f0-9]{40}$/i;
|
|
1474
|
+
const circuitArtifacts = [];
|
|
1475
|
+
for (let i = 0; i < data.circuits.length; i++) {
|
|
1476
|
+
const circuitData = data.circuits[i];
|
|
1477
|
+
const { artifacts } = circuitData;
|
|
1478
|
+
circuitArtifacts.push({
|
|
1479
|
+
artifacts
|
|
1480
|
+
});
|
|
1481
|
+
// where we storing the r1cs downloaded
|
|
1482
|
+
const localR1csPath = `./${circuitData.name}.r1cs`;
|
|
1483
|
+
// where we storing the wasm downloaded
|
|
1484
|
+
const localWasmPath = `./${circuitData.name}.wasm`;
|
|
1485
|
+
// download the r1cs to extract the metadata
|
|
1486
|
+
const streamPipeline = util.promisify(stream.pipeline);
|
|
1487
|
+
// Make the call.
|
|
1488
|
+
const responseR1CS = await fetch(artifacts.r1csStoragePath);
|
|
1489
|
+
// Handle errors.
|
|
1490
|
+
if (!responseR1CS.ok && responseR1CS.status !== 200)
|
|
1491
|
+
throw new Error(`There was an error while trying to download the r1cs file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
|
|
1492
|
+
await streamPipeline(responseR1CS.body, fs.createWriteStream(localR1csPath));
|
|
1493
|
+
// Write the file locally
|
|
1494
|
+
// extract the metadata from the r1cs
|
|
1495
|
+
const metadata = getR1CSInfo(localR1csPath);
|
|
1496
|
+
// download wasm too to ensure it's available
|
|
1497
|
+
const responseWASM = await fetch(artifacts.wasmStoragePath);
|
|
1498
|
+
if (!responseWASM.ok && responseWASM.status !== 200)
|
|
1499
|
+
throw new Error(`There was an error while trying to download the WASM file for circuit ${circuitData.name}. Please check that the file has the correct permissions (public) set.`);
|
|
1500
|
+
await streamPipeline(responseWASM.body, fs.createWriteStream(localWasmPath));
|
|
1501
|
+
// validate that the circuit hash and template links are valid
|
|
1502
|
+
const { template } = circuitData;
|
|
1503
|
+
const URLMatch = template.source.match(urlPattern);
|
|
1504
|
+
if (!URLMatch || URLMatch.length === 0 || URLMatch.length > 1)
|
|
1505
|
+
throw new Error("You should provide the URL to the circuits templates on GitHub.");
|
|
1506
|
+
const hashMatch = template.commitHash.match(commitHashPattern);
|
|
1507
|
+
if (!hashMatch || hashMatch.length === 0 || hashMatch.length > 1)
|
|
1508
|
+
throw new Error("You should provide a valid commit hash of the circuit templates.");
|
|
1509
|
+
// calculate the hash of the r1cs file
|
|
1510
|
+
const r1csBlake2bHash = await blake512FromPath(localR1csPath);
|
|
1511
|
+
const circuitPrefix = extractPrefix(circuitData.name);
|
|
1512
|
+
// filenames
|
|
1513
|
+
const doubleDigitsPowers = convertToDoubleDigits(metadata.pot);
|
|
1514
|
+
const r1csCompleteFilename = `${circuitData.name}.r1cs`;
|
|
1515
|
+
const wasmCompleteFilename = `${circuitData.name}.wasm`;
|
|
1516
|
+
const smallestPowersOfTauCompleteFilenameForCircuit = `${potFilenameTemplate}${doubleDigitsPowers}.ptau`;
|
|
1517
|
+
const firstZkeyCompleteFilename = `${circuitPrefix}_${genesisZkeyIndex}.zkey`;
|
|
1518
|
+
// storage paths
|
|
1519
|
+
const r1csStorageFilePath = getR1csStorageFilePath(circuitPrefix, r1csCompleteFilename);
|
|
1520
|
+
const wasmStorageFilePath = getWasmStorageFilePath(circuitPrefix, wasmCompleteFilename);
|
|
1521
|
+
const potStorageFilePath = getPotStorageFilePath(smallestPowersOfTauCompleteFilenameForCircuit);
|
|
1522
|
+
const zkeyStorageFilePath = getZkeyStorageFilePath(circuitPrefix, firstZkeyCompleteFilename);
|
|
1523
|
+
const files = {
|
|
1524
|
+
potFilename: smallestPowersOfTauCompleteFilenameForCircuit,
|
|
1525
|
+
r1csFilename: r1csCompleteFilename,
|
|
1526
|
+
wasmFilename: wasmCompleteFilename,
|
|
1527
|
+
initialZkeyFilename: firstZkeyCompleteFilename,
|
|
1528
|
+
potStoragePath: potStorageFilePath,
|
|
1529
|
+
r1csStoragePath: r1csStorageFilePath,
|
|
1530
|
+
wasmStoragePath: wasmStorageFilePath,
|
|
1531
|
+
initialZkeyStoragePath: zkeyStorageFilePath,
|
|
1532
|
+
r1csBlake2bHash
|
|
1533
|
+
};
|
|
1534
|
+
// validate that the compiler hash is a valid hash
|
|
1535
|
+
const { compiler } = circuitData;
|
|
1536
|
+
const compilerHashMatch = compiler.commitHash.match(commitHashPattern);
|
|
1537
|
+
if (!compilerHashMatch || compilerHashMatch.length === 0 || compilerHashMatch.length > 1)
|
|
1538
|
+
throw new Error("You should provide a valid commit hash of the circuit compiler.");
|
|
1539
|
+
// validate that the verification options are valid
|
|
1540
|
+
const { verification } = circuitData;
|
|
1541
|
+
if (verification.cfOrVm !== "CF" && verification.cfOrVm !== "VM")
|
|
1542
|
+
throw new Error("Please enter a valid verification mechanism: either CF or VM");
|
|
1543
|
+
// @todo VM parameters verification
|
|
1544
|
+
// if (verification['cfOrVM'] === "VM") {}
|
|
1545
|
+
// check that the timeout is provided for the correct configuration
|
|
1546
|
+
let dynamicThreshold;
|
|
1547
|
+
let fixedTimeWindow;
|
|
1548
|
+
let circuit = {};
|
|
1549
|
+
if (data.timeoutMechanismType === "DYNAMIC" /* CeremonyTimeoutType.DYNAMIC */) {
|
|
1550
|
+
if (circuitData.dynamicThreshold <= 0)
|
|
1551
|
+
throw new Error("The dynamic threshold should be > 0.");
|
|
1552
|
+
dynamicThreshold = circuitData.dynamicThreshold;
|
|
1553
|
+
// the Circuit data for the ceremony setup
|
|
1554
|
+
circuit = {
|
|
1555
|
+
name: circuitData.name,
|
|
1556
|
+
description: circuitData.description,
|
|
1557
|
+
prefix: circuitPrefix,
|
|
1558
|
+
sequencePosition: i + 1,
|
|
1559
|
+
metadata,
|
|
1560
|
+
files,
|
|
1561
|
+
template,
|
|
1562
|
+
compiler,
|
|
1563
|
+
verification,
|
|
1564
|
+
dynamicThreshold,
|
|
1565
|
+
avgTimings: {
|
|
1566
|
+
contributionComputation: 0,
|
|
1567
|
+
fullContribution: 0,
|
|
1568
|
+
verifyCloudFunction: 0
|
|
1569
|
+
}
|
|
1570
|
+
};
|
|
1571
|
+
}
|
|
1572
|
+
if (data.timeoutMechanismType === "FIXED" /* CeremonyTimeoutType.FIXED */) {
|
|
1573
|
+
if (circuitData.fixedTimeWindow <= 0)
|
|
1574
|
+
throw new Error("The fixed time window threshold should be > 0.");
|
|
1575
|
+
fixedTimeWindow = circuitData.fixedTimeWindow;
|
|
1576
|
+
// the Circuit data for the ceremony setup
|
|
1577
|
+
circuit = {
|
|
1578
|
+
name: circuitData.name,
|
|
1579
|
+
description: circuitData.description,
|
|
1580
|
+
prefix: circuitPrefix,
|
|
1581
|
+
sequencePosition: i + 1,
|
|
1582
|
+
metadata,
|
|
1583
|
+
files,
|
|
1584
|
+
template,
|
|
1585
|
+
compiler,
|
|
1586
|
+
verification,
|
|
1587
|
+
fixedTimeWindow,
|
|
1588
|
+
avgTimings: {
|
|
1589
|
+
contributionComputation: 0,
|
|
1590
|
+
fullContribution: 0,
|
|
1591
|
+
verifyCloudFunction: 0
|
|
1592
|
+
}
|
|
1593
|
+
};
|
|
1594
|
+
}
|
|
1595
|
+
circuits.push(circuit);
|
|
1596
|
+
// remove the local r1cs and wasm downloads (if used for verifying the config only vs setup)
|
|
1597
|
+
if (cleanup) {
|
|
1598
|
+
fs.unlinkSync(localR1csPath);
|
|
1599
|
+
fs.unlinkSync(localWasmPath);
|
|
1600
|
+
}
|
|
1601
|
+
}
|
|
1602
|
+
const setupData = {
|
|
1603
|
+
ceremonyInputData: {
|
|
1604
|
+
title: data.title,
|
|
1605
|
+
description: data.description,
|
|
1606
|
+
startDate: startDate.valueOf(),
|
|
1607
|
+
endDate: endDate.valueOf(),
|
|
1608
|
+
timeoutMechanismType: data.timeoutMechanismType,
|
|
1609
|
+
penalty: data.penalty
|
|
1610
|
+
},
|
|
1611
|
+
ceremonyPrefix: extractPrefix(data.title),
|
|
1612
|
+
circuits,
|
|
1613
|
+
circuitArtifacts
|
|
1614
|
+
};
|
|
1615
|
+
return setupData;
|
|
1616
|
+
}
|
|
1617
|
+
catch (error) {
|
|
1618
|
+
throw new Error(`Error while parsing up the ceremony setup file. ${error.message}`);
|
|
1619
|
+
}
|
|
1620
|
+
};
|
|
1585
1621
|
|
|
1586
1622
|
/**
|
|
1587
1623
|
* Verify that a zKey is valid
|
|
@@ -1830,7 +1866,7 @@ const getFirestoreDatabase = (app) => firestore.getFirestore(app);
|
|
|
1830
1866
|
* @param app <FirebaseApp> - the Firebase application.
|
|
1831
1867
|
* @returns <Functions> - the Cloud Functions associated to the application.
|
|
1832
1868
|
*/
|
|
1833
|
-
const getFirebaseFunctions = (app) => functions.getFunctions(app,
|
|
1869
|
+
const getFirebaseFunctions = (app) => functions.getFunctions(app, "europe-west1");
|
|
1834
1870
|
/**
|
|
1835
1871
|
* Retrieve the configuration variables for the AWS services (S3, EC2).
|
|
1836
1872
|
* @returns <AWSVariables> - the values of the AWS services configuration variables.
|
|
@@ -1839,14 +1875,14 @@ const getAWSVariables = () => {
|
|
|
1839
1875
|
if (!process.env.AWS_ACCESS_KEY_ID ||
|
|
1840
1876
|
!process.env.AWS_SECRET_ACCESS_KEY ||
|
|
1841
1877
|
!process.env.AWS_REGION ||
|
|
1842
|
-
!process.env.
|
|
1878
|
+
!process.env.AWS_INSTANCE_PROFILE_ARN ||
|
|
1843
1879
|
!process.env.AWS_AMI_ID)
|
|
1844
1880
|
throw new Error("Could not retrieve the AWS environment variables. Please, verify your environment configuration and retry");
|
|
1845
1881
|
return {
|
|
1846
1882
|
accessKeyId: process.env.AWS_ACCESS_KEY_ID,
|
|
1847
1883
|
secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY,
|
|
1848
1884
|
region: process.env.AWS_REGION || "us-east-1",
|
|
1849
|
-
|
|
1885
|
+
instanceProfileArn: process.env.AWS_INSTANCE_PROFILE_ARN,
|
|
1850
1886
|
amiId: process.env.AWS_AMI_ID
|
|
1851
1887
|
};
|
|
1852
1888
|
};
|
|
@@ -1927,11 +1963,11 @@ const p256 = (proofPart) => {
|
|
|
1927
1963
|
*/
|
|
1928
1964
|
const formatSolidityCalldata = (circuitInput, _proof) => {
|
|
1929
1965
|
try {
|
|
1930
|
-
const proof =
|
|
1966
|
+
const proof = unstringifyBigInts(_proof);
|
|
1931
1967
|
// format the public inputs to the circuit
|
|
1932
1968
|
const formattedCircuitInput = [];
|
|
1933
1969
|
for (const cInput of circuitInput) {
|
|
1934
|
-
formattedCircuitInput.push(p256(
|
|
1970
|
+
formattedCircuitInput.push(p256(unstringifyBigInts(cInput)));
|
|
1935
1971
|
}
|
|
1936
1972
|
// construct calldata
|
|
1937
1973
|
const calldata = {
|
|
@@ -2081,55 +2117,28 @@ const verifyCeremony = async (functions, firestore$1, ceremonyPrefix, outputDire
|
|
|
2081
2117
|
};
|
|
2082
2118
|
|
|
2083
2119
|
/**
|
|
2084
|
-
* This function
|
|
2085
|
-
* @param user
|
|
2086
|
-
* @returns
|
|
2120
|
+
* This function queries the GitHub API to fetch users statistics
|
|
2121
|
+
* @param user {string} the user uid
|
|
2122
|
+
* @returns {any} the stats from the GitHub API
|
|
2087
2123
|
*/
|
|
2088
|
-
const
|
|
2089
|
-
const response = await fetch(`https://api.github.com/user/${user}
|
|
2124
|
+
const getGitHubStats = async (user) => {
|
|
2125
|
+
const response = await fetch(`https://api.github.com/user/${user}`, {
|
|
2090
2126
|
method: "GET",
|
|
2091
2127
|
headers: {
|
|
2092
2128
|
Authorization: `token ${process.env.GITHUB_ACCESS_TOKEN}`
|
|
2093
2129
|
}
|
|
2094
2130
|
});
|
|
2095
2131
|
if (response.status !== 200)
|
|
2096
|
-
throw new Error("It was not possible to retrieve the
|
|
2132
|
+
throw new Error("It was not possible to retrieve the user's statistic. Please try again.");
|
|
2097
2133
|
const jsonData = await response.json();
|
|
2098
|
-
|
|
2099
|
-
|
|
2100
|
-
|
|
2101
|
-
|
|
2102
|
-
|
|
2103
|
-
|
|
2104
|
-
|
|
2105
|
-
|
|
2106
|
-
const response = await fetch(`https://api.github.com/user/${user}/followers`, {
|
|
2107
|
-
method: "GET",
|
|
2108
|
-
headers: {
|
|
2109
|
-
Authorization: `token ${process.env.GITHUB_ACCESS_TOKEN}`
|
|
2110
|
-
}
|
|
2111
|
-
});
|
|
2112
|
-
if (response.status !== 200)
|
|
2113
|
-
throw new Error("It was not possible to retrieve the number of followers. Please try again.");
|
|
2114
|
-
const jsonData = await response.json();
|
|
2115
|
-
return jsonData.length;
|
|
2116
|
-
};
|
|
2117
|
-
/**
|
|
2118
|
-
* This function will return the number of following of a user
|
|
2119
|
-
* @param user <string> The username of the user
|
|
2120
|
-
* @returns <number> The number of following users
|
|
2121
|
-
*/
|
|
2122
|
-
const getNumberOfFollowingGitHub = async (user) => {
|
|
2123
|
-
const response = await fetch(`https://api.github.com/user/${user}/following`, {
|
|
2124
|
-
method: "GET",
|
|
2125
|
-
headers: {
|
|
2126
|
-
Authorization: `token ${process.env.GITHUB_ACCESS_TOKEN}`
|
|
2127
|
-
}
|
|
2128
|
-
});
|
|
2129
|
-
if (response.status !== 200)
|
|
2130
|
-
throw new Error("It was not possible to retrieve the number of following. Please try again.");
|
|
2131
|
-
const jsonData = await response.json();
|
|
2132
|
-
return jsonData.length;
|
|
2134
|
+
const data = {
|
|
2135
|
+
following: jsonData.following,
|
|
2136
|
+
followers: jsonData.followers,
|
|
2137
|
+
publicRepos: jsonData.public_repos,
|
|
2138
|
+
avatarUrl: jsonData.avatar_url,
|
|
2139
|
+
age: jsonData.created_at
|
|
2140
|
+
};
|
|
2141
|
+
return data;
|
|
2133
2142
|
};
|
|
2134
2143
|
/**
|
|
2135
2144
|
* This function will check if the user is reputable enough to be able to use the app
|
|
@@ -2137,19 +2146,24 @@ const getNumberOfFollowingGitHub = async (user) => {
|
|
|
2137
2146
|
* @param minimumAmountOfFollowing <number> The minimum amount of following the user should have
|
|
2138
2147
|
* @param minimumAmountOfFollowers <number> The minimum amount of followers the user should have
|
|
2139
2148
|
* @param minimumAmountOfPublicRepos <number> The minimum amount of public repos the user should have
|
|
2140
|
-
* @returns <
|
|
2149
|
+
* @returns <any> Return the avatar URL of the user if the user is reputable, false otherwise
|
|
2141
2150
|
*/
|
|
2142
|
-
const githubReputation = async (userLogin, minimumAmountOfFollowing, minimumAmountOfFollowers, minimumAmountOfPublicRepos) => {
|
|
2151
|
+
const githubReputation = async (userLogin, minimumAmountOfFollowing, minimumAmountOfFollowers, minimumAmountOfPublicRepos, minimumAge) => {
|
|
2143
2152
|
if (!process.env.GITHUB_ACCESS_TOKEN)
|
|
2144
2153
|
throw new Error("The GitHub access token is missing. Please insert a valid token to be used for anti-sybil checks on user registation, and then try again.");
|
|
2145
|
-
const following = await
|
|
2146
|
-
const repos = await getNumberOfPublicReposGitHub(userLogin);
|
|
2147
|
-
const followers = await getNumberOfFollowersGitHub(userLogin);
|
|
2154
|
+
const { following, followers, publicRepos, avatarUrl, age } = await getGitHubStats(userLogin);
|
|
2148
2155
|
if (following < minimumAmountOfFollowing ||
|
|
2149
|
-
|
|
2150
|
-
followers < minimumAmountOfFollowers
|
|
2151
|
-
|
|
2152
|
-
|
|
2156
|
+
publicRepos < minimumAmountOfPublicRepos ||
|
|
2157
|
+
followers < minimumAmountOfFollowers ||
|
|
2158
|
+
new Date(age) > new Date(Date.now() - minimumAge))
|
|
2159
|
+
return {
|
|
2160
|
+
reputable: false,
|
|
2161
|
+
avatarUrl: ""
|
|
2162
|
+
};
|
|
2163
|
+
return {
|
|
2164
|
+
reputable: true,
|
|
2165
|
+
avatarUrl
|
|
2166
|
+
};
|
|
2153
2167
|
};
|
|
2154
2168
|
|
|
2155
2169
|
/**
|
|
@@ -2357,8 +2371,13 @@ const vmDependenciesAndCacheArtifactsCommand = (zKeyPath, potPath, snsTopic, reg
|
|
|
2357
2371
|
// eslint-disable-next-line no-template-curly-in-string
|
|
2358
2372
|
"touch ${MARKER_FILE}",
|
|
2359
2373
|
"sudo yum update -y",
|
|
2360
|
-
"curl -
|
|
2361
|
-
"
|
|
2374
|
+
"curl -O https://nodejs.org/dist/v16.13.0/node-v16.13.0-linux-x64.tar.xz",
|
|
2375
|
+
"tar -xf node-v16.13.0-linux-x64.tar.xz",
|
|
2376
|
+
"mv node-v16.13.0-linux-x64 nodejs",
|
|
2377
|
+
"sudo mv nodejs /opt/",
|
|
2378
|
+
"echo 'export NODEJS_HOME=/opt/nodejs' >> /etc/profile",
|
|
2379
|
+
"echo 'export PATH=$NODEJS_HOME/bin:$PATH' >> /etc/profile",
|
|
2380
|
+
"source /etc/profile",
|
|
2362
2381
|
"npm install -g snarkjs",
|
|
2363
2382
|
`aws s3 cp s3://${zKeyPath} /var/tmp/genesisZkey.zkey`,
|
|
2364
2383
|
`aws s3 cp s3://${potPath} /var/tmp/pot.ptau`,
|
|
@@ -2377,6 +2396,7 @@ const vmDependenciesAndCacheArtifactsCommand = (zKeyPath, potPath, snsTopic, reg
|
|
|
2377
2396
|
* @returns Array<string> - the list of commands for contribution verification.
|
|
2378
2397
|
*/
|
|
2379
2398
|
const vmContributionVerificationCommand = (bucketName, lastZkeyStoragePath, verificationTranscriptStoragePathAndFilename) => [
|
|
2399
|
+
`source /etc/profile`,
|
|
2380
2400
|
`aws s3 cp s3://${bucketName}/${lastZkeyStoragePath} /var/tmp/lastZKey.zkey > /var/tmp/log.txt`,
|
|
2381
2401
|
`snarkjs zkvi /var/tmp/genesisZkey.zkey /var/tmp/pot.ptau /var/tmp/lastZKey.zkey > /var/tmp/verification_transcript.log`,
|
|
2382
2402
|
`aws s3 cp /var/tmp/verification_transcript.log s3://${bucketName}/${verificationTranscriptStoragePathAndFilename} &>/dev/null`,
|
|
@@ -2403,7 +2423,7 @@ const computeDiskSizeForVM = (zKeySizeInBytes, pot) => Math.ceil(2 * convertByte
|
|
|
2403
2423
|
*/
|
|
2404
2424
|
const createEC2Instance = async (ec2, commands, instanceType, volumeSize, diskType) => {
|
|
2405
2425
|
// Get the AWS variables.
|
|
2406
|
-
const { amiId,
|
|
2426
|
+
const { amiId, instanceProfileArn } = getAWSVariables();
|
|
2407
2427
|
// Parametrize the VM EC2 instance.
|
|
2408
2428
|
const params = {
|
|
2409
2429
|
ImageId: amiId,
|
|
@@ -2412,7 +2432,7 @@ const createEC2Instance = async (ec2, commands, instanceType, volumeSize, diskTy
|
|
|
2412
2432
|
MinCount: 1,
|
|
2413
2433
|
// nb. to find this: iam -> roles -> role_name.
|
|
2414
2434
|
IamInstanceProfile: {
|
|
2415
|
-
Arn:
|
|
2435
|
+
Arn: instanceProfileArn
|
|
2416
2436
|
},
|
|
2417
2437
|
// nb. for running commands at the startup.
|
|
2418
2438
|
UserData: Buffer.from(commands.join("\n")).toString("base64"),
|