@aztec/blob-lib 0.0.1-commit.b655e406 → 0.0.1-commit.b6e433891
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/batched_blob.d.ts +31 -0
- package/dest/batched_blob.d.ts.map +1 -0
- package/dest/batched_blob.js +20 -0
- package/dest/blob.d.ts +14 -14
- package/dest/blob.d.ts.map +1 -1
- package/dest/blob.js +21 -20
- package/dest/blob_batching.d.ts +35 -72
- package/dest/blob_batching.d.ts.map +1 -1
- package/dest/blob_batching.js +75 -110
- package/dest/blob_utils.d.ts +22 -12
- package/dest/blob_utils.d.ts.map +1 -1
- package/dest/blob_utils.js +32 -23
- package/dest/circuit_types/blob_accumulator.d.ts +4 -2
- package/dest/circuit_types/blob_accumulator.d.ts.map +1 -1
- package/dest/circuit_types/blob_accumulator.js +5 -1
- package/dest/circuit_types/final_blob_accumulator.d.ts +3 -2
- package/dest/circuit_types/final_blob_accumulator.d.ts.map +1 -1
- package/dest/circuit_types/final_blob_accumulator.js +5 -2
- package/dest/circuit_types/final_blob_batching_challenges.d.ts +3 -2
- package/dest/circuit_types/final_blob_batching_challenges.d.ts.map +1 -1
- package/dest/circuit_types/final_blob_batching_challenges.js +2 -1
- package/dest/circuit_types/index.d.ts +1 -1
- package/dest/encoding/block_blob_data.d.ts +30 -0
- package/dest/encoding/block_blob_data.d.ts.map +1 -0
- package/dest/encoding/block_blob_data.js +75 -0
- package/dest/encoding/block_end_marker.d.ts +11 -0
- package/dest/encoding/block_end_marker.d.ts.map +1 -0
- package/dest/encoding/block_end_marker.js +41 -0
- package/dest/encoding/block_end_state_field.d.ts +12 -0
- package/dest/encoding/block_end_state_field.d.ts.map +1 -0
- package/dest/encoding/block_end_state_field.js +39 -0
- package/dest/encoding/checkpoint_blob_data.d.ts +15 -0
- package/dest/encoding/checkpoint_blob_data.d.ts.map +1 -0
- package/dest/encoding/checkpoint_blob_data.js +67 -0
- package/dest/encoding/checkpoint_end_marker.d.ts +8 -0
- package/dest/encoding/checkpoint_end_marker.d.ts.map +1 -0
- package/dest/encoding/checkpoint_end_marker.js +28 -0
- package/dest/encoding/fixtures.d.ts +41 -0
- package/dest/encoding/fixtures.d.ts.map +1 -0
- package/dest/encoding/fixtures.js +140 -0
- package/dest/encoding/index.d.ts +10 -0
- package/dest/encoding/index.d.ts.map +1 -0
- package/dest/encoding/index.js +9 -0
- package/dest/encoding/tx_blob_data.d.ts +19 -0
- package/dest/encoding/tx_blob_data.d.ts.map +1 -0
- package/dest/encoding/tx_blob_data.js +79 -0
- package/dest/encoding/tx_start_marker.d.ts +16 -0
- package/dest/encoding/tx_start_marker.d.ts.map +1 -0
- package/dest/{encoding.js → encoding/tx_start_marker.js} +14 -60
- package/dest/errors.d.ts +1 -1
- package/dest/errors.d.ts.map +1 -1
- package/dest/hash.d.ts +14 -6
- package/dest/hash.d.ts.map +1 -1
- package/dest/hash.js +25 -14
- package/dest/index.d.ts +4 -4
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +3 -3
- package/dest/interface.d.ts +1 -2
- package/dest/interface.d.ts.map +1 -1
- package/dest/kzg_context.d.ts +10 -4
- package/dest/kzg_context.d.ts.map +1 -1
- package/dest/kzg_context.js +29 -5
- package/dest/sponge_blob.d.ts +9 -13
- package/dest/sponge_blob.d.ts.map +1 -1
- package/dest/sponge_blob.js +21 -36
- package/dest/testing.d.ts +9 -17
- package/dest/testing.d.ts.map +1 -1
- package/dest/testing.js +35 -64
- package/dest/types.d.ts +2 -1
- package/dest/types.d.ts.map +1 -1
- package/dest/types.js +1 -0
- package/package.json +8 -7
- package/src/batched_blob.ts +26 -0
- package/src/blob.ts +21 -20
- package/src/blob_batching.ts +93 -128
- package/src/blob_utils.ts +38 -25
- package/src/circuit_types/blob_accumulator.ts +13 -1
- package/src/circuit_types/final_blob_accumulator.ts +2 -1
- package/src/circuit_types/final_blob_batching_challenges.ts +2 -1
- package/src/encoding/block_blob_data.ts +114 -0
- package/src/encoding/block_end_marker.ts +55 -0
- package/src/encoding/block_end_state_field.ts +59 -0
- package/src/encoding/checkpoint_blob_data.ts +102 -0
- package/src/encoding/checkpoint_end_marker.ts +40 -0
- package/src/encoding/fixtures.ts +210 -0
- package/src/encoding/index.ts +9 -0
- package/src/encoding/tx_blob_data.ts +116 -0
- package/src/{encoding.ts → encoding/tx_start_marker.ts} +20 -77
- package/src/hash.ts +26 -14
- package/src/index.ts +3 -3
- package/src/interface.ts +0 -1
- package/src/kzg_context.ts +39 -3
- package/src/sponge_blob.ts +23 -36
- package/src/testing.ts +48 -74
- package/src/types.ts +1 -0
- package/dest/deserialize.d.ts +0 -14
- package/dest/deserialize.d.ts.map +0 -1
- package/dest/deserialize.js +0 -33
- package/dest/encoding.d.ts +0 -26
- package/dest/encoding.d.ts.map +0 -1
- package/src/deserialize.ts +0 -38
package/dest/blob_batching.js
CHANGED
|
@@ -1,24 +1,52 @@
|
|
|
1
|
-
import {
|
|
2
|
-
import { poseidon2Hash
|
|
3
|
-
import {
|
|
4
|
-
import {
|
|
1
|
+
import { BLOBS_PER_CHECKPOINT, MAX_CHECKPOINTS_PER_EPOCH } from '@aztec/constants';
|
|
2
|
+
import { poseidon2Hash } from '@aztec/foundation/crypto/poseidon';
|
|
3
|
+
import { sha256ToField } from '@aztec/foundation/crypto/sha256';
|
|
4
|
+
import { BLS12Fr, BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
5
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
6
|
+
import { BatchedBlob } from './batched_blob.js';
|
|
7
|
+
import { getBlobsPerL1Block } from './blob_utils.js';
|
|
5
8
|
import { BlobAccumulator, FinalBlobAccumulator, FinalBlobBatchingChallenges } from './circuit_types/index.js';
|
|
6
|
-
import {
|
|
7
|
-
import {
|
|
9
|
+
import { computeBlobFieldsHash, hashNoirBigNumLimbs } from './hash.js';
|
|
10
|
+
import { getKzg } from './kzg_context.js';
|
|
8
11
|
/**
|
|
9
12
|
* A class to create, manage, and prove batched EVM blobs.
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
13
|
+
* See noir-projects/noir-protocol-circuits/crates/blob/src/abis/blob_accumulator.nr
|
|
14
|
+
*/ export class BatchedBlobAccumulator {
|
|
15
|
+
blobCommitmentsHashAcc;
|
|
16
|
+
zAcc;
|
|
17
|
+
yAcc;
|
|
18
|
+
cAcc;
|
|
19
|
+
qAcc;
|
|
20
|
+
gammaAcc;
|
|
21
|
+
gammaPow;
|
|
22
|
+
finalBlobChallenges;
|
|
23
|
+
constructor(/** Hash of Cs (to link to L1 blob hashes). */ blobCommitmentsHashAcc, /** Challenge point z_acc. Final value used such that p_i(z) = y_i. */ zAcc, /** Evaluation y_acc. Final value is is linear combination of all evaluations y_i = p_i(z) with gamma. */ yAcc, /** Commitment c_acc. Final value is linear combination of all commitments C_i = [p_i] with gamma. */ cAcc, /** KZG opening q_acc. Final value is linear combination of all blob kzg 'proofs' Q_i with gamma. */ qAcc, /**
|
|
24
|
+
* Challenge point gamma_acc for multi opening. Used with y, C, and kzg 'proof' Q above.
|
|
25
|
+
* TODO(#13608): We calculate this by hashing natively in the circuit (hence Fr representation), but it's actually used
|
|
26
|
+
* as a BLS12Fr field elt. Is this safe? Is there a skew?
|
|
27
|
+
*/ gammaAcc, /** Simply gamma^(i + 1) at blob i. Used for calculating the i'th element of the above linear comb.s */ gammaPow, /** Final challenge values used in evaluation. Optimistically input and checked in the final acc. */ finalBlobChallenges){
|
|
28
|
+
this.blobCommitmentsHashAcc = blobCommitmentsHashAcc;
|
|
29
|
+
this.zAcc = zAcc;
|
|
30
|
+
this.yAcc = yAcc;
|
|
31
|
+
this.cAcc = cAcc;
|
|
32
|
+
this.qAcc = qAcc;
|
|
33
|
+
this.gammaAcc = gammaAcc;
|
|
34
|
+
this.gammaPow = gammaPow;
|
|
35
|
+
this.finalBlobChallenges = finalBlobChallenges;
|
|
36
|
+
}
|
|
37
|
+
/**
|
|
38
|
+
* Create the empty accumulation state of the epoch.
|
|
39
|
+
* @returns An empty blob accumulator with challenges.
|
|
40
|
+
*/ static newWithChallenges(finalBlobChallenges) {
|
|
41
|
+
return new BatchedBlobAccumulator(Fr.ZERO, Fr.ZERO, BLS12Fr.ZERO, BLS12Point.ZERO, BLS12Point.ZERO, Fr.ZERO, BLS12Fr.ZERO, finalBlobChallenges);
|
|
42
|
+
}
|
|
43
|
+
/**
|
|
44
|
+
* Returns an empty BatchedBlobAccumulator with precomputed challenges from all blobs in the epoch.
|
|
45
|
+
* @dev MUST input all blobs to be broadcast. Does not work in multiple calls because z and gamma are calculated
|
|
46
|
+
* beforehand from ALL blobs.
|
|
47
|
+
*/ static async fromBlobFields(blobFieldsPerCheckpoint) {
|
|
48
|
+
const finalBlobChallenges = await this.precomputeBatchedBlobChallenges(blobFieldsPerCheckpoint);
|
|
49
|
+
return BatchedBlobAccumulator.newWithChallenges(finalBlobChallenges);
|
|
22
50
|
}
|
|
23
51
|
/**
|
|
24
52
|
* Get the final batched opening proof from multiple blobs.
|
|
@@ -26,25 +54,18 @@ import { kzg } from './kzg_context.js';
|
|
|
26
54
|
* beforehand from ALL blobs.
|
|
27
55
|
*
|
|
28
56
|
* @returns A batched blob.
|
|
29
|
-
*/ static async batch(
|
|
30
|
-
|
|
31
|
-
|
|
57
|
+
*/ static async batch(blobFieldsPerCheckpoint, verifyProof = false) {
|
|
58
|
+
const numCheckpoints = blobFieldsPerCheckpoint.length;
|
|
59
|
+
if (numCheckpoints > MAX_CHECKPOINTS_PER_EPOCH) {
|
|
60
|
+
throw new Error(`Too many checkpoints sent to batch(). The maximum is ${MAX_CHECKPOINTS_PER_EPOCH}. Got ${numCheckpoints}.`);
|
|
32
61
|
}
|
|
33
62
|
// Precalculate the values (z and gamma) and initialize the accumulator:
|
|
34
|
-
let acc = await this.
|
|
63
|
+
let acc = await this.fromBlobFields(blobFieldsPerCheckpoint);
|
|
35
64
|
// Now we can create a multi opening proof of all input blobs:
|
|
36
|
-
for (const
|
|
37
|
-
acc = await acc.
|
|
65
|
+
for (const blobFields of blobFieldsPerCheckpoint){
|
|
66
|
+
acc = await acc.accumulateFields(blobFields);
|
|
38
67
|
}
|
|
39
|
-
return await acc.finalize();
|
|
40
|
-
}
|
|
41
|
-
/**
|
|
42
|
-
* Returns an empty BatchedBlobAccumulator with precomputed challenges from all blobs in the epoch.
|
|
43
|
-
* @dev MUST input all blobs to be broadcast. Does not work in multiple calls because z and gamma are calculated
|
|
44
|
-
* beforehand from ALL blobs.
|
|
45
|
-
*/ static async newAccumulator(blobs) {
|
|
46
|
-
const finalBlobChallenges = await this.precomputeBatchedBlobChallenges(blobs);
|
|
47
|
-
return BatchedBlobAccumulator.newWithChallenges(finalBlobChallenges);
|
|
68
|
+
return await acc.finalize(verifyProof);
|
|
48
69
|
}
|
|
49
70
|
/**
|
|
50
71
|
* Gets the final challenges based on all blobs and their elements to perform a multi opening proof.
|
|
@@ -57,13 +78,15 @@ import { kzg } from './kzg_context.js';
|
|
|
57
78
|
*
|
|
58
79
|
* @param blobs - The blobs to precompute the challenges for. Each sub-array is the blobs for an L1 block.
|
|
59
80
|
* @returns Challenges z and gamma.
|
|
60
|
-
*/ static async precomputeBatchedBlobChallenges(
|
|
81
|
+
*/ static async precomputeBatchedBlobChallenges(blobFieldsPerCheckpoint) {
|
|
61
82
|
// Compute the final challenge z to evaluate the blobs.
|
|
62
83
|
let z;
|
|
63
|
-
|
|
84
|
+
const allBlobs = [];
|
|
85
|
+
for (const blobFields of blobFieldsPerCheckpoint){
|
|
64
86
|
// Compute the hash of all the fields in the block.
|
|
65
|
-
const blobFieldsHash = await
|
|
66
|
-
|
|
87
|
+
const blobFieldsHash = await computeBlobFieldsHash(blobFields);
|
|
88
|
+
const blobs = await getBlobsPerL1Block(blobFields);
|
|
89
|
+
for (const blob of blobs){
|
|
67
90
|
// Compute the challenge z for each blob and accumulate it.
|
|
68
91
|
const challengeZ = await blob.computeChallengeZ(blobFieldsHash);
|
|
69
92
|
if (!z) {
|
|
@@ -75,13 +98,13 @@ import { kzg } from './kzg_context.js';
|
|
|
75
98
|
]);
|
|
76
99
|
}
|
|
77
100
|
}
|
|
101
|
+
allBlobs.push(...blobs);
|
|
78
102
|
}
|
|
79
103
|
if (!z) {
|
|
80
104
|
throw new Error('No blobs to precompute challenges for.');
|
|
81
105
|
}
|
|
82
106
|
// Now we have a shared challenge for all blobs, evaluate them...
|
|
83
|
-
const
|
|
84
|
-
const proofObjects = allBlobs.map((b)=>b.evaluate(z));
|
|
107
|
+
const proofObjects = await Promise.all(allBlobs.map((b)=>b.evaluate(z)));
|
|
85
108
|
const evaluations = await Promise.all(proofObjects.map(({ y })=>hashNoirBigNumLimbs(y)));
|
|
86
109
|
// ...and find the challenge for the linear combination of blobs.
|
|
87
110
|
let gamma = evaluations[0];
|
|
@@ -98,74 +121,12 @@ import { kzg } from './kzg_context.js';
|
|
|
98
121
|
]);
|
|
99
122
|
return new FinalBlobBatchingChallenges(z, BLS12Fr.fromBN254Fr(gamma));
|
|
100
123
|
}
|
|
101
|
-
verify() {
|
|
102
|
-
return kzg.verifyKzgProof(this.commitment.compress(), this.z.toBuffer(), this.y.toBuffer(), this.q.compress());
|
|
103
|
-
}
|
|
104
|
-
// Returns ethereum's versioned blob hash, following kzg_to_versioned_hash: https://eips.ethereum.org/EIPS/eip-4844#helpers
|
|
105
|
-
getEthVersionedBlobHash() {
|
|
106
|
-
return computeEthVersionedBlobHash(this.commitment.compress());
|
|
107
|
-
}
|
|
108
|
-
/**
|
|
109
|
-
* Returns a proof of opening of the blobs to verify on L1 using the point evaluation precompile:
|
|
110
|
-
*
|
|
111
|
-
* input[:32] - versioned_hash
|
|
112
|
-
* input[32:64] - z
|
|
113
|
-
* input[64:96] - y
|
|
114
|
-
* input[96:144] - commitment C
|
|
115
|
-
* input[144:192] - commitment Q (a 'proof' committing to the quotient polynomial q(X))
|
|
116
|
-
*
|
|
117
|
-
* See https://eips.ethereum.org/EIPS/eip-4844#point-evaluation-precompile
|
|
118
|
-
*/ getEthBlobEvaluationInputs() {
|
|
119
|
-
const buf = Buffer.concat([
|
|
120
|
-
this.getEthVersionedBlobHash(),
|
|
121
|
-
this.z.toBuffer(),
|
|
122
|
-
this.y.toBuffer(),
|
|
123
|
-
this.commitment.compress(),
|
|
124
|
-
this.q.compress()
|
|
125
|
-
]);
|
|
126
|
-
return `0x${buf.toString('hex')}`;
|
|
127
|
-
}
|
|
128
|
-
toFinalBlobAccumulator() {
|
|
129
|
-
return new FinalBlobAccumulator(this.blobCommitmentsHash, this.z, this.y, this.commitment);
|
|
130
|
-
}
|
|
131
|
-
}
|
|
132
|
-
/**
|
|
133
|
-
* See noir-projects/noir-protocol-circuits/crates/blob/src/abis/blob_accumulator.nr
|
|
134
|
-
*/ export class BatchedBlobAccumulator {
|
|
135
|
-
blobCommitmentsHashAcc;
|
|
136
|
-
zAcc;
|
|
137
|
-
yAcc;
|
|
138
|
-
cAcc;
|
|
139
|
-
qAcc;
|
|
140
|
-
gammaAcc;
|
|
141
|
-
gammaPow;
|
|
142
|
-
finalBlobChallenges;
|
|
143
|
-
constructor(/** Hash of Cs (to link to L1 blob hashes). */ blobCommitmentsHashAcc, /** Challenge point z_acc. Final value used such that p_i(z) = y_i. */ zAcc, /** Evaluation y_acc. Final value is is linear combination of all evaluations y_i = p_i(z) with gamma. */ yAcc, /** Commitment c_acc. Final value is linear combination of all commitments C_i = [p_i] with gamma. */ cAcc, /** KZG opening q_acc. Final value is linear combination of all blob kzg 'proofs' Q_i with gamma. */ qAcc, /**
|
|
144
|
-
* Challenge point gamma_acc for multi opening. Used with y, C, and kzg 'proof' Q above.
|
|
145
|
-
* TODO(#13608): We calculate this by hashing natively in the circuit (hence Fr representation), but it's actually used
|
|
146
|
-
* as a BLS12Fr field elt. Is this safe? Is there a skew?
|
|
147
|
-
*/ gammaAcc, /** Simply gamma^(i + 1) at blob i. Used for calculating the i'th element of the above linear comb.s */ gammaPow, /** Final challenge values used in evaluation. Optimistically input and checked in the final acc. */ finalBlobChallenges){
|
|
148
|
-
this.blobCommitmentsHashAcc = blobCommitmentsHashAcc;
|
|
149
|
-
this.zAcc = zAcc;
|
|
150
|
-
this.yAcc = yAcc;
|
|
151
|
-
this.cAcc = cAcc;
|
|
152
|
-
this.qAcc = qAcc;
|
|
153
|
-
this.gammaAcc = gammaAcc;
|
|
154
|
-
this.gammaPow = gammaPow;
|
|
155
|
-
this.finalBlobChallenges = finalBlobChallenges;
|
|
156
|
-
}
|
|
157
|
-
/**
|
|
158
|
-
* Create the empty accumulation state of the epoch.
|
|
159
|
-
* @returns An empty blob accumulator with challenges.
|
|
160
|
-
*/ static newWithChallenges(finalBlobChallenges) {
|
|
161
|
-
return new BatchedBlobAccumulator(Fr.ZERO, Fr.ZERO, BLS12Fr.ZERO, BLS12Point.ZERO, BLS12Point.ZERO, Fr.ZERO, BLS12Fr.ZERO, finalBlobChallenges);
|
|
162
|
-
}
|
|
163
124
|
/**
|
|
164
125
|
* Given blob i, accumulate all state.
|
|
165
126
|
* We assume the input blob has not been evaluated at z.
|
|
166
127
|
* @returns An updated blob accumulator.
|
|
167
|
-
*/ async
|
|
168
|
-
const { proof, y: thisY } = blob.evaluate(this.finalBlobChallenges.z);
|
|
128
|
+
*/ async accumulateBlob(blob, blobFieldsHash) {
|
|
129
|
+
const { proof, y: thisY } = await blob.evaluate(this.finalBlobChallenges.z);
|
|
169
130
|
const thisC = BLS12Point.decompress(blob.commitment);
|
|
170
131
|
const thisQ = BLS12Point.decompress(proof);
|
|
171
132
|
const blobChallengeZ = await blob.computeChallengeZ(blobFieldsHash);
|
|
@@ -198,18 +159,19 @@ import { kzg } from './kzg_context.js';
|
|
|
198
159
|
/**
|
|
199
160
|
* Given blobs, accumulate all state.
|
|
200
161
|
* We assume the input blobs have not been evaluated at z.
|
|
201
|
-
* @param
|
|
162
|
+
* @param blobFields - The blob fields of a checkpoint to accumulate.
|
|
202
163
|
* @returns An updated blob accumulator.
|
|
203
|
-
*/ async
|
|
204
|
-
|
|
205
|
-
|
|
164
|
+
*/ async accumulateFields(blobFields) {
|
|
165
|
+
const blobs = await getBlobsPerL1Block(blobFields);
|
|
166
|
+
if (blobs.length > BLOBS_PER_CHECKPOINT) {
|
|
167
|
+
throw new Error(`Too many blobs to accumulate. The maximum is ${BLOBS_PER_CHECKPOINT} per checkpoint. Got ${blobs.length}.`);
|
|
206
168
|
}
|
|
207
169
|
// Compute the hash of all the fields in the block.
|
|
208
|
-
const blobFieldsHash = await
|
|
170
|
+
const blobFieldsHash = await computeBlobFieldsHash(blobFields);
|
|
209
171
|
// Initialize the acc to iterate over:
|
|
210
172
|
let acc = this.clone();
|
|
211
173
|
for (const blob of blobs){
|
|
212
|
-
acc = await acc.
|
|
174
|
+
acc = await acc.accumulateBlob(blob, blobFieldsHash);
|
|
213
175
|
}
|
|
214
176
|
return acc;
|
|
215
177
|
}
|
|
@@ -240,11 +202,14 @@ import { kzg } from './kzg_context.js';
|
|
|
240
202
|
throw new Error(`Blob batching mismatch: accumulated gamma ${calculatedGamma} does not equal injected gamma ${this.finalBlobChallenges.gamma.toBN254Fr()}`);
|
|
241
203
|
}
|
|
242
204
|
const batchedBlob = new BatchedBlob(this.blobCommitmentsHashAcc, this.zAcc, this.yAcc, this.cAcc, this.qAcc);
|
|
243
|
-
if (verifyProof && !
|
|
205
|
+
if (verifyProof && !this.verify()) {
|
|
244
206
|
throw new Error(`KZG proof did not verify.`);
|
|
245
207
|
}
|
|
246
208
|
return batchedBlob;
|
|
247
209
|
}
|
|
210
|
+
verify() {
|
|
211
|
+
return getKzg().verifyKzgProof(this.cAcc.compress(), this.zAcc.toBuffer(), this.yAcc.toBuffer(), this.qAcc.compress());
|
|
212
|
+
}
|
|
248
213
|
isEmptyState() {
|
|
249
214
|
return this.blobCommitmentsHashAcc.isZero() && this.zAcc.isZero() && this.yAcc.isZero() && this.cAcc.isZero() && this.qAcc.isZero() && this.gammaAcc.isZero() && this.gammaPow.isZero();
|
|
250
215
|
}
|
package/dest/blob_utils.d.ts
CHANGED
|
@@ -1,5 +1,8 @@
|
|
|
1
|
-
import { BLS12Point
|
|
1
|
+
import { BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
2
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
3
|
+
import type { BatchedBlob } from './batched_blob.js';
|
|
2
4
|
import { Blob } from './blob.js';
|
|
5
|
+
import { type CheckpointBlobData } from './encoding/index.js';
|
|
3
6
|
/**
|
|
4
7
|
* @param blobs - The blobs to emit.
|
|
5
8
|
* @returns The blobs' compressed commitments in hex prefixed by the number of blobs. 1 byte for the prefix, 48 bytes
|
|
@@ -13,18 +16,25 @@ export declare function getPrefixedEthBlobCommitments(blobs: Blob[]): `0x${strin
|
|
|
13
16
|
*
|
|
14
17
|
* @throws If the number of fields does not match what's indicated by the checkpoint prefix.
|
|
15
18
|
*/
|
|
16
|
-
export declare function getBlobsPerL1Block(fields: Fr[]): Blob[]
|
|
19
|
+
export declare function getBlobsPerL1Block(fields: Fr[]): Promise<Blob[]>;
|
|
17
20
|
/**
|
|
18
|
-
* Get the
|
|
19
|
-
*
|
|
20
|
-
*
|
|
21
|
-
* @param blobs - The blobs to read fields from. Should be all the blobs in the L1 block proposing the checkpoint.
|
|
22
|
-
* @param checkEncoding - Whether to check if the entire encoded blob fields are valid. If false, it will still check
|
|
23
|
-
* the checkpoint prefix and throw if there's not enough fields.
|
|
24
|
-
* @returns The fields added throughout the checkpoint.
|
|
21
|
+
* Get the encoded data from all blobs in the checkpoint.
|
|
22
|
+
* @param blobs - The blobs to read data from. Should be all the blobs for the L1 block proposing the checkpoint.
|
|
23
|
+
* @returns The encoded data of the checkpoint.
|
|
25
24
|
*/
|
|
26
|
-
export declare function
|
|
27
|
-
export declare function computeBlobFieldsHashFromBlobs(blobs: Blob[]): Promise<Fr>;
|
|
25
|
+
export declare function decodeCheckpointBlobDataFromBlobs(blobs: Blob[]): CheckpointBlobData;
|
|
28
26
|
export declare function computeBlobsHashFromBlobs(blobs: Blob[]): Fr;
|
|
29
27
|
export declare function getBlobCommitmentsFromBlobs(blobs: Blob[]): BLS12Point[];
|
|
30
|
-
|
|
28
|
+
/**
|
|
29
|
+
* Returns a proof of opening of the blobs to verify on L1 using the point evaluation precompile:
|
|
30
|
+
*
|
|
31
|
+
* input[:32] - versioned_hash
|
|
32
|
+
* input[32:64] - z
|
|
33
|
+
* input[64:96] - y
|
|
34
|
+
* input[96:144] - commitment C
|
|
35
|
+
* input[144:192] - commitment Q (a 'proof' committing to the quotient polynomial q(X))
|
|
36
|
+
*
|
|
37
|
+
* See https://eips.ethereum.org/EIPS/eip-4844#point-evaluation-precompile
|
|
38
|
+
*/
|
|
39
|
+
export declare function getEthBlobEvaluationInputs(batchedBlob: BatchedBlob): `0x${string}`;
|
|
40
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiYmxvYl91dGlscy5kLnRzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vc3JjL2Jsb2JfdXRpbHMudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQ0EsT0FBTyxFQUFFLFVBQVUsRUFBRSxNQUFNLGdDQUFnQyxDQUFDO0FBQzVELE9BQU8sRUFBRSxFQUFFLEVBQUUsTUFBTSxnQ0FBZ0MsQ0FBQztBQUVwRCxPQUFPLEtBQUssRUFBRSxXQUFXLEVBQUUsTUFBTSxtQkFBbUIsQ0FBQztBQUNyRCxPQUFPLEVBQUUsSUFBSSxFQUFFLE1BQU0sV0FBVyxDQUFDO0FBQ2pDLE9BQU8sRUFBRSxLQUFLLGtCQUFrQixFQUFzQyxNQUFNLHFCQUFxQixDQUFDO0FBR2xHOzs7OztHQUtHO0FBQ0gsd0JBQWdCLDZCQUE2QixDQUFDLEtBQUssRUFBRSxJQUFJLEVBQUUsR0FBRyxLQUFLLE1BQU0sRUFBRSxDQVMxRTtBQUVEOzs7OztHQUtHO0FBQ0gsd0JBQXNCLGtCQUFrQixDQUFDLE1BQU0sRUFBRSxFQUFFLEVBQUUsR0FBRyxPQUFPLENBQUMsSUFBSSxFQUFFLENBQUMsQ0FXdEU7QUFFRDs7OztHQUlHO0FBQ0gsd0JBQWdCLGlDQUFpQyxDQUFDLEtBQUssRUFBRSxJQUFJLEVBQUUsR0FBRyxrQkFBa0IsQ0FHbkY7QUFFRCx3QkFBZ0IseUJBQXlCLENBQUMsS0FBSyxFQUFFLElBQUksRUFBRSxHQUFHLEVBQUUsQ0FFM0Q7QUFFRCx3QkFBZ0IsMkJBQTJCLENBQUMsS0FBSyxFQUFFLElBQUksRUFBRSxHQUFHLFVBQVUsRUFBRSxDQUV2RTtBQUVEOzs7Ozs7Ozs7O0dBVUc7QUFDSCx3QkFBZ0IsMEJBQTBCLENBQUMsV0FBVyxFQUFFLFdBQVcsR0FBRyxLQUFLLE1BQU0sRUFBRSxDQVNsRiJ9
|
package/dest/blob_utils.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"blob_utils.d.ts","sourceRoot":"","sources":["../src/blob_utils.ts"],"names":[],"mappings":"AACA,OAAO,EAAE,UAAU,EAAE,EAAE,EAAE,MAAM,
|
|
1
|
+
{"version":3,"file":"blob_utils.d.ts","sourceRoot":"","sources":["../src/blob_utils.ts"],"names":[],"mappings":"AACA,OAAO,EAAE,UAAU,EAAE,MAAM,gCAAgC,CAAC;AAC5D,OAAO,EAAE,EAAE,EAAE,MAAM,gCAAgC,CAAC;AAEpD,OAAO,KAAK,EAAE,WAAW,EAAE,MAAM,mBAAmB,CAAC;AACrD,OAAO,EAAE,IAAI,EAAE,MAAM,WAAW,CAAC;AACjC,OAAO,EAAE,KAAK,kBAAkB,EAAsC,MAAM,qBAAqB,CAAC;AAGlG;;;;;GAKG;AACH,wBAAgB,6BAA6B,CAAC,KAAK,EAAE,IAAI,EAAE,GAAG,KAAK,MAAM,EAAE,CAS1E;AAED;;;;;GAKG;AACH,wBAAsB,kBAAkB,CAAC,MAAM,EAAE,EAAE,EAAE,GAAG,OAAO,CAAC,IAAI,EAAE,CAAC,CAWtE;AAED;;;;GAIG;AACH,wBAAgB,iCAAiC,CAAC,KAAK,EAAE,IAAI,EAAE,GAAG,kBAAkB,CAGnF;AAED,wBAAgB,yBAAyB,CAAC,KAAK,EAAE,IAAI,EAAE,GAAG,EAAE,CAE3D;AAED,wBAAgB,2BAA2B,CAAC,KAAK,EAAE,IAAI,EAAE,GAAG,UAAU,EAAE,CAEvE;AAED;;;;;;;;;;GAUG;AACH,wBAAgB,0BAA0B,CAAC,WAAW,EAAE,WAAW,GAAG,KAAK,MAAM,EAAE,CASlF"}
|
package/dest/blob_utils.js
CHANGED
|
@@ -1,8 +1,8 @@
|
|
|
1
1
|
import { FIELDS_PER_BLOB } from '@aztec/constants';
|
|
2
|
-
import { BLS12Point } from '@aztec/foundation/
|
|
2
|
+
import { BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
3
3
|
import { Blob } from './blob.js';
|
|
4
|
-
import {
|
|
5
|
-
import {
|
|
4
|
+
import { decodeCheckpointBlobDataFromBuffer } from './encoding/index.js';
|
|
5
|
+
import { computeBlobsHash, computeEthVersionedBlobHash } from './hash.js';
|
|
6
6
|
/**
|
|
7
7
|
* @param blobs - The blobs to emit.
|
|
8
8
|
* @returns The blobs' compressed commitments in hex prefixed by the number of blobs. 1 byte for the prefix, 48 bytes
|
|
@@ -24,33 +24,22 @@ import { computeBlobFieldsHash, computeBlobsHash } from './hash.js';
|
|
|
24
24
|
* @returns As many blobs as required to broadcast the given fields to an L1 block.
|
|
25
25
|
*
|
|
26
26
|
* @throws If the number of fields does not match what's indicated by the checkpoint prefix.
|
|
27
|
-
*/ export function getBlobsPerL1Block(fields) {
|
|
27
|
+
*/ export async function getBlobsPerL1Block(fields) {
|
|
28
28
|
if (!fields.length) {
|
|
29
29
|
throw new Error('Cannot create blobs from empty fields.');
|
|
30
30
|
}
|
|
31
31
|
const numBlobs = Math.ceil(fields.length / FIELDS_PER_BLOB);
|
|
32
|
-
return Array.from({
|
|
32
|
+
return await Promise.all(Array.from({
|
|
33
33
|
length: numBlobs
|
|
34
|
-
}, (_, i)=>Blob.fromFields(fields.slice(i * FIELDS_PER_BLOB, (i + 1) * FIELDS_PER_BLOB)));
|
|
34
|
+
}, (_, i)=>Blob.fromFields(fields.slice(i * FIELDS_PER_BLOB, (i + 1) * FIELDS_PER_BLOB))));
|
|
35
35
|
}
|
|
36
36
|
/**
|
|
37
|
-
* Get the
|
|
38
|
-
*
|
|
39
|
-
*
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
* @returns The fields added throughout the checkpoint.
|
|
44
|
-
*/ export function getBlobFieldsInCheckpoint(blobs, checkEncoding = false) {
|
|
45
|
-
return deserializeEncodedBlobToFields(Buffer.concat(blobs.map((b)=>b.data)), checkEncoding);
|
|
46
|
-
}
|
|
47
|
-
export async function computeBlobFieldsHashFromBlobs(blobs) {
|
|
48
|
-
const fields = blobs.map((b)=>b.toFields()).flat();
|
|
49
|
-
const numBlobFields = fields[0].toNumber();
|
|
50
|
-
if (numBlobFields > fields.length) {
|
|
51
|
-
throw new Error(`The prefix indicates ${numBlobFields} fields. Got ${fields.length}.`);
|
|
52
|
-
}
|
|
53
|
-
return await computeBlobFieldsHash(fields.slice(0, numBlobFields));
|
|
37
|
+
* Get the encoded data from all blobs in the checkpoint.
|
|
38
|
+
* @param blobs - The blobs to read data from. Should be all the blobs for the L1 block proposing the checkpoint.
|
|
39
|
+
* @returns The encoded data of the checkpoint.
|
|
40
|
+
*/ export function decodeCheckpointBlobDataFromBlobs(blobs) {
|
|
41
|
+
const buf = Buffer.concat(blobs.map((b)=>b.data));
|
|
42
|
+
return decodeCheckpointBlobDataFromBuffer(buf);
|
|
54
43
|
}
|
|
55
44
|
export function computeBlobsHashFromBlobs(blobs) {
|
|
56
45
|
return computeBlobsHash(blobs.map((b)=>b.getEthVersionedBlobHash()));
|
|
@@ -58,3 +47,23 @@ export function computeBlobsHashFromBlobs(blobs) {
|
|
|
58
47
|
export function getBlobCommitmentsFromBlobs(blobs) {
|
|
59
48
|
return blobs.map((b)=>BLS12Point.decompress(b.commitment));
|
|
60
49
|
}
|
|
50
|
+
/**
|
|
51
|
+
* Returns a proof of opening of the blobs to verify on L1 using the point evaluation precompile:
|
|
52
|
+
*
|
|
53
|
+
* input[:32] - versioned_hash
|
|
54
|
+
* input[32:64] - z
|
|
55
|
+
* input[64:96] - y
|
|
56
|
+
* input[96:144] - commitment C
|
|
57
|
+
* input[144:192] - commitment Q (a 'proof' committing to the quotient polynomial q(X))
|
|
58
|
+
*
|
|
59
|
+
* See https://eips.ethereum.org/EIPS/eip-4844#point-evaluation-precompile
|
|
60
|
+
*/ export function getEthBlobEvaluationInputs(batchedBlob) {
|
|
61
|
+
const buf = Buffer.concat([
|
|
62
|
+
computeEthVersionedBlobHash(batchedBlob.commitment.compress()),
|
|
63
|
+
batchedBlob.z.toBuffer(),
|
|
64
|
+
batchedBlob.y.toBuffer(),
|
|
65
|
+
batchedBlob.commitment.compress(),
|
|
66
|
+
batchedBlob.q.compress()
|
|
67
|
+
]);
|
|
68
|
+
return `0x${buf.toString('hex')}`;
|
|
69
|
+
}
|
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import { BLS12Fr, BLS12Point
|
|
1
|
+
import { BLS12Fr, BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
2
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
2
3
|
import { BufferReader, FieldReader } from '@aztec/foundation/serialize';
|
|
3
4
|
/**
|
|
4
5
|
* See `noir-projects/noir-protocol-circuits/crates/blob/src/abis/blob_accumulator.nr` for documentation.
|
|
@@ -17,5 +18,6 @@ export declare class BlobAccumulator {
|
|
|
17
18
|
toBuffer(): Buffer<ArrayBufferLike>;
|
|
18
19
|
toFields(): Fr[];
|
|
19
20
|
static fromFields(fields: Fr[] | FieldReader): BlobAccumulator;
|
|
21
|
+
static random(): BlobAccumulator;
|
|
20
22
|
}
|
|
21
|
-
//# sourceMappingURL=
|
|
23
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiYmxvYl9hY2N1bXVsYXRvci5kLnRzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vLi4vc3JjL2NpcmN1aXRfdHlwZXMvYmxvYl9hY2N1bXVsYXRvci50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFDQSxPQUFPLEVBQVcsT0FBTyxFQUFFLFVBQVUsRUFBRSxNQUFNLGdDQUFnQyxDQUFDO0FBQzlFLE9BQU8sRUFBRSxFQUFFLEVBQUUsTUFBTSxnQ0FBZ0MsQ0FBQztBQUNwRCxPQUFPLEVBQUUsWUFBWSxFQUFFLFdBQVcsRUFBcUIsTUFBTSw2QkFBNkIsQ0FBQztBQUUzRjs7R0FFRztBQUNILHFCQUFhLGVBQWU7SUFFakIsc0JBQXNCLEVBQUUsRUFBRTtJQUMxQixJQUFJLEVBQUUsRUFBRTtJQUNSLElBQUksRUFBRSxPQUFPO0lBQ2IsSUFBSSxFQUFFLFVBQVU7SUFDaEIsUUFBUSxFQUFFLEVBQUU7SUFDWixXQUFXLEVBQUUsT0FBTztJQU43QixZQUNTLHNCQUFzQixFQUFFLEVBQUUsRUFDMUIsSUFBSSxFQUFFLEVBQUUsRUFDUixJQUFJLEVBQUUsT0FBTyxFQUNiLElBQUksRUFBRSxVQUFVLEVBQ2hCLFFBQVEsRUFBRSxFQUFFLEVBQ1osV0FBVyxFQUFFLE9BQU8sRUFDekI7SUFFSixNQUFNLENBQUMsS0FBSyxJQUFJLGVBQWUsQ0FFOUI7SUFFRCxNQUFNLENBQUMsS0FBSyxFQUFFLGVBQWUsV0FTNUI7SUFFRCxNQUFNLENBQUMsVUFBVSxDQUFDLE1BQU0sRUFBRSxNQUFNLEdBQUcsWUFBWSxHQUFHLGVBQWUsQ0FVaEU7SUFFRCxRQUFRLDRCQVNQO0lBRUQsUUFBUSxTQVdQO0lBRUQsTUFBTSxDQUFDLFVBQVUsQ0FBQyxNQUFNLEVBQUUsRUFBRSxFQUFFLEdBQUcsV0FBVyxHQUFHLGVBQWUsQ0FjN0Q7SUFFRCxNQUFNLENBQUMsTUFBTSxvQkFTWjtDQUNGIn0=
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"blob_accumulator.d.ts","sourceRoot":"","sources":["../../src/circuit_types/blob_accumulator.ts"],"names":[],"mappings":"AACA,OAAO,EAAW,OAAO,EAAE,UAAU,EAAE,EAAE,EAAE,MAAM,
|
|
1
|
+
{"version":3,"file":"blob_accumulator.d.ts","sourceRoot":"","sources":["../../src/circuit_types/blob_accumulator.ts"],"names":[],"mappings":"AACA,OAAO,EAAW,OAAO,EAAE,UAAU,EAAE,MAAM,gCAAgC,CAAC;AAC9E,OAAO,EAAE,EAAE,EAAE,MAAM,gCAAgC,CAAC;AACpD,OAAO,EAAE,YAAY,EAAE,WAAW,EAAqB,MAAM,6BAA6B,CAAC;AAE3F;;GAEG;AACH,qBAAa,eAAe;IAEjB,sBAAsB,EAAE,EAAE;IAC1B,IAAI,EAAE,EAAE;IACR,IAAI,EAAE,OAAO;IACb,IAAI,EAAE,UAAU;IAChB,QAAQ,EAAE,EAAE;IACZ,WAAW,EAAE,OAAO;IAN7B,YACS,sBAAsB,EAAE,EAAE,EAC1B,IAAI,EAAE,EAAE,EACR,IAAI,EAAE,OAAO,EACb,IAAI,EAAE,UAAU,EAChB,QAAQ,EAAE,EAAE,EACZ,WAAW,EAAE,OAAO,EACzB;IAEJ,MAAM,CAAC,KAAK,IAAI,eAAe,CAE9B;IAED,MAAM,CAAC,KAAK,EAAE,eAAe,WAS5B;IAED,MAAM,CAAC,UAAU,CAAC,MAAM,EAAE,MAAM,GAAG,YAAY,GAAG,eAAe,CAUhE;IAED,QAAQ,4BASP;IAED,QAAQ,SAWP;IAED,MAAM,CAAC,UAAU,CAAC,MAAM,EAAE,EAAE,EAAE,GAAG,WAAW,GAAG,eAAe,CAc7D;IAED,MAAM,CAAC,MAAM,oBASZ;CACF"}
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
import { BLS12_FQ_LIMBS, BLS12_FR_LIMBS } from '@aztec/constants';
|
|
2
|
-
import { BLS12Fq, BLS12Fr, BLS12Point
|
|
2
|
+
import { BLS12Fq, BLS12Fr, BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
3
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
3
4
|
import { BufferReader, FieldReader, serializeToBuffer } from '@aztec/foundation/serialize';
|
|
4
5
|
/**
|
|
5
6
|
* See `noir-projects/noir-protocol-circuits/crates/blob/src/abis/blob_accumulator.nr` for documentation.
|
|
@@ -55,4 +56,7 @@ import { BufferReader, FieldReader, serializeToBuffer } from '@aztec/foundation/
|
|
|
55
56
|
limbs: reader.readFieldArray(BLS12_FR_LIMBS).map((f)=>f.toString())
|
|
56
57
|
}));
|
|
57
58
|
}
|
|
59
|
+
static random() {
|
|
60
|
+
return new BlobAccumulator(Fr.random(), Fr.random(), BLS12Fr.random(), BLS12Point.random(), Fr.random(), BLS12Fr.random());
|
|
61
|
+
}
|
|
58
62
|
}
|
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import { BLS12Fr, BLS12Point
|
|
1
|
+
import { BLS12Fr, BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
2
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
2
3
|
import { BufferReader } from '@aztec/foundation/serialize';
|
|
3
4
|
import { inspect } from 'util';
|
|
4
5
|
/**
|
|
@@ -19,4 +20,4 @@ export declare class FinalBlobAccumulator {
|
|
|
19
20
|
static random(): FinalBlobAccumulator;
|
|
20
21
|
[inspect.custom](): string;
|
|
21
22
|
}
|
|
22
|
-
//# sourceMappingURL=
|
|
23
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiZmluYWxfYmxvYl9hY2N1bXVsYXRvci5kLnRzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vLi4vc3JjL2NpcmN1aXRfdHlwZXMvZmluYWxfYmxvYl9hY2N1bXVsYXRvci50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxPQUFPLEVBQUUsT0FBTyxFQUFFLFVBQVUsRUFBRSxNQUFNLGdDQUFnQyxDQUFDO0FBQ3JFLE9BQU8sRUFBRSxFQUFFLEVBQUUsTUFBTSxnQ0FBZ0MsQ0FBQztBQUNwRCxPQUFPLEVBQUUsWUFBWSxFQUFxQixNQUFNLDZCQUE2QixDQUFDO0FBRTlFLE9BQU8sRUFBRSxPQUFPLEVBQUUsTUFBTSxNQUFNLENBQUM7QUFFL0I7O0dBRUc7QUFDSCxxQkFBYSxvQkFBb0I7SUFFdEIsbUJBQW1CLEVBQUUsRUFBRTtJQUN2QixDQUFDLEVBQUUsRUFBRTtJQUNMLENBQUMsRUFBRSxPQUFPO0lBQ1YsQ0FBQyxFQUFFLFVBQVU7SUFKdEIsWUFDUyxtQkFBbUIsRUFBRSxFQUFFLEVBQ3ZCLENBQUMsRUFBRSxFQUFFLEVBQ0wsQ0FBQyxFQUFFLE9BQU8sRUFDVixDQUFDLEVBQUUsVUFBVSxFQUNsQjtJQUVKLE1BQU0sQ0FBQyxLQUFLLElBQUksb0JBQW9CLENBRW5DO0lBRUQsTUFBTSxDQUFDLFVBQVUsQ0FBQyxNQUFNLEVBQUUsTUFBTSxHQUFHLFlBQVksR0FBRyxvQkFBb0IsQ0FRckU7SUFFRCxRQUFRLDRCQUVQO0lBRUQsUUFBUSxTQU9QO0lBR0QsUUFBUSxXQU1QO0lBRUQsTUFBTSxDQUFDLEtBQUssRUFBRSxvQkFBb0IsV0FPakM7SUFHRCxNQUFNLENBQUMsTUFBTSx5QkFFWjtJQUVELENBQUMsT0FBTyxDQUFDLE1BQU0sQ0FBQyxXQU9mO0NBQ0YifQ==
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"final_blob_accumulator.d.ts","sourceRoot":"","sources":["../../src/circuit_types/final_blob_accumulator.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,UAAU,EAAE,EAAE,EAAE,MAAM,
|
|
1
|
+
{"version":3,"file":"final_blob_accumulator.d.ts","sourceRoot":"","sources":["../../src/circuit_types/final_blob_accumulator.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,UAAU,EAAE,MAAM,gCAAgC,CAAC;AACrE,OAAO,EAAE,EAAE,EAAE,MAAM,gCAAgC,CAAC;AACpD,OAAO,EAAE,YAAY,EAAqB,MAAM,6BAA6B,CAAC;AAE9E,OAAO,EAAE,OAAO,EAAE,MAAM,MAAM,CAAC;AAE/B;;GAEG;AACH,qBAAa,oBAAoB;IAEtB,mBAAmB,EAAE,EAAE;IACvB,CAAC,EAAE,EAAE;IACL,CAAC,EAAE,OAAO;IACV,CAAC,EAAE,UAAU;IAJtB,YACS,mBAAmB,EAAE,EAAE,EACvB,CAAC,EAAE,EAAE,EACL,CAAC,EAAE,OAAO,EACV,CAAC,EAAE,UAAU,EAClB;IAEJ,MAAM,CAAC,KAAK,IAAI,oBAAoB,CAEnC;IAED,MAAM,CAAC,UAAU,CAAC,MAAM,EAAE,MAAM,GAAG,YAAY,GAAG,oBAAoB,CAQrE;IAED,QAAQ,4BAEP;IAED,QAAQ,SAOP;IAGD,QAAQ,WAMP;IAED,MAAM,CAAC,KAAK,EAAE,oBAAoB,WAOjC;IAGD,MAAM,CAAC,MAAM,yBAEZ;IAED,CAAC,OAAO,CAAC,MAAM,CAAC,WAOf;CACF"}
|
|
@@ -1,6 +1,9 @@
|
|
|
1
|
-
|
|
1
|
+
var _computedKey;
|
|
2
|
+
import { BLS12Fr, BLS12Point } from '@aztec/foundation/curves/bls12';
|
|
3
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
2
4
|
import { BufferReader, serializeToBuffer } from '@aztec/foundation/serialize';
|
|
3
5
|
import { inspect } from 'util';
|
|
6
|
+
_computedKey = inspect.custom;
|
|
4
7
|
/**
|
|
5
8
|
* See `noir-projects/noir-protocol-circuits/crates/blob/src/abis/final_blob_accumulator.nr` for documentation.
|
|
6
9
|
*/ export class FinalBlobAccumulator {
|
|
@@ -52,7 +55,7 @@ import { inspect } from 'util';
|
|
|
52
55
|
static random() {
|
|
53
56
|
return new FinalBlobAccumulator(Fr.random(), Fr.random(), BLS12Fr.random(), BLS12Point.random());
|
|
54
57
|
}
|
|
55
|
-
[
|
|
58
|
+
[_computedKey]() {
|
|
56
59
|
return `FinalBlobAccumulator {
|
|
57
60
|
blobCommitmentsHash: ${inspect(this.blobCommitmentsHash)},
|
|
58
61
|
z: ${inspect(this.z)},
|
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import { BLS12Fr
|
|
1
|
+
import { BLS12Fr } from '@aztec/foundation/curves/bls12';
|
|
2
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
2
3
|
import { BufferReader } from '@aztec/foundation/serialize';
|
|
3
4
|
/**
|
|
4
5
|
* See `noir-projects/noir-protocol-circuits/crates/blob/src/abis/final_blob_batching_challenges.nr` for documentation.
|
|
@@ -12,4 +13,4 @@ export declare class FinalBlobBatchingChallenges {
|
|
|
12
13
|
static fromBuffer(buffer: Buffer | BufferReader): FinalBlobBatchingChallenges;
|
|
13
14
|
toBuffer(): Buffer<ArrayBufferLike>;
|
|
14
15
|
}
|
|
15
|
-
//# sourceMappingURL=
|
|
16
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiZmluYWxfYmxvYl9iYXRjaGluZ19jaGFsbGVuZ2VzLmQudHMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY2lyY3VpdF90eXBlcy9maW5hbF9ibG9iX2JhdGNoaW5nX2NoYWxsZW5nZXMudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQUEsT0FBTyxFQUFFLE9BQU8sRUFBRSxNQUFNLGdDQUFnQyxDQUFDO0FBQ3pELE9BQU8sRUFBRSxFQUFFLEVBQUUsTUFBTSxnQ0FBZ0MsQ0FBQztBQUNwRCxPQUFPLEVBQUUsWUFBWSxFQUFxQixNQUFNLDZCQUE2QixDQUFDO0FBRTlFOztHQUVHO0FBQ0gscUJBQWEsMkJBQTJCO2FBRXBCLENBQUMsRUFBRSxFQUFFO2FBQ0wsS0FBSyxFQUFFLE9BQU87SUFGaEMsWUFDa0IsQ0FBQyxFQUFFLEVBQUUsRUFDTCxLQUFLLEVBQUUsT0FBTyxFQUM1QjtJQUVKLE1BQU0sQ0FBQyxLQUFLLEVBQUUsMkJBQTJCLFdBRXhDO0lBRUQsTUFBTSxDQUFDLEtBQUssSUFBSSwyQkFBMkIsQ0FFMUM7SUFFRCxNQUFNLENBQUMsVUFBVSxDQUFDLE1BQU0sRUFBRSxNQUFNLEdBQUcsWUFBWSxHQUFHLDJCQUEyQixDQUc1RTtJQUVELFFBQVEsNEJBRVA7Q0FDRiJ9
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"final_blob_batching_challenges.d.ts","sourceRoot":"","sources":["../../src/circuit_types/final_blob_batching_challenges.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,EAAE,EAAE,MAAM,
|
|
1
|
+
{"version":3,"file":"final_blob_batching_challenges.d.ts","sourceRoot":"","sources":["../../src/circuit_types/final_blob_batching_challenges.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,MAAM,gCAAgC,CAAC;AACzD,OAAO,EAAE,EAAE,EAAE,MAAM,gCAAgC,CAAC;AACpD,OAAO,EAAE,YAAY,EAAqB,MAAM,6BAA6B,CAAC;AAE9E;;GAEG;AACH,qBAAa,2BAA2B;aAEpB,CAAC,EAAE,EAAE;aACL,KAAK,EAAE,OAAO;IAFhC,YACkB,CAAC,EAAE,EAAE,EACL,KAAK,EAAE,OAAO,EAC5B;IAEJ,MAAM,CAAC,KAAK,EAAE,2BAA2B,WAExC;IAED,MAAM,CAAC,KAAK,IAAI,2BAA2B,CAE1C;IAED,MAAM,CAAC,UAAU,CAAC,MAAM,EAAE,MAAM,GAAG,YAAY,GAAG,2BAA2B,CAG5E;IAED,QAAQ,4BAEP;CACF"}
|
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import { BLS12Fr
|
|
1
|
+
import { BLS12Fr } from '@aztec/foundation/curves/bls12';
|
|
2
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
2
3
|
import { BufferReader, serializeToBuffer } from '@aztec/foundation/serialize';
|
|
3
4
|
/**
|
|
4
5
|
* See `noir-projects/noir-protocol-circuits/crates/blob/src/abis/final_blob_batching_challenges.nr` for documentation.
|
|
@@ -1,4 +1,4 @@
|
|
|
1
1
|
export * from './blob_accumulator.js';
|
|
2
2
|
export * from './final_blob_accumulator.js';
|
|
3
3
|
export * from './final_blob_batching_challenges.js';
|
|
4
|
-
//# sourceMappingURL=
|
|
4
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiaW5kZXguZC50cyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uL3NyYy9jaXJjdWl0X3R5cGVzL2luZGV4LnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUNBLGNBQWMsdUJBQXVCLENBQUM7QUFDdEMsY0FBYyw2QkFBNkIsQ0FBQztBQUM1QyxjQUFjLHFDQUFxQyxDQUFDIn0=
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
import { Fr } from '@aztec/foundation/curves/bn254';
|
|
2
|
+
import { FieldReader } from '@aztec/foundation/serialize';
|
|
3
|
+
import { type BlockEndMarker } from './block_end_marker.js';
|
|
4
|
+
import { type BlockEndStateField } from './block_end_state_field.js';
|
|
5
|
+
import { type TxBlobData } from './tx_blob_data.js';
|
|
6
|
+
export declare const NUM_BLOCK_END_BLOB_FIELDS = 6;
|
|
7
|
+
export declare const NUM_FIRST_BLOCK_END_BLOB_FIELDS = 7;
|
|
8
|
+
export declare const NUM_CHECKPOINT_END_MARKER_FIELDS = 1;
|
|
9
|
+
/**
|
|
10
|
+
* Returns the number of blob fields used for block end data.
|
|
11
|
+
* @param isFirstBlockInCheckpoint - Whether this is the first block in a checkpoint.
|
|
12
|
+
*/
|
|
13
|
+
export declare function getNumBlockEndBlobFields(isFirstBlockInCheckpoint: boolean): number;
|
|
14
|
+
export interface BlockEndBlobData {
|
|
15
|
+
blockEndMarker: BlockEndMarker;
|
|
16
|
+
blockEndStateField: BlockEndStateField;
|
|
17
|
+
lastArchiveRoot: Fr;
|
|
18
|
+
noteHashRoot: Fr;
|
|
19
|
+
nullifierRoot: Fr;
|
|
20
|
+
publicDataRoot: Fr;
|
|
21
|
+
l1ToL2MessageRoot: Fr | undefined;
|
|
22
|
+
}
|
|
23
|
+
export interface BlockBlobData extends BlockEndBlobData {
|
|
24
|
+
txs: TxBlobData[];
|
|
25
|
+
}
|
|
26
|
+
export declare function encodeBlockEndBlobData(blockEndBlobData: BlockEndBlobData): Fr[];
|
|
27
|
+
export declare function decodeBlockEndBlobData(fields: Fr[] | FieldReader, isFirstBlock: boolean): BlockEndBlobData;
|
|
28
|
+
export declare function encodeBlockBlobData(blockBlobData: BlockBlobData): Fr[];
|
|
29
|
+
export declare function decodeBlockBlobData(fields: Fr[] | FieldReader, isFirstBlock: boolean): BlockBlobData;
|
|
30
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiYmxvY2tfYmxvYl9kYXRhLmQudHMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvZW5jb2RpbmcvYmxvY2tfYmxvYl9kYXRhLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBLE9BQU8sRUFBRSxFQUFFLEVBQUUsTUFBTSxnQ0FBZ0MsQ0FBQztBQUNwRCxPQUFPLEVBQUUsV0FBVyxFQUFFLE1BQU0sNkJBQTZCLENBQUM7QUFHMUQsT0FBTyxFQUNMLEtBQUssY0FBYyxFQUlwQixNQUFNLHVCQUF1QixDQUFDO0FBQy9CLE9BQU8sRUFDTCxLQUFLLGtCQUFrQixFQUd4QixNQUFNLDRCQUE0QixDQUFDO0FBQ3BDLE9BQU8sRUFBRSxLQUFLLFVBQVUsRUFBc0MsTUFBTSxtQkFBbUIsQ0FBQztBQUl4RixlQUFPLE1BQU0seUJBQXlCLElBQUksQ0FBQztBQUMzQyxlQUFPLE1BQU0sK0JBQStCLElBQUksQ0FBQztBQUNqRCxlQUFPLE1BQU0sZ0NBQWdDLElBQUksQ0FBQztBQUVsRDs7O0dBR0c7QUFDSCx3QkFBZ0Isd0JBQXdCLENBQUMsd0JBQXdCLEVBQUUsT0FBTyxHQUFHLE1BQU0sQ0FFbEY7QUFFRCxNQUFNLFdBQVcsZ0JBQWdCO0lBQy9CLGNBQWMsRUFBRSxjQUFjLENBQUM7SUFDL0Isa0JBQWtCLEVBQUUsa0JBQWtCLENBQUM7SUFDdkMsZUFBZSxFQUFFLEVBQUUsQ0FBQztJQUNwQixZQUFZLEVBQUUsRUFBRSxDQUFDO0lBQ2pCLGFBQWEsRUFBRSxFQUFFLENBQUM7SUFDbEIsY0FBYyxFQUFFLEVBQUUsQ0FBQztJQUNuQixpQkFBaUIsRUFBRSxFQUFFLEdBQUcsU0FBUyxDQUFDO0NBQ25DO0FBRUQsTUFBTSxXQUFXLGFBQWMsU0FBUSxnQkFBZ0I7SUFDckQsR0FBRyxFQUFFLFVBQVUsRUFBRSxDQUFDO0NBQ25CO0FBRUQsd0JBQWdCLHNCQUFzQixDQUFDLGdCQUFnQixFQUFFLGdCQUFnQixHQUFHLEVBQUUsRUFBRSxDQVUvRTtBQUVELHdCQUFnQixzQkFBc0IsQ0FBQyxNQUFNLEVBQUUsRUFBRSxFQUFFLEdBQUcsV0FBVyxFQUFFLFlBQVksRUFBRSxPQUFPLEdBQUcsZ0JBQWdCLENBbUIxRztBQUVELHdCQUFnQixtQkFBbUIsQ0FBQyxhQUFhLEVBQUUsYUFBYSxHQUFHLEVBQUUsRUFBRSxDQUV0RTtBQUVELHdCQUFnQixtQkFBbUIsQ0FBQyxNQUFNLEVBQUUsRUFBRSxFQUFFLEdBQUcsV0FBVyxFQUFFLFlBQVksRUFBRSxPQUFPLEdBQUcsYUFBYSxDQStCcEcifQ==
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"block_blob_data.d.ts","sourceRoot":"","sources":["../../src/encoding/block_blob_data.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,EAAE,EAAE,MAAM,gCAAgC,CAAC;AACpD,OAAO,EAAE,WAAW,EAAE,MAAM,6BAA6B,CAAC;AAG1D,OAAO,EACL,KAAK,cAAc,EAIpB,MAAM,uBAAuB,CAAC;AAC/B,OAAO,EACL,KAAK,kBAAkB,EAGxB,MAAM,4BAA4B,CAAC;AACpC,OAAO,EAAE,KAAK,UAAU,EAAsC,MAAM,mBAAmB,CAAC;AAIxF,eAAO,MAAM,yBAAyB,IAAI,CAAC;AAC3C,eAAO,MAAM,+BAA+B,IAAI,CAAC;AACjD,eAAO,MAAM,gCAAgC,IAAI,CAAC;AAElD;;;GAGG;AACH,wBAAgB,wBAAwB,CAAC,wBAAwB,EAAE,OAAO,GAAG,MAAM,CAElF;AAED,MAAM,WAAW,gBAAgB;IAC/B,cAAc,EAAE,cAAc,CAAC;IAC/B,kBAAkB,EAAE,kBAAkB,CAAC;IACvC,eAAe,EAAE,EAAE,CAAC;IACpB,YAAY,EAAE,EAAE,CAAC;IACjB,aAAa,EAAE,EAAE,CAAC;IAClB,cAAc,EAAE,EAAE,CAAC;IACnB,iBAAiB,EAAE,EAAE,GAAG,SAAS,CAAC;CACnC;AAED,MAAM,WAAW,aAAc,SAAQ,gBAAgB;IACrD,GAAG,EAAE,UAAU,EAAE,CAAC;CACnB;AAED,wBAAgB,sBAAsB,CAAC,gBAAgB,EAAE,gBAAgB,GAAG,EAAE,EAAE,CAU/E;AAED,wBAAgB,sBAAsB,CAAC,MAAM,EAAE,EAAE,EAAE,GAAG,WAAW,EAAE,YAAY,EAAE,OAAO,GAAG,gBAAgB,CAmB1G;AAED,wBAAgB,mBAAmB,CAAC,aAAa,EAAE,aAAa,GAAG,EAAE,EAAE,CAEtE;AAED,wBAAgB,mBAAmB,CAAC,MAAM,EAAE,EAAE,EAAE,GAAG,WAAW,EAAE,YAAY,EAAE,OAAO,GAAG,aAAa,CA+BpG"}
|
|
@@ -0,0 +1,75 @@
|
|
|
1
|
+
import { FieldReader } from '@aztec/foundation/serialize';
|
|
2
|
+
import { BlobDeserializationError } from '../errors.js';
|
|
3
|
+
import { decodeBlockEndMarker, encodeBlockEndMarker, isBlockEndMarker } from './block_end_marker.js';
|
|
4
|
+
import { decodeBlockEndStateField, encodeBlockEndStateField } from './block_end_state_field.js';
|
|
5
|
+
import { decodeTxBlobData, encodeTxBlobData } from './tx_blob_data.js';
|
|
6
|
+
// Must match the implementation in `noir-protocol-circuits/crates/types/src/blob_data/block_blob_data.nr`.
|
|
7
|
+
export const NUM_BLOCK_END_BLOB_FIELDS = 6;
|
|
8
|
+
export const NUM_FIRST_BLOCK_END_BLOB_FIELDS = 7;
|
|
9
|
+
export const NUM_CHECKPOINT_END_MARKER_FIELDS = 1;
|
|
10
|
+
/**
|
|
11
|
+
* Returns the number of blob fields used for block end data.
|
|
12
|
+
* @param isFirstBlockInCheckpoint - Whether this is the first block in a checkpoint.
|
|
13
|
+
*/ export function getNumBlockEndBlobFields(isFirstBlockInCheckpoint) {
|
|
14
|
+
return isFirstBlockInCheckpoint ? NUM_FIRST_BLOCK_END_BLOB_FIELDS : NUM_BLOCK_END_BLOB_FIELDS;
|
|
15
|
+
}
|
|
16
|
+
export function encodeBlockEndBlobData(blockEndBlobData) {
|
|
17
|
+
return [
|
|
18
|
+
encodeBlockEndMarker(blockEndBlobData.blockEndMarker),
|
|
19
|
+
encodeBlockEndStateField(blockEndBlobData.blockEndStateField),
|
|
20
|
+
blockEndBlobData.lastArchiveRoot,
|
|
21
|
+
blockEndBlobData.noteHashRoot,
|
|
22
|
+
blockEndBlobData.nullifierRoot,
|
|
23
|
+
blockEndBlobData.publicDataRoot,
|
|
24
|
+
...blockEndBlobData.l1ToL2MessageRoot ? [
|
|
25
|
+
blockEndBlobData.l1ToL2MessageRoot
|
|
26
|
+
] : []
|
|
27
|
+
];
|
|
28
|
+
}
|
|
29
|
+
export function decodeBlockEndBlobData(fields, isFirstBlock) {
|
|
30
|
+
const reader = FieldReader.asReader(fields);
|
|
31
|
+
const numBlockEndData = getNumBlockEndBlobFields(isFirstBlock);
|
|
32
|
+
if (numBlockEndData > reader.remainingFields()) {
|
|
33
|
+
throw new BlobDeserializationError(`Incorrect encoding of blob fields: not enough fields for block end data. Expected ${numBlockEndData} fields, only ${reader.remainingFields()} remaining.`);
|
|
34
|
+
}
|
|
35
|
+
return {
|
|
36
|
+
blockEndMarker: decodeBlockEndMarker(reader.readField()),
|
|
37
|
+
blockEndStateField: decodeBlockEndStateField(reader.readField()),
|
|
38
|
+
lastArchiveRoot: reader.readField(),
|
|
39
|
+
noteHashRoot: reader.readField(),
|
|
40
|
+
nullifierRoot: reader.readField(),
|
|
41
|
+
publicDataRoot: reader.readField(),
|
|
42
|
+
l1ToL2MessageRoot: isFirstBlock ? reader.readField() : undefined
|
|
43
|
+
};
|
|
44
|
+
}
|
|
45
|
+
export function encodeBlockBlobData(blockBlobData) {
|
|
46
|
+
return [
|
|
47
|
+
...blockBlobData.txs.map((tx)=>encodeTxBlobData(tx)).flat(),
|
|
48
|
+
...encodeBlockEndBlobData(blockBlobData)
|
|
49
|
+
];
|
|
50
|
+
}
|
|
51
|
+
export function decodeBlockBlobData(fields, isFirstBlock) {
|
|
52
|
+
const reader = FieldReader.asReader(fields);
|
|
53
|
+
const txs = [];
|
|
54
|
+
let hasReachedBlockEnd = false;
|
|
55
|
+
while(!hasReachedBlockEnd){
|
|
56
|
+
if (reader.isFinished()) {
|
|
57
|
+
throw new BlobDeserializationError(`Incorrect encoding of blob fields: not enough fields for block end marker.`);
|
|
58
|
+
}
|
|
59
|
+
const currentField = reader.peekField();
|
|
60
|
+
if (isBlockEndMarker(currentField)) {
|
|
61
|
+
hasReachedBlockEnd = true;
|
|
62
|
+
} else {
|
|
63
|
+
txs.push(decodeTxBlobData(reader));
|
|
64
|
+
}
|
|
65
|
+
}
|
|
66
|
+
const blockEndBlobData = decodeBlockEndBlobData(reader, isFirstBlock);
|
|
67
|
+
const blockEndMarker = blockEndBlobData.blockEndMarker;
|
|
68
|
+
if (blockEndMarker.numTxs !== txs.length) {
|
|
69
|
+
throw new BlobDeserializationError(`Incorrect encoding of blob fields: expected ${blockEndMarker.numTxs} txs, but got ${txs.length}.`);
|
|
70
|
+
}
|
|
71
|
+
return {
|
|
72
|
+
txs,
|
|
73
|
+
...blockEndBlobData
|
|
74
|
+
};
|
|
75
|
+
}
|