split-hash 0.1.5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +11 -0
- package/LICENSE +21 -0
- package/README.md +123 -0
- package/lib/es2015/index.d.ts +3 -0
- package/lib/es2015/index.js +16 -0
- package/lib/es2015/index.js.map +1 -0
- package/lib/es2015/split-hash-validator.d.ts +13 -0
- package/lib/es2015/split-hash-validator.js +88 -0
- package/lib/es2015/split-hash-validator.js.map +1 -0
- package/lib/es2015/split-hash.d.ts +7 -0
- package/lib/es2015/split-hash.js +78 -0
- package/lib/es2015/split-hash.js.map +1 -0
- package/lib/es2015/types.d.ts +6 -0
- package/lib/es2015/types.js +3 -0
- package/lib/es2015/types.js.map +1 -0
- package/lib/es2018/index.d.ts +3 -0
- package/lib/es2018/index.js +16 -0
- package/lib/es2018/index.js.map +1 -0
- package/lib/es2018/split-hash-validator.d.ts +13 -0
- package/lib/es2018/split-hash-validator.js +88 -0
- package/lib/es2018/split-hash-validator.js.map +1 -0
- package/lib/es2018/split-hash.d.ts +7 -0
- package/lib/es2018/split-hash.js +46 -0
- package/lib/es2018/split-hash.js.map +1 -0
- package/lib/es2018/types.d.ts +6 -0
- package/lib/es2018/types.js +3 -0
- package/lib/es2018/types.js.map +1 -0
- package/package.json +67 -0
package/CHANGELOG.md
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
# Changelog
|
|
2
|
+
|
|
3
|
+
All notable changes to this project will be documented in this file. See [standard-version](https://github.com/conventional-changelog/standard-version) for commit guidelines.
|
|
4
|
+
|
|
5
|
+
### [0.1.5](https://github.com/BlackGlory/split-hash/compare/v0.1.4...v0.1.5) (2021-10-14)
|
|
6
|
+
|
|
7
|
+
### [0.1.4](https://github.com/BlackGlory/split-hash/compare/v0.1.3...v0.1.4) (2021-02-04)
|
|
8
|
+
|
|
9
|
+
### [0.1.3](https://github.com/BlackGlory/split-hash/compare/v0.1.2...v0.1.3) (2021-02-03)
|
|
10
|
+
|
|
11
|
+
### [0.1.2](https://github.com/BlackGlory/split-hash/compare/v0.1.1...v0.1.2) (2020-10-10)
|
package/LICENSE
ADDED
|
@@ -0,0 +1,21 @@
|
|
|
1
|
+
MIT License
|
|
2
|
+
|
|
3
|
+
Copyright (c) 2020 BlackGlory <woshenmedoubuzhidao@blackglory.me>
|
|
4
|
+
|
|
5
|
+
Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
6
|
+
of this software and associated documentation files (the "Software"), to deal
|
|
7
|
+
in the Software without restriction, including without limitation the rights
|
|
8
|
+
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
9
|
+
copies of the Software, and to permit persons to whom the Software is
|
|
10
|
+
furnished to do so, subject to the following conditions:
|
|
11
|
+
|
|
12
|
+
The above copyright notice and this permission notice shall be included in all
|
|
13
|
+
copies or substantial portions of the Software.
|
|
14
|
+
|
|
15
|
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
16
|
+
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
17
|
+
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
18
|
+
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
19
|
+
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
20
|
+
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
21
|
+
SOFTWARE.
|
package/README.md
ADDED
|
@@ -0,0 +1,123 @@
|
|
|
1
|
+
# split-hash
|
|
2
|
+
|
|
3
|
+
Split the stream based on bytes and get digests from each part.
|
|
4
|
+
|
|
5
|
+
## Install
|
|
6
|
+
|
|
7
|
+
```sh
|
|
8
|
+
npm install --save split-hash
|
|
9
|
+
# or
|
|
10
|
+
yarn add split-hash
|
|
11
|
+
```
|
|
12
|
+
|
|
13
|
+
## Usage
|
|
14
|
+
|
|
15
|
+
### Hash
|
|
16
|
+
|
|
17
|
+
```js
|
|
18
|
+
import { splitHash } from 'split-hash'
|
|
19
|
+
import * as crypto from 'crypto'
|
|
20
|
+
|
|
21
|
+
const KiB = 1024
|
|
22
|
+
|
|
23
|
+
const createHash = () => {
|
|
24
|
+
const hash = crypto.createHash('sha256')
|
|
25
|
+
return {
|
|
26
|
+
update(buffer) {
|
|
27
|
+
hash.update(buffer)
|
|
28
|
+
}
|
|
29
|
+
, digest() {
|
|
30
|
+
return hash.digest('hex')
|
|
31
|
+
}
|
|
32
|
+
}
|
|
33
|
+
}
|
|
34
|
+
|
|
35
|
+
const stream = fs.createReadStream('filename.bin')
|
|
36
|
+
const iter = splitHash(stream, 512 * KiB, createHash)
|
|
37
|
+
|
|
38
|
+
for await (const hash of iter) {
|
|
39
|
+
console.log(hash)
|
|
40
|
+
}
|
|
41
|
+
```
|
|
42
|
+
|
|
43
|
+
### Validate
|
|
44
|
+
|
|
45
|
+
```js
|
|
46
|
+
import { SplitHashValidator } from 'split-hash'
|
|
47
|
+
import * as crypto from 'crypto'
|
|
48
|
+
|
|
49
|
+
const KiB = 1024
|
|
50
|
+
|
|
51
|
+
const createHash = () => {
|
|
52
|
+
const hash = crypto.createHash('sha256')
|
|
53
|
+
return {
|
|
54
|
+
update(buffer) {
|
|
55
|
+
hash.update(buffer)
|
|
56
|
+
}
|
|
57
|
+
, digest() {
|
|
58
|
+
return hash.digest('hex')
|
|
59
|
+
}
|
|
60
|
+
}
|
|
61
|
+
}
|
|
62
|
+
|
|
63
|
+
const hashList = [/* ... */]
|
|
64
|
+
const validator = new SplitHashValidator(hashList, 512 * KiB, createHash)
|
|
65
|
+
|
|
66
|
+
const stream = fs.createReadStream('filename.bin')
|
|
67
|
+
stream
|
|
68
|
+
.pipe(validator)
|
|
69
|
+
.on('data', /* same as stream */)
|
|
70
|
+
.on('error', err => console.error('not matched'))
|
|
71
|
+
```
|
|
72
|
+
|
|
73
|
+
## Interface
|
|
74
|
+
|
|
75
|
+
```ts
|
|
76
|
+
type ProgressiveHashFactory<T> = () => ProgressiveHash<T>
|
|
77
|
+
|
|
78
|
+
interface ProgressiveHash<T> {
|
|
79
|
+
update(buffer: Buffer): void
|
|
80
|
+
digest(): T
|
|
81
|
+
}
|
|
82
|
+
```
|
|
83
|
+
|
|
84
|
+
## API
|
|
85
|
+
|
|
86
|
+
### splitHash
|
|
87
|
+
|
|
88
|
+
```ts
|
|
89
|
+
function splitHash<T>(
|
|
90
|
+
stream: NodeJS.ReadableStream
|
|
91
|
+
, blockSize: number
|
|
92
|
+
, createHash: ProgressiveHashFactory<T>
|
|
93
|
+
): AsyncIterable<T>
|
|
94
|
+
```
|
|
95
|
+
|
|
96
|
+
It throws `StreamEncodingError` when the `stream` encoding is set.
|
|
97
|
+
|
|
98
|
+
### SplitHashValidator
|
|
99
|
+
|
|
100
|
+
```ts
|
|
101
|
+
class SplitHashValidator<T> extends Stream.Transform {
|
|
102
|
+
constructor(
|
|
103
|
+
digests: T[]
|
|
104
|
+
, blockSize: number
|
|
105
|
+
, createHash: ProgressiveHashFactory<T>
|
|
106
|
+
, equals: (a: T, b: T) => boolean = Object.is
|
|
107
|
+
)
|
|
108
|
+
}
|
|
109
|
+
```
|
|
110
|
+
|
|
111
|
+
It throws `NotMatchedError` when the `stream` does not match digests.
|
|
112
|
+
|
|
113
|
+
### StreamEncodingError
|
|
114
|
+
|
|
115
|
+
```ts
|
|
116
|
+
class StreamEncodingError extends Error
|
|
117
|
+
```
|
|
118
|
+
|
|
119
|
+
### NotMatchedError
|
|
120
|
+
|
|
121
|
+
```ts
|
|
122
|
+
class NotMatchedError extends Error
|
|
123
|
+
```
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
3
|
+
if (k2 === undefined) k2 = k;
|
|
4
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
5
|
+
}) : (function(o, m, k, k2) {
|
|
6
|
+
if (k2 === undefined) k2 = k;
|
|
7
|
+
o[k2] = m[k];
|
|
8
|
+
}));
|
|
9
|
+
var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
10
|
+
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
11
|
+
};
|
|
12
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
13
|
+
__exportStar(require("./types"), exports);
|
|
14
|
+
__exportStar(require("./split-hash"), exports);
|
|
15
|
+
__exportStar(require("./split-hash-validator"), exports);
|
|
16
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;AAAA,6CAA0B;AAC1B,kDAA+B;AAC/B,4DAAyC"}
|
|
@@ -0,0 +1,13 @@
|
|
|
1
|
+
/// <reference types="node" />
|
|
2
|
+
import { ProgressiveHashFactory } from './types';
|
|
3
|
+
import { Transform, TransformCallback } from 'stream';
|
|
4
|
+
export declare class SplitHashValidator<T> extends Transform {
|
|
5
|
+
#private;
|
|
6
|
+
constructor(digests: T[], blockSize: number, createHash: ProgressiveHashFactory<T>, equals?: (a: T, b: T) => boolean);
|
|
7
|
+
_transform(chunk: Buffer, encoding: BufferEncoding, callback: TransformCallback): void;
|
|
8
|
+
_flush(callback: TransformCallback): void;
|
|
9
|
+
}
|
|
10
|
+
export declare class NotMatchedError extends Error {
|
|
11
|
+
name: string;
|
|
12
|
+
constructor();
|
|
13
|
+
}
|
|
@@ -0,0 +1,88 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __classPrivateFieldSet = (this && this.__classPrivateFieldSet) || function (receiver, state, value, kind, f) {
|
|
3
|
+
if (kind === "m") throw new TypeError("Private method is not writable");
|
|
4
|
+
if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a setter");
|
|
5
|
+
if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot write private member to an object whose class did not declare it");
|
|
6
|
+
return (kind === "a" ? f.call(receiver, value) : f ? f.value = value : state.set(receiver, value)), value;
|
|
7
|
+
};
|
|
8
|
+
var __classPrivateFieldGet = (this && this.__classPrivateFieldGet) || function (receiver, state, kind, f) {
|
|
9
|
+
if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a getter");
|
|
10
|
+
if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot read private member from an object whose class did not declare it");
|
|
11
|
+
return kind === "m" ? f : kind === "a" ? f.call(receiver) : f ? f.value : state.get(receiver);
|
|
12
|
+
};
|
|
13
|
+
var _SplitHashValidator_digests, _SplitHashValidator_blockSize, _SplitHashValidator_createHash, _SplitHashValidator_hash, _SplitHashValidator_equals, _SplitHashValidator_accu, _SplitHashValidator_digestIndex;
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
exports.NotMatchedError = exports.SplitHashValidator = void 0;
|
|
16
|
+
const stream_1 = require("stream");
|
|
17
|
+
class SplitHashValidator extends stream_1.Transform {
|
|
18
|
+
constructor(digests, blockSize, createHash, equals = Object.is) {
|
|
19
|
+
super();
|
|
20
|
+
_SplitHashValidator_digests.set(this, void 0);
|
|
21
|
+
_SplitHashValidator_blockSize.set(this, void 0);
|
|
22
|
+
_SplitHashValidator_createHash.set(this, void 0);
|
|
23
|
+
_SplitHashValidator_hash.set(this, void 0);
|
|
24
|
+
_SplitHashValidator_equals.set(this, void 0);
|
|
25
|
+
_SplitHashValidator_accu.set(this, 0);
|
|
26
|
+
_SplitHashValidator_digestIndex.set(this, 0);
|
|
27
|
+
__classPrivateFieldSet(this, _SplitHashValidator_digests, digests, "f");
|
|
28
|
+
__classPrivateFieldSet(this, _SplitHashValidator_blockSize, blockSize, "f");
|
|
29
|
+
__classPrivateFieldSet(this, _SplitHashValidator_createHash, createHash, "f");
|
|
30
|
+
__classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
|
|
31
|
+
__classPrivateFieldSet(this, _SplitHashValidator_equals, equals, "f");
|
|
32
|
+
}
|
|
33
|
+
_transform(chunk, encoding, callback) {
|
|
34
|
+
var _a;
|
|
35
|
+
if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length < __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f")) {
|
|
36
|
+
__classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(chunk);
|
|
37
|
+
__classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length, "f");
|
|
38
|
+
}
|
|
39
|
+
else {
|
|
40
|
+
let offset = 0;
|
|
41
|
+
while (true) {
|
|
42
|
+
const needed = __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f") - __classPrivateFieldGet(this, _SplitHashValidator_accu, "f");
|
|
43
|
+
const slice = chunk.slice(offset, offset + needed);
|
|
44
|
+
if (slice.length === needed) {
|
|
45
|
+
__classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
|
|
46
|
+
const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
|
|
47
|
+
if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
|
|
48
|
+
return callback(new NotMatchedError());
|
|
49
|
+
}
|
|
50
|
+
__classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
|
|
51
|
+
__classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
|
|
52
|
+
__classPrivateFieldSet(this, _SplitHashValidator_accu, 0, "f");
|
|
53
|
+
offset += slice.length;
|
|
54
|
+
}
|
|
55
|
+
else {
|
|
56
|
+
__classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
|
|
57
|
+
__classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + slice.length, "f");
|
|
58
|
+
break;
|
|
59
|
+
}
|
|
60
|
+
}
|
|
61
|
+
}
|
|
62
|
+
callback(null, chunk);
|
|
63
|
+
}
|
|
64
|
+
_flush(callback) {
|
|
65
|
+
var _a;
|
|
66
|
+
if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") > 0) {
|
|
67
|
+
const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
|
|
68
|
+
if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
|
|
69
|
+
return callback(new NotMatchedError());
|
|
70
|
+
}
|
|
71
|
+
__classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
|
|
72
|
+
}
|
|
73
|
+
if (__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f") !== __classPrivateFieldGet(this, _SplitHashValidator_digests, "f").length) {
|
|
74
|
+
return callback(new NotMatchedError());
|
|
75
|
+
}
|
|
76
|
+
callback();
|
|
77
|
+
}
|
|
78
|
+
}
|
|
79
|
+
exports.SplitHashValidator = SplitHashValidator;
|
|
80
|
+
_SplitHashValidator_digests = new WeakMap(), _SplitHashValidator_blockSize = new WeakMap(), _SplitHashValidator_createHash = new WeakMap(), _SplitHashValidator_hash = new WeakMap(), _SplitHashValidator_equals = new WeakMap(), _SplitHashValidator_accu = new WeakMap(), _SplitHashValidator_digestIndex = new WeakMap();
|
|
81
|
+
class NotMatchedError extends Error {
|
|
82
|
+
constructor() {
|
|
83
|
+
super('hashes do not match');
|
|
84
|
+
this.name = this.constructor.name;
|
|
85
|
+
}
|
|
86
|
+
}
|
|
87
|
+
exports.NotMatchedError = NotMatchedError;
|
|
88
|
+
//# sourceMappingURL=split-hash-validator.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"split-hash-validator.js","sourceRoot":"","sources":["../../src/split-hash-validator.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;AACA,mCAAqD;AAErD,MAAa,kBAAsB,SAAQ,kBAAS;IASlD,YACE,OAAY,EACZ,SAAiB,EACjB,UAAqC,EACrC,SAAkC,MAAM,CAAC,EAAE;QAE3C,KAAK,EAAE,CAAA;QAdT,8CAAa;QACb,gDAAkB;QAClB,iDAAsC;QACtC,2CAAyB;QACzB,6CAAgC;QAChC,mCAAQ,CAAC,EAAA;QACT,0CAAe,CAAC,EAAA;QAUd,uBAAA,IAAI,+BAAY,OAAO,MAAA,CAAA;QACvB,uBAAA,IAAI,iCAAc,SAAS,MAAA,CAAA;QAC3B,uBAAA,IAAI,kCAAe,UAAU,MAAA,CAAA;QAC7B,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;QAC/B,uBAAA,IAAI,8BAAW,MAAM,MAAA,CAAA;IACvB,CAAC;IAED,UAAU,CAAC,KAAa,EAAE,QAAwB,EAAE,QAA2B;;QAG7E,IAAI,uBAAA,IAAI,gCAAM,GAAG,KAAK,CAAC,MAAM,GAAG,uBAAA,IAAI,qCAAW,EAAE;YAC/C,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;YACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;SAC3B;aAAM;YACL,IAAI,MAAM,GAAG,CAAC,CAAA;YACd,OAAO,IAAI,EAAE;gBACX,MAAM,MAAM,GAAG,uBAAA,IAAI,qCAAW,GAAG,uBAAA,IAAI,gCAAM,CAAA;gBAC3C,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;gBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;oBAC3B,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;oBAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;wBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;qBACvC;oBACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;oBAEnB,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;oBAC/B,uBAAA,IAAI,4BAAS,CAAC,MAAA,CAAA;oBACd,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;iBACvB;qBAAM;oBAEL,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;oBAC1B,MAAK;iBACN;aACF;SACF;QAED,QAAQ,CAAC,IAAI,EAAE,KAAK,CAAC,CAAA;IACvB,CAAC;IAED,MAAM,CAAC,QAA2B;;QAChC,IAAI,uBAAA,IAAI,gCAAM,GAAG,CAAC,EAAE;YAClB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;YAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;gBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;aACvC;YACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;SACpB;QAED,IAAI,uBAAA,IAAI,uCAAa,KAAK,uBAAA,IAAI,mCAAS,CAAC,MAAM,EAAE;YAC9C,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;SACvC;QAED,QAAQ,EAAE,CAAA;IACZ,CAAC;CACF;AAzED,gDAyEC;;AAED,MAAa,eAAgB,SAAQ,KAAK;IAGxC;QACE,KAAK,CAAC,qBAAqB,CAAC,CAAA;QAH9B,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,0CAMC"}
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
/// <reference types="node" />
|
|
2
|
+
import { ProgressiveHashFactory } from './types';
|
|
3
|
+
export declare function splitHash<T>(stream: NodeJS.ReadableStream, blockSize: number, createHash: ProgressiveHashFactory<T>): AsyncIterable<T>;
|
|
4
|
+
export declare class StreamEncodingError extends Error {
|
|
5
|
+
name: string;
|
|
6
|
+
constructor();
|
|
7
|
+
}
|
|
@@ -0,0 +1,78 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __asyncValues = (this && this.__asyncValues) || function (o) {
|
|
3
|
+
if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
|
|
4
|
+
var m = o[Symbol.asyncIterator], i;
|
|
5
|
+
return m ? m.call(o) : (o = typeof __values === "function" ? __values(o) : o[Symbol.iterator](), i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i);
|
|
6
|
+
function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
|
|
7
|
+
function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
|
|
8
|
+
};
|
|
9
|
+
var __await = (this && this.__await) || function (v) { return this instanceof __await ? (this.v = v, this) : new __await(v); }
|
|
10
|
+
var __asyncGenerator = (this && this.__asyncGenerator) || function (thisArg, _arguments, generator) {
|
|
11
|
+
if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
|
|
12
|
+
var g = generator.apply(thisArg, _arguments || []), i, q = [];
|
|
13
|
+
return i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i;
|
|
14
|
+
function verb(n) { if (g[n]) i[n] = function (v) { return new Promise(function (a, b) { q.push([n, v, a, b]) > 1 || resume(n, v); }); }; }
|
|
15
|
+
function resume(n, v) { try { step(g[n](v)); } catch (e) { settle(q[0][3], e); } }
|
|
16
|
+
function step(r) { r.value instanceof __await ? Promise.resolve(r.value.v).then(fulfill, reject) : settle(q[0][2], r); }
|
|
17
|
+
function fulfill(value) { resume("next", value); }
|
|
18
|
+
function reject(value) { resume("throw", value); }
|
|
19
|
+
function settle(f, v) { if (f(v), q.shift(), q.length) resume(q[0][0], q[0][1]); }
|
|
20
|
+
};
|
|
21
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
22
|
+
exports.StreamEncodingError = exports.splitHash = void 0;
|
|
23
|
+
function splitHash(stream, blockSize, createHash) {
|
|
24
|
+
return __asyncGenerator(this, arguments, function* splitHash_1() {
|
|
25
|
+
var e_1, _a;
|
|
26
|
+
let hash = createHash();
|
|
27
|
+
let accu = 0;
|
|
28
|
+
try {
|
|
29
|
+
for (var stream_1 = __asyncValues(stream), stream_1_1; stream_1_1 = yield __await(stream_1.next()), !stream_1_1.done;) {
|
|
30
|
+
const chunk = stream_1_1.value;
|
|
31
|
+
if (!Buffer.isBuffer(chunk))
|
|
32
|
+
throw new StreamEncodingError();
|
|
33
|
+
if (accu + chunk.length < blockSize) {
|
|
34
|
+
hash.update(chunk);
|
|
35
|
+
accu += chunk.length;
|
|
36
|
+
}
|
|
37
|
+
else {
|
|
38
|
+
let offset = 0;
|
|
39
|
+
while (true) {
|
|
40
|
+
const needed = blockSize - accu;
|
|
41
|
+
const slice = chunk.slice(offset, offset + needed);
|
|
42
|
+
if (slice.length === needed) {
|
|
43
|
+
hash.update(slice);
|
|
44
|
+
const digest = hash.digest();
|
|
45
|
+
yield yield __await(digest);
|
|
46
|
+
hash = createHash();
|
|
47
|
+
accu = 0;
|
|
48
|
+
offset += slice.length;
|
|
49
|
+
}
|
|
50
|
+
else {
|
|
51
|
+
hash.update(slice);
|
|
52
|
+
accu += slice.length;
|
|
53
|
+
break;
|
|
54
|
+
}
|
|
55
|
+
}
|
|
56
|
+
}
|
|
57
|
+
}
|
|
58
|
+
}
|
|
59
|
+
catch (e_1_1) { e_1 = { error: e_1_1 }; }
|
|
60
|
+
finally {
|
|
61
|
+
try {
|
|
62
|
+
if (stream_1_1 && !stream_1_1.done && (_a = stream_1.return)) yield __await(_a.call(stream_1));
|
|
63
|
+
}
|
|
64
|
+
finally { if (e_1) throw e_1.error; }
|
|
65
|
+
}
|
|
66
|
+
if (accu > 0)
|
|
67
|
+
yield yield __await(hash.digest());
|
|
68
|
+
});
|
|
69
|
+
}
|
|
70
|
+
exports.splitHash = splitHash;
|
|
71
|
+
class StreamEncodingError extends Error {
|
|
72
|
+
constructor() {
|
|
73
|
+
super('stream encoding must not be set.');
|
|
74
|
+
this.name = this.constructor.name;
|
|
75
|
+
}
|
|
76
|
+
}
|
|
77
|
+
exports.StreamEncodingError = StreamEncodingError;
|
|
78
|
+
//# sourceMappingURL=split-hash.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"split-hash.js","sourceRoot":"","sources":["../../src/split-hash.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;AAEA,SAAuB,SAAS,CAAI,MAA6B,EAAE,SAAiB,EAAE,UAAqC;;;QACzH,IAAI,IAAI,GAAG,UAAU,EAAE,CAAA;QACvB,IAAI,IAAI,GAAG,CAAC,CAAA;;YACZ,KAA0B,IAAA,WAAA,cAAA,MAAM,CAAA,YAAA;gBAArB,MAAM,KAAK,mBAAA,CAAA;gBACpB,IAAI,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC;oBAAE,MAAM,IAAI,mBAAmB,EAAE,CAAA;gBAC5D,IAAI,IAAI,GAAG,KAAK,CAAC,MAAM,GAAG,SAAS,EAAE;oBACnC,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;iBACrB;qBAAM;oBACL,IAAI,MAAM,GAAG,CAAC,CAAA;oBACd,OAAO,IAAI,EAAE;wBACX,MAAM,MAAM,GAAG,SAAS,GAAG,IAAI,CAAA;wBAC/B,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;wBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;4BAC3B,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;4BAClB,MAAM,MAAM,GAAG,IAAI,CAAC,MAAM,EAAE,CAAA;4BAC5B,oBAAM,MAAM,CAAA,CAAA;4BAEZ,IAAI,GAAG,UAAU,EAAE,CAAA;4BACnB,IAAI,GAAG,CAAC,CAAA;4BACR,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;yBACvB;6BAAM;4BAEL,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;4BAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;4BACpB,MAAK;yBACN;qBACF;iBACF;aACF;;;;;;;;;QAED,IAAI,IAAI,GAAG,CAAC;YAAE,oBAAM,IAAI,CAAC,MAAM,EAAE,CAAA,CAAA;IACnC,CAAC;CAAA;AAhCD,8BAgCC;AAED,MAAa,mBAAoB,SAAQ,KAAK;IAG5C;QACE,KAAK,CAAC,kCAAkC,CAAC,CAAA;QAH3C,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,kDAMC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"types.js","sourceRoot":"","sources":["../../src/types.ts"],"names":[],"mappings":""}
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
3
|
+
if (k2 === undefined) k2 = k;
|
|
4
|
+
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
5
|
+
}) : (function(o, m, k, k2) {
|
|
6
|
+
if (k2 === undefined) k2 = k;
|
|
7
|
+
o[k2] = m[k];
|
|
8
|
+
}));
|
|
9
|
+
var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
10
|
+
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
11
|
+
};
|
|
12
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
13
|
+
__exportStar(require("./types"), exports);
|
|
14
|
+
__exportStar(require("./split-hash"), exports);
|
|
15
|
+
__exportStar(require("./split-hash-validator"), exports);
|
|
16
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;AAAA,6CAA0B;AAC1B,kDAA+B;AAC/B,4DAAyC"}
|
|
@@ -0,0 +1,13 @@
|
|
|
1
|
+
/// <reference types="node" />
|
|
2
|
+
import { ProgressiveHashFactory } from './types';
|
|
3
|
+
import { Transform, TransformCallback } from 'stream';
|
|
4
|
+
export declare class SplitHashValidator<T> extends Transform {
|
|
5
|
+
#private;
|
|
6
|
+
constructor(digests: T[], blockSize: number, createHash: ProgressiveHashFactory<T>, equals?: (a: T, b: T) => boolean);
|
|
7
|
+
_transform(chunk: Buffer, encoding: BufferEncoding, callback: TransformCallback): void;
|
|
8
|
+
_flush(callback: TransformCallback): void;
|
|
9
|
+
}
|
|
10
|
+
export declare class NotMatchedError extends Error {
|
|
11
|
+
name: string;
|
|
12
|
+
constructor();
|
|
13
|
+
}
|
|
@@ -0,0 +1,88 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __classPrivateFieldSet = (this && this.__classPrivateFieldSet) || function (receiver, state, value, kind, f) {
|
|
3
|
+
if (kind === "m") throw new TypeError("Private method is not writable");
|
|
4
|
+
if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a setter");
|
|
5
|
+
if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot write private member to an object whose class did not declare it");
|
|
6
|
+
return (kind === "a" ? f.call(receiver, value) : f ? f.value = value : state.set(receiver, value)), value;
|
|
7
|
+
};
|
|
8
|
+
var __classPrivateFieldGet = (this && this.__classPrivateFieldGet) || function (receiver, state, kind, f) {
|
|
9
|
+
if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a getter");
|
|
10
|
+
if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot read private member from an object whose class did not declare it");
|
|
11
|
+
return kind === "m" ? f : kind === "a" ? f.call(receiver) : f ? f.value : state.get(receiver);
|
|
12
|
+
};
|
|
13
|
+
var _SplitHashValidator_digests, _SplitHashValidator_blockSize, _SplitHashValidator_createHash, _SplitHashValidator_hash, _SplitHashValidator_equals, _SplitHashValidator_accu, _SplitHashValidator_digestIndex;
|
|
14
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
15
|
+
exports.NotMatchedError = exports.SplitHashValidator = void 0;
|
|
16
|
+
const stream_1 = require("stream");
|
|
17
|
+
class SplitHashValidator extends stream_1.Transform {
|
|
18
|
+
constructor(digests, blockSize, createHash, equals = Object.is) {
|
|
19
|
+
super();
|
|
20
|
+
_SplitHashValidator_digests.set(this, void 0);
|
|
21
|
+
_SplitHashValidator_blockSize.set(this, void 0);
|
|
22
|
+
_SplitHashValidator_createHash.set(this, void 0);
|
|
23
|
+
_SplitHashValidator_hash.set(this, void 0);
|
|
24
|
+
_SplitHashValidator_equals.set(this, void 0);
|
|
25
|
+
_SplitHashValidator_accu.set(this, 0);
|
|
26
|
+
_SplitHashValidator_digestIndex.set(this, 0);
|
|
27
|
+
__classPrivateFieldSet(this, _SplitHashValidator_digests, digests, "f");
|
|
28
|
+
__classPrivateFieldSet(this, _SplitHashValidator_blockSize, blockSize, "f");
|
|
29
|
+
__classPrivateFieldSet(this, _SplitHashValidator_createHash, createHash, "f");
|
|
30
|
+
__classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
|
|
31
|
+
__classPrivateFieldSet(this, _SplitHashValidator_equals, equals, "f");
|
|
32
|
+
}
|
|
33
|
+
_transform(chunk, encoding, callback) {
|
|
34
|
+
var _a;
|
|
35
|
+
if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length < __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f")) {
|
|
36
|
+
__classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(chunk);
|
|
37
|
+
__classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length, "f");
|
|
38
|
+
}
|
|
39
|
+
else {
|
|
40
|
+
let offset = 0;
|
|
41
|
+
while (true) {
|
|
42
|
+
const needed = __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f") - __classPrivateFieldGet(this, _SplitHashValidator_accu, "f");
|
|
43
|
+
const slice = chunk.slice(offset, offset + needed);
|
|
44
|
+
if (slice.length === needed) {
|
|
45
|
+
__classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
|
|
46
|
+
const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
|
|
47
|
+
if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
|
|
48
|
+
return callback(new NotMatchedError());
|
|
49
|
+
}
|
|
50
|
+
__classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
|
|
51
|
+
__classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
|
|
52
|
+
__classPrivateFieldSet(this, _SplitHashValidator_accu, 0, "f");
|
|
53
|
+
offset += slice.length;
|
|
54
|
+
}
|
|
55
|
+
else {
|
|
56
|
+
__classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
|
|
57
|
+
__classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + slice.length, "f");
|
|
58
|
+
break;
|
|
59
|
+
}
|
|
60
|
+
}
|
|
61
|
+
}
|
|
62
|
+
callback(null, chunk);
|
|
63
|
+
}
|
|
64
|
+
_flush(callback) {
|
|
65
|
+
var _a;
|
|
66
|
+
if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") > 0) {
|
|
67
|
+
const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
|
|
68
|
+
if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
|
|
69
|
+
return callback(new NotMatchedError());
|
|
70
|
+
}
|
|
71
|
+
__classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
|
|
72
|
+
}
|
|
73
|
+
if (__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f") !== __classPrivateFieldGet(this, _SplitHashValidator_digests, "f").length) {
|
|
74
|
+
return callback(new NotMatchedError());
|
|
75
|
+
}
|
|
76
|
+
callback();
|
|
77
|
+
}
|
|
78
|
+
}
|
|
79
|
+
exports.SplitHashValidator = SplitHashValidator;
|
|
80
|
+
_SplitHashValidator_digests = new WeakMap(), _SplitHashValidator_blockSize = new WeakMap(), _SplitHashValidator_createHash = new WeakMap(), _SplitHashValidator_hash = new WeakMap(), _SplitHashValidator_equals = new WeakMap(), _SplitHashValidator_accu = new WeakMap(), _SplitHashValidator_digestIndex = new WeakMap();
|
|
81
|
+
class NotMatchedError extends Error {
|
|
82
|
+
constructor() {
|
|
83
|
+
super('hashes do not match');
|
|
84
|
+
this.name = this.constructor.name;
|
|
85
|
+
}
|
|
86
|
+
}
|
|
87
|
+
exports.NotMatchedError = NotMatchedError;
|
|
88
|
+
//# sourceMappingURL=split-hash-validator.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"split-hash-validator.js","sourceRoot":"","sources":["../../src/split-hash-validator.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;AACA,mCAAqD;AAErD,MAAa,kBAAsB,SAAQ,kBAAS;IASlD,YACE,OAAY,EACZ,SAAiB,EACjB,UAAqC,EACrC,SAAkC,MAAM,CAAC,EAAE;QAE3C,KAAK,EAAE,CAAA;QAdT,8CAAa;QACb,gDAAkB;QAClB,iDAAsC;QACtC,2CAAyB;QACzB,6CAAgC;QAChC,mCAAQ,CAAC,EAAA;QACT,0CAAe,CAAC,EAAA;QAUd,uBAAA,IAAI,+BAAY,OAAO,MAAA,CAAA;QACvB,uBAAA,IAAI,iCAAc,SAAS,MAAA,CAAA;QAC3B,uBAAA,IAAI,kCAAe,UAAU,MAAA,CAAA;QAC7B,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;QAC/B,uBAAA,IAAI,8BAAW,MAAM,MAAA,CAAA;IACvB,CAAC;IAED,UAAU,CAAC,KAAa,EAAE,QAAwB,EAAE,QAA2B;;QAG7E,IAAI,uBAAA,IAAI,gCAAM,GAAG,KAAK,CAAC,MAAM,GAAG,uBAAA,IAAI,qCAAW,EAAE;YAC/C,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;YACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;SAC3B;aAAM;YACL,IAAI,MAAM,GAAG,CAAC,CAAA;YACd,OAAO,IAAI,EAAE;gBACX,MAAM,MAAM,GAAG,uBAAA,IAAI,qCAAW,GAAG,uBAAA,IAAI,gCAAM,CAAA;gBAC3C,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;gBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;oBAC3B,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;oBAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;wBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;qBACvC;oBACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;oBAEnB,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;oBAC/B,uBAAA,IAAI,4BAAS,CAAC,MAAA,CAAA;oBACd,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;iBACvB;qBAAM;oBAEL,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;oBAC1B,MAAK;iBACN;aACF;SACF;QAED,QAAQ,CAAC,IAAI,EAAE,KAAK,CAAC,CAAA;IACvB,CAAC;IAED,MAAM,CAAC,QAA2B;;QAChC,IAAI,uBAAA,IAAI,gCAAM,GAAG,CAAC,EAAE;YAClB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;YAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;gBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;aACvC;YACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;SACpB;QAED,IAAI,uBAAA,IAAI,uCAAa,KAAK,uBAAA,IAAI,mCAAS,CAAC,MAAM,EAAE;YAC9C,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;SACvC;QAED,QAAQ,EAAE,CAAA;IACZ,CAAC;CACF;AAzED,gDAyEC;;AAED,MAAa,eAAgB,SAAQ,KAAK;IAGxC;QACE,KAAK,CAAC,qBAAqB,CAAC,CAAA;QAH9B,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,0CAMC"}
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
/// <reference types="node" />
|
|
2
|
+
import { ProgressiveHashFactory } from './types';
|
|
3
|
+
export declare function splitHash<T>(stream: NodeJS.ReadableStream, blockSize: number, createHash: ProgressiveHashFactory<T>): AsyncIterable<T>;
|
|
4
|
+
export declare class StreamEncodingError extends Error {
|
|
5
|
+
name: string;
|
|
6
|
+
constructor();
|
|
7
|
+
}
|
|
@@ -0,0 +1,46 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.StreamEncodingError = exports.splitHash = void 0;
|
|
4
|
+
async function* splitHash(stream, blockSize, createHash) {
|
|
5
|
+
let hash = createHash();
|
|
6
|
+
let accu = 0;
|
|
7
|
+
for await (const chunk of stream) {
|
|
8
|
+
if (!Buffer.isBuffer(chunk))
|
|
9
|
+
throw new StreamEncodingError();
|
|
10
|
+
if (accu + chunk.length < blockSize) {
|
|
11
|
+
hash.update(chunk);
|
|
12
|
+
accu += chunk.length;
|
|
13
|
+
}
|
|
14
|
+
else {
|
|
15
|
+
let offset = 0;
|
|
16
|
+
while (true) {
|
|
17
|
+
const needed = blockSize - accu;
|
|
18
|
+
const slice = chunk.slice(offset, offset + needed);
|
|
19
|
+
if (slice.length === needed) {
|
|
20
|
+
hash.update(slice);
|
|
21
|
+
const digest = hash.digest();
|
|
22
|
+
yield digest;
|
|
23
|
+
hash = createHash();
|
|
24
|
+
accu = 0;
|
|
25
|
+
offset += slice.length;
|
|
26
|
+
}
|
|
27
|
+
else {
|
|
28
|
+
hash.update(slice);
|
|
29
|
+
accu += slice.length;
|
|
30
|
+
break;
|
|
31
|
+
}
|
|
32
|
+
}
|
|
33
|
+
}
|
|
34
|
+
}
|
|
35
|
+
if (accu > 0)
|
|
36
|
+
yield hash.digest();
|
|
37
|
+
}
|
|
38
|
+
exports.splitHash = splitHash;
|
|
39
|
+
class StreamEncodingError extends Error {
|
|
40
|
+
constructor() {
|
|
41
|
+
super('stream encoding must not be set.');
|
|
42
|
+
this.name = this.constructor.name;
|
|
43
|
+
}
|
|
44
|
+
}
|
|
45
|
+
exports.StreamEncodingError = StreamEncodingError;
|
|
46
|
+
//# sourceMappingURL=split-hash.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"split-hash.js","sourceRoot":"","sources":["../../src/split-hash.ts"],"names":[],"mappings":";;;AAEO,KAAK,SAAS,CAAC,CAAC,SAAS,CAAI,MAA6B,EAAE,SAAiB,EAAE,UAAqC;IACzH,IAAI,IAAI,GAAG,UAAU,EAAE,CAAA;IACvB,IAAI,IAAI,GAAG,CAAC,CAAA;IACZ,IAAI,KAAK,EAAE,MAAM,KAAK,IAAI,MAAM,EAAE;QAChC,IAAI,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC;YAAE,MAAM,IAAI,mBAAmB,EAAE,CAAA;QAC5D,IAAI,IAAI,GAAG,KAAK,CAAC,MAAM,GAAG,SAAS,EAAE;YACnC,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;YAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;SACrB;aAAM;YACL,IAAI,MAAM,GAAG,CAAC,CAAA;YACd,OAAO,IAAI,EAAE;gBACX,MAAM,MAAM,GAAG,SAAS,GAAG,IAAI,CAAA;gBAC/B,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;gBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;oBAC3B,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBAClB,MAAM,MAAM,GAAG,IAAI,CAAC,MAAM,EAAE,CAAA;oBAC5B,MAAM,MAAM,CAAA;oBAEZ,IAAI,GAAG,UAAU,EAAE,CAAA;oBACnB,IAAI,GAAG,CAAC,CAAA;oBACR,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;iBACvB;qBAAM;oBAEL,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;oBACpB,MAAK;iBACN;aACF;SACF;KACF;IAED,IAAI,IAAI,GAAG,CAAC;QAAE,MAAM,IAAI,CAAC,MAAM,EAAE,CAAA;AACnC,CAAC;AAhCD,8BAgCC;AAED,MAAa,mBAAoB,SAAQ,KAAK;IAG5C;QACE,KAAK,CAAC,kCAAkC,CAAC,CAAA;QAH3C,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,kDAMC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"types.js","sourceRoot":"","sources":["../../src/types.ts"],"names":[],"mappings":""}
|
package/package.json
ADDED
|
@@ -0,0 +1,67 @@
|
|
|
1
|
+
{
|
|
2
|
+
"name": "split-hash",
|
|
3
|
+
"version": "0.1.5",
|
|
4
|
+
"description": "Split the stream based on bytes and get digests from each part.",
|
|
5
|
+
"keywords": [
|
|
6
|
+
"split",
|
|
7
|
+
"hash",
|
|
8
|
+
"HashList",
|
|
9
|
+
"stream",
|
|
10
|
+
"block",
|
|
11
|
+
"chunk",
|
|
12
|
+
"slice"
|
|
13
|
+
],
|
|
14
|
+
"files": [
|
|
15
|
+
"lib"
|
|
16
|
+
],
|
|
17
|
+
"main": "lib/es2018/index.js",
|
|
18
|
+
"typings": "lib/es2018/index.d.ts",
|
|
19
|
+
"repository": "git@github.com:BlackGlory/split-hash.git",
|
|
20
|
+
"author": "BlackGlory <woshenmedoubuzhidao@blackglory.me>",
|
|
21
|
+
"license": "MIT",
|
|
22
|
+
"scripts": {
|
|
23
|
+
"lint": "eslint --ext .js,.jsx,.ts,.tsx --quiet src __tests__",
|
|
24
|
+
"test": "jest --config jest.config.js",
|
|
25
|
+
"test:debug": "node --inspect node_modules/.bin/jest --runInBand",
|
|
26
|
+
"test:coverage": "jest --coverage --config jest.config.js",
|
|
27
|
+
"prepublishOnly": "run-s clean build",
|
|
28
|
+
"clean": "run-p clean:*",
|
|
29
|
+
"clean:build": "rimraf lib",
|
|
30
|
+
"build": "run-p build:*",
|
|
31
|
+
"build:es2015": "run-s build:es2015:*",
|
|
32
|
+
"build:es2015:compile": "tsc --project tsconfig.build.json --module commonjs --target es2015 --outDir lib/es2015",
|
|
33
|
+
"build:es2015:patch": "tscpaths -p tsconfig.build.json -s ./src -o ./lib/es2015",
|
|
34
|
+
"build:es2018": "run-s build:es2018:*",
|
|
35
|
+
"build:es2018:compile": "tsc --project tsconfig.build.json --module commonjs --target es2018 --outDir lib/es2018",
|
|
36
|
+
"build:es2018:patch": "tscpaths -p tsconfig.build.json -s ./src -o ./lib/es2018",
|
|
37
|
+
"release": "standard-version"
|
|
38
|
+
},
|
|
39
|
+
"husky": {
|
|
40
|
+
"hooks": {
|
|
41
|
+
"pre-commit": "run-s lint test",
|
|
42
|
+
"commit-msg": "commitlint -E HUSKY_GIT_PARAMS"
|
|
43
|
+
}
|
|
44
|
+
},
|
|
45
|
+
"devDependencies": {
|
|
46
|
+
"@blackglory/go": "^0.1.0",
|
|
47
|
+
"@blackglory/jest-matchers": "^0.1.21",
|
|
48
|
+
"@blackglory/types": "^0.5.1",
|
|
49
|
+
"@commitlint/cli": "^13.2.1",
|
|
50
|
+
"@commitlint/config-conventional": "^13.2.0",
|
|
51
|
+
"@types/jest": "^27.0.2",
|
|
52
|
+
"@types/node": "^16.10.9",
|
|
53
|
+
"@typescript-eslint/eslint-plugin": "^5.0.0",
|
|
54
|
+
"@typescript-eslint/parser": "^5.0.0",
|
|
55
|
+
"eslint": "^8.0.1",
|
|
56
|
+
"husky": "^4.3.0",
|
|
57
|
+
"iterable-operator": "^0.14.4",
|
|
58
|
+
"jest": "^27.2.5",
|
|
59
|
+
"npm-run-all": "^4.1.5",
|
|
60
|
+
"return-style": "^0.12.4",
|
|
61
|
+
"rimraf": "^3.0.2",
|
|
62
|
+
"standard-version": "^9.3.1",
|
|
63
|
+
"ts-jest": "^27.0.5",
|
|
64
|
+
"tscpaths": "^0.0.9",
|
|
65
|
+
"typescript": "^4.4.4"
|
|
66
|
+
}
|
|
67
|
+
}
|