split-hash 0.1.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md ADDED
@@ -0,0 +1,11 @@
1
+ # Changelog
2
+
3
+ All notable changes to this project will be documented in this file. See [standard-version](https://github.com/conventional-changelog/standard-version) for commit guidelines.
4
+
5
+ ### [0.1.5](https://github.com/BlackGlory/split-hash/compare/v0.1.4...v0.1.5) (2021-10-14)
6
+
7
+ ### [0.1.4](https://github.com/BlackGlory/split-hash/compare/v0.1.3...v0.1.4) (2021-02-04)
8
+
9
+ ### [0.1.3](https://github.com/BlackGlory/split-hash/compare/v0.1.2...v0.1.3) (2021-02-03)
10
+
11
+ ### [0.1.2](https://github.com/BlackGlory/split-hash/compare/v0.1.1...v0.1.2) (2020-10-10)
package/LICENSE ADDED
@@ -0,0 +1,21 @@
1
+ MIT License
2
+
3
+ Copyright (c) 2020 BlackGlory <woshenmedoubuzhidao@blackglory.me>
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in all
13
+ copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ SOFTWARE.
package/README.md ADDED
@@ -0,0 +1,123 @@
1
+ # split-hash
2
+
3
+ Split the stream based on bytes and get digests from each part.
4
+
5
+ ## Install
6
+
7
+ ```sh
8
+ npm install --save split-hash
9
+ # or
10
+ yarn add split-hash
11
+ ```
12
+
13
+ ## Usage
14
+
15
+ ### Hash
16
+
17
+ ```js
18
+ import { splitHash } from 'split-hash'
19
+ import * as crypto from 'crypto'
20
+
21
+ const KiB = 1024
22
+
23
+ const createHash = () => {
24
+ const hash = crypto.createHash('sha256')
25
+ return {
26
+ update(buffer) {
27
+ hash.update(buffer)
28
+ }
29
+ , digest() {
30
+ return hash.digest('hex')
31
+ }
32
+ }
33
+ }
34
+
35
+ const stream = fs.createReadStream('filename.bin')
36
+ const iter = splitHash(stream, 512 * KiB, createHash)
37
+
38
+ for await (const hash of iter) {
39
+ console.log(hash)
40
+ }
41
+ ```
42
+
43
+ ### Validate
44
+
45
+ ```js
46
+ import { SplitHashValidator } from 'split-hash'
47
+ import * as crypto from 'crypto'
48
+
49
+ const KiB = 1024
50
+
51
+ const createHash = () => {
52
+ const hash = crypto.createHash('sha256')
53
+ return {
54
+ update(buffer) {
55
+ hash.update(buffer)
56
+ }
57
+ , digest() {
58
+ return hash.digest('hex')
59
+ }
60
+ }
61
+ }
62
+
63
+ const hashList = [/* ... */]
64
+ const validator = new SplitHashValidator(hashList, 512 * KiB, createHash)
65
+
66
+ const stream = fs.createReadStream('filename.bin')
67
+ stream
68
+ .pipe(validator)
69
+ .on('data', /* same as stream */)
70
+ .on('error', err => console.error('not matched'))
71
+ ```
72
+
73
+ ## Interface
74
+
75
+ ```ts
76
+ type ProgressiveHashFactory<T> = () => ProgressiveHash<T>
77
+
78
+ interface ProgressiveHash<T> {
79
+ update(buffer: Buffer): void
80
+ digest(): T
81
+ }
82
+ ```
83
+
84
+ ## API
85
+
86
+ ### splitHash
87
+
88
+ ```ts
89
+ function splitHash<T>(
90
+ stream: NodeJS.ReadableStream
91
+ , blockSize: number
92
+ , createHash: ProgressiveHashFactory<T>
93
+ ): AsyncIterable<T>
94
+ ```
95
+
96
+ It throws `StreamEncodingError` when the `stream` encoding is set.
97
+
98
+ ### SplitHashValidator
99
+
100
+ ```ts
101
+ class SplitHashValidator<T> extends Stream.Transform {
102
+ constructor(
103
+ digests: T[]
104
+ , blockSize: number
105
+ , createHash: ProgressiveHashFactory<T>
106
+ , equals: (a: T, b: T) => boolean = Object.is
107
+ )
108
+ }
109
+ ```
110
+
111
+ It throws `NotMatchedError` when the `stream` does not match digests.
112
+
113
+ ### StreamEncodingError
114
+
115
+ ```ts
116
+ class StreamEncodingError extends Error
117
+ ```
118
+
119
+ ### NotMatchedError
120
+
121
+ ```ts
122
+ class NotMatchedError extends Error
123
+ ```
@@ -0,0 +1,3 @@
1
+ export * from './types';
2
+ export * from './split-hash';
3
+ export * from './split-hash-validator';
@@ -0,0 +1,16 @@
1
+ "use strict";
2
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
3
+ if (k2 === undefined) k2 = k;
4
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
5
+ }) : (function(o, m, k, k2) {
6
+ if (k2 === undefined) k2 = k;
7
+ o[k2] = m[k];
8
+ }));
9
+ var __exportStar = (this && this.__exportStar) || function(m, exports) {
10
+ for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
11
+ };
12
+ Object.defineProperty(exports, "__esModule", { value: true });
13
+ __exportStar(require("./types"), exports);
14
+ __exportStar(require("./split-hash"), exports);
15
+ __exportStar(require("./split-hash-validator"), exports);
16
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;AAAA,6CAA0B;AAC1B,kDAA+B;AAC/B,4DAAyC"}
@@ -0,0 +1,13 @@
1
+ /// <reference types="node" />
2
+ import { ProgressiveHashFactory } from './types';
3
+ import { Transform, TransformCallback } from 'stream';
4
+ export declare class SplitHashValidator<T> extends Transform {
5
+ #private;
6
+ constructor(digests: T[], blockSize: number, createHash: ProgressiveHashFactory<T>, equals?: (a: T, b: T) => boolean);
7
+ _transform(chunk: Buffer, encoding: BufferEncoding, callback: TransformCallback): void;
8
+ _flush(callback: TransformCallback): void;
9
+ }
10
+ export declare class NotMatchedError extends Error {
11
+ name: string;
12
+ constructor();
13
+ }
@@ -0,0 +1,88 @@
1
+ "use strict";
2
+ var __classPrivateFieldSet = (this && this.__classPrivateFieldSet) || function (receiver, state, value, kind, f) {
3
+ if (kind === "m") throw new TypeError("Private method is not writable");
4
+ if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a setter");
5
+ if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot write private member to an object whose class did not declare it");
6
+ return (kind === "a" ? f.call(receiver, value) : f ? f.value = value : state.set(receiver, value)), value;
7
+ };
8
+ var __classPrivateFieldGet = (this && this.__classPrivateFieldGet) || function (receiver, state, kind, f) {
9
+ if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a getter");
10
+ if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot read private member from an object whose class did not declare it");
11
+ return kind === "m" ? f : kind === "a" ? f.call(receiver) : f ? f.value : state.get(receiver);
12
+ };
13
+ var _SplitHashValidator_digests, _SplitHashValidator_blockSize, _SplitHashValidator_createHash, _SplitHashValidator_hash, _SplitHashValidator_equals, _SplitHashValidator_accu, _SplitHashValidator_digestIndex;
14
+ Object.defineProperty(exports, "__esModule", { value: true });
15
+ exports.NotMatchedError = exports.SplitHashValidator = void 0;
16
+ const stream_1 = require("stream");
17
+ class SplitHashValidator extends stream_1.Transform {
18
+ constructor(digests, blockSize, createHash, equals = Object.is) {
19
+ super();
20
+ _SplitHashValidator_digests.set(this, void 0);
21
+ _SplitHashValidator_blockSize.set(this, void 0);
22
+ _SplitHashValidator_createHash.set(this, void 0);
23
+ _SplitHashValidator_hash.set(this, void 0);
24
+ _SplitHashValidator_equals.set(this, void 0);
25
+ _SplitHashValidator_accu.set(this, 0);
26
+ _SplitHashValidator_digestIndex.set(this, 0);
27
+ __classPrivateFieldSet(this, _SplitHashValidator_digests, digests, "f");
28
+ __classPrivateFieldSet(this, _SplitHashValidator_blockSize, blockSize, "f");
29
+ __classPrivateFieldSet(this, _SplitHashValidator_createHash, createHash, "f");
30
+ __classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
31
+ __classPrivateFieldSet(this, _SplitHashValidator_equals, equals, "f");
32
+ }
33
+ _transform(chunk, encoding, callback) {
34
+ var _a;
35
+ if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length < __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f")) {
36
+ __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(chunk);
37
+ __classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length, "f");
38
+ }
39
+ else {
40
+ let offset = 0;
41
+ while (true) {
42
+ const needed = __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f") - __classPrivateFieldGet(this, _SplitHashValidator_accu, "f");
43
+ const slice = chunk.slice(offset, offset + needed);
44
+ if (slice.length === needed) {
45
+ __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
46
+ const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
47
+ if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
48
+ return callback(new NotMatchedError());
49
+ }
50
+ __classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
51
+ __classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
52
+ __classPrivateFieldSet(this, _SplitHashValidator_accu, 0, "f");
53
+ offset += slice.length;
54
+ }
55
+ else {
56
+ __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
57
+ __classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + slice.length, "f");
58
+ break;
59
+ }
60
+ }
61
+ }
62
+ callback(null, chunk);
63
+ }
64
+ _flush(callback) {
65
+ var _a;
66
+ if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") > 0) {
67
+ const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
68
+ if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
69
+ return callback(new NotMatchedError());
70
+ }
71
+ __classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
72
+ }
73
+ if (__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f") !== __classPrivateFieldGet(this, _SplitHashValidator_digests, "f").length) {
74
+ return callback(new NotMatchedError());
75
+ }
76
+ callback();
77
+ }
78
+ }
79
+ exports.SplitHashValidator = SplitHashValidator;
80
+ _SplitHashValidator_digests = new WeakMap(), _SplitHashValidator_blockSize = new WeakMap(), _SplitHashValidator_createHash = new WeakMap(), _SplitHashValidator_hash = new WeakMap(), _SplitHashValidator_equals = new WeakMap(), _SplitHashValidator_accu = new WeakMap(), _SplitHashValidator_digestIndex = new WeakMap();
81
+ class NotMatchedError extends Error {
82
+ constructor() {
83
+ super('hashes do not match');
84
+ this.name = this.constructor.name;
85
+ }
86
+ }
87
+ exports.NotMatchedError = NotMatchedError;
88
+ //# sourceMappingURL=split-hash-validator.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"split-hash-validator.js","sourceRoot":"","sources":["../../src/split-hash-validator.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;AACA,mCAAqD;AAErD,MAAa,kBAAsB,SAAQ,kBAAS;IASlD,YACE,OAAY,EACZ,SAAiB,EACjB,UAAqC,EACrC,SAAkC,MAAM,CAAC,EAAE;QAE3C,KAAK,EAAE,CAAA;QAdT,8CAAa;QACb,gDAAkB;QAClB,iDAAsC;QACtC,2CAAyB;QACzB,6CAAgC;QAChC,mCAAQ,CAAC,EAAA;QACT,0CAAe,CAAC,EAAA;QAUd,uBAAA,IAAI,+BAAY,OAAO,MAAA,CAAA;QACvB,uBAAA,IAAI,iCAAc,SAAS,MAAA,CAAA;QAC3B,uBAAA,IAAI,kCAAe,UAAU,MAAA,CAAA;QAC7B,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;QAC/B,uBAAA,IAAI,8BAAW,MAAM,MAAA,CAAA;IACvB,CAAC;IAED,UAAU,CAAC,KAAa,EAAE,QAAwB,EAAE,QAA2B;;QAG7E,IAAI,uBAAA,IAAI,gCAAM,GAAG,KAAK,CAAC,MAAM,GAAG,uBAAA,IAAI,qCAAW,EAAE;YAC/C,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;YACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;SAC3B;aAAM;YACL,IAAI,MAAM,GAAG,CAAC,CAAA;YACd,OAAO,IAAI,EAAE;gBACX,MAAM,MAAM,GAAG,uBAAA,IAAI,qCAAW,GAAG,uBAAA,IAAI,gCAAM,CAAA;gBAC3C,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;gBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;oBAC3B,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;oBAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;wBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;qBACvC;oBACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;oBAEnB,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;oBAC/B,uBAAA,IAAI,4BAAS,CAAC,MAAA,CAAA;oBACd,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;iBACvB;qBAAM;oBAEL,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;oBAC1B,MAAK;iBACN;aACF;SACF;QAED,QAAQ,CAAC,IAAI,EAAE,KAAK,CAAC,CAAA;IACvB,CAAC;IAED,MAAM,CAAC,QAA2B;;QAChC,IAAI,uBAAA,IAAI,gCAAM,GAAG,CAAC,EAAE;YAClB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;YAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;gBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;aACvC;YACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;SACpB;QAED,IAAI,uBAAA,IAAI,uCAAa,KAAK,uBAAA,IAAI,mCAAS,CAAC,MAAM,EAAE;YAC9C,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;SACvC;QAED,QAAQ,EAAE,CAAA;IACZ,CAAC;CACF;AAzED,gDAyEC;;AAED,MAAa,eAAgB,SAAQ,KAAK;IAGxC;QACE,KAAK,CAAC,qBAAqB,CAAC,CAAA;QAH9B,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,0CAMC"}
@@ -0,0 +1,7 @@
1
+ /// <reference types="node" />
2
+ import { ProgressiveHashFactory } from './types';
3
+ export declare function splitHash<T>(stream: NodeJS.ReadableStream, blockSize: number, createHash: ProgressiveHashFactory<T>): AsyncIterable<T>;
4
+ export declare class StreamEncodingError extends Error {
5
+ name: string;
6
+ constructor();
7
+ }
@@ -0,0 +1,78 @@
1
+ "use strict";
2
+ var __asyncValues = (this && this.__asyncValues) || function (o) {
3
+ if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
4
+ var m = o[Symbol.asyncIterator], i;
5
+ return m ? m.call(o) : (o = typeof __values === "function" ? __values(o) : o[Symbol.iterator](), i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i);
6
+ function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }
7
+ function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }
8
+ };
9
+ var __await = (this && this.__await) || function (v) { return this instanceof __await ? (this.v = v, this) : new __await(v); }
10
+ var __asyncGenerator = (this && this.__asyncGenerator) || function (thisArg, _arguments, generator) {
11
+ if (!Symbol.asyncIterator) throw new TypeError("Symbol.asyncIterator is not defined.");
12
+ var g = generator.apply(thisArg, _arguments || []), i, q = [];
13
+ return i = {}, verb("next"), verb("throw"), verb("return"), i[Symbol.asyncIterator] = function () { return this; }, i;
14
+ function verb(n) { if (g[n]) i[n] = function (v) { return new Promise(function (a, b) { q.push([n, v, a, b]) > 1 || resume(n, v); }); }; }
15
+ function resume(n, v) { try { step(g[n](v)); } catch (e) { settle(q[0][3], e); } }
16
+ function step(r) { r.value instanceof __await ? Promise.resolve(r.value.v).then(fulfill, reject) : settle(q[0][2], r); }
17
+ function fulfill(value) { resume("next", value); }
18
+ function reject(value) { resume("throw", value); }
19
+ function settle(f, v) { if (f(v), q.shift(), q.length) resume(q[0][0], q[0][1]); }
20
+ };
21
+ Object.defineProperty(exports, "__esModule", { value: true });
22
+ exports.StreamEncodingError = exports.splitHash = void 0;
23
+ function splitHash(stream, blockSize, createHash) {
24
+ return __asyncGenerator(this, arguments, function* splitHash_1() {
25
+ var e_1, _a;
26
+ let hash = createHash();
27
+ let accu = 0;
28
+ try {
29
+ for (var stream_1 = __asyncValues(stream), stream_1_1; stream_1_1 = yield __await(stream_1.next()), !stream_1_1.done;) {
30
+ const chunk = stream_1_1.value;
31
+ if (!Buffer.isBuffer(chunk))
32
+ throw new StreamEncodingError();
33
+ if (accu + chunk.length < blockSize) {
34
+ hash.update(chunk);
35
+ accu += chunk.length;
36
+ }
37
+ else {
38
+ let offset = 0;
39
+ while (true) {
40
+ const needed = blockSize - accu;
41
+ const slice = chunk.slice(offset, offset + needed);
42
+ if (slice.length === needed) {
43
+ hash.update(slice);
44
+ const digest = hash.digest();
45
+ yield yield __await(digest);
46
+ hash = createHash();
47
+ accu = 0;
48
+ offset += slice.length;
49
+ }
50
+ else {
51
+ hash.update(slice);
52
+ accu += slice.length;
53
+ break;
54
+ }
55
+ }
56
+ }
57
+ }
58
+ }
59
+ catch (e_1_1) { e_1 = { error: e_1_1 }; }
60
+ finally {
61
+ try {
62
+ if (stream_1_1 && !stream_1_1.done && (_a = stream_1.return)) yield __await(_a.call(stream_1));
63
+ }
64
+ finally { if (e_1) throw e_1.error; }
65
+ }
66
+ if (accu > 0)
67
+ yield yield __await(hash.digest());
68
+ });
69
+ }
70
+ exports.splitHash = splitHash;
71
+ class StreamEncodingError extends Error {
72
+ constructor() {
73
+ super('stream encoding must not be set.');
74
+ this.name = this.constructor.name;
75
+ }
76
+ }
77
+ exports.StreamEncodingError = StreamEncodingError;
78
+ //# sourceMappingURL=split-hash.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"split-hash.js","sourceRoot":"","sources":["../../src/split-hash.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;AAEA,SAAuB,SAAS,CAAI,MAA6B,EAAE,SAAiB,EAAE,UAAqC;;;QACzH,IAAI,IAAI,GAAG,UAAU,EAAE,CAAA;QACvB,IAAI,IAAI,GAAG,CAAC,CAAA;;YACZ,KAA0B,IAAA,WAAA,cAAA,MAAM,CAAA,YAAA;gBAArB,MAAM,KAAK,mBAAA,CAAA;gBACpB,IAAI,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC;oBAAE,MAAM,IAAI,mBAAmB,EAAE,CAAA;gBAC5D,IAAI,IAAI,GAAG,KAAK,CAAC,MAAM,GAAG,SAAS,EAAE;oBACnC,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;iBACrB;qBAAM;oBACL,IAAI,MAAM,GAAG,CAAC,CAAA;oBACd,OAAO,IAAI,EAAE;wBACX,MAAM,MAAM,GAAG,SAAS,GAAG,IAAI,CAAA;wBAC/B,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;wBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;4BAC3B,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;4BAClB,MAAM,MAAM,GAAG,IAAI,CAAC,MAAM,EAAE,CAAA;4BAC5B,oBAAM,MAAM,CAAA,CAAA;4BAEZ,IAAI,GAAG,UAAU,EAAE,CAAA;4BACnB,IAAI,GAAG,CAAC,CAAA;4BACR,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;yBACvB;6BAAM;4BAEL,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;4BAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;4BACpB,MAAK;yBACN;qBACF;iBACF;aACF;;;;;;;;;QAED,IAAI,IAAI,GAAG,CAAC;YAAE,oBAAM,IAAI,CAAC,MAAM,EAAE,CAAA,CAAA;IACnC,CAAC;CAAA;AAhCD,8BAgCC;AAED,MAAa,mBAAoB,SAAQ,KAAK;IAG5C;QACE,KAAK,CAAC,kCAAkC,CAAC,CAAA;QAH3C,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,kDAMC"}
@@ -0,0 +1,6 @@
1
+ /// <reference types="node" />
2
+ export declare type ProgressiveHashFactory<T> = () => ProgressiveHash<T>;
3
+ export interface ProgressiveHash<T> {
4
+ update(buffer: Buffer): void;
5
+ digest(): T;
6
+ }
@@ -0,0 +1,3 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ //# sourceMappingURL=types.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"types.js","sourceRoot":"","sources":["../../src/types.ts"],"names":[],"mappings":""}
@@ -0,0 +1,3 @@
1
+ export * from './types';
2
+ export * from './split-hash';
3
+ export * from './split-hash-validator';
@@ -0,0 +1,16 @@
1
+ "use strict";
2
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
3
+ if (k2 === undefined) k2 = k;
4
+ Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
5
+ }) : (function(o, m, k, k2) {
6
+ if (k2 === undefined) k2 = k;
7
+ o[k2] = m[k];
8
+ }));
9
+ var __exportStar = (this && this.__exportStar) || function(m, exports) {
10
+ for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
11
+ };
12
+ Object.defineProperty(exports, "__esModule", { value: true });
13
+ __exportStar(require("./types"), exports);
14
+ __exportStar(require("./split-hash"), exports);
15
+ __exportStar(require("./split-hash-validator"), exports);
16
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;AAAA,6CAA0B;AAC1B,kDAA+B;AAC/B,4DAAyC"}
@@ -0,0 +1,13 @@
1
+ /// <reference types="node" />
2
+ import { ProgressiveHashFactory } from './types';
3
+ import { Transform, TransformCallback } from 'stream';
4
+ export declare class SplitHashValidator<T> extends Transform {
5
+ #private;
6
+ constructor(digests: T[], blockSize: number, createHash: ProgressiveHashFactory<T>, equals?: (a: T, b: T) => boolean);
7
+ _transform(chunk: Buffer, encoding: BufferEncoding, callback: TransformCallback): void;
8
+ _flush(callback: TransformCallback): void;
9
+ }
10
+ export declare class NotMatchedError extends Error {
11
+ name: string;
12
+ constructor();
13
+ }
@@ -0,0 +1,88 @@
1
+ "use strict";
2
+ var __classPrivateFieldSet = (this && this.__classPrivateFieldSet) || function (receiver, state, value, kind, f) {
3
+ if (kind === "m") throw new TypeError("Private method is not writable");
4
+ if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a setter");
5
+ if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot write private member to an object whose class did not declare it");
6
+ return (kind === "a" ? f.call(receiver, value) : f ? f.value = value : state.set(receiver, value)), value;
7
+ };
8
+ var __classPrivateFieldGet = (this && this.__classPrivateFieldGet) || function (receiver, state, kind, f) {
9
+ if (kind === "a" && !f) throw new TypeError("Private accessor was defined without a getter");
10
+ if (typeof state === "function" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError("Cannot read private member from an object whose class did not declare it");
11
+ return kind === "m" ? f : kind === "a" ? f.call(receiver) : f ? f.value : state.get(receiver);
12
+ };
13
+ var _SplitHashValidator_digests, _SplitHashValidator_blockSize, _SplitHashValidator_createHash, _SplitHashValidator_hash, _SplitHashValidator_equals, _SplitHashValidator_accu, _SplitHashValidator_digestIndex;
14
+ Object.defineProperty(exports, "__esModule", { value: true });
15
+ exports.NotMatchedError = exports.SplitHashValidator = void 0;
16
+ const stream_1 = require("stream");
17
+ class SplitHashValidator extends stream_1.Transform {
18
+ constructor(digests, blockSize, createHash, equals = Object.is) {
19
+ super();
20
+ _SplitHashValidator_digests.set(this, void 0);
21
+ _SplitHashValidator_blockSize.set(this, void 0);
22
+ _SplitHashValidator_createHash.set(this, void 0);
23
+ _SplitHashValidator_hash.set(this, void 0);
24
+ _SplitHashValidator_equals.set(this, void 0);
25
+ _SplitHashValidator_accu.set(this, 0);
26
+ _SplitHashValidator_digestIndex.set(this, 0);
27
+ __classPrivateFieldSet(this, _SplitHashValidator_digests, digests, "f");
28
+ __classPrivateFieldSet(this, _SplitHashValidator_blockSize, blockSize, "f");
29
+ __classPrivateFieldSet(this, _SplitHashValidator_createHash, createHash, "f");
30
+ __classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
31
+ __classPrivateFieldSet(this, _SplitHashValidator_equals, equals, "f");
32
+ }
33
+ _transform(chunk, encoding, callback) {
34
+ var _a;
35
+ if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length < __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f")) {
36
+ __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(chunk);
37
+ __classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + chunk.length, "f");
38
+ }
39
+ else {
40
+ let offset = 0;
41
+ while (true) {
42
+ const needed = __classPrivateFieldGet(this, _SplitHashValidator_blockSize, "f") - __classPrivateFieldGet(this, _SplitHashValidator_accu, "f");
43
+ const slice = chunk.slice(offset, offset + needed);
44
+ if (slice.length === needed) {
45
+ __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
46
+ const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
47
+ if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
48
+ return callback(new NotMatchedError());
49
+ }
50
+ __classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
51
+ __classPrivateFieldSet(this, _SplitHashValidator_hash, __classPrivateFieldGet(this, _SplitHashValidator_createHash, "f").call(this), "f");
52
+ __classPrivateFieldSet(this, _SplitHashValidator_accu, 0, "f");
53
+ offset += slice.length;
54
+ }
55
+ else {
56
+ __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").update(slice);
57
+ __classPrivateFieldSet(this, _SplitHashValidator_accu, __classPrivateFieldGet(this, _SplitHashValidator_accu, "f") + slice.length, "f");
58
+ break;
59
+ }
60
+ }
61
+ }
62
+ callback(null, chunk);
63
+ }
64
+ _flush(callback) {
65
+ var _a;
66
+ if (__classPrivateFieldGet(this, _SplitHashValidator_accu, "f") > 0) {
67
+ const digest = __classPrivateFieldGet(this, _SplitHashValidator_hash, "f").digest();
68
+ if (!__classPrivateFieldGet(this, _SplitHashValidator_equals, "f").call(this, __classPrivateFieldGet(this, _SplitHashValidator_digests, "f")[__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f")], digest)) {
69
+ return callback(new NotMatchedError());
70
+ }
71
+ __classPrivateFieldSet(this, _SplitHashValidator_digestIndex, (_a = __classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f"), _a++, _a), "f");
72
+ }
73
+ if (__classPrivateFieldGet(this, _SplitHashValidator_digestIndex, "f") !== __classPrivateFieldGet(this, _SplitHashValidator_digests, "f").length) {
74
+ return callback(new NotMatchedError());
75
+ }
76
+ callback();
77
+ }
78
+ }
79
+ exports.SplitHashValidator = SplitHashValidator;
80
+ _SplitHashValidator_digests = new WeakMap(), _SplitHashValidator_blockSize = new WeakMap(), _SplitHashValidator_createHash = new WeakMap(), _SplitHashValidator_hash = new WeakMap(), _SplitHashValidator_equals = new WeakMap(), _SplitHashValidator_accu = new WeakMap(), _SplitHashValidator_digestIndex = new WeakMap();
81
+ class NotMatchedError extends Error {
82
+ constructor() {
83
+ super('hashes do not match');
84
+ this.name = this.constructor.name;
85
+ }
86
+ }
87
+ exports.NotMatchedError = NotMatchedError;
88
+ //# sourceMappingURL=split-hash-validator.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"split-hash-validator.js","sourceRoot":"","sources":["../../src/split-hash-validator.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;AACA,mCAAqD;AAErD,MAAa,kBAAsB,SAAQ,kBAAS;IASlD,YACE,OAAY,EACZ,SAAiB,EACjB,UAAqC,EACrC,SAAkC,MAAM,CAAC,EAAE;QAE3C,KAAK,EAAE,CAAA;QAdT,8CAAa;QACb,gDAAkB;QAClB,iDAAsC;QACtC,2CAAyB;QACzB,6CAAgC;QAChC,mCAAQ,CAAC,EAAA;QACT,0CAAe,CAAC,EAAA;QAUd,uBAAA,IAAI,+BAAY,OAAO,MAAA,CAAA;QACvB,uBAAA,IAAI,iCAAc,SAAS,MAAA,CAAA;QAC3B,uBAAA,IAAI,kCAAe,UAAU,MAAA,CAAA;QAC7B,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;QAC/B,uBAAA,IAAI,8BAAW,MAAM,MAAA,CAAA;IACvB,CAAC;IAED,UAAU,CAAC,KAAa,EAAE,QAAwB,EAAE,QAA2B;;QAG7E,IAAI,uBAAA,IAAI,gCAAM,GAAG,KAAK,CAAC,MAAM,GAAG,uBAAA,IAAI,qCAAW,EAAE;YAC/C,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;YACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;SAC3B;aAAM;YACL,IAAI,MAAM,GAAG,CAAC,CAAA;YACd,OAAO,IAAI,EAAE;gBACX,MAAM,MAAM,GAAG,uBAAA,IAAI,qCAAW,GAAG,uBAAA,IAAI,gCAAM,CAAA;gBAC3C,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;gBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;oBAC3B,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;oBAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;wBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;qBACvC;oBACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;oBAEnB,uBAAA,IAAI,4BAAS,uBAAA,IAAI,sCAAY,MAAhB,IAAI,CAAc,MAAA,CAAA;oBAC/B,uBAAA,IAAI,4BAAS,CAAC,MAAA,CAAA;oBACd,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;iBACvB;qBAAM;oBAEL,uBAAA,IAAI,gCAAM,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBACxB,qHAAc,KAAK,CAAC,MAAM,MAAA,CAAA;oBAC1B,MAAK;iBACN;aACF;SACF;QAED,QAAQ,CAAC,IAAI,EAAE,KAAK,CAAC,CAAA;IACvB,CAAC;IAED,MAAM,CAAC,QAA2B;;QAChC,IAAI,uBAAA,IAAI,gCAAM,GAAG,CAAC,EAAE;YAClB,MAAM,MAAM,GAAG,uBAAA,IAAI,gCAAM,CAAC,MAAM,EAAE,CAAA;YAClC,IAAI,CAAC,uBAAA,IAAI,kCAAQ,MAAZ,IAAI,EAAS,uBAAA,IAAI,mCAAS,CAAC,uBAAA,IAAI,uCAAa,CAAC,EAAE,MAAM,CAAC,EAAE;gBAC3D,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;aACvC;YACD,8DAAA,CAAA,uEAAiB,EAAjB,IAAmB,IAAA,CAAA,MAAA,CAAA;SACpB;QAED,IAAI,uBAAA,IAAI,uCAAa,KAAK,uBAAA,IAAI,mCAAS,CAAC,MAAM,EAAE;YAC9C,OAAO,QAAQ,CAAC,IAAI,eAAe,EAAE,CAAC,CAAA;SACvC;QAED,QAAQ,EAAE,CAAA;IACZ,CAAC;CACF;AAzED,gDAyEC;;AAED,MAAa,eAAgB,SAAQ,KAAK;IAGxC;QACE,KAAK,CAAC,qBAAqB,CAAC,CAAA;QAH9B,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,0CAMC"}
@@ -0,0 +1,7 @@
1
+ /// <reference types="node" />
2
+ import { ProgressiveHashFactory } from './types';
3
+ export declare function splitHash<T>(stream: NodeJS.ReadableStream, blockSize: number, createHash: ProgressiveHashFactory<T>): AsyncIterable<T>;
4
+ export declare class StreamEncodingError extends Error {
5
+ name: string;
6
+ constructor();
7
+ }
@@ -0,0 +1,46 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.StreamEncodingError = exports.splitHash = void 0;
4
+ async function* splitHash(stream, blockSize, createHash) {
5
+ let hash = createHash();
6
+ let accu = 0;
7
+ for await (const chunk of stream) {
8
+ if (!Buffer.isBuffer(chunk))
9
+ throw new StreamEncodingError();
10
+ if (accu + chunk.length < blockSize) {
11
+ hash.update(chunk);
12
+ accu += chunk.length;
13
+ }
14
+ else {
15
+ let offset = 0;
16
+ while (true) {
17
+ const needed = blockSize - accu;
18
+ const slice = chunk.slice(offset, offset + needed);
19
+ if (slice.length === needed) {
20
+ hash.update(slice);
21
+ const digest = hash.digest();
22
+ yield digest;
23
+ hash = createHash();
24
+ accu = 0;
25
+ offset += slice.length;
26
+ }
27
+ else {
28
+ hash.update(slice);
29
+ accu += slice.length;
30
+ break;
31
+ }
32
+ }
33
+ }
34
+ }
35
+ if (accu > 0)
36
+ yield hash.digest();
37
+ }
38
+ exports.splitHash = splitHash;
39
+ class StreamEncodingError extends Error {
40
+ constructor() {
41
+ super('stream encoding must not be set.');
42
+ this.name = this.constructor.name;
43
+ }
44
+ }
45
+ exports.StreamEncodingError = StreamEncodingError;
46
+ //# sourceMappingURL=split-hash.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"split-hash.js","sourceRoot":"","sources":["../../src/split-hash.ts"],"names":[],"mappings":";;;AAEO,KAAK,SAAS,CAAC,CAAC,SAAS,CAAI,MAA6B,EAAE,SAAiB,EAAE,UAAqC;IACzH,IAAI,IAAI,GAAG,UAAU,EAAE,CAAA;IACvB,IAAI,IAAI,GAAG,CAAC,CAAA;IACZ,IAAI,KAAK,EAAE,MAAM,KAAK,IAAI,MAAM,EAAE;QAChC,IAAI,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC;YAAE,MAAM,IAAI,mBAAmB,EAAE,CAAA;QAC5D,IAAI,IAAI,GAAG,KAAK,CAAC,MAAM,GAAG,SAAS,EAAE;YACnC,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;YAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;SACrB;aAAM;YACL,IAAI,MAAM,GAAG,CAAC,CAAA;YACd,OAAO,IAAI,EAAE;gBACX,MAAM,MAAM,GAAG,SAAS,GAAG,IAAI,CAAA;gBAC/B,MAAM,KAAK,GAAG,KAAK,CAAC,KAAK,CAAC,MAAM,EAAE,MAAM,GAAG,MAAM,CAAC,CAAA;gBAClD,IAAI,KAAK,CAAC,MAAM,KAAK,MAAM,EAAE;oBAC3B,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBAClB,MAAM,MAAM,GAAG,IAAI,CAAC,MAAM,EAAE,CAAA;oBAC5B,MAAM,MAAM,CAAA;oBAEZ,IAAI,GAAG,UAAU,EAAE,CAAA;oBACnB,IAAI,GAAG,CAAC,CAAA;oBACR,MAAM,IAAI,KAAK,CAAC,MAAM,CAAA;iBACvB;qBAAM;oBAEL,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,CAAA;oBAClB,IAAI,IAAI,KAAK,CAAC,MAAM,CAAA;oBACpB,MAAK;iBACN;aACF;SACF;KACF;IAED,IAAI,IAAI,GAAG,CAAC;QAAE,MAAM,IAAI,CAAC,MAAM,EAAE,CAAA;AACnC,CAAC;AAhCD,8BAgCC;AAED,MAAa,mBAAoB,SAAQ,KAAK;IAG5C;QACE,KAAK,CAAC,kCAAkC,CAAC,CAAA;QAH3C,SAAI,GAAG,IAAI,CAAC,WAAW,CAAC,IAAI,CAAA;IAI5B,CAAC;CACF;AAND,kDAMC"}
@@ -0,0 +1,6 @@
1
+ /// <reference types="node" />
2
+ export declare type ProgressiveHashFactory<T> = () => ProgressiveHash<T>;
3
+ export interface ProgressiveHash<T> {
4
+ update(buffer: Buffer): void;
5
+ digest(): T;
6
+ }
@@ -0,0 +1,3 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ //# sourceMappingURL=types.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"types.js","sourceRoot":"","sources":["../../src/types.ts"],"names":[],"mappings":""}
package/package.json ADDED
@@ -0,0 +1,67 @@
1
+ {
2
+ "name": "split-hash",
3
+ "version": "0.1.5",
4
+ "description": "Split the stream based on bytes and get digests from each part.",
5
+ "keywords": [
6
+ "split",
7
+ "hash",
8
+ "HashList",
9
+ "stream",
10
+ "block",
11
+ "chunk",
12
+ "slice"
13
+ ],
14
+ "files": [
15
+ "lib"
16
+ ],
17
+ "main": "lib/es2018/index.js",
18
+ "typings": "lib/es2018/index.d.ts",
19
+ "repository": "git@github.com:BlackGlory/split-hash.git",
20
+ "author": "BlackGlory <woshenmedoubuzhidao@blackglory.me>",
21
+ "license": "MIT",
22
+ "scripts": {
23
+ "lint": "eslint --ext .js,.jsx,.ts,.tsx --quiet src __tests__",
24
+ "test": "jest --config jest.config.js",
25
+ "test:debug": "node --inspect node_modules/.bin/jest --runInBand",
26
+ "test:coverage": "jest --coverage --config jest.config.js",
27
+ "prepublishOnly": "run-s clean build",
28
+ "clean": "run-p clean:*",
29
+ "clean:build": "rimraf lib",
30
+ "build": "run-p build:*",
31
+ "build:es2015": "run-s build:es2015:*",
32
+ "build:es2015:compile": "tsc --project tsconfig.build.json --module commonjs --target es2015 --outDir lib/es2015",
33
+ "build:es2015:patch": "tscpaths -p tsconfig.build.json -s ./src -o ./lib/es2015",
34
+ "build:es2018": "run-s build:es2018:*",
35
+ "build:es2018:compile": "tsc --project tsconfig.build.json --module commonjs --target es2018 --outDir lib/es2018",
36
+ "build:es2018:patch": "tscpaths -p tsconfig.build.json -s ./src -o ./lib/es2018",
37
+ "release": "standard-version"
38
+ },
39
+ "husky": {
40
+ "hooks": {
41
+ "pre-commit": "run-s lint test",
42
+ "commit-msg": "commitlint -E HUSKY_GIT_PARAMS"
43
+ }
44
+ },
45
+ "devDependencies": {
46
+ "@blackglory/go": "^0.1.0",
47
+ "@blackglory/jest-matchers": "^0.1.21",
48
+ "@blackglory/types": "^0.5.1",
49
+ "@commitlint/cli": "^13.2.1",
50
+ "@commitlint/config-conventional": "^13.2.0",
51
+ "@types/jest": "^27.0.2",
52
+ "@types/node": "^16.10.9",
53
+ "@typescript-eslint/eslint-plugin": "^5.0.0",
54
+ "@typescript-eslint/parser": "^5.0.0",
55
+ "eslint": "^8.0.1",
56
+ "husky": "^4.3.0",
57
+ "iterable-operator": "^0.14.4",
58
+ "jest": "^27.2.5",
59
+ "npm-run-all": "^4.1.5",
60
+ "return-style": "^0.12.4",
61
+ "rimraf": "^3.0.2",
62
+ "standard-version": "^9.3.1",
63
+ "ts-jest": "^27.0.5",
64
+ "tscpaths": "^0.0.9",
65
+ "typescript": "^4.4.4"
66
+ }
67
+ }