@loaders.gl/shapefile 3.4.0-alpha.2 → 3.4.0-alpha.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/dbf-worker.js +1 -1
- package/dist/dist.min.js +47 -0
- package/dist/es5/dbf-loader.js +7 -11
- package/dist/es5/dbf-loader.js.map +1 -1
- package/dist/es5/index.js.map +1 -1
- package/dist/es5/lib/parsers/parse-dbf.js +47 -75
- package/dist/es5/lib/parsers/parse-dbf.js.map +1 -1
- package/dist/es5/lib/parsers/parse-shapefile.js +239 -249
- package/dist/es5/lib/parsers/parse-shapefile.js.map +1 -1
- package/dist/es5/lib/parsers/parse-shp-geometry.js +0 -16
- package/dist/es5/lib/parsers/parse-shp-geometry.js.map +1 -1
- package/dist/es5/lib/parsers/parse-shp-header.js +0 -1
- package/dist/es5/lib/parsers/parse-shp-header.js.map +1 -1
- package/dist/es5/lib/parsers/parse-shp.js +74 -79
- package/dist/es5/lib/parsers/parse-shp.js.map +1 -1
- package/dist/es5/lib/parsers/parse-shx.js +0 -1
- package/dist/es5/lib/parsers/parse-shx.js.map +1 -1
- package/dist/es5/lib/streaming/binary-chunk-reader.js +9 -33
- package/dist/es5/lib/streaming/binary-chunk-reader.js.map +1 -1
- package/dist/es5/lib/streaming/binary-reader.js +4 -11
- package/dist/es5/lib/streaming/binary-reader.js.map +1 -1
- package/dist/es5/lib/streaming/zip-batch-iterators.js +54 -58
- package/dist/es5/lib/streaming/zip-batch-iterators.js.map +1 -1
- package/dist/es5/shapefile-loader.js +1 -2
- package/dist/es5/shapefile-loader.js.map +1 -1
- package/dist/es5/shp-loader.js +7 -11
- package/dist/es5/shp-loader.js.map +1 -1
- package/dist/es5/workers/dbf-worker.js.map +1 -1
- package/dist/es5/workers/shp-worker.js.map +1 -1
- package/dist/esm/bundle.js +0 -1
- package/dist/esm/bundle.js.map +1 -1
- package/dist/esm/dbf-loader.js +1 -4
- package/dist/esm/dbf-loader.js.map +1 -1
- package/dist/esm/lib/parsers/parse-dbf.js +34 -45
- package/dist/esm/lib/parsers/parse-dbf.js.map +1 -1
- package/dist/esm/lib/parsers/parse-shapefile.js +0 -16
- package/dist/esm/lib/parsers/parse-shapefile.js.map +1 -1
- package/dist/esm/lib/parsers/parse-shp-geometry.js +0 -18
- package/dist/esm/lib/parsers/parse-shp-geometry.js.map +1 -1
- package/dist/esm/lib/parsers/parse-shp-header.js +0 -1
- package/dist/esm/lib/parsers/parse-shp-header.js.map +1 -1
- package/dist/esm/lib/parsers/parse-shp.js +0 -4
- package/dist/esm/lib/parsers/parse-shp.js.map +1 -1
- package/dist/esm/lib/parsers/parse-shx.js +0 -1
- package/dist/esm/lib/parsers/parse-shx.js.map +1 -1
- package/dist/esm/lib/streaming/binary-chunk-reader.js +0 -16
- package/dist/esm/lib/streaming/binary-chunk-reader.js.map +1 -1
- package/dist/esm/lib/streaming/binary-reader.js +0 -3
- package/dist/esm/lib/streaming/binary-reader.js.map +1 -1
- package/dist/esm/lib/streaming/zip-batch-iterators.js +0 -5
- package/dist/esm/lib/streaming/zip-batch-iterators.js.map +1 -1
- package/dist/esm/shapefile-loader.js +1 -3
- package/dist/esm/shapefile-loader.js.map +1 -1
- package/dist/esm/shp-loader.js +1 -4
- package/dist/esm/shp-loader.js.map +1 -1
- package/dist/shp-worker.js +1 -1
- package/package.json +5 -5
|
@@ -11,7 +11,7 @@ var _createClass2 = _interopRequireDefault(require("@babel/runtime/helpers/creat
|
|
|
11
11
|
var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/defineProperty"));
|
|
12
12
|
function _createForOfIteratorHelper(o, allowArrayLike) { var it = typeof Symbol !== "undefined" && o[Symbol.iterator] || o["@@iterator"]; if (!it) { if (Array.isArray(o) || (it = _unsupportedIterableToArray(o)) || allowArrayLike && o && typeof o.length === "number") { if (it) o = it; var i = 0; var F = function F() {}; return { s: F, n: function n() { if (i >= o.length) return { done: true }; return { done: false, value: o[i++] }; }, e: function e(_e) { throw _e; }, f: F }; } throw new TypeError("Invalid attempt to iterate non-iterable instance.\nIn order to be iterable, non-array objects must have a [Symbol.iterator]() method."); } var normalCompletion = true, didErr = false, err; return { s: function s() { it = it.call(o); }, n: function n() { var step = it.next(); normalCompletion = step.done; return step; }, e: function e(_e2) { didErr = true; err = _e2; }, f: function f() { try { if (!normalCompletion && it.return != null) it.return(); } finally { if (didErr) throw err; } } }; }
|
|
13
13
|
function _unsupportedIterableToArray(o, minLen) { if (!o) return; if (typeof o === "string") return _arrayLikeToArray(o, minLen); var n = Object.prototype.toString.call(o).slice(8, -1); if (n === "Object" && o.constructor) n = o.constructor.name; if (n === "Map" || n === "Set") return Array.from(o); if (n === "Arguments" || /^(?:Ui|I)nt(?:8|16|32)(?:Clamped)?Array$/.test(n)) return _arrayLikeToArray(o, minLen); }
|
|
14
|
-
function _arrayLikeToArray(arr, len) { if (len == null || len > arr.length) len = arr.length; for (var i = 0, arr2 = new Array(len); i < len; i++)
|
|
14
|
+
function _arrayLikeToArray(arr, len) { if (len == null || len > arr.length) len = arr.length; for (var i = 0, arr2 = new Array(len); i < len; i++) arr2[i] = arr[i]; return arr2; }
|
|
15
15
|
var BinaryChunkReader = function () {
|
|
16
16
|
function BinaryChunkReader(options) {
|
|
17
17
|
(0, _classCallCheck2.default)(this, BinaryChunkReader);
|
|
@@ -22,17 +22,14 @@ var BinaryChunkReader = function () {
|
|
|
22
22
|
var _ref = options || {},
|
|
23
23
|
_ref$maxRewindBytes = _ref.maxRewindBytes,
|
|
24
24
|
maxRewindBytes = _ref$maxRewindBytes === void 0 ? 0 : _ref$maxRewindBytes;
|
|
25
|
-
|
|
26
25
|
this.offset = 0;
|
|
27
26
|
this.arrayBuffers = [];
|
|
28
27
|
this.ended = false;
|
|
29
|
-
|
|
30
28
|
this.maxRewindBytes = maxRewindBytes;
|
|
31
29
|
}
|
|
32
30
|
(0, _createClass2.default)(BinaryChunkReader, [{
|
|
33
31
|
key: "write",
|
|
34
|
-
value:
|
|
35
|
-
function write(arrayBuffer) {
|
|
32
|
+
value: function write(arrayBuffer) {
|
|
36
33
|
this.arrayBuffers.push(arrayBuffer);
|
|
37
34
|
}
|
|
38
35
|
}, {
|
|
@@ -41,11 +38,9 @@ var BinaryChunkReader = function () {
|
|
|
41
38
|
this.arrayBuffers = [];
|
|
42
39
|
this.ended = true;
|
|
43
40
|
}
|
|
44
|
-
|
|
45
41
|
}, {
|
|
46
42
|
key: "hasAvailableBytes",
|
|
47
|
-
value:
|
|
48
|
-
function hasAvailableBytes(bytes) {
|
|
43
|
+
value: function hasAvailableBytes(bytes) {
|
|
49
44
|
var bytesAvailable = -this.offset;
|
|
50
45
|
var _iterator = _createForOfIteratorHelper(this.arrayBuffers),
|
|
51
46
|
_step;
|
|
@@ -64,44 +59,34 @@ var BinaryChunkReader = function () {
|
|
|
64
59
|
}
|
|
65
60
|
return false;
|
|
66
61
|
}
|
|
67
|
-
|
|
68
62
|
}, {
|
|
69
63
|
key: "findBufferOffsets",
|
|
70
|
-
value:
|
|
71
|
-
function findBufferOffsets(bytes) {
|
|
64
|
+
value: function findBufferOffsets(bytes) {
|
|
72
65
|
var offset = -this.offset;
|
|
73
66
|
var selectedBuffers = [];
|
|
74
67
|
for (var i = 0; i < this.arrayBuffers.length; i++) {
|
|
75
68
|
var buf = this.arrayBuffers[i];
|
|
76
|
-
|
|
77
69
|
if (offset + buf.byteLength <= 0) {
|
|
78
70
|
offset += buf.byteLength;
|
|
79
71
|
continue;
|
|
80
72
|
}
|
|
81
|
-
|
|
82
73
|
var start = offset <= 0 ? Math.abs(offset) : 0;
|
|
83
74
|
var end = void 0;
|
|
84
|
-
|
|
85
75
|
if (start + bytes <= buf.byteLength) {
|
|
86
76
|
end = start + bytes;
|
|
87
77
|
selectedBuffers.push([i, [start, end]]);
|
|
88
78
|
return selectedBuffers;
|
|
89
79
|
}
|
|
90
|
-
|
|
91
80
|
end = buf.byteLength;
|
|
92
81
|
selectedBuffers.push([i, [start, end]]);
|
|
93
|
-
|
|
94
82
|
bytes -= buf.byteLength - start;
|
|
95
83
|
offset += buf.byteLength;
|
|
96
84
|
}
|
|
97
|
-
|
|
98
85
|
return null;
|
|
99
86
|
}
|
|
100
|
-
|
|
101
87
|
}, {
|
|
102
88
|
key: "getDataView",
|
|
103
|
-
value:
|
|
104
|
-
function getDataView(bytes) {
|
|
89
|
+
value: function getDataView(bytes) {
|
|
105
90
|
var bufferOffsets = this.findBufferOffsets(bytes);
|
|
106
91
|
if (!bufferOffsets && this.ended) {
|
|
107
92
|
throw new Error('binary data exhausted');
|
|
@@ -109,7 +94,6 @@ var BinaryChunkReader = function () {
|
|
|
109
94
|
if (!bufferOffsets) {
|
|
110
95
|
return null;
|
|
111
96
|
}
|
|
112
|
-
|
|
113
97
|
if (bufferOffsets.length === 1) {
|
|
114
98
|
var _bufferOffsets$ = (0, _slicedToArray2.default)(bufferOffsets[0], 2),
|
|
115
99
|
bufferIndex = _bufferOffsets$[0],
|
|
@@ -122,27 +106,22 @@ var BinaryChunkReader = function () {
|
|
|
122
106
|
this.disposeBuffers();
|
|
123
107
|
return _view;
|
|
124
108
|
}
|
|
125
|
-
|
|
126
109
|
var view = new DataView(this._combineArrayBuffers(bufferOffsets));
|
|
127
110
|
this.offset += bytes;
|
|
128
111
|
this.disposeBuffers();
|
|
129
112
|
return view;
|
|
130
113
|
}
|
|
131
|
-
|
|
132
114
|
}, {
|
|
133
115
|
key: "disposeBuffers",
|
|
134
|
-
value:
|
|
135
|
-
function disposeBuffers() {
|
|
116
|
+
value: function disposeBuffers() {
|
|
136
117
|
while (this.arrayBuffers.length > 0 && this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength) {
|
|
137
118
|
this.offset -= this.arrayBuffers[0].byteLength;
|
|
138
119
|
this.arrayBuffers.shift();
|
|
139
120
|
}
|
|
140
121
|
}
|
|
141
|
-
|
|
142
122
|
}, {
|
|
143
123
|
key: "_combineArrayBuffers",
|
|
144
|
-
value:
|
|
145
|
-
function _combineArrayBuffers(bufferOffsets) {
|
|
124
|
+
value: function _combineArrayBuffers(bufferOffsets) {
|
|
146
125
|
var byteLength = 0;
|
|
147
126
|
var _iterator2 = _createForOfIteratorHelper(bufferOffsets),
|
|
148
127
|
_step2;
|
|
@@ -160,7 +139,6 @@ var BinaryChunkReader = function () {
|
|
|
160
139
|
_iterator2.f();
|
|
161
140
|
}
|
|
162
141
|
var result = new Uint8Array(byteLength);
|
|
163
|
-
|
|
164
142
|
var resultOffset = 0;
|
|
165
143
|
var _iterator3 = _createForOfIteratorHelper(bufferOffsets),
|
|
166
144
|
_step3;
|
|
@@ -185,14 +163,12 @@ var BinaryChunkReader = function () {
|
|
|
185
163
|
}
|
|
186
164
|
}, {
|
|
187
165
|
key: "skip",
|
|
188
|
-
value:
|
|
189
|
-
function skip(bytes) {
|
|
166
|
+
value: function skip(bytes) {
|
|
190
167
|
this.offset += bytes;
|
|
191
168
|
}
|
|
192
169
|
}, {
|
|
193
170
|
key: "rewind",
|
|
194
|
-
value:
|
|
195
|
-
function rewind(bytes) {
|
|
171
|
+
value: function rewind(bytes) {
|
|
196
172
|
this.offset -= bytes;
|
|
197
173
|
}
|
|
198
174
|
}]);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"binary-chunk-reader.js","names":["BinaryChunkReader","options","maxRewindBytes","offset","arrayBuffers","ended","arrayBuffer","push","bytes","bytesAvailable","byteLength","selectedBuffers","i","length","buf","start","Math","abs","end","bufferOffsets","findBufferOffsets","Error","bufferIndex","view","DataView","disposeBuffers","_combineArrayBuffers","shift","bufferOffset","result","Uint8Array","resultOffset","sourceArray","set","subarray","buffer"],"sources":["../../../../src/lib/streaming/binary-chunk-reader.ts"],"sourcesContent":["type BinaryChunkReaderOptions = {\n maxRewindBytes: number;\n};\n\nexport default class BinaryChunkReader {\n offset: number;\n arrayBuffers: ArrayBuffer[];\n ended: boolean;\n maxRewindBytes: number;\n\n constructor(options?: BinaryChunkReaderOptions) {\n const {maxRewindBytes = 0} = options || {};\n\n /** current global offset into current array buffer*/\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffers = [];\n this.ended = false;\n\n /** bytes behind offset to hold on to */\n this.maxRewindBytes = maxRewindBytes;\n }\n /**\n * @param arrayBuffer\n */\n write(arrayBuffer: ArrayBuffer): void {\n this.arrayBuffers.push(arrayBuffer);\n }\n\n end(): void {\n this.arrayBuffers = [];\n this.ended = true;\n }\n\n /**\n * Has enough bytes available in array buffers\n *\n * @param bytes Number of bytes\n * @return boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n let bytesAvailable = -this.offset;\n for (const arrayBuffer of this.arrayBuffers) {\n bytesAvailable += arrayBuffer.byteLength;\n if (bytesAvailable >= bytes) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * Find offsets of byte ranges within this.arrayBuffers\n *\n * @param bytes Byte length to read\n * @return Arrays with byte ranges pointing to this.arrayBuffers, Output type is nested array, e.g. [ [0, [1, 2]], ...]\n */\n findBufferOffsets(bytes: number): any[] | null {\n let offset = -this.offset;\n const selectedBuffers: any = [];\n\n for (let i = 0; i < this.arrayBuffers.length; i++) {\n const buf = this.arrayBuffers[i];\n\n // Current buffer isn't long enough to reach global offset\n if (offset + buf.byteLength <= 0) {\n offset += buf.byteLength;\n // eslint-disable-next-line no-continue\n continue;\n }\n\n // Find start/end offsets for this buffer\n // When offset < 0, need to skip over Math.abs(offset) bytes\n // When offset > 0, implies bytes in previous buffer, start at 0\n const start = offset <= 0 ? Math.abs(offset) : 0;\n let end: number;\n\n // Length of requested bytes is contained in current buffer\n if (start + bytes <= buf.byteLength) {\n end = start + bytes;\n selectedBuffers.push([i, [start, end]]);\n return selectedBuffers;\n }\n\n // Will need to look into next buffer\n end = buf.byteLength;\n selectedBuffers.push([i, [start, end]]);\n\n // Need to read fewer bytes in next iter\n bytes -= buf.byteLength - start;\n offset += buf.byteLength;\n }\n\n // Should only finish loop if exhausted all arrays\n return null;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes Number of bytes\n * @return DataView with data\n */\n getDataView(bytes: number): DataView | null {\n const bufferOffsets = this.findBufferOffsets(bytes);\n // return `null` if not enough data, except if end() already called, in\n // which case throw an error.\n if (!bufferOffsets && this.ended) {\n throw new Error('binary data exhausted');\n }\n\n if (!bufferOffsets) {\n // @ts-ignore\n return null;\n }\n\n // If only one arrayBuffer needed, return DataView directly\n if (bufferOffsets.length === 1) {\n const [bufferIndex, [start, end]] = bufferOffsets[0];\n const arrayBuffer = this.arrayBuffers[bufferIndex];\n const view = new DataView(arrayBuffer, start, end - start);\n\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n // Concatenate portions of multiple ArrayBuffers\n const view = new DataView(this._combineArrayBuffers(bufferOffsets));\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n /**\n * Dispose of old array buffers\n */\n disposeBuffers(): void {\n while (\n this.arrayBuffers.length > 0 &&\n this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength\n ) {\n this.offset -= this.arrayBuffers[0].byteLength;\n this.arrayBuffers.shift();\n }\n }\n\n /**\n * Copy multiple ArrayBuffers into one contiguous ArrayBuffer\n *\n * In contrast to concatenateArrayBuffers, this only copies the necessary\n * portions of the source arrays, rather than first copying the entire arrays\n * then taking a part of them.\n *\n * @param bufferOffsets List of internal array offsets\n * @return New contiguous ArrayBuffer\n */\n _combineArrayBuffers(bufferOffsets: any[]): ArrayBufferLike {\n let byteLength: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [start, end] = bufferOffset[1];\n byteLength += end - start;\n }\n\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let resultOffset: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [bufferIndex, [start, end]] = bufferOffset;\n const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);\n result.set(sourceArray.subarray(start, end), resultOffset);\n resultOffset += end - start;\n }\n\n return result.buffer;\n }\n /**\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n /**\n * @param bytes\n */\n rewind(bytes: number): void {\n // TODO - only works if offset is already set\n this.offset -= bytes;\n }\n}\n"],"mappings":";;;;;;;;;;;;;;IAIqBA,iBAAiB;EAMpC,2BAAYC,OAAkC,EAAE;IAAA;IAAA;IAAA;IAAA;IAAA;IAC9C,WAA6BA,OAAO,IAAI,CAAC,CAAC;MAAA,2BAAnCC,cAAc;MAAdA,cAAc,oCAAG,CAAC;;IAGzB,IAAI,CAACC,MAAM,GAAG,CAAC;IAEf,IAAI,CAACC,YAAY,GAAG,EAAE;IACtB,IAAI,CAACC,KAAK,GAAG,KAAK;;IAGlB,IAAI,CAACH,cAAc,GAAGA,cAAc;EACtC;EAAC;IAAA;IAAA;IAID,eAAMI,WAAwB,EAAQ;MACpC,IAAI,CAACF,YAAY,CAACG,IAAI,CAACD,WAAW,CAAC;IACrC;EAAC;IAAA;IAAA,OAED,eAAY;MACV,IAAI,CAACF,YAAY,GAAG,EAAE;MACtB,IAAI,CAACC,KAAK,GAAG,IAAI;IACnB;;EAAC;IAAA;IAAA;IAQD,2BAAkBG,KAAa,EAAW;MACxC,IAAIC,cAAc,GAAG,CAAC,IAAI,CAACN,MAAM;MAAC,2CACR,IAAI,CAACC,YAAY;QAAA;MAAA;QAA3C,oDAA6C;UAAA,IAAlCE,WAAW;UACpBG,cAAc,IAAIH,WAAW,CAACI,UAAU;UACxC,IAAID,cAAc,IAAID,KAAK,EAAE;YAC3B,OAAO,IAAI;UACb;QACF;MAAC;QAAA;MAAA;QAAA;MAAA;MACD,OAAO,KAAK;IACd;;EAAC;IAAA;IAAA;IAQD,2BAAkBA,KAAa,EAAgB;MAC7C,IAAIL,MAAM,GAAG,CAAC,IAAI,CAACA,MAAM;MACzB,IAAMQ,eAAoB,GAAG,EAAE;MAE/B,KAAK,IAAIC,CAAC,GAAG,CAAC,EAAEA,CAAC,GAAG,IAAI,CAACR,YAAY,CAACS,MAAM,EAAED,CAAC,EAAE,EAAE;QACjD,IAAME,GAAG,GAAG,IAAI,CAACV,YAAY,CAACQ,CAAC,CAAC;;QAGhC,IAAIT,MAAM,GAAGW,GAAG,CAACJ,UAAU,IAAI,CAAC,EAAE;UAChCP,MAAM,IAAIW,GAAG,CAACJ,UAAU;UAExB;QACF;;QAKA,IAAMK,KAAK,GAAGZ,MAAM,IAAI,CAAC,GAAGa,IAAI,CAACC,GAAG,CAACd,MAAM,CAAC,GAAG,CAAC;QAChD,IAAIe,GAAW;;QAGf,IAAIH,KAAK,GAAGP,KAAK,IAAIM,GAAG,CAACJ,UAAU,EAAE;UACnCQ,GAAG,GAAGH,KAAK,GAAGP,KAAK;UACnBG,eAAe,CAACJ,IAAI,CAAC,CAACK,CAAC,EAAE,CAACG,KAAK,EAAEG,GAAG,CAAC,CAAC,CAAC;UACvC,OAAOP,eAAe;QACxB;;QAGAO,GAAG,GAAGJ,GAAG,CAACJ,UAAU;QACpBC,eAAe,CAACJ,IAAI,CAAC,CAACK,CAAC,EAAE,CAACG,KAAK,EAAEG,GAAG,CAAC,CAAC,CAAC;;QAGvCV,KAAK,IAAIM,GAAG,CAACJ,UAAU,GAAGK,KAAK;QAC/BZ,MAAM,IAAIW,GAAG,CAACJ,UAAU;MAC1B;;MAGA,OAAO,IAAI;IACb;;EAAC;IAAA;IAAA;IAQD,qBAAYF,KAAa,EAAmB;MAC1C,IAAMW,aAAa,GAAG,IAAI,CAACC,iBAAiB,CAACZ,KAAK,CAAC;MAGnD,IAAI,CAACW,aAAa,IAAI,IAAI,CAACd,KAAK,EAAE;QAChC,MAAM,IAAIgB,KAAK,CAAC,uBAAuB,CAAC;MAC1C;MAEA,IAAI,CAACF,aAAa,EAAE;QAElB,OAAO,IAAI;MACb;;MAGA,IAAIA,aAAa,CAACN,MAAM,KAAK,CAAC,EAAE;QAC9B,mDAAoCM,aAAa,CAAC,CAAC,CAAC;UAA7CG,WAAW;UAAA;UAAGP,KAAK;UAAEG,GAAG;QAC/B,IAAMZ,WAAW,GAAG,IAAI,CAACF,YAAY,CAACkB,WAAW,CAAC;QAClD,IAAMC,KAAI,GAAG,IAAIC,QAAQ,CAAClB,WAAW,EAAES,KAAK,EAAEG,GAAG,GAAGH,KAAK,CAAC;QAE1D,IAAI,CAACZ,MAAM,IAAIK,KAAK;QACpB,IAAI,CAACiB,cAAc,EAAE;QACrB,OAAOF,KAAI;MACb;;MAGA,IAAMA,IAAI,GAAG,IAAIC,QAAQ,CAAC,IAAI,CAACE,oBAAoB,CAACP,aAAa,CAAC,CAAC;MACnE,IAAI,CAAChB,MAAM,IAAIK,KAAK;MACpB,IAAI,CAACiB,cAAc,EAAE;MACrB,OAAOF,IAAI;IACb;;EAAC;IAAA;IAAA;IAKD,0BAAuB;MACrB,OACE,IAAI,CAACnB,YAAY,CAACS,MAAM,GAAG,CAAC,IAC5B,IAAI,CAACV,MAAM,GAAG,IAAI,CAACD,cAAc,IAAI,IAAI,CAACE,YAAY,CAAC,CAAC,CAAC,CAACM,UAAU,EACpE;QACA,IAAI,CAACP,MAAM,IAAI,IAAI,CAACC,YAAY,CAAC,CAAC,CAAC,CAACM,UAAU;QAC9C,IAAI,CAACN,YAAY,CAACuB,KAAK,EAAE;MAC3B;IACF;;EAAC;IAAA;IAAA;IAYD,8BAAqBR,aAAoB,EAAmB;MAC1D,IAAIT,UAAkB,GAAG,CAAC;MAAC,4CACAS,aAAa;QAAA;MAAA;QAAxC,uDAA0C;UAAA,IAA/BS,YAAY;UACrB,kDAAqBA,YAAY,CAAC,CAAC,CAAC;YAA7Bb,KAAK;YAAEG,GAAG;UACjBR,UAAU,IAAIQ,GAAG,GAAGH,KAAK;QAC3B;MAAC;QAAA;MAAA;QAAA;MAAA;MAED,IAAMc,MAAM,GAAG,IAAIC,UAAU,CAACpB,UAAU,CAAC;;MAGzC,IAAIqB,YAAoB,GAAG,CAAC;MAAC,4CACFZ,aAAa;QAAA;MAAA;QAAxC,uDAA0C;UAAA,IAA/BS,aAAY;UACrB,kDAAoCA,aAAY;YAAzCN,WAAW;YAAA;YAAGP,MAAK;YAAEG,IAAG;UAC/B,IAAMc,WAAW,GAAG,IAAIF,UAAU,CAAC,IAAI,CAAC1B,YAAY,CAACkB,WAAW,CAAC,CAAC;UAClEO,MAAM,CAACI,GAAG,CAACD,WAAW,CAACE,QAAQ,CAACnB,MAAK,EAAEG,IAAG,CAAC,EAAEa,YAAY,CAAC;UAC1DA,YAAY,IAAIb,IAAG,GAAGH,MAAK;QAC7B;MAAC;QAAA;MAAA;QAAA;MAAA;MAED,OAAOc,MAAM,CAACM,MAAM;IACtB;EAAC;IAAA;IAAA;IAID,cAAK3B,KAAa,EAAQ;MACxB,IAAI,CAACL,MAAM,IAAIK,KAAK;IACtB;EAAC;IAAA;IAAA;IAID,gBAAOA,KAAa,EAAQ;MAE1B,IAAI,CAACL,MAAM,IAAIK,KAAK;IACtB;EAAC;EAAA;AAAA;AAAA"}
|
|
1
|
+
{"version":3,"file":"binary-chunk-reader.js","names":["BinaryChunkReader","options","_classCallCheck2","default","_defineProperty2","_ref","_ref$maxRewindBytes","maxRewindBytes","offset","arrayBuffers","ended","_createClass2","key","value","write","arrayBuffer","push","end","hasAvailableBytes","bytes","bytesAvailable","_iterator","_createForOfIteratorHelper","_step","s","n","done","byteLength","err","e","f","findBufferOffsets","selectedBuffers","i","length","buf","start","Math","abs","getDataView","bufferOffsets","Error","_bufferOffsets$","_slicedToArray2","bufferIndex","_bufferOffsets$$","view","DataView","disposeBuffers","_combineArrayBuffers","shift","_iterator2","_step2","bufferOffset","_bufferOffset$","result","Uint8Array","resultOffset","_iterator3","_step3","_bufferOffset2","_bufferOffset2$","sourceArray","set","subarray","buffer","skip","rewind","exports"],"sources":["../../../../src/lib/streaming/binary-chunk-reader.ts"],"sourcesContent":["type BinaryChunkReaderOptions = {\n maxRewindBytes: number;\n};\n\nexport default class BinaryChunkReader {\n offset: number;\n arrayBuffers: ArrayBuffer[];\n ended: boolean;\n maxRewindBytes: number;\n\n constructor(options?: BinaryChunkReaderOptions) {\n const {maxRewindBytes = 0} = options || {};\n\n /** current global offset into current array buffer*/\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffers = [];\n this.ended = false;\n\n /** bytes behind offset to hold on to */\n this.maxRewindBytes = maxRewindBytes;\n }\n /**\n * @param arrayBuffer\n */\n write(arrayBuffer: ArrayBuffer): void {\n this.arrayBuffers.push(arrayBuffer);\n }\n\n end(): void {\n this.arrayBuffers = [];\n this.ended = true;\n }\n\n /**\n * Has enough bytes available in array buffers\n *\n * @param bytes Number of bytes\n * @return boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n let bytesAvailable = -this.offset;\n for (const arrayBuffer of this.arrayBuffers) {\n bytesAvailable += arrayBuffer.byteLength;\n if (bytesAvailable >= bytes) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * Find offsets of byte ranges within this.arrayBuffers\n *\n * @param bytes Byte length to read\n * @return Arrays with byte ranges pointing to this.arrayBuffers, Output type is nested array, e.g. [ [0, [1, 2]], ...]\n */\n findBufferOffsets(bytes: number): any[] | null {\n let offset = -this.offset;\n const selectedBuffers: any = [];\n\n for (let i = 0; i < this.arrayBuffers.length; i++) {\n const buf = this.arrayBuffers[i];\n\n // Current buffer isn't long enough to reach global offset\n if (offset + buf.byteLength <= 0) {\n offset += buf.byteLength;\n // eslint-disable-next-line no-continue\n continue;\n }\n\n // Find start/end offsets for this buffer\n // When offset < 0, need to skip over Math.abs(offset) bytes\n // When offset > 0, implies bytes in previous buffer, start at 0\n const start = offset <= 0 ? Math.abs(offset) : 0;\n let end: number;\n\n // Length of requested bytes is contained in current buffer\n if (start + bytes <= buf.byteLength) {\n end = start + bytes;\n selectedBuffers.push([i, [start, end]]);\n return selectedBuffers;\n }\n\n // Will need to look into next buffer\n end = buf.byteLength;\n selectedBuffers.push([i, [start, end]]);\n\n // Need to read fewer bytes in next iter\n bytes -= buf.byteLength - start;\n offset += buf.byteLength;\n }\n\n // Should only finish loop if exhausted all arrays\n return null;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes Number of bytes\n * @return DataView with data\n */\n getDataView(bytes: number): DataView | null {\n const bufferOffsets = this.findBufferOffsets(bytes);\n // return `null` if not enough data, except if end() already called, in\n // which case throw an error.\n if (!bufferOffsets && this.ended) {\n throw new Error('binary data exhausted');\n }\n\n if (!bufferOffsets) {\n // @ts-ignore\n return null;\n }\n\n // If only one arrayBuffer needed, return DataView directly\n if (bufferOffsets.length === 1) {\n const [bufferIndex, [start, end]] = bufferOffsets[0];\n const arrayBuffer = this.arrayBuffers[bufferIndex];\n const view = new DataView(arrayBuffer, start, end - start);\n\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n // Concatenate portions of multiple ArrayBuffers\n const view = new DataView(this._combineArrayBuffers(bufferOffsets));\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n /**\n * Dispose of old array buffers\n */\n disposeBuffers(): void {\n while (\n this.arrayBuffers.length > 0 &&\n this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength\n ) {\n this.offset -= this.arrayBuffers[0].byteLength;\n this.arrayBuffers.shift();\n }\n }\n\n /**\n * Copy multiple ArrayBuffers into one contiguous ArrayBuffer\n *\n * In contrast to concatenateArrayBuffers, this only copies the necessary\n * portions of the source arrays, rather than first copying the entire arrays\n * then taking a part of them.\n *\n * @param bufferOffsets List of internal array offsets\n * @return New contiguous ArrayBuffer\n */\n _combineArrayBuffers(bufferOffsets: any[]): ArrayBufferLike {\n let byteLength: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [start, end] = bufferOffset[1];\n byteLength += end - start;\n }\n\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let resultOffset: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [bufferIndex, [start, end]] = bufferOffset;\n const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);\n result.set(sourceArray.subarray(start, end), resultOffset);\n resultOffset += end - start;\n }\n\n return result.buffer;\n }\n /**\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n /**\n * @param bytes\n */\n rewind(bytes: number): void {\n // TODO - only works if offset is already set\n this.offset -= bytes;\n }\n}\n"],"mappings":";;;;;;;;;;;;;;IAIqBA,iBAAiB;EAMpC,SAAAA,kBAAYC,OAAkC,EAAE;IAAA,IAAAC,gBAAA,CAAAC,OAAA,QAAAH,iBAAA;IAAA,IAAAI,gBAAA,CAAAD,OAAA;IAAA,IAAAC,gBAAA,CAAAD,OAAA;IAAA,IAAAC,gBAAA,CAAAD,OAAA;IAAA,IAAAC,gBAAA,CAAAD,OAAA;IAC9C,IAAAE,IAAA,GAA6BJ,OAAO,IAAI,CAAC,CAAC;MAAAK,mBAAA,GAAAD,IAAA,CAAnCE,cAAc;MAAdA,cAAc,GAAAD,mBAAA,cAAG,CAAC,GAAAA,mBAAA;IAGzB,IAAI,CAACE,MAAM,GAAG,CAAC;IAEf,IAAI,CAACC,YAAY,GAAG,EAAE;IACtB,IAAI,CAACC,KAAK,GAAG,KAAK;IAGlB,IAAI,CAACH,cAAc,GAAGA,cAAc;EACtC;EAAC,IAAAI,aAAA,CAAAR,OAAA,EAAAH,iBAAA;IAAAY,GAAA;IAAAC,KAAA,EAID,SAAAC,MAAMC,WAAwB,EAAQ;MACpC,IAAI,CAACN,YAAY,CAACO,IAAI,CAACD,WAAW,CAAC;IACrC;EAAC;IAAAH,GAAA;IAAAC,KAAA,EAED,SAAAI,IAAA,EAAY;MACV,IAAI,CAACR,YAAY,GAAG,EAAE;MACtB,IAAI,CAACC,KAAK,GAAG,IAAI;IACnB;EAAC;IAAAE,GAAA;IAAAC,KAAA,EAQD,SAAAK,kBAAkBC,KAAa,EAAW;MACxC,IAAIC,cAAc,GAAG,CAAC,IAAI,CAACZ,MAAM;MAAC,IAAAa,SAAA,GAAAC,0BAAA,CACR,IAAI,CAACb,YAAY;QAAAc,KAAA;MAAA;QAA3C,KAAAF,SAAA,CAAAG,CAAA,MAAAD,KAAA,GAAAF,SAAA,CAAAI,CAAA,IAAAC,IAAA,GAA6C;UAAA,IAAlCX,WAAW,GAAAQ,KAAA,CAAAV,KAAA;UACpBO,cAAc,IAAIL,WAAW,CAACY,UAAU;UACxC,IAAIP,cAAc,IAAID,KAAK,EAAE;YAC3B,OAAO,IAAI;UACb;QACF;MAAC,SAAAS,GAAA;QAAAP,SAAA,CAAAQ,CAAA,CAAAD,GAAA;MAAA;QAAAP,SAAA,CAAAS,CAAA;MAAA;MACD,OAAO,KAAK;IACd;EAAC;IAAAlB,GAAA;IAAAC,KAAA,EAQD,SAAAkB,kBAAkBZ,KAAa,EAAgB;MAC7C,IAAIX,MAAM,GAAG,CAAC,IAAI,CAACA,MAAM;MACzB,IAAMwB,eAAoB,GAAG,EAAE;MAE/B,KAAK,IAAIC,CAAC,GAAG,CAAC,EAAEA,CAAC,GAAG,IAAI,CAACxB,YAAY,CAACyB,MAAM,EAAED,CAAC,EAAE,EAAE;QACjD,IAAME,GAAG,GAAG,IAAI,CAAC1B,YAAY,CAACwB,CAAC,CAAC;QAGhC,IAAIzB,MAAM,GAAG2B,GAAG,CAACR,UAAU,IAAI,CAAC,EAAE;UAChCnB,MAAM,IAAI2B,GAAG,CAACR,UAAU;UAExB;QACF;QAKA,IAAMS,KAAK,GAAG5B,MAAM,IAAI,CAAC,GAAG6B,IAAI,CAACC,GAAG,CAAC9B,MAAM,CAAC,GAAG,CAAC;QAChD,IAAIS,GAAW;QAGf,IAAImB,KAAK,GAAGjB,KAAK,IAAIgB,GAAG,CAACR,UAAU,EAAE;UACnCV,GAAG,GAAGmB,KAAK,GAAGjB,KAAK;UACnBa,eAAe,CAAChB,IAAI,CAAC,CAACiB,CAAC,EAAE,CAACG,KAAK,EAAEnB,GAAG,CAAC,CAAC,CAAC;UACvC,OAAOe,eAAe;QACxB;QAGAf,GAAG,GAAGkB,GAAG,CAACR,UAAU;QACpBK,eAAe,CAAChB,IAAI,CAAC,CAACiB,CAAC,EAAE,CAACG,KAAK,EAAEnB,GAAG,CAAC,CAAC,CAAC;QAGvCE,KAAK,IAAIgB,GAAG,CAACR,UAAU,GAAGS,KAAK;QAC/B5B,MAAM,IAAI2B,GAAG,CAACR,UAAU;MAC1B;MAGA,OAAO,IAAI;IACb;EAAC;IAAAf,GAAA;IAAAC,KAAA,EAQD,SAAA0B,YAAYpB,KAAa,EAAmB;MAC1C,IAAMqB,aAAa,GAAG,IAAI,CAACT,iBAAiB,CAACZ,KAAK,CAAC;MAGnD,IAAI,CAACqB,aAAa,IAAI,IAAI,CAAC9B,KAAK,EAAE;QAChC,MAAM,IAAI+B,KAAK,CAAC,uBAAuB,CAAC;MAC1C;MAEA,IAAI,CAACD,aAAa,EAAE;QAElB,OAAO,IAAI;MACb;MAGA,IAAIA,aAAa,CAACN,MAAM,KAAK,CAAC,EAAE;QAC9B,IAAAQ,eAAA,OAAAC,eAAA,CAAAxC,OAAA,EAAoCqC,aAAa,CAAC,CAAC,CAAC;UAA7CI,WAAW,GAAAF,eAAA;UAAAG,gBAAA,OAAAF,eAAA,CAAAxC,OAAA,EAAAuC,eAAA;UAAGN,KAAK,GAAAS,gBAAA;UAAE5B,GAAG,GAAA4B,gBAAA;QAC/B,IAAM9B,WAAW,GAAG,IAAI,CAACN,YAAY,CAACmC,WAAW,CAAC;QAClD,IAAME,KAAI,GAAG,IAAIC,QAAQ,CAAChC,WAAW,EAAEqB,KAAK,EAAEnB,GAAG,GAAGmB,KAAK,CAAC;QAE1D,IAAI,CAAC5B,MAAM,IAAIW,KAAK;QACpB,IAAI,CAAC6B,cAAc,CAAC,CAAC;QACrB,OAAOF,KAAI;MACb;MAGA,IAAMA,IAAI,GAAG,IAAIC,QAAQ,CAAC,IAAI,CAACE,oBAAoB,CAACT,aAAa,CAAC,CAAC;MACnE,IAAI,CAAChC,MAAM,IAAIW,KAAK;MACpB,IAAI,CAAC6B,cAAc,CAAC,CAAC;MACrB,OAAOF,IAAI;IACb;EAAC;IAAAlC,GAAA;IAAAC,KAAA,EAKD,SAAAmC,eAAA,EAAuB;MACrB,OACE,IAAI,CAACvC,YAAY,CAACyB,MAAM,GAAG,CAAC,IAC5B,IAAI,CAAC1B,MAAM,GAAG,IAAI,CAACD,cAAc,IAAI,IAAI,CAACE,YAAY,CAAC,CAAC,CAAC,CAACkB,UAAU,EACpE;QACA,IAAI,CAACnB,MAAM,IAAI,IAAI,CAACC,YAAY,CAAC,CAAC,CAAC,CAACkB,UAAU;QAC9C,IAAI,CAAClB,YAAY,CAACyC,KAAK,CAAC,CAAC;MAC3B;IACF;EAAC;IAAAtC,GAAA;IAAAC,KAAA,EAYD,SAAAoC,qBAAqBT,aAAoB,EAAmB;MAC1D,IAAIb,UAAkB,GAAG,CAAC;MAAC,IAAAwB,UAAA,GAAA7B,0BAAA,CACAkB,aAAa;QAAAY,MAAA;MAAA;QAAxC,KAAAD,UAAA,CAAA3B,CAAA,MAAA4B,MAAA,GAAAD,UAAA,CAAA1B,CAAA,IAAAC,IAAA,GAA0C;UAAA,IAA/B2B,YAAY,GAAAD,MAAA,CAAAvC,KAAA;UACrB,IAAAyC,cAAA,OAAAX,eAAA,CAAAxC,OAAA,EAAqBkD,YAAY,CAAC,CAAC,CAAC;YAA7BjB,KAAK,GAAAkB,cAAA;YAAErC,GAAG,GAAAqC,cAAA;UACjB3B,UAAU,IAAIV,GAAG,GAAGmB,KAAK;QAC3B;MAAC,SAAAR,GAAA;QAAAuB,UAAA,CAAAtB,CAAA,CAAAD,GAAA;MAAA;QAAAuB,UAAA,CAAArB,CAAA;MAAA;MAED,IAAMyB,MAAM,GAAG,IAAIC,UAAU,CAAC7B,UAAU,CAAC;MAGzC,IAAI8B,YAAoB,GAAG,CAAC;MAAC,IAAAC,UAAA,GAAApC,0BAAA,CACFkB,aAAa;QAAAmB,MAAA;MAAA;QAAxC,KAAAD,UAAA,CAAAlC,CAAA,MAAAmC,MAAA,GAAAD,UAAA,CAAAjC,CAAA,IAAAC,IAAA,GAA0C;UAAA,IAA/B2B,aAAY,GAAAM,MAAA,CAAA9C,KAAA;UACrB,IAAA+C,cAAA,OAAAjB,eAAA,CAAAxC,OAAA,EAAoCkD,aAAY;YAAzCT,WAAW,GAAAgB,cAAA;YAAAC,eAAA,OAAAlB,eAAA,CAAAxC,OAAA,EAAAyD,cAAA;YAAGxB,MAAK,GAAAyB,eAAA;YAAE5C,IAAG,GAAA4C,eAAA;UAC/B,IAAMC,WAAW,GAAG,IAAIN,UAAU,CAAC,IAAI,CAAC/C,YAAY,CAACmC,WAAW,CAAC,CAAC;UAClEW,MAAM,CAACQ,GAAG,CAACD,WAAW,CAACE,QAAQ,CAAC5B,MAAK,EAAEnB,IAAG,CAAC,EAAEwC,YAAY,CAAC;UAC1DA,YAAY,IAAIxC,IAAG,GAAGmB,MAAK;QAC7B;MAAC,SAAAR,GAAA;QAAA8B,UAAA,CAAA7B,CAAA,CAAAD,GAAA;MAAA;QAAA8B,UAAA,CAAA5B,CAAA;MAAA;MAED,OAAOyB,MAAM,CAACU,MAAM;IACtB;EAAC;IAAArD,GAAA;IAAAC,KAAA,EAID,SAAAqD,KAAK/C,KAAa,EAAQ;MACxB,IAAI,CAACX,MAAM,IAAIW,KAAK;IACtB;EAAC;IAAAP,GAAA;IAAAC,KAAA,EAID,SAAAsD,OAAOhD,KAAa,EAAQ;MAE1B,IAAI,CAACX,MAAM,IAAIW,KAAK;IACtB;EAAC;EAAA,OAAAnB,iBAAA;AAAA;AAAAoE,OAAA,CAAAjE,OAAA,GAAAH,iBAAA"}
|
|
@@ -18,15 +18,12 @@ var BinaryReader = function () {
|
|
|
18
18
|
}
|
|
19
19
|
(0, _createClass2.default)(BinaryReader, [{
|
|
20
20
|
key: "hasAvailableBytes",
|
|
21
|
-
value:
|
|
22
|
-
function hasAvailableBytes(bytes) {
|
|
21
|
+
value: function hasAvailableBytes(bytes) {
|
|
23
22
|
return this.arrayBuffer.byteLength - this.offset >= bytes;
|
|
24
23
|
}
|
|
25
|
-
|
|
26
24
|
}, {
|
|
27
25
|
key: "getDataView",
|
|
28
|
-
value:
|
|
29
|
-
function getDataView(bytes) {
|
|
26
|
+
value: function getDataView(bytes) {
|
|
30
27
|
if (bytes && !this.hasAvailableBytes(bytes)) {
|
|
31
28
|
throw new Error('binary data exhausted');
|
|
32
29
|
}
|
|
@@ -34,18 +31,14 @@ var BinaryReader = function () {
|
|
|
34
31
|
this.offset += bytes;
|
|
35
32
|
return dataView;
|
|
36
33
|
}
|
|
37
|
-
|
|
38
34
|
}, {
|
|
39
35
|
key: "skip",
|
|
40
|
-
value:
|
|
41
|
-
function skip(bytes) {
|
|
36
|
+
value: function skip(bytes) {
|
|
42
37
|
this.offset += bytes;
|
|
43
38
|
}
|
|
44
|
-
|
|
45
39
|
}, {
|
|
46
40
|
key: "rewind",
|
|
47
|
-
value:
|
|
48
|
-
function rewind(bytes) {
|
|
41
|
+
value: function rewind(bytes) {
|
|
49
42
|
this.offset -= bytes;
|
|
50
43
|
}
|
|
51
44
|
}]);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"binary-reader.js","names":["BinaryReader","arrayBuffer","offset","bytes","byteLength","
|
|
1
|
+
{"version":3,"file":"binary-reader.js","names":["BinaryReader","arrayBuffer","_classCallCheck2","default","_defineProperty2","offset","_createClass2","key","value","hasAvailableBytes","bytes","byteLength","getDataView","Error","dataView","DataView","skip","rewind","exports"],"sources":["../../../../src/lib/streaming/binary-reader.ts"],"sourcesContent":["export default class BinaryReader {\n offset: number;\n arrayBuffer: ArrayBuffer;\n\n constructor(arrayBuffer: ArrayBuffer) {\n /** current global (stream) offset */\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffer = arrayBuffer;\n }\n /**\n * Checks if there are available bytes in data\n *\n * @param bytes\n * @returns boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n return this.arrayBuffer.byteLength - this.offset >= bytes;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes\n * @returns Dataview\n */\n getDataView(bytes: number): DataView {\n if (bytes && !this.hasAvailableBytes(bytes)) {\n throw new Error('binary data exhausted');\n }\n\n const dataView = bytes\n ? new DataView(this.arrayBuffer, this.offset, bytes)\n : new DataView(this.arrayBuffer, this.offset);\n this.offset += bytes;\n return dataView;\n }\n\n /**\n * Skipping\n *\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n\n /**\n * Rewinding\n *\n * @param bytes\n */\n rewind(bytes: number): void {\n this.offset -= bytes;\n }\n}\n"],"mappings":";;;;;;;;;;IAAqBA,YAAY;EAI/B,SAAAA,aAAYC,WAAwB,EAAE;IAAA,IAAAC,gBAAA,CAAAC,OAAA,QAAAH,YAAA;IAAA,IAAAI,gBAAA,CAAAD,OAAA;IAAA,IAAAC,gBAAA,CAAAD,OAAA;IAEpC,IAAI,CAACE,MAAM,GAAG,CAAC;IAEf,IAAI,CAACJ,WAAW,GAAGA,WAAW;EAChC;EAAC,IAAAK,aAAA,CAAAH,OAAA,EAAAH,YAAA;IAAAO,GAAA;IAAAC,KAAA,EAOD,SAAAC,kBAAkBC,KAAa,EAAW;MACxC,OAAO,IAAI,CAACT,WAAW,CAACU,UAAU,GAAG,IAAI,CAACN,MAAM,IAAIK,KAAK;IAC3D;EAAC;IAAAH,GAAA;IAAAC,KAAA,EAQD,SAAAI,YAAYF,KAAa,EAAY;MACnC,IAAIA,KAAK,IAAI,CAAC,IAAI,CAACD,iBAAiB,CAACC,KAAK,CAAC,EAAE;QAC3C,MAAM,IAAIG,KAAK,CAAC,uBAAuB,CAAC;MAC1C;MAEA,IAAMC,QAAQ,GAAGJ,KAAK,GAClB,IAAIK,QAAQ,CAAC,IAAI,CAACd,WAAW,EAAE,IAAI,CAACI,MAAM,EAAEK,KAAK,CAAC,GAClD,IAAIK,QAAQ,CAAC,IAAI,CAACd,WAAW,EAAE,IAAI,CAACI,MAAM,CAAC;MAC/C,IAAI,CAACA,MAAM,IAAIK,KAAK;MACpB,OAAOI,QAAQ;IACjB;EAAC;IAAAP,GAAA;IAAAC,KAAA,EAOD,SAAAQ,KAAKN,KAAa,EAAQ;MACxB,IAAI,CAACL,MAAM,IAAIK,KAAK;IACtB;EAAC;IAAAH,GAAA;IAAAC,KAAA,EAOD,SAAAS,OAAOP,KAAa,EAAQ;MAC1B,IAAI,CAACL,MAAM,IAAIK,KAAK;IACtB;EAAC;EAAA,OAAAV,YAAA;AAAA;AAAAkB,OAAA,CAAAf,OAAA,GAAAH,YAAA"}
|
|
@@ -15,66 +15,64 @@ function _zipBatchIterators() {
|
|
|
15
15
|
_zipBatchIterators = (0, _wrapAsyncGenerator2.default)(_regenerator.default.mark(function _callee(iterator1, iterator2) {
|
|
16
16
|
var batch1, batch2, iterator1Done, iterator2Done, _yield$_awaitAsyncGen, value, done, _yield$_awaitAsyncGen2, _value, _done, batch;
|
|
17
17
|
return _regenerator.default.wrap(function _callee$(_context) {
|
|
18
|
-
while (1) {
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
18
|
+
while (1) switch (_context.prev = _context.next) {
|
|
19
|
+
case 0:
|
|
20
|
+
batch1 = [];
|
|
21
|
+
batch2 = [];
|
|
22
|
+
iterator1Done = false;
|
|
23
|
+
iterator2Done = false;
|
|
24
|
+
case 4:
|
|
25
|
+
if (!(!iterator1Done && !iterator2Done)) {
|
|
26
|
+
_context.next = 27;
|
|
27
|
+
break;
|
|
28
|
+
}
|
|
29
|
+
if (!(batch1.length === 0 && !iterator1Done)) {
|
|
30
|
+
_context.next = 14;
|
|
31
|
+
break;
|
|
32
|
+
}
|
|
33
|
+
_context.next = 8;
|
|
34
|
+
return (0, _awaitAsyncGenerator2.default)(iterator1.next());
|
|
35
|
+
case 8:
|
|
36
|
+
_yield$_awaitAsyncGen = _context.sent;
|
|
37
|
+
value = _yield$_awaitAsyncGen.value;
|
|
38
|
+
done = _yield$_awaitAsyncGen.done;
|
|
39
|
+
if (done) {
|
|
40
|
+
iterator1Done = true;
|
|
41
|
+
} else {
|
|
42
|
+
batch1 = value;
|
|
43
|
+
}
|
|
44
|
+
_context.next = 21;
|
|
45
|
+
break;
|
|
46
|
+
case 14:
|
|
47
|
+
if (!(batch2.length === 0 && !iterator2Done)) {
|
|
45
48
|
_context.next = 21;
|
|
46
49
|
break;
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
}
|
|
63
|
-
case 21:
|
|
64
|
-
batch = extractBatch(batch1, batch2);
|
|
65
|
-
if (!batch) {
|
|
66
|
-
_context.next = 25;
|
|
67
|
-
break;
|
|
68
|
-
}
|
|
50
|
+
}
|
|
51
|
+
_context.next = 17;
|
|
52
|
+
return (0, _awaitAsyncGenerator2.default)(iterator2.next());
|
|
53
|
+
case 17:
|
|
54
|
+
_yield$_awaitAsyncGen2 = _context.sent;
|
|
55
|
+
_value = _yield$_awaitAsyncGen2.value;
|
|
56
|
+
_done = _yield$_awaitAsyncGen2.done;
|
|
57
|
+
if (_done) {
|
|
58
|
+
iterator2Done = true;
|
|
59
|
+
} else {
|
|
60
|
+
batch2 = _value;
|
|
61
|
+
}
|
|
62
|
+
case 21:
|
|
63
|
+
batch = extractBatch(batch1, batch2);
|
|
64
|
+
if (!batch) {
|
|
69
65
|
_context.next = 25;
|
|
70
|
-
return batch;
|
|
71
|
-
case 25:
|
|
72
|
-
_context.next = 4;
|
|
73
66
|
break;
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
67
|
+
}
|
|
68
|
+
_context.next = 25;
|
|
69
|
+
return batch;
|
|
70
|
+
case 25:
|
|
71
|
+
_context.next = 4;
|
|
72
|
+
break;
|
|
73
|
+
case 27:
|
|
74
|
+
case "end":
|
|
75
|
+
return _context.stop();
|
|
78
76
|
}
|
|
79
77
|
}, _callee);
|
|
80
78
|
}));
|
|
@@ -85,9 +83,7 @@ function extractBatch(batch1, batch2) {
|
|
|
85
83
|
if (batchLength === 0) {
|
|
86
84
|
return null;
|
|
87
85
|
}
|
|
88
|
-
|
|
89
86
|
var batch = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];
|
|
90
|
-
|
|
91
87
|
batch1.splice(0, batchLength);
|
|
92
88
|
batch2.splice(0, batchLength);
|
|
93
89
|
return batch;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"zip-batch-iterators.js","names":["zipBatchIterators","iterator1","iterator2","batch1","batch2","iterator1Done","iterator2Done","
|
|
1
|
+
{"version":3,"file":"zip-batch-iterators.js","names":["zipBatchIterators","_x","_x2","_zipBatchIterators","apply","arguments","_wrapAsyncGenerator2","default","_regenerator","mark","_callee","iterator1","iterator2","batch1","batch2","iterator1Done","iterator2Done","_yield$_awaitAsyncGen","value","done","_yield$_awaitAsyncGen2","_value","_done","batch","wrap","_callee$","_context","prev","next","length","_awaitAsyncGenerator2","sent","extractBatch","stop","batchLength","Math","min","slice","splice"],"sources":["../../../../src/lib/streaming/zip-batch-iterators.ts"],"sourcesContent":["/**\n * Zip two iterators together\n *\n * @param iterator1\n * @param iterator2\n */\nexport async function* zipBatchIterators(\n iterator1: AsyncIterator<any[]>,\n iterator2: AsyncIterator<any[]>\n): AsyncGenerator<number[][], void, unknown> {\n let batch1: number[] = [];\n let batch2: number[] = [];\n let iterator1Done: boolean = false;\n let iterator2Done: boolean = false;\n\n // TODO - one could let all iterators flow at full speed using `Promise.race`\n // however we might end up with a big temporary buffer\n while (!iterator1Done && !iterator2Done) {\n if (batch1.length === 0 && !iterator1Done) {\n const {value, done} = await iterator1.next();\n if (done) {\n iterator1Done = true;\n } else {\n batch1 = value;\n }\n } else if (batch2.length === 0 && !iterator2Done) {\n const {value, done} = await iterator2.next();\n if (done) {\n iterator2Done = true;\n } else {\n batch2 = value;\n }\n }\n\n const batch = extractBatch(batch1, batch2);\n if (batch) {\n yield batch;\n }\n }\n}\n\n/**\n * Extract batch of same length from two batches\n *\n * @param batch1\n * @param batch2\n * @return array | null\n */\nfunction extractBatch(batch1: number[], batch2: number[]): number[][] | null {\n const batchLength: number = Math.min(batch1.length, batch2.length);\n if (batchLength === 0) {\n return null;\n }\n\n // Non interleaved arrays\n const batch: number[][] = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];\n\n // Modify the 2 batches\n batch1.splice(0, batchLength);\n batch2.splice(0, batchLength);\n return batch;\n}\n"],"mappings":";;;;;;;;;;SAMuBA,iBAAiBA,CAAAC,EAAA,EAAAC,GAAA;EAAA,OAAAC,kBAAA,CAAAC,KAAA,OAAAC,SAAA;AAAA;AAAA,SAAAF,mBAAA;EAAAA,kBAAA,OAAAG,oBAAA,CAAAC,OAAA,EAAAC,YAAA,CAAAD,OAAA,CAAAE,IAAA,CAAjC,SAAAC,QACLC,SAA+B,EAC/BC,SAA+B;IAAA,IAAAC,MAAA,EAAAC,MAAA,EAAAC,aAAA,EAAAC,aAAA,EAAAC,qBAAA,EAAAC,KAAA,EAAAC,IAAA,EAAAC,sBAAA,EAAAC,MAAA,EAAAC,KAAA,EAAAC,KAAA;IAAA,OAAAf,YAAA,CAAAD,OAAA,CAAAiB,IAAA,UAAAC,SAAAC,QAAA;MAAA,kBAAAA,QAAA,CAAAC,IAAA,GAAAD,QAAA,CAAAE,IAAA;QAAA;UAE3Bf,MAAgB,GAAG,EAAE;UACrBC,MAAgB,GAAG,EAAE;UACrBC,aAAsB,GAAG,KAAK;UAC9BC,aAAsB,GAAG,KAAK;QAAA;UAAA,MAI3B,CAACD,aAAa,IAAI,CAACC,aAAa;YAAAU,QAAA,CAAAE,IAAA;YAAA;UAAA;UAAA,MACjCf,MAAM,CAACgB,MAAM,KAAK,CAAC,IAAI,CAACd,aAAa;YAAAW,QAAA,CAAAE,IAAA;YAAA;UAAA;UAAAF,QAAA,CAAAE,IAAA;UAAA,WAAAE,qBAAA,CAAAvB,OAAA,EACXI,SAAS,CAACiB,IAAI,CAAC,CAAC;QAAA;UAAAX,qBAAA,GAAAS,QAAA,CAAAK,IAAA;UAArCb,KAAK,GAAAD,qBAAA,CAALC,KAAK;UAAEC,IAAI,GAAAF,qBAAA,CAAJE,IAAI;UAClB,IAAIA,IAAI,EAAE;YACRJ,aAAa,GAAG,IAAI;UACtB,CAAC,MAAM;YACLF,MAAM,GAAGK,KAAK;UAChB;UAACQ,QAAA,CAAAE,IAAA;UAAA;QAAA;UAAA,MACQd,MAAM,CAACe,MAAM,KAAK,CAAC,IAAI,CAACb,aAAa;YAAAU,QAAA,CAAAE,IAAA;YAAA;UAAA;UAAAF,QAAA,CAAAE,IAAA;UAAA,WAAAE,qBAAA,CAAAvB,OAAA,EAClBK,SAAS,CAACgB,IAAI,CAAC,CAAC;QAAA;UAAAR,sBAAA,GAAAM,QAAA,CAAAK,IAAA;UAArCb,MAAK,GAAAE,sBAAA,CAALF,KAAK;UAAEC,KAAI,GAAAC,sBAAA,CAAJD,IAAI;UAClB,IAAIA,KAAI,EAAE;YACRH,aAAa,GAAG,IAAI;UACtB,CAAC,MAAM;YACLF,MAAM,GAAGI,MAAK;UAChB;QAAC;UAGGK,KAAK,GAAGS,YAAY,CAACnB,MAAM,EAAEC,MAAM,CAAC;UAAA,KACtCS,KAAK;YAAAG,QAAA,CAAAE,IAAA;YAAA;UAAA;UAAAF,QAAA,CAAAE,IAAA;UACP,OAAML,KAAK;QAAA;UAAAG,QAAA,CAAAE,IAAA;UAAA;QAAA;QAAA;UAAA,OAAAF,QAAA,CAAAO,IAAA;MAAA;IAAA,GAAAvB,OAAA;EAAA,CAGhB;EAAA,OAAAP,kBAAA,CAAAC,KAAA,OAAAC,SAAA;AAAA;AASD,SAAS2B,YAAYA,CAACnB,MAAgB,EAAEC,MAAgB,EAAqB;EAC3E,IAAMoB,WAAmB,GAAGC,IAAI,CAACC,GAAG,CAACvB,MAAM,CAACgB,MAAM,EAAEf,MAAM,CAACe,MAAM,CAAC;EAClE,IAAIK,WAAW,KAAK,CAAC,EAAE;IACrB,OAAO,IAAI;EACb;EAGA,IAAMX,KAAiB,GAAG,CAACV,MAAM,CAACwB,KAAK,CAAC,CAAC,EAAEH,WAAW,CAAC,EAAEpB,MAAM,CAACuB,KAAK,CAAC,CAAC,EAAEH,WAAW,CAAC,CAAC;EAGtFrB,MAAM,CAACyB,MAAM,CAAC,CAAC,EAAEJ,WAAW,CAAC;EAC7BpB,MAAM,CAACwB,MAAM,CAAC,CAAC,EAAEJ,WAAW,CAAC;EAC7B,OAAOX,KAAK;AACd"}
|
|
@@ -6,8 +6,7 @@ Object.defineProperty(exports, "__esModule", {
|
|
|
6
6
|
exports._typecheckShapefileLoader = exports.ShapefileLoader = void 0;
|
|
7
7
|
var _shpLoader = require("./shp-loader");
|
|
8
8
|
var _parseShapefile = require("./lib/parsers/parse-shapefile");
|
|
9
|
-
var VERSION = typeof "3.4.0-alpha.
|
|
10
|
-
|
|
9
|
+
var VERSION = typeof "3.4.0-alpha.4" !== 'undefined' ? "3.4.0-alpha.4" : 'latest';
|
|
11
10
|
var ShapefileLoader = {
|
|
12
11
|
name: 'Shapefile',
|
|
13
12
|
id: 'shapefile',
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"shapefile-loader.js","names":["VERSION","ShapefileLoader","name","id","module","version","category","extensions","mimeTypes","tests","Uint8Array","SHP_MAGIC_NUMBER","buffer","options","shapefile","shp","_maxDimensions","parse","parseShapefile","parseInBatches","parseShapefileInBatches","_typecheckShapefileLoader"],"sources":["../../src/shapefile-loader.ts"],"sourcesContent":["import type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {SHP_MAGIC_NUMBER} from './shp-loader';\nimport {parseShapefile, parseShapefileInBatches} from './lib/parsers/parse-shapefile';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * Shapefile loader\n * @note Shapefile is multifile format and requires providing additional files\n */\nexport const ShapefileLoader = {\n name: 'Shapefile',\n id: 'shapefile',\n module: 'shapefile',\n version: VERSION,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shapefile: {},\n shp: {\n _maxDimensions: 4\n }\n },\n parse: parseShapefile,\n parseInBatches: parseShapefileInBatches\n};\n\nexport const _typecheckShapefileLoader: LoaderWithParser = ShapefileLoader;\n"],"mappings":";;;;;;AACA;AACA;AAIA,
|
|
1
|
+
{"version":3,"file":"shapefile-loader.js","names":["_shpLoader","require","_parseShapefile","VERSION","ShapefileLoader","name","id","module","version","category","extensions","mimeTypes","tests","Uint8Array","SHP_MAGIC_NUMBER","buffer","options","shapefile","shp","_maxDimensions","parse","parseShapefile","parseInBatches","parseShapefileInBatches","exports","_typecheckShapefileLoader"],"sources":["../../src/shapefile-loader.ts"],"sourcesContent":["import type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {SHP_MAGIC_NUMBER} from './shp-loader';\nimport {parseShapefile, parseShapefileInBatches} from './lib/parsers/parse-shapefile';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * Shapefile loader\n * @note Shapefile is multifile format and requires providing additional files\n */\nexport const ShapefileLoader = {\n name: 'Shapefile',\n id: 'shapefile',\n module: 'shapefile',\n version: VERSION,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shapefile: {},\n shp: {\n _maxDimensions: 4\n }\n },\n parse: parseShapefile,\n parseInBatches: parseShapefileInBatches\n};\n\nexport const _typecheckShapefileLoader: LoaderWithParser = ShapefileLoader;\n"],"mappings":";;;;;;AACA,IAAAA,UAAA,GAAAC,OAAA;AACA,IAAAC,eAAA,GAAAD,OAAA;AAIA,IAAME,OAAO,GAAG,sBAAkB,KAAK,WAAW,qBAAiB,QAAQ;AAMpE,IAAMC,eAAe,GAAG;EAC7BC,IAAI,EAAE,WAAW;EACjBC,EAAE,EAAE,WAAW;EACfC,MAAM,EAAE,WAAW;EACnBC,OAAO,EAAEL,OAAO;EAChBM,QAAQ,EAAE,UAAU;EACpBC,UAAU,EAAE,CAAC,KAAK,CAAC;EACnBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EACvCC,KAAK,EAAE,CAAC,IAAIC,UAAU,CAACC,2BAAgB,CAAC,CAACC,MAAM,CAAC;EAChDC,OAAO,EAAE;IACPC,SAAS,EAAE,CAAC,CAAC;IACbC,GAAG,EAAE;MACHC,cAAc,EAAE;IAClB;EACF,CAAC;EACDC,KAAK,EAAEC,8BAAc;EACrBC,cAAc,EAAEC;AAClB,CAAC;AAACC,OAAA,CAAApB,eAAA,GAAAA,eAAA;AAEK,IAAMqB,yBAA2C,GAAGrB,eAAe;AAACoB,OAAA,CAAAC,yBAAA,GAAAA,yBAAA"}
|
package/dist/es5/shp-loader.js
CHANGED
|
@@ -11,9 +11,8 @@ var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/de
|
|
|
11
11
|
var _parseShp = require("./lib/parsers/parse-shp");
|
|
12
12
|
function ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); enumerableOnly && (symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; })), keys.push.apply(keys, symbols); } return keys; }
|
|
13
13
|
function _objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = null != arguments[i] ? arguments[i] : {}; i % 2 ? ownKeys(Object(source), !0).forEach(function (key) { (0, _defineProperty2.default)(target, key, source[key]); }) : Object.getOwnPropertyDescriptors ? Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)) : ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } return target; }
|
|
14
|
-
var VERSION = typeof "3.4.0-alpha.
|
|
14
|
+
var VERSION = typeof "3.4.0-alpha.4" !== 'undefined' ? "3.4.0-alpha.4" : 'latest';
|
|
15
15
|
var SHP_MAGIC_NUMBER = [0x00, 0x00, 0x27, 0x0a];
|
|
16
|
-
|
|
17
16
|
exports.SHP_MAGIC_NUMBER = SHP_MAGIC_NUMBER;
|
|
18
17
|
var SHPWorkerLoader = {
|
|
19
18
|
name: 'SHP',
|
|
@@ -31,20 +30,17 @@ var SHPWorkerLoader = {
|
|
|
31
30
|
}
|
|
32
31
|
}
|
|
33
32
|
};
|
|
34
|
-
|
|
35
33
|
exports.SHPWorkerLoader = SHPWorkerLoader;
|
|
36
34
|
var SHPLoader = _objectSpread(_objectSpread({}, SHPWorkerLoader), {}, {
|
|
37
35
|
parse: function () {
|
|
38
36
|
var _parse = (0, _asyncToGenerator2.default)(_regenerator.default.mark(function _callee(arrayBuffer, options) {
|
|
39
37
|
return _regenerator.default.wrap(function _callee$(_context) {
|
|
40
|
-
while (1) {
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
return _context.stop();
|
|
47
|
-
}
|
|
38
|
+
while (1) switch (_context.prev = _context.next) {
|
|
39
|
+
case 0:
|
|
40
|
+
return _context.abrupt("return", (0, _parseShp.parseSHP)(arrayBuffer, options));
|
|
41
|
+
case 1:
|
|
42
|
+
case "end":
|
|
43
|
+
return _context.stop();
|
|
48
44
|
}
|
|
49
45
|
}, _callee);
|
|
50
46
|
}));
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"shp-loader.js","names":["VERSION","SHP_MAGIC_NUMBER","SHPWorkerLoader","name","id","module","version","worker","category","extensions","mimeTypes","tests","Uint8Array","buffer","options","shp","_maxDimensions","SHPLoader","parse","arrayBuffer","parseSHP","parseSync","parseInBatches","parseSHPInBatches"],"sources":["../../src/shp-loader.ts"],"sourcesContent":["import type {Loader, LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {parseSHP, parseSHPInBatches} from './lib/parsers/parse-shp';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\nexport const SHP_MAGIC_NUMBER = [0x00, 0x00, 0x27, 0x0a];\n\n/**\n * SHP file loader\n */\nexport const SHPWorkerLoader: Loader = {\n name: 'SHP',\n id: 'shp',\n module: 'shapefile',\n version: VERSION,\n worker: true,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n // ISSUE: This also identifies SHX files, which are identical to SHP for the first 100 bytes...\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shp: {\n _maxDimensions: 4\n }\n }\n};\n\n/** SHP file loader */\nexport const SHPLoader: LoaderWithParser = {\n ...SHPWorkerLoader,\n parse: async (arrayBuffer, options?) => parseSHP(arrayBuffer, options),\n parseSync: parseSHP,\n parseInBatches: parseSHPInBatches\n};\n"],"mappings":";;;;;;;;;;AACA;AAAoE;AAAA;AAIpE,
|
|
1
|
+
{"version":3,"file":"shp-loader.js","names":["_parseShp","require","ownKeys","object","enumerableOnly","keys","Object","getOwnPropertySymbols","symbols","filter","sym","getOwnPropertyDescriptor","enumerable","push","apply","_objectSpread","target","i","arguments","length","source","forEach","key","_defineProperty2","default","getOwnPropertyDescriptors","defineProperties","defineProperty","VERSION","SHP_MAGIC_NUMBER","exports","SHPWorkerLoader","name","id","module","version","worker","category","extensions","mimeTypes","tests","Uint8Array","buffer","options","shp","_maxDimensions","SHPLoader","parse","_parse","_asyncToGenerator2","_regenerator","mark","_callee","arrayBuffer","wrap","_callee$","_context","prev","next","abrupt","parseSHP","stop","_x","_x2","parseSync","parseInBatches","parseSHPInBatches"],"sources":["../../src/shp-loader.ts"],"sourcesContent":["import type {Loader, LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {parseSHP, parseSHPInBatches} from './lib/parsers/parse-shp';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\nexport const SHP_MAGIC_NUMBER = [0x00, 0x00, 0x27, 0x0a];\n\n/**\n * SHP file loader\n */\nexport const SHPWorkerLoader: Loader = {\n name: 'SHP',\n id: 'shp',\n module: 'shapefile',\n version: VERSION,\n worker: true,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n // ISSUE: This also identifies SHX files, which are identical to SHP for the first 100 bytes...\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shp: {\n _maxDimensions: 4\n }\n }\n};\n\n/** SHP file loader */\nexport const SHPLoader: LoaderWithParser = {\n ...SHPWorkerLoader,\n parse: async (arrayBuffer, options?) => parseSHP(arrayBuffer, options),\n parseSync: parseSHP,\n parseInBatches: parseSHPInBatches\n};\n"],"mappings":";;;;;;;;;;AACA,IAAAA,SAAA,GAAAC,OAAA;AAAoE,SAAAC,QAAAC,MAAA,EAAAC,cAAA,QAAAC,IAAA,GAAAC,MAAA,CAAAD,IAAA,CAAAF,MAAA,OAAAG,MAAA,CAAAC,qBAAA,QAAAC,OAAA,GAAAF,MAAA,CAAAC,qBAAA,CAAAJ,MAAA,GAAAC,cAAA,KAAAI,OAAA,GAAAA,OAAA,CAAAC,MAAA,WAAAC,GAAA,WAAAJ,MAAA,CAAAK,wBAAA,CAAAR,MAAA,EAAAO,GAAA,EAAAE,UAAA,OAAAP,IAAA,CAAAQ,IAAA,CAAAC,KAAA,CAAAT,IAAA,EAAAG,OAAA,YAAAH,IAAA;AAAA,SAAAU,cAAAC,MAAA,aAAAC,CAAA,MAAAA,CAAA,GAAAC,SAAA,CAAAC,MAAA,EAAAF,CAAA,UAAAG,MAAA,WAAAF,SAAA,CAAAD,CAAA,IAAAC,SAAA,CAAAD,CAAA,QAAAA,CAAA,OAAAf,OAAA,CAAAI,MAAA,CAAAc,MAAA,OAAAC,OAAA,WAAAC,GAAA,QAAAC,gBAAA,CAAAC,OAAA,EAAAR,MAAA,EAAAM,GAAA,EAAAF,MAAA,CAAAE,GAAA,SAAAhB,MAAA,CAAAmB,yBAAA,GAAAnB,MAAA,CAAAoB,gBAAA,CAAAV,MAAA,EAAAV,MAAA,CAAAmB,yBAAA,CAAAL,MAAA,KAAAlB,OAAA,CAAAI,MAAA,CAAAc,MAAA,GAAAC,OAAA,WAAAC,GAAA,IAAAhB,MAAA,CAAAqB,cAAA,CAAAX,MAAA,EAAAM,GAAA,EAAAhB,MAAA,CAAAK,wBAAA,CAAAS,MAAA,EAAAE,GAAA,iBAAAN,MAAA;AAIpE,IAAMY,OAAO,GAAG,sBAAkB,KAAK,WAAW,qBAAiB,QAAQ;AAEpE,IAAMC,gBAAgB,GAAG,CAAC,IAAI,EAAE,IAAI,EAAE,IAAI,EAAE,IAAI,CAAC;AAACC,OAAA,CAAAD,gBAAA,GAAAA,gBAAA;AAKlD,IAAME,eAAuB,GAAG;EACrCC,IAAI,EAAE,KAAK;EACXC,EAAE,EAAE,KAAK;EACTC,MAAM,EAAE,WAAW;EACnBC,OAAO,EAAEP,OAAO;EAChBQ,MAAM,EAAE,IAAI;EACZC,QAAQ,EAAE,UAAU;EACpBC,UAAU,EAAE,CAAC,KAAK,CAAC;EACnBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EAEvCC,KAAK,EAAE,CAAC,IAAIC,UAAU,CAACZ,gBAAgB,CAAC,CAACa,MAAM,CAAC;EAChDC,OAAO,EAAE;IACPC,GAAG,EAAE;MACHC,cAAc,EAAE;IAClB;EACF;AACF,CAAC;AAACf,OAAA,CAAAC,eAAA,GAAAA,eAAA;AAGK,IAAMe,SAA2B,GAAA/B,aAAA,CAAAA,aAAA,KACnCgB,eAAe;EAClBgB,KAAK;IAAA,IAAAC,MAAA,OAAAC,kBAAA,CAAAzB,OAAA,EAAA0B,YAAA,CAAA1B,OAAA,CAAA2B,IAAA,CAAE,SAAAC,QAAOC,WAAW,EAAEV,OAAQ;MAAA,OAAAO,YAAA,CAAA1B,OAAA,CAAA8B,IAAA,UAAAC,SAAAC,QAAA;QAAA,kBAAAA,QAAA,CAAAC,IAAA,GAAAD,QAAA,CAAAE,IAAA;UAAA;YAAA,OAAAF,QAAA,CAAAG,MAAA,WAAK,IAAAC,kBAAQ,EAACP,WAAW,EAAEV,OAAO,CAAC;UAAA;UAAA;YAAA,OAAAa,QAAA,CAAAK,IAAA;QAAA;MAAA,GAAAT,OAAA;IAAA;IAAA,SAAAL,MAAAe,EAAA,EAAAC,GAAA;MAAA,OAAAf,MAAA,CAAAlC,KAAA,OAAAI,SAAA;IAAA;IAAA,OAAA6B,KAAA;EAAA;EACtEiB,SAAS,EAAEJ,kBAAQ;EACnBK,cAAc,EAAEC;AAAiB,EAClC;AAACpC,OAAA,CAAAgB,SAAA,GAAAA,SAAA"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"dbf-worker.js","names":["createLoaderWorker","DBFLoader"],"sources":["../../../src/workers/dbf-worker.ts"],"sourcesContent":["import {DBFLoader} from '../dbf-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(DBFLoader);\n"],"mappings":";;AAAA;AACA;AAEA,
|
|
1
|
+
{"version":3,"file":"dbf-worker.js","names":["_dbfLoader","require","_loaderUtils","createLoaderWorker","DBFLoader"],"sources":["../../../src/workers/dbf-worker.ts"],"sourcesContent":["import {DBFLoader} from '../dbf-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(DBFLoader);\n"],"mappings":";;AAAA,IAAAA,UAAA,GAAAC,OAAA;AACA,IAAAC,YAAA,GAAAD,OAAA;AAEA,IAAAE,+BAAkB,EAACC,oBAAS,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"shp-worker.js","names":["createLoaderWorker","SHPLoader"],"sources":["../../../src/workers/shp-worker.ts"],"sourcesContent":["import {SHPLoader} from '../shp-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(SHPLoader);\n"],"mappings":";;AAAA;AACA;AAEA,
|
|
1
|
+
{"version":3,"file":"shp-worker.js","names":["_shpLoader","require","_loaderUtils","createLoaderWorker","SHPLoader"],"sources":["../../../src/workers/shp-worker.ts"],"sourcesContent":["import {SHPLoader} from '../shp-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(SHPLoader);\n"],"mappings":";;AAAA,IAAAA,UAAA,GAAAC,OAAA;AACA,IAAAC,YAAA,GAAAD,OAAA;AAEA,IAAAE,+BAAkB,EAACC,oBAAS,CAAC"}
|
package/dist/esm/bundle.js
CHANGED
package/dist/esm/bundle.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"bundle.js","names":["moduleExports","require","globalThis","loaders","module","exports","Object","assign"],"sources":["../../src/bundle.ts"],"sourcesContent":["// @ts-nocheck\nconst moduleExports = require('./index');\nglobalThis.loaders = globalThis.loaders || {};\nmodule.exports = Object.assign(globalThis.loaders, moduleExports);\n"],"mappings":"
|
|
1
|
+
{"version":3,"file":"bundle.js","names":["moduleExports","require","globalThis","loaders","module","exports","Object","assign"],"sources":["../../src/bundle.ts"],"sourcesContent":["// @ts-nocheck\nconst moduleExports = require('./index');\nglobalThis.loaders = globalThis.loaders || {};\nmodule.exports = Object.assign(globalThis.loaders, moduleExports);\n"],"mappings":"AACA,MAAMA,aAAa,GAAGC,OAAO,CAAC,SAAS,CAAC;AACxCC,UAAU,CAACC,OAAO,GAAGD,UAAU,CAACC,OAAO,IAAI,CAAC,CAAC;AAC7CC,MAAM,CAACC,OAAO,GAAGC,MAAM,CAACC,MAAM,CAACL,UAAU,CAACC,OAAO,EAAEH,aAAa,CAAC"}
|
package/dist/esm/dbf-loader.js
CHANGED
|
@@ -1,7 +1,5 @@
|
|
|
1
1
|
import { parseDBF, parseDBFInBatches } from './lib/parsers/parse-dbf';
|
|
2
|
-
|
|
3
|
-
const VERSION = typeof "3.4.0-alpha.2" !== 'undefined' ? "3.4.0-alpha.2" : 'latest';
|
|
4
|
-
|
|
2
|
+
const VERSION = typeof "3.4.0-alpha.4" !== 'undefined' ? "3.4.0-alpha.4" : 'latest';
|
|
5
3
|
export const DBFWorkerLoader = {
|
|
6
4
|
name: 'DBF',
|
|
7
5
|
id: 'dbf',
|
|
@@ -17,7 +15,6 @@ export const DBFWorkerLoader = {
|
|
|
17
15
|
}
|
|
18
16
|
}
|
|
19
17
|
};
|
|
20
|
-
|
|
21
18
|
export const DBFLoader = {
|
|
22
19
|
...DBFWorkerLoader,
|
|
23
20
|
parse: async (arrayBuffer, options) => parseDBF(arrayBuffer, options),
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"dbf-loader.js","names":["parseDBF","parseDBFInBatches","VERSION","DBFWorkerLoader","name","id","module","version","worker","category","extensions","mimeTypes","options","dbf","encoding","DBFLoader","parse","arrayBuffer","parseSync","parseInBatches"],"sources":["../../src/dbf-loader.ts"],"sourcesContent":["import type {Loader, LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {parseDBF, parseDBFInBatches} from './lib/parsers/parse-dbf';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * DBFLoader - DBF files are used to contain non-geometry columns in Shapefiles\n */\nexport const DBFWorkerLoader: Loader = {\n name: 'DBF',\n id: 'dbf',\n module: 'shapefile',\n version: VERSION,\n worker: true,\n category: 'table',\n extensions: ['dbf'],\n mimeTypes: ['application/x-dbf'],\n options: {\n dbf: {\n encoding: 'latin1'\n }\n }\n};\n\n/** DBF file loader */\nexport const DBFLoader: LoaderWithParser = {\n ...DBFWorkerLoader,\n parse: async (arrayBuffer, options) => parseDBF(arrayBuffer, options),\n parseSync: parseDBF,\n parseInBatches: parseDBFInBatches\n};\n"],"mappings":"AACA,SAAQA,QAAQ,EAAEC,iBAAiB,QAAO,yBAAyB
|
|
1
|
+
{"version":3,"file":"dbf-loader.js","names":["parseDBF","parseDBFInBatches","VERSION","DBFWorkerLoader","name","id","module","version","worker","category","extensions","mimeTypes","options","dbf","encoding","DBFLoader","parse","arrayBuffer","parseSync","parseInBatches"],"sources":["../../src/dbf-loader.ts"],"sourcesContent":["import type {Loader, LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {parseDBF, parseDBFInBatches} from './lib/parsers/parse-dbf';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * DBFLoader - DBF files are used to contain non-geometry columns in Shapefiles\n */\nexport const DBFWorkerLoader: Loader = {\n name: 'DBF',\n id: 'dbf',\n module: 'shapefile',\n version: VERSION,\n worker: true,\n category: 'table',\n extensions: ['dbf'],\n mimeTypes: ['application/x-dbf'],\n options: {\n dbf: {\n encoding: 'latin1'\n }\n }\n};\n\n/** DBF file loader */\nexport const DBFLoader: LoaderWithParser = {\n ...DBFWorkerLoader,\n parse: async (arrayBuffer, options) => parseDBF(arrayBuffer, options),\n parseSync: parseDBF,\n parseInBatches: parseDBFInBatches\n};\n"],"mappings":"AACA,SAAQA,QAAQ,EAAEC,iBAAiB,QAAO,yBAAyB;AAInE,MAAMC,OAAO,GAAG,sBAAkB,KAAK,WAAW,qBAAiB,QAAQ;AAK3E,OAAO,MAAMC,eAAuB,GAAG;EACrCC,IAAI,EAAE,KAAK;EACXC,EAAE,EAAE,KAAK;EACTC,MAAM,EAAE,WAAW;EACnBC,OAAO,EAAEL,OAAO;EAChBM,MAAM,EAAE,IAAI;EACZC,QAAQ,EAAE,OAAO;EACjBC,UAAU,EAAE,CAAC,KAAK,CAAC;EACnBC,SAAS,EAAE,CAAC,mBAAmB,CAAC;EAChCC,OAAO,EAAE;IACPC,GAAG,EAAE;MACHC,QAAQ,EAAE;IACZ;EACF;AACF,CAAC;AAGD,OAAO,MAAMC,SAA2B,GAAG;EACzC,GAAGZ,eAAe;EAClBa,KAAK,EAAE,MAAAA,CAAOC,WAAW,EAAEL,OAAO,KAAKZ,QAAQ,CAACiB,WAAW,EAAEL,OAAO,CAAC;EACrEM,SAAS,EAAElB,QAAQ;EACnBmB,cAAc,EAAElB;AAClB,CAAC"}
|