@loaders.gl/shapefile 3.3.0-alpha.5 → 3.3.0-alpha.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (68) hide show
  1. package/dist/dbf-worker.js +1 -1
  2. package/dist/dist.min.js +47 -11
  3. package/dist/es5/bundle.js +0 -1
  4. package/dist/es5/bundle.js.map +1 -1
  5. package/dist/es5/dbf-loader.js +5 -16
  6. package/dist/es5/dbf-loader.js.map +1 -1
  7. package/dist/es5/index.js +6 -9
  8. package/dist/es5/index.js.map +1 -1
  9. package/dist/es5/lib/parsers/parse-dbf.js +69 -154
  10. package/dist/es5/lib/parsers/parse-dbf.js.map +1 -1
  11. package/dist/es5/lib/parsers/parse-shapefile.js +34 -148
  12. package/dist/es5/lib/parsers/parse-shapefile.js.map +1 -1
  13. package/dist/es5/lib/parsers/parse-shp-geometry.js +4 -57
  14. package/dist/es5/lib/parsers/parse-shp-geometry.js.map +1 -1
  15. package/dist/es5/lib/parsers/parse-shp-header.js +0 -3
  16. package/dist/es5/lib/parsers/parse-shp-header.js.map +1 -1
  17. package/dist/es5/lib/parsers/parse-shp.js +39 -98
  18. package/dist/es5/lib/parsers/parse-shp.js.map +1 -1
  19. package/dist/es5/lib/parsers/parse-shx.js +0 -4
  20. package/dist/es5/lib/parsers/parse-shx.js.map +1 -1
  21. package/dist/es5/lib/parsers/types.js.map +1 -1
  22. package/dist/es5/lib/streaming/binary-chunk-reader.js +39 -55
  23. package/dist/es5/lib/streaming/binary-chunk-reader.js.map +1 -1
  24. package/dist/es5/lib/streaming/binary-reader.js +11 -12
  25. package/dist/es5/lib/streaming/binary-reader.js.map +1 -1
  26. package/dist/es5/lib/streaming/zip-batch-iterators.js +1 -24
  27. package/dist/es5/lib/streaming/zip-batch-iterators.js.map +1 -1
  28. package/dist/es5/shapefile-loader.js +1 -3
  29. package/dist/es5/shapefile-loader.js.map +1 -1
  30. package/dist/es5/shp-loader.js +6 -17
  31. package/dist/es5/shp-loader.js.map +1 -1
  32. package/dist/es5/workers/dbf-worker.js +0 -2
  33. package/dist/es5/workers/dbf-worker.js.map +1 -1
  34. package/dist/es5/workers/shp-worker.js +0 -2
  35. package/dist/es5/workers/shp-worker.js.map +1 -1
  36. package/dist/esm/bundle.js +1 -1
  37. package/dist/esm/bundle.js.map +1 -1
  38. package/dist/esm/dbf-loader.js +6 -2
  39. package/dist/esm/dbf-loader.js.map +1 -1
  40. package/dist/esm/index.js.map +1 -1
  41. package/dist/esm/lib/parsers/parse-dbf.js +18 -58
  42. package/dist/esm/lib/parsers/parse-dbf.js.map +1 -1
  43. package/dist/esm/lib/parsers/parse-shapefile.js +12 -35
  44. package/dist/esm/lib/parsers/parse-shapefile.js.map +1 -1
  45. package/dist/esm/lib/parsers/parse-shp-geometry.js +3 -28
  46. package/dist/esm/lib/parsers/parse-shp-geometry.js.map +1 -1
  47. package/dist/esm/lib/parsers/parse-shp-header.js +1 -3
  48. package/dist/esm/lib/parsers/parse-shp-header.js.map +1 -1
  49. package/dist/esm/lib/parsers/parse-shp.js +2 -26
  50. package/dist/esm/lib/parsers/parse-shp.js.map +1 -1
  51. package/dist/esm/lib/parsers/parse-shx.js +1 -2
  52. package/dist/esm/lib/parsers/parse-shx.js.map +1 -1
  53. package/dist/esm/lib/parsers/types.js.map +1 -1
  54. package/dist/esm/lib/streaming/binary-chunk-reader.js +4 -19
  55. package/dist/esm/lib/streaming/binary-chunk-reader.js.map +1 -1
  56. package/dist/esm/lib/streaming/binary-reader.js +0 -5
  57. package/dist/esm/lib/streaming/binary-reader.js.map +1 -1
  58. package/dist/esm/lib/streaming/zip-batch-iterators.js +2 -5
  59. package/dist/esm/lib/streaming/zip-batch-iterators.js.map +1 -1
  60. package/dist/esm/shapefile-loader.js +3 -1
  61. package/dist/esm/shapefile-loader.js.map +1 -1
  62. package/dist/esm/shp-loader.js +6 -2
  63. package/dist/esm/shp-loader.js.map +1 -1
  64. package/dist/esm/workers/dbf-worker.js.map +1 -1
  65. package/dist/esm/workers/shp-worker.js.map +1 -1
  66. package/dist/shp-worker.js +1 -1
  67. package/package.json +5 -5
  68. package/src/lib/parsers/parse-dbf.ts +1 -1
@@ -12,17 +12,13 @@ const STATE = {
12
12
  END: 2,
13
13
  ERROR: 3
14
14
  };
15
-
16
15
  class SHPParser {
17
16
  constructor(options) {
18
17
  _defineProperty(this, "options", {});
19
-
20
18
  _defineProperty(this, "binaryReader", new BinaryChunkReader({
21
19
  maxRewindBytes: SHP_RECORD_HEADER_SIZE
22
20
  }));
23
-
24
21
  _defineProperty(this, "state", STATE.EXPECTING_HEADER);
25
-
26
22
  _defineProperty(this, "result", {
27
23
  geometries: [],
28
24
  progress: {
@@ -32,57 +28,47 @@ class SHPParser {
32
28
  },
33
29
  currentIndex: NaN
34
30
  });
35
-
36
31
  this.options = options;
37
32
  }
38
-
39
33
  write(arrayBuffer) {
40
34
  this.binaryReader.write(arrayBuffer);
41
35
  this.state = parseState(this.state, this.result, this.binaryReader, this.options);
42
36
  }
43
-
44
37
  end() {
45
38
  this.binaryReader.end();
46
39
  this.state = parseState(this.state, this.result, this.binaryReader, this.options);
47
-
48
40
  if (this.state !== STATE.END) {
49
41
  this.state = STATE.ERROR;
50
42
  this.result.error = 'SHP incomplete file';
51
43
  }
52
44
  }
53
-
54
45
  }
55
-
56
46
  export function parseSHP(arrayBuffer, options) {
57
47
  const shpParser = new SHPParser(options);
58
48
  shpParser.write(arrayBuffer);
59
49
  shpParser.end();
50
+
60
51
  return shpParser.result;
61
52
  }
53
+
62
54
  export async function* parseSHPInBatches(asyncIterator, options) {
63
55
  const parser = new SHPParser(options);
64
56
  let headerReturned = false;
65
-
66
57
  for await (const arrayBuffer of asyncIterator) {
67
58
  parser.write(arrayBuffer);
68
-
69
59
  if (!headerReturned && parser.result.header) {
70
60
  headerReturned = true;
71
61
  yield parser.result.header;
72
62
  }
73
-
74
63
  if (parser.result.geometries.length > 0) {
75
64
  yield parser.result.geometries;
76
65
  parser.result.geometries = [];
77
66
  }
78
67
  }
79
-
80
68
  parser.end();
81
-
82
69
  if (parser.result.geometries.length > 0) {
83
70
  yield parser.result.geometries;
84
71
  }
85
-
86
72
  return;
87
73
  }
88
74
 
@@ -93,14 +79,11 @@ function parseState(state, result, binaryReader, options) {
93
79
  case STATE.ERROR:
94
80
  case STATE.END:
95
81
  return state;
96
-
97
82
  case STATE.EXPECTING_HEADER:
98
83
  const dataView = binaryReader.getDataView(SHP_HEADER_SIZE);
99
-
100
84
  if (!dataView) {
101
85
  return state;
102
86
  }
103
-
104
87
  result.header = parseSHPHeader(dataView);
105
88
  result.progress = {
106
89
  bytesUsed: 0,
@@ -110,23 +93,19 @@ function parseState(state, result, binaryReader, options) {
110
93
  result.currentIndex = 1;
111
94
  state = STATE.EXPECTING_RECORD;
112
95
  break;
113
-
114
96
  case STATE.EXPECTING_RECORD:
115
97
  while (binaryReader.hasAvailableBytes(SHP_RECORD_HEADER_SIZE)) {
116
98
  var _result$header;
117
-
118
99
  const recordHeaderView = binaryReader.getDataView(SHP_RECORD_HEADER_SIZE);
119
100
  const recordHeader = {
120
101
  recordNumber: recordHeaderView.getInt32(0, BIG_ENDIAN),
121
102
  byteLength: recordHeaderView.getInt32(4, BIG_ENDIAN) * 2,
122
103
  type: recordHeaderView.getInt32(8, LITTLE_ENDIAN)
123
104
  };
124
-
125
105
  if (!binaryReader.hasAvailableBytes(recordHeader.byteLength - 4)) {
126
106
  binaryReader.rewind(SHP_RECORD_HEADER_SIZE);
127
107
  return state;
128
108
  }
129
-
130
109
  const invalidRecord = recordHeader.byteLength < 4 || recordHeader.type !== ((_result$header = result.header) === null || _result$header === void 0 ? void 0 : _result$header.type) || recordHeader.recordNumber !== result.currentIndex;
131
110
 
132
111
  if (invalidRecord) {
@@ -140,13 +119,10 @@ function parseState(state, result, binaryReader, options) {
140
119
  result.progress.rows = result.currentIndex - 1;
141
120
  }
142
121
  }
143
-
144
122
  if (binaryReader.ended) {
145
123
  state = STATE.END;
146
124
  }
147
-
148
125
  return state;
149
-
150
126
  default:
151
127
  state = STATE.ERROR;
152
128
  result.error = "illegal parser state ".concat(state);
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/parsers/parse-shp.ts"],"names":["BinaryChunkReader","parseSHPHeader","parseRecord","LITTLE_ENDIAN","BIG_ENDIAN","SHP_HEADER_SIZE","SHP_RECORD_HEADER_SIZE","STATE","EXPECTING_HEADER","EXPECTING_RECORD","END","ERROR","SHPParser","constructor","options","maxRewindBytes","geometries","progress","bytesTotal","NaN","bytesUsed","rows","currentIndex","write","arrayBuffer","binaryReader","state","parseState","result","end","error","parseSHP","shpParser","parseSHPInBatches","asyncIterator","parser","headerReturned","header","length","dataView","getDataView","hasAvailableBytes","recordHeaderView","recordHeader","recordNumber","getInt32","byteLength","type","rewind","invalidRecord","recordView","geometry","push","ended","message"],"mappings":";AACA,OAAOA,iBAAP,MAA8B,kCAA9B;AACA,SAAQC,cAAR,QAAwC,oBAAxC;AACA,SAAQC,WAAR,QAA0B,sBAA1B;AAGA,MAAMC,aAAa,GAAG,IAAtB;AACA,MAAMC,UAAU,GAAG,KAAnB;AAEA,MAAMC,eAAe,GAAG,GAAxB;AAGA,MAAMC,sBAAsB,GAAG,EAA/B;AAEA,MAAMC,KAAK,GAAG;AACZC,EAAAA,gBAAgB,EAAE,CADN;AAEZC,EAAAA,gBAAgB,EAAE,CAFN;AAGZC,EAAAA,GAAG,EAAE,CAHO;AAIZC,EAAAA,KAAK,EAAE;AAJK,CAAd;;AAmBA,MAAMC,SAAN,CAAgB;AAgBdC,EAAAA,WAAW,CAACC,OAAD,EAA6B;AAAA,qCAfX,EAeW;;AAAA,0CAdzB,IAAId,iBAAJ,CAAsB;AAACe,MAAAA,cAAc,EAAET;AAAjB,KAAtB,CAcyB;;AAAA,mCAbhCC,KAAK,CAACC,gBAa0B;;AAAA,oCAZpB;AAClBQ,MAAAA,UAAU,EAAE,EADM;AAIlBC,MAAAA,QAAQ,EAAE;AACRC,QAAAA,UAAU,EAAEC,GADJ;AAERC,QAAAA,SAAS,EAAED,GAFH;AAGRE,QAAAA,IAAI,EAAEF;AAHE,OAJQ;AASlBG,MAAAA,YAAY,EAAEH;AATI,KAYoB;;AACtC,SAAKL,OAAL,GAAeA,OAAf;AACD;;AAEDS,EAAAA,KAAK,CAACC,WAAD,EAA2B;AAC9B,SAAKC,YAAL,CAAkBF,KAAlB,CAAwBC,WAAxB;AACA,SAAKE,KAAL,GAAaC,UAAU,CAAC,KAAKD,KAAN,EAAa,KAAKE,MAAlB,EAA0B,KAAKH,YAA/B,EAA6C,KAAKX,OAAlD,CAAvB;AACD;;AAEDe,EAAAA,GAAG,GAAG;AACJ,SAAKJ,YAAL,CAAkBI,GAAlB;AACA,SAAKH,KAAL,GAAaC,UAAU,CAAC,KAAKD,KAAN,EAAa,KAAKE,MAAlB,EAA0B,KAAKH,YAA/B,EAA6C,KAAKX,OAAlD,CAAvB;;AAEA,QAAI,KAAKY,KAAL,KAAenB,KAAK,CAACG,GAAzB,EAA8B;AAC5B,WAAKgB,KAAL,GAAanB,KAAK,CAACI,KAAnB;AACA,WAAKiB,MAAL,CAAYE,KAAZ,GAAoB,qBAApB;AACD;AACF;;AAjCa;;AAoChB,OAAO,SAASC,QAAT,CAAkBP,WAAlB,EAA4CV,OAA5C,EAAgF;AACrF,QAAMkB,SAAS,GAAG,IAAIpB,SAAJ,CAAcE,OAAd,CAAlB;AACAkB,EAAAA,SAAS,CAACT,KAAV,CAAgBC,WAAhB;AACAQ,EAAAA,SAAS,CAACH,GAAV;AAGA,SAAOG,SAAS,CAACJ,MAAjB;AACD;AAOD,OAAO,gBAAgBK,iBAAhB,CACLC,aADK,EAELpB,OAFK,EAGmC;AACxC,QAAMqB,MAAM,GAAG,IAAIvB,SAAJ,CAAcE,OAAd,CAAf;AACA,MAAIsB,cAAc,GAAG,KAArB;;AACA,aAAW,MAAMZ,WAAjB,IAAgCU,aAAhC,EAA+C;AAC7CC,IAAAA,MAAM,CAACZ,KAAP,CAAaC,WAAb;;AACA,QAAI,CAACY,cAAD,IAAmBD,MAAM,CAACP,MAAP,CAAcS,MAArC,EAA6C;AAC3CD,MAAAA,cAAc,GAAG,IAAjB;AACA,YAAMD,MAAM,CAACP,MAAP,CAAcS,MAApB;AACD;;AAED,QAAIF,MAAM,CAACP,MAAP,CAAcZ,UAAd,CAAyBsB,MAAzB,GAAkC,CAAtC,EAAyC;AACvC,YAAMH,MAAM,CAACP,MAAP,CAAcZ,UAApB;AACAmB,MAAAA,MAAM,CAACP,MAAP,CAAcZ,UAAd,GAA2B,EAA3B;AACD;AACF;;AACDmB,EAAAA,MAAM,CAACN,GAAP;;AACA,MAAIM,MAAM,CAACP,MAAP,CAAcZ,UAAd,CAAyBsB,MAAzB,GAAkC,CAAtC,EAAyC;AACvC,UAAMH,MAAM,CAACP,MAAP,CAAcZ,UAApB;AACD;;AAED;AACD;;AAgBD,SAASW,UAAT,CACED,KADF,EAEEE,MAFF,EAGEH,YAHF,EAIEX,OAJF,EAKU;AAER,SAAO,IAAP,EAAa;AACX,QAAI;AACF,cAAQY,KAAR;AACE,aAAKnB,KAAK,CAACI,KAAX;AACA,aAAKJ,KAAK,CAACG,GAAX;AACE,iBAAOgB,KAAP;;AAEF,aAAKnB,KAAK,CAACC,gBAAX;AAEE,gBAAM+B,QAAQ,GAAGd,YAAY,CAACe,WAAb,CAAyBnC,eAAzB,CAAjB;;AACA,cAAI,CAACkC,QAAL,EAAe;AACb,mBAAOb,KAAP;AACD;;AAEDE,UAAAA,MAAM,CAACS,MAAP,GAAgBpC,cAAc,CAACsC,QAAD,CAA9B;AACAX,UAAAA,MAAM,CAACX,QAAP,GAAkB;AAChBG,YAAAA,SAAS,EAAE,CADK;AAEhBF,YAAAA,UAAU,EAAEU,MAAM,CAACS,MAAP,CAAcC,MAFV;AAGhBjB,YAAAA,IAAI,EAAE;AAHU,WAAlB;AAMAO,UAAAA,MAAM,CAACN,YAAP,GAAsB,CAAtB;AACAI,UAAAA,KAAK,GAAGnB,KAAK,CAACE,gBAAd;AACA;;AAEF,aAAKF,KAAK,CAACE,gBAAX;AACE,iBAAOgB,YAAY,CAACgB,iBAAb,CAA+BnC,sBAA/B,CAAP,EAA+D;AAAA;;AAC7D,kBAAMoC,gBAAgB,GAAGjB,YAAY,CAACe,WAAb,CAAyBlC,sBAAzB,CAAzB;AACA,kBAAMqC,YAAY,GAAG;AACnBC,cAAAA,YAAY,EAAEF,gBAAgB,CAACG,QAAjB,CAA0B,CAA1B,EAA6BzC,UAA7B,CADK;AAGnB0C,cAAAA,UAAU,EAAEJ,gBAAgB,CAACG,QAAjB,CAA0B,CAA1B,EAA6BzC,UAA7B,IAA2C,CAHpC;AAKnB2C,cAAAA,IAAI,EAAEL,gBAAgB,CAACG,QAAjB,CAA0B,CAA1B,EAA6B1C,aAA7B;AALa,aAArB;;AAQA,gBAAI,CAACsB,YAAY,CAACgB,iBAAb,CAA+BE,YAAY,CAACG,UAAb,GAA0B,CAAzD,CAAL,EAAkE;AAChErB,cAAAA,YAAY,CAACuB,MAAb,CAAoB1C,sBAApB;AACA,qBAAOoB,KAAP;AACD;;AAED,kBAAMuB,aAAa,GACjBN,YAAY,CAACG,UAAb,GAA0B,CAA1B,IACAH,YAAY,CAACI,IAAb,wBAAsBnB,MAAM,CAACS,MAA7B,mDAAsB,eAAeU,IAArC,CADA,IAEAJ,YAAY,CAACC,YAAb,KAA8BhB,MAAM,CAACN,YAHvC;;AAMA,gBAAI2B,aAAJ,EAAmB;AAKjBxB,cAAAA,YAAY,CAACuB,MAAb,CAAoB1C,sBAAsB,GAAG,CAA7C;AACD,aAND,MAMO;AAGLmB,cAAAA,YAAY,CAACuB,MAAb,CAAoB,CAApB;AAEA,oBAAME,UAAU,GAAGzB,YAAY,CAACe,WAAb,CAAyBG,YAAY,CAACG,UAAtC,CAAnB;AACA,oBAAMK,QAAQ,GAAGjD,WAAW,CAACgD,UAAD,EAAapC,OAAb,CAA5B;AACAc,cAAAA,MAAM,CAACZ,UAAP,CAAkBoC,IAAlB,CAAuBD,QAAvB;AAEAvB,cAAAA,MAAM,CAACN,YAAP;AACAM,cAAAA,MAAM,CAACX,QAAP,CAAgBI,IAAhB,GAAuBO,MAAM,CAACN,YAAP,GAAsB,CAA7C;AACD;AACF;;AAED,cAAIG,YAAY,CAAC4B,KAAjB,EAAwB;AACtB3B,YAAAA,KAAK,GAAGnB,KAAK,CAACG,GAAd;AACD;;AAED,iBAAOgB,KAAP;;AAEF;AACEA,UAAAA,KAAK,GAAGnB,KAAK,CAACI,KAAd;AACAiB,UAAAA,MAAM,CAACE,KAAP,kCAAuCJ,KAAvC;AACA,iBAAOA,KAAP;AA1EJ;AA4ED,KA7ED,CA6EE,OAAOI,KAAP,EAAc;AACdJ,MAAAA,KAAK,GAAGnB,KAAK,CAACI,KAAd;AACAiB,MAAAA,MAAM,CAACE,KAAP,iCAAuCA,KAAvC,aAAuCA,KAAvC,uBAAuCA,KAAD,CAAkBwB,OAAxD;AACA,aAAO5B,KAAP;AACD;AACF;AACF","sourcesContent":["import type {BinaryGeometry} from '@loaders.gl/schema';\nimport BinaryChunkReader from '../streaming/binary-chunk-reader';\nimport {parseSHPHeader, SHPHeader} from './parse-shp-header';\nimport {parseRecord} from './parse-shp-geometry';\nimport {SHPLoaderOptions} from './types';\n\nconst LITTLE_ENDIAN = true;\nconst BIG_ENDIAN = false;\n\nconst SHP_HEADER_SIZE = 100;\n// According to the spec, the record header is just 8 bytes, but here we set it\n// to 12 so that we can also access the record's type\nconst SHP_RECORD_HEADER_SIZE = 12;\n\nconst STATE = {\n EXPECTING_HEADER: 0,\n EXPECTING_RECORD: 1,\n END: 2,\n ERROR: 3\n};\n\ntype SHPResult = {\n geometries: (BinaryGeometry | null)[];\n header?: SHPHeader;\n error?: string;\n progress: {\n bytesUsed: number;\n bytesTotal: number;\n rows: number;\n };\n currentIndex: number;\n};\n\nclass SHPParser {\n options?: SHPLoaderOptions = {};\n binaryReader = new BinaryChunkReader({maxRewindBytes: SHP_RECORD_HEADER_SIZE});\n state = STATE.EXPECTING_HEADER;\n result: SHPResult = {\n geometries: [],\n // Initialize with number values to make TS happy\n // These are initialized for real in STATE.EXPECTING_HEADER\n progress: {\n bytesTotal: NaN,\n bytesUsed: NaN,\n rows: NaN\n },\n currentIndex: NaN\n };\n\n constructor(options?: SHPLoaderOptions) {\n this.options = options;\n }\n\n write(arrayBuffer: ArrayBuffer) {\n this.binaryReader.write(arrayBuffer);\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n }\n\n end() {\n this.binaryReader.end();\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n // this.result.progress.bytesUsed = this.binaryReader.bytesUsed();\n if (this.state !== STATE.END) {\n this.state = STATE.ERROR;\n this.result.error = 'SHP incomplete file';\n }\n }\n}\n\nexport function parseSHP(arrayBuffer: ArrayBuffer, options?: object): BinaryGeometry[] {\n const shpParser = new SHPParser(options);\n shpParser.write(arrayBuffer);\n shpParser.end();\n\n // @ts-ignore\n return shpParser.result;\n}\n\n/**\n * @param asyncIterator\n * @param options\n * @returns\n */\nexport async function* parseSHPInBatches(\n asyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options?: object\n): AsyncIterable<BinaryGeometry | object> {\n const parser = new SHPParser(options);\n let headerReturned = false;\n for await (const arrayBuffer of asyncIterator) {\n parser.write(arrayBuffer);\n if (!headerReturned && parser.result.header) {\n headerReturned = true;\n yield parser.result.header;\n }\n\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n parser.result.geometries = [];\n }\n }\n parser.end();\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n }\n\n return;\n}\n\n/**\n * State-machine parser for SHP data\n *\n * Note that whenever more data is needed, a `return`, not a `break`, is\n * necessary, as the `break` keeps the context within `parseState`, while\n * `return` releases context so that more data can be written into the\n * BinaryChunkReader.\n *\n * @param state Current state\n * @param result An object to hold result data\n * @param binaryReader\n * @return State at end of current parsing\n */\n/* eslint-disable complexity, max-depth */\nfunction parseState(\n state: number,\n result: SHPResult,\n binaryReader: BinaryChunkReader,\n options?: SHPLoaderOptions\n): number {\n // eslint-disable-next-line no-constant-condition\n while (true) {\n try {\n switch (state) {\n case STATE.ERROR:\n case STATE.END:\n return state;\n\n case STATE.EXPECTING_HEADER:\n // Parse initial file header\n const dataView = binaryReader.getDataView(SHP_HEADER_SIZE);\n if (!dataView) {\n return state;\n }\n\n result.header = parseSHPHeader(dataView);\n result.progress = {\n bytesUsed: 0,\n bytesTotal: result.header.length,\n rows: 0\n };\n // index numbering starts at 1\n result.currentIndex = 1;\n state = STATE.EXPECTING_RECORD;\n break;\n\n case STATE.EXPECTING_RECORD:\n while (binaryReader.hasAvailableBytes(SHP_RECORD_HEADER_SIZE)) {\n const recordHeaderView = binaryReader.getDataView(SHP_RECORD_HEADER_SIZE) as DataView;\n const recordHeader = {\n recordNumber: recordHeaderView.getInt32(0, BIG_ENDIAN),\n // 2 byte words; includes the four words of record header\n byteLength: recordHeaderView.getInt32(4, BIG_ENDIAN) * 2,\n // This is actually part of the record, not the header...\n type: recordHeaderView.getInt32(8, LITTLE_ENDIAN)\n };\n\n if (!binaryReader.hasAvailableBytes(recordHeader.byteLength - 4)) {\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE);\n return state;\n }\n\n const invalidRecord =\n recordHeader.byteLength < 4 ||\n recordHeader.type !== result.header?.type ||\n recordHeader.recordNumber !== result.currentIndex;\n\n // All records must have at least four bytes (for the record shape type)\n if (invalidRecord) {\n // Malformed record, try again, advancing just 4 bytes\n // Note: this is a rewind because binaryReader.getDataView above\n // moved the pointer forward 12 bytes, so rewinding 8 bytes still\n // leaves us 4 bytes ahead\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE - 4);\n } else {\n // Note: type is actually part of the record, not the header, so\n // rewind 4 bytes before reading record\n binaryReader.rewind(4);\n\n const recordView = binaryReader.getDataView(recordHeader.byteLength) as DataView;\n const geometry = parseRecord(recordView, options);\n result.geometries.push(geometry);\n\n result.currentIndex++;\n result.progress.rows = result.currentIndex - 1;\n }\n }\n\n if (binaryReader.ended) {\n state = STATE.END;\n }\n\n return state;\n\n default:\n state = STATE.ERROR;\n result.error = `illegal parser state ${state}`;\n return state;\n }\n } catch (error) {\n state = STATE.ERROR;\n result.error = `SHP parsing failed: ${(error as Error)?.message}`;\n return state;\n }\n }\n}\n"],"file":"parse-shp.js"}
1
+ {"version":3,"file":"parse-shp.js","names":["BinaryChunkReader","parseSHPHeader","parseRecord","LITTLE_ENDIAN","BIG_ENDIAN","SHP_HEADER_SIZE","SHP_RECORD_HEADER_SIZE","STATE","EXPECTING_HEADER","EXPECTING_RECORD","END","ERROR","SHPParser","constructor","options","maxRewindBytes","geometries","progress","bytesTotal","NaN","bytesUsed","rows","currentIndex","write","arrayBuffer","binaryReader","state","parseState","result","end","error","parseSHP","shpParser","parseSHPInBatches","asyncIterator","parser","headerReturned","header","length","dataView","getDataView","hasAvailableBytes","recordHeaderView","recordHeader","recordNumber","getInt32","byteLength","type","rewind","invalidRecord","recordView","geometry","push","ended","message"],"sources":["../../../../src/lib/parsers/parse-shp.ts"],"sourcesContent":["import type {BinaryGeometry} from '@loaders.gl/schema';\nimport BinaryChunkReader from '../streaming/binary-chunk-reader';\nimport {parseSHPHeader, SHPHeader} from './parse-shp-header';\nimport {parseRecord} from './parse-shp-geometry';\nimport {SHPLoaderOptions} from './types';\n\nconst LITTLE_ENDIAN = true;\nconst BIG_ENDIAN = false;\n\nconst SHP_HEADER_SIZE = 100;\n// According to the spec, the record header is just 8 bytes, but here we set it\n// to 12 so that we can also access the record's type\nconst SHP_RECORD_HEADER_SIZE = 12;\n\nconst STATE = {\n EXPECTING_HEADER: 0,\n EXPECTING_RECORD: 1,\n END: 2,\n ERROR: 3\n};\n\ntype SHPResult = {\n geometries: (BinaryGeometry | null)[];\n header?: SHPHeader;\n error?: string;\n progress: {\n bytesUsed: number;\n bytesTotal: number;\n rows: number;\n };\n currentIndex: number;\n};\n\nclass SHPParser {\n options?: SHPLoaderOptions = {};\n binaryReader = new BinaryChunkReader({maxRewindBytes: SHP_RECORD_HEADER_SIZE});\n state = STATE.EXPECTING_HEADER;\n result: SHPResult = {\n geometries: [],\n // Initialize with number values to make TS happy\n // These are initialized for real in STATE.EXPECTING_HEADER\n progress: {\n bytesTotal: NaN,\n bytesUsed: NaN,\n rows: NaN\n },\n currentIndex: NaN\n };\n\n constructor(options?: SHPLoaderOptions) {\n this.options = options;\n }\n\n write(arrayBuffer: ArrayBuffer) {\n this.binaryReader.write(arrayBuffer);\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n }\n\n end() {\n this.binaryReader.end();\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n // this.result.progress.bytesUsed = this.binaryReader.bytesUsed();\n if (this.state !== STATE.END) {\n this.state = STATE.ERROR;\n this.result.error = 'SHP incomplete file';\n }\n }\n}\n\nexport function parseSHP(arrayBuffer: ArrayBuffer, options?: object): BinaryGeometry[] {\n const shpParser = new SHPParser(options);\n shpParser.write(arrayBuffer);\n shpParser.end();\n\n // @ts-ignore\n return shpParser.result;\n}\n\n/**\n * @param asyncIterator\n * @param options\n * @returns\n */\nexport async function* parseSHPInBatches(\n asyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options?: object\n): AsyncIterable<BinaryGeometry | object> {\n const parser = new SHPParser(options);\n let headerReturned = false;\n for await (const arrayBuffer of asyncIterator) {\n parser.write(arrayBuffer);\n if (!headerReturned && parser.result.header) {\n headerReturned = true;\n yield parser.result.header;\n }\n\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n parser.result.geometries = [];\n }\n }\n parser.end();\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n }\n\n return;\n}\n\n/**\n * State-machine parser for SHP data\n *\n * Note that whenever more data is needed, a `return`, not a `break`, is\n * necessary, as the `break` keeps the context within `parseState`, while\n * `return` releases context so that more data can be written into the\n * BinaryChunkReader.\n *\n * @param state Current state\n * @param result An object to hold result data\n * @param binaryReader\n * @return State at end of current parsing\n */\n/* eslint-disable complexity, max-depth */\nfunction parseState(\n state: number,\n result: SHPResult,\n binaryReader: BinaryChunkReader,\n options?: SHPLoaderOptions\n): number {\n // eslint-disable-next-line no-constant-condition\n while (true) {\n try {\n switch (state) {\n case STATE.ERROR:\n case STATE.END:\n return state;\n\n case STATE.EXPECTING_HEADER:\n // Parse initial file header\n const dataView = binaryReader.getDataView(SHP_HEADER_SIZE);\n if (!dataView) {\n return state;\n }\n\n result.header = parseSHPHeader(dataView);\n result.progress = {\n bytesUsed: 0,\n bytesTotal: result.header.length,\n rows: 0\n };\n // index numbering starts at 1\n result.currentIndex = 1;\n state = STATE.EXPECTING_RECORD;\n break;\n\n case STATE.EXPECTING_RECORD:\n while (binaryReader.hasAvailableBytes(SHP_RECORD_HEADER_SIZE)) {\n const recordHeaderView = binaryReader.getDataView(SHP_RECORD_HEADER_SIZE) as DataView;\n const recordHeader = {\n recordNumber: recordHeaderView.getInt32(0, BIG_ENDIAN),\n // 2 byte words; includes the four words of record header\n byteLength: recordHeaderView.getInt32(4, BIG_ENDIAN) * 2,\n // This is actually part of the record, not the header...\n type: recordHeaderView.getInt32(8, LITTLE_ENDIAN)\n };\n\n if (!binaryReader.hasAvailableBytes(recordHeader.byteLength - 4)) {\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE);\n return state;\n }\n\n const invalidRecord =\n recordHeader.byteLength < 4 ||\n recordHeader.type !== result.header?.type ||\n recordHeader.recordNumber !== result.currentIndex;\n\n // All records must have at least four bytes (for the record shape type)\n if (invalidRecord) {\n // Malformed record, try again, advancing just 4 bytes\n // Note: this is a rewind because binaryReader.getDataView above\n // moved the pointer forward 12 bytes, so rewinding 8 bytes still\n // leaves us 4 bytes ahead\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE - 4);\n } else {\n // Note: type is actually part of the record, not the header, so\n // rewind 4 bytes before reading record\n binaryReader.rewind(4);\n\n const recordView = binaryReader.getDataView(recordHeader.byteLength) as DataView;\n const geometry = parseRecord(recordView, options);\n result.geometries.push(geometry);\n\n result.currentIndex++;\n result.progress.rows = result.currentIndex - 1;\n }\n }\n\n if (binaryReader.ended) {\n state = STATE.END;\n }\n\n return state;\n\n default:\n state = STATE.ERROR;\n result.error = `illegal parser state ${state}`;\n return state;\n }\n } catch (error) {\n state = STATE.ERROR;\n result.error = `SHP parsing failed: ${(error as Error)?.message}`;\n return state;\n }\n }\n}\n"],"mappings":";AACA,OAAOA,iBAAiB,MAAM,kCAAkC;AAChE,SAAQC,cAAc,QAAkB,oBAAoB;AAC5D,SAAQC,WAAW,QAAO,sBAAsB;AAGhD,MAAMC,aAAa,GAAG,IAAI;AAC1B,MAAMC,UAAU,GAAG,KAAK;AAExB,MAAMC,eAAe,GAAG,GAAG;AAG3B,MAAMC,sBAAsB,GAAG,EAAE;AAEjC,MAAMC,KAAK,GAAG;EACZC,gBAAgB,EAAE,CAAC;EACnBC,gBAAgB,EAAE,CAAC;EACnBC,GAAG,EAAE,CAAC;EACNC,KAAK,EAAE;AACT,CAAC;AAcD,MAAMC,SAAS,CAAC;EAgBdC,WAAW,CAACC,OAA0B,EAAE;IAAA,iCAfX,CAAC,CAAC;IAAA,sCAChB,IAAId,iBAAiB,CAAC;MAACe,cAAc,EAAET;IAAsB,CAAC,CAAC;IAAA,+BACtEC,KAAK,CAACC,gBAAgB;IAAA,gCACV;MAClBQ,UAAU,EAAE,EAAE;MAGdC,QAAQ,EAAE;QACRC,UAAU,EAAEC,GAAG;QACfC,SAAS,EAAED,GAAG;QACdE,IAAI,EAAEF;MACR,CAAC;MACDG,YAAY,EAAEH;IAChB,CAAC;IAGC,IAAI,CAACL,OAAO,GAAGA,OAAO;EACxB;EAEAS,KAAK,CAACC,WAAwB,EAAE;IAC9B,IAAI,CAACC,YAAY,CAACF,KAAK,CAACC,WAAW,CAAC;IACpC,IAAI,CAACE,KAAK,GAAGC,UAAU,CAAC,IAAI,CAACD,KAAK,EAAE,IAAI,CAACE,MAAM,EAAE,IAAI,CAACH,YAAY,EAAE,IAAI,CAACX,OAAO,CAAC;EACnF;EAEAe,GAAG,GAAG;IACJ,IAAI,CAACJ,YAAY,CAACI,GAAG,EAAE;IACvB,IAAI,CAACH,KAAK,GAAGC,UAAU,CAAC,IAAI,CAACD,KAAK,EAAE,IAAI,CAACE,MAAM,EAAE,IAAI,CAACH,YAAY,EAAE,IAAI,CAACX,OAAO,CAAC;IAEjF,IAAI,IAAI,CAACY,KAAK,KAAKnB,KAAK,CAACG,GAAG,EAAE;MAC5B,IAAI,CAACgB,KAAK,GAAGnB,KAAK,CAACI,KAAK;MACxB,IAAI,CAACiB,MAAM,CAACE,KAAK,GAAG,qBAAqB;IAC3C;EACF;AACF;AAEA,OAAO,SAASC,QAAQ,CAACP,WAAwB,EAAEV,OAAgB,EAAoB;EACrF,MAAMkB,SAAS,GAAG,IAAIpB,SAAS,CAACE,OAAO,CAAC;EACxCkB,SAAS,CAACT,KAAK,CAACC,WAAW,CAAC;EAC5BQ,SAAS,CAACH,GAAG,EAAE;;EAGf,OAAOG,SAAS,CAACJ,MAAM;AACzB;;AAOA,OAAO,gBAAgBK,iBAAiB,CACtCC,aAAiE,EACjEpB,OAAgB,EACwB;EACxC,MAAMqB,MAAM,GAAG,IAAIvB,SAAS,CAACE,OAAO,CAAC;EACrC,IAAIsB,cAAc,GAAG,KAAK;EAC1B,WAAW,MAAMZ,WAAW,IAAIU,aAAa,EAAE;IAC7CC,MAAM,CAACZ,KAAK,CAACC,WAAW,CAAC;IACzB,IAAI,CAACY,cAAc,IAAID,MAAM,CAACP,MAAM,CAACS,MAAM,EAAE;MAC3CD,cAAc,GAAG,IAAI;MACrB,MAAMD,MAAM,CAACP,MAAM,CAACS,MAAM;IAC5B;IAEA,IAAIF,MAAM,CAACP,MAAM,CAACZ,UAAU,CAACsB,MAAM,GAAG,CAAC,EAAE;MACvC,MAAMH,MAAM,CAACP,MAAM,CAACZ,UAAU;MAC9BmB,MAAM,CAACP,MAAM,CAACZ,UAAU,GAAG,EAAE;IAC/B;EACF;EACAmB,MAAM,CAACN,GAAG,EAAE;EACZ,IAAIM,MAAM,CAACP,MAAM,CAACZ,UAAU,CAACsB,MAAM,GAAG,CAAC,EAAE;IACvC,MAAMH,MAAM,CAACP,MAAM,CAACZ,UAAU;EAChC;EAEA;AACF;;AAgBA,SAASW,UAAU,CACjBD,KAAa,EACbE,MAAiB,EACjBH,YAA+B,EAC/BX,OAA0B,EAClB;EAER,OAAO,IAAI,EAAE;IACX,IAAI;MACF,QAAQY,KAAK;QACX,KAAKnB,KAAK,CAACI,KAAK;QAChB,KAAKJ,KAAK,CAACG,GAAG;UACZ,OAAOgB,KAAK;QAEd,KAAKnB,KAAK,CAACC,gBAAgB;UAEzB,MAAM+B,QAAQ,GAAGd,YAAY,CAACe,WAAW,CAACnC,eAAe,CAAC;UAC1D,IAAI,CAACkC,QAAQ,EAAE;YACb,OAAOb,KAAK;UACd;UAEAE,MAAM,CAACS,MAAM,GAAGpC,cAAc,CAACsC,QAAQ,CAAC;UACxCX,MAAM,CAACX,QAAQ,GAAG;YAChBG,SAAS,EAAE,CAAC;YACZF,UAAU,EAAEU,MAAM,CAACS,MAAM,CAACC,MAAM;YAChCjB,IAAI,EAAE;UACR,CAAC;UAEDO,MAAM,CAACN,YAAY,GAAG,CAAC;UACvBI,KAAK,GAAGnB,KAAK,CAACE,gBAAgB;UAC9B;QAEF,KAAKF,KAAK,CAACE,gBAAgB;UACzB,OAAOgB,YAAY,CAACgB,iBAAiB,CAACnC,sBAAsB,CAAC,EAAE;YAAA;YAC7D,MAAMoC,gBAAgB,GAAGjB,YAAY,CAACe,WAAW,CAAClC,sBAAsB,CAAa;YACrF,MAAMqC,YAAY,GAAG;cACnBC,YAAY,EAAEF,gBAAgB,CAACG,QAAQ,CAAC,CAAC,EAAEzC,UAAU,CAAC;cAEtD0C,UAAU,EAAEJ,gBAAgB,CAACG,QAAQ,CAAC,CAAC,EAAEzC,UAAU,CAAC,GAAG,CAAC;cAExD2C,IAAI,EAAEL,gBAAgB,CAACG,QAAQ,CAAC,CAAC,EAAE1C,aAAa;YAClD,CAAC;YAED,IAAI,CAACsB,YAAY,CAACgB,iBAAiB,CAACE,YAAY,CAACG,UAAU,GAAG,CAAC,CAAC,EAAE;cAChErB,YAAY,CAACuB,MAAM,CAAC1C,sBAAsB,CAAC;cAC3C,OAAOoB,KAAK;YACd;YAEA,MAAMuB,aAAa,GACjBN,YAAY,CAACG,UAAU,GAAG,CAAC,IAC3BH,YAAY,CAACI,IAAI,wBAAKnB,MAAM,CAACS,MAAM,mDAAb,eAAeU,IAAI,KACzCJ,YAAY,CAACC,YAAY,KAAKhB,MAAM,CAACN,YAAY;;YAGnD,IAAI2B,aAAa,EAAE;cAKjBxB,YAAY,CAACuB,MAAM,CAAC1C,sBAAsB,GAAG,CAAC,CAAC;YACjD,CAAC,MAAM;cAGLmB,YAAY,CAACuB,MAAM,CAAC,CAAC,CAAC;cAEtB,MAAME,UAAU,GAAGzB,YAAY,CAACe,WAAW,CAACG,YAAY,CAACG,UAAU,CAAa;cAChF,MAAMK,QAAQ,GAAGjD,WAAW,CAACgD,UAAU,EAAEpC,OAAO,CAAC;cACjDc,MAAM,CAACZ,UAAU,CAACoC,IAAI,CAACD,QAAQ,CAAC;cAEhCvB,MAAM,CAACN,YAAY,EAAE;cACrBM,MAAM,CAACX,QAAQ,CAACI,IAAI,GAAGO,MAAM,CAACN,YAAY,GAAG,CAAC;YAChD;UACF;UAEA,IAAIG,YAAY,CAAC4B,KAAK,EAAE;YACtB3B,KAAK,GAAGnB,KAAK,CAACG,GAAG;UACnB;UAEA,OAAOgB,KAAK;QAEd;UACEA,KAAK,GAAGnB,KAAK,CAACI,KAAK;UACnBiB,MAAM,CAACE,KAAK,kCAA2BJ,KAAK,CAAE;UAC9C,OAAOA,KAAK;MAAC;IAEnB,CAAC,CAAC,OAAOI,KAAK,EAAE;MACdJ,KAAK,GAAGnB,KAAK,CAACI,KAAK;MACnBiB,MAAM,CAACE,KAAK,iCAA2BA,KAAK,aAALA,KAAK,uBAALA,KAAK,CAAYwB,OAAO,CAAE;MACjE,OAAO5B,KAAK;IACd;EACF;AACF"}
@@ -1,6 +1,7 @@
1
1
  import { parseSHPHeader } from './parse-shp-header';
2
2
  const SHX_HEADER_SIZE = 100;
3
3
  const BIG_ENDIAN = false;
4
+
4
5
  export function parseShx(arrayBuffer) {
5
6
  const headerView = new DataView(arrayBuffer, 0, SHX_HEADER_SIZE);
6
7
  const header = parseSHPHeader(headerView);
@@ -8,12 +9,10 @@ export function parseShx(arrayBuffer) {
8
9
  const contentView = new DataView(arrayBuffer, SHX_HEADER_SIZE, contentLength);
9
10
  const offsets = new Int32Array(contentLength);
10
11
  const lengths = new Int32Array(contentLength);
11
-
12
12
  for (let i = 0; i < contentLength / 8; i++) {
13
13
  offsets[i] = contentView.getInt32(i * 8, BIG_ENDIAN);
14
14
  lengths[i] = contentView.getInt32(i * 8 + 4, BIG_ENDIAN);
15
15
  }
16
-
17
16
  return {
18
17
  offsets,
19
18
  lengths
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/parsers/parse-shx.ts"],"names":["parseSHPHeader","SHX_HEADER_SIZE","BIG_ENDIAN","parseShx","arrayBuffer","headerView","DataView","header","contentLength","length","contentView","offsets","Int32Array","lengths","i","getInt32"],"mappings":"AAAA,SAAQA,cAAR,QAA6B,oBAA7B;AAOA,MAAMC,eAAe,GAAG,GAAxB;AACA,MAAMC,UAAU,GAAG,KAAnB;AAMA,OAAO,SAASC,QAAT,CAAkBC,WAAlB,EAAuD;AAE5D,QAAMC,UAAU,GAAG,IAAIC,QAAJ,CAAaF,WAAb,EAA0B,CAA1B,EAA6BH,eAA7B,CAAnB;AACA,QAAMM,MAAM,GAAGP,cAAc,CAACK,UAAD,CAA7B;AACA,QAAMG,aAAa,GAAGD,MAAM,CAACE,MAAP,GAAgBR,eAAtC;AAEA,QAAMS,WAAW,GAAG,IAAIJ,QAAJ,CAAaF,WAAb,EAA0BH,eAA1B,EAA2CO,aAA3C,CAApB;AAEA,QAAMG,OAAO,GAAG,IAAIC,UAAJ,CAAeJ,aAAf,CAAhB;AACA,QAAMK,OAAO,GAAG,IAAID,UAAJ,CAAeJ,aAAf,CAAhB;;AAEA,OAAK,IAAIM,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGN,aAAa,GAAG,CAApC,EAAuCM,CAAC,EAAxC,EAA4C;AAC1CH,IAAAA,OAAO,CAACG,CAAD,CAAP,GAAaJ,WAAW,CAACK,QAAZ,CAAqBD,CAAC,GAAG,CAAzB,EAA4BZ,UAA5B,CAAb;AACAW,IAAAA,OAAO,CAACC,CAAD,CAAP,GAAaJ,WAAW,CAACK,QAAZ,CAAqBD,CAAC,GAAG,CAAJ,GAAQ,CAA7B,EAAgCZ,UAAhC,CAAb;AACD;;AAED,SAAO;AACLS,IAAAA,OADK;AAELE,IAAAA;AAFK,GAAP;AAID","sourcesContent":["import {parseSHPHeader} from './parse-shp-header';\n\nexport interface SHXOutput {\n offsets: Int32Array;\n lengths: Int32Array;\n}\n\nconst SHX_HEADER_SIZE = 100;\nconst BIG_ENDIAN = false;\n\n/**\n * @param arrayBuffer\n * @returns SHXOutput\n */\nexport function parseShx(arrayBuffer: ArrayBuffer): SHXOutput {\n // SHX header is identical to SHP Header\n const headerView = new DataView(arrayBuffer, 0, SHX_HEADER_SIZE);\n const header = parseSHPHeader(headerView);\n const contentLength = header.length - SHX_HEADER_SIZE;\n\n const contentView = new DataView(arrayBuffer, SHX_HEADER_SIZE, contentLength);\n\n const offsets = new Int32Array(contentLength);\n const lengths = new Int32Array(contentLength);\n\n for (let i = 0; i < contentLength / 8; i++) {\n offsets[i] = contentView.getInt32(i * 8, BIG_ENDIAN);\n lengths[i] = contentView.getInt32(i * 8 + 4, BIG_ENDIAN);\n }\n\n return {\n offsets,\n lengths\n };\n}\n"],"file":"parse-shx.js"}
1
+ {"version":3,"file":"parse-shx.js","names":["parseSHPHeader","SHX_HEADER_SIZE","BIG_ENDIAN","parseShx","arrayBuffer","headerView","DataView","header","contentLength","length","contentView","offsets","Int32Array","lengths","i","getInt32"],"sources":["../../../../src/lib/parsers/parse-shx.ts"],"sourcesContent":["import {parseSHPHeader} from './parse-shp-header';\n\nexport interface SHXOutput {\n offsets: Int32Array;\n lengths: Int32Array;\n}\n\nconst SHX_HEADER_SIZE = 100;\nconst BIG_ENDIAN = false;\n\n/**\n * @param arrayBuffer\n * @returns SHXOutput\n */\nexport function parseShx(arrayBuffer: ArrayBuffer): SHXOutput {\n // SHX header is identical to SHP Header\n const headerView = new DataView(arrayBuffer, 0, SHX_HEADER_SIZE);\n const header = parseSHPHeader(headerView);\n const contentLength = header.length - SHX_HEADER_SIZE;\n\n const contentView = new DataView(arrayBuffer, SHX_HEADER_SIZE, contentLength);\n\n const offsets = new Int32Array(contentLength);\n const lengths = new Int32Array(contentLength);\n\n for (let i = 0; i < contentLength / 8; i++) {\n offsets[i] = contentView.getInt32(i * 8, BIG_ENDIAN);\n lengths[i] = contentView.getInt32(i * 8 + 4, BIG_ENDIAN);\n }\n\n return {\n offsets,\n lengths\n };\n}\n"],"mappings":"AAAA,SAAQA,cAAc,QAAO,oBAAoB;AAOjD,MAAMC,eAAe,GAAG,GAAG;AAC3B,MAAMC,UAAU,GAAG,KAAK;;AAMxB,OAAO,SAASC,QAAQ,CAACC,WAAwB,EAAa;EAE5D,MAAMC,UAAU,GAAG,IAAIC,QAAQ,CAACF,WAAW,EAAE,CAAC,EAAEH,eAAe,CAAC;EAChE,MAAMM,MAAM,GAAGP,cAAc,CAACK,UAAU,CAAC;EACzC,MAAMG,aAAa,GAAGD,MAAM,CAACE,MAAM,GAAGR,eAAe;EAErD,MAAMS,WAAW,GAAG,IAAIJ,QAAQ,CAACF,WAAW,EAAEH,eAAe,EAAEO,aAAa,CAAC;EAE7E,MAAMG,OAAO,GAAG,IAAIC,UAAU,CAACJ,aAAa,CAAC;EAC7C,MAAMK,OAAO,GAAG,IAAID,UAAU,CAACJ,aAAa,CAAC;EAE7C,KAAK,IAAIM,CAAC,GAAG,CAAC,EAAEA,CAAC,GAAGN,aAAa,GAAG,CAAC,EAAEM,CAAC,EAAE,EAAE;IAC1CH,OAAO,CAACG,CAAC,CAAC,GAAGJ,WAAW,CAACK,QAAQ,CAACD,CAAC,GAAG,CAAC,EAAEZ,UAAU,CAAC;IACpDW,OAAO,CAACC,CAAC,CAAC,GAAGJ,WAAW,CAACK,QAAQ,CAACD,CAAC,GAAG,CAAC,GAAG,CAAC,EAAEZ,UAAU,CAAC;EAC1D;EAEA,OAAO;IACLS,OAAO;IACPE;EACF,CAAC;AACH"}
@@ -1 +1 @@
1
- {"version":3,"sources":[],"names":[],"mappings":"","sourcesContent":[],"file":"types.js"}
1
+ {"version":3,"file":"types.js","names":[],"sources":["../../../../src/lib/parsers/types.ts"],"sourcesContent":["import {Schema, ObjectRowTable} from '@loaders.gl/schema';\nimport type {LoaderOptions} from '@loaders.gl/loader-utils';\n\nexport type SHPLoaderOptions = LoaderOptions & {\n shp?: {\n _maxDimensions?: number;\n };\n};\n\nexport type DBFLoaderOptions = LoaderOptions & {\n dbf?: {\n encoding?: string;\n shape?: 'rows' | 'table' | 'object-row-table';\n };\n};\n\nexport type ShapefileLoaderOptions = LoaderOptions &\n SHPLoaderOptions & {\n shapefile?: {\n shape?: 'geojson';\n };\n gis?: {\n reproject?: boolean;\n _targetCrs?: string;\n /** @deprecated. Use options.shapefile.shape */\n format?: 'geojson';\n };\n };\n\nexport type DBFRowsOutput = ObjectRowTable['data'];\n\n/**\n * DBF Table output. Deprecated in favor of ObjectRowTable\n * @deprecated\n */\nexport interface DBFTableOutput {\n schema?: Schema;\n rows: DBFRowsOutput;\n}\n\nexport type DBFHeader = {\n // Last updated date\n year: number;\n month: number;\n day: number;\n // Number of records in data file\n nRecords: number;\n // Length of header in bytes\n headerLength: number;\n // Length of each record\n recordLength: number;\n // Not sure if this is usually set\n languageDriver: number;\n};\n\nexport type DBFField = {\n name: string;\n dataType: string;\n fieldLength: number;\n decimal: number;\n};\n\nexport type DBFResult = {\n data: {[key: string]: any}[];\n schema?: Schema;\n error?: string;\n dbfHeader?: DBFHeader;\n dbfFields?: DBFField[];\n progress?: {\n bytesUsed: number;\n rowsTotal: number;\n rows: number;\n };\n};\n"],"mappings":""}
@@ -2,26 +2,22 @@ import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
2
  export default class BinaryChunkReader {
3
3
  constructor(options) {
4
4
  _defineProperty(this, "offset", void 0);
5
-
6
5
  _defineProperty(this, "arrayBuffers", void 0);
7
-
8
6
  _defineProperty(this, "ended", void 0);
9
-
10
7
  _defineProperty(this, "maxRewindBytes", void 0);
11
-
12
8
  const {
13
9
  maxRewindBytes = 0
14
10
  } = options || {};
11
+
15
12
  this.offset = 0;
16
13
  this.arrayBuffers = [];
17
14
  this.ended = false;
15
+
18
16
  this.maxRewindBytes = maxRewindBytes;
19
17
  }
20
-
21
18
  write(arrayBuffer) {
22
19
  this.arrayBuffers.push(arrayBuffer);
23
20
  }
24
-
25
21
  end() {
26
22
  this.arrayBuffers = [];
27
23
  this.ended = true;
@@ -29,22 +25,18 @@ export default class BinaryChunkReader {
29
25
 
30
26
  hasAvailableBytes(bytes) {
31
27
  let bytesAvailable = -this.offset;
32
-
33
28
  for (const arrayBuffer of this.arrayBuffers) {
34
29
  bytesAvailable += arrayBuffer.byteLength;
35
-
36
30
  if (bytesAvailable >= bytes) {
37
31
  return true;
38
32
  }
39
33
  }
40
-
41
34
  return false;
42
35
  }
43
36
 
44
37
  findBufferOffsets(bytes) {
45
38
  let offset = -this.offset;
46
39
  const selectedBuffers = [];
47
-
48
40
  for (let i = 0; i < this.arrayBuffers.length; i++) {
49
41
  const buf = this.arrayBuffers[i];
50
42
 
@@ -64,6 +56,7 @@ export default class BinaryChunkReader {
64
56
 
65
57
  end = buf.byteLength;
66
58
  selectedBuffers.push([i, [start, end]]);
59
+
67
60
  bytes -= buf.byteLength - start;
68
61
  offset += buf.byteLength;
69
62
  }
@@ -73,11 +66,9 @@ export default class BinaryChunkReader {
73
66
 
74
67
  getDataView(bytes) {
75
68
  const bufferOffsets = this.findBufferOffsets(bytes);
76
-
77
69
  if (!bufferOffsets && this.ended) {
78
70
  throw new Error('binary data exhausted');
79
71
  }
80
-
81
72
  if (!bufferOffsets) {
82
73
  return null;
83
74
  }
@@ -106,32 +97,26 @@ export default class BinaryChunkReader {
106
97
 
107
98
  _combineArrayBuffers(bufferOffsets) {
108
99
  let byteLength = 0;
109
-
110
100
  for (const bufferOffset of bufferOffsets) {
111
101
  const [start, end] = bufferOffset[1];
112
102
  byteLength += end - start;
113
103
  }
114
-
115
104
  const result = new Uint8Array(byteLength);
116
- let resultOffset = 0;
117
105
 
106
+ let resultOffset = 0;
118
107
  for (const bufferOffset of bufferOffsets) {
119
108
  const [bufferIndex, [start, end]] = bufferOffset;
120
109
  const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);
121
110
  result.set(sourceArray.subarray(start, end), resultOffset);
122
111
  resultOffset += end - start;
123
112
  }
124
-
125
113
  return result.buffer;
126
114
  }
127
-
128
115
  skip(bytes) {
129
116
  this.offset += bytes;
130
117
  }
131
-
132
118
  rewind(bytes) {
133
119
  this.offset -= bytes;
134
120
  }
135
-
136
121
  }
137
122
  //# sourceMappingURL=binary-chunk-reader.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/streaming/binary-chunk-reader.ts"],"names":["BinaryChunkReader","constructor","options","maxRewindBytes","offset","arrayBuffers","ended","write","arrayBuffer","push","end","hasAvailableBytes","bytes","bytesAvailable","byteLength","findBufferOffsets","selectedBuffers","i","length","buf","start","Math","abs","getDataView","bufferOffsets","Error","bufferIndex","view","DataView","disposeBuffers","_combineArrayBuffers","shift","bufferOffset","result","Uint8Array","resultOffset","sourceArray","set","subarray","buffer","skip","rewind"],"mappings":";AAIA,eAAe,MAAMA,iBAAN,CAAwB;AAMrCC,EAAAA,WAAW,CAACC,OAAD,EAAqC;AAAA;;AAAA;;AAAA;;AAAA;;AAC9C,UAAM;AAACC,MAAAA,cAAc,GAAG;AAAlB,QAAuBD,OAAO,IAAI,EAAxC;AAGA,SAAKE,MAAL,GAAc,CAAd;AAEA,SAAKC,YAAL,GAAoB,EAApB;AACA,SAAKC,KAAL,GAAa,KAAb;AAGA,SAAKH,cAAL,GAAsBA,cAAtB;AACD;;AAIDI,EAAAA,KAAK,CAACC,WAAD,EAAiC;AACpC,SAAKH,YAAL,CAAkBI,IAAlB,CAAuBD,WAAvB;AACD;;AAEDE,EAAAA,GAAG,GAAS;AACV,SAAKL,YAAL,GAAoB,EAApB;AACA,SAAKC,KAAL,GAAa,IAAb;AACD;;AAQDK,EAAAA,iBAAiB,CAACC,KAAD,EAAyB;AACxC,QAAIC,cAAc,GAAG,CAAC,KAAKT,MAA3B;;AACA,SAAK,MAAMI,WAAX,IAA0B,KAAKH,YAA/B,EAA6C;AAC3CQ,MAAAA,cAAc,IAAIL,WAAW,CAACM,UAA9B;;AACA,UAAID,cAAc,IAAID,KAAtB,EAA6B;AAC3B,eAAO,IAAP;AACD;AACF;;AACD,WAAO,KAAP;AACD;;AAQDG,EAAAA,iBAAiB,CAACH,KAAD,EAA8B;AAC7C,QAAIR,MAAM,GAAG,CAAC,KAAKA,MAAnB;AACA,UAAMY,eAAoB,GAAG,EAA7B;;AAEA,SAAK,IAAIC,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAG,KAAKZ,YAAL,CAAkBa,MAAtC,EAA8CD,CAAC,EAA/C,EAAmD;AACjD,YAAME,GAAG,GAAG,KAAKd,YAAL,CAAkBY,CAAlB,CAAZ;;AAGA,UAAIb,MAAM,GAAGe,GAAG,CAACL,UAAb,IAA2B,CAA/B,EAAkC;AAChCV,QAAAA,MAAM,IAAIe,GAAG,CAACL,UAAd;AAEA;AACD;;AAKD,YAAMM,KAAK,GAAGhB,MAAM,IAAI,CAAV,GAAciB,IAAI,CAACC,GAAL,CAASlB,MAAT,CAAd,GAAiC,CAA/C;AACA,UAAIM,GAAJ;;AAGA,UAAIU,KAAK,GAAGR,KAAR,IAAiBO,GAAG,CAACL,UAAzB,EAAqC;AACnCJ,QAAAA,GAAG,GAAGU,KAAK,GAAGR,KAAd;AACAI,QAAAA,eAAe,CAACP,IAAhB,CAAqB,CAACQ,CAAD,EAAI,CAACG,KAAD,EAAQV,GAAR,CAAJ,CAArB;AACA,eAAOM,eAAP;AACD;;AAGDN,MAAAA,GAAG,GAAGS,GAAG,CAACL,UAAV;AACAE,MAAAA,eAAe,CAACP,IAAhB,CAAqB,CAACQ,CAAD,EAAI,CAACG,KAAD,EAAQV,GAAR,CAAJ,CAArB;AAGAE,MAAAA,KAAK,IAAIO,GAAG,CAACL,UAAJ,GAAiBM,KAA1B;AACAhB,MAAAA,MAAM,IAAIe,GAAG,CAACL,UAAd;AACD;;AAGD,WAAO,IAAP;AACD;;AAQDS,EAAAA,WAAW,CAACX,KAAD,EAAiC;AAC1C,UAAMY,aAAa,GAAG,KAAKT,iBAAL,CAAuBH,KAAvB,CAAtB;;AAGA,QAAI,CAACY,aAAD,IAAkB,KAAKlB,KAA3B,EAAkC;AAChC,YAAM,IAAImB,KAAJ,CAAU,uBAAV,CAAN;AACD;;AAED,QAAI,CAACD,aAAL,EAAoB;AAElB,aAAO,IAAP;AACD;;AAGD,QAAIA,aAAa,CAACN,MAAd,KAAyB,CAA7B,EAAgC;AAC9B,YAAM,CAACQ,WAAD,EAAc,CAACN,KAAD,EAAQV,GAAR,CAAd,IAA8Bc,aAAa,CAAC,CAAD,CAAjD;AACA,YAAMhB,WAAW,GAAG,KAAKH,YAAL,CAAkBqB,WAAlB,CAApB;AACA,YAAMC,IAAI,GAAG,IAAIC,QAAJ,CAAapB,WAAb,EAA0BY,KAA1B,EAAiCV,GAAG,GAAGU,KAAvC,CAAb;AAEA,WAAKhB,MAAL,IAAeQ,KAAf;AACA,WAAKiB,cAAL;AACA,aAAOF,IAAP;AACD;;AAGD,UAAMA,IAAI,GAAG,IAAIC,QAAJ,CAAa,KAAKE,oBAAL,CAA0BN,aAA1B,CAAb,CAAb;AACA,SAAKpB,MAAL,IAAeQ,KAAf;AACA,SAAKiB,cAAL;AACA,WAAOF,IAAP;AACD;;AAKDE,EAAAA,cAAc,GAAS;AACrB,WACE,KAAKxB,YAAL,CAAkBa,MAAlB,GAA2B,CAA3B,IACA,KAAKd,MAAL,GAAc,KAAKD,cAAnB,IAAqC,KAAKE,YAAL,CAAkB,CAAlB,EAAqBS,UAF5D,EAGE;AACA,WAAKV,MAAL,IAAe,KAAKC,YAAL,CAAkB,CAAlB,EAAqBS,UAApC;AACA,WAAKT,YAAL,CAAkB0B,KAAlB;AACD;AACF;;AAYDD,EAAAA,oBAAoB,CAACN,aAAD,EAAwC;AAC1D,QAAIV,UAAkB,GAAG,CAAzB;;AACA,SAAK,MAAMkB,YAAX,IAA2BR,aAA3B,EAA0C;AACxC,YAAM,CAACJ,KAAD,EAAQV,GAAR,IAAesB,YAAY,CAAC,CAAD,CAAjC;AACAlB,MAAAA,UAAU,IAAIJ,GAAG,GAAGU,KAApB;AACD;;AAED,UAAMa,MAAM,GAAG,IAAIC,UAAJ,CAAepB,UAAf,CAAf;AAGA,QAAIqB,YAAoB,GAAG,CAA3B;;AACA,SAAK,MAAMH,YAAX,IAA2BR,aAA3B,EAA0C;AACxC,YAAM,CAACE,WAAD,EAAc,CAACN,KAAD,EAAQV,GAAR,CAAd,IAA8BsB,YAApC;AACA,YAAMI,WAAW,GAAG,IAAIF,UAAJ,CAAe,KAAK7B,YAAL,CAAkBqB,WAAlB,CAAf,CAApB;AACAO,MAAAA,MAAM,CAACI,GAAP,CAAWD,WAAW,CAACE,QAAZ,CAAqBlB,KAArB,EAA4BV,GAA5B,CAAX,EAA6CyB,YAA7C;AACAA,MAAAA,YAAY,IAAIzB,GAAG,GAAGU,KAAtB;AACD;;AAED,WAAOa,MAAM,CAACM,MAAd;AACD;;AAIDC,EAAAA,IAAI,CAAC5B,KAAD,EAAsB;AACxB,SAAKR,MAAL,IAAeQ,KAAf;AACD;;AAID6B,EAAAA,MAAM,CAAC7B,KAAD,EAAsB;AAE1B,SAAKR,MAAL,IAAeQ,KAAf;AACD;;AAzLoC","sourcesContent":["type BinaryChunkReaderOptions = {\n maxRewindBytes: number;\n};\n\nexport default class BinaryChunkReader {\n offset: number;\n arrayBuffers: ArrayBuffer[];\n ended: boolean;\n maxRewindBytes: number;\n\n constructor(options?: BinaryChunkReaderOptions) {\n const {maxRewindBytes = 0} = options || {};\n\n /** current global offset into current array buffer*/\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffers = [];\n this.ended = false;\n\n /** bytes behind offset to hold on to */\n this.maxRewindBytes = maxRewindBytes;\n }\n /**\n * @param arrayBuffer\n */\n write(arrayBuffer: ArrayBuffer): void {\n this.arrayBuffers.push(arrayBuffer);\n }\n\n end(): void {\n this.arrayBuffers = [];\n this.ended = true;\n }\n\n /**\n * Has enough bytes available in array buffers\n *\n * @param bytes Number of bytes\n * @return boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n let bytesAvailable = -this.offset;\n for (const arrayBuffer of this.arrayBuffers) {\n bytesAvailable += arrayBuffer.byteLength;\n if (bytesAvailable >= bytes) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * Find offsets of byte ranges within this.arrayBuffers\n *\n * @param bytes Byte length to read\n * @return Arrays with byte ranges pointing to this.arrayBuffers, Output type is nested array, e.g. [ [0, [1, 2]], ...]\n */\n findBufferOffsets(bytes: number): any[] | null {\n let offset = -this.offset;\n const selectedBuffers: any = [];\n\n for (let i = 0; i < this.arrayBuffers.length; i++) {\n const buf = this.arrayBuffers[i];\n\n // Current buffer isn't long enough to reach global offset\n if (offset + buf.byteLength <= 0) {\n offset += buf.byteLength;\n // eslint-disable-next-line no-continue\n continue;\n }\n\n // Find start/end offsets for this buffer\n // When offset < 0, need to skip over Math.abs(offset) bytes\n // When offset > 0, implies bytes in previous buffer, start at 0\n const start = offset <= 0 ? Math.abs(offset) : 0;\n let end: number;\n\n // Length of requested bytes is contained in current buffer\n if (start + bytes <= buf.byteLength) {\n end = start + bytes;\n selectedBuffers.push([i, [start, end]]);\n return selectedBuffers;\n }\n\n // Will need to look into next buffer\n end = buf.byteLength;\n selectedBuffers.push([i, [start, end]]);\n\n // Need to read fewer bytes in next iter\n bytes -= buf.byteLength - start;\n offset += buf.byteLength;\n }\n\n // Should only finish loop if exhausted all arrays\n return null;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes Number of bytes\n * @return DataView with data\n */\n getDataView(bytes: number): DataView | null {\n const bufferOffsets = this.findBufferOffsets(bytes);\n // return `null` if not enough data, except if end() already called, in\n // which case throw an error.\n if (!bufferOffsets && this.ended) {\n throw new Error('binary data exhausted');\n }\n\n if (!bufferOffsets) {\n // @ts-ignore\n return null;\n }\n\n // If only one arrayBuffer needed, return DataView directly\n if (bufferOffsets.length === 1) {\n const [bufferIndex, [start, end]] = bufferOffsets[0];\n const arrayBuffer = this.arrayBuffers[bufferIndex];\n const view = new DataView(arrayBuffer, start, end - start);\n\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n // Concatenate portions of multiple ArrayBuffers\n const view = new DataView(this._combineArrayBuffers(bufferOffsets));\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n /**\n * Dispose of old array buffers\n */\n disposeBuffers(): void {\n while (\n this.arrayBuffers.length > 0 &&\n this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength\n ) {\n this.offset -= this.arrayBuffers[0].byteLength;\n this.arrayBuffers.shift();\n }\n }\n\n /**\n * Copy multiple ArrayBuffers into one contiguous ArrayBuffer\n *\n * In contrast to concatenateArrayBuffers, this only copies the necessary\n * portions of the source arrays, rather than first copying the entire arrays\n * then taking a part of them.\n *\n * @param bufferOffsets List of internal array offsets\n * @return New contiguous ArrayBuffer\n */\n _combineArrayBuffers(bufferOffsets: any[]): ArrayBufferLike {\n let byteLength: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [start, end] = bufferOffset[1];\n byteLength += end - start;\n }\n\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let resultOffset: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [bufferIndex, [start, end]] = bufferOffset;\n const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);\n result.set(sourceArray.subarray(start, end), resultOffset);\n resultOffset += end - start;\n }\n\n return result.buffer;\n }\n /**\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n /**\n * @param bytes\n */\n rewind(bytes: number): void {\n // TODO - only works if offset is already set\n this.offset -= bytes;\n }\n}\n"],"file":"binary-chunk-reader.js"}
1
+ {"version":3,"file":"binary-chunk-reader.js","names":["BinaryChunkReader","constructor","options","maxRewindBytes","offset","arrayBuffers","ended","write","arrayBuffer","push","end","hasAvailableBytes","bytes","bytesAvailable","byteLength","findBufferOffsets","selectedBuffers","i","length","buf","start","Math","abs","getDataView","bufferOffsets","Error","bufferIndex","view","DataView","disposeBuffers","_combineArrayBuffers","shift","bufferOffset","result","Uint8Array","resultOffset","sourceArray","set","subarray","buffer","skip","rewind"],"sources":["../../../../src/lib/streaming/binary-chunk-reader.ts"],"sourcesContent":["type BinaryChunkReaderOptions = {\n maxRewindBytes: number;\n};\n\nexport default class BinaryChunkReader {\n offset: number;\n arrayBuffers: ArrayBuffer[];\n ended: boolean;\n maxRewindBytes: number;\n\n constructor(options?: BinaryChunkReaderOptions) {\n const {maxRewindBytes = 0} = options || {};\n\n /** current global offset into current array buffer*/\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffers = [];\n this.ended = false;\n\n /** bytes behind offset to hold on to */\n this.maxRewindBytes = maxRewindBytes;\n }\n /**\n * @param arrayBuffer\n */\n write(arrayBuffer: ArrayBuffer): void {\n this.arrayBuffers.push(arrayBuffer);\n }\n\n end(): void {\n this.arrayBuffers = [];\n this.ended = true;\n }\n\n /**\n * Has enough bytes available in array buffers\n *\n * @param bytes Number of bytes\n * @return boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n let bytesAvailable = -this.offset;\n for (const arrayBuffer of this.arrayBuffers) {\n bytesAvailable += arrayBuffer.byteLength;\n if (bytesAvailable >= bytes) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * Find offsets of byte ranges within this.arrayBuffers\n *\n * @param bytes Byte length to read\n * @return Arrays with byte ranges pointing to this.arrayBuffers, Output type is nested array, e.g. [ [0, [1, 2]], ...]\n */\n findBufferOffsets(bytes: number): any[] | null {\n let offset = -this.offset;\n const selectedBuffers: any = [];\n\n for (let i = 0; i < this.arrayBuffers.length; i++) {\n const buf = this.arrayBuffers[i];\n\n // Current buffer isn't long enough to reach global offset\n if (offset + buf.byteLength <= 0) {\n offset += buf.byteLength;\n // eslint-disable-next-line no-continue\n continue;\n }\n\n // Find start/end offsets for this buffer\n // When offset < 0, need to skip over Math.abs(offset) bytes\n // When offset > 0, implies bytes in previous buffer, start at 0\n const start = offset <= 0 ? Math.abs(offset) : 0;\n let end: number;\n\n // Length of requested bytes is contained in current buffer\n if (start + bytes <= buf.byteLength) {\n end = start + bytes;\n selectedBuffers.push([i, [start, end]]);\n return selectedBuffers;\n }\n\n // Will need to look into next buffer\n end = buf.byteLength;\n selectedBuffers.push([i, [start, end]]);\n\n // Need to read fewer bytes in next iter\n bytes -= buf.byteLength - start;\n offset += buf.byteLength;\n }\n\n // Should only finish loop if exhausted all arrays\n return null;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes Number of bytes\n * @return DataView with data\n */\n getDataView(bytes: number): DataView | null {\n const bufferOffsets = this.findBufferOffsets(bytes);\n // return `null` if not enough data, except if end() already called, in\n // which case throw an error.\n if (!bufferOffsets && this.ended) {\n throw new Error('binary data exhausted');\n }\n\n if (!bufferOffsets) {\n // @ts-ignore\n return null;\n }\n\n // If only one arrayBuffer needed, return DataView directly\n if (bufferOffsets.length === 1) {\n const [bufferIndex, [start, end]] = bufferOffsets[0];\n const arrayBuffer = this.arrayBuffers[bufferIndex];\n const view = new DataView(arrayBuffer, start, end - start);\n\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n // Concatenate portions of multiple ArrayBuffers\n const view = new DataView(this._combineArrayBuffers(bufferOffsets));\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n /**\n * Dispose of old array buffers\n */\n disposeBuffers(): void {\n while (\n this.arrayBuffers.length > 0 &&\n this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength\n ) {\n this.offset -= this.arrayBuffers[0].byteLength;\n this.arrayBuffers.shift();\n }\n }\n\n /**\n * Copy multiple ArrayBuffers into one contiguous ArrayBuffer\n *\n * In contrast to concatenateArrayBuffers, this only copies the necessary\n * portions of the source arrays, rather than first copying the entire arrays\n * then taking a part of them.\n *\n * @param bufferOffsets List of internal array offsets\n * @return New contiguous ArrayBuffer\n */\n _combineArrayBuffers(bufferOffsets: any[]): ArrayBufferLike {\n let byteLength: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [start, end] = bufferOffset[1];\n byteLength += end - start;\n }\n\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let resultOffset: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [bufferIndex, [start, end]] = bufferOffset;\n const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);\n result.set(sourceArray.subarray(start, end), resultOffset);\n resultOffset += end - start;\n }\n\n return result.buffer;\n }\n /**\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n /**\n * @param bytes\n */\n rewind(bytes: number): void {\n // TODO - only works if offset is already set\n this.offset -= bytes;\n }\n}\n"],"mappings":";AAIA,eAAe,MAAMA,iBAAiB,CAAC;EAMrCC,WAAW,CAACC,OAAkC,EAAE;IAAA;IAAA;IAAA;IAAA;IAC9C,MAAM;MAACC,cAAc,GAAG;IAAC,CAAC,GAAGD,OAAO,IAAI,CAAC,CAAC;;IAG1C,IAAI,CAACE,MAAM,GAAG,CAAC;IAEf,IAAI,CAACC,YAAY,GAAG,EAAE;IACtB,IAAI,CAACC,KAAK,GAAG,KAAK;;IAGlB,IAAI,CAACH,cAAc,GAAGA,cAAc;EACtC;EAIAI,KAAK,CAACC,WAAwB,EAAQ;IACpC,IAAI,CAACH,YAAY,CAACI,IAAI,CAACD,WAAW,CAAC;EACrC;EAEAE,GAAG,GAAS;IACV,IAAI,CAACL,YAAY,GAAG,EAAE;IACtB,IAAI,CAACC,KAAK,GAAG,IAAI;EACnB;;EAQAK,iBAAiB,CAACC,KAAa,EAAW;IACxC,IAAIC,cAAc,GAAG,CAAC,IAAI,CAACT,MAAM;IACjC,KAAK,MAAMI,WAAW,IAAI,IAAI,CAACH,YAAY,EAAE;MAC3CQ,cAAc,IAAIL,WAAW,CAACM,UAAU;MACxC,IAAID,cAAc,IAAID,KAAK,EAAE;QAC3B,OAAO,IAAI;MACb;IACF;IACA,OAAO,KAAK;EACd;;EAQAG,iBAAiB,CAACH,KAAa,EAAgB;IAC7C,IAAIR,MAAM,GAAG,CAAC,IAAI,CAACA,MAAM;IACzB,MAAMY,eAAoB,GAAG,EAAE;IAE/B,KAAK,IAAIC,CAAC,GAAG,CAAC,EAAEA,CAAC,GAAG,IAAI,CAACZ,YAAY,CAACa,MAAM,EAAED,CAAC,EAAE,EAAE;MACjD,MAAME,GAAG,GAAG,IAAI,CAACd,YAAY,CAACY,CAAC,CAAC;;MAGhC,IAAIb,MAAM,GAAGe,GAAG,CAACL,UAAU,IAAI,CAAC,EAAE;QAChCV,MAAM,IAAIe,GAAG,CAACL,UAAU;QAExB;MACF;;MAKA,MAAMM,KAAK,GAAGhB,MAAM,IAAI,CAAC,GAAGiB,IAAI,CAACC,GAAG,CAAClB,MAAM,CAAC,GAAG,CAAC;MAChD,IAAIM,GAAW;;MAGf,IAAIU,KAAK,GAAGR,KAAK,IAAIO,GAAG,CAACL,UAAU,EAAE;QACnCJ,GAAG,GAAGU,KAAK,GAAGR,KAAK;QACnBI,eAAe,CAACP,IAAI,CAAC,CAACQ,CAAC,EAAE,CAACG,KAAK,EAAEV,GAAG,CAAC,CAAC,CAAC;QACvC,OAAOM,eAAe;MACxB;;MAGAN,GAAG,GAAGS,GAAG,CAACL,UAAU;MACpBE,eAAe,CAACP,IAAI,CAAC,CAACQ,CAAC,EAAE,CAACG,KAAK,EAAEV,GAAG,CAAC,CAAC,CAAC;;MAGvCE,KAAK,IAAIO,GAAG,CAACL,UAAU,GAAGM,KAAK;MAC/BhB,MAAM,IAAIe,GAAG,CAACL,UAAU;IAC1B;;IAGA,OAAO,IAAI;EACb;;EAQAS,WAAW,CAACX,KAAa,EAAmB;IAC1C,MAAMY,aAAa,GAAG,IAAI,CAACT,iBAAiB,CAACH,KAAK,CAAC;IAGnD,IAAI,CAACY,aAAa,IAAI,IAAI,CAAClB,KAAK,EAAE;MAChC,MAAM,IAAImB,KAAK,CAAC,uBAAuB,CAAC;IAC1C;IAEA,IAAI,CAACD,aAAa,EAAE;MAElB,OAAO,IAAI;IACb;;IAGA,IAAIA,aAAa,CAACN,MAAM,KAAK,CAAC,EAAE;MAC9B,MAAM,CAACQ,WAAW,EAAE,CAACN,KAAK,EAAEV,GAAG,CAAC,CAAC,GAAGc,aAAa,CAAC,CAAC,CAAC;MACpD,MAAMhB,WAAW,GAAG,IAAI,CAACH,YAAY,CAACqB,WAAW,CAAC;MAClD,MAAMC,IAAI,GAAG,IAAIC,QAAQ,CAACpB,WAAW,EAAEY,KAAK,EAAEV,GAAG,GAAGU,KAAK,CAAC;MAE1D,IAAI,CAAChB,MAAM,IAAIQ,KAAK;MACpB,IAAI,CAACiB,cAAc,EAAE;MACrB,OAAOF,IAAI;IACb;;IAGA,MAAMA,IAAI,GAAG,IAAIC,QAAQ,CAAC,IAAI,CAACE,oBAAoB,CAACN,aAAa,CAAC,CAAC;IACnE,IAAI,CAACpB,MAAM,IAAIQ,KAAK;IACpB,IAAI,CAACiB,cAAc,EAAE;IACrB,OAAOF,IAAI;EACb;;EAKAE,cAAc,GAAS;IACrB,OACE,IAAI,CAACxB,YAAY,CAACa,MAAM,GAAG,CAAC,IAC5B,IAAI,CAACd,MAAM,GAAG,IAAI,CAACD,cAAc,IAAI,IAAI,CAACE,YAAY,CAAC,CAAC,CAAC,CAACS,UAAU,EACpE;MACA,IAAI,CAACV,MAAM,IAAI,IAAI,CAACC,YAAY,CAAC,CAAC,CAAC,CAACS,UAAU;MAC9C,IAAI,CAACT,YAAY,CAAC0B,KAAK,EAAE;IAC3B;EACF;;EAYAD,oBAAoB,CAACN,aAAoB,EAAmB;IAC1D,IAAIV,UAAkB,GAAG,CAAC;IAC1B,KAAK,MAAMkB,YAAY,IAAIR,aAAa,EAAE;MACxC,MAAM,CAACJ,KAAK,EAAEV,GAAG,CAAC,GAAGsB,YAAY,CAAC,CAAC,CAAC;MACpClB,UAAU,IAAIJ,GAAG,GAAGU,KAAK;IAC3B;IAEA,MAAMa,MAAM,GAAG,IAAIC,UAAU,CAACpB,UAAU,CAAC;;IAGzC,IAAIqB,YAAoB,GAAG,CAAC;IAC5B,KAAK,MAAMH,YAAY,IAAIR,aAAa,EAAE;MACxC,MAAM,CAACE,WAAW,EAAE,CAACN,KAAK,EAAEV,GAAG,CAAC,CAAC,GAAGsB,YAAY;MAChD,MAAMI,WAAW,GAAG,IAAIF,UAAU,CAAC,IAAI,CAAC7B,YAAY,CAACqB,WAAW,CAAC,CAAC;MAClEO,MAAM,CAACI,GAAG,CAACD,WAAW,CAACE,QAAQ,CAAClB,KAAK,EAAEV,GAAG,CAAC,EAAEyB,YAAY,CAAC;MAC1DA,YAAY,IAAIzB,GAAG,GAAGU,KAAK;IAC7B;IAEA,OAAOa,MAAM,CAACM,MAAM;EACtB;EAIAC,IAAI,CAAC5B,KAAa,EAAQ;IACxB,IAAI,CAACR,MAAM,IAAIQ,KAAK;EACtB;EAIA6B,MAAM,CAAC7B,KAAa,EAAQ;IAE1B,IAAI,CAACR,MAAM,IAAIQ,KAAK;EACtB;AACF"}
@@ -2,13 +2,10 @@ import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
2
  export default class BinaryReader {
3
3
  constructor(arrayBuffer) {
4
4
  _defineProperty(this, "offset", void 0);
5
-
6
5
  _defineProperty(this, "arrayBuffer", void 0);
7
-
8
6
  this.offset = 0;
9
7
  this.arrayBuffer = arrayBuffer;
10
8
  }
11
-
12
9
  hasAvailableBytes(bytes) {
13
10
  return this.arrayBuffer.byteLength - this.offset >= bytes;
14
11
  }
@@ -17,7 +14,6 @@ export default class BinaryReader {
17
14
  if (bytes && !this.hasAvailableBytes(bytes)) {
18
15
  throw new Error('binary data exhausted');
19
16
  }
20
-
21
17
  const dataView = bytes ? new DataView(this.arrayBuffer, this.offset, bytes) : new DataView(this.arrayBuffer, this.offset);
22
18
  this.offset += bytes;
23
19
  return dataView;
@@ -30,6 +26,5 @@ export default class BinaryReader {
30
26
  rewind(bytes) {
31
27
  this.offset -= bytes;
32
28
  }
33
-
34
29
  }
35
30
  //# sourceMappingURL=binary-reader.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/streaming/binary-reader.ts"],"names":["BinaryReader","constructor","arrayBuffer","offset","hasAvailableBytes","bytes","byteLength","getDataView","Error","dataView","DataView","skip","rewind"],"mappings":";AAAA,eAAe,MAAMA,YAAN,CAAmB;AAIhCC,EAAAA,WAAW,CAACC,WAAD,EAA2B;AAAA;;AAAA;;AAEpC,SAAKC,MAAL,GAAc,CAAd;AAEA,SAAKD,WAAL,GAAmBA,WAAnB;AACD;;AAODE,EAAAA,iBAAiB,CAACC,KAAD,EAAyB;AACxC,WAAO,KAAKH,WAAL,CAAiBI,UAAjB,GAA8B,KAAKH,MAAnC,IAA6CE,KAApD;AACD;;AAQDE,EAAAA,WAAW,CAACF,KAAD,EAA0B;AACnC,QAAIA,KAAK,IAAI,CAAC,KAAKD,iBAAL,CAAuBC,KAAvB,CAAd,EAA6C;AAC3C,YAAM,IAAIG,KAAJ,CAAU,uBAAV,CAAN;AACD;;AAED,UAAMC,QAAQ,GAAGJ,KAAK,GAClB,IAAIK,QAAJ,CAAa,KAAKR,WAAlB,EAA+B,KAAKC,MAApC,EAA4CE,KAA5C,CADkB,GAElB,IAAIK,QAAJ,CAAa,KAAKR,WAAlB,EAA+B,KAAKC,MAApC,CAFJ;AAGA,SAAKA,MAAL,IAAeE,KAAf;AACA,WAAOI,QAAP;AACD;;AAODE,EAAAA,IAAI,CAACN,KAAD,EAAsB;AACxB,SAAKF,MAAL,IAAeE,KAAf;AACD;;AAODO,EAAAA,MAAM,CAACP,KAAD,EAAsB;AAC1B,SAAKF,MAAL,IAAeE,KAAf;AACD;;AAtD+B","sourcesContent":["export default class BinaryReader {\n offset: number;\n arrayBuffer: ArrayBuffer;\n\n constructor(arrayBuffer: ArrayBuffer) {\n /** current global (stream) offset */\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffer = arrayBuffer;\n }\n /**\n * Checks if there are available bytes in data\n *\n * @param bytes\n * @returns boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n return this.arrayBuffer.byteLength - this.offset >= bytes;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes\n * @returns Dataview\n */\n getDataView(bytes: number): DataView {\n if (bytes && !this.hasAvailableBytes(bytes)) {\n throw new Error('binary data exhausted');\n }\n\n const dataView = bytes\n ? new DataView(this.arrayBuffer, this.offset, bytes)\n : new DataView(this.arrayBuffer, this.offset);\n this.offset += bytes;\n return dataView;\n }\n\n /**\n * Skipping\n *\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n\n /**\n * Rewinding\n *\n * @param bytes\n */\n rewind(bytes: number): void {\n this.offset -= bytes;\n }\n}\n"],"file":"binary-reader.js"}
1
+ {"version":3,"file":"binary-reader.js","names":["BinaryReader","constructor","arrayBuffer","offset","hasAvailableBytes","bytes","byteLength","getDataView","Error","dataView","DataView","skip","rewind"],"sources":["../../../../src/lib/streaming/binary-reader.ts"],"sourcesContent":["export default class BinaryReader {\n offset: number;\n arrayBuffer: ArrayBuffer;\n\n constructor(arrayBuffer: ArrayBuffer) {\n /** current global (stream) offset */\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffer = arrayBuffer;\n }\n /**\n * Checks if there are available bytes in data\n *\n * @param bytes\n * @returns boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n return this.arrayBuffer.byteLength - this.offset >= bytes;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes\n * @returns Dataview\n */\n getDataView(bytes: number): DataView {\n if (bytes && !this.hasAvailableBytes(bytes)) {\n throw new Error('binary data exhausted');\n }\n\n const dataView = bytes\n ? new DataView(this.arrayBuffer, this.offset, bytes)\n : new DataView(this.arrayBuffer, this.offset);\n this.offset += bytes;\n return dataView;\n }\n\n /**\n * Skipping\n *\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n\n /**\n * Rewinding\n *\n * @param bytes\n */\n rewind(bytes: number): void {\n this.offset -= bytes;\n }\n}\n"],"mappings":";AAAA,eAAe,MAAMA,YAAY,CAAC;EAIhCC,WAAW,CAACC,WAAwB,EAAE;IAAA;IAAA;IAEpC,IAAI,CAACC,MAAM,GAAG,CAAC;IAEf,IAAI,CAACD,WAAW,GAAGA,WAAW;EAChC;EAOAE,iBAAiB,CAACC,KAAa,EAAW;IACxC,OAAO,IAAI,CAACH,WAAW,CAACI,UAAU,GAAG,IAAI,CAACH,MAAM,IAAIE,KAAK;EAC3D;;EAQAE,WAAW,CAACF,KAAa,EAAY;IACnC,IAAIA,KAAK,IAAI,CAAC,IAAI,CAACD,iBAAiB,CAACC,KAAK,CAAC,EAAE;MAC3C,MAAM,IAAIG,KAAK,CAAC,uBAAuB,CAAC;IAC1C;IAEA,MAAMC,QAAQ,GAAGJ,KAAK,GAClB,IAAIK,QAAQ,CAAC,IAAI,CAACR,WAAW,EAAE,IAAI,CAACC,MAAM,EAAEE,KAAK,CAAC,GAClD,IAAIK,QAAQ,CAAC,IAAI,CAACR,WAAW,EAAE,IAAI,CAACC,MAAM,CAAC;IAC/C,IAAI,CAACA,MAAM,IAAIE,KAAK;IACpB,OAAOI,QAAQ;EACjB;;EAOAE,IAAI,CAACN,KAAa,EAAQ;IACxB,IAAI,CAACF,MAAM,IAAIE,KAAK;EACtB;;EAOAO,MAAM,CAACP,KAAa,EAAQ;IAC1B,IAAI,CAACF,MAAM,IAAIE,KAAK;EACtB;AACF"}
@@ -1,3 +1,4 @@
1
+
1
2
  export async function* zipBatchIterators(iterator1, iterator2) {
2
3
  let batch1 = [];
3
4
  let batch2 = [];
@@ -10,7 +11,6 @@ export async function* zipBatchIterators(iterator1, iterator2) {
10
11
  value,
11
12
  done
12
13
  } = await iterator1.next();
13
-
14
14
  if (done) {
15
15
  iterator1Done = true;
16
16
  } else {
@@ -21,16 +21,13 @@ export async function* zipBatchIterators(iterator1, iterator2) {
21
21
  value,
22
22
  done
23
23
  } = await iterator2.next();
24
-
25
24
  if (done) {
26
25
  iterator2Done = true;
27
26
  } else {
28
27
  batch2 = value;
29
28
  }
30
29
  }
31
-
32
30
  const batch = extractBatch(batch1, batch2);
33
-
34
31
  if (batch) {
35
32
  yield batch;
36
33
  }
@@ -39,12 +36,12 @@ export async function* zipBatchIterators(iterator1, iterator2) {
39
36
 
40
37
  function extractBatch(batch1, batch2) {
41
38
  const batchLength = Math.min(batch1.length, batch2.length);
42
-
43
39
  if (batchLength === 0) {
44
40
  return null;
45
41
  }
46
42
 
47
43
  const batch = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];
44
+
48
45
  batch1.splice(0, batchLength);
49
46
  batch2.splice(0, batchLength);
50
47
  return batch;
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/streaming/zip-batch-iterators.ts"],"names":["zipBatchIterators","iterator1","iterator2","batch1","batch2","iterator1Done","iterator2Done","length","value","done","next","batch","extractBatch","batchLength","Math","min","slice","splice"],"mappings":"AAMA,OAAO,gBAAgBA,iBAAhB,CACLC,SADK,EAELC,SAFK,EAGsC;AAC3C,MAAIC,MAAgB,GAAG,EAAvB;AACA,MAAIC,MAAgB,GAAG,EAAvB;AACA,MAAIC,aAAsB,GAAG,KAA7B;AACA,MAAIC,aAAsB,GAAG,KAA7B;;AAIA,SAAO,CAACD,aAAD,IAAkB,CAACC,aAA1B,EAAyC;AACvC,QAAIH,MAAM,CAACI,MAAP,KAAkB,CAAlB,IAAuB,CAACF,aAA5B,EAA2C;AACzC,YAAM;AAACG,QAAAA,KAAD;AAAQC,QAAAA;AAAR,UAAgB,MAAMR,SAAS,CAACS,IAAV,EAA5B;;AACA,UAAID,IAAJ,EAAU;AACRJ,QAAAA,aAAa,GAAG,IAAhB;AACD,OAFD,MAEO;AACLF,QAAAA,MAAM,GAAGK,KAAT;AACD;AACF,KAPD,MAOO,IAAIJ,MAAM,CAACG,MAAP,KAAkB,CAAlB,IAAuB,CAACD,aAA5B,EAA2C;AAChD,YAAM;AAACE,QAAAA,KAAD;AAAQC,QAAAA;AAAR,UAAgB,MAAMP,SAAS,CAACQ,IAAV,EAA5B;;AACA,UAAID,IAAJ,EAAU;AACRH,QAAAA,aAAa,GAAG,IAAhB;AACD,OAFD,MAEO;AACLF,QAAAA,MAAM,GAAGI,KAAT;AACD;AACF;;AAED,UAAMG,KAAK,GAAGC,YAAY,CAACT,MAAD,EAASC,MAAT,CAA1B;;AACA,QAAIO,KAAJ,EAAW;AACT,YAAMA,KAAN;AACD;AACF;AACF;;AASD,SAASC,YAAT,CAAsBT,MAAtB,EAAwCC,MAAxC,EAA6E;AAC3E,QAAMS,WAAmB,GAAGC,IAAI,CAACC,GAAL,CAASZ,MAAM,CAACI,MAAhB,EAAwBH,MAAM,CAACG,MAA/B,CAA5B;;AACA,MAAIM,WAAW,KAAK,CAApB,EAAuB;AACrB,WAAO,IAAP;AACD;;AAGD,QAAMF,KAAiB,GAAG,CAACR,MAAM,CAACa,KAAP,CAAa,CAAb,EAAgBH,WAAhB,CAAD,EAA+BT,MAAM,CAACY,KAAP,CAAa,CAAb,EAAgBH,WAAhB,CAA/B,CAA1B;AAGAV,EAAAA,MAAM,CAACc,MAAP,CAAc,CAAd,EAAiBJ,WAAjB;AACAT,EAAAA,MAAM,CAACa,MAAP,CAAc,CAAd,EAAiBJ,WAAjB;AACA,SAAOF,KAAP;AACD","sourcesContent":["/**\n * Zip two iterators together\n *\n * @param iterator1\n * @param iterator2\n */\nexport async function* zipBatchIterators(\n iterator1: AsyncIterator<any[]>,\n iterator2: AsyncIterator<any[]>\n): AsyncGenerator<number[][], void, unknown> {\n let batch1: number[] = [];\n let batch2: number[] = [];\n let iterator1Done: boolean = false;\n let iterator2Done: boolean = false;\n\n // TODO - one could let all iterators flow at full speed using `Promise.race`\n // however we might end up with a big temporary buffer\n while (!iterator1Done && !iterator2Done) {\n if (batch1.length === 0 && !iterator1Done) {\n const {value, done} = await iterator1.next();\n if (done) {\n iterator1Done = true;\n } else {\n batch1 = value;\n }\n } else if (batch2.length === 0 && !iterator2Done) {\n const {value, done} = await iterator2.next();\n if (done) {\n iterator2Done = true;\n } else {\n batch2 = value;\n }\n }\n\n const batch = extractBatch(batch1, batch2);\n if (batch) {\n yield batch;\n }\n }\n}\n\n/**\n * Extract batch of same length from two batches\n *\n * @param batch1\n * @param batch2\n * @return array | null\n */\nfunction extractBatch(batch1: number[], batch2: number[]): number[][] | null {\n const batchLength: number = Math.min(batch1.length, batch2.length);\n if (batchLength === 0) {\n return null;\n }\n\n // Non interleaved arrays\n const batch: number[][] = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];\n\n // Modify the 2 batches\n batch1.splice(0, batchLength);\n batch2.splice(0, batchLength);\n return batch;\n}\n"],"file":"zip-batch-iterators.js"}
1
+ {"version":3,"file":"zip-batch-iterators.js","names":["zipBatchIterators","iterator1","iterator2","batch1","batch2","iterator1Done","iterator2Done","length","value","done","next","batch","extractBatch","batchLength","Math","min","slice","splice"],"sources":["../../../../src/lib/streaming/zip-batch-iterators.ts"],"sourcesContent":["/**\n * Zip two iterators together\n *\n * @param iterator1\n * @param iterator2\n */\nexport async function* zipBatchIterators(\n iterator1: AsyncIterator<any[]>,\n iterator2: AsyncIterator<any[]>\n): AsyncGenerator<number[][], void, unknown> {\n let batch1: number[] = [];\n let batch2: number[] = [];\n let iterator1Done: boolean = false;\n let iterator2Done: boolean = false;\n\n // TODO - one could let all iterators flow at full speed using `Promise.race`\n // however we might end up with a big temporary buffer\n while (!iterator1Done && !iterator2Done) {\n if (batch1.length === 0 && !iterator1Done) {\n const {value, done} = await iterator1.next();\n if (done) {\n iterator1Done = true;\n } else {\n batch1 = value;\n }\n } else if (batch2.length === 0 && !iterator2Done) {\n const {value, done} = await iterator2.next();\n if (done) {\n iterator2Done = true;\n } else {\n batch2 = value;\n }\n }\n\n const batch = extractBatch(batch1, batch2);\n if (batch) {\n yield batch;\n }\n }\n}\n\n/**\n * Extract batch of same length from two batches\n *\n * @param batch1\n * @param batch2\n * @return array | null\n */\nfunction extractBatch(batch1: number[], batch2: number[]): number[][] | null {\n const batchLength: number = Math.min(batch1.length, batch2.length);\n if (batchLength === 0) {\n return null;\n }\n\n // Non interleaved arrays\n const batch: number[][] = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];\n\n // Modify the 2 batches\n batch1.splice(0, batchLength);\n batch2.splice(0, batchLength);\n return batch;\n}\n"],"mappings":";AAMA,OAAO,gBAAgBA,iBAAiB,CACtCC,SAA+B,EAC/BC,SAA+B,EACY;EAC3C,IAAIC,MAAgB,GAAG,EAAE;EACzB,IAAIC,MAAgB,GAAG,EAAE;EACzB,IAAIC,aAAsB,GAAG,KAAK;EAClC,IAAIC,aAAsB,GAAG,KAAK;;EAIlC,OAAO,CAACD,aAAa,IAAI,CAACC,aAAa,EAAE;IACvC,IAAIH,MAAM,CAACI,MAAM,KAAK,CAAC,IAAI,CAACF,aAAa,EAAE;MACzC,MAAM;QAACG,KAAK;QAAEC;MAAI,CAAC,GAAG,MAAMR,SAAS,CAACS,IAAI,EAAE;MAC5C,IAAID,IAAI,EAAE;QACRJ,aAAa,GAAG,IAAI;MACtB,CAAC,MAAM;QACLF,MAAM,GAAGK,KAAK;MAChB;IACF,CAAC,MAAM,IAAIJ,MAAM,CAACG,MAAM,KAAK,CAAC,IAAI,CAACD,aAAa,EAAE;MAChD,MAAM;QAACE,KAAK;QAAEC;MAAI,CAAC,GAAG,MAAMP,SAAS,CAACQ,IAAI,EAAE;MAC5C,IAAID,IAAI,EAAE;QACRH,aAAa,GAAG,IAAI;MACtB,CAAC,MAAM;QACLF,MAAM,GAAGI,KAAK;MAChB;IACF;IAEA,MAAMG,KAAK,GAAGC,YAAY,CAACT,MAAM,EAAEC,MAAM,CAAC;IAC1C,IAAIO,KAAK,EAAE;MACT,MAAMA,KAAK;IACb;EACF;AACF;;AASA,SAASC,YAAY,CAACT,MAAgB,EAAEC,MAAgB,EAAqB;EAC3E,MAAMS,WAAmB,GAAGC,IAAI,CAACC,GAAG,CAACZ,MAAM,CAACI,MAAM,EAAEH,MAAM,CAACG,MAAM,CAAC;EAClE,IAAIM,WAAW,KAAK,CAAC,EAAE;IACrB,OAAO,IAAI;EACb;;EAGA,MAAMF,KAAiB,GAAG,CAACR,MAAM,CAACa,KAAK,CAAC,CAAC,EAAEH,WAAW,CAAC,EAAET,MAAM,CAACY,KAAK,CAAC,CAAC,EAAEH,WAAW,CAAC,CAAC;;EAGtFV,MAAM,CAACc,MAAM,CAAC,CAAC,EAAEJ,WAAW,CAAC;EAC7BT,MAAM,CAACa,MAAM,CAAC,CAAC,EAAEJ,WAAW,CAAC;EAC7B,OAAOF,KAAK;AACd"}
@@ -1,6 +1,8 @@
1
1
  import { SHP_MAGIC_NUMBER } from './shp-loader';
2
2
  import { parseShapefile, parseShapefileInBatches } from './lib/parsers/parse-shapefile';
3
- const VERSION = typeof "3.3.0-alpha.5" !== 'undefined' ? "3.3.0-alpha.5" : 'latest';
3
+
4
+ const VERSION = typeof "3.3.0-alpha.7" !== 'undefined' ? "3.3.0-alpha.7" : 'latest';
5
+
4
6
  export const ShapefileLoader = {
5
7
  name: 'Shapefile',
6
8
  id: 'shapefile',
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/shapefile-loader.ts"],"names":["SHP_MAGIC_NUMBER","parseShapefile","parseShapefileInBatches","VERSION","ShapefileLoader","name","id","module","version","category","extensions","mimeTypes","tests","Uint8Array","buffer","options","shapefile","shp","_maxDimensions","parse","parseInBatches","_typecheckShapefileLoader"],"mappings":"AACA,SAAQA,gBAAR,QAA+B,cAA/B;AACA,SAAQC,cAAR,EAAwBC,uBAAxB,QAAsD,+BAAtD;AAIA,MAAMC,OAAO,GAAG,2BAAuB,WAAvB,qBAAmD,QAAnE;AAMA,OAAO,MAAMC,eAAe,GAAG;AAC7BC,EAAAA,IAAI,EAAE,WADuB;AAE7BC,EAAAA,EAAE,EAAE,WAFyB;AAG7BC,EAAAA,MAAM,EAAE,WAHqB;AAI7BC,EAAAA,OAAO,EAAEL,OAJoB;AAK7BM,EAAAA,QAAQ,EAAE,UALmB;AAM7BC,EAAAA,UAAU,EAAE,CAAC,KAAD,CANiB;AAO7BC,EAAAA,SAAS,EAAE,CAAC,0BAAD,CAPkB;AAQ7BC,EAAAA,KAAK,EAAE,CAAC,IAAIC,UAAJ,CAAeb,gBAAf,EAAiCc,MAAlC,CARsB;AAS7BC,EAAAA,OAAO,EAAE;AACPC,IAAAA,SAAS,EAAE,EADJ;AAEPC,IAAAA,GAAG,EAAE;AACHC,MAAAA,cAAc,EAAE;AADb;AAFE,GAToB;AAe7BC,EAAAA,KAAK,EAAElB,cAfsB;AAgB7BmB,EAAAA,cAAc,EAAElB;AAhBa,CAAxB;AAmBP,OAAO,MAAMmB,yBAA2C,GAAGjB,eAApD","sourcesContent":["import type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {SHP_MAGIC_NUMBER} from './shp-loader';\nimport {parseShapefile, parseShapefileInBatches} from './lib/parsers/parse-shapefile';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * Shapefile loader\n * @note Shapefile is multifile format and requires providing additional files\n */\nexport const ShapefileLoader = {\n name: 'Shapefile',\n id: 'shapefile',\n module: 'shapefile',\n version: VERSION,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shapefile: {},\n shp: {\n _maxDimensions: 4\n }\n },\n parse: parseShapefile,\n parseInBatches: parseShapefileInBatches\n};\n\nexport const _typecheckShapefileLoader: LoaderWithParser = ShapefileLoader;\n"],"file":"shapefile-loader.js"}
1
+ {"version":3,"file":"shapefile-loader.js","names":["SHP_MAGIC_NUMBER","parseShapefile","parseShapefileInBatches","VERSION","ShapefileLoader","name","id","module","version","category","extensions","mimeTypes","tests","Uint8Array","buffer","options","shapefile","shp","_maxDimensions","parse","parseInBatches","_typecheckShapefileLoader"],"sources":["../../src/shapefile-loader.ts"],"sourcesContent":["import type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {SHP_MAGIC_NUMBER} from './shp-loader';\nimport {parseShapefile, parseShapefileInBatches} from './lib/parsers/parse-shapefile';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * Shapefile loader\n * @note Shapefile is multifile format and requires providing additional files\n */\nexport const ShapefileLoader = {\n name: 'Shapefile',\n id: 'shapefile',\n module: 'shapefile',\n version: VERSION,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shapefile: {},\n shp: {\n _maxDimensions: 4\n }\n },\n parse: parseShapefile,\n parseInBatches: parseShapefileInBatches\n};\n\nexport const _typecheckShapefileLoader: LoaderWithParser = ShapefileLoader;\n"],"mappings":"AACA,SAAQA,gBAAgB,QAAO,cAAc;AAC7C,SAAQC,cAAc,EAAEC,uBAAuB,QAAO,+BAA+B;;AAIrF,MAAMC,OAAO,GAAG,sBAAkB,KAAK,WAAW,qBAAiB,QAAQ;;AAM3E,OAAO,MAAMC,eAAe,GAAG;EAC7BC,IAAI,EAAE,WAAW;EACjBC,EAAE,EAAE,WAAW;EACfC,MAAM,EAAE,WAAW;EACnBC,OAAO,EAAEL,OAAO;EAChBM,QAAQ,EAAE,UAAU;EACpBC,UAAU,EAAE,CAAC,KAAK,CAAC;EACnBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EACvCC,KAAK,EAAE,CAAC,IAAIC,UAAU,CAACb,gBAAgB,CAAC,CAACc,MAAM,CAAC;EAChDC,OAAO,EAAE;IACPC,SAAS,EAAE,CAAC,CAAC;IACbC,GAAG,EAAE;MACHC,cAAc,EAAE;IAClB;EACF,CAAC;EACDC,KAAK,EAAElB,cAAc;EACrBmB,cAAc,EAAElB;AAClB,CAAC;AAED,OAAO,MAAMmB,yBAA2C,GAAGjB,eAAe"}
@@ -1,6 +1,8 @@
1
1
  import { parseSHP, parseSHPInBatches } from './lib/parsers/parse-shp';
2
- const VERSION = typeof "3.3.0-alpha.5" !== 'undefined' ? "3.3.0-alpha.5" : 'latest';
2
+
3
+ const VERSION = typeof "3.3.0-alpha.7" !== 'undefined' ? "3.3.0-alpha.7" : 'latest';
3
4
  export const SHP_MAGIC_NUMBER = [0x00, 0x00, 0x27, 0x0a];
5
+
4
6
  export const SHPWorkerLoader = {
5
7
  name: 'SHP',
6
8
  id: 'shp',
@@ -17,7 +19,9 @@ export const SHPWorkerLoader = {
17
19
  }
18
20
  }
19
21
  };
20
- export const SHPLoader = { ...SHPWorkerLoader,
22
+
23
+ export const SHPLoader = {
24
+ ...SHPWorkerLoader,
21
25
  parse: async (arrayBuffer, options) => parseSHP(arrayBuffer, options),
22
26
  parseSync: parseSHP,
23
27
  parseInBatches: parseSHPInBatches
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/shp-loader.ts"],"names":["parseSHP","parseSHPInBatches","VERSION","SHP_MAGIC_NUMBER","SHPWorkerLoader","name","id","module","version","worker","category","extensions","mimeTypes","tests","Uint8Array","buffer","options","shp","_maxDimensions","SHPLoader","parse","arrayBuffer","parseSync","parseInBatches"],"mappings":"AACA,SAAQA,QAAR,EAAkBC,iBAAlB,QAA0C,yBAA1C;AAIA,MAAMC,OAAO,GAAG,2BAAuB,WAAvB,qBAAmD,QAAnE;AAEA,OAAO,MAAMC,gBAAgB,GAAG,CAAC,IAAD,EAAO,IAAP,EAAa,IAAb,EAAmB,IAAnB,CAAzB;AAKP,OAAO,MAAMC,eAAuB,GAAG;AACrCC,EAAAA,IAAI,EAAE,KAD+B;AAErCC,EAAAA,EAAE,EAAE,KAFiC;AAGrCC,EAAAA,MAAM,EAAE,WAH6B;AAIrCC,EAAAA,OAAO,EAAEN,OAJ4B;AAKrCO,EAAAA,MAAM,EAAE,IAL6B;AAMrCC,EAAAA,QAAQ,EAAE,UAN2B;AAOrCC,EAAAA,UAAU,EAAE,CAAC,KAAD,CAPyB;AAQrCC,EAAAA,SAAS,EAAE,CAAC,0BAAD,CAR0B;AAUrCC,EAAAA,KAAK,EAAE,CAAC,IAAIC,UAAJ,CAAeX,gBAAf,EAAiCY,MAAlC,CAV8B;AAWrCC,EAAAA,OAAO,EAAE;AACPC,IAAAA,GAAG,EAAE;AACHC,MAAAA,cAAc,EAAE;AADb;AADE;AAX4B,CAAhC;AAmBP,OAAO,MAAMC,SAA2B,GAAG,EACzC,GAAGf,eADsC;AAEzCgB,EAAAA,KAAK,EAAE,OAAOC,WAAP,EAAoBL,OAApB,KAAiChB,QAAQ,CAACqB,WAAD,EAAcL,OAAd,CAFP;AAGzCM,EAAAA,SAAS,EAAEtB,QAH8B;AAIzCuB,EAAAA,cAAc,EAAEtB;AAJyB,CAApC","sourcesContent":["import type {Loader, LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {parseSHP, parseSHPInBatches} from './lib/parsers/parse-shp';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\nexport const SHP_MAGIC_NUMBER = [0x00, 0x00, 0x27, 0x0a];\n\n/**\n * SHP file loader\n */\nexport const SHPWorkerLoader: Loader = {\n name: 'SHP',\n id: 'shp',\n module: 'shapefile',\n version: VERSION,\n worker: true,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n // ISSUE: This also identifies SHX files, which are identical to SHP for the first 100 bytes...\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shp: {\n _maxDimensions: 4\n }\n }\n};\n\n/** SHP file loader */\nexport const SHPLoader: LoaderWithParser = {\n ...SHPWorkerLoader,\n parse: async (arrayBuffer, options?) => parseSHP(arrayBuffer, options),\n parseSync: parseSHP,\n parseInBatches: parseSHPInBatches\n};\n"],"file":"shp-loader.js"}
1
+ {"version":3,"file":"shp-loader.js","names":["parseSHP","parseSHPInBatches","VERSION","SHP_MAGIC_NUMBER","SHPWorkerLoader","name","id","module","version","worker","category","extensions","mimeTypes","tests","Uint8Array","buffer","options","shp","_maxDimensions","SHPLoader","parse","arrayBuffer","parseSync","parseInBatches"],"sources":["../../src/shp-loader.ts"],"sourcesContent":["import type {Loader, LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {parseSHP, parseSHPInBatches} from './lib/parsers/parse-shp';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\nexport const SHP_MAGIC_NUMBER = [0x00, 0x00, 0x27, 0x0a];\n\n/**\n * SHP file loader\n */\nexport const SHPWorkerLoader: Loader = {\n name: 'SHP',\n id: 'shp',\n module: 'shapefile',\n version: VERSION,\n worker: true,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n // ISSUE: This also identifies SHX files, which are identical to SHP for the first 100 bytes...\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shp: {\n _maxDimensions: 4\n }\n }\n};\n\n/** SHP file loader */\nexport const SHPLoader: LoaderWithParser = {\n ...SHPWorkerLoader,\n parse: async (arrayBuffer, options?) => parseSHP(arrayBuffer, options),\n parseSync: parseSHP,\n parseInBatches: parseSHPInBatches\n};\n"],"mappings":"AACA,SAAQA,QAAQ,EAAEC,iBAAiB,QAAO,yBAAyB;;AAInE,MAAMC,OAAO,GAAG,sBAAkB,KAAK,WAAW,qBAAiB,QAAQ;AAE3E,OAAO,MAAMC,gBAAgB,GAAG,CAAC,IAAI,EAAE,IAAI,EAAE,IAAI,EAAE,IAAI,CAAC;;AAKxD,OAAO,MAAMC,eAAuB,GAAG;EACrCC,IAAI,EAAE,KAAK;EACXC,EAAE,EAAE,KAAK;EACTC,MAAM,EAAE,WAAW;EACnBC,OAAO,EAAEN,OAAO;EAChBO,MAAM,EAAE,IAAI;EACZC,QAAQ,EAAE,UAAU;EACpBC,UAAU,EAAE,CAAC,KAAK,CAAC;EACnBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EAEvCC,KAAK,EAAE,CAAC,IAAIC,UAAU,CAACX,gBAAgB,CAAC,CAACY,MAAM,CAAC;EAChDC,OAAO,EAAE;IACPC,GAAG,EAAE;MACHC,cAAc,EAAE;IAClB;EACF;AACF,CAAC;;AAGD,OAAO,MAAMC,SAA2B,GAAG;EACzC,GAAGf,eAAe;EAClBgB,KAAK,EAAE,OAAOC,WAAW,EAAEL,OAAQ,KAAKhB,QAAQ,CAACqB,WAAW,EAAEL,OAAO,CAAC;EACtEM,SAAS,EAAEtB,QAAQ;EACnBuB,cAAc,EAAEtB;AAClB,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../src/workers/dbf-worker.ts"],"names":["DBFLoader","createLoaderWorker"],"mappings":"AAAA,SAAQA,SAAR,QAAwB,eAAxB;AACA,SAAQC,kBAAR,QAAiC,0BAAjC;AAEAA,kBAAkB,CAACD,SAAD,CAAlB","sourcesContent":["import {DBFLoader} from '../dbf-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(DBFLoader);\n"],"file":"dbf-worker.js"}
1
+ {"version":3,"file":"dbf-worker.js","names":["DBFLoader","createLoaderWorker"],"sources":["../../../src/workers/dbf-worker.ts"],"sourcesContent":["import {DBFLoader} from '../dbf-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(DBFLoader);\n"],"mappings":"AAAA,SAAQA,SAAS,QAAO,eAAe;AACvC,SAAQC,kBAAkB,QAAO,0BAA0B;AAE3DA,kBAAkB,CAACD,SAAS,CAAC"}
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../src/workers/shp-worker.ts"],"names":["SHPLoader","createLoaderWorker"],"mappings":"AAAA,SAAQA,SAAR,QAAwB,eAAxB;AACA,SAAQC,kBAAR,QAAiC,0BAAjC;AAEAA,kBAAkB,CAACD,SAAD,CAAlB","sourcesContent":["import {SHPLoader} from '../shp-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(SHPLoader);\n"],"file":"shp-worker.js"}
1
+ {"version":3,"file":"shp-worker.js","names":["SHPLoader","createLoaderWorker"],"sources":["../../../src/workers/shp-worker.ts"],"sourcesContent":["import {SHPLoader} from '../shp-loader';\nimport {createLoaderWorker} from '@loaders.gl/loader-utils';\n\ncreateLoaderWorker(SHPLoader);\n"],"mappings":"AAAA,SAAQA,SAAS,QAAO,eAAe;AACvC,SAAQC,kBAAkB,QAAO,0BAA0B;AAE3DA,kBAAkB,CAACD,SAAS,CAAC"}
@@ -430,7 +430,7 @@
430
430
  }
431
431
 
432
432
  // src/shp-loader.ts
433
- var VERSION = true ? "3.3.0-alpha.5" : "latest";
433
+ var VERSION = true ? "3.3.0-alpha.7" : "latest";
434
434
  var SHP_MAGIC_NUMBER2 = [0, 0, 39, 10];
435
435
  var SHPWorkerLoader = {
436
436
  name: "SHP",
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@loaders.gl/shapefile",
3
3
  "description": "Loader for the Shapefile Format",
4
- "version": "3.3.0-alpha.5",
4
+ "version": "3.3.0-alpha.7",
5
5
  "license": "MIT",
6
6
  "publishConfig": {
7
7
  "access": "public"
@@ -37,10 +37,10 @@
37
37
  "build-worker-dbf": "esbuild src/workers/dbf-worker.ts --bundle --outfile=dist/dbf-worker.js --define:__VERSION__=\\\"$npm_package_version\\\""
38
38
  },
39
39
  "dependencies": {
40
- "@loaders.gl/gis": "3.3.0-alpha.5",
41
- "@loaders.gl/loader-utils": "3.3.0-alpha.5",
42
- "@loaders.gl/schema": "3.3.0-alpha.5",
40
+ "@loaders.gl/gis": "3.3.0-alpha.7",
41
+ "@loaders.gl/loader-utils": "3.3.0-alpha.7",
42
+ "@loaders.gl/schema": "3.3.0-alpha.7",
43
43
  "@math.gl/proj4": "^3.5.1"
44
44
  },
45
- "gitHead": "d2df3bead97710c45fd2974cd51ecd7d5f7f5ea4"
45
+ "gitHead": "29b08f3519c50984e84bf4234e607cab7c7d1c3e"
46
46
  }
@@ -290,7 +290,7 @@ function parseRow(
290
290
  fields: DBFField[],
291
291
  textDecoder: TextDecoder
292
292
  ): {[key: string]: any} {
293
- const out = {};
293
+ const out: {[key: string]: string | number | boolean | null} = {};
294
294
  let offset = 0;
295
295
  for (const field of fields) {
296
296
  const text = textDecoder.decode(