@loaders.gl/shapefile 3.3.0-alpha.5 → 3.3.0-alpha.6

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (67) hide show
  1. package/dist/dbf-worker.js +1 -1
  2. package/dist/dist.min.js +47 -11
  3. package/dist/es5/bundle.js +0 -1
  4. package/dist/es5/bundle.js.map +1 -1
  5. package/dist/es5/dbf-loader.js +5 -16
  6. package/dist/es5/dbf-loader.js.map +1 -1
  7. package/dist/es5/index.js +6 -9
  8. package/dist/es5/index.js.map +1 -1
  9. package/dist/es5/lib/parsers/parse-dbf.js +69 -154
  10. package/dist/es5/lib/parsers/parse-dbf.js.map +1 -1
  11. package/dist/es5/lib/parsers/parse-shapefile.js +34 -148
  12. package/dist/es5/lib/parsers/parse-shapefile.js.map +1 -1
  13. package/dist/es5/lib/parsers/parse-shp-geometry.js +4 -57
  14. package/dist/es5/lib/parsers/parse-shp-geometry.js.map +1 -1
  15. package/dist/es5/lib/parsers/parse-shp-header.js +0 -3
  16. package/dist/es5/lib/parsers/parse-shp-header.js.map +1 -1
  17. package/dist/es5/lib/parsers/parse-shp.js +39 -98
  18. package/dist/es5/lib/parsers/parse-shp.js.map +1 -1
  19. package/dist/es5/lib/parsers/parse-shx.js +0 -4
  20. package/dist/es5/lib/parsers/parse-shx.js.map +1 -1
  21. package/dist/es5/lib/parsers/types.js.map +1 -1
  22. package/dist/es5/lib/streaming/binary-chunk-reader.js +39 -55
  23. package/dist/es5/lib/streaming/binary-chunk-reader.js.map +1 -1
  24. package/dist/es5/lib/streaming/binary-reader.js +11 -12
  25. package/dist/es5/lib/streaming/binary-reader.js.map +1 -1
  26. package/dist/es5/lib/streaming/zip-batch-iterators.js +1 -24
  27. package/dist/es5/lib/streaming/zip-batch-iterators.js.map +1 -1
  28. package/dist/es5/shapefile-loader.js +1 -3
  29. package/dist/es5/shapefile-loader.js.map +1 -1
  30. package/dist/es5/shp-loader.js +6 -17
  31. package/dist/es5/shp-loader.js.map +1 -1
  32. package/dist/es5/workers/dbf-worker.js +0 -2
  33. package/dist/es5/workers/dbf-worker.js.map +1 -1
  34. package/dist/es5/workers/shp-worker.js +0 -2
  35. package/dist/es5/workers/shp-worker.js.map +1 -1
  36. package/dist/esm/bundle.js +1 -1
  37. package/dist/esm/bundle.js.map +1 -1
  38. package/dist/esm/dbf-loader.js +6 -2
  39. package/dist/esm/dbf-loader.js.map +1 -1
  40. package/dist/esm/index.js.map +1 -1
  41. package/dist/esm/lib/parsers/parse-dbf.js +18 -58
  42. package/dist/esm/lib/parsers/parse-dbf.js.map +1 -1
  43. package/dist/esm/lib/parsers/parse-shapefile.js +12 -35
  44. package/dist/esm/lib/parsers/parse-shapefile.js.map +1 -1
  45. package/dist/esm/lib/parsers/parse-shp-geometry.js +3 -28
  46. package/dist/esm/lib/parsers/parse-shp-geometry.js.map +1 -1
  47. package/dist/esm/lib/parsers/parse-shp-header.js +1 -3
  48. package/dist/esm/lib/parsers/parse-shp-header.js.map +1 -1
  49. package/dist/esm/lib/parsers/parse-shp.js +2 -26
  50. package/dist/esm/lib/parsers/parse-shp.js.map +1 -1
  51. package/dist/esm/lib/parsers/parse-shx.js +1 -2
  52. package/dist/esm/lib/parsers/parse-shx.js.map +1 -1
  53. package/dist/esm/lib/parsers/types.js.map +1 -1
  54. package/dist/esm/lib/streaming/binary-chunk-reader.js +4 -19
  55. package/dist/esm/lib/streaming/binary-chunk-reader.js.map +1 -1
  56. package/dist/esm/lib/streaming/binary-reader.js +0 -5
  57. package/dist/esm/lib/streaming/binary-reader.js.map +1 -1
  58. package/dist/esm/lib/streaming/zip-batch-iterators.js +2 -5
  59. package/dist/esm/lib/streaming/zip-batch-iterators.js.map +1 -1
  60. package/dist/esm/shapefile-loader.js +3 -1
  61. package/dist/esm/shapefile-loader.js.map +1 -1
  62. package/dist/esm/shp-loader.js +6 -2
  63. package/dist/esm/shp-loader.js.map +1 -1
  64. package/dist/esm/workers/dbf-worker.js.map +1 -1
  65. package/dist/esm/workers/shp-worker.js.map +1 -1
  66. package/dist/shp-worker.js +1 -1
  67. package/package.json +5 -5
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/parsers/parse-shp.ts"],"names":["LITTLE_ENDIAN","BIG_ENDIAN","SHP_HEADER_SIZE","SHP_RECORD_HEADER_SIZE","STATE","EXPECTING_HEADER","EXPECTING_RECORD","END","ERROR","SHPParser","options","BinaryChunkReader","maxRewindBytes","geometries","progress","bytesTotal","NaN","bytesUsed","rows","currentIndex","arrayBuffer","binaryReader","write","state","parseState","result","end","error","parseSHP","shpParser","parseSHPInBatches","asyncIterator","parser","headerReturned","header","length","dataView","getDataView","hasAvailableBytes","recordHeaderView","recordHeader","recordNumber","getInt32","byteLength","type","rewind","invalidRecord","recordView","geometry","push","ended","message"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;AACA;;AACA;;AACA;;AAGA,IAAMA,aAAa,GAAG,IAAtB;AACA,IAAMC,UAAU,GAAG,KAAnB;AAEA,IAAMC,eAAe,GAAG,GAAxB;AAGA,IAAMC,sBAAsB,GAAG,EAA/B;AAEA,IAAMC,KAAK,GAAG;AACZC,EAAAA,gBAAgB,EAAE,CADN;AAEZC,EAAAA,gBAAgB,EAAE,CAFN;AAGZC,EAAAA,GAAG,EAAE,CAHO;AAIZC,EAAAA,KAAK,EAAE;AAJK,CAAd;;IAmBMC,S;AAgBJ,qBAAYC,OAAZ,EAAwC;AAAA;AAAA,mDAfX,EAeW;AAAA,wDAdzB,IAAIC,0BAAJ,CAAsB;AAACC,MAAAA,cAAc,EAAET;AAAjB,KAAtB,CAcyB;AAAA,iDAbhCC,KAAK,CAACC,gBAa0B;AAAA,kDAZpB;AAClBQ,MAAAA,UAAU,EAAE,EADM;AAIlBC,MAAAA,QAAQ,EAAE;AACRC,QAAAA,UAAU,EAAEC,GADJ;AAERC,QAAAA,SAAS,EAAED,GAFH;AAGRE,QAAAA,IAAI,EAAEF;AAHE,OAJQ;AASlBG,MAAAA,YAAY,EAAEH;AATI,KAYoB;AACtC,SAAKN,OAAL,GAAeA,OAAf;AACD;;;;WAED,eAAMU,WAAN,EAAgC;AAC9B,WAAKC,YAAL,CAAkBC,KAAlB,CAAwBF,WAAxB;AACA,WAAKG,KAAL,GAAaC,UAAU,CAAC,KAAKD,KAAN,EAAa,KAAKE,MAAlB,EAA0B,KAAKJ,YAA/B,EAA6C,KAAKX,OAAlD,CAAvB;AACD;;;WAED,eAAM;AACJ,WAAKW,YAAL,CAAkBK,GAAlB;AACA,WAAKH,KAAL,GAAaC,UAAU,CAAC,KAAKD,KAAN,EAAa,KAAKE,MAAlB,EAA0B,KAAKJ,YAA/B,EAA6C,KAAKX,OAAlD,CAAvB;;AAEA,UAAI,KAAKa,KAAL,KAAenB,KAAK,CAACG,GAAzB,EAA8B;AAC5B,aAAKgB,KAAL,GAAanB,KAAK,CAACI,KAAnB;AACA,aAAKiB,MAAL,CAAYE,KAAZ,GAAoB,qBAApB;AACD;AACF;;;;;AAGI,SAASC,QAAT,CAAkBR,WAAlB,EAA4CV,OAA5C,EAAgF;AACrF,MAAMmB,SAAS,GAAG,IAAIpB,SAAJ,CAAcC,OAAd,CAAlB;AACAmB,EAAAA,SAAS,CAACP,KAAV,CAAgBF,WAAhB;AACAS,EAAAA,SAAS,CAACH,GAAV;AAGA,SAAOG,SAAS,CAACJ,MAAjB;AACD;;SAOsBK,iB;;;;;mFAAhB,iBACLC,aADK,EAELrB,OAFK;AAAA;;AAAA;AAAA;AAAA;AAAA;AAICsB,YAAAA,MAJD,GAIU,IAAIvB,SAAJ,CAAcC,OAAd,CAJV;AAKDuB,YAAAA,cALC,GAKgB,KALhB;AAAA;AAAA;AAAA;AAAA,qDAM2BF,aAN3B;;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;;AAMYX,YAAAA,WANZ;AAOHY,YAAAA,MAAM,CAACV,KAAP,CAAaF,WAAb;;AAPG,kBAQC,CAACa,cAAD,IAAmBD,MAAM,CAACP,MAAP,CAAcS,MARlC;AAAA;AAAA;AAAA;;AASDD,YAAAA,cAAc,GAAG,IAAjB;AATC;AAUD,mBAAMD,MAAM,CAACP,MAAP,CAAcS,MAApB;;AAVC;AAAA,kBAaCF,MAAM,CAACP,MAAP,CAAcZ,UAAd,CAAyBsB,MAAzB,GAAkC,CAbnC;AAAA;AAAA;AAAA;;AAAA;AAcD,mBAAMH,MAAM,CAACP,MAAP,CAAcZ,UAApB;;AAdC;AAeDmB,YAAAA,MAAM,CAACP,MAAP,CAAcZ,UAAd,GAA2B,EAA3B;;AAfC;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;;AAAA;;AAAA;AAAA;;AAAA;AAAA;;AAAA;AAkBLmB,YAAAA,MAAM,CAACN,GAAP;;AAlBK,kBAmBDM,MAAM,CAACP,MAAP,CAAcZ,UAAd,CAAyBsB,MAAzB,GAAkC,CAnBjC;AAAA;AAAA;AAAA;;AAAA;AAoBH,mBAAMH,MAAM,CAACP,MAAP,CAAcZ,UAApB;;AApBG;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,G;;;;AAwCP,SAASW,UAAT,CACED,KADF,EAEEE,MAFF,EAGEJ,YAHF,EAIEX,OAJF,EAKU;AAER,SAAO,IAAP,EAAa;AACX,QAAI;AACF,cAAQa,KAAR;AACE,aAAKnB,KAAK,CAACI,KAAX;AACA,aAAKJ,KAAK,CAACG,GAAX;AACE,iBAAOgB,KAAP;;AAEF,aAAKnB,KAAK,CAACC,gBAAX;AAEE,cAAM+B,QAAQ,GAAGf,YAAY,CAACgB,WAAb,CAAyBnC,eAAzB,CAAjB;;AACA,cAAI,CAACkC,QAAL,EAAe;AACb,mBAAOb,KAAP;AACD;;AAEDE,UAAAA,MAAM,CAACS,MAAP,GAAgB,oCAAeE,QAAf,CAAhB;AACAX,UAAAA,MAAM,CAACX,QAAP,GAAkB;AAChBG,YAAAA,SAAS,EAAE,CADK;AAEhBF,YAAAA,UAAU,EAAEU,MAAM,CAACS,MAAP,CAAcC,MAFV;AAGhBjB,YAAAA,IAAI,EAAE;AAHU,WAAlB;AAMAO,UAAAA,MAAM,CAACN,YAAP,GAAsB,CAAtB;AACAI,UAAAA,KAAK,GAAGnB,KAAK,CAACE,gBAAd;AACA;;AAEF,aAAKF,KAAK,CAACE,gBAAX;AACE,iBAAOe,YAAY,CAACiB,iBAAb,CAA+BnC,sBAA/B,CAAP,EAA+D;AAAA;;AAC7D,gBAAMoC,gBAAgB,GAAGlB,YAAY,CAACgB,WAAb,CAAyBlC,sBAAzB,CAAzB;AACA,gBAAMqC,YAAY,GAAG;AACnBC,cAAAA,YAAY,EAAEF,gBAAgB,CAACG,QAAjB,CAA0B,CAA1B,EAA6BzC,UAA7B,CADK;AAGnB0C,cAAAA,UAAU,EAAEJ,gBAAgB,CAACG,QAAjB,CAA0B,CAA1B,EAA6BzC,UAA7B,IAA2C,CAHpC;AAKnB2C,cAAAA,IAAI,EAAEL,gBAAgB,CAACG,QAAjB,CAA0B,CAA1B,EAA6B1C,aAA7B;AALa,aAArB;;AAQA,gBAAI,CAACqB,YAAY,CAACiB,iBAAb,CAA+BE,YAAY,CAACG,UAAb,GAA0B,CAAzD,CAAL,EAAkE;AAChEtB,cAAAA,YAAY,CAACwB,MAAb,CAAoB1C,sBAApB;AACA,qBAAOoB,KAAP;AACD;;AAED,gBAAMuB,aAAa,GACjBN,YAAY,CAACG,UAAb,GAA0B,CAA1B,IACAH,YAAY,CAACI,IAAb,wBAAsBnB,MAAM,CAACS,MAA7B,mDAAsB,eAAeU,IAArC,CADA,IAEAJ,YAAY,CAACC,YAAb,KAA8BhB,MAAM,CAACN,YAHvC;;AAMA,gBAAI2B,aAAJ,EAAmB;AAKjBzB,cAAAA,YAAY,CAACwB,MAAb,CAAoB1C,sBAAsB,GAAG,CAA7C;AACD,aAND,MAMO;AAGLkB,cAAAA,YAAY,CAACwB,MAAb,CAAoB,CAApB;AAEA,kBAAME,UAAU,GAAG1B,YAAY,CAACgB,WAAb,CAAyBG,YAAY,CAACG,UAAtC,CAAnB;AACA,kBAAMK,QAAQ,GAAG,mCAAYD,UAAZ,EAAwBrC,OAAxB,CAAjB;AACAe,cAAAA,MAAM,CAACZ,UAAP,CAAkBoC,IAAlB,CAAuBD,QAAvB;AAEAvB,cAAAA,MAAM,CAACN,YAAP;AACAM,cAAAA,MAAM,CAACX,QAAP,CAAgBI,IAAhB,GAAuBO,MAAM,CAACN,YAAP,GAAsB,CAA7C;AACD;AACF;;AAED,cAAIE,YAAY,CAAC6B,KAAjB,EAAwB;AACtB3B,YAAAA,KAAK,GAAGnB,KAAK,CAACG,GAAd;AACD;;AAED,iBAAOgB,KAAP;;AAEF;AACEA,UAAAA,KAAK,GAAGnB,KAAK,CAACI,KAAd;AACAiB,UAAAA,MAAM,CAACE,KAAP,kCAAuCJ,KAAvC;AACA,iBAAOA,KAAP;AA1EJ;AA4ED,KA7ED,CA6EE,OAAOI,KAAP,EAAc;AACdJ,MAAAA,KAAK,GAAGnB,KAAK,CAACI,KAAd;AACAiB,MAAAA,MAAM,CAACE,KAAP,iCAAuCA,KAAvC,aAAuCA,KAAvC,uBAAuCA,KAAD,CAAkBwB,OAAxD;AACA,aAAO5B,KAAP;AACD;AACF;AACF","sourcesContent":["import type {BinaryGeometry} from '@loaders.gl/schema';\nimport BinaryChunkReader from '../streaming/binary-chunk-reader';\nimport {parseSHPHeader, SHPHeader} from './parse-shp-header';\nimport {parseRecord} from './parse-shp-geometry';\nimport {SHPLoaderOptions} from './types';\n\nconst LITTLE_ENDIAN = true;\nconst BIG_ENDIAN = false;\n\nconst SHP_HEADER_SIZE = 100;\n// According to the spec, the record header is just 8 bytes, but here we set it\n// to 12 so that we can also access the record's type\nconst SHP_RECORD_HEADER_SIZE = 12;\n\nconst STATE = {\n EXPECTING_HEADER: 0,\n EXPECTING_RECORD: 1,\n END: 2,\n ERROR: 3\n};\n\ntype SHPResult = {\n geometries: (BinaryGeometry | null)[];\n header?: SHPHeader;\n error?: string;\n progress: {\n bytesUsed: number;\n bytesTotal: number;\n rows: number;\n };\n currentIndex: number;\n};\n\nclass SHPParser {\n options?: SHPLoaderOptions = {};\n binaryReader = new BinaryChunkReader({maxRewindBytes: SHP_RECORD_HEADER_SIZE});\n state = STATE.EXPECTING_HEADER;\n result: SHPResult = {\n geometries: [],\n // Initialize with number values to make TS happy\n // These are initialized for real in STATE.EXPECTING_HEADER\n progress: {\n bytesTotal: NaN,\n bytesUsed: NaN,\n rows: NaN\n },\n currentIndex: NaN\n };\n\n constructor(options?: SHPLoaderOptions) {\n this.options = options;\n }\n\n write(arrayBuffer: ArrayBuffer) {\n this.binaryReader.write(arrayBuffer);\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n }\n\n end() {\n this.binaryReader.end();\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n // this.result.progress.bytesUsed = this.binaryReader.bytesUsed();\n if (this.state !== STATE.END) {\n this.state = STATE.ERROR;\n this.result.error = 'SHP incomplete file';\n }\n }\n}\n\nexport function parseSHP(arrayBuffer: ArrayBuffer, options?: object): BinaryGeometry[] {\n const shpParser = new SHPParser(options);\n shpParser.write(arrayBuffer);\n shpParser.end();\n\n // @ts-ignore\n return shpParser.result;\n}\n\n/**\n * @param asyncIterator\n * @param options\n * @returns\n */\nexport async function* parseSHPInBatches(\n asyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options?: object\n): AsyncIterable<BinaryGeometry | object> {\n const parser = new SHPParser(options);\n let headerReturned = false;\n for await (const arrayBuffer of asyncIterator) {\n parser.write(arrayBuffer);\n if (!headerReturned && parser.result.header) {\n headerReturned = true;\n yield parser.result.header;\n }\n\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n parser.result.geometries = [];\n }\n }\n parser.end();\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n }\n\n return;\n}\n\n/**\n * State-machine parser for SHP data\n *\n * Note that whenever more data is needed, a `return`, not a `break`, is\n * necessary, as the `break` keeps the context within `parseState`, while\n * `return` releases context so that more data can be written into the\n * BinaryChunkReader.\n *\n * @param state Current state\n * @param result An object to hold result data\n * @param binaryReader\n * @return State at end of current parsing\n */\n/* eslint-disable complexity, max-depth */\nfunction parseState(\n state: number,\n result: SHPResult,\n binaryReader: BinaryChunkReader,\n options?: SHPLoaderOptions\n): number {\n // eslint-disable-next-line no-constant-condition\n while (true) {\n try {\n switch (state) {\n case STATE.ERROR:\n case STATE.END:\n return state;\n\n case STATE.EXPECTING_HEADER:\n // Parse initial file header\n const dataView = binaryReader.getDataView(SHP_HEADER_SIZE);\n if (!dataView) {\n return state;\n }\n\n result.header = parseSHPHeader(dataView);\n result.progress = {\n bytesUsed: 0,\n bytesTotal: result.header.length,\n rows: 0\n };\n // index numbering starts at 1\n result.currentIndex = 1;\n state = STATE.EXPECTING_RECORD;\n break;\n\n case STATE.EXPECTING_RECORD:\n while (binaryReader.hasAvailableBytes(SHP_RECORD_HEADER_SIZE)) {\n const recordHeaderView = binaryReader.getDataView(SHP_RECORD_HEADER_SIZE) as DataView;\n const recordHeader = {\n recordNumber: recordHeaderView.getInt32(0, BIG_ENDIAN),\n // 2 byte words; includes the four words of record header\n byteLength: recordHeaderView.getInt32(4, BIG_ENDIAN) * 2,\n // This is actually part of the record, not the header...\n type: recordHeaderView.getInt32(8, LITTLE_ENDIAN)\n };\n\n if (!binaryReader.hasAvailableBytes(recordHeader.byteLength - 4)) {\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE);\n return state;\n }\n\n const invalidRecord =\n recordHeader.byteLength < 4 ||\n recordHeader.type !== result.header?.type ||\n recordHeader.recordNumber !== result.currentIndex;\n\n // All records must have at least four bytes (for the record shape type)\n if (invalidRecord) {\n // Malformed record, try again, advancing just 4 bytes\n // Note: this is a rewind because binaryReader.getDataView above\n // moved the pointer forward 12 bytes, so rewinding 8 bytes still\n // leaves us 4 bytes ahead\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE - 4);\n } else {\n // Note: type is actually part of the record, not the header, so\n // rewind 4 bytes before reading record\n binaryReader.rewind(4);\n\n const recordView = binaryReader.getDataView(recordHeader.byteLength) as DataView;\n const geometry = parseRecord(recordView, options);\n result.geometries.push(geometry);\n\n result.currentIndex++;\n result.progress.rows = result.currentIndex - 1;\n }\n }\n\n if (binaryReader.ended) {\n state = STATE.END;\n }\n\n return state;\n\n default:\n state = STATE.ERROR;\n result.error = `illegal parser state ${state}`;\n return state;\n }\n } catch (error) {\n state = STATE.ERROR;\n result.error = `SHP parsing failed: ${(error as Error)?.message}`;\n return state;\n }\n }\n}\n"],"file":"parse-shp.js"}
1
+ {"version":3,"file":"parse-shp.js","names":["LITTLE_ENDIAN","BIG_ENDIAN","SHP_HEADER_SIZE","SHP_RECORD_HEADER_SIZE","STATE","EXPECTING_HEADER","EXPECTING_RECORD","END","ERROR","SHPParser","options","BinaryChunkReader","maxRewindBytes","geometries","progress","bytesTotal","NaN","bytesUsed","rows","currentIndex","arrayBuffer","binaryReader","write","state","parseState","result","end","error","parseSHP","shpParser","parseSHPInBatches","asyncIterator","parser","headerReturned","header","length","dataView","getDataView","parseSHPHeader","hasAvailableBytes","recordHeaderView","recordHeader","recordNumber","getInt32","byteLength","type","rewind","invalidRecord","recordView","geometry","parseRecord","push","ended","message"],"sources":["../../../../src/lib/parsers/parse-shp.ts"],"sourcesContent":["import type {BinaryGeometry} from '@loaders.gl/schema';\nimport BinaryChunkReader from '../streaming/binary-chunk-reader';\nimport {parseSHPHeader, SHPHeader} from './parse-shp-header';\nimport {parseRecord} from './parse-shp-geometry';\nimport {SHPLoaderOptions} from './types';\n\nconst LITTLE_ENDIAN = true;\nconst BIG_ENDIAN = false;\n\nconst SHP_HEADER_SIZE = 100;\n// According to the spec, the record header is just 8 bytes, but here we set it\n// to 12 so that we can also access the record's type\nconst SHP_RECORD_HEADER_SIZE = 12;\n\nconst STATE = {\n EXPECTING_HEADER: 0,\n EXPECTING_RECORD: 1,\n END: 2,\n ERROR: 3\n};\n\ntype SHPResult = {\n geometries: (BinaryGeometry | null)[];\n header?: SHPHeader;\n error?: string;\n progress: {\n bytesUsed: number;\n bytesTotal: number;\n rows: number;\n };\n currentIndex: number;\n};\n\nclass SHPParser {\n options?: SHPLoaderOptions = {};\n binaryReader = new BinaryChunkReader({maxRewindBytes: SHP_RECORD_HEADER_SIZE});\n state = STATE.EXPECTING_HEADER;\n result: SHPResult = {\n geometries: [],\n // Initialize with number values to make TS happy\n // These are initialized for real in STATE.EXPECTING_HEADER\n progress: {\n bytesTotal: NaN,\n bytesUsed: NaN,\n rows: NaN\n },\n currentIndex: NaN\n };\n\n constructor(options?: SHPLoaderOptions) {\n this.options = options;\n }\n\n write(arrayBuffer: ArrayBuffer) {\n this.binaryReader.write(arrayBuffer);\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n }\n\n end() {\n this.binaryReader.end();\n this.state = parseState(this.state, this.result, this.binaryReader, this.options);\n // this.result.progress.bytesUsed = this.binaryReader.bytesUsed();\n if (this.state !== STATE.END) {\n this.state = STATE.ERROR;\n this.result.error = 'SHP incomplete file';\n }\n }\n}\n\nexport function parseSHP(arrayBuffer: ArrayBuffer, options?: object): BinaryGeometry[] {\n const shpParser = new SHPParser(options);\n shpParser.write(arrayBuffer);\n shpParser.end();\n\n // @ts-ignore\n return shpParser.result;\n}\n\n/**\n * @param asyncIterator\n * @param options\n * @returns\n */\nexport async function* parseSHPInBatches(\n asyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options?: object\n): AsyncIterable<BinaryGeometry | object> {\n const parser = new SHPParser(options);\n let headerReturned = false;\n for await (const arrayBuffer of asyncIterator) {\n parser.write(arrayBuffer);\n if (!headerReturned && parser.result.header) {\n headerReturned = true;\n yield parser.result.header;\n }\n\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n parser.result.geometries = [];\n }\n }\n parser.end();\n if (parser.result.geometries.length > 0) {\n yield parser.result.geometries;\n }\n\n return;\n}\n\n/**\n * State-machine parser for SHP data\n *\n * Note that whenever more data is needed, a `return`, not a `break`, is\n * necessary, as the `break` keeps the context within `parseState`, while\n * `return` releases context so that more data can be written into the\n * BinaryChunkReader.\n *\n * @param state Current state\n * @param result An object to hold result data\n * @param binaryReader\n * @return State at end of current parsing\n */\n/* eslint-disable complexity, max-depth */\nfunction parseState(\n state: number,\n result: SHPResult,\n binaryReader: BinaryChunkReader,\n options?: SHPLoaderOptions\n): number {\n // eslint-disable-next-line no-constant-condition\n while (true) {\n try {\n switch (state) {\n case STATE.ERROR:\n case STATE.END:\n return state;\n\n case STATE.EXPECTING_HEADER:\n // Parse initial file header\n const dataView = binaryReader.getDataView(SHP_HEADER_SIZE);\n if (!dataView) {\n return state;\n }\n\n result.header = parseSHPHeader(dataView);\n result.progress = {\n bytesUsed: 0,\n bytesTotal: result.header.length,\n rows: 0\n };\n // index numbering starts at 1\n result.currentIndex = 1;\n state = STATE.EXPECTING_RECORD;\n break;\n\n case STATE.EXPECTING_RECORD:\n while (binaryReader.hasAvailableBytes(SHP_RECORD_HEADER_SIZE)) {\n const recordHeaderView = binaryReader.getDataView(SHP_RECORD_HEADER_SIZE) as DataView;\n const recordHeader = {\n recordNumber: recordHeaderView.getInt32(0, BIG_ENDIAN),\n // 2 byte words; includes the four words of record header\n byteLength: recordHeaderView.getInt32(4, BIG_ENDIAN) * 2,\n // This is actually part of the record, not the header...\n type: recordHeaderView.getInt32(8, LITTLE_ENDIAN)\n };\n\n if (!binaryReader.hasAvailableBytes(recordHeader.byteLength - 4)) {\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE);\n return state;\n }\n\n const invalidRecord =\n recordHeader.byteLength < 4 ||\n recordHeader.type !== result.header?.type ||\n recordHeader.recordNumber !== result.currentIndex;\n\n // All records must have at least four bytes (for the record shape type)\n if (invalidRecord) {\n // Malformed record, try again, advancing just 4 bytes\n // Note: this is a rewind because binaryReader.getDataView above\n // moved the pointer forward 12 bytes, so rewinding 8 bytes still\n // leaves us 4 bytes ahead\n binaryReader.rewind(SHP_RECORD_HEADER_SIZE - 4);\n } else {\n // Note: type is actually part of the record, not the header, so\n // rewind 4 bytes before reading record\n binaryReader.rewind(4);\n\n const recordView = binaryReader.getDataView(recordHeader.byteLength) as DataView;\n const geometry = parseRecord(recordView, options);\n result.geometries.push(geometry);\n\n result.currentIndex++;\n result.progress.rows = result.currentIndex - 1;\n }\n }\n\n if (binaryReader.ended) {\n state = STATE.END;\n }\n\n return state;\n\n default:\n state = STATE.ERROR;\n result.error = `illegal parser state ${state}`;\n return state;\n }\n } catch (error) {\n state = STATE.ERROR;\n result.error = `SHP parsing failed: ${(error as Error)?.message}`;\n return state;\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;AACA;AACA;AACA;AAAiD;AAAA;AAGjD,IAAMA,aAAa,GAAG,IAAI;AAC1B,IAAMC,UAAU,GAAG,KAAK;AAExB,IAAMC,eAAe,GAAG,GAAG;AAG3B,IAAMC,sBAAsB,GAAG,EAAE;AAEjC,IAAMC,KAAK,GAAG;EACZC,gBAAgB,EAAE,CAAC;EACnBC,gBAAgB,EAAE,CAAC;EACnBC,GAAG,EAAE,CAAC;EACNC,KAAK,EAAE;AACT,CAAC;AAAC,IAcIC,SAAS;EAgBb,mBAAYC,OAA0B,EAAE;IAAA;IAAA,+CAfX,CAAC,CAAC;IAAA,oDAChB,IAAIC,0BAAiB,CAAC;MAACC,cAAc,EAAET;IAAsB,CAAC,CAAC;IAAA,6CACtEC,KAAK,CAACC,gBAAgB;IAAA,8CACV;MAClBQ,UAAU,EAAE,EAAE;MAGdC,QAAQ,EAAE;QACRC,UAAU,EAAEC,GAAG;QACfC,SAAS,EAAED,GAAG;QACdE,IAAI,EAAEF;MACR,CAAC;MACDG,YAAY,EAAEH;IAChB,CAAC;IAGC,IAAI,CAACN,OAAO,GAAGA,OAAO;EACxB;EAAC;IAAA;IAAA,OAED,eAAMU,WAAwB,EAAE;MAC9B,IAAI,CAACC,YAAY,CAACC,KAAK,CAACF,WAAW,CAAC;MACpC,IAAI,CAACG,KAAK,GAAGC,UAAU,CAAC,IAAI,CAACD,KAAK,EAAE,IAAI,CAACE,MAAM,EAAE,IAAI,CAACJ,YAAY,EAAE,IAAI,CAACX,OAAO,CAAC;IACnF;EAAC;IAAA;IAAA,OAED,eAAM;MACJ,IAAI,CAACW,YAAY,CAACK,GAAG,EAAE;MACvB,IAAI,CAACH,KAAK,GAAGC,UAAU,CAAC,IAAI,CAACD,KAAK,EAAE,IAAI,CAACE,MAAM,EAAE,IAAI,CAACJ,YAAY,EAAE,IAAI,CAACX,OAAO,CAAC;MAEjF,IAAI,IAAI,CAACa,KAAK,KAAKnB,KAAK,CAACG,GAAG,EAAE;QAC5B,IAAI,CAACgB,KAAK,GAAGnB,KAAK,CAACI,KAAK;QACxB,IAAI,CAACiB,MAAM,CAACE,KAAK,GAAG,qBAAqB;MAC3C;IACF;EAAC;EAAA;AAAA;AAGI,SAASC,QAAQ,CAACR,WAAwB,EAAEV,OAAgB,EAAoB;EACrF,IAAMmB,SAAS,GAAG,IAAIpB,SAAS,CAACC,OAAO,CAAC;EACxCmB,SAAS,CAACP,KAAK,CAACF,WAAW,CAAC;EAC5BS,SAAS,CAACH,GAAG,EAAE;;EAGf,OAAOG,SAAS,CAACJ,MAAM;AACzB;;AAAC,SAOsBK,iBAAiB;EAAA;AAAA;AAAA;EAAA,iFAAjC,iBACLC,aAAiE,EACjErB,OAAgB;IAAA;IAAA;MAAA;QAAA;UAAA;YAEVsB,MAAM,GAAG,IAAIvB,SAAS,CAACC,OAAO,CAAC;YACjCuB,cAAc,GAAG,KAAK;YAAA;YAAA;YAAA;YAAA,2BACMF,aAAa;UAAA;YAAA;YAAA;UAAA;YAAA;cAAA;cAAA;YAAA;YAA5BX,WAAW;YAC1BY,MAAM,CAACV,KAAK,CAACF,WAAW,CAAC;YAAC,MACtB,CAACa,cAAc,IAAID,MAAM,CAACP,MAAM,CAACS,MAAM;cAAA;cAAA;YAAA;YACzCD,cAAc,GAAG,IAAI;YAAC;YACtB,OAAMD,MAAM,CAACP,MAAM,CAACS,MAAM;UAAA;YAAA,MAGxBF,MAAM,CAACP,MAAM,CAACZ,UAAU,CAACsB,MAAM,GAAG,CAAC;cAAA;cAAA;YAAA;YAAA;YACrC,OAAMH,MAAM,CAACP,MAAM,CAACZ,UAAU;UAAA;YAC9BmB,MAAM,CAACP,MAAM,CAACZ,UAAU,GAAG,EAAE;UAAC;YAAA;YAAA;YAAA;UAAA;YAAA;YAAA;UAAA;YAAA;YAAA;YAAA;YAAA;UAAA;YAAA;YAAA;YAAA;cAAA;cAAA;YAAA;YAAA;YAAA;UAAA;YAAA;YAAA;cAAA;cAAA;YAAA;YAAA;UAAA;YAAA;UAAA;YAAA;UAAA;YAGlCmB,MAAM,CAACN,GAAG,EAAE;YAAC,MACTM,MAAM,CAACP,MAAM,CAACZ,UAAU,CAACsB,MAAM,GAAG,CAAC;cAAA;cAAA;YAAA;YAAA;YACrC,OAAMH,MAAM,CAACP,MAAM,CAACZ,UAAU;UAAA;YAAA;UAAA;UAAA;YAAA;QAAA;MAAA;IAAA;EAAA,CAIjC;EAAA;AAAA;AAgBD,SAASW,UAAU,CACjBD,KAAa,EACbE,MAAiB,EACjBJ,YAA+B,EAC/BX,OAA0B,EAClB;EAER,OAAO,IAAI,EAAE;IACX,IAAI;MACF,QAAQa,KAAK;QACX,KAAKnB,KAAK,CAACI,KAAK;QAChB,KAAKJ,KAAK,CAACG,GAAG;UACZ,OAAOgB,KAAK;QAEd,KAAKnB,KAAK,CAACC,gBAAgB;UAEzB,IAAM+B,QAAQ,GAAGf,YAAY,CAACgB,WAAW,CAACnC,eAAe,CAAC;UAC1D,IAAI,CAACkC,QAAQ,EAAE;YACb,OAAOb,KAAK;UACd;UAEAE,MAAM,CAACS,MAAM,GAAG,IAAAI,8BAAc,EAACF,QAAQ,CAAC;UACxCX,MAAM,CAACX,QAAQ,GAAG;YAChBG,SAAS,EAAE,CAAC;YACZF,UAAU,EAAEU,MAAM,CAACS,MAAM,CAACC,MAAM;YAChCjB,IAAI,EAAE;UACR,CAAC;UAEDO,MAAM,CAACN,YAAY,GAAG,CAAC;UACvBI,KAAK,GAAGnB,KAAK,CAACE,gBAAgB;UAC9B;QAEF,KAAKF,KAAK,CAACE,gBAAgB;UACzB,OAAOe,YAAY,CAACkB,iBAAiB,CAACpC,sBAAsB,CAAC,EAAE;YAAA;YAC7D,IAAMqC,gBAAgB,GAAGnB,YAAY,CAACgB,WAAW,CAAClC,sBAAsB,CAAa;YACrF,IAAMsC,YAAY,GAAG;cACnBC,YAAY,EAAEF,gBAAgB,CAACG,QAAQ,CAAC,CAAC,EAAE1C,UAAU,CAAC;cAEtD2C,UAAU,EAAEJ,gBAAgB,CAACG,QAAQ,CAAC,CAAC,EAAE1C,UAAU,CAAC,GAAG,CAAC;cAExD4C,IAAI,EAAEL,gBAAgB,CAACG,QAAQ,CAAC,CAAC,EAAE3C,aAAa;YAClD,CAAC;YAED,IAAI,CAACqB,YAAY,CAACkB,iBAAiB,CAACE,YAAY,CAACG,UAAU,GAAG,CAAC,CAAC,EAAE;cAChEvB,YAAY,CAACyB,MAAM,CAAC3C,sBAAsB,CAAC;cAC3C,OAAOoB,KAAK;YACd;YAEA,IAAMwB,aAAa,GACjBN,YAAY,CAACG,UAAU,GAAG,CAAC,IAC3BH,YAAY,CAACI,IAAI,wBAAKpB,MAAM,CAACS,MAAM,mDAAb,eAAeW,IAAI,KACzCJ,YAAY,CAACC,YAAY,KAAKjB,MAAM,CAACN,YAAY;;YAGnD,IAAI4B,aAAa,EAAE;cAKjB1B,YAAY,CAACyB,MAAM,CAAC3C,sBAAsB,GAAG,CAAC,CAAC;YACjD,CAAC,MAAM;cAGLkB,YAAY,CAACyB,MAAM,CAAC,CAAC,CAAC;cAEtB,IAAME,UAAU,GAAG3B,YAAY,CAACgB,WAAW,CAACI,YAAY,CAACG,UAAU,CAAa;cAChF,IAAMK,QAAQ,GAAG,IAAAC,6BAAW,EAACF,UAAU,EAAEtC,OAAO,CAAC;cACjDe,MAAM,CAACZ,UAAU,CAACsC,IAAI,CAACF,QAAQ,CAAC;cAEhCxB,MAAM,CAACN,YAAY,EAAE;cACrBM,MAAM,CAACX,QAAQ,CAACI,IAAI,GAAGO,MAAM,CAACN,YAAY,GAAG,CAAC;YAChD;UACF;UAEA,IAAIE,YAAY,CAAC+B,KAAK,EAAE;YACtB7B,KAAK,GAAGnB,KAAK,CAACG,GAAG;UACnB;UAEA,OAAOgB,KAAK;QAEd;UACEA,KAAK,GAAGnB,KAAK,CAACI,KAAK;UACnBiB,MAAM,CAACE,KAAK,kCAA2BJ,KAAK,CAAE;UAC9C,OAAOA,KAAK;MAAC;IAEnB,CAAC,CAAC,OAAOI,KAAK,EAAE;MACdJ,KAAK,GAAGnB,KAAK,CAACI,KAAK;MACnBiB,MAAM,CAACE,KAAK,iCAA2BA,KAAK,aAALA,KAAK,uBAALA,KAAK,CAAY0B,OAAO,CAAE;MACjE,OAAO9B,KAAK;IACd;EACF;AACF"}
@@ -4,9 +4,7 @@ Object.defineProperty(exports, "__esModule", {
4
4
  value: true
5
5
  });
6
6
  exports.parseShx = parseShx;
7
-
8
7
  var _parseShpHeader = require("./parse-shp-header");
9
-
10
8
  var SHX_HEADER_SIZE = 100;
11
9
  var BIG_ENDIAN = false;
12
10
 
@@ -17,12 +15,10 @@ function parseShx(arrayBuffer) {
17
15
  var contentView = new DataView(arrayBuffer, SHX_HEADER_SIZE, contentLength);
18
16
  var offsets = new Int32Array(contentLength);
19
17
  var lengths = new Int32Array(contentLength);
20
-
21
18
  for (var i = 0; i < contentLength / 8; i++) {
22
19
  offsets[i] = contentView.getInt32(i * 8, BIG_ENDIAN);
23
20
  lengths[i] = contentView.getInt32(i * 8 + 4, BIG_ENDIAN);
24
21
  }
25
-
26
22
  return {
27
23
  offsets: offsets,
28
24
  lengths: lengths
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/parsers/parse-shx.ts"],"names":["SHX_HEADER_SIZE","BIG_ENDIAN","parseShx","arrayBuffer","headerView","DataView","header","contentLength","length","contentView","offsets","Int32Array","lengths","i","getInt32"],"mappings":";;;;;;;AAAA;;AAOA,IAAMA,eAAe,GAAG,GAAxB;AACA,IAAMC,UAAU,GAAG,KAAnB;;AAMO,SAASC,QAAT,CAAkBC,WAAlB,EAAuD;AAE5D,MAAMC,UAAU,GAAG,IAAIC,QAAJ,CAAaF,WAAb,EAA0B,CAA1B,EAA6BH,eAA7B,CAAnB;AACA,MAAMM,MAAM,GAAG,oCAAeF,UAAf,CAAf;AACA,MAAMG,aAAa,GAAGD,MAAM,CAACE,MAAP,GAAgBR,eAAtC;AAEA,MAAMS,WAAW,GAAG,IAAIJ,QAAJ,CAAaF,WAAb,EAA0BH,eAA1B,EAA2CO,aAA3C,CAApB;AAEA,MAAMG,OAAO,GAAG,IAAIC,UAAJ,CAAeJ,aAAf,CAAhB;AACA,MAAMK,OAAO,GAAG,IAAID,UAAJ,CAAeJ,aAAf,CAAhB;;AAEA,OAAK,IAAIM,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAGN,aAAa,GAAG,CAApC,EAAuCM,CAAC,EAAxC,EAA4C;AAC1CH,IAAAA,OAAO,CAACG,CAAD,CAAP,GAAaJ,WAAW,CAACK,QAAZ,CAAqBD,CAAC,GAAG,CAAzB,EAA4BZ,UAA5B,CAAb;AACAW,IAAAA,OAAO,CAACC,CAAD,CAAP,GAAaJ,WAAW,CAACK,QAAZ,CAAqBD,CAAC,GAAG,CAAJ,GAAQ,CAA7B,EAAgCZ,UAAhC,CAAb;AACD;;AAED,SAAO;AACLS,IAAAA,OAAO,EAAPA,OADK;AAELE,IAAAA,OAAO,EAAPA;AAFK,GAAP;AAID","sourcesContent":["import {parseSHPHeader} from './parse-shp-header';\n\nexport interface SHXOutput {\n offsets: Int32Array;\n lengths: Int32Array;\n}\n\nconst SHX_HEADER_SIZE = 100;\nconst BIG_ENDIAN = false;\n\n/**\n * @param arrayBuffer\n * @returns SHXOutput\n */\nexport function parseShx(arrayBuffer: ArrayBuffer): SHXOutput {\n // SHX header is identical to SHP Header\n const headerView = new DataView(arrayBuffer, 0, SHX_HEADER_SIZE);\n const header = parseSHPHeader(headerView);\n const contentLength = header.length - SHX_HEADER_SIZE;\n\n const contentView = new DataView(arrayBuffer, SHX_HEADER_SIZE, contentLength);\n\n const offsets = new Int32Array(contentLength);\n const lengths = new Int32Array(contentLength);\n\n for (let i = 0; i < contentLength / 8; i++) {\n offsets[i] = contentView.getInt32(i * 8, BIG_ENDIAN);\n lengths[i] = contentView.getInt32(i * 8 + 4, BIG_ENDIAN);\n }\n\n return {\n offsets,\n lengths\n };\n}\n"],"file":"parse-shx.js"}
1
+ {"version":3,"file":"parse-shx.js","names":["SHX_HEADER_SIZE","BIG_ENDIAN","parseShx","arrayBuffer","headerView","DataView","header","parseSHPHeader","contentLength","length","contentView","offsets","Int32Array","lengths","i","getInt32"],"sources":["../../../../src/lib/parsers/parse-shx.ts"],"sourcesContent":["import {parseSHPHeader} from './parse-shp-header';\n\nexport interface SHXOutput {\n offsets: Int32Array;\n lengths: Int32Array;\n}\n\nconst SHX_HEADER_SIZE = 100;\nconst BIG_ENDIAN = false;\n\n/**\n * @param arrayBuffer\n * @returns SHXOutput\n */\nexport function parseShx(arrayBuffer: ArrayBuffer): SHXOutput {\n // SHX header is identical to SHP Header\n const headerView = new DataView(arrayBuffer, 0, SHX_HEADER_SIZE);\n const header = parseSHPHeader(headerView);\n const contentLength = header.length - SHX_HEADER_SIZE;\n\n const contentView = new DataView(arrayBuffer, SHX_HEADER_SIZE, contentLength);\n\n const offsets = new Int32Array(contentLength);\n const lengths = new Int32Array(contentLength);\n\n for (let i = 0; i < contentLength / 8; i++) {\n offsets[i] = contentView.getInt32(i * 8, BIG_ENDIAN);\n lengths[i] = contentView.getInt32(i * 8 + 4, BIG_ENDIAN);\n }\n\n return {\n offsets,\n lengths\n };\n}\n"],"mappings":";;;;;;AAAA;AAOA,IAAMA,eAAe,GAAG,GAAG;AAC3B,IAAMC,UAAU,GAAG,KAAK;;AAMjB,SAASC,QAAQ,CAACC,WAAwB,EAAa;EAE5D,IAAMC,UAAU,GAAG,IAAIC,QAAQ,CAACF,WAAW,EAAE,CAAC,EAAEH,eAAe,CAAC;EAChE,IAAMM,MAAM,GAAG,IAAAC,8BAAc,EAACH,UAAU,CAAC;EACzC,IAAMI,aAAa,GAAGF,MAAM,CAACG,MAAM,GAAGT,eAAe;EAErD,IAAMU,WAAW,GAAG,IAAIL,QAAQ,CAACF,WAAW,EAAEH,eAAe,EAAEQ,aAAa,CAAC;EAE7E,IAAMG,OAAO,GAAG,IAAIC,UAAU,CAACJ,aAAa,CAAC;EAC7C,IAAMK,OAAO,GAAG,IAAID,UAAU,CAACJ,aAAa,CAAC;EAE7C,KAAK,IAAIM,CAAC,GAAG,CAAC,EAAEA,CAAC,GAAGN,aAAa,GAAG,CAAC,EAAEM,CAAC,EAAE,EAAE;IAC1CH,OAAO,CAACG,CAAC,CAAC,GAAGJ,WAAW,CAACK,QAAQ,CAACD,CAAC,GAAG,CAAC,EAAEb,UAAU,CAAC;IACpDY,OAAO,CAACC,CAAC,CAAC,GAAGJ,WAAW,CAACK,QAAQ,CAACD,CAAC,GAAG,CAAC,GAAG,CAAC,EAAEb,UAAU,CAAC;EAC1D;EAEA,OAAO;IACLU,OAAO,EAAPA,OAAO;IACPE,OAAO,EAAPA;EACF,CAAC;AACH"}
@@ -1 +1 @@
1
- {"version":3,"sources":[],"names":[],"mappings":"","sourcesContent":[],"file":"types.js"}
1
+ {"version":3,"file":"types.js","names":[],"sources":["../../../../src/lib/parsers/types.ts"],"sourcesContent":["import {Schema, ObjectRowTable} from '@loaders.gl/schema';\nimport type {LoaderOptions} from '@loaders.gl/loader-utils';\n\nexport type SHPLoaderOptions = LoaderOptions & {\n shp?: {\n _maxDimensions?: number;\n };\n};\n\nexport type DBFLoaderOptions = LoaderOptions & {\n dbf?: {\n encoding?: string;\n shape?: 'rows' | 'table' | 'object-row-table';\n };\n};\n\nexport type ShapefileLoaderOptions = LoaderOptions &\n SHPLoaderOptions & {\n shapefile?: {\n shape?: 'geojson';\n };\n gis?: {\n reproject?: boolean;\n _targetCrs?: string;\n /** @deprecated. Use options.shapefile.shape */\n format?: 'geojson';\n };\n };\n\nexport type DBFRowsOutput = ObjectRowTable['data'];\n\n/**\n * DBF Table output. Deprecated in favor of ObjectRowTable\n * @deprecated\n */\nexport interface DBFTableOutput {\n schema?: Schema;\n rows: DBFRowsOutput;\n}\n\nexport type DBFHeader = {\n // Last updated date\n year: number;\n month: number;\n day: number;\n // Number of records in data file\n nRecords: number;\n // Length of header in bytes\n headerLength: number;\n // Length of each record\n recordLength: number;\n // Not sure if this is usually set\n languageDriver: number;\n};\n\nexport type DBFField = {\n name: string;\n dataType: string;\n fieldLength: number;\n decimal: number;\n};\n\nexport type DBFResult = {\n data: {[key: string]: any}[];\n schema?: Schema;\n error?: string;\n dbfHeader?: DBFHeader;\n dbfFields?: DBFField[];\n progress?: {\n bytesUsed: number;\n rowsTotal: number;\n rows: number;\n };\n};\n"],"mappings":""}
@@ -1,26 +1,17 @@
1
1
  "use strict";
2
2
 
3
3
  var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
4
-
5
4
  Object.defineProperty(exports, "__esModule", {
6
5
  value: true
7
6
  });
8
7
  exports.default = void 0;
9
-
10
8
  var _slicedToArray2 = _interopRequireDefault(require("@babel/runtime/helpers/slicedToArray"));
11
-
12
9
  var _classCallCheck2 = _interopRequireDefault(require("@babel/runtime/helpers/classCallCheck"));
13
-
14
10
  var _createClass2 = _interopRequireDefault(require("@babel/runtime/helpers/createClass"));
15
-
16
11
  var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/defineProperty"));
17
-
18
12
  function _createForOfIteratorHelper(o, allowArrayLike) { var it = typeof Symbol !== "undefined" && o[Symbol.iterator] || o["@@iterator"]; if (!it) { if (Array.isArray(o) || (it = _unsupportedIterableToArray(o)) || allowArrayLike && o && typeof o.length === "number") { if (it) o = it; var i = 0; var F = function F() {}; return { s: F, n: function n() { if (i >= o.length) return { done: true }; return { done: false, value: o[i++] }; }, e: function e(_e) { throw _e; }, f: F }; } throw new TypeError("Invalid attempt to iterate non-iterable instance.\nIn order to be iterable, non-array objects must have a [Symbol.iterator]() method."); } var normalCompletion = true, didErr = false, err; return { s: function s() { it = it.call(o); }, n: function n() { var step = it.next(); normalCompletion = step.done; return step; }, e: function e(_e2) { didErr = true; err = _e2; }, f: function f() { try { if (!normalCompletion && it.return != null) it.return(); } finally { if (didErr) throw err; } } }; }
19
-
20
13
  function _unsupportedIterableToArray(o, minLen) { if (!o) return; if (typeof o === "string") return _arrayLikeToArray(o, minLen); var n = Object.prototype.toString.call(o).slice(8, -1); if (n === "Object" && o.constructor) n = o.constructor.name; if (n === "Map" || n === "Set") return Array.from(o); if (n === "Arguments" || /^(?:Ui|I)nt(?:8|16|32)(?:Clamped)?Array$/.test(n)) return _arrayLikeToArray(o, minLen); }
21
-
22
14
  function _arrayLikeToArray(arr, len) { if (len == null || len > arr.length) len = arr.length; for (var i = 0, arr2 = new Array(len); i < len; i++) { arr2[i] = arr[i]; } return arr2; }
23
-
24
15
  var BinaryChunkReader = function () {
25
16
  function BinaryChunkReader(options) {
26
17
  (0, _classCallCheck2.default)(this, BinaryChunkReader);
@@ -28,20 +19,20 @@ var BinaryChunkReader = function () {
28
19
  (0, _defineProperty2.default)(this, "arrayBuffers", void 0);
29
20
  (0, _defineProperty2.default)(this, "ended", void 0);
30
21
  (0, _defineProperty2.default)(this, "maxRewindBytes", void 0);
31
-
32
22
  var _ref = options || {},
33
- _ref$maxRewindBytes = _ref.maxRewindBytes,
34
- maxRewindBytes = _ref$maxRewindBytes === void 0 ? 0 : _ref$maxRewindBytes;
23
+ _ref$maxRewindBytes = _ref.maxRewindBytes,
24
+ maxRewindBytes = _ref$maxRewindBytes === void 0 ? 0 : _ref$maxRewindBytes;
35
25
 
36
26
  this.offset = 0;
37
27
  this.arrayBuffers = [];
38
28
  this.ended = false;
29
+
39
30
  this.maxRewindBytes = maxRewindBytes;
40
31
  }
41
-
42
32
  (0, _createClass2.default)(BinaryChunkReader, [{
43
33
  key: "write",
44
- value: function write(arrayBuffer) {
34
+ value:
35
+ function write(arrayBuffer) {
45
36
  this.arrayBuffers.push(arrayBuffer);
46
37
  }
47
38
  }, {
@@ -50,19 +41,18 @@ var BinaryChunkReader = function () {
50
41
  this.arrayBuffers = [];
51
42
  this.ended = true;
52
43
  }
44
+
53
45
  }, {
54
46
  key: "hasAvailableBytes",
55
- value: function hasAvailableBytes(bytes) {
47
+ value:
48
+ function hasAvailableBytes(bytes) {
56
49
  var bytesAvailable = -this.offset;
57
-
58
50
  var _iterator = _createForOfIteratorHelper(this.arrayBuffers),
59
- _step;
60
-
51
+ _step;
61
52
  try {
62
53
  for (_iterator.s(); !(_step = _iterator.n()).done;) {
63
54
  var arrayBuffer = _step.value;
64
55
  bytesAvailable += arrayBuffer.byteLength;
65
-
66
56
  if (bytesAvailable >= bytes) {
67
57
  return true;
68
58
  }
@@ -72,15 +62,15 @@ var BinaryChunkReader = function () {
72
62
  } finally {
73
63
  _iterator.f();
74
64
  }
75
-
76
65
  return false;
77
66
  }
67
+
78
68
  }, {
79
69
  key: "findBufferOffsets",
80
- value: function findBufferOffsets(bytes) {
70
+ value:
71
+ function findBufferOffsets(bytes) {
81
72
  var offset = -this.offset;
82
73
  var selectedBuffers = [];
83
-
84
74
  for (var i = 0; i < this.arrayBuffers.length; i++) {
85
75
  var buf = this.arrayBuffers[i];
86
76
 
@@ -100,36 +90,34 @@ var BinaryChunkReader = function () {
100
90
 
101
91
  end = buf.byteLength;
102
92
  selectedBuffers.push([i, [start, end]]);
93
+
103
94
  bytes -= buf.byteLength - start;
104
95
  offset += buf.byteLength;
105
96
  }
106
97
 
107
98
  return null;
108
99
  }
100
+
109
101
  }, {
110
102
  key: "getDataView",
111
- value: function getDataView(bytes) {
103
+ value:
104
+ function getDataView(bytes) {
112
105
  var bufferOffsets = this.findBufferOffsets(bytes);
113
-
114
106
  if (!bufferOffsets && this.ended) {
115
107
  throw new Error('binary data exhausted');
116
108
  }
117
-
118
109
  if (!bufferOffsets) {
119
110
  return null;
120
111
  }
121
112
 
122
113
  if (bufferOffsets.length === 1) {
123
114
  var _bufferOffsets$ = (0, _slicedToArray2.default)(bufferOffsets[0], 2),
124
- bufferIndex = _bufferOffsets$[0],
125
- _bufferOffsets$$ = (0, _slicedToArray2.default)(_bufferOffsets$[1], 2),
126
- start = _bufferOffsets$$[0],
127
- end = _bufferOffsets$$[1];
128
-
115
+ bufferIndex = _bufferOffsets$[0],
116
+ _bufferOffsets$$ = (0, _slicedToArray2.default)(_bufferOffsets$[1], 2),
117
+ start = _bufferOffsets$$[0],
118
+ end = _bufferOffsets$$[1];
129
119
  var arrayBuffer = this.arrayBuffers[bufferIndex];
130
-
131
120
  var _view = new DataView(arrayBuffer, start, end - start);
132
-
133
121
  this.offset += bytes;
134
122
  this.disposeBuffers();
135
123
  return _view;
@@ -140,30 +128,30 @@ var BinaryChunkReader = function () {
140
128
  this.disposeBuffers();
141
129
  return view;
142
130
  }
131
+
143
132
  }, {
144
133
  key: "disposeBuffers",
145
- value: function disposeBuffers() {
134
+ value:
135
+ function disposeBuffers() {
146
136
  while (this.arrayBuffers.length > 0 && this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength) {
147
137
  this.offset -= this.arrayBuffers[0].byteLength;
148
138
  this.arrayBuffers.shift();
149
139
  }
150
140
  }
141
+
151
142
  }, {
152
143
  key: "_combineArrayBuffers",
153
- value: function _combineArrayBuffers(bufferOffsets) {
144
+ value:
145
+ function _combineArrayBuffers(bufferOffsets) {
154
146
  var byteLength = 0;
155
-
156
147
  var _iterator2 = _createForOfIteratorHelper(bufferOffsets),
157
- _step2;
158
-
148
+ _step2;
159
149
  try {
160
150
  for (_iterator2.s(); !(_step2 = _iterator2.n()).done;) {
161
151
  var bufferOffset = _step2.value;
162
-
163
152
  var _bufferOffset$ = (0, _slicedToArray2.default)(bufferOffset[1], 2),
164
- start = _bufferOffset$[0],
165
- end = _bufferOffset$[1];
166
-
153
+ start = _bufferOffset$[0],
154
+ end = _bufferOffset$[1];
167
155
  byteLength += end - start;
168
156
  }
169
157
  } catch (err) {
@@ -171,23 +159,19 @@ var BinaryChunkReader = function () {
171
159
  } finally {
172
160
  _iterator2.f();
173
161
  }
174
-
175
162
  var result = new Uint8Array(byteLength);
176
- var resultOffset = 0;
177
163
 
164
+ var resultOffset = 0;
178
165
  var _iterator3 = _createForOfIteratorHelper(bufferOffsets),
179
- _step3;
180
-
166
+ _step3;
181
167
  try {
182
168
  for (_iterator3.s(); !(_step3 = _iterator3.n()).done;) {
183
169
  var _bufferOffset = _step3.value;
184
-
185
170
  var _bufferOffset2 = (0, _slicedToArray2.default)(_bufferOffset, 2),
186
- bufferIndex = _bufferOffset2[0],
187
- _bufferOffset2$ = (0, _slicedToArray2.default)(_bufferOffset2[1], 2),
188
- _start = _bufferOffset2$[0],
189
- _end = _bufferOffset2$[1];
190
-
171
+ bufferIndex = _bufferOffset2[0],
172
+ _bufferOffset2$ = (0, _slicedToArray2.default)(_bufferOffset2[1], 2),
173
+ _start = _bufferOffset2$[0],
174
+ _end = _bufferOffset2$[1];
191
175
  var sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);
192
176
  result.set(sourceArray.subarray(_start, _end), resultOffset);
193
177
  resultOffset += _end - _start;
@@ -197,22 +181,22 @@ var BinaryChunkReader = function () {
197
181
  } finally {
198
182
  _iterator3.f();
199
183
  }
200
-
201
184
  return result.buffer;
202
185
  }
203
186
  }, {
204
187
  key: "skip",
205
- value: function skip(bytes) {
188
+ value:
189
+ function skip(bytes) {
206
190
  this.offset += bytes;
207
191
  }
208
192
  }, {
209
193
  key: "rewind",
210
- value: function rewind(bytes) {
194
+ value:
195
+ function rewind(bytes) {
211
196
  this.offset -= bytes;
212
197
  }
213
198
  }]);
214
199
  return BinaryChunkReader;
215
200
  }();
216
-
217
201
  exports.default = BinaryChunkReader;
218
202
  //# sourceMappingURL=binary-chunk-reader.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/streaming/binary-chunk-reader.ts"],"names":["BinaryChunkReader","options","maxRewindBytes","offset","arrayBuffers","ended","arrayBuffer","push","bytes","bytesAvailable","byteLength","selectedBuffers","i","length","buf","start","Math","abs","end","bufferOffsets","findBufferOffsets","Error","bufferIndex","view","DataView","disposeBuffers","_combineArrayBuffers","shift","bufferOffset","result","Uint8Array","resultOffset","sourceArray","set","subarray","buffer"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;IAIqBA,iB;AAMnB,6BAAYC,OAAZ,EAAgD;AAAA;AAAA;AAAA;AAAA;AAAA;;AAC9C,eAA6BA,OAAO,IAAI,EAAxC;AAAA,mCAAOC,cAAP;AAAA,QAAOA,cAAP,oCAAwB,CAAxB;;AAGA,SAAKC,MAAL,GAAc,CAAd;AAEA,SAAKC,YAAL,GAAoB,EAApB;AACA,SAAKC,KAAL,GAAa,KAAb;AAGA,SAAKH,cAAL,GAAsBA,cAAtB;AACD;;;;WAID,eAAMI,WAAN,EAAsC;AACpC,WAAKF,YAAL,CAAkBG,IAAlB,CAAuBD,WAAvB;AACD;;;WAED,eAAY;AACV,WAAKF,YAAL,GAAoB,EAApB;AACA,WAAKC,KAAL,GAAa,IAAb;AACD;;;WAQD,2BAAkBG,KAAlB,EAA0C;AACxC,UAAIC,cAAc,GAAG,CAAC,KAAKN,MAA3B;;AADwC,iDAEd,KAAKC,YAFS;AAAA;;AAAA;AAExC,4DAA6C;AAAA,cAAlCE,WAAkC;AAC3CG,UAAAA,cAAc,IAAIH,WAAW,CAACI,UAA9B;;AACA,cAAID,cAAc,IAAID,KAAtB,EAA6B;AAC3B,mBAAO,IAAP;AACD;AACF;AAPuC;AAAA;AAAA;AAAA;AAAA;;AAQxC,aAAO,KAAP;AACD;;;WAQD,2BAAkBA,KAAlB,EAA+C;AAC7C,UAAIL,MAAM,GAAG,CAAC,KAAKA,MAAnB;AACA,UAAMQ,eAAoB,GAAG,EAA7B;;AAEA,WAAK,IAAIC,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAG,KAAKR,YAAL,CAAkBS,MAAtC,EAA8CD,CAAC,EAA/C,EAAmD;AACjD,YAAME,GAAG,GAAG,KAAKV,YAAL,CAAkBQ,CAAlB,CAAZ;;AAGA,YAAIT,MAAM,GAAGW,GAAG,CAACJ,UAAb,IAA2B,CAA/B,EAAkC;AAChCP,UAAAA,MAAM,IAAIW,GAAG,CAACJ,UAAd;AAEA;AACD;;AAKD,YAAMK,KAAK,GAAGZ,MAAM,IAAI,CAAV,GAAca,IAAI,CAACC,GAAL,CAASd,MAAT,CAAd,GAAiC,CAA/C;AACA,YAAIe,GAAW,SAAf;;AAGA,YAAIH,KAAK,GAAGP,KAAR,IAAiBM,GAAG,CAACJ,UAAzB,EAAqC;AACnCQ,UAAAA,GAAG,GAAGH,KAAK,GAAGP,KAAd;AACAG,UAAAA,eAAe,CAACJ,IAAhB,CAAqB,CAACK,CAAD,EAAI,CAACG,KAAD,EAAQG,GAAR,CAAJ,CAArB;AACA,iBAAOP,eAAP;AACD;;AAGDO,QAAAA,GAAG,GAAGJ,GAAG,CAACJ,UAAV;AACAC,QAAAA,eAAe,CAACJ,IAAhB,CAAqB,CAACK,CAAD,EAAI,CAACG,KAAD,EAAQG,GAAR,CAAJ,CAArB;AAGAV,QAAAA,KAAK,IAAIM,GAAG,CAACJ,UAAJ,GAAiBK,KAA1B;AACAZ,QAAAA,MAAM,IAAIW,GAAG,CAACJ,UAAd;AACD;;AAGD,aAAO,IAAP;AACD;;;WAQD,qBAAYF,KAAZ,EAA4C;AAC1C,UAAMW,aAAa,GAAG,KAAKC,iBAAL,CAAuBZ,KAAvB,CAAtB;;AAGA,UAAI,CAACW,aAAD,IAAkB,KAAKd,KAA3B,EAAkC;AAChC,cAAM,IAAIgB,KAAJ,CAAU,uBAAV,CAAN;AACD;;AAED,UAAI,CAACF,aAAL,EAAoB;AAElB,eAAO,IAAP;AACD;;AAGD,UAAIA,aAAa,CAACN,MAAd,KAAyB,CAA7B,EAAgC;AAC9B,2DAAoCM,aAAa,CAAC,CAAD,CAAjD;AAAA,YAAOG,WAAP;AAAA;AAAA,YAAqBP,KAArB;AAAA,YAA4BG,GAA5B;;AACA,YAAMZ,WAAW,GAAG,KAAKF,YAAL,CAAkBkB,WAAlB,CAApB;;AACA,YAAMC,KAAI,GAAG,IAAIC,QAAJ,CAAalB,WAAb,EAA0BS,KAA1B,EAAiCG,GAAG,GAAGH,KAAvC,CAAb;;AAEA,aAAKZ,MAAL,IAAeK,KAAf;AACA,aAAKiB,cAAL;AACA,eAAOF,KAAP;AACD;;AAGD,UAAMA,IAAI,GAAG,IAAIC,QAAJ,CAAa,KAAKE,oBAAL,CAA0BP,aAA1B,CAAb,CAAb;AACA,WAAKhB,MAAL,IAAeK,KAAf;AACA,WAAKiB,cAAL;AACA,aAAOF,IAAP;AACD;;;WAKD,0BAAuB;AACrB,aACE,KAAKnB,YAAL,CAAkBS,MAAlB,GAA2B,CAA3B,IACA,KAAKV,MAAL,GAAc,KAAKD,cAAnB,IAAqC,KAAKE,YAAL,CAAkB,CAAlB,EAAqBM,UAF5D,EAGE;AACA,aAAKP,MAAL,IAAe,KAAKC,YAAL,CAAkB,CAAlB,EAAqBM,UAApC;AACA,aAAKN,YAAL,CAAkBuB,KAAlB;AACD;AACF;;;WAYD,8BAAqBR,aAArB,EAA4D;AAC1D,UAAIT,UAAkB,GAAG,CAAzB;;AAD0D,kDAE/BS,aAF+B;AAAA;;AAAA;AAE1D,+DAA0C;AAAA,cAA/BS,YAA+B;;AACxC,4DAAqBA,YAAY,CAAC,CAAD,CAAjC;AAAA,cAAOb,KAAP;AAAA,cAAcG,GAAd;;AACAR,UAAAA,UAAU,IAAIQ,GAAG,GAAGH,KAApB;AACD;AALyD;AAAA;AAAA;AAAA;AAAA;;AAO1D,UAAMc,MAAM,GAAG,IAAIC,UAAJ,CAAepB,UAAf,CAAf;AAGA,UAAIqB,YAAoB,GAAG,CAA3B;;AAV0D,kDAW/BZ,aAX+B;AAAA;;AAAA;AAW1D,+DAA0C;AAAA,cAA/BS,aAA+B;;AACxC,4DAAoCA,aAApC;AAAA,cAAON,WAAP;AAAA;AAAA,cAAqBP,MAArB;AAAA,cAA4BG,IAA5B;;AACA,cAAMc,WAAW,GAAG,IAAIF,UAAJ,CAAe,KAAK1B,YAAL,CAAkBkB,WAAlB,CAAf,CAApB;AACAO,UAAAA,MAAM,CAACI,GAAP,CAAWD,WAAW,CAACE,QAAZ,CAAqBnB,MAArB,EAA4BG,IAA5B,CAAX,EAA6Ca,YAA7C;AACAA,UAAAA,YAAY,IAAIb,IAAG,GAAGH,MAAtB;AACD;AAhByD;AAAA;AAAA;AAAA;AAAA;;AAkB1D,aAAOc,MAAM,CAACM,MAAd;AACD;;;WAID,cAAK3B,KAAL,EAA0B;AACxB,WAAKL,MAAL,IAAeK,KAAf;AACD;;;WAID,gBAAOA,KAAP,EAA4B;AAE1B,WAAKL,MAAL,IAAeK,KAAf;AACD","sourcesContent":["type BinaryChunkReaderOptions = {\n maxRewindBytes: number;\n};\n\nexport default class BinaryChunkReader {\n offset: number;\n arrayBuffers: ArrayBuffer[];\n ended: boolean;\n maxRewindBytes: number;\n\n constructor(options?: BinaryChunkReaderOptions) {\n const {maxRewindBytes = 0} = options || {};\n\n /** current global offset into current array buffer*/\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffers = [];\n this.ended = false;\n\n /** bytes behind offset to hold on to */\n this.maxRewindBytes = maxRewindBytes;\n }\n /**\n * @param arrayBuffer\n */\n write(arrayBuffer: ArrayBuffer): void {\n this.arrayBuffers.push(arrayBuffer);\n }\n\n end(): void {\n this.arrayBuffers = [];\n this.ended = true;\n }\n\n /**\n * Has enough bytes available in array buffers\n *\n * @param bytes Number of bytes\n * @return boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n let bytesAvailable = -this.offset;\n for (const arrayBuffer of this.arrayBuffers) {\n bytesAvailable += arrayBuffer.byteLength;\n if (bytesAvailable >= bytes) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * Find offsets of byte ranges within this.arrayBuffers\n *\n * @param bytes Byte length to read\n * @return Arrays with byte ranges pointing to this.arrayBuffers, Output type is nested array, e.g. [ [0, [1, 2]], ...]\n */\n findBufferOffsets(bytes: number): any[] | null {\n let offset = -this.offset;\n const selectedBuffers: any = [];\n\n for (let i = 0; i < this.arrayBuffers.length; i++) {\n const buf = this.arrayBuffers[i];\n\n // Current buffer isn't long enough to reach global offset\n if (offset + buf.byteLength <= 0) {\n offset += buf.byteLength;\n // eslint-disable-next-line no-continue\n continue;\n }\n\n // Find start/end offsets for this buffer\n // When offset < 0, need to skip over Math.abs(offset) bytes\n // When offset > 0, implies bytes in previous buffer, start at 0\n const start = offset <= 0 ? Math.abs(offset) : 0;\n let end: number;\n\n // Length of requested bytes is contained in current buffer\n if (start + bytes <= buf.byteLength) {\n end = start + bytes;\n selectedBuffers.push([i, [start, end]]);\n return selectedBuffers;\n }\n\n // Will need to look into next buffer\n end = buf.byteLength;\n selectedBuffers.push([i, [start, end]]);\n\n // Need to read fewer bytes in next iter\n bytes -= buf.byteLength - start;\n offset += buf.byteLength;\n }\n\n // Should only finish loop if exhausted all arrays\n return null;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes Number of bytes\n * @return DataView with data\n */\n getDataView(bytes: number): DataView | null {\n const bufferOffsets = this.findBufferOffsets(bytes);\n // return `null` if not enough data, except if end() already called, in\n // which case throw an error.\n if (!bufferOffsets && this.ended) {\n throw new Error('binary data exhausted');\n }\n\n if (!bufferOffsets) {\n // @ts-ignore\n return null;\n }\n\n // If only one arrayBuffer needed, return DataView directly\n if (bufferOffsets.length === 1) {\n const [bufferIndex, [start, end]] = bufferOffsets[0];\n const arrayBuffer = this.arrayBuffers[bufferIndex];\n const view = new DataView(arrayBuffer, start, end - start);\n\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n // Concatenate portions of multiple ArrayBuffers\n const view = new DataView(this._combineArrayBuffers(bufferOffsets));\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n /**\n * Dispose of old array buffers\n */\n disposeBuffers(): void {\n while (\n this.arrayBuffers.length > 0 &&\n this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength\n ) {\n this.offset -= this.arrayBuffers[0].byteLength;\n this.arrayBuffers.shift();\n }\n }\n\n /**\n * Copy multiple ArrayBuffers into one contiguous ArrayBuffer\n *\n * In contrast to concatenateArrayBuffers, this only copies the necessary\n * portions of the source arrays, rather than first copying the entire arrays\n * then taking a part of them.\n *\n * @param bufferOffsets List of internal array offsets\n * @return New contiguous ArrayBuffer\n */\n _combineArrayBuffers(bufferOffsets: any[]): ArrayBufferLike {\n let byteLength: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [start, end] = bufferOffset[1];\n byteLength += end - start;\n }\n\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let resultOffset: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [bufferIndex, [start, end]] = bufferOffset;\n const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);\n result.set(sourceArray.subarray(start, end), resultOffset);\n resultOffset += end - start;\n }\n\n return result.buffer;\n }\n /**\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n /**\n * @param bytes\n */\n rewind(bytes: number): void {\n // TODO - only works if offset is already set\n this.offset -= bytes;\n }\n}\n"],"file":"binary-chunk-reader.js"}
1
+ {"version":3,"file":"binary-chunk-reader.js","names":["BinaryChunkReader","options","maxRewindBytes","offset","arrayBuffers","ended","arrayBuffer","push","bytes","bytesAvailable","byteLength","selectedBuffers","i","length","buf","start","Math","abs","end","bufferOffsets","findBufferOffsets","Error","bufferIndex","view","DataView","disposeBuffers","_combineArrayBuffers","shift","bufferOffset","result","Uint8Array","resultOffset","sourceArray","set","subarray","buffer"],"sources":["../../../../src/lib/streaming/binary-chunk-reader.ts"],"sourcesContent":["type BinaryChunkReaderOptions = {\n maxRewindBytes: number;\n};\n\nexport default class BinaryChunkReader {\n offset: number;\n arrayBuffers: ArrayBuffer[];\n ended: boolean;\n maxRewindBytes: number;\n\n constructor(options?: BinaryChunkReaderOptions) {\n const {maxRewindBytes = 0} = options || {};\n\n /** current global offset into current array buffer*/\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffers = [];\n this.ended = false;\n\n /** bytes behind offset to hold on to */\n this.maxRewindBytes = maxRewindBytes;\n }\n /**\n * @param arrayBuffer\n */\n write(arrayBuffer: ArrayBuffer): void {\n this.arrayBuffers.push(arrayBuffer);\n }\n\n end(): void {\n this.arrayBuffers = [];\n this.ended = true;\n }\n\n /**\n * Has enough bytes available in array buffers\n *\n * @param bytes Number of bytes\n * @return boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n let bytesAvailable = -this.offset;\n for (const arrayBuffer of this.arrayBuffers) {\n bytesAvailable += arrayBuffer.byteLength;\n if (bytesAvailable >= bytes) {\n return true;\n }\n }\n return false;\n }\n\n /**\n * Find offsets of byte ranges within this.arrayBuffers\n *\n * @param bytes Byte length to read\n * @return Arrays with byte ranges pointing to this.arrayBuffers, Output type is nested array, e.g. [ [0, [1, 2]], ...]\n */\n findBufferOffsets(bytes: number): any[] | null {\n let offset = -this.offset;\n const selectedBuffers: any = [];\n\n for (let i = 0; i < this.arrayBuffers.length; i++) {\n const buf = this.arrayBuffers[i];\n\n // Current buffer isn't long enough to reach global offset\n if (offset + buf.byteLength <= 0) {\n offset += buf.byteLength;\n // eslint-disable-next-line no-continue\n continue;\n }\n\n // Find start/end offsets for this buffer\n // When offset < 0, need to skip over Math.abs(offset) bytes\n // When offset > 0, implies bytes in previous buffer, start at 0\n const start = offset <= 0 ? Math.abs(offset) : 0;\n let end: number;\n\n // Length of requested bytes is contained in current buffer\n if (start + bytes <= buf.byteLength) {\n end = start + bytes;\n selectedBuffers.push([i, [start, end]]);\n return selectedBuffers;\n }\n\n // Will need to look into next buffer\n end = buf.byteLength;\n selectedBuffers.push([i, [start, end]]);\n\n // Need to read fewer bytes in next iter\n bytes -= buf.byteLength - start;\n offset += buf.byteLength;\n }\n\n // Should only finish loop if exhausted all arrays\n return null;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes Number of bytes\n * @return DataView with data\n */\n getDataView(bytes: number): DataView | null {\n const bufferOffsets = this.findBufferOffsets(bytes);\n // return `null` if not enough data, except if end() already called, in\n // which case throw an error.\n if (!bufferOffsets && this.ended) {\n throw new Error('binary data exhausted');\n }\n\n if (!bufferOffsets) {\n // @ts-ignore\n return null;\n }\n\n // If only one arrayBuffer needed, return DataView directly\n if (bufferOffsets.length === 1) {\n const [bufferIndex, [start, end]] = bufferOffsets[0];\n const arrayBuffer = this.arrayBuffers[bufferIndex];\n const view = new DataView(arrayBuffer, start, end - start);\n\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n // Concatenate portions of multiple ArrayBuffers\n const view = new DataView(this._combineArrayBuffers(bufferOffsets));\n this.offset += bytes;\n this.disposeBuffers();\n return view;\n }\n\n /**\n * Dispose of old array buffers\n */\n disposeBuffers(): void {\n while (\n this.arrayBuffers.length > 0 &&\n this.offset - this.maxRewindBytes >= this.arrayBuffers[0].byteLength\n ) {\n this.offset -= this.arrayBuffers[0].byteLength;\n this.arrayBuffers.shift();\n }\n }\n\n /**\n * Copy multiple ArrayBuffers into one contiguous ArrayBuffer\n *\n * In contrast to concatenateArrayBuffers, this only copies the necessary\n * portions of the source arrays, rather than first copying the entire arrays\n * then taking a part of them.\n *\n * @param bufferOffsets List of internal array offsets\n * @return New contiguous ArrayBuffer\n */\n _combineArrayBuffers(bufferOffsets: any[]): ArrayBufferLike {\n let byteLength: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [start, end] = bufferOffset[1];\n byteLength += end - start;\n }\n\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let resultOffset: number = 0;\n for (const bufferOffset of bufferOffsets) {\n const [bufferIndex, [start, end]] = bufferOffset;\n const sourceArray = new Uint8Array(this.arrayBuffers[bufferIndex]);\n result.set(sourceArray.subarray(start, end), resultOffset);\n resultOffset += end - start;\n }\n\n return result.buffer;\n }\n /**\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n /**\n * @param bytes\n */\n rewind(bytes: number): void {\n // TODO - only works if offset is already set\n this.offset -= bytes;\n }\n}\n"],"mappings":";;;;;;;;;;;;;;IAIqBA,iBAAiB;EAMpC,2BAAYC,OAAkC,EAAE;IAAA;IAAA;IAAA;IAAA;IAAA;IAC9C,WAA6BA,OAAO,IAAI,CAAC,CAAC;MAAA,2BAAnCC,cAAc;MAAdA,cAAc,oCAAG,CAAC;;IAGzB,IAAI,CAACC,MAAM,GAAG,CAAC;IAEf,IAAI,CAACC,YAAY,GAAG,EAAE;IACtB,IAAI,CAACC,KAAK,GAAG,KAAK;;IAGlB,IAAI,CAACH,cAAc,GAAGA,cAAc;EACtC;EAAC;IAAA;IAAA;IAID,eAAMI,WAAwB,EAAQ;MACpC,IAAI,CAACF,YAAY,CAACG,IAAI,CAACD,WAAW,CAAC;IACrC;EAAC;IAAA;IAAA,OAED,eAAY;MACV,IAAI,CAACF,YAAY,GAAG,EAAE;MACtB,IAAI,CAACC,KAAK,GAAG,IAAI;IACnB;;EAAC;IAAA;IAAA;IAQD,2BAAkBG,KAAa,EAAW;MACxC,IAAIC,cAAc,GAAG,CAAC,IAAI,CAACN,MAAM;MAAC,2CACR,IAAI,CAACC,YAAY;QAAA;MAAA;QAA3C,oDAA6C;UAAA,IAAlCE,WAAW;UACpBG,cAAc,IAAIH,WAAW,CAACI,UAAU;UACxC,IAAID,cAAc,IAAID,KAAK,EAAE;YAC3B,OAAO,IAAI;UACb;QACF;MAAC;QAAA;MAAA;QAAA;MAAA;MACD,OAAO,KAAK;IACd;;EAAC;IAAA;IAAA;IAQD,2BAAkBA,KAAa,EAAgB;MAC7C,IAAIL,MAAM,GAAG,CAAC,IAAI,CAACA,MAAM;MACzB,IAAMQ,eAAoB,GAAG,EAAE;MAE/B,KAAK,IAAIC,CAAC,GAAG,CAAC,EAAEA,CAAC,GAAG,IAAI,CAACR,YAAY,CAACS,MAAM,EAAED,CAAC,EAAE,EAAE;QACjD,IAAME,GAAG,GAAG,IAAI,CAACV,YAAY,CAACQ,CAAC,CAAC;;QAGhC,IAAIT,MAAM,GAAGW,GAAG,CAACJ,UAAU,IAAI,CAAC,EAAE;UAChCP,MAAM,IAAIW,GAAG,CAACJ,UAAU;UAExB;QACF;;QAKA,IAAMK,KAAK,GAAGZ,MAAM,IAAI,CAAC,GAAGa,IAAI,CAACC,GAAG,CAACd,MAAM,CAAC,GAAG,CAAC;QAChD,IAAIe,GAAW;;QAGf,IAAIH,KAAK,GAAGP,KAAK,IAAIM,GAAG,CAACJ,UAAU,EAAE;UACnCQ,GAAG,GAAGH,KAAK,GAAGP,KAAK;UACnBG,eAAe,CAACJ,IAAI,CAAC,CAACK,CAAC,EAAE,CAACG,KAAK,EAAEG,GAAG,CAAC,CAAC,CAAC;UACvC,OAAOP,eAAe;QACxB;;QAGAO,GAAG,GAAGJ,GAAG,CAACJ,UAAU;QACpBC,eAAe,CAACJ,IAAI,CAAC,CAACK,CAAC,EAAE,CAACG,KAAK,EAAEG,GAAG,CAAC,CAAC,CAAC;;QAGvCV,KAAK,IAAIM,GAAG,CAACJ,UAAU,GAAGK,KAAK;QAC/BZ,MAAM,IAAIW,GAAG,CAACJ,UAAU;MAC1B;;MAGA,OAAO,IAAI;IACb;;EAAC;IAAA;IAAA;IAQD,qBAAYF,KAAa,EAAmB;MAC1C,IAAMW,aAAa,GAAG,IAAI,CAACC,iBAAiB,CAACZ,KAAK,CAAC;MAGnD,IAAI,CAACW,aAAa,IAAI,IAAI,CAACd,KAAK,EAAE;QAChC,MAAM,IAAIgB,KAAK,CAAC,uBAAuB,CAAC;MAC1C;MAEA,IAAI,CAACF,aAAa,EAAE;QAElB,OAAO,IAAI;MACb;;MAGA,IAAIA,aAAa,CAACN,MAAM,KAAK,CAAC,EAAE;QAC9B,mDAAoCM,aAAa,CAAC,CAAC,CAAC;UAA7CG,WAAW;UAAA;UAAGP,KAAK;UAAEG,GAAG;QAC/B,IAAMZ,WAAW,GAAG,IAAI,CAACF,YAAY,CAACkB,WAAW,CAAC;QAClD,IAAMC,KAAI,GAAG,IAAIC,QAAQ,CAAClB,WAAW,EAAES,KAAK,EAAEG,GAAG,GAAGH,KAAK,CAAC;QAE1D,IAAI,CAACZ,MAAM,IAAIK,KAAK;QACpB,IAAI,CAACiB,cAAc,EAAE;QACrB,OAAOF,KAAI;MACb;;MAGA,IAAMA,IAAI,GAAG,IAAIC,QAAQ,CAAC,IAAI,CAACE,oBAAoB,CAACP,aAAa,CAAC,CAAC;MACnE,IAAI,CAAChB,MAAM,IAAIK,KAAK;MACpB,IAAI,CAACiB,cAAc,EAAE;MACrB,OAAOF,IAAI;IACb;;EAAC;IAAA;IAAA;IAKD,0BAAuB;MACrB,OACE,IAAI,CAACnB,YAAY,CAACS,MAAM,GAAG,CAAC,IAC5B,IAAI,CAACV,MAAM,GAAG,IAAI,CAACD,cAAc,IAAI,IAAI,CAACE,YAAY,CAAC,CAAC,CAAC,CAACM,UAAU,EACpE;QACA,IAAI,CAACP,MAAM,IAAI,IAAI,CAACC,YAAY,CAAC,CAAC,CAAC,CAACM,UAAU;QAC9C,IAAI,CAACN,YAAY,CAACuB,KAAK,EAAE;MAC3B;IACF;;EAAC;IAAA;IAAA;IAYD,8BAAqBR,aAAoB,EAAmB;MAC1D,IAAIT,UAAkB,GAAG,CAAC;MAAC,4CACAS,aAAa;QAAA;MAAA;QAAxC,uDAA0C;UAAA,IAA/BS,YAAY;UACrB,kDAAqBA,YAAY,CAAC,CAAC,CAAC;YAA7Bb,KAAK;YAAEG,GAAG;UACjBR,UAAU,IAAIQ,GAAG,GAAGH,KAAK;QAC3B;MAAC;QAAA;MAAA;QAAA;MAAA;MAED,IAAMc,MAAM,GAAG,IAAIC,UAAU,CAACpB,UAAU,CAAC;;MAGzC,IAAIqB,YAAoB,GAAG,CAAC;MAAC,4CACFZ,aAAa;QAAA;MAAA;QAAxC,uDAA0C;UAAA,IAA/BS,aAAY;UACrB,kDAAoCA,aAAY;YAAzCN,WAAW;YAAA;YAAGP,MAAK;YAAEG,IAAG;UAC/B,IAAMc,WAAW,GAAG,IAAIF,UAAU,CAAC,IAAI,CAAC1B,YAAY,CAACkB,WAAW,CAAC,CAAC;UAClEO,MAAM,CAACI,GAAG,CAACD,WAAW,CAACE,QAAQ,CAACnB,MAAK,EAAEG,IAAG,CAAC,EAAEa,YAAY,CAAC;UAC1DA,YAAY,IAAIb,IAAG,GAAGH,MAAK;QAC7B;MAAC;QAAA;MAAA;QAAA;MAAA;MAED,OAAOc,MAAM,CAACM,MAAM;IACtB;EAAC;IAAA;IAAA;IAID,cAAK3B,KAAa,EAAQ;MACxB,IAAI,CAACL,MAAM,IAAIK,KAAK;IACtB;EAAC;IAAA;IAAA;IAID,gBAAOA,KAAa,EAAQ;MAE1B,IAAI,CAACL,MAAM,IAAIK,KAAK;IACtB;EAAC;EAAA;AAAA;AAAA"}
@@ -1,18 +1,13 @@
1
1
  "use strict";
2
2
 
3
3
  var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
4
-
5
4
  Object.defineProperty(exports, "__esModule", {
6
5
  value: true
7
6
  });
8
7
  exports.default = void 0;
9
-
10
8
  var _classCallCheck2 = _interopRequireDefault(require("@babel/runtime/helpers/classCallCheck"));
11
-
12
9
  var _createClass2 = _interopRequireDefault(require("@babel/runtime/helpers/createClass"));
13
-
14
10
  var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/defineProperty"));
15
-
16
11
  var BinaryReader = function () {
17
12
  function BinaryReader(arrayBuffer) {
18
13
  (0, _classCallCheck2.default)(this, BinaryReader);
@@ -21,36 +16,40 @@ var BinaryReader = function () {
21
16
  this.offset = 0;
22
17
  this.arrayBuffer = arrayBuffer;
23
18
  }
24
-
25
19
  (0, _createClass2.default)(BinaryReader, [{
26
20
  key: "hasAvailableBytes",
27
- value: function hasAvailableBytes(bytes) {
21
+ value:
22
+ function hasAvailableBytes(bytes) {
28
23
  return this.arrayBuffer.byteLength - this.offset >= bytes;
29
24
  }
25
+
30
26
  }, {
31
27
  key: "getDataView",
32
- value: function getDataView(bytes) {
28
+ value:
29
+ function getDataView(bytes) {
33
30
  if (bytes && !this.hasAvailableBytes(bytes)) {
34
31
  throw new Error('binary data exhausted');
35
32
  }
36
-
37
33
  var dataView = bytes ? new DataView(this.arrayBuffer, this.offset, bytes) : new DataView(this.arrayBuffer, this.offset);
38
34
  this.offset += bytes;
39
35
  return dataView;
40
36
  }
37
+
41
38
  }, {
42
39
  key: "skip",
43
- value: function skip(bytes) {
40
+ value:
41
+ function skip(bytes) {
44
42
  this.offset += bytes;
45
43
  }
44
+
46
45
  }, {
47
46
  key: "rewind",
48
- value: function rewind(bytes) {
47
+ value:
48
+ function rewind(bytes) {
49
49
  this.offset -= bytes;
50
50
  }
51
51
  }]);
52
52
  return BinaryReader;
53
53
  }();
54
-
55
54
  exports.default = BinaryReader;
56
55
  //# sourceMappingURL=binary-reader.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/streaming/binary-reader.ts"],"names":["BinaryReader","arrayBuffer","offset","bytes","byteLength","hasAvailableBytes","Error","dataView","DataView"],"mappings":";;;;;;;;;;;;;;;IAAqBA,Y;AAInB,wBAAYC,WAAZ,EAAsC;AAAA;AAAA;AAAA;AAEpC,SAAKC,MAAL,GAAc,CAAd;AAEA,SAAKD,WAAL,GAAmBA,WAAnB;AACD;;;;WAOD,2BAAkBE,KAAlB,EAA0C;AACxC,aAAO,KAAKF,WAAL,CAAiBG,UAAjB,GAA8B,KAAKF,MAAnC,IAA6CC,KAApD;AACD;;;WAQD,qBAAYA,KAAZ,EAAqC;AACnC,UAAIA,KAAK,IAAI,CAAC,KAAKE,iBAAL,CAAuBF,KAAvB,CAAd,EAA6C;AAC3C,cAAM,IAAIG,KAAJ,CAAU,uBAAV,CAAN;AACD;;AAED,UAAMC,QAAQ,GAAGJ,KAAK,GAClB,IAAIK,QAAJ,CAAa,KAAKP,WAAlB,EAA+B,KAAKC,MAApC,EAA4CC,KAA5C,CADkB,GAElB,IAAIK,QAAJ,CAAa,KAAKP,WAAlB,EAA+B,KAAKC,MAApC,CAFJ;AAGA,WAAKA,MAAL,IAAeC,KAAf;AACA,aAAOI,QAAP;AACD;;;WAOD,cAAKJ,KAAL,EAA0B;AACxB,WAAKD,MAAL,IAAeC,KAAf;AACD;;;WAOD,gBAAOA,KAAP,EAA4B;AAC1B,WAAKD,MAAL,IAAeC,KAAf;AACD","sourcesContent":["export default class BinaryReader {\n offset: number;\n arrayBuffer: ArrayBuffer;\n\n constructor(arrayBuffer: ArrayBuffer) {\n /** current global (stream) offset */\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffer = arrayBuffer;\n }\n /**\n * Checks if there are available bytes in data\n *\n * @param bytes\n * @returns boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n return this.arrayBuffer.byteLength - this.offset >= bytes;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes\n * @returns Dataview\n */\n getDataView(bytes: number): DataView {\n if (bytes && !this.hasAvailableBytes(bytes)) {\n throw new Error('binary data exhausted');\n }\n\n const dataView = bytes\n ? new DataView(this.arrayBuffer, this.offset, bytes)\n : new DataView(this.arrayBuffer, this.offset);\n this.offset += bytes;\n return dataView;\n }\n\n /**\n * Skipping\n *\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n\n /**\n * Rewinding\n *\n * @param bytes\n */\n rewind(bytes: number): void {\n this.offset -= bytes;\n }\n}\n"],"file":"binary-reader.js"}
1
+ {"version":3,"file":"binary-reader.js","names":["BinaryReader","arrayBuffer","offset","bytes","byteLength","hasAvailableBytes","Error","dataView","DataView"],"sources":["../../../../src/lib/streaming/binary-reader.ts"],"sourcesContent":["export default class BinaryReader {\n offset: number;\n arrayBuffer: ArrayBuffer;\n\n constructor(arrayBuffer: ArrayBuffer) {\n /** current global (stream) offset */\n this.offset = 0;\n /** current buffer from iterator */\n this.arrayBuffer = arrayBuffer;\n }\n /**\n * Checks if there are available bytes in data\n *\n * @param bytes\n * @returns boolean\n */\n hasAvailableBytes(bytes: number): boolean {\n return this.arrayBuffer.byteLength - this.offset >= bytes;\n }\n\n /**\n * Get the required number of bytes from the iterator\n *\n * @param bytes\n * @returns Dataview\n */\n getDataView(bytes: number): DataView {\n if (bytes && !this.hasAvailableBytes(bytes)) {\n throw new Error('binary data exhausted');\n }\n\n const dataView = bytes\n ? new DataView(this.arrayBuffer, this.offset, bytes)\n : new DataView(this.arrayBuffer, this.offset);\n this.offset += bytes;\n return dataView;\n }\n\n /**\n * Skipping\n *\n * @param bytes\n */\n skip(bytes: number): void {\n this.offset += bytes;\n }\n\n /**\n * Rewinding\n *\n * @param bytes\n */\n rewind(bytes: number): void {\n this.offset -= bytes;\n }\n}\n"],"mappings":";;;;;;;;;;IAAqBA,YAAY;EAI/B,sBAAYC,WAAwB,EAAE;IAAA;IAAA;IAAA;IAEpC,IAAI,CAACC,MAAM,GAAG,CAAC;IAEf,IAAI,CAACD,WAAW,GAAGA,WAAW;EAChC;EAAC;IAAA;IAAA;IAOD,2BAAkBE,KAAa,EAAW;MACxC,OAAO,IAAI,CAACF,WAAW,CAACG,UAAU,GAAG,IAAI,CAACF,MAAM,IAAIC,KAAK;IAC3D;;EAAC;IAAA;IAAA;IAQD,qBAAYA,KAAa,EAAY;MACnC,IAAIA,KAAK,IAAI,CAAC,IAAI,CAACE,iBAAiB,CAACF,KAAK,CAAC,EAAE;QAC3C,MAAM,IAAIG,KAAK,CAAC,uBAAuB,CAAC;MAC1C;MAEA,IAAMC,QAAQ,GAAGJ,KAAK,GAClB,IAAIK,QAAQ,CAAC,IAAI,CAACP,WAAW,EAAE,IAAI,CAACC,MAAM,EAAEC,KAAK,CAAC,GAClD,IAAIK,QAAQ,CAAC,IAAI,CAACP,WAAW,EAAE,IAAI,CAACC,MAAM,CAAC;MAC/C,IAAI,CAACA,MAAM,IAAIC,KAAK;MACpB,OAAOI,QAAQ;IACjB;;EAAC;IAAA;IAAA;IAOD,cAAKJ,KAAa,EAAQ;MACxB,IAAI,CAACD,MAAM,IAAIC,KAAK;IACtB;;EAAC;IAAA;IAAA;IAOD,gBAAOA,KAAa,EAAQ;MAC1B,IAAI,CAACD,MAAM,IAAIC,KAAK;IACtB;EAAC;EAAA;AAAA;AAAA"}
@@ -1,26 +1,19 @@
1
1
  "use strict";
2
2
 
3
3
  var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefault");
4
-
5
4
  Object.defineProperty(exports, "__esModule", {
6
5
  value: true
7
6
  });
8
7
  exports.zipBatchIterators = zipBatchIterators;
9
-
10
8
  var _regenerator = _interopRequireDefault(require("@babel/runtime/regenerator"));
11
-
12
9
  var _awaitAsyncGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/awaitAsyncGenerator"));
13
-
14
10
  var _wrapAsyncGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/wrapAsyncGenerator"));
15
-
16
11
  function zipBatchIterators(_x, _x2) {
17
12
  return _zipBatchIterators.apply(this, arguments);
18
13
  }
19
-
20
14
  function _zipBatchIterators() {
21
15
  _zipBatchIterators = (0, _wrapAsyncGenerator2.default)(_regenerator.default.mark(function _callee(iterator1, iterator2) {
22
16
  var batch1, batch2, iterator1Done, iterator2Done, _yield$_awaitAsyncGen, value, done, _yield$_awaitAsyncGen2, _value, _done, batch;
23
-
24
17
  return _regenerator.default.wrap(function _callee$(_context) {
25
18
  while (1) {
26
19
  switch (_context.prev = _context.next) {
@@ -29,70 +22,55 @@ function _zipBatchIterators() {
29
22
  batch2 = [];
30
23
  iterator1Done = false;
31
24
  iterator2Done = false;
32
-
33
25
  case 4:
34
26
  if (!(!iterator1Done && !iterator2Done)) {
35
27
  _context.next = 27;
36
28
  break;
37
29
  }
38
-
39
30
  if (!(batch1.length === 0 && !iterator1Done)) {
40
31
  _context.next = 14;
41
32
  break;
42
33
  }
43
-
44
34
  _context.next = 8;
45
35
  return (0, _awaitAsyncGenerator2.default)(iterator1.next());
46
-
47
36
  case 8:
48
37
  _yield$_awaitAsyncGen = _context.sent;
49
38
  value = _yield$_awaitAsyncGen.value;
50
39
  done = _yield$_awaitAsyncGen.done;
51
-
52
40
  if (done) {
53
41
  iterator1Done = true;
54
42
  } else {
55
43
  batch1 = value;
56
44
  }
57
-
58
45
  _context.next = 21;
59
46
  break;
60
-
61
47
  case 14:
62
48
  if (!(batch2.length === 0 && !iterator2Done)) {
63
49
  _context.next = 21;
64
50
  break;
65
51
  }
66
-
67
52
  _context.next = 17;
68
53
  return (0, _awaitAsyncGenerator2.default)(iterator2.next());
69
-
70
54
  case 17:
71
55
  _yield$_awaitAsyncGen2 = _context.sent;
72
56
  _value = _yield$_awaitAsyncGen2.value;
73
57
  _done = _yield$_awaitAsyncGen2.done;
74
-
75
58
  if (_done) {
76
59
  iterator2Done = true;
77
60
  } else {
78
61
  batch2 = _value;
79
62
  }
80
-
81
63
  case 21:
82
64
  batch = extractBatch(batch1, batch2);
83
-
84
65
  if (!batch) {
85
66
  _context.next = 25;
86
67
  break;
87
68
  }
88
-
89
69
  _context.next = 25;
90
70
  return batch;
91
-
92
71
  case 25:
93
72
  _context.next = 4;
94
73
  break;
95
-
96
74
  case 27:
97
75
  case "end":
98
76
  return _context.stop();
@@ -102,15 +80,14 @@ function _zipBatchIterators() {
102
80
  }));
103
81
  return _zipBatchIterators.apply(this, arguments);
104
82
  }
105
-
106
83
  function extractBatch(batch1, batch2) {
107
84
  var batchLength = Math.min(batch1.length, batch2.length);
108
-
109
85
  if (batchLength === 0) {
110
86
  return null;
111
87
  }
112
88
 
113
89
  var batch = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];
90
+
114
91
  batch1.splice(0, batchLength);
115
92
  batch2.splice(0, batchLength);
116
93
  return batch;
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/lib/streaming/zip-batch-iterators.ts"],"names":["zipBatchIterators","iterator1","iterator2","batch1","batch2","iterator1Done","iterator2Done","length","next","value","done","batch","extractBatch","batchLength","Math","min","slice","splice"],"mappings":";;;;;;;;;;;;;;;SAMuBA,iB;;;;;mFAAhB,iBACLC,SADK,EAELC,SAFK;AAAA;;AAAA;AAAA;AAAA;AAAA;AAIDC,YAAAA,MAJC,GAIkB,EAJlB;AAKDC,YAAAA,MALC,GAKkB,EALlB;AAMDC,YAAAA,aANC,GAMwB,KANxB;AAODC,YAAAA,aAPC,GAOwB,KAPxB;;AAAA;AAAA,kBAWE,CAACD,aAAD,IAAkB,CAACC,aAXrB;AAAA;AAAA;AAAA;;AAAA,kBAYCH,MAAM,CAACI,MAAP,KAAkB,CAAlB,IAAuB,CAACF,aAZzB;AAAA;AAAA;AAAA;;AAAA;AAAA,sDAa2BJ,SAAS,CAACO,IAAV,EAb3B;;AAAA;AAAA;AAaMC,YAAAA,KAbN,yBAaMA,KAbN;AAaaC,YAAAA,IAbb,yBAaaA,IAbb;;AAcD,gBAAIA,IAAJ,EAAU;AACRL,cAAAA,aAAa,GAAG,IAAhB;AACD,aAFD,MAEO;AACLF,cAAAA,MAAM,GAAGM,KAAT;AACD;;AAlBA;AAAA;;AAAA;AAAA,kBAmBQL,MAAM,CAACG,MAAP,KAAkB,CAAlB,IAAuB,CAACD,aAnBhC;AAAA;AAAA;AAAA;;AAAA;AAAA,sDAoB2BJ,SAAS,CAACM,IAAV,EApB3B;;AAAA;AAAA;AAoBMC,YAAAA,MApBN,0BAoBMA,KApBN;AAoBaC,YAAAA,KApBb,0BAoBaA,IApBb;;AAqBD,gBAAIA,KAAJ,EAAU;AACRJ,cAAAA,aAAa,GAAG,IAAhB;AACD,aAFD,MAEO;AACLF,cAAAA,MAAM,GAAGK,MAAT;AACD;;AAzBA;AA4BGE,YAAAA,KA5BH,GA4BWC,YAAY,CAACT,MAAD,EAASC,MAAT,CA5BvB;;AAAA,iBA6BCO,KA7BD;AAAA;AAAA;AAAA;;AAAA;AA8BD,mBAAMA,KAAN;;AA9BC;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,G;;;;AA0CP,SAASC,YAAT,CAAsBT,MAAtB,EAAwCC,MAAxC,EAA6E;AAC3E,MAAMS,WAAmB,GAAGC,IAAI,CAACC,GAAL,CAASZ,MAAM,CAACI,MAAhB,EAAwBH,MAAM,CAACG,MAA/B,CAA5B;;AACA,MAAIM,WAAW,KAAK,CAApB,EAAuB;AACrB,WAAO,IAAP;AACD;;AAGD,MAAMF,KAAiB,GAAG,CAACR,MAAM,CAACa,KAAP,CAAa,CAAb,EAAgBH,WAAhB,CAAD,EAA+BT,MAAM,CAACY,KAAP,CAAa,CAAb,EAAgBH,WAAhB,CAA/B,CAA1B;AAGAV,EAAAA,MAAM,CAACc,MAAP,CAAc,CAAd,EAAiBJ,WAAjB;AACAT,EAAAA,MAAM,CAACa,MAAP,CAAc,CAAd,EAAiBJ,WAAjB;AACA,SAAOF,KAAP;AACD","sourcesContent":["/**\n * Zip two iterators together\n *\n * @param iterator1\n * @param iterator2\n */\nexport async function* zipBatchIterators(\n iterator1: AsyncIterator<any[]>,\n iterator2: AsyncIterator<any[]>\n): AsyncGenerator<number[][], void, unknown> {\n let batch1: number[] = [];\n let batch2: number[] = [];\n let iterator1Done: boolean = false;\n let iterator2Done: boolean = false;\n\n // TODO - one could let all iterators flow at full speed using `Promise.race`\n // however we might end up with a big temporary buffer\n while (!iterator1Done && !iterator2Done) {\n if (batch1.length === 0 && !iterator1Done) {\n const {value, done} = await iterator1.next();\n if (done) {\n iterator1Done = true;\n } else {\n batch1 = value;\n }\n } else if (batch2.length === 0 && !iterator2Done) {\n const {value, done} = await iterator2.next();\n if (done) {\n iterator2Done = true;\n } else {\n batch2 = value;\n }\n }\n\n const batch = extractBatch(batch1, batch2);\n if (batch) {\n yield batch;\n }\n }\n}\n\n/**\n * Extract batch of same length from two batches\n *\n * @param batch1\n * @param batch2\n * @return array | null\n */\nfunction extractBatch(batch1: number[], batch2: number[]): number[][] | null {\n const batchLength: number = Math.min(batch1.length, batch2.length);\n if (batchLength === 0) {\n return null;\n }\n\n // Non interleaved arrays\n const batch: number[][] = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];\n\n // Modify the 2 batches\n batch1.splice(0, batchLength);\n batch2.splice(0, batchLength);\n return batch;\n}\n"],"file":"zip-batch-iterators.js"}
1
+ {"version":3,"file":"zip-batch-iterators.js","names":["zipBatchIterators","iterator1","iterator2","batch1","batch2","iterator1Done","iterator2Done","length","next","value","done","batch","extractBatch","batchLength","Math","min","slice","splice"],"sources":["../../../../src/lib/streaming/zip-batch-iterators.ts"],"sourcesContent":["/**\n * Zip two iterators together\n *\n * @param iterator1\n * @param iterator2\n */\nexport async function* zipBatchIterators(\n iterator1: AsyncIterator<any[]>,\n iterator2: AsyncIterator<any[]>\n): AsyncGenerator<number[][], void, unknown> {\n let batch1: number[] = [];\n let batch2: number[] = [];\n let iterator1Done: boolean = false;\n let iterator2Done: boolean = false;\n\n // TODO - one could let all iterators flow at full speed using `Promise.race`\n // however we might end up with a big temporary buffer\n while (!iterator1Done && !iterator2Done) {\n if (batch1.length === 0 && !iterator1Done) {\n const {value, done} = await iterator1.next();\n if (done) {\n iterator1Done = true;\n } else {\n batch1 = value;\n }\n } else if (batch2.length === 0 && !iterator2Done) {\n const {value, done} = await iterator2.next();\n if (done) {\n iterator2Done = true;\n } else {\n batch2 = value;\n }\n }\n\n const batch = extractBatch(batch1, batch2);\n if (batch) {\n yield batch;\n }\n }\n}\n\n/**\n * Extract batch of same length from two batches\n *\n * @param batch1\n * @param batch2\n * @return array | null\n */\nfunction extractBatch(batch1: number[], batch2: number[]): number[][] | null {\n const batchLength: number = Math.min(batch1.length, batch2.length);\n if (batchLength === 0) {\n return null;\n }\n\n // Non interleaved arrays\n const batch: number[][] = [batch1.slice(0, batchLength), batch2.slice(0, batchLength)];\n\n // Modify the 2 batches\n batch1.splice(0, batchLength);\n batch2.splice(0, batchLength);\n return batch;\n}\n"],"mappings":";;;;;;;;;;SAMuBA,iBAAiB;EAAA;AAAA;AAAA;EAAA,iFAAjC,iBACLC,SAA+B,EAC/BC,SAA+B;IAAA;IAAA;MAAA;QAAA;UAAA;YAE3BC,MAAgB,GAAG,EAAE;YACrBC,MAAgB,GAAG,EAAE;YACrBC,aAAsB,GAAG,KAAK;YAC9BC,aAAsB,GAAG,KAAK;UAAA;YAAA,MAI3B,CAACD,aAAa,IAAI,CAACC,aAAa;cAAA;cAAA;YAAA;YAAA,MACjCH,MAAM,CAACI,MAAM,KAAK,CAAC,IAAI,CAACF,aAAa;cAAA;cAAA;YAAA;YAAA;YAAA,0CACXJ,SAAS,CAACO,IAAI,EAAE;UAAA;YAAA;YAArCC,KAAK,yBAALA,KAAK;YAAEC,IAAI,yBAAJA,IAAI;YAClB,IAAIA,IAAI,EAAE;cACRL,aAAa,GAAG,IAAI;YACtB,CAAC,MAAM;cACLF,MAAM,GAAGM,KAAK;YAChB;YAAC;YAAA;UAAA;YAAA,MACQL,MAAM,CAACG,MAAM,KAAK,CAAC,IAAI,CAACD,aAAa;cAAA;cAAA;YAAA;YAAA;YAAA,0CAClBJ,SAAS,CAACM,IAAI,EAAE;UAAA;YAAA;YAArCC,MAAK,0BAALA,KAAK;YAAEC,KAAI,0BAAJA,IAAI;YAClB,IAAIA,KAAI,EAAE;cACRJ,aAAa,GAAG,IAAI;YACtB,CAAC,MAAM;cACLF,MAAM,GAAGK,MAAK;YAChB;UAAC;YAGGE,KAAK,GAAGC,YAAY,CAACT,MAAM,EAAEC,MAAM,CAAC;YAAA,KACtCO,KAAK;cAAA;cAAA;YAAA;YAAA;YACP,OAAMA,KAAK;UAAA;YAAA;YAAA;UAAA;UAAA;YAAA;QAAA;MAAA;IAAA;EAAA,CAGhB;EAAA;AAAA;AASD,SAASC,YAAY,CAACT,MAAgB,EAAEC,MAAgB,EAAqB;EAC3E,IAAMS,WAAmB,GAAGC,IAAI,CAACC,GAAG,CAACZ,MAAM,CAACI,MAAM,EAAEH,MAAM,CAACG,MAAM,CAAC;EAClE,IAAIM,WAAW,KAAK,CAAC,EAAE;IACrB,OAAO,IAAI;EACb;;EAGA,IAAMF,KAAiB,GAAG,CAACR,MAAM,CAACa,KAAK,CAAC,CAAC,EAAEH,WAAW,CAAC,EAAET,MAAM,CAACY,KAAK,CAAC,CAAC,EAAEH,WAAW,CAAC,CAAC;;EAGtFV,MAAM,CAACc,MAAM,CAAC,CAAC,EAAEJ,WAAW,CAAC;EAC7BT,MAAM,CAACa,MAAM,CAAC,CAAC,EAAEJ,WAAW,CAAC;EAC7B,OAAOF,KAAK;AACd"}
@@ -4,12 +4,10 @@ Object.defineProperty(exports, "__esModule", {
4
4
  value: true
5
5
  });
6
6
  exports._typecheckShapefileLoader = exports.ShapefileLoader = void 0;
7
-
8
7
  var _shpLoader = require("./shp-loader");
9
-
10
8
  var _parseShapefile = require("./lib/parsers/parse-shapefile");
9
+ var VERSION = typeof "3.3.0-alpha.6" !== 'undefined' ? "3.3.0-alpha.6" : 'latest';
11
10
 
12
- var VERSION = typeof "3.3.0-alpha.5" !== 'undefined' ? "3.3.0-alpha.5" : 'latest';
13
11
  var ShapefileLoader = {
14
12
  name: 'Shapefile',
15
13
  id: 'shapefile',
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/shapefile-loader.ts"],"names":["VERSION","ShapefileLoader","name","id","module","version","category","extensions","mimeTypes","tests","Uint8Array","SHP_MAGIC_NUMBER","buffer","options","shapefile","shp","_maxDimensions","parse","parseShapefile","parseInBatches","parseShapefileInBatches","_typecheckShapefileLoader"],"mappings":";;;;;;;AACA;;AACA;;AAIA,IAAMA,OAAO,GAAG,2BAAuB,WAAvB,qBAAmD,QAAnE;AAMO,IAAMC,eAAe,GAAG;AAC7BC,EAAAA,IAAI,EAAE,WADuB;AAE7BC,EAAAA,EAAE,EAAE,WAFyB;AAG7BC,EAAAA,MAAM,EAAE,WAHqB;AAI7BC,EAAAA,OAAO,EAAEL,OAJoB;AAK7BM,EAAAA,QAAQ,EAAE,UALmB;AAM7BC,EAAAA,UAAU,EAAE,CAAC,KAAD,CANiB;AAO7BC,EAAAA,SAAS,EAAE,CAAC,0BAAD,CAPkB;AAQ7BC,EAAAA,KAAK,EAAE,CAAC,IAAIC,UAAJ,CAAeC,2BAAf,EAAiCC,MAAlC,CARsB;AAS7BC,EAAAA,OAAO,EAAE;AACPC,IAAAA,SAAS,EAAE,EADJ;AAEPC,IAAAA,GAAG,EAAE;AACHC,MAAAA,cAAc,EAAE;AADb;AAFE,GAToB;AAe7BC,EAAAA,KAAK,EAAEC,8BAfsB;AAgB7BC,EAAAA,cAAc,EAAEC;AAhBa,CAAxB;;AAmBA,IAAMC,yBAA2C,GAAGpB,eAApD","sourcesContent":["import type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {SHP_MAGIC_NUMBER} from './shp-loader';\nimport {parseShapefile, parseShapefileInBatches} from './lib/parsers/parse-shapefile';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * Shapefile loader\n * @note Shapefile is multifile format and requires providing additional files\n */\nexport const ShapefileLoader = {\n name: 'Shapefile',\n id: 'shapefile',\n module: 'shapefile',\n version: VERSION,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shapefile: {},\n shp: {\n _maxDimensions: 4\n }\n },\n parse: parseShapefile,\n parseInBatches: parseShapefileInBatches\n};\n\nexport const _typecheckShapefileLoader: LoaderWithParser = ShapefileLoader;\n"],"file":"shapefile-loader.js"}
1
+ {"version":3,"file":"shapefile-loader.js","names":["VERSION","ShapefileLoader","name","id","module","version","category","extensions","mimeTypes","tests","Uint8Array","SHP_MAGIC_NUMBER","buffer","options","shapefile","shp","_maxDimensions","parse","parseShapefile","parseInBatches","parseShapefileInBatches","_typecheckShapefileLoader"],"sources":["../../src/shapefile-loader.ts"],"sourcesContent":["import type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport {SHP_MAGIC_NUMBER} from './shp-loader';\nimport {parseShapefile, parseShapefileInBatches} from './lib/parsers/parse-shapefile';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/**\n * Shapefile loader\n * @note Shapefile is multifile format and requires providing additional files\n */\nexport const ShapefileLoader = {\n name: 'Shapefile',\n id: 'shapefile',\n module: 'shapefile',\n version: VERSION,\n category: 'geometry',\n extensions: ['shp'],\n mimeTypes: ['application/octet-stream'],\n tests: [new Uint8Array(SHP_MAGIC_NUMBER).buffer],\n options: {\n shapefile: {},\n shp: {\n _maxDimensions: 4\n }\n },\n parse: parseShapefile,\n parseInBatches: parseShapefileInBatches\n};\n\nexport const _typecheckShapefileLoader: LoaderWithParser = ShapefileLoader;\n"],"mappings":";;;;;;AACA;AACA;AAIA,IAAMA,OAAO,GAAG,sBAAkB,KAAK,WAAW,qBAAiB,QAAQ;;AAMpE,IAAMC,eAAe,GAAG;EAC7BC,IAAI,EAAE,WAAW;EACjBC,EAAE,EAAE,WAAW;EACfC,MAAM,EAAE,WAAW;EACnBC,OAAO,EAAEL,OAAO;EAChBM,QAAQ,EAAE,UAAU;EACpBC,UAAU,EAAE,CAAC,KAAK,CAAC;EACnBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EACvCC,KAAK,EAAE,CAAC,IAAIC,UAAU,CAACC,2BAAgB,CAAC,CAACC,MAAM,CAAC;EAChDC,OAAO,EAAE;IACPC,SAAS,EAAE,CAAC,CAAC;IACbC,GAAG,EAAE;MACHC,cAAc,EAAE;IAClB;EACF,CAAC;EACDC,KAAK,EAAEC,8BAAc;EACrBC,cAAc,EAAEC;AAClB,CAAC;AAAC;AAEK,IAAMC,yBAA2C,GAAGpB,eAAe;AAAC"}