@loaders.gl/json 3.1.3 → 4.0.0-alpha.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (93) hide show
  1. package/dist/bundle.js +2 -2
  2. package/dist/bundle.js.map +1 -0
  3. package/dist/geojson-loader.js +69 -66
  4. package/dist/geojson-loader.js.map +1 -0
  5. package/dist/geojson-worker.js +1 -1
  6. package/dist/index.js +7 -19
  7. package/dist/index.js.map +1 -0
  8. package/dist/json-loader.js +38 -48
  9. package/dist/json-loader.js.map +1 -0
  10. package/dist/jsonl-loader.js +1 -53
  11. package/dist/{es5/jsonl-loader.js.map → jsonl-loader.js.map} +0 -0
  12. package/dist/{es5/lib → lib}/clarinet/LICENSE +0 -0
  13. package/dist/lib/clarinet/clarinet.js +493 -510
  14. package/dist/lib/clarinet/clarinet.js.map +1 -0
  15. package/dist/lib/jsonpath/jsonpath.js +78 -81
  16. package/dist/lib/jsonpath/jsonpath.js.map +1 -0
  17. package/dist/lib/parse-json-in-batches.js +98 -94
  18. package/dist/lib/parse-json-in-batches.js.map +1 -0
  19. package/dist/lib/parse-json.js +29 -24
  20. package/dist/lib/parse-json.js.map +1 -0
  21. package/dist/lib/parse-ndjson-in-batches.js +34 -32
  22. package/dist/lib/parse-ndjson-in-batches.js.map +1 -0
  23. package/dist/lib/parse-ndjson.js +10 -13
  24. package/dist/lib/parse-ndjson.js.map +1 -0
  25. package/dist/lib/parser/json-parser.js +125 -95
  26. package/dist/lib/parser/json-parser.js.map +1 -0
  27. package/dist/lib/parser/streaming-json-parser.js +88 -95
  28. package/dist/lib/parser/streaming-json-parser.js.map +1 -0
  29. package/dist/ndjson-loader.js +25 -28
  30. package/dist/ndjson-loader.js.map +1 -0
  31. package/dist/workers/geojson-worker.js +4 -5
  32. package/dist/workers/geojson-worker.js.map +1 -0
  33. package/package.json +8 -8
  34. package/dist/es5/bundle.js +0 -7
  35. package/dist/es5/bundle.js.map +0 -1
  36. package/dist/es5/geojson-loader.js +0 -209
  37. package/dist/es5/geojson-loader.js.map +0 -1
  38. package/dist/es5/index.js +0 -62
  39. package/dist/es5/index.js.map +0 -1
  40. package/dist/es5/json-loader.js +0 -85
  41. package/dist/es5/json-loader.js.map +0 -1
  42. package/dist/es5/jsonl-loader.js +0 -2
  43. package/dist/es5/lib/clarinet/clarinet.js +0 -535
  44. package/dist/es5/lib/clarinet/clarinet.js.map +0 -1
  45. package/dist/es5/lib/jsonpath/jsonpath.js +0 -144
  46. package/dist/es5/lib/jsonpath/jsonpath.js.map +0 -1
  47. package/dist/es5/lib/parse-json-in-batches.js +0 -278
  48. package/dist/es5/lib/parse-json-in-batches.js.map +0 -1
  49. package/dist/es5/lib/parse-json.js +0 -46
  50. package/dist/es5/lib/parse-json.js.map +0 -1
  51. package/dist/es5/lib/parse-ndjson-in-batches.js +0 -158
  52. package/dist/es5/lib/parse-ndjson-in-batches.js.map +0 -1
  53. package/dist/es5/lib/parse-ndjson.js +0 -18
  54. package/dist/es5/lib/parse-ndjson.js.map +0 -1
  55. package/dist/es5/lib/parser/json-parser.js +0 -159
  56. package/dist/es5/lib/parser/json-parser.js.map +0 -1
  57. package/dist/es5/lib/parser/streaming-json-parser.js +0 -158
  58. package/dist/es5/lib/parser/streaming-json-parser.js.map +0 -1
  59. package/dist/es5/ndjson-loader.js +0 -67
  60. package/dist/es5/ndjson-loader.js.map +0 -1
  61. package/dist/es5/workers/geojson-worker.js +0 -8
  62. package/dist/es5/workers/geojson-worker.js.map +0 -1
  63. package/dist/esm/bundle.js +0 -5
  64. package/dist/esm/bundle.js.map +0 -1
  65. package/dist/esm/geojson-loader.js +0 -81
  66. package/dist/esm/geojson-loader.js.map +0 -1
  67. package/dist/esm/index.js +0 -7
  68. package/dist/esm/index.js.map +0 -1
  69. package/dist/esm/json-loader.js +0 -47
  70. package/dist/esm/json-loader.js.map +0 -1
  71. package/dist/esm/jsonl-loader.js +0 -2
  72. package/dist/esm/jsonl-loader.js.map +0 -1
  73. package/dist/esm/lib/clarinet/LICENSE +0 -28
  74. package/dist/esm/lib/clarinet/clarinet.js +0 -518
  75. package/dist/esm/lib/clarinet/clarinet.js.map +0 -1
  76. package/dist/esm/lib/jsonpath/jsonpath.js +0 -86
  77. package/dist/esm/lib/jsonpath/jsonpath.js.map +0 -1
  78. package/dist/esm/lib/parse-json-in-batches.js +0 -104
  79. package/dist/esm/lib/parse-json-in-batches.js.map +0 -1
  80. package/dist/esm/lib/parse-json.js +0 -34
  81. package/dist/esm/lib/parse-json.js.map +0 -1
  82. package/dist/esm/lib/parse-ndjson-in-batches.js +0 -37
  83. package/dist/esm/lib/parse-ndjson-in-batches.js.map +0 -1
  84. package/dist/esm/lib/parse-ndjson.js +0 -11
  85. package/dist/esm/lib/parse-ndjson.js.map +0 -1
  86. package/dist/esm/lib/parser/json-parser.js +0 -128
  87. package/dist/esm/lib/parser/json-parser.js.map +0 -1
  88. package/dist/esm/lib/parser/streaming-json-parser.js +0 -93
  89. package/dist/esm/lib/parser/streaming-json-parser.js.map +0 -1
  90. package/dist/esm/ndjson-loader.js +0 -32
  91. package/dist/esm/ndjson-loader.js.map +0 -1
  92. package/dist/esm/workers/geojson-worker.js +0 -4
  93. package/dist/esm/workers/geojson-worker.js.map +0 -1
@@ -1,86 +0,0 @@
1
- import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
- export default class JSONPath {
3
- constructor(path = null) {
4
- _defineProperty(this, "path", void 0);
5
-
6
- this.path = ['$'];
7
-
8
- if (path instanceof JSONPath) {
9
- this.path = [...path.path];
10
- return;
11
- }
12
-
13
- if (Array.isArray(path)) {
14
- this.path.push(...path);
15
- return;
16
- }
17
-
18
- if (typeof path === 'string') {
19
- this.path = path.split('.');
20
-
21
- if (this.path[0] !== '$') {
22
- throw new Error('JSONPaths must start with $');
23
- }
24
- }
25
- }
26
-
27
- clone() {
28
- return new JSONPath(this);
29
- }
30
-
31
- toString() {
32
- return this.path.join('.');
33
- }
34
-
35
- push(name) {
36
- this.path.push(name);
37
- }
38
-
39
- pop() {
40
- return this.path.pop();
41
- }
42
-
43
- set(name) {
44
- this.path[this.path.length - 1] = name;
45
- }
46
-
47
- equals(other) {
48
- if (!this || !other || this.path.length !== other.path.length) {
49
- return false;
50
- }
51
-
52
- for (let i = 0; i < this.path.length; ++i) {
53
- if (this.path[i] !== other.path[i]) {
54
- return false;
55
- }
56
- }
57
-
58
- return true;
59
- }
60
-
61
- setFieldAtPath(object, value) {
62
- const path = [...this.path];
63
- path.shift();
64
- const field = path.pop();
65
-
66
- for (const component of path) {
67
- object = object[component];
68
- }
69
-
70
- object[field] = value;
71
- }
72
-
73
- getFieldAtPath(object) {
74
- const path = [...this.path];
75
- path.shift();
76
- const field = path.pop();
77
-
78
- for (const component of path) {
79
- object = object[component];
80
- }
81
-
82
- return object[field];
83
- }
84
-
85
- }
86
- //# sourceMappingURL=jsonpath.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../../src/lib/jsonpath/jsonpath.ts"],"names":["JSONPath","constructor","path","Array","isArray","push","split","Error","clone","toString","join","name","pop","set","length","equals","other","i","setFieldAtPath","object","value","shift","field","component","getFieldAtPath"],"mappings":";AAQA,eAAe,MAAMA,QAAN,CAAe;AAG5BC,EAAAA,WAAW,CAACC,IAAyC,GAAG,IAA7C,EAAmD;AAAA;;AAC5D,SAAKA,IAAL,GAAY,CAAC,GAAD,CAAZ;;AAEA,QAAIA,IAAI,YAAYF,QAApB,EAA8B;AAE5B,WAAKE,IAAL,GAAY,CAAC,GAAGA,IAAI,CAACA,IAAT,CAAZ;AACA;AACD;;AAED,QAAIC,KAAK,CAACC,OAAN,CAAcF,IAAd,CAAJ,EAAyB;AACvB,WAAKA,IAAL,CAAUG,IAAV,CAAe,GAAGH,IAAlB;AACA;AACD;;AAGD,QAAI,OAAOA,IAAP,KAAgB,QAApB,EAA8B;AAC5B,WAAKA,IAAL,GAAYA,IAAI,CAACI,KAAL,CAAW,GAAX,CAAZ;;AACA,UAAI,KAAKJ,IAAL,CAAU,CAAV,MAAiB,GAArB,EAA0B;AACxB,cAAM,IAAIK,KAAJ,CAAU,6BAAV,CAAN;AACD;AACF;AACF;;AAEDC,EAAAA,KAAK,GAAa;AAChB,WAAO,IAAIR,QAAJ,CAAa,IAAb,CAAP;AACD;;AAEDS,EAAAA,QAAQ,GAAW;AACjB,WAAO,KAAKP,IAAL,CAAUQ,IAAV,CAAe,GAAf,CAAP;AACD;;AAEDL,EAAAA,IAAI,CAACM,IAAD,EAAqB;AACvB,SAAKT,IAAL,CAAUG,IAAV,CAAeM,IAAf;AACD;;AAEDC,EAAAA,GAAG,GAAG;AACJ,WAAO,KAAKV,IAAL,CAAUU,GAAV,EAAP;AACD;;AAEDC,EAAAA,GAAG,CAACF,IAAD,EAAqB;AACtB,SAAKT,IAAL,CAAU,KAAKA,IAAL,CAAUY,MAAV,GAAmB,CAA7B,IAAkCH,IAAlC;AACD;;AAEDI,EAAAA,MAAM,CAACC,KAAD,EAA2B;AAC/B,QAAI,CAAC,IAAD,IAAS,CAACA,KAAV,IAAmB,KAAKd,IAAL,CAAUY,MAAV,KAAqBE,KAAK,CAACd,IAAN,CAAWY,MAAvD,EAA+D;AAC7D,aAAO,KAAP;AACD;;AAED,SAAK,IAAIG,CAAC,GAAG,CAAb,EAAgBA,CAAC,GAAG,KAAKf,IAAL,CAAUY,MAA9B,EAAsC,EAAEG,CAAxC,EAA2C;AACzC,UAAI,KAAKf,IAAL,CAAUe,CAAV,MAAiBD,KAAK,CAACd,IAAN,CAAWe,CAAX,CAArB,EAAoC;AAClC,eAAO,KAAP;AACD;AACF;;AAED,WAAO,IAAP;AACD;;AAQDC,EAAAA,cAAc,CAACC,MAAD,EAASC,KAAT,EAAgB;AAC5B,UAAMlB,IAAI,GAAG,CAAC,GAAG,KAAKA,IAAT,CAAb;AACAA,IAAAA,IAAI,CAACmB,KAAL;AACA,UAAMC,KAAK,GAAGpB,IAAI,CAACU,GAAL,EAAd;;AACA,SAAK,MAAMW,SAAX,IAAwBrB,IAAxB,EAA8B;AAC5BiB,MAAAA,MAAM,GAAGA,MAAM,CAACI,SAAD,CAAf;AACD;;AAEDJ,IAAAA,MAAM,CAACG,KAAD,CAAN,GAAgBF,KAAhB;AACD;;AAODI,EAAAA,cAAc,CAACL,MAAD,EAAS;AACrB,UAAMjB,IAAI,GAAG,CAAC,GAAG,KAAKA,IAAT,CAAb;AACAA,IAAAA,IAAI,CAACmB,KAAL;AACA,UAAMC,KAAK,GAAGpB,IAAI,CAACU,GAAL,EAAd;;AACA,SAAK,MAAMW,SAAX,IAAwBrB,IAAxB,EAA8B;AAC5BiB,MAAAA,MAAM,GAAGA,MAAM,CAACI,SAAD,CAAf;AACD;;AAED,WAAOJ,MAAM,CAACG,KAAD,CAAb;AACD;;AA3F2B","sourcesContent":["/**\n * A parser for a minimal subset of the jsonpath standard\n * Full JSON path parsers for JS exist but are quite large (bundle size)\n *\n * Supports\n *\n * `$.component.component.component`\n */\nexport default class JSONPath {\n path: string[];\n\n constructor(path: JSONPath | string[] | string | null = null) {\n this.path = ['$'];\n\n if (path instanceof JSONPath) {\n // @ts-ignore\n this.path = [...path.path];\n return;\n }\n\n if (Array.isArray(path)) {\n this.path.push(...path);\n return;\n }\n\n // Parse a string as a JSONPath\n if (typeof path === 'string') {\n this.path = path.split('.');\n if (this.path[0] !== '$') {\n throw new Error('JSONPaths must start with $');\n }\n }\n }\n\n clone(): JSONPath {\n return new JSONPath(this);\n }\n\n toString(): string {\n return this.path.join('.');\n }\n\n push(name: string): void {\n this.path.push(name);\n }\n\n pop() {\n return this.path.pop();\n }\n\n set(name: string): void {\n this.path[this.path.length - 1] = name;\n }\n\n equals(other: JSONPath): boolean {\n if (!this || !other || this.path.length !== other.path.length) {\n return false;\n }\n\n for (let i = 0; i < this.path.length; ++i) {\n if (this.path[i] !== other.path[i]) {\n return false;\n }\n }\n\n return true;\n }\n\n /**\n * Sets the value pointed at by path\n * TODO - handle root path\n * @param object\n * @param value\n */\n setFieldAtPath(object, value) {\n const path = [...this.path];\n path.shift();\n const field = path.pop();\n for (const component of path) {\n object = object[component];\n }\n // @ts-ignore\n object[field] = value;\n }\n\n /**\n * Gets the value pointed at by path\n * TODO - handle root path\n * @param object\n */\n getFieldAtPath(object) {\n const path = [...this.path];\n path.shift();\n const field = path.pop();\n for (const component of path) {\n object = object[component];\n }\n // @ts-ignore\n return object[field];\n }\n}\n"],"file":"jsonpath.js"}
@@ -1,104 +0,0 @@
1
- import { TableBatchBuilder } from '@loaders.gl/schema';
2
- import { assert, makeTextDecoderIterator } from '@loaders.gl/loader-utils';
3
- import StreamingJSONParser from './parser/streaming-json-parser';
4
- import JSONPath from './jsonpath/jsonpath';
5
- export default async function* parseJSONInBatches(binaryAsyncIterator, options) {
6
- var _options$json;
7
-
8
- const asyncIterator = makeTextDecoderIterator(binaryAsyncIterator);
9
- const {
10
- metadata
11
- } = options;
12
- const {
13
- jsonpaths
14
- } = options.json || {};
15
- let isFirstChunk = true;
16
- const schema = null;
17
- const shape = (options === null || options === void 0 ? void 0 : (_options$json = options.json) === null || _options$json === void 0 ? void 0 : _options$json.shape) || 'row-table';
18
- const tableBatchBuilder = new TableBatchBuilder(schema, { ...options,
19
- shape
20
- });
21
- const parser = new StreamingJSONParser({
22
- jsonpaths
23
- });
24
-
25
- for await (const chunk of asyncIterator) {
26
- const rows = parser.write(chunk);
27
- const jsonpath = rows.length > 0 && parser.getStreamingJsonPathAsString();
28
-
29
- if (rows.length > 0 && isFirstChunk) {
30
- if (metadata) {
31
- const initialBatch = {
32
- shape,
33
- batchType: 'partial-result',
34
- data: [],
35
- length: 0,
36
- bytesUsed: 0,
37
- container: parser.getPartialResult(),
38
- jsonpath
39
- };
40
- yield initialBatch;
41
- }
42
-
43
- isFirstChunk = false;
44
- }
45
-
46
- for (const row of rows) {
47
- tableBatchBuilder.addRow(row);
48
- const batch = tableBatchBuilder.getFullBatch({
49
- jsonpath
50
- });
51
-
52
- if (batch) {
53
- yield batch;
54
- }
55
- }
56
-
57
- tableBatchBuilder.chunkComplete(chunk);
58
- const batch = tableBatchBuilder.getFullBatch({
59
- jsonpath
60
- });
61
-
62
- if (batch) {
63
- yield batch;
64
- }
65
- }
66
-
67
- const jsonpath = parser.getStreamingJsonPathAsString();
68
- const batch = tableBatchBuilder.getFinalBatch({
69
- jsonpath
70
- });
71
-
72
- if (batch) {
73
- yield batch;
74
- }
75
-
76
- if (metadata) {
77
- const finalBatch = {
78
- shape,
79
- batchType: 'final-result',
80
- container: parser.getPartialResult(),
81
- jsonpath: parser.getStreamingJsonPathAsString(),
82
- data: [],
83
- length: 0
84
- };
85
- yield finalBatch;
86
- }
87
- }
88
- export function rebuildJsonObject(batch, data) {
89
- assert(batch.batchType === 'final-result');
90
-
91
- if (batch.jsonpath === '$') {
92
- return data;
93
- }
94
-
95
- if (batch.jsonpath && batch.jsonpath.length > 1) {
96
- const topLevelObject = batch.container;
97
- const streamingPath = new JSONPath(batch.jsonpath);
98
- streamingPath.setFieldAtPath(topLevelObject, data);
99
- return topLevelObject;
100
- }
101
-
102
- return batch.container;
103
- }
104
- //# sourceMappingURL=parse-json-in-batches.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../src/lib/parse-json-in-batches.ts"],"names":["TableBatchBuilder","assert","makeTextDecoderIterator","StreamingJSONParser","JSONPath","parseJSONInBatches","binaryAsyncIterator","options","asyncIterator","metadata","jsonpaths","json","isFirstChunk","schema","shape","tableBatchBuilder","parser","chunk","rows","write","jsonpath","length","getStreamingJsonPathAsString","initialBatch","batchType","data","bytesUsed","container","getPartialResult","row","addRow","batch","getFullBatch","chunkComplete","getFinalBatch","finalBatch","rebuildJsonObject","topLevelObject","streamingPath","setFieldAtPath"],"mappings":"AAEA,SAAQA,iBAAR,QAAgC,oBAAhC;AACA,SAAQC,MAAR,EAAgBC,uBAAhB,QAA8C,0BAA9C;AACA,OAAOC,mBAAP,MAAgC,gCAAhC;AACA,OAAOC,QAAP,MAAqB,qBAArB;AAIA,eAAe,gBAAgBC,kBAAhB,CACbC,mBADa,EAEbC,OAFa,EAGS;AAAA;;AACtB,QAAMC,aAAa,GAAGN,uBAAuB,CAACI,mBAAD,CAA7C;AAEA,QAAM;AAACG,IAAAA;AAAD,MAAaF,OAAnB;AACA,QAAM;AAACG,IAAAA;AAAD,MAAcH,OAAO,CAACI,IAAR,IAAgB,EAApC;AAEA,MAAIC,YAAqB,GAAG,IAA5B;AAGA,QAAMC,MAAM,GAAG,IAAf;AACA,QAAMC,KAAK,GAAG,CAAAP,OAAO,SAAP,IAAAA,OAAO,WAAP,6BAAAA,OAAO,CAAEI,IAAT,gEAAeG,KAAf,KAAwB,WAAtC;AAEA,QAAMC,iBAAiB,GAAG,IAAIf,iBAAJ,CAAsBa,MAAtB,EAA8B,EACtD,GAAGN,OADmD;AAEtDO,IAAAA;AAFsD,GAA9B,CAA1B;AAKA,QAAME,MAAM,GAAG,IAAIb,mBAAJ,CAAwB;AAACO,IAAAA;AAAD,GAAxB,CAAf;;AAEA,aAAW,MAAMO,KAAjB,IAA0BT,aAA1B,EAAyC;AACvC,UAAMU,IAAI,GAAGF,MAAM,CAACG,KAAP,CAAaF,KAAb,CAAb;AAEA,UAAMG,QAAQ,GAAGF,IAAI,CAACG,MAAL,GAAc,CAAd,IAAmBL,MAAM,CAACM,4BAAP,EAApC;;AAEA,QAAIJ,IAAI,CAACG,MAAL,GAAc,CAAd,IAAmBT,YAAvB,EAAqC;AACnC,UAAIH,QAAJ,EAAc;AACZ,cAAMc,YAAmB,GAAG;AAE1BT,UAAAA,KAF0B;AAG1BU,UAAAA,SAAS,EAAE,gBAHe;AAI1BC,UAAAA,IAAI,EAAE,EAJoB;AAK1BJ,UAAAA,MAAM,EAAE,CALkB;AAM1BK,UAAAA,SAAS,EAAE,CANe;AAQ1BC,UAAAA,SAAS,EAAEX,MAAM,CAACY,gBAAP,EARe;AAS1BR,UAAAA;AAT0B,SAA5B;AAWA,cAAMG,YAAN;AACD;;AACDX,MAAAA,YAAY,GAAG,KAAf;AAED;;AAGD,SAAK,MAAMiB,GAAX,IAAkBX,IAAlB,EAAwB;AACtBH,MAAAA,iBAAiB,CAACe,MAAlB,CAAyBD,GAAzB;AAEA,YAAME,KAAK,GAAGhB,iBAAiB,CAACiB,YAAlB,CAA+B;AAACZ,QAAAA;AAAD,OAA/B,CAAd;;AACA,UAAIW,KAAJ,EAAW;AACT,cAAMA,KAAN;AACD;AACF;;AAEDhB,IAAAA,iBAAiB,CAACkB,aAAlB,CAAgChB,KAAhC;AACA,UAAMc,KAAK,GAAGhB,iBAAiB,CAACiB,YAAlB,CAA+B;AAACZ,MAAAA;AAAD,KAA/B,CAAd;;AACA,QAAIW,KAAJ,EAAW;AACT,YAAMA,KAAN;AACD;AACF;;AAGD,QAAMX,QAAQ,GAAGJ,MAAM,CAACM,4BAAP,EAAjB;AACA,QAAMS,KAAK,GAAGhB,iBAAiB,CAACmB,aAAlB,CAAgC;AAACd,IAAAA;AAAD,GAAhC,CAAd;;AACA,MAAIW,KAAJ,EAAW;AACT,UAAMA,KAAN;AACD;;AAED,MAAItB,QAAJ,EAAc;AACZ,UAAM0B,UAAiB,GAAG;AACxBrB,MAAAA,KADwB;AAExBU,MAAAA,SAAS,EAAE,cAFa;AAGxBG,MAAAA,SAAS,EAAEX,MAAM,CAACY,gBAAP,EAHa;AAIxBR,MAAAA,QAAQ,EAAEJ,MAAM,CAACM,4BAAP,EAJc;AAKxBG,MAAAA,IAAI,EAAE,EALkB;AAMxBJ,MAAAA,MAAM,EAAE;AANgB,KAA1B;AASA,UAAMc,UAAN;AACD;AACF;AAED,OAAO,SAASC,iBAAT,CAA2BL,KAA3B,EAAkCN,IAAlC,EAAwC;AAE7CxB,EAAAA,MAAM,CAAC8B,KAAK,CAACP,SAAN,KAAoB,cAArB,CAAN;;AAGA,MAAIO,KAAK,CAACX,QAAN,KAAmB,GAAvB,EAA4B;AAC1B,WAAOK,IAAP;AACD;;AAGD,MAAIM,KAAK,CAACX,QAAN,IAAkBW,KAAK,CAACX,QAAN,CAAeC,MAAf,GAAwB,CAA9C,EAAiD;AAC/C,UAAMgB,cAAc,GAAGN,KAAK,CAACJ,SAA7B;AACA,UAAMW,aAAa,GAAG,IAAIlC,QAAJ,CAAa2B,KAAK,CAACX,QAAnB,CAAtB;AACAkB,IAAAA,aAAa,CAACC,cAAd,CAA6BF,cAA7B,EAA6CZ,IAA7C;AACA,WAAOY,cAAP;AACD;;AAGD,SAAON,KAAK,CAACJ,SAAb;AACD","sourcesContent":["import type {Batch} from '@loaders.gl/schema';\nimport type {JSONLoaderOptions} from '../json-loader';\nimport {TableBatchBuilder} from '@loaders.gl/schema';\nimport {assert, makeTextDecoderIterator} from '@loaders.gl/loader-utils';\nimport StreamingJSONParser from './parser/streaming-json-parser';\nimport JSONPath from './jsonpath/jsonpath';\n\n// TODO - support batch size 0 = no batching/single batch?\n// eslint-disable-next-line max-statements, complexity\nexport default async function* parseJSONInBatches(\n binaryAsyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options: JSONLoaderOptions\n): AsyncIterable<Batch> {\n const asyncIterator = makeTextDecoderIterator(binaryAsyncIterator);\n\n const {metadata} = options;\n const {jsonpaths} = options.json || {};\n\n let isFirstChunk: boolean = true;\n\n // TODO fix Schema deduction\n const schema = null; // new Schema([]);\n const shape = options?.json?.shape || 'row-table';\n // @ts-ignore\n const tableBatchBuilder = new TableBatchBuilder(schema, {\n ...options,\n shape\n });\n\n const parser = new StreamingJSONParser({jsonpaths});\n\n for await (const chunk of asyncIterator) {\n const rows = parser.write(chunk);\n\n const jsonpath = rows.length > 0 && parser.getStreamingJsonPathAsString();\n\n if (rows.length > 0 && isFirstChunk) {\n if (metadata) {\n const initialBatch: Batch = {\n // Common fields\n shape,\n batchType: 'partial-result',\n data: [],\n length: 0,\n bytesUsed: 0,\n // JSON additions\n container: parser.getPartialResult(),\n jsonpath\n };\n yield initialBatch;\n }\n isFirstChunk = false;\n // schema = deduceSchema(rows);\n }\n\n // Add the row\n for (const row of rows) {\n tableBatchBuilder.addRow(row);\n // If a batch has been completed, emit it\n const batch = tableBatchBuilder.getFullBatch({jsonpath});\n if (batch) {\n yield batch;\n }\n }\n\n tableBatchBuilder.chunkComplete(chunk);\n const batch = tableBatchBuilder.getFullBatch({jsonpath});\n if (batch) {\n yield batch;\n }\n }\n\n // yield final batch\n const jsonpath = parser.getStreamingJsonPathAsString();\n const batch = tableBatchBuilder.getFinalBatch({jsonpath});\n if (batch) {\n yield batch;\n }\n\n if (metadata) {\n const finalBatch: Batch = {\n shape,\n batchType: 'final-result',\n container: parser.getPartialResult(),\n jsonpath: parser.getStreamingJsonPathAsString(),\n data: [],\n length: 0\n // schema: null\n };\n yield finalBatch;\n }\n}\n\nexport function rebuildJsonObject(batch, data) {\n // Last batch will have this special type and will provide all the root object of the parsed file\n assert(batch.batchType === 'final-result');\n\n // The streamed JSON data is a top level array (jsonpath = '$'), just return the array of row objects\n if (batch.jsonpath === '$') {\n return data;\n }\n\n // (jsonpath !== '$') The streamed data is not a top level array, so stitch it back in to the top-level object\n if (batch.jsonpath && batch.jsonpath.length > 1) {\n const topLevelObject = batch.container;\n const streamingPath = new JSONPath(batch.jsonpath);\n streamingPath.setFieldAtPath(topLevelObject, data);\n return topLevelObject;\n }\n\n // No jsonpath, in this case nothing was streamed.\n return batch.container;\n}\n"],"file":"parse-json-in-batches.js"}
@@ -1,34 +0,0 @@
1
- export default function parseJSONSync(jsonText, options) {
2
- try {
3
- var _options$json;
4
-
5
- const json = JSON.parse(jsonText);
6
-
7
- if ((_options$json = options.json) !== null && _options$json !== void 0 && _options$json.table) {
8
- return getFirstArray(json) || json;
9
- }
10
-
11
- return json;
12
- } catch (error) {
13
- throw new Error('JSONLoader: failed to parse JSON');
14
- }
15
- }
16
-
17
- function getFirstArray(json) {
18
- if (Array.isArray(json)) {
19
- return json;
20
- }
21
-
22
- if (json && typeof json === 'object') {
23
- for (const value of Object.values(json)) {
24
- const array = getFirstArray(value);
25
-
26
- if (array) {
27
- return array;
28
- }
29
- }
30
- }
31
-
32
- return null;
33
- }
34
- //# sourceMappingURL=parse-json.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../src/lib/parse-json.ts"],"names":["parseJSONSync","jsonText","options","json","JSON","parse","table","getFirstArray","error","Error","Array","isArray","value","Object","values","array"],"mappings":"AAEA,eAAe,SAASA,aAAT,CAAuBC,QAAvB,EAAyCC,OAAzC,EAAqE;AAClF,MAAI;AAAA;;AACF,UAAMC,IAAI,GAAGC,IAAI,CAACC,KAAL,CAAWJ,QAAX,CAAb;;AACA,yBAAIC,OAAO,CAACC,IAAZ,0CAAI,cAAcG,KAAlB,EAAyB;AACvB,aAAOC,aAAa,CAACJ,IAAD,CAAb,IAAuBA,IAA9B;AACD;;AACD,WAAOA,IAAP;AACD,GAND,CAME,OAAOK,KAAP,EAAc;AACd,UAAM,IAAIC,KAAJ,CAAU,kCAAV,CAAN;AACD;AACF;;AAED,SAASF,aAAT,CAAuBJ,IAAvB,EAA6B;AAC3B,MAAIO,KAAK,CAACC,OAAN,CAAcR,IAAd,CAAJ,EAAyB;AACvB,WAAOA,IAAP;AACD;;AACD,MAAIA,IAAI,IAAI,OAAOA,IAAP,KAAgB,QAA5B,EAAsC;AACpC,SAAK,MAAMS,KAAX,IAAoBC,MAAM,CAACC,MAAP,CAAcX,IAAd,CAApB,EAAyC;AACvC,YAAMY,KAAK,GAAGR,aAAa,CAACK,KAAD,CAA3B;;AACA,UAAIG,KAAJ,EAAW;AACT,eAAOA,KAAP;AACD;AACF;AACF;;AACD,SAAO,IAAP;AACD","sourcesContent":["import type {JSONLoaderOptions} from '../json-loader';\n\nexport default function parseJSONSync(jsonText: string, options: JSONLoaderOptions) {\n try {\n const json = JSON.parse(jsonText);\n if (options.json?.table) {\n return getFirstArray(json) || json;\n }\n return json;\n } catch (error) {\n throw new Error('JSONLoader: failed to parse JSON');\n }\n}\n\nfunction getFirstArray(json) {\n if (Array.isArray(json)) {\n return json;\n }\n if (json && typeof json === 'object') {\n for (const value of Object.values(json)) {\n const array = getFirstArray(value);\n if (array) {\n return array;\n }\n }\n }\n return null;\n}\n"],"file":"parse-json.js"}
@@ -1,37 +0,0 @@
1
- import { TableBatchBuilder } from '@loaders.gl/schema';
2
- import { makeLineIterator, makeNumberedLineIterator, makeTextDecoderIterator } from '@loaders.gl/loader-utils';
3
- export default async function* parseNDJSONInBatches(binaryAsyncIterator, options) {
4
- const textIterator = makeTextDecoderIterator(binaryAsyncIterator);
5
- const lineIterator = makeLineIterator(textIterator);
6
- const numberedLineIterator = makeNumberedLineIterator(lineIterator);
7
- const schema = null;
8
- const shape = 'row-table';
9
- const tableBatchBuilder = new TableBatchBuilder(schema, { ...options,
10
- shape
11
- });
12
-
13
- for await (const {
14
- counter,
15
- line
16
- } of numberedLineIterator) {
17
- try {
18
- const row = JSON.parse(line);
19
- tableBatchBuilder.addRow(row);
20
- tableBatchBuilder.chunkComplete(line);
21
- const batch = tableBatchBuilder.getFullBatch();
22
-
23
- if (batch) {
24
- yield batch;
25
- }
26
- } catch (error) {
27
- throw new Error("NDJSONLoader: failed to parse JSON on line ".concat(counter));
28
- }
29
- }
30
-
31
- const batch = tableBatchBuilder.getFinalBatch();
32
-
33
- if (batch) {
34
- yield batch;
35
- }
36
- }
37
- //# sourceMappingURL=parse-ndjson-in-batches.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../src/lib/parse-ndjson-in-batches.ts"],"names":["TableBatchBuilder","makeLineIterator","makeNumberedLineIterator","makeTextDecoderIterator","parseNDJSONInBatches","binaryAsyncIterator","options","textIterator","lineIterator","numberedLineIterator","schema","shape","tableBatchBuilder","counter","line","row","JSON","parse","addRow","chunkComplete","batch","getFullBatch","error","Error","getFinalBatch"],"mappings":"AACA,SAAQA,iBAAR,QAAgC,oBAAhC;AACA,SAEEC,gBAFF,EAGEC,wBAHF,EAIEC,uBAJF,QAKO,0BALP;AAOA,eAAe,gBAAgBC,oBAAhB,CACbC,mBADa,EAEbC,OAFa,EAGS;AACtB,QAAMC,YAAY,GAAGJ,uBAAuB,CAACE,mBAAD,CAA5C;AACA,QAAMG,YAAY,GAAGP,gBAAgB,CAACM,YAAD,CAArC;AACA,QAAME,oBAAoB,GAAGP,wBAAwB,CAACM,YAAD,CAArD;AAEA,QAAME,MAAM,GAAG,IAAf;AACA,QAAMC,KAAK,GAAG,WAAd;AAEA,QAAMC,iBAAiB,GAAG,IAAIZ,iBAAJ,CAAsBU,MAAtB,EAA8B,EACtD,GAAGJ,OADmD;AAEtDK,IAAAA;AAFsD,GAA9B,CAA1B;;AAKA,aAAW,MAAM;AAACE,IAAAA,OAAD;AAAUC,IAAAA;AAAV,GAAjB,IAAoCL,oBAApC,EAA0D;AACxD,QAAI;AACF,YAAMM,GAAG,GAAGC,IAAI,CAACC,KAAL,CAAWH,IAAX,CAAZ;AACAF,MAAAA,iBAAiB,CAACM,MAAlB,CAAyBH,GAAzB;AACAH,MAAAA,iBAAiB,CAACO,aAAlB,CAAgCL,IAAhC;AACA,YAAMM,KAAK,GAAGR,iBAAiB,CAACS,YAAlB,EAAd;;AACA,UAAID,KAAJ,EAAW;AACT,cAAMA,KAAN;AACD;AACF,KARD,CAQE,OAAOE,KAAP,EAAc;AACd,YAAM,IAAIC,KAAJ,sDAAwDV,OAAxD,EAAN;AACD;AACF;;AAED,QAAMO,KAAK,GAAGR,iBAAiB,CAACY,aAAlB,EAAd;;AACA,MAAIJ,KAAJ,EAAW;AACT,UAAMA,KAAN;AACD;AACF","sourcesContent":["import type {Batch} from '@loaders.gl/schema';\nimport {TableBatchBuilder} from '@loaders.gl/schema';\nimport {\n LoaderOptions,\n makeLineIterator,\n makeNumberedLineIterator,\n makeTextDecoderIterator\n} from '@loaders.gl/loader-utils';\n\nexport default async function* parseNDJSONInBatches(\n binaryAsyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options?: LoaderOptions\n): AsyncIterable<Batch> {\n const textIterator = makeTextDecoderIterator(binaryAsyncIterator);\n const lineIterator = makeLineIterator(textIterator);\n const numberedLineIterator = makeNumberedLineIterator(lineIterator);\n\n const schema = null;\n const shape = 'row-table';\n // @ts-ignore\n const tableBatchBuilder = new TableBatchBuilder(schema, {\n ...options,\n shape\n });\n\n for await (const {counter, line} of numberedLineIterator) {\n try {\n const row = JSON.parse(line);\n tableBatchBuilder.addRow(row);\n tableBatchBuilder.chunkComplete(line);\n const batch = tableBatchBuilder.getFullBatch();\n if (batch) {\n yield batch;\n }\n } catch (error) {\n throw new Error(`NDJSONLoader: failed to parse JSON on line ${counter}`);\n }\n }\n\n const batch = tableBatchBuilder.getFinalBatch();\n if (batch) {\n yield batch;\n }\n}\n"],"file":"parse-ndjson-in-batches.js"}
@@ -1,11 +0,0 @@
1
- export default function parseNDJSONSync(ndjsonText) {
2
- const lines = ndjsonText.trim().split('\n');
3
- return lines.map((line, counter) => {
4
- try {
5
- return JSON.parse(line);
6
- } catch (error) {
7
- throw new Error("NDJSONLoader: failed to parse JSON on line ".concat(counter + 1));
8
- }
9
- });
10
- }
11
- //# sourceMappingURL=parse-ndjson.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../src/lib/parse-ndjson.ts"],"names":["parseNDJSONSync","ndjsonText","lines","trim","split","map","line","counter","JSON","parse","error","Error"],"mappings":"AAAA,eAAe,SAASA,eAAT,CAAyBC,UAAzB,EAA6C;AAC1D,QAAMC,KAAK,GAAGD,UAAU,CAACE,IAAX,GAAkBC,KAAlB,CAAwB,IAAxB,CAAd;AACA,SAAOF,KAAK,CAACG,GAAN,CAAU,CAACC,IAAD,EAAOC,OAAP,KAAmB;AAClC,QAAI;AACF,aAAOC,IAAI,CAACC,KAAL,CAAWH,IAAX,CAAP;AACD,KAFD,CAEE,OAAOI,KAAP,EAAc;AACd,YAAM,IAAIC,KAAJ,sDAAwDJ,OAAO,GAAG,CAAlE,EAAN;AACD;AACF,GANM,CAAP;AAOD","sourcesContent":["export default function parseNDJSONSync(ndjsonText: string) {\n const lines = ndjsonText.trim().split('\\n');\n return lines.map((line, counter) => {\n try {\n return JSON.parse(line);\n } catch (error) {\n throw new Error(`NDJSONLoader: failed to parse JSON on line ${counter + 1}`);\n }\n });\n}\n"],"file":"parse-ndjson.js"}
@@ -1,128 +0,0 @@
1
- import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
- import ClarinetParser from '../clarinet/clarinet';
3
- import JSONPath from '../jsonpath/jsonpath';
4
- export default class JSONParser {
5
- constructor(options) {
6
- _defineProperty(this, "parser", void 0);
7
-
8
- _defineProperty(this, "result", undefined);
9
-
10
- _defineProperty(this, "previousStates", []);
11
-
12
- _defineProperty(this, "currentState", Object.freeze({
13
- container: [],
14
- key: null
15
- }));
16
-
17
- _defineProperty(this, "jsonpath", new JSONPath());
18
-
19
- this.reset();
20
- this.parser = new ClarinetParser({
21
- onready: () => {
22
- this.jsonpath = new JSONPath();
23
- this.previousStates.length = 0;
24
- this.currentState.container.length = 0;
25
- },
26
- onopenobject: name => {
27
- this._openObject({});
28
-
29
- if (typeof name !== 'undefined') {
30
- this.parser.emit('onkey', name);
31
- }
32
- },
33
- onkey: name => {
34
- this.jsonpath.set(name);
35
- this.currentState.key = name;
36
- },
37
- oncloseobject: () => {
38
- this._closeObject();
39
- },
40
- onopenarray: () => {
41
- this._openArray();
42
- },
43
- onclosearray: () => {
44
- this._closeArray();
45
- },
46
- onvalue: value => {
47
- this._pushOrSet(value);
48
- },
49
- onerror: error => {
50
- throw error;
51
- },
52
- onend: () => {
53
- this.result = this.currentState.container.pop();
54
- },
55
- ...options
56
- });
57
- }
58
-
59
- reset() {
60
- this.result = undefined;
61
- this.previousStates = [];
62
- this.currentState = Object.freeze({
63
- container: [],
64
- key: null
65
- });
66
- this.jsonpath = new JSONPath();
67
- }
68
-
69
- write(chunk) {
70
- this.parser.write(chunk);
71
- }
72
-
73
- close() {
74
- this.parser.close();
75
- }
76
-
77
- _pushOrSet(value) {
78
- const {
79
- container,
80
- key
81
- } = this.currentState;
82
-
83
- if (key !== null) {
84
- container[key] = value;
85
- this.currentState.key = null;
86
- } else {
87
- container.push(value);
88
- }
89
- }
90
-
91
- _openArray(newContainer = []) {
92
- this.jsonpath.push(null);
93
-
94
- this._pushOrSet(newContainer);
95
-
96
- this.previousStates.push(this.currentState);
97
- this.currentState = {
98
- container: newContainer,
99
- isArray: true,
100
- key: null
101
- };
102
- }
103
-
104
- _closeArray() {
105
- this.jsonpath.pop();
106
- this.currentState = this.previousStates.pop();
107
- }
108
-
109
- _openObject(newContainer = {}) {
110
- this.jsonpath.push(null);
111
-
112
- this._pushOrSet(newContainer);
113
-
114
- this.previousStates.push(this.currentState);
115
- this.currentState = {
116
- container: newContainer,
117
- isArray: false,
118
- key: null
119
- };
120
- }
121
-
122
- _closeObject() {
123
- this.jsonpath.pop();
124
- this.currentState = this.previousStates.pop();
125
- }
126
-
127
- }
128
- //# sourceMappingURL=json-parser.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../../src/lib/parser/json-parser.ts"],"names":["ClarinetParser","JSONPath","JSONParser","constructor","options","undefined","Object","freeze","container","key","reset","parser","onready","jsonpath","previousStates","length","currentState","onopenobject","name","_openObject","emit","onkey","set","oncloseobject","_closeObject","onopenarray","_openArray","onclosearray","_closeArray","onvalue","value","_pushOrSet","onerror","error","onend","result","pop","write","chunk","close","push","newContainer","isArray"],"mappings":";AAEA,OAAOA,cAAP,MAAoD,sBAApD;AACA,OAAOC,QAAP,MAAqB,sBAArB;AAIA,eAAe,MAAMC,UAAN,CAAiB;AAO9BC,EAAAA,WAAW,CAACC,OAAD,EAAiC;AAAA;;AAAA,oCALnCC,SAKmC;;AAAA,4CAJ3B,EAI2B;;AAAA,0CAH7BC,MAAM,CAACC,MAAP,CAAc;AAACC,MAAAA,SAAS,EAAE,EAAZ;AAAgBC,MAAAA,GAAG,EAAE;AAArB,KAAd,CAG6B;;AAAA,sCAFvB,IAAIR,QAAJ,EAEuB;;AAC1C,SAAKS,KAAL;AACA,SAAKC,MAAL,GAAc,IAAIX,cAAJ,CAAmB;AAC/BY,MAAAA,OAAO,EAAE,MAAM;AACb,aAAKC,QAAL,GAAgB,IAAIZ,QAAJ,EAAhB;AACA,aAAKa,cAAL,CAAoBC,MAApB,GAA6B,CAA7B;AACA,aAAKC,YAAL,CAAkBR,SAAlB,CAA4BO,MAA5B,GAAqC,CAArC;AACD,OAL8B;AAO/BE,MAAAA,YAAY,EAAGC,IAAD,IAAU;AACtB,aAAKC,WAAL,CAAiB,EAAjB;;AACA,YAAI,OAAOD,IAAP,KAAgB,WAApB,EAAiC;AAC/B,eAAKP,MAAL,CAAYS,IAAZ,CAAiB,OAAjB,EAA0BF,IAA1B;AACD;AACF,OAZ8B;AAc/BG,MAAAA,KAAK,EAAGH,IAAD,IAAU;AACf,aAAKL,QAAL,CAAcS,GAAd,CAAkBJ,IAAlB;AACA,aAAKF,YAAL,CAAkBP,GAAlB,GAAwBS,IAAxB;AACD,OAjB8B;AAmB/BK,MAAAA,aAAa,EAAE,MAAM;AACnB,aAAKC,YAAL;AACD,OArB8B;AAuB/BC,MAAAA,WAAW,EAAE,MAAM;AACjB,aAAKC,UAAL;AACD,OAzB8B;AA2B/BC,MAAAA,YAAY,EAAE,MAAM;AAClB,aAAKC,WAAL;AACD,OA7B8B;AA+B/BC,MAAAA,OAAO,EAAGC,KAAD,IAAW;AAClB,aAAKC,UAAL,CAAgBD,KAAhB;AACD,OAjC8B;AAmC/BE,MAAAA,OAAO,EAAGC,KAAD,IAAW;AAClB,cAAMA,KAAN;AACD,OArC8B;AAuC/BC,MAAAA,KAAK,EAAE,MAAM;AACX,aAAKC,MAAL,GAAc,KAAKnB,YAAL,CAAkBR,SAAlB,CAA4B4B,GAA5B,EAAd;AACD,OAzC8B;AA2C/B,SAAGhC;AA3C4B,KAAnB,CAAd;AA6CD;;AAEDM,EAAAA,KAAK,GAAS;AACZ,SAAKyB,MAAL,GAAc9B,SAAd;AACA,SAAKS,cAAL,GAAsB,EAAtB;AACA,SAAKE,YAAL,GAAoBV,MAAM,CAACC,MAAP,CAAc;AAACC,MAAAA,SAAS,EAAE,EAAZ;AAAgBC,MAAAA,GAAG,EAAE;AAArB,KAAd,CAApB;AACA,SAAKI,QAAL,GAAgB,IAAIZ,QAAJ,EAAhB;AACD;;AAEDoC,EAAAA,KAAK,CAACC,KAAD,EAAc;AACjB,SAAK3B,MAAL,CAAY0B,KAAZ,CAAkBC,KAAlB;AACD;;AAEDC,EAAAA,KAAK,GAAS;AACZ,SAAK5B,MAAL,CAAY4B,KAAZ;AACD;;AAIDR,EAAAA,UAAU,CAACD,KAAD,EAAc;AACtB,UAAM;AAACtB,MAAAA,SAAD;AAAYC,MAAAA;AAAZ,QAAmB,KAAKO,YAA9B;;AACA,QAAIP,GAAG,KAAK,IAAZ,EAAkB;AAChBD,MAAAA,SAAS,CAACC,GAAD,CAAT,GAAiBqB,KAAjB;AACA,WAAKd,YAAL,CAAkBP,GAAlB,GAAwB,IAAxB;AACD,KAHD,MAGO;AACLD,MAAAA,SAAS,CAACgC,IAAV,CAAeV,KAAf;AACD;AACF;;AAEDJ,EAAAA,UAAU,CAACe,YAAY,GAAG,EAAhB,EAA0B;AAClC,SAAK5B,QAAL,CAAc2B,IAAd,CAAmB,IAAnB;;AACA,SAAKT,UAAL,CAAgBU,YAAhB;;AACA,SAAK3B,cAAL,CAAoB0B,IAApB,CAAyB,KAAKxB,YAA9B;AACA,SAAKA,YAAL,GAAoB;AAACR,MAAAA,SAAS,EAAEiC,YAAZ;AAA0BC,MAAAA,OAAO,EAAE,IAAnC;AAAyCjC,MAAAA,GAAG,EAAE;AAA9C,KAApB;AACD;;AAEDmB,EAAAA,WAAW,GAAS;AAClB,SAAKf,QAAL,CAAcuB,GAAd;AACA,SAAKpB,YAAL,GAAoB,KAAKF,cAAL,CAAoBsB,GAApB,EAApB;AACD;;AAEDjB,EAAAA,WAAW,CAACsB,YAAY,GAAG,EAAhB,EAA0B;AACnC,SAAK5B,QAAL,CAAc2B,IAAd,CAAmB,IAAnB;;AACA,SAAKT,UAAL,CAAgBU,YAAhB;;AACA,SAAK3B,cAAL,CAAoB0B,IAApB,CAAyB,KAAKxB,YAA9B;AACA,SAAKA,YAAL,GAAoB;AAACR,MAAAA,SAAS,EAAEiC,YAAZ;AAA0BC,MAAAA,OAAO,EAAE,KAAnC;AAA0CjC,MAAAA,GAAG,EAAE;AAA/C,KAApB;AACD;;AAEDe,EAAAA,YAAY,GAAS;AACnB,SAAKX,QAAL,CAAcuB,GAAd;AACA,SAAKpB,YAAL,GAAoB,KAAKF,cAAL,CAAoBsB,GAApB,EAApB;AACD;;AAzG6B","sourcesContent":["// @ts-nocheck\n\nimport ClarinetParser, {ClarinetParserOptions} from '../clarinet/clarinet';\nimport JSONPath from '../jsonpath/jsonpath';\n\n// JSONParser builds a JSON object using the events emitted by the Clarinet parser\n\nexport default class JSONParser {\n readonly parser: ClarinetParser;\n result = undefined;\n previousStates = [];\n currentState = Object.freeze({container: [], key: null});\n jsonpath: JSONPath = new JSONPath();\n\n constructor(options: ClarinetParserOptions) {\n this.reset();\n this.parser = new ClarinetParser({\n onready: () => {\n this.jsonpath = new JSONPath();\n this.previousStates.length = 0;\n this.currentState.container.length = 0;\n },\n\n onopenobject: (name) => {\n this._openObject({});\n if (typeof name !== 'undefined') {\n this.parser.emit('onkey', name);\n }\n },\n\n onkey: (name) => {\n this.jsonpath.set(name);\n this.currentState.key = name;\n },\n\n oncloseobject: () => {\n this._closeObject();\n },\n\n onopenarray: () => {\n this._openArray();\n },\n\n onclosearray: () => {\n this._closeArray();\n },\n\n onvalue: (value) => {\n this._pushOrSet(value);\n },\n\n onerror: (error) => {\n throw error;\n },\n\n onend: () => {\n this.result = this.currentState.container.pop();\n },\n\n ...options\n });\n }\n\n reset(): void {\n this.result = undefined;\n this.previousStates = [];\n this.currentState = Object.freeze({container: [], key: null});\n this.jsonpath = new JSONPath();\n }\n\n write(chunk): void {\n this.parser.write(chunk);\n }\n\n close(): void {\n this.parser.close();\n }\n\n // PRIVATE METHODS\n\n _pushOrSet(value): void {\n const {container, key} = this.currentState;\n if (key !== null) {\n container[key] = value;\n this.currentState.key = null;\n } else {\n container.push(value);\n }\n }\n\n _openArray(newContainer = []): void {\n this.jsonpath.push(null);\n this._pushOrSet(newContainer);\n this.previousStates.push(this.currentState);\n this.currentState = {container: newContainer, isArray: true, key: null};\n }\n\n _closeArray(): void {\n this.jsonpath.pop();\n this.currentState = this.previousStates.pop();\n }\n\n _openObject(newContainer = {}): void {\n this.jsonpath.push(null);\n this._pushOrSet(newContainer);\n this.previousStates.push(this.currentState);\n this.currentState = {container: newContainer, isArray: false, key: null};\n }\n\n _closeObject(): void {\n this.jsonpath.pop();\n this.currentState = this.previousStates.pop();\n }\n}\n"],"file":"json-parser.js"}
@@ -1,93 +0,0 @@
1
- import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
- import { default as JSONParser } from './json-parser';
3
- import JSONPath from '../jsonpath/jsonpath';
4
- export default class StreamingJSONParser extends JSONParser {
5
- constructor(options = {}) {
6
- super({
7
- onopenarray: () => {
8
- if (!this.streamingArray) {
9
- if (this._matchJSONPath()) {
10
- this.streamingJsonPath = this.getJsonPath().clone();
11
- this.streamingArray = [];
12
-
13
- this._openArray(this.streamingArray);
14
-
15
- return;
16
- }
17
- }
18
-
19
- this._openArray();
20
- },
21
- onopenobject: name => {
22
- if (!this.topLevelObject) {
23
- this.topLevelObject = {};
24
-
25
- this._openObject(this.topLevelObject);
26
- } else {
27
- this._openObject({});
28
- }
29
-
30
- if (typeof name !== 'undefined') {
31
- this.parser.emit('onkey', name);
32
- }
33
- }
34
- });
35
-
36
- _defineProperty(this, "jsonPaths", void 0);
37
-
38
- _defineProperty(this, "streamingJsonPath", null);
39
-
40
- _defineProperty(this, "streamingArray", null);
41
-
42
- _defineProperty(this, "topLevelObject", null);
43
-
44
- const jsonpaths = options.jsonpaths || [];
45
- this.jsonPaths = jsonpaths.map(jsonpath => new JSONPath(jsonpath));
46
- }
47
-
48
- write(chunk) {
49
- super.write(chunk);
50
- let array = [];
51
-
52
- if (this.streamingArray) {
53
- array = [...this.streamingArray];
54
- this.streamingArray.length = 0;
55
- }
56
-
57
- return array;
58
- }
59
-
60
- getPartialResult() {
61
- return this.topLevelObject;
62
- }
63
-
64
- getStreamingJsonPath() {
65
- return this.streamingJsonPath;
66
- }
67
-
68
- getStreamingJsonPathAsString() {
69
- return this.streamingJsonPath && this.streamingJsonPath.toString();
70
- }
71
-
72
- getJsonPath() {
73
- return this.jsonpath;
74
- }
75
-
76
- _matchJSONPath() {
77
- const currentPath = this.getJsonPath();
78
-
79
- if (this.jsonPaths.length === 0) {
80
- return true;
81
- }
82
-
83
- for (const jsonPath of this.jsonPaths) {
84
- if (jsonPath.equals(currentPath)) {
85
- return true;
86
- }
87
- }
88
-
89
- return false;
90
- }
91
-
92
- }
93
- //# sourceMappingURL=streaming-json-parser.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../../src/lib/parser/streaming-json-parser.ts"],"names":["default","JSONParser","JSONPath","StreamingJSONParser","constructor","options","onopenarray","streamingArray","_matchJSONPath","streamingJsonPath","getJsonPath","clone","_openArray","onopenobject","name","topLevelObject","_openObject","parser","emit","jsonpaths","jsonPaths","map","jsonpath","write","chunk","array","length","getPartialResult","getStreamingJsonPath","getStreamingJsonPathAsString","toString","currentPath","jsonPath","equals"],"mappings":";AAAA,SAAQA,OAAO,IAAIC,UAAnB,QAAoC,eAApC;AACA,OAAOC,QAAP,MAAqB,sBAArB;AAMA,eAAe,MAAMC,mBAAN,SAAkCF,UAAlC,CAA6C;AAM1DG,EAAAA,WAAW,CAACC,OAA6B,GAAG,EAAjC,EAAqC;AAC9C,UAAM;AACJC,MAAAA,WAAW,EAAE,MAAM;AACjB,YAAI,CAAC,KAAKC,cAAV,EAA0B;AACxB,cAAI,KAAKC,cAAL,EAAJ,EAA2B;AAEzB,iBAAKC,iBAAL,GAAyB,KAAKC,WAAL,GAAmBC,KAAnB,EAAzB;AACA,iBAAKJ,cAAL,GAAsB,EAAtB;;AACA,iBAAKK,UAAL,CAAgB,KAAKL,cAArB;;AACA;AACD;AACF;;AAED,aAAKK,UAAL;AACD,OAbG;AAgBJC,MAAAA,YAAY,EAAGC,IAAD,IAAU;AACtB,YAAI,CAAC,KAAKC,cAAV,EAA0B;AACxB,eAAKA,cAAL,GAAsB,EAAtB;;AACA,eAAKC,WAAL,CAAiB,KAAKD,cAAtB;AACD,SAHD,MAGO;AACL,eAAKC,WAAL,CAAiB,EAAjB;AACD;;AACD,YAAI,OAAOF,IAAP,KAAgB,WAApB,EAAiC;AAC/B,eAAKG,MAAL,CAAYC,IAAZ,CAAiB,OAAjB,EAA0BJ,IAA1B;AACD;AACF;AA1BG,KAAN;;AAD8C;;AAAA,+CAJH,IAIG;;AAAA,4CAHT,IAGS;;AAAA,4CAFR,IAEQ;;AA6B9C,UAAMK,SAAS,GAAGd,OAAO,CAACc,SAAR,IAAqB,EAAvC;AACA,SAAKC,SAAL,GAAiBD,SAAS,CAACE,GAAV,CAAeC,QAAD,IAAc,IAAIpB,QAAJ,CAAaoB,QAAb,CAA5B,CAAjB;AACD;;AASDC,EAAAA,KAAK,CAACC,KAAD,EAAQ;AACX,UAAMD,KAAN,CAAYC,KAAZ;AACA,QAAIC,KAAY,GAAG,EAAnB;;AACA,QAAI,KAAKlB,cAAT,EAAyB;AACvBkB,MAAAA,KAAK,GAAG,CAAC,GAAG,KAAKlB,cAAT,CAAR;AACA,WAAKA,cAAL,CAAoBmB,MAApB,GAA6B,CAA7B;AACD;;AACD,WAAOD,KAAP;AACD;;AAODE,EAAAA,gBAAgB,GAAG;AACjB,WAAO,KAAKZ,cAAZ;AACD;;AAEDa,EAAAA,oBAAoB,GAAG;AACrB,WAAO,KAAKnB,iBAAZ;AACD;;AAEDoB,EAAAA,4BAA4B,GAAG;AAC7B,WAAO,KAAKpB,iBAAL,IAA0B,KAAKA,iBAAL,CAAuBqB,QAAvB,EAAjC;AACD;;AAEDpB,EAAAA,WAAW,GAAG;AACZ,WAAO,KAAKY,QAAZ;AACD;;AAODd,EAAAA,cAAc,GAAG;AACf,UAAMuB,WAAW,GAAG,KAAKrB,WAAL,EAApB;;AAKA,QAAI,KAAKU,SAAL,CAAeM,MAAf,KAA0B,CAA9B,EAAiC;AAC/B,aAAO,IAAP;AACD;;AAED,SAAK,MAAMM,QAAX,IAAuB,KAAKZ,SAA5B,EAAuC;AACrC,UAAIY,QAAQ,CAACC,MAAT,CAAgBF,WAAhB,CAAJ,EAAkC;AAChC,eAAO,IAAP;AACD;AACF;;AAED,WAAO,KAAP;AACD;;AAnGyD","sourcesContent":["import {default as JSONParser} from './json-parser';\nimport JSONPath from '../jsonpath/jsonpath';\n\n/**\n * The `StreamingJSONParser` looks for the first array in the JSON structure.\n * and emits an array of chunks\n */\nexport default class StreamingJSONParser extends JSONParser {\n private jsonPaths: JSONPath[];\n private streamingJsonPath: JSONPath | null = null;\n private streamingArray: any[] | null = null;\n private topLevelObject: object | null = null;\n\n constructor(options: {[key: string]: any} = {}) {\n super({\n onopenarray: () => {\n if (!this.streamingArray) {\n if (this._matchJSONPath()) {\n // @ts-ignore\n this.streamingJsonPath = this.getJsonPath().clone();\n this.streamingArray = [];\n this._openArray(this.streamingArray as []);\n return;\n }\n }\n\n this._openArray();\n },\n\n // Redefine onopenarray to inject value for top-level object\n onopenobject: (name) => {\n if (!this.topLevelObject) {\n this.topLevelObject = {};\n this._openObject(this.topLevelObject);\n } else {\n this._openObject({});\n }\n if (typeof name !== 'undefined') {\n this.parser.emit('onkey', name);\n }\n }\n });\n const jsonpaths = options.jsonpaths || [];\n this.jsonPaths = jsonpaths.map((jsonpath) => new JSONPath(jsonpath));\n }\n\n /**\n * write REDEFINITION\n * - super.write() chunk to parser\n * - get the contents (so far) of \"topmost-level\" array as batch of rows\n * - clear top-level array\n * - return the batch of rows\\\n */\n write(chunk) {\n super.write(chunk);\n let array: any[] = [];\n if (this.streamingArray) {\n array = [...this.streamingArray];\n this.streamingArray.length = 0;\n }\n return array;\n }\n\n /**\n * Returns a partially formed result object\n * Useful for returning the \"wrapper\" object when array is not top level\n * e.g. GeoJSON\n */\n getPartialResult() {\n return this.topLevelObject;\n }\n\n getStreamingJsonPath() {\n return this.streamingJsonPath;\n }\n\n getStreamingJsonPathAsString() {\n return this.streamingJsonPath && this.streamingJsonPath.toString();\n }\n\n getJsonPath() {\n return this.jsonpath;\n }\n\n // PRIVATE METHODS\n\n /**\n * Checks is this.getJsonPath matches the jsonpaths provided in options\n */\n _matchJSONPath() {\n const currentPath = this.getJsonPath();\n // console.debug(`Testing JSONPath`, currentPath);\n\n // Backwards compatibility, match any array\n // TODO implement using wildcard once that is supported\n if (this.jsonPaths.length === 0) {\n return true;\n }\n\n for (const jsonPath of this.jsonPaths) {\n if (jsonPath.equals(currentPath)) {\n return true;\n }\n }\n\n return false;\n }\n}\n"],"file":"streaming-json-parser.js"}
@@ -1,32 +0,0 @@
1
- import parseNDJSONSync from './lib/parse-ndjson';
2
- import parseNDJSONInBatches from './lib/parse-ndjson-in-batches';
3
- const VERSION = typeof "3.1.3" !== 'undefined' ? "3.1.3" : 'latest';
4
- export const NDJSONLoader = {
5
- name: 'NDJSON',
6
- id: 'ndjson',
7
- module: 'json',
8
- version: VERSION,
9
- extensions: ['ndjson'],
10
- mimeTypes: ['application/x-ndjson'],
11
- category: 'table',
12
- text: true,
13
- parse,
14
- parseTextSync,
15
- parseInBatches,
16
- options: {}
17
- };
18
-
19
- async function parse(arrayBuffer) {
20
- return parseTextSync(new TextDecoder().decode(arrayBuffer));
21
- }
22
-
23
- function parseTextSync(text) {
24
- return parseNDJSONSync(text);
25
- }
26
-
27
- function parseInBatches(asyncIterator, options) {
28
- return parseNDJSONInBatches(asyncIterator, options);
29
- }
30
-
31
- export const _typecheckNDJSONLoader = NDJSONLoader;
32
- //# sourceMappingURL=ndjson-loader.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../src/ndjson-loader.ts"],"names":["parseNDJSONSync","parseNDJSONInBatches","VERSION","NDJSONLoader","name","id","module","version","extensions","mimeTypes","category","text","parse","parseTextSync","parseInBatches","options","arrayBuffer","TextDecoder","decode","asyncIterator","_typecheckNDJSONLoader"],"mappings":"AAEA,OAAOA,eAAP,MAA4B,oBAA5B;AACA,OAAOC,oBAAP,MAAiC,+BAAjC;AAIA,MAAMC,OAAO,GAAG,mBAAuB,WAAvB,aAAmD,QAAnE;AAEA,OAAO,MAAMC,YAAY,GAAG;AAC1BC,EAAAA,IAAI,EAAE,QADoB;AAE1BC,EAAAA,EAAE,EAAE,QAFsB;AAG1BC,EAAAA,MAAM,EAAE,MAHkB;AAI1BC,EAAAA,OAAO,EAAEL,OAJiB;AAK1BM,EAAAA,UAAU,EAAE,CAAC,QAAD,CALc;AAM1BC,EAAAA,SAAS,EAAE,CAAC,sBAAD,CANe;AAO1BC,EAAAA,QAAQ,EAAE,OAPgB;AAQ1BC,EAAAA,IAAI,EAAE,IARoB;AAS1BC,EAAAA,KAT0B;AAU1BC,EAAAA,aAV0B;AAW1BC,EAAAA,cAX0B;AAY1BC,EAAAA,OAAO,EAAE;AAZiB,CAArB;;AAeP,eAAeH,KAAf,CAAqBI,WAArB,EAA+C;AAC7C,SAAOH,aAAa,CAAC,IAAII,WAAJ,GAAkBC,MAAlB,CAAyBF,WAAzB,CAAD,CAApB;AACD;;AAED,SAASH,aAAT,CAAuBF,IAAvB,EAAqC;AACnC,SAAOX,eAAe,CAACW,IAAD,CAAtB;AACD;;AAED,SAASG,cAAT,CACEK,aADF,EAEEJ,OAFF,EAGwB;AACtB,SAAOd,oBAAoB,CAACkB,aAAD,EAAgBJ,OAAhB,CAA3B;AACD;;AAED,OAAO,MAAMK,sBAAwC,GAAGjB,YAAjD","sourcesContent":["import type {Batch} from '@loaders.gl/schema';\nimport type {LoaderWithParser, LoaderOptions} from '@loaders.gl/loader-utils';\nimport parseNDJSONSync from './lib/parse-ndjson';\nimport parseNDJSONInBatches from './lib/parse-ndjson-in-batches';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\nexport const NDJSONLoader = {\n name: 'NDJSON',\n id: 'ndjson',\n module: 'json',\n version: VERSION,\n extensions: ['ndjson'],\n mimeTypes: ['application/x-ndjson'],\n category: 'table',\n text: true,\n parse,\n parseTextSync,\n parseInBatches,\n options: {}\n};\n\nasync function parse(arrayBuffer: ArrayBuffer) {\n return parseTextSync(new TextDecoder().decode(arrayBuffer));\n}\n\nfunction parseTextSync(text: string) {\n return parseNDJSONSync(text);\n}\n\nfunction parseInBatches(\n asyncIterator: AsyncIterable<ArrayBuffer> | Iterable<ArrayBuffer>,\n options?: LoaderOptions\n): AsyncIterable<Batch> {\n return parseNDJSONInBatches(asyncIterator, options);\n}\n\nexport const _typecheckNDJSONLoader: LoaderWithParser = NDJSONLoader;\n"],"file":"ndjson-loader.js"}
@@ -1,4 +0,0 @@
1
- import { createLoaderWorker } from '@loaders.gl/loader-utils';
2
- import { GeoJSONLoader } from '../geojson-loader';
3
- createLoaderWorker(GeoJSONLoader);
4
- //# sourceMappingURL=geojson-worker.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../src/workers/geojson-worker.ts"],"names":["createLoaderWorker","GeoJSONLoader"],"mappings":"AAAA,SAAQA,kBAAR,QAAiC,0BAAjC;AACA,SAAQC,aAAR,QAA4B,mBAA5B;AAEAD,kBAAkB,CAACC,aAAD,CAAlB","sourcesContent":["import {createLoaderWorker} from '@loaders.gl/loader-utils';\nimport {GeoJSONLoader} from '../geojson-loader';\n\ncreateLoaderWorker(GeoJSONLoader);\n"],"file":"geojson-worker.js"}