@loaders.gl/parquet 4.0.0 → 4.0.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +10 -205
- package/dist/index.d.ts +0 -3
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +0 -2
- package/dist/index.js.map +1 -1
- package/dist/lib/parsers/get-parquet-schema.js +1 -1
- package/dist/lib/parsers/get-parquet-schema.js.map +1 -1
- package/dist/lib/parsers/parse-parquet-to-rows.d.ts.map +1 -1
- package/dist/lib/parsers/parse-parquet-to-rows.js +3 -2
- package/dist/lib/parsers/parse-parquet-to-rows.js.map +1 -1
- package/dist/lib/wasm/encode-parquet-wasm.d.ts +7 -8
- package/dist/lib/wasm/encode-parquet-wasm.d.ts.map +1 -1
- package/dist/lib/wasm/encode-parquet-wasm.js +2 -2
- package/dist/lib/wasm/encode-parquet-wasm.js.map +1 -1
- package/dist/lib/wasm/parse-parquet-wasm.d.ts +2 -2
- package/dist/lib/wasm/parse-parquet-wasm.d.ts.map +1 -1
- package/dist/lib/wasm/parse-parquet-wasm.js +3 -3
- package/dist/lib/wasm/parse-parquet-wasm.js.map +1 -1
- package/dist/parquet-wasm-loader.d.ts +2 -2
- package/dist/parquet-wasm-loader.d.ts.map +1 -1
- package/dist/parquet-wasm-loader.js.map +1 -1
- package/dist/parquet-wasm-writer.d.ts +2 -2
- package/dist/parquet-wasm-writer.d.ts.map +1 -1
- package/dist/parquet-wasm-writer.js.map +1 -1
- package/dist/parquet-writer.d.ts.map +1 -1
- package/dist/parquet-writer.js +3 -2
- package/dist/parquet-writer.js.map +1 -1
- package/package.json +8 -8
- package/src/index.ts +2 -10
- package/src/lib/parsers/get-parquet-schema.ts +1 -1
- package/src/lib/parsers/parse-parquet-to-rows.ts +4 -2
- package/src/lib/wasm/encode-parquet-wasm.ts +11 -10
- package/src/lib/wasm/parse-parquet-wasm.ts +6 -7
- package/src/parquet-wasm-loader.ts +2 -2
- package/src/parquet-wasm-writer.ts +2 -2
- package/src/parquet-writer.ts +3 -2
- package/dist/lib/geo/decode-geo-column.d.ts +0 -4
- package/dist/lib/geo/decode-geo-column.d.ts.map +0 -1
- package/dist/lib/geo/decode-geo-column.js +0 -47
- package/dist/lib/geo/decode-geo-column.js.map +0 -1
- package/dist/lib/geo/decode-geo-metadata.d.ts +0 -44
- package/dist/lib/geo/decode-geo-metadata.d.ts.map +0 -1
- package/dist/lib/geo/decode-geo-metadata.js +0 -89
- package/dist/lib/geo/decode-geo-metadata.js.map +0 -1
- package/dist/lib/geo/geoparquet-metadata-schema.d.ts +0 -79
- package/dist/lib/geo/geoparquet-metadata-schema.d.ts.map +0 -1
- package/dist/lib/geo/geoparquet-metadata-schema.js +0 -76
- package/dist/lib/geo/geoparquet-metadata-schema.js.map +0 -1
- package/dist/lib/geo/geoparquet-metadata-schema.json +0 -60
- package/src/lib/geo/decode-geo-column.ts +0 -55
- package/src/lib/geo/decode-geo-metadata.ts +0 -177
- package/src/lib/geo/geoparquet-metadata-schema.json +0 -60
- package/src/lib/geo/geoparquet-metadata-schema.ts +0 -70
package/dist/index.cjs
CHANGED
|
@@ -32,7 +32,6 @@ var src_exports = {};
|
|
|
32
32
|
__export(src_exports, {
|
|
33
33
|
Buffer: () => Buffer3,
|
|
34
34
|
BufferPolyfill: () => Buffer2,
|
|
35
|
-
GEOPARQUET_METADATA_JSON_SCHEMA: () => GEOPARQUET_METADATA_JSON_SCHEMA,
|
|
36
35
|
ParquetColumnarLoader: () => ParquetColumnarLoader2,
|
|
37
36
|
ParquetEncoder: () => ParquetEncoder,
|
|
38
37
|
ParquetLoader: () => ParquetLoader2,
|
|
@@ -42,11 +41,8 @@ __export(src_exports, {
|
|
|
42
41
|
_ParquetWriter: () => ParquetWriter,
|
|
43
42
|
convertParquetSchema: () => convertParquetSchema,
|
|
44
43
|
convertParquetToArrowSchema: () => convertParquetSchema,
|
|
45
|
-
getGeoMetadata: () => getGeoMetadata,
|
|
46
44
|
installBufferPolyfill: () => installBufferPolyfill,
|
|
47
|
-
preloadCompressions: () => preloadCompressions
|
|
48
|
-
setGeoMetadata: () => setGeoMetadata,
|
|
49
|
-
unpackGeoMetadata: () => unpackGeoMetadata
|
|
45
|
+
preloadCompressions: () => preloadCompressions
|
|
50
46
|
});
|
|
51
47
|
module.exports = __toCommonJS(src_exports);
|
|
52
48
|
|
|
@@ -1821,6 +1817,8 @@ ParquetColumnarLoader.Buffer = Buffer;
|
|
|
1821
1817
|
|
|
1822
1818
|
// src/lib/parsers/parse-parquet-to-rows.ts
|
|
1823
1819
|
var import_loader_utils = require("@loaders.gl/loader-utils");
|
|
1820
|
+
var import_gis2 = require("@loaders.gl/gis");
|
|
1821
|
+
var import_wkt = require("@loaders.gl/wkt");
|
|
1824
1822
|
|
|
1825
1823
|
// src/parquetjs/codecs/plain.ts
|
|
1826
1824
|
var import_int53 = __toESM(require("int53"), 1);
|
|
@@ -7188,149 +7186,17 @@ function getSchemaMetadata(parquetMetadata) {
|
|
|
7188
7186
|
return metadata;
|
|
7189
7187
|
}
|
|
7190
7188
|
|
|
7191
|
-
// src/lib/geo/decode-geo-metadata.ts
|
|
7192
|
-
function parseJSONStringMetadata(schema, metadataKey) {
|
|
7193
|
-
const stringifiedMetadata = schema.metadata[metadataKey];
|
|
7194
|
-
if (!stringifiedMetadata) {
|
|
7195
|
-
return null;
|
|
7196
|
-
}
|
|
7197
|
-
try {
|
|
7198
|
-
const metadata = JSON.parse(stringifiedMetadata);
|
|
7199
|
-
if (!metadata || typeof metadata !== "object") {
|
|
7200
|
-
return null;
|
|
7201
|
-
}
|
|
7202
|
-
return metadata;
|
|
7203
|
-
} catch {
|
|
7204
|
-
return null;
|
|
7205
|
-
}
|
|
7206
|
-
}
|
|
7207
|
-
function unpackJSONStringMetadata(schema, metadataKey) {
|
|
7208
|
-
const json = parseJSONStringMetadata(schema, metadataKey);
|
|
7209
|
-
for (const [key, value] of Object.entries(json || {})) {
|
|
7210
|
-
schema.metadata[`${metadataKey}.${key}`] = typeof value === "string" ? value : JSON.stringify(value);
|
|
7211
|
-
}
|
|
7212
|
-
}
|
|
7213
|
-
function getGeoMetadata(schema) {
|
|
7214
|
-
const geoMetadata = parseJSONStringMetadata(schema, "geo");
|
|
7215
|
-
return geoMetadata;
|
|
7216
|
-
}
|
|
7217
|
-
function setGeoMetadata(schema, geoMetadata) {
|
|
7218
|
-
const stringifiedGeoMetadata = JSON.stringify(geoMetadata);
|
|
7219
|
-
schema.metadata.geo = stringifiedGeoMetadata;
|
|
7220
|
-
}
|
|
7221
|
-
function unpackGeoMetadata(schema) {
|
|
7222
|
-
const geoMetadata = getGeoMetadata(schema);
|
|
7223
|
-
if (!geoMetadata) {
|
|
7224
|
-
return;
|
|
7225
|
-
}
|
|
7226
|
-
const { version, primary_column, columns } = geoMetadata;
|
|
7227
|
-
if (version) {
|
|
7228
|
-
schema.metadata["geo.version"] = version;
|
|
7229
|
-
}
|
|
7230
|
-
if (primary_column) {
|
|
7231
|
-
schema.metadata["geo.primary_column"] = primary_column;
|
|
7232
|
-
}
|
|
7233
|
-
schema.metadata["geo.columns"] = Object.keys(columns || {}).join("");
|
|
7234
|
-
for (const [columnName, columnMetadata] of Object.entries(columns || {})) {
|
|
7235
|
-
const field = schema.fields.find((field2) => field2.name === columnName);
|
|
7236
|
-
if (field) {
|
|
7237
|
-
if (field.name === primary_column) {
|
|
7238
|
-
setFieldMetadata(field, "geo.primary_field", "true");
|
|
7239
|
-
}
|
|
7240
|
-
unpackGeoFieldMetadata(field, columnMetadata);
|
|
7241
|
-
}
|
|
7242
|
-
}
|
|
7243
|
-
}
|
|
7244
|
-
function unpackGeoFieldMetadata(field, columnMetadata) {
|
|
7245
|
-
for (const [key, value] of Object.entries(columnMetadata || {})) {
|
|
7246
|
-
switch (key) {
|
|
7247
|
-
case "geometry_type":
|
|
7248
|
-
setFieldMetadata(field, `geo.${key}`, value.join(","));
|
|
7249
|
-
break;
|
|
7250
|
-
case "bbox":
|
|
7251
|
-
setFieldMetadata(field, `geo.crs.${key}`, JSON.stringify(value));
|
|
7252
|
-
break;
|
|
7253
|
-
case "crs":
|
|
7254
|
-
for (const [crsKey, crsValue] of Object.entries(value || {})) {
|
|
7255
|
-
switch (crsKey) {
|
|
7256
|
-
case "id":
|
|
7257
|
-
const crsId = typeof crsValue === "object" ? (
|
|
7258
|
-
// @ts-ignore
|
|
7259
|
-
`${crsValue == null ? void 0 : crsValue.authority}:${crsValue == null ? void 0 : crsValue.code}`
|
|
7260
|
-
) : JSON.stringify(crsValue);
|
|
7261
|
-
setFieldMetadata(field, `geo.crs.${crsKey}`, crsId);
|
|
7262
|
-
break;
|
|
7263
|
-
default:
|
|
7264
|
-
setFieldMetadata(
|
|
7265
|
-
field,
|
|
7266
|
-
`geo.crs.${crsKey}`,
|
|
7267
|
-
typeof crsValue === "string" ? crsValue : JSON.stringify(crsValue)
|
|
7268
|
-
);
|
|
7269
|
-
break;
|
|
7270
|
-
}
|
|
7271
|
-
}
|
|
7272
|
-
break;
|
|
7273
|
-
case "edges":
|
|
7274
|
-
default:
|
|
7275
|
-
setFieldMetadata(
|
|
7276
|
-
field,
|
|
7277
|
-
`geo.${key}`,
|
|
7278
|
-
typeof value === "string" ? value : JSON.stringify(value)
|
|
7279
|
-
);
|
|
7280
|
-
}
|
|
7281
|
-
}
|
|
7282
|
-
}
|
|
7283
|
-
function setFieldMetadata(field, key, value) {
|
|
7284
|
-
field.metadata = field.metadata || {};
|
|
7285
|
-
field.metadata[key] = value;
|
|
7286
|
-
}
|
|
7287
|
-
|
|
7288
7189
|
// src/lib/parsers/get-parquet-schema.ts
|
|
7190
|
+
var import_gis = require("@loaders.gl/gis");
|
|
7289
7191
|
async function getSchemaFromParquetReader(reader) {
|
|
7290
7192
|
const parquetSchema = await reader.getSchema();
|
|
7291
7193
|
const parquetMetadata = await reader.getFileMetadata();
|
|
7292
7194
|
const schema = convertParquetSchema(parquetSchema, parquetMetadata);
|
|
7293
|
-
unpackGeoMetadata(schema);
|
|
7294
|
-
unpackJSONStringMetadata(schema, "pandas");
|
|
7195
|
+
(0, import_gis.unpackGeoMetadata)(schema);
|
|
7196
|
+
(0, import_gis.unpackJSONStringMetadata)(schema, "pandas");
|
|
7295
7197
|
return schema;
|
|
7296
7198
|
}
|
|
7297
7199
|
|
|
7298
|
-
// src/lib/geo/decode-geo-column.ts
|
|
7299
|
-
var import_schema2 = require("@loaders.gl/schema");
|
|
7300
|
-
var import_wkt = require("@loaders.gl/wkt");
|
|
7301
|
-
function convertWKBTableToGeoJSON(table, schema) {
|
|
7302
|
-
const geoMetadata = getGeoMetadata(schema);
|
|
7303
|
-
const primaryColumn = geoMetadata == null ? void 0 : geoMetadata.primary_column;
|
|
7304
|
-
if (!primaryColumn) {
|
|
7305
|
-
throw new Error("no geometry column");
|
|
7306
|
-
}
|
|
7307
|
-
const columnMetadata = geoMetadata.columns[primaryColumn];
|
|
7308
|
-
const features = [];
|
|
7309
|
-
const length = (0, import_schema2.getTableLength)(table);
|
|
7310
|
-
for (let rowIndex = 0; rowIndex < length; rowIndex++) {
|
|
7311
|
-
const row = (0, import_schema2.getTableRowAsObject)(table, rowIndex);
|
|
7312
|
-
const geometry = parseGeometry(row[primaryColumn], columnMetadata);
|
|
7313
|
-
delete row[primaryColumn];
|
|
7314
|
-
const feature = { type: "Feature", geometry, properties: row };
|
|
7315
|
-
features.push(feature);
|
|
7316
|
-
}
|
|
7317
|
-
return { shape: "geojson-table", schema, type: "FeatureCollection", features };
|
|
7318
|
-
}
|
|
7319
|
-
function parseGeometry(geometry, columnMetadata) {
|
|
7320
|
-
var _a, _b, _c, _d;
|
|
7321
|
-
switch (columnMetadata.encoding) {
|
|
7322
|
-
case "wkt":
|
|
7323
|
-
return ((_b = (_a = import_wkt.WKTLoader).parseTextSync) == null ? void 0 : _b.call(_a, geometry)) || null;
|
|
7324
|
-
case "wkb":
|
|
7325
|
-
default:
|
|
7326
|
-
const arrayBuffer = ArrayBuffer.isView(geometry) ? geometry.buffer.slice(geometry.byteOffset, geometry.byteOffset + geometry.byteLength) : geometry;
|
|
7327
|
-
const geojson = (_d = (_c = import_wkt.WKBLoader).parseSync) == null ? void 0 : _d.call(_c, arrayBuffer, {
|
|
7328
|
-
wkb: { shape: "geometry" }
|
|
7329
|
-
});
|
|
7330
|
-
return geojson;
|
|
7331
|
-
}
|
|
7332
|
-
}
|
|
7333
|
-
|
|
7334
7200
|
// src/lib/parsers/parse-parquet-to-rows.ts
|
|
7335
7201
|
async function parseParquet(arrayBuffer, options) {
|
|
7336
7202
|
var _a, _b;
|
|
@@ -7359,7 +7225,7 @@ async function parseParquet(arrayBuffer, options) {
|
|
|
7359
7225
|
return objectRowTable;
|
|
7360
7226
|
case "geojson-table":
|
|
7361
7227
|
try {
|
|
7362
|
-
return convertWKBTableToGeoJSON(objectRowTable, schema);
|
|
7228
|
+
return (0, import_gis2.convertWKBTableToGeoJSON)(objectRowTable, schema, [import_wkt.WKTLoader, import_wkt.WKBLoader]);
|
|
7363
7229
|
} catch (error) {
|
|
7364
7230
|
return objectRowTable;
|
|
7365
7231
|
}
|
|
@@ -7425,9 +7291,10 @@ var ParquetWriter = {
|
|
|
7425
7291
|
version: VERSION2,
|
|
7426
7292
|
extensions: ["parquet"],
|
|
7427
7293
|
mimeTypes: ["application/octet-stream"],
|
|
7428
|
-
encodeSync,
|
|
7429
7294
|
binary: true,
|
|
7430
|
-
options: {}
|
|
7295
|
+
options: {},
|
|
7296
|
+
encode: async (data, options) => encodeSync(data, options),
|
|
7297
|
+
encodeSync
|
|
7431
7298
|
};
|
|
7432
7299
|
function encodeSync(data, options) {
|
|
7433
7300
|
return new ArrayBuffer(0);
|
|
@@ -7810,68 +7677,6 @@ function encodeFooter(schema, rowCount, rowGroups, userMetadata) {
|
|
|
7810
7677
|
return footerEncoded;
|
|
7811
7678
|
}
|
|
7812
7679
|
|
|
7813
|
-
// src/lib/geo/geoparquet-metadata-schema.ts
|
|
7814
|
-
var GEOPARQUET_METADATA_JSON_SCHEMA = {
|
|
7815
|
-
$schema: "http://json-schema.org/draft-07/schema#",
|
|
7816
|
-
title: "GeoParquet",
|
|
7817
|
-
description: "Parquet metadata included in the geo field.",
|
|
7818
|
-
type: "object",
|
|
7819
|
-
required: ["version", "primary_column", "columns"],
|
|
7820
|
-
properties: {
|
|
7821
|
-
version: { type: "string", const: "1.0.0-beta.1" },
|
|
7822
|
-
primary_column: { type: "string", minLength: 1 },
|
|
7823
|
-
columns: {
|
|
7824
|
-
type: "object",
|
|
7825
|
-
minProperties: 1,
|
|
7826
|
-
patternProperties: {
|
|
7827
|
-
".+": {
|
|
7828
|
-
type: "object",
|
|
7829
|
-
required: ["encoding", "geometry_types"],
|
|
7830
|
-
properties: {
|
|
7831
|
-
encoding: { type: "string", const: "WKB" },
|
|
7832
|
-
geometry_types: {
|
|
7833
|
-
type: "array",
|
|
7834
|
-
uniqueItems: true,
|
|
7835
|
-
items: {
|
|
7836
|
-
type: "string",
|
|
7837
|
-
pattern: "^(GeometryCollection|(Multi)?(Point|LineString|Polygon))( Z)?$"
|
|
7838
|
-
}
|
|
7839
|
-
},
|
|
7840
|
-
crs: {
|
|
7841
|
-
oneOf: [
|
|
7842
|
-
{
|
|
7843
|
-
$ref: "https://proj.org/schemas/v0.5/projjson.schema.json"
|
|
7844
|
-
},
|
|
7845
|
-
{ type: "null" }
|
|
7846
|
-
]
|
|
7847
|
-
},
|
|
7848
|
-
edges: { type: "string", enum: ["planar", "spherical"] },
|
|
7849
|
-
orientation: { type: "string", const: "counterclockwise" },
|
|
7850
|
-
bbox: {
|
|
7851
|
-
type: "array",
|
|
7852
|
-
items: { type: "number" },
|
|
7853
|
-
oneOf: [
|
|
7854
|
-
{
|
|
7855
|
-
description: "2D bbox consisting of (xmin, ymin, xmax, ymax)",
|
|
7856
|
-
minItems: 4,
|
|
7857
|
-
maxItems: 4
|
|
7858
|
-
},
|
|
7859
|
-
{
|
|
7860
|
-
description: "3D bbox consisting of (xmin, ymin, zmin, xmax, ymax, zmax)",
|
|
7861
|
-
minItems: 6,
|
|
7862
|
-
maxItems: 6
|
|
7863
|
-
}
|
|
7864
|
-
]
|
|
7865
|
-
},
|
|
7866
|
-
epoch: { type: "number" }
|
|
7867
|
-
}
|
|
7868
|
-
}
|
|
7869
|
-
},
|
|
7870
|
-
additionalProperties: false
|
|
7871
|
-
}
|
|
7872
|
-
}
|
|
7873
|
-
};
|
|
7874
|
-
|
|
7875
7680
|
// src/index.ts
|
|
7876
7681
|
var ParquetLoader2 = {
|
|
7877
7682
|
...ParquetLoader,
|
package/dist/index.d.ts
CHANGED
|
@@ -13,8 +13,5 @@ export { ParquetSchema } from './parquetjs/schema/schema';
|
|
|
13
13
|
export { ParquetReader } from './parquetjs/parser/parquet-reader';
|
|
14
14
|
export { ParquetEncoder } from './parquetjs/encoder/parquet-encoder';
|
|
15
15
|
export { convertParquetSchema, convertParquetSchema as convertParquetToArrowSchema } from './lib/arrow/convert-schema-from-parquet';
|
|
16
|
-
export { GEOPARQUET_METADATA_JSON_SCHEMA } from './lib/geo/geoparquet-metadata-schema';
|
|
17
|
-
export type { GeoMetadata } from './lib/geo/decode-geo-metadata';
|
|
18
|
-
export { getGeoMetadata, setGeoMetadata, unpackGeoMetadata } from './lib/geo/decode-geo-metadata';
|
|
19
16
|
export { BufferPolyfill, installBufferPolyfill } from './buffer-polyfill';
|
|
20
17
|
//# sourceMappingURL=index.d.ts.map
|
package/dist/index.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAGA,OAAO,EAAC,MAAM,EAAC,MAAM,2CAA2C,CAAC;AAEjE,OAAO,KAAK,EAAC,gBAAgB,EAAC,MAAM,0BAA0B,CAAC;AAC/D,OAAO,KAAK,EACV,cAAc,EACd,mBAAmB,EACnB,aAAa,EACb,kBAAkB,EAClB,YAAY,EACZ,iBAAiB,EAClB,MAAM,oBAAoB,CAAC;AAK5B,OAAO,EACL,aAAa,IAAI,mBAAmB,EAEpC,oBAAoB,EACrB,MAAM,kBAAkB,CAAC;AAW1B,OAAO,EAAC,mBAAmB,EAAC,CAAC;AAG7B,6BAA6B;AAC7B,eAAO,MAAM,aAAa,EAAE,gBAAgB,CAC1C,cAAc,GAAG,YAAY,EAC7B,mBAAmB,GAAG,iBAAiB,EACvC,oBAAoB,CAMrB,CAAC;AAEF,6BAA6B;AAC7B,eAAO,MAAM,qBAAqB,EAAE,gBAAgB,CAClD,aAAa,EACb,kBAAkB,EAClB,oBAAoB,CAMrB,CAAC;AAcF,OAAO,EAAC,aAAa,IAAI,cAAc,EAAC,MAAM,kBAAkB,CAAC;AAKjE,OAAO,EAAC,mBAAmB,EAAC,MAAM,yBAAyB,CAAC;AAE5D,OAAO,EAAC,aAAa,EAAC,MAAM,2BAA2B,CAAC;AACxD,OAAO,EAAC,aAAa,EAAC,MAAM,mCAAmC,CAAC;AAChE,OAAO,EAAC,cAAc,EAAC,MAAM,qCAAqC,CAAC;AAEnE,OAAO,EACL,oBAAoB,EACpB,oBAAoB,IAAI,2BAA2B,EACpD,MAAM,yCAAyC,CAAC;
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAGA,OAAO,EAAC,MAAM,EAAC,MAAM,2CAA2C,CAAC;AAEjE,OAAO,KAAK,EAAC,gBAAgB,EAAC,MAAM,0BAA0B,CAAC;AAC/D,OAAO,KAAK,EACV,cAAc,EACd,mBAAmB,EACnB,aAAa,EACb,kBAAkB,EAClB,YAAY,EACZ,iBAAiB,EAClB,MAAM,oBAAoB,CAAC;AAK5B,OAAO,EACL,aAAa,IAAI,mBAAmB,EAEpC,oBAAoB,EACrB,MAAM,kBAAkB,CAAC;AAW1B,OAAO,EAAC,mBAAmB,EAAC,CAAC;AAG7B,6BAA6B;AAC7B,eAAO,MAAM,aAAa,EAAE,gBAAgB,CAC1C,cAAc,GAAG,YAAY,EAC7B,mBAAmB,GAAG,iBAAiB,EACvC,oBAAoB,CAMrB,CAAC;AAEF,6BAA6B;AAC7B,eAAO,MAAM,qBAAqB,EAAE,gBAAgB,CAClD,aAAa,EACb,kBAAkB,EAClB,oBAAoB,CAMrB,CAAC;AAcF,OAAO,EAAC,aAAa,IAAI,cAAc,EAAC,MAAM,kBAAkB,CAAC;AAKjE,OAAO,EAAC,mBAAmB,EAAC,MAAM,yBAAyB,CAAC;AAE5D,OAAO,EAAC,aAAa,EAAC,MAAM,2BAA2B,CAAC;AACxD,OAAO,EAAC,aAAa,EAAC,MAAM,mCAAmC,CAAC;AAChE,OAAO,EAAC,cAAc,EAAC,MAAM,qCAAqC,CAAC;AAEnE,OAAO,EACL,oBAAoB,EACpB,oBAAoB,IAAI,2BAA2B,EACpD,MAAM,yCAAyC,CAAC;AAGjD,OAAO,EAAC,cAAc,EAAE,qBAAqB,EAAC,MAAM,mBAAmB,CAAC"}
|
package/dist/index.js
CHANGED
|
@@ -19,7 +19,5 @@ export { ParquetSchema } from "./parquetjs/schema/schema.js";
|
|
|
19
19
|
export { ParquetReader } from "./parquetjs/parser/parquet-reader.js";
|
|
20
20
|
export { ParquetEncoder } from "./parquetjs/encoder/parquet-encoder.js";
|
|
21
21
|
export { convertParquetSchema, convertParquetSchema as convertParquetToArrowSchema } from "./lib/arrow/convert-schema-from-parquet.js";
|
|
22
|
-
export { GEOPARQUET_METADATA_JSON_SCHEMA } from "./lib/geo/geoparquet-metadata-schema.js";
|
|
23
|
-
export { getGeoMetadata, setGeoMetadata, unpackGeoMetadata } from "./lib/geo/decode-geo-metadata.js";
|
|
24
22
|
export { BufferPolyfill, installBufferPolyfill } from "./buffer-polyfill/index.js";
|
|
25
23
|
//# sourceMappingURL=index.js.map
|
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","names":["Buffer","ParquetLoader","ParquetWorkerLoader","ParquetColumnarWorkerLoader","parseParquet","parseParquetFileInBatches","parseParquetInColumns","parseParquetFileInColumnarBatches","parse","parseFileInBatches","ParquetColumnarLoader","ParquetWriter","_ParquetWriter","preloadCompressions","ParquetSchema","ParquetReader","ParquetEncoder","convertParquetSchema","convertParquetToArrowSchema","
|
|
1
|
+
{"version":3,"file":"index.js","names":["Buffer","ParquetLoader","ParquetWorkerLoader","ParquetColumnarWorkerLoader","parseParquet","parseParquetFileInBatches","parseParquetInColumns","parseParquetFileInColumnarBatches","parse","parseFileInBatches","ParquetColumnarLoader","ParquetWriter","_ParquetWriter","preloadCompressions","ParquetSchema","ParquetReader","ParquetEncoder","convertParquetSchema","convertParquetToArrowSchema","BufferPolyfill","installBufferPolyfill"],"sources":["../src/index.ts"],"sourcesContent":["// loaders.gl, MIT license\n// Copyright (c) vis.gl contributors\n\nexport {Buffer} from './buffer-polyfill/install-buffer-polyfill';\n\nimport type {LoaderWithParser} from '@loaders.gl/loader-utils';\nimport type {\n ObjectRowTable,\n ObjectRowTableBatch,\n ColumnarTable,\n ColumnarTableBatch,\n GeoJSONTable,\n GeoJSONTableBatch\n} from '@loaders.gl/schema';\n// import type * as arrow from 'apache-arrow';\n\n// ParquetLoader\n\nimport {\n ParquetLoader as ParquetWorkerLoader,\n ParquetLoader as ParquetColumnarWorkerLoader,\n ParquetLoaderOptions\n} from './parquet-loader';\nimport {parseParquet, parseParquetFileInBatches} from './lib/parsers/parse-parquet-to-rows';\nimport {\n parseParquetInColumns,\n parseParquetFileInColumnarBatches\n} from './lib/parsers/parse-parquet-to-columns';\n\n// import type {ParquetWasmLoaderOptions} from './lib/wasm/parse-parquet-wasm';\n// import {parseParquetWasm} from './lib/wasm/parse-parquet-wasm';\n// import {ParquetWasmLoader as ParquetWasmWorkerLoader} from './parquet-wasm-loader';\n\nexport {ParquetWorkerLoader};\n// export {ParquetWasmWorkerLoader};\n\n/** ParquetJS table loader */\nexport const ParquetLoader: LoaderWithParser<\n ObjectRowTable | GeoJSONTable,\n ObjectRowTableBatch | GeoJSONTableBatch,\n ParquetLoaderOptions\n> = {\n ...ParquetWorkerLoader,\n parse: parseParquet,\n // @ts-expect-error\n parseFileInBatches: parseParquetFileInBatches\n};\n\n/** ParquetJS table loader */\nexport const ParquetColumnarLoader: LoaderWithParser<\n ColumnarTable,\n ColumnarTableBatch,\n ParquetLoaderOptions\n> = {\n ...ParquetColumnarWorkerLoader,\n parse: parseParquetInColumns,\n // @ts-expect-error\n parseFileInBatches: parseParquetFileInColumnarBatches\n};\n\n// export const ParquetWasmLoader: LoaderWithParser<\n// arrow.Table,\n// never,\n// ParquetWasmLoaderOptions\n// > = {\n// ...ParquetWasmWorkerLoader,\n// // @ts-expect-error Getting strange errors in wasm\n// parse: () => {} // parseParquetWasm\n// };\n\n// ParquetWriter\n\nexport {ParquetWriter as _ParquetWriter} from './parquet-writer';\n// export {ParquetWasmWriter} from './parquet-wasm-writer';\n\n// EXPERIMENTAL - expose the internal parquetjs API\n\nexport {preloadCompressions} from './parquetjs/compression';\n\nexport {ParquetSchema} from './parquetjs/schema/schema';\nexport {ParquetReader} from './parquetjs/parser/parquet-reader';\nexport {ParquetEncoder} from './parquetjs/encoder/parquet-encoder';\n\nexport {\n convertParquetSchema,\n convertParquetSchema as convertParquetToArrowSchema\n} from './lib/arrow/convert-schema-from-parquet';\n\n// Experimental\nexport {BufferPolyfill, installBufferPolyfill} from './buffer-polyfill';\n"],"mappings":"SAGQA,MAAM;AAAA,SAgBZC,aAAa,IAAIC,mBAAmB,EACpCD,aAAa,IAAIE,2BAA2B;AAAA,SAGtCC,YAAY,EAAEC,yBAAyB;AAAA,SAE7CC,qBAAqB,EACrBC,iCAAiC;AAOnC,SAAQL,mBAAmB;AAI3B,OAAO,MAAMD,aAIZ,GAAG;EACF,GAAGC,mBAAmB;EACtBM,KAAK,EAAEJ,YAAY;EAEnBK,kBAAkB,EAAEJ;AACtB,CAAC;AAGD,OAAO,MAAMK,qBAIZ,GAAG;EACF,GAAGP,2BAA2B;EAC9BK,KAAK,EAAEF,qBAAqB;EAE5BG,kBAAkB,EAAEF;AACtB,CAAC;AAAC,SAcMI,aAAa,IAAIC,cAAc;AAAA,SAK/BC,mBAAmB;AAAA,SAEnBC,aAAa;AAAA,SACbC,aAAa;AAAA,SACbC,cAAc;AAAA,SAGpBC,oBAAoB,EACpBA,oBAAoB,IAAIC,2BAA2B;AAAA,SAI7CC,cAAc,EAAEC,qBAAqB"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { convertParquetSchema } from "../arrow/convert-schema-from-parquet.js";
|
|
2
|
-
import { unpackGeoMetadata, unpackJSONStringMetadata } from
|
|
2
|
+
import { unpackGeoMetadata, unpackJSONStringMetadata } from '@loaders.gl/gis';
|
|
3
3
|
export async function getSchemaFromParquetReader(reader) {
|
|
4
4
|
const parquetSchema = await reader.getSchema();
|
|
5
5
|
const parquetMetadata = await reader.getFileMetadata();
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"get-parquet-schema.js","names":["convertParquetSchema","unpackGeoMetadata","unpackJSONStringMetadata","getSchemaFromParquetReader","reader","parquetSchema","getSchema","parquetMetadata","getFileMetadata","schema"],"sources":["../../../src/lib/parsers/get-parquet-schema.ts"],"sourcesContent":["// loaders.gl\nimport {Schema} from '@loaders.gl/schema';\nimport {ParquetReader} from '../../parquetjs/parser/parquet-reader';\nimport {convertParquetSchema} from '../arrow/convert-schema-from-parquet';\nimport {unpackGeoMetadata, unpackJSONStringMetadata} from '
|
|
1
|
+
{"version":3,"file":"get-parquet-schema.js","names":["convertParquetSchema","unpackGeoMetadata","unpackJSONStringMetadata","getSchemaFromParquetReader","reader","parquetSchema","getSchema","parquetMetadata","getFileMetadata","schema"],"sources":["../../../src/lib/parsers/get-parquet-schema.ts"],"sourcesContent":["// loaders.gl\nimport {Schema} from '@loaders.gl/schema';\nimport {ParquetReader} from '../../parquetjs/parser/parquet-reader';\nimport {convertParquetSchema} from '../arrow/convert-schema-from-parquet';\nimport {unpackGeoMetadata, unpackJSONStringMetadata} from '@loaders.gl/gis';\n\nexport async function getSchemaFromParquetReader(reader: ParquetReader): Promise<Schema> {\n const parquetSchema = await reader.getSchema();\n const parquetMetadata = await reader.getFileMetadata();\n const schema = convertParquetSchema(parquetSchema, parquetMetadata);\n unpackGeoMetadata(schema);\n unpackJSONStringMetadata(schema, 'pandas');\n return schema;\n}\n"],"mappings":"SAGQA,oBAAoB;AAC5B,SAAQC,iBAAiB,EAAEC,wBAAwB,QAAO,iBAAiB;AAE3E,OAAO,eAAeC,0BAA0BA,CAACC,MAAqB,EAAmB;EACvF,MAAMC,aAAa,GAAG,MAAMD,MAAM,CAACE,SAAS,CAAC,CAAC;EAC9C,MAAMC,eAAe,GAAG,MAAMH,MAAM,CAACI,eAAe,CAAC,CAAC;EACtD,MAAMC,MAAM,GAAGT,oBAAoB,CAACK,aAAa,EAAEE,eAAe,CAAC;EACnEN,iBAAiB,CAACQ,MAAM,CAAC;EACzBP,wBAAwB,CAACO,MAAM,EAAE,QAAQ,CAAC;EAC1C,OAAOA,MAAM;AACf"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parse-parquet-to-rows.d.ts","sourceRoot":"","sources":["../../../src/lib/parsers/parse-parquet-to-rows.ts"],"names":[],"mappings":"AAGA,OAAO,EAAC,YAAY,EAAE,cAAc,EAAE,mBAAmB,EAAC,MAAM,oBAAoB,CAAC;
|
|
1
|
+
{"version":3,"file":"parse-parquet-to-rows.d.ts","sourceRoot":"","sources":["../../../src/lib/parsers/parse-parquet-to-rows.ts"],"names":[],"mappings":"AAGA,OAAO,EAAC,YAAY,EAAE,cAAc,EAAE,mBAAmB,EAAC,MAAM,oBAAoB,CAAC;AAIrF,OAAO,KAAK,EAAC,oBAAoB,EAAC,MAAM,sBAAsB,CAAC;AAE/D,OAAO,EAAC,aAAa,EAAC,MAAM,uCAAuC,CAAC;AAIpE,wBAAsB,YAAY,CAChC,WAAW,EAAE,WAAW,EACxB,OAAO,CAAC,EAAE,oBAAoB,GAC7B,OAAO,CAAC,cAAc,GAAG,YAAY,CAAC,CAyCxC;AAED,wBAAuB,yBAAyB,CAC9C,MAAM,EAAE,aAAa,EACrB,OAAO,CAAC,EAAE,oBAAoB,GAC7B,aAAa,CAAC,mBAAmB,CAAC,CAYpC"}
|
|
@@ -1,8 +1,9 @@
|
|
|
1
1
|
import { BlobFile } from '@loaders.gl/loader-utils';
|
|
2
|
+
import { convertWKBTableToGeoJSON } from '@loaders.gl/gis';
|
|
3
|
+
import { WKTLoader, WKBLoader } from '@loaders.gl/wkt';
|
|
2
4
|
import { ParquetReader } from "../../parquetjs/parser/parquet-reader.js";
|
|
3
5
|
import { getSchemaFromParquetReader } from "./get-parquet-schema.js";
|
|
4
6
|
import { installBufferPolyfill } from "../../buffer-polyfill/index.js";
|
|
5
|
-
import { convertWKBTableToGeoJSON } from "../geo/decode-geo-column.js";
|
|
6
7
|
export async function parseParquet(arrayBuffer, options) {
|
|
7
8
|
var _options$parquet, _options$parquet2;
|
|
8
9
|
installBufferPolyfill();
|
|
@@ -30,7 +31,7 @@ export async function parseParquet(arrayBuffer, options) {
|
|
|
30
31
|
return objectRowTable;
|
|
31
32
|
case 'geojson-table':
|
|
32
33
|
try {
|
|
33
|
-
return convertWKBTableToGeoJSON(objectRowTable, schema);
|
|
34
|
+
return convertWKBTableToGeoJSON(objectRowTable, schema, [WKTLoader, WKBLoader]);
|
|
34
35
|
} catch (error) {
|
|
35
36
|
return objectRowTable;
|
|
36
37
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parse-parquet-to-rows.js","names":["BlobFile","ParquetReader","getSchemaFromParquetReader","installBufferPolyfill","
|
|
1
|
+
{"version":3,"file":"parse-parquet-to-rows.js","names":["BlobFile","convertWKBTableToGeoJSON","WKTLoader","WKBLoader","ParquetReader","getSchemaFromParquetReader","installBufferPolyfill","parseParquet","arrayBuffer","options","_options$parquet","_options$parquet2","blob","Blob","file","reader","preserveBinary","parquet","schema","rows","rowBatches","rowBatchIterator","rowBatch","row","push","objectRowTable","shape","data","error","Error","parseParquetFileInBatches","batchType","length"],"sources":["../../../src/lib/parsers/parse-parquet-to-rows.ts"],"sourcesContent":["// import type {LoaderWithParser, Loader, LoaderOptions} from '@loaders.gl/loader-utils';\n// import {ColumnarTableBatch} from '@loaders.gl/schema';\nimport {BlobFile} from '@loaders.gl/loader-utils';\nimport {GeoJSONTable, ObjectRowTable, ObjectRowTableBatch} from '@loaders.gl/schema';\nimport {convertWKBTableToGeoJSON} from '@loaders.gl/gis';\nimport {WKTLoader, WKBLoader} from '@loaders.gl/wkt';\n\nimport type {ParquetLoaderOptions} from '../../parquet-loader';\nimport type {ParquetRow} from '../../parquetjs/schema/declare';\nimport {ParquetReader} from '../../parquetjs/parser/parquet-reader';\nimport {getSchemaFromParquetReader} from './get-parquet-schema';\nimport {installBufferPolyfill} from '../../buffer-polyfill';\n\nexport async function parseParquet(\n arrayBuffer: ArrayBuffer,\n options?: ParquetLoaderOptions\n): Promise<ObjectRowTable | GeoJSONTable> {\n installBufferPolyfill();\n\n const blob = new Blob([arrayBuffer]);\n const file = new BlobFile(blob);\n const reader = new ParquetReader(file, {\n preserveBinary: options?.parquet?.preserveBinary\n });\n\n const schema = await getSchemaFromParquetReader(reader);\n\n const rows: ParquetRow[] = [];\n\n const rowBatches = reader.rowBatchIterator(options?.parquet);\n for await (const rowBatch of rowBatches) {\n // we have only one input batch so return\n for (const row of rowBatch) {\n rows.push(row);\n }\n }\n const objectRowTable: ObjectRowTable = {\n shape: 'object-row-table',\n schema,\n data: rows\n };\n\n const shape = options?.parquet?.shape;\n switch (shape) {\n case 'object-row-table':\n return objectRowTable;\n\n case 'geojson-table':\n try {\n return convertWKBTableToGeoJSON(objectRowTable, schema, [WKTLoader, WKBLoader]);\n } catch (error) {\n return objectRowTable;\n }\n\n default:\n throw new Error(shape);\n }\n}\n\nexport async function* parseParquetFileInBatches(\n reader: ParquetReader,\n options?: ParquetLoaderOptions\n): AsyncIterable<ObjectRowTableBatch> {\n const schema = await getSchemaFromParquetReader(reader);\n const rowBatches = reader.rowBatchIterator(options?.parquet);\n for await (const rows of rowBatches) {\n yield {\n batchType: 'data',\n shape: 'object-row-table',\n schema,\n data: rows,\n length: rows.length\n };\n }\n}\n"],"mappings":"AAEA,SAAQA,QAAQ,QAAO,0BAA0B;AAEjD,SAAQC,wBAAwB,QAAO,iBAAiB;AACxD,SAAQC,SAAS,EAAEC,SAAS,QAAO,iBAAiB;AAAC,SAI7CC,aAAa;AAAA,SACbC,0BAA0B;AAAA,SAC1BC,qBAAqB;AAE7B,OAAO,eAAeC,YAAYA,CAChCC,WAAwB,EACxBC,OAA8B,EACU;EAAA,IAAAC,gBAAA,EAAAC,iBAAA;EACxCL,qBAAqB,CAAC,CAAC;EAEvB,MAAMM,IAAI,GAAG,IAAIC,IAAI,CAAC,CAACL,WAAW,CAAC,CAAC;EACpC,MAAMM,IAAI,GAAG,IAAId,QAAQ,CAACY,IAAI,CAAC;EAC/B,MAAMG,MAAM,GAAG,IAAIX,aAAa,CAACU,IAAI,EAAE;IACrCE,cAAc,EAAEP,OAAO,aAAPA,OAAO,wBAAAC,gBAAA,GAAPD,OAAO,CAAEQ,OAAO,cAAAP,gBAAA,uBAAhBA,gBAAA,CAAkBM;EACpC,CAAC,CAAC;EAEF,MAAME,MAAM,GAAG,MAAMb,0BAA0B,CAACU,MAAM,CAAC;EAEvD,MAAMI,IAAkB,GAAG,EAAE;EAE7B,MAAMC,UAAU,GAAGL,MAAM,CAACM,gBAAgB,CAACZ,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEQ,OAAO,CAAC;EAC5D,WAAW,MAAMK,QAAQ,IAAIF,UAAU,EAAE;IAEvC,KAAK,MAAMG,GAAG,IAAID,QAAQ,EAAE;MAC1BH,IAAI,CAACK,IAAI,CAACD,GAAG,CAAC;IAChB;EACF;EACA,MAAME,cAA8B,GAAG;IACrCC,KAAK,EAAE,kBAAkB;IACzBR,MAAM;IACNS,IAAI,EAAER;EACR,CAAC;EAED,MAAMO,KAAK,GAAGjB,OAAO,aAAPA,OAAO,wBAAAE,iBAAA,GAAPF,OAAO,CAAEQ,OAAO,cAAAN,iBAAA,uBAAhBA,iBAAA,CAAkBe,KAAK;EACrC,QAAQA,KAAK;IACX,KAAK,kBAAkB;MACrB,OAAOD,cAAc;IAEvB,KAAK,eAAe;MAClB,IAAI;QACF,OAAOxB,wBAAwB,CAACwB,cAAc,EAAEP,MAAM,EAAE,CAAChB,SAAS,EAAEC,SAAS,CAAC,CAAC;MACjF,CAAC,CAAC,OAAOyB,KAAK,EAAE;QACd,OAAOH,cAAc;MACvB;IAEF;MACE,MAAM,IAAII,KAAK,CAACH,KAAK,CAAC;EAC1B;AACF;AAEA,OAAO,gBAAgBI,yBAAyBA,CAC9Cf,MAAqB,EACrBN,OAA8B,EACM;EACpC,MAAMS,MAAM,GAAG,MAAMb,0BAA0B,CAACU,MAAM,CAAC;EACvD,MAAMK,UAAU,GAAGL,MAAM,CAACM,gBAAgB,CAACZ,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEQ,OAAO,CAAC;EAC5D,WAAW,MAAME,IAAI,IAAIC,UAAU,EAAE;IACnC,MAAM;MACJW,SAAS,EAAE,MAAM;MACjBL,KAAK,EAAE,kBAAkB;MACzBR,MAAM;MACNS,IAAI,EAAER,IAAI;MACVa,MAAM,EAAEb,IAAI,CAACa;IACf,CAAC;EACH;AACF"}
|
|
@@ -1,21 +1,20 @@
|
|
|
1
|
-
import type { Table } from 'apache-arrow';
|
|
2
1
|
import type { WriterOptions } from '@loaders.gl/loader-utils';
|
|
2
|
+
import * as arrow from 'apache-arrow';
|
|
3
3
|
export type ParquetWriterOptions = WriterOptions & {
|
|
4
4
|
parquet?: {
|
|
5
5
|
wasmUrl?: string;
|
|
6
6
|
};
|
|
7
7
|
};
|
|
8
8
|
/**
|
|
9
|
-
* Encode Arrow Table to Parquet buffer
|
|
9
|
+
* Encode Arrow arrow.Table to Parquet buffer
|
|
10
10
|
*/
|
|
11
|
-
export declare function encode(table: Table, options?: ParquetWriterOptions): Promise<ArrayBuffer>;
|
|
11
|
+
export declare function encode(table: arrow.Table, options?: ParquetWriterOptions): Promise<ArrayBuffer>;
|
|
12
12
|
/**
|
|
13
|
-
* Serialize a
|
|
14
|
-
* wrapper for {@link RecordBatchStreamWriter} and {@link RecordBatchFileWriter}.
|
|
13
|
+
* Serialize a table to the IPC format. This function is a convenience
|
|
15
14
|
* Opposite of {@link tableFromIPC}.
|
|
16
15
|
*
|
|
17
|
-
* @param table The Table to serialize.
|
|
18
|
-
* @param type Whether to serialize the Table as a file or a stream.
|
|
16
|
+
* @param table The arrow.Table to serialize.
|
|
17
|
+
* @param type Whether to serialize the arrow.Table as a file or a stream.
|
|
19
18
|
*/
|
|
20
|
-
export declare function tableToIPC(table: Table): Uint8Array;
|
|
19
|
+
export declare function tableToIPC(table: arrow.Table): Uint8Array;
|
|
21
20
|
//# sourceMappingURL=encode-parquet-wasm.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"encode-parquet-wasm.d.ts","sourceRoot":"","sources":["../../../src/lib/wasm/encode-parquet-wasm.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAC,
|
|
1
|
+
{"version":3,"file":"encode-parquet-wasm.d.ts","sourceRoot":"","sources":["../../../src/lib/wasm/encode-parquet-wasm.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAC,aAAa,EAAC,MAAM,0BAA0B,CAAC;AAE5D,OAAO,KAAK,KAAK,MAAM,cAAc,CAAC;AAGtC,MAAM,MAAM,oBAAoB,GAAG,aAAa,GAAG;IACjD,OAAO,CAAC,EAAE;QACR,OAAO,CAAC,EAAE,MAAM,CAAC;KAClB,CAAC;CACH,CAAC;AAEF;;GAEG;AACH,wBAAsB,MAAM,CAC1B,KAAK,EAAE,KAAK,CAAC,KAAK,EAClB,OAAO,CAAC,EAAE,oBAAoB,GAC7B,OAAO,CAAC,WAAW,CAAC,CAYtB;AAED;;;;;;GAMG;AACH,wBAAgB,UAAU,CAAC,KAAK,EAAE,KAAK,CAAC,KAAK,GAAG,UAAU,CAEzD"}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import
|
|
1
|
+
import * as arrow from 'apache-arrow';
|
|
2
2
|
import { loadWasm } from "./load-wasm/index.js";
|
|
3
3
|
export async function encode(table, options) {
|
|
4
4
|
var _options$parquet;
|
|
@@ -10,6 +10,6 @@ export async function encode(table, options) {
|
|
|
10
10
|
return parquetBytes.buffer.slice(parquetBytes.byteOffset, parquetBytes.byteLength + parquetBytes.byteOffset);
|
|
11
11
|
}
|
|
12
12
|
export function tableToIPC(table) {
|
|
13
|
-
return RecordBatchStreamWriter.writeAll(table).toUint8Array(true);
|
|
13
|
+
return arrow.RecordBatchStreamWriter.writeAll(table).toUint8Array(true);
|
|
14
14
|
}
|
|
15
15
|
//# sourceMappingURL=encode-parquet-wasm.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"encode-parquet-wasm.js","names":["
|
|
1
|
+
{"version":3,"file":"encode-parquet-wasm.js","names":["arrow","loadWasm","encode","table","options","_options$parquet","wasmUrl","parquet","wasm","arrowIPCBytes","tableToIPC","writerProperties","WriterPropertiesBuilder","build","parquetBytes","writeParquet","buffer","slice","byteOffset","byteLength","RecordBatchStreamWriter","writeAll","toUint8Array"],"sources":["../../../src/lib/wasm/encode-parquet-wasm.ts"],"sourcesContent":["import type {WriterOptions} from '@loaders.gl/loader-utils';\n\nimport * as arrow from 'apache-arrow';\nimport {loadWasm} from './load-wasm';\n\nexport type ParquetWriterOptions = WriterOptions & {\n parquet?: {\n wasmUrl?: string;\n };\n};\n\n/**\n * Encode Arrow arrow.Table to Parquet buffer\n */\nexport async function encode(\n table: arrow.Table,\n options?: ParquetWriterOptions\n): Promise<ArrayBuffer> {\n const wasmUrl = options?.parquet?.wasmUrl;\n const wasm = await loadWasm(wasmUrl);\n\n const arrowIPCBytes = tableToIPC(table);\n // TODO: provide options for how to write table.\n const writerProperties = new wasm.WriterPropertiesBuilder().build();\n const parquetBytes = wasm.writeParquet(arrowIPCBytes, writerProperties);\n return parquetBytes.buffer.slice(\n parquetBytes.byteOffset,\n parquetBytes.byteLength + parquetBytes.byteOffset\n );\n}\n\n/**\n * Serialize a table to the IPC format. This function is a convenience\n * Opposite of {@link tableFromIPC}.\n *\n * @param table The arrow.Table to serialize.\n * @param type Whether to serialize the arrow.Table as a file or a stream.\n */\nexport function tableToIPC(table: arrow.Table): Uint8Array {\n return arrow.RecordBatchStreamWriter.writeAll(table).toUint8Array(true);\n}\n"],"mappings":"AAEA,OAAO,KAAKA,KAAK,MAAM,cAAc;AAAC,SAC9BC,QAAQ;AAWhB,OAAO,eAAeC,MAAMA,CAC1BC,KAAkB,EAClBC,OAA8B,EACR;EAAA,IAAAC,gBAAA;EACtB,MAAMC,OAAO,GAAGF,OAAO,aAAPA,OAAO,wBAAAC,gBAAA,GAAPD,OAAO,CAAEG,OAAO,cAAAF,gBAAA,uBAAhBA,gBAAA,CAAkBC,OAAO;EACzC,MAAME,IAAI,GAAG,MAAMP,QAAQ,CAACK,OAAO,CAAC;EAEpC,MAAMG,aAAa,GAAGC,UAAU,CAACP,KAAK,CAAC;EAEvC,MAAMQ,gBAAgB,GAAG,IAAIH,IAAI,CAACI,uBAAuB,CAAC,CAAC,CAACC,KAAK,CAAC,CAAC;EACnE,MAAMC,YAAY,GAAGN,IAAI,CAACO,YAAY,CAACN,aAAa,EAAEE,gBAAgB,CAAC;EACvE,OAAOG,YAAY,CAACE,MAAM,CAACC,KAAK,CAC9BH,YAAY,CAACI,UAAU,EACvBJ,YAAY,CAACK,UAAU,GAAGL,YAAY,CAACI,UACzC,CAAC;AACH;AASA,OAAO,SAASR,UAAUA,CAACP,KAAkB,EAAc;EACzD,OAAOH,KAAK,CAACoB,uBAAuB,CAACC,QAAQ,CAAClB,KAAK,CAAC,CAACmB,YAAY,CAAC,IAAI,CAAC;AACzE"}
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
import type { LoaderOptions } from '@loaders.gl/loader-utils';
|
|
2
|
-
import
|
|
2
|
+
import * as arrow from 'apache-arrow';
|
|
3
3
|
export type ParquetWasmLoaderOptions = LoaderOptions & {
|
|
4
4
|
parquet?: {
|
|
5
5
|
type?: 'arrow-table';
|
|
6
6
|
wasmUrl?: string;
|
|
7
7
|
};
|
|
8
8
|
};
|
|
9
|
-
export declare function parseParquetWasm(arrayBuffer: ArrayBuffer, options?: ParquetWasmLoaderOptions): Promise<
|
|
9
|
+
export declare function parseParquetWasm(arrayBuffer: ArrayBuffer, options?: ParquetWasmLoaderOptions): Promise<arrow.Table>;
|
|
10
10
|
//# sourceMappingURL=parse-parquet-wasm.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parse-parquet-wasm.d.ts","sourceRoot":"","sources":["../../../src/lib/wasm/parse-parquet-wasm.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"parse-parquet-wasm.d.ts","sourceRoot":"","sources":["../../../src/lib/wasm/parse-parquet-wasm.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAC,aAAa,EAAC,MAAM,0BAA0B,CAAC;AAC5D,OAAO,KAAK,KAAK,MAAM,cAAc,CAAC;AAGtC,MAAM,MAAM,wBAAwB,GAAG,aAAa,GAAG;IACrD,OAAO,CAAC,EAAE;QACR,IAAI,CAAC,EAAE,aAAa,CAAC;QACrB,OAAO,CAAC,EAAE,MAAM,CAAC;KAClB,CAAC;CACH,CAAC;AAEF,wBAAsB,gBAAgB,CACpC,WAAW,EAAE,WAAW,EACxB,OAAO,CAAC,EAAE,wBAAwB,GACjC,OAAO,CAAC,KAAK,CAAC,KAAK,CAAC,CAYtB"}
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import
|
|
1
|
+
import * as arrow from 'apache-arrow';
|
|
2
2
|
import { loadWasm } from "./load-wasm/load-wasm-node.js";
|
|
3
3
|
export async function parseParquetWasm(arrayBuffer, options) {
|
|
4
4
|
var _options$parquet;
|
|
@@ -11,11 +11,11 @@ export async function parseParquetWasm(arrayBuffer, options) {
|
|
|
11
11
|
return arrowTable;
|
|
12
12
|
}
|
|
13
13
|
function tableFromIPC(input) {
|
|
14
|
-
const reader = RecordBatchStreamReader.from(input);
|
|
14
|
+
const reader = arrow.RecordBatchStreamReader.from(input);
|
|
15
15
|
const recordBatches = [];
|
|
16
16
|
for (const recordBatch of reader) {
|
|
17
17
|
recordBatches.push(recordBatch);
|
|
18
18
|
}
|
|
19
|
-
return new
|
|
19
|
+
return new arrow.Table(recordBatches);
|
|
20
20
|
}
|
|
21
21
|
//# sourceMappingURL=parse-parquet-wasm.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parse-parquet-wasm.js","names":["
|
|
1
|
+
{"version":3,"file":"parse-parquet-wasm.js","names":["arrow","loadWasm","parseParquetWasm","arrayBuffer","options","_options$parquet","wasmUrl","parquet","wasm","arr","Uint8Array","arrowIPCUint8Arr","readParquet","arrowIPCBuffer","buffer","slice","byteOffset","byteLength","arrowTable","tableFromIPC","input","reader","RecordBatchStreamReader","from","recordBatches","recordBatch","push","Table"],"sources":["../../../src/lib/wasm/parse-parquet-wasm.ts"],"sourcesContent":["// eslint-disable\nimport type {LoaderOptions} from '@loaders.gl/loader-utils';\nimport * as arrow from 'apache-arrow';\nimport {loadWasm} from './load-wasm/load-wasm-node';\n\nexport type ParquetWasmLoaderOptions = LoaderOptions & {\n parquet?: {\n type?: 'arrow-table';\n wasmUrl?: string;\n };\n};\n\nexport async function parseParquetWasm(\n arrayBuffer: ArrayBuffer,\n options?: ParquetWasmLoaderOptions\n): Promise<arrow.Table> {\n const wasmUrl = options?.parquet?.wasmUrl;\n const wasm = await loadWasm(wasmUrl);\n\n const arr = new Uint8Array(arrayBuffer);\n const arrowIPCUint8Arr = wasm.readParquet(arr);\n const arrowIPCBuffer = arrowIPCUint8Arr.buffer.slice(\n arrowIPCUint8Arr.byteOffset,\n arrowIPCUint8Arr.byteLength + arrowIPCUint8Arr.byteOffset\n );\n const arrowTable = tableFromIPC(arrowIPCBuffer);\n return arrowTable;\n}\n\n/**\n * Deserialize the IPC format into a {@link Table}. This function is a\n * convenience wrapper for {@link RecordBatchReader}. Opposite of {@link tableToIPC}.\n */\nfunction tableFromIPC(input: ArrayBuffer): arrow.Table {\n const reader = arrow.RecordBatchStreamReader.from(input);\n const recordBatches: arrow.RecordBatch[] = [];\n for (const recordBatch of reader) {\n recordBatches.push(recordBatch);\n }\n return new arrow.Table(recordBatches);\n}\n"],"mappings":"AAEA,OAAO,KAAKA,KAAK,MAAM,cAAc;AAAC,SAC9BC,QAAQ;AAShB,OAAO,eAAeC,gBAAgBA,CACpCC,WAAwB,EACxBC,OAAkC,EACZ;EAAA,IAAAC,gBAAA;EACtB,MAAMC,OAAO,GAAGF,OAAO,aAAPA,OAAO,wBAAAC,gBAAA,GAAPD,OAAO,CAAEG,OAAO,cAAAF,gBAAA,uBAAhBA,gBAAA,CAAkBC,OAAO;EACzC,MAAME,IAAI,GAAG,MAAMP,QAAQ,CAACK,OAAO,CAAC;EAEpC,MAAMG,GAAG,GAAG,IAAIC,UAAU,CAACP,WAAW,CAAC;EACvC,MAAMQ,gBAAgB,GAAGH,IAAI,CAACI,WAAW,CAACH,GAAG,CAAC;EAC9C,MAAMI,cAAc,GAAGF,gBAAgB,CAACG,MAAM,CAACC,KAAK,CAClDJ,gBAAgB,CAACK,UAAU,EAC3BL,gBAAgB,CAACM,UAAU,GAAGN,gBAAgB,CAACK,UACjD,CAAC;EACD,MAAME,UAAU,GAAGC,YAAY,CAACN,cAAc,CAAC;EAC/C,OAAOK,UAAU;AACnB;AAMA,SAASC,YAAYA,CAACC,KAAkB,EAAe;EACrD,MAAMC,MAAM,GAAGrB,KAAK,CAACsB,uBAAuB,CAACC,IAAI,CAACH,KAAK,CAAC;EACxD,MAAMI,aAAkC,GAAG,EAAE;EAC7C,KAAK,MAAMC,WAAW,IAAIJ,MAAM,EAAE;IAChCG,aAAa,CAACE,IAAI,CAACD,WAAW,CAAC;EACjC;EACA,OAAO,IAAIzB,KAAK,CAAC2B,KAAK,CAACH,aAAa,CAAC;AACvC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import type { Loader, LoaderOptions } from '@loaders.gl/loader-utils';
|
|
2
|
-
import type
|
|
2
|
+
import type * as arrow from 'apache-arrow';
|
|
3
3
|
/** Parquet WASM loader options */
|
|
4
4
|
export type ParquetWasmLoaderOptions = LoaderOptions & {
|
|
5
5
|
parquet?: {
|
|
@@ -8,5 +8,5 @@ export type ParquetWasmLoaderOptions = LoaderOptions & {
|
|
|
8
8
|
};
|
|
9
9
|
};
|
|
10
10
|
/** Parquet WASM table loader */
|
|
11
|
-
export declare const ParquetWasmLoader: Loader<
|
|
11
|
+
export declare const ParquetWasmLoader: Loader<arrow.Table, never, ParquetWasmLoaderOptions>;
|
|
12
12
|
//# sourceMappingURL=parquet-wasm-loader.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parquet-wasm-loader.d.ts","sourceRoot":"","sources":["../src/parquet-wasm-loader.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAC,MAAM,EAAE,aAAa,EAAC,MAAM,0BAA0B,CAAC;AACpE,OAAO,KAAK,
|
|
1
|
+
{"version":3,"file":"parquet-wasm-loader.d.ts","sourceRoot":"","sources":["../src/parquet-wasm-loader.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAC,MAAM,EAAE,aAAa,EAAC,MAAM,0BAA0B,CAAC;AACpE,OAAO,KAAK,KAAK,KAAK,MAAM,cAAc,CAAC;AAM3C,kCAAkC;AAClC,MAAM,MAAM,wBAAwB,GAAG,aAAa,GAAG;IACrD,OAAO,CAAC,EAAE;QACR,IAAI,CAAC,EAAE,aAAa,CAAC;QACrB,OAAO,CAAC,EAAE,MAAM,CAAC;KAClB,CAAC;CACH,CAAC;AAEF,gCAAgC;AAChC,eAAO,MAAM,iBAAiB,EAAE,MAAM,CAAC,KAAK,CAAC,KAAK,EAAE,KAAK,EAAE,wBAAwB,CAiBlF,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parquet-wasm-loader.js","names":["VERSION","__VERSION__","ParquetWasmLoader","name","id","module","version","worker","category","extensions","mimeTypes","binary","tests","options","parquet","type","wasmUrl"],"sources":["../src/parquet-wasm-loader.ts"],"sourcesContent":["// loaders.gl, MIT license\n// Copyright (c) vis.gl contributors\n\nimport type {Loader, LoaderOptions} from '@loaders.gl/loader-utils';\nimport type
|
|
1
|
+
{"version":3,"file":"parquet-wasm-loader.js","names":["VERSION","__VERSION__","ParquetWasmLoader","name","id","module","version","worker","category","extensions","mimeTypes","binary","tests","options","parquet","type","wasmUrl"],"sources":["../src/parquet-wasm-loader.ts"],"sourcesContent":["// loaders.gl, MIT license\n// Copyright (c) vis.gl contributors\n\nimport type {Loader, LoaderOptions} from '@loaders.gl/loader-utils';\nimport type * as arrow from 'apache-arrow';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/** Parquet WASM loader options */\nexport type ParquetWasmLoaderOptions = LoaderOptions & {\n parquet?: {\n type?: 'arrow-table';\n wasmUrl?: string;\n };\n};\n\n/** Parquet WASM table loader */\nexport const ParquetWasmLoader: Loader<arrow.Table, never, ParquetWasmLoaderOptions> = {\n name: 'Apache Parquet',\n id: 'parquet-wasm',\n module: 'parquet',\n version: VERSION,\n worker: false,\n category: 'table',\n extensions: ['parquet'],\n mimeTypes: ['application/octet-stream'],\n binary: true,\n tests: ['PAR1', 'PARE'],\n options: {\n parquet: {\n type: 'arrow-table',\n wasmUrl: 'https://unpkg.com/parquet-wasm@0.3.1/esm2/arrow1_bg.wasm'\n }\n }\n};\n"],"mappings":"AAQA,MAAMA,OAAO,GAAG,OAAOC,WAAW,KAAK,WAAW,GAAGA,WAAW,GAAG,QAAQ;AAW3E,OAAO,MAAMC,iBAAuE,GAAG;EACrFC,IAAI,EAAE,gBAAgB;EACtBC,EAAE,EAAE,cAAc;EAClBC,MAAM,EAAE,SAAS;EACjBC,OAAO,EAAEN,OAAO;EAChBO,MAAM,EAAE,KAAK;EACbC,QAAQ,EAAE,OAAO;EACjBC,UAAU,EAAE,CAAC,SAAS,CAAC;EACvBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EACvCC,MAAM,EAAE,IAAI;EACZC,KAAK,EAAE,CAAC,MAAM,EAAE,MAAM,CAAC;EACvBC,OAAO,EAAE;IACPC,OAAO,EAAE;MACPC,IAAI,EAAE,aAAa;MACnBC,OAAO,EAAE;IACX;EACF;AACF,CAAC"}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import type { Writer } from '@loaders.gl/loader-utils';
|
|
2
2
|
import { ParquetWriterOptions } from './lib/wasm/encode-parquet-wasm';
|
|
3
|
-
import type
|
|
3
|
+
import type * as arrow from 'apache-arrow';
|
|
4
4
|
/** Parquet WASM writer */
|
|
5
|
-
export declare const ParquetWasmWriter: Writer<
|
|
5
|
+
export declare const ParquetWasmWriter: Writer<arrow.Table, never, ParquetWriterOptions>;
|
|
6
6
|
//# sourceMappingURL=parquet-wasm-writer.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parquet-wasm-writer.d.ts","sourceRoot":"","sources":["../src/parquet-wasm-writer.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAC,MAAM,EAAC,MAAM,0BAA0B,CAAC;AACrD,OAAO,EAAS,oBAAoB,EAAC,MAAM,gCAAgC,CAAC;AAC5E,OAAO,KAAK,
|
|
1
|
+
{"version":3,"file":"parquet-wasm-writer.d.ts","sourceRoot":"","sources":["../src/parquet-wasm-writer.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAC,MAAM,EAAC,MAAM,0BAA0B,CAAC;AACrD,OAAO,EAAS,oBAAoB,EAAC,MAAM,gCAAgC,CAAC;AAC5E,OAAO,KAAK,KAAK,KAAK,MAAM,cAAc,CAAC;AAM3C,0BAA0B;AAC1B,eAAO,MAAM,iBAAiB,EAAE,MAAM,CAAC,KAAK,CAAC,KAAK,EAAE,KAAK,EAAE,oBAAoB,CAc9E,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parquet-wasm-writer.js","names":["encode","VERSION","__VERSION__","ParquetWasmWriter","name","id","module","version","extensions","mimeTypes","binary","options","parquet","wasmUrl"],"sources":["../src/parquet-wasm-writer.ts"],"sourcesContent":["// loaders.gl, MIT license\n// Copyright (c) vis.gl contributors\n\nimport type {Writer} from '@loaders.gl/loader-utils';\nimport {encode, ParquetWriterOptions} from './lib/wasm/encode-parquet-wasm';\nimport type
|
|
1
|
+
{"version":3,"file":"parquet-wasm-writer.js","names":["encode","VERSION","__VERSION__","ParquetWasmWriter","name","id","module","version","extensions","mimeTypes","binary","options","parquet","wasmUrl"],"sources":["../src/parquet-wasm-writer.ts"],"sourcesContent":["// loaders.gl, MIT license\n// Copyright (c) vis.gl contributors\n\nimport type {Writer} from '@loaders.gl/loader-utils';\nimport {encode, ParquetWriterOptions} from './lib/wasm/encode-parquet-wasm';\nimport type * as arrow from 'apache-arrow';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\n/** Parquet WASM writer */\nexport const ParquetWasmWriter: Writer<arrow.Table, never, ParquetWriterOptions> = {\n name: 'Apache Parquet',\n id: 'parquet-wasm',\n module: 'parquet',\n version: VERSION,\n extensions: ['parquet'],\n mimeTypes: ['application/octet-stream'],\n encode,\n binary: true,\n options: {\n parquet: {\n wasmUrl: 'https://unpkg.com/parquet-wasm@0.3.1/esm2/arrow1_bg.wasm'\n }\n }\n};\n"],"mappings":"SAIQA,MAAM;AAKd,MAAMC,OAAO,GAAG,OAAOC,WAAW,KAAK,WAAW,GAAGA,WAAW,GAAG,QAAQ;AAG3E,OAAO,MAAMC,iBAAmE,GAAG;EACjFC,IAAI,EAAE,gBAAgB;EACtBC,EAAE,EAAE,cAAc;EAClBC,MAAM,EAAE,SAAS;EACjBC,OAAO,EAAEN,OAAO;EAChBO,UAAU,EAAE,CAAC,SAAS,CAAC;EACvBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EACvCT,MAAM;EACNU,MAAM,EAAE,IAAI;EACZC,OAAO,EAAE;IACPC,OAAO,EAAE;MACPC,OAAO,EAAE;IACX;EACF;AACF,CAAC"}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parquet-writer.d.ts","sourceRoot":"","sources":["../src/parquet-writer.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAC,MAAM,EAAC,MAAM,0BAA0B,CAAC;AACrD,OAAO,EAAC,KAAK,EAAE,UAAU,EAAC,MAAM,oBAAoB,CAAC;AAMrD,MAAM,MAAM,oBAAoB,GAAG,EAAE,CAAC;AAEtC,eAAO,MAAM,aAAa,EAAE,MAAM,CAAC,KAAK,EAAE,UAAU,EAAE,oBAAoB,
|
|
1
|
+
{"version":3,"file":"parquet-writer.d.ts","sourceRoot":"","sources":["../src/parquet-writer.ts"],"names":[],"mappings":"AAGA,OAAO,KAAK,EAAC,MAAM,EAAC,MAAM,0BAA0B,CAAC;AACrD,OAAO,EAAC,KAAK,EAAE,UAAU,EAAC,MAAM,oBAAoB,CAAC;AAMrD,MAAM,MAAM,oBAAoB,GAAG,EAAE,CAAC;AAEtC,eAAO,MAAM,aAAa,EAAE,MAAM,CAAC,KAAK,EAAE,UAAU,EAAE,oBAAoB,CAWzE,CAAC"}
|
package/dist/parquet-writer.js
CHANGED
|
@@ -6,9 +6,10 @@ export const ParquetWriter = {
|
|
|
6
6
|
version: VERSION,
|
|
7
7
|
extensions: ['parquet'],
|
|
8
8
|
mimeTypes: ['application/octet-stream'],
|
|
9
|
-
encodeSync,
|
|
10
9
|
binary: true,
|
|
11
|
-
options: {}
|
|
10
|
+
options: {},
|
|
11
|
+
encode: async (data, options) => encodeSync(data, options),
|
|
12
|
+
encodeSync
|
|
12
13
|
};
|
|
13
14
|
function encodeSync(data, options) {
|
|
14
15
|
return new ArrayBuffer(0);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parquet-writer.js","names":["VERSION","__VERSION__","ParquetWriter","name","id","module","version","extensions","mimeTypes","
|
|
1
|
+
{"version":3,"file":"parquet-writer.js","names":["VERSION","__VERSION__","ParquetWriter","name","id","module","version","extensions","mimeTypes","binary","options","encode","data","encodeSync","ArrayBuffer"],"sources":["../src/parquet-writer.ts"],"sourcesContent":["// loaders.gl, MIT license\n// Copyright (c) vis.gl contributors\n\nimport type {Writer} from '@loaders.gl/loader-utils';\nimport {Table, TableBatch} from '@loaders.gl/schema';\n\n// __VERSION__ is injected by babel-plugin-version-inline\n// @ts-ignore TS2304: Cannot find name '__VERSION__'.\nconst VERSION = typeof __VERSION__ !== 'undefined' ? __VERSION__ : 'latest';\n\nexport type ParquetWriterOptions = {};\n\nexport const ParquetWriter: Writer<Table, TableBatch, ParquetWriterOptions> = {\n name: 'Apache Parquet',\n id: 'parquet',\n module: 'parquet',\n version: VERSION,\n extensions: ['parquet'],\n mimeTypes: ['application/octet-stream'],\n binary: true,\n options: {},\n encode: async (data, options) => encodeSync(data, options),\n encodeSync\n};\n\nfunction encodeSync(data, options?: ParquetWriterOptions) {\n return new ArrayBuffer(0);\n}\n"],"mappings":"AAQA,MAAMA,OAAO,GAAG,OAAOC,WAAW,KAAK,WAAW,GAAGA,WAAW,GAAG,QAAQ;AAI3E,OAAO,MAAMC,aAA8D,GAAG;EAC5EC,IAAI,EAAE,gBAAgB;EACtBC,EAAE,EAAE,SAAS;EACbC,MAAM,EAAE,SAAS;EACjBC,OAAO,EAAEN,OAAO;EAChBO,UAAU,EAAE,CAAC,SAAS,CAAC;EACvBC,SAAS,EAAE,CAAC,0BAA0B,CAAC;EACvCC,MAAM,EAAE,IAAI;EACZC,OAAO,EAAE,CAAC,CAAC;EACXC,MAAM,EAAE,MAAAA,CAAOC,IAAI,EAAEF,OAAO,KAAKG,UAAU,CAACD,IAAI,EAAEF,OAAO,CAAC;EAC1DG;AACF,CAAC;AAED,SAASA,UAAUA,CAACD,IAAI,EAAEF,OAA8B,EAAE;EACxD,OAAO,IAAII,WAAW,CAAC,CAAC,CAAC;AAC3B"}
|