parquetlens 0.2.0 → 0.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,166 +0,0 @@
1
- var __create = Object.create;
2
- var __defProp = Object.defineProperty;
3
- var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
- var __getOwnPropNames = Object.getOwnPropertyNames;
5
- var __getProtoOf = Object.getPrototypeOf;
6
- var __hasOwnProp = Object.prototype.hasOwnProperty;
7
- var __require = /* @__PURE__ */ ((x) => typeof require !== "undefined" ? require : typeof Proxy !== "undefined" ? new Proxy(x, {
8
- get: (a, b) => (typeof require !== "undefined" ? require : a)[b]
9
- }) : x)(function(x) {
10
- if (typeof require !== "undefined") return require.apply(this, arguments);
11
- throw Error('Dynamic require of "' + x + '" is not supported');
12
- });
13
- var __commonJS = (cb, mod) => function __require2() {
14
- return mod || (0, cb[__getOwnPropNames(cb)[0]])((mod = { exports: {} }).exports, mod), mod.exports;
15
- };
16
- var __copyProps = (to, from, except, desc) => {
17
- if (from && typeof from === "object" || typeof from === "function") {
18
- for (let key of __getOwnPropNames(from))
19
- if (!__hasOwnProp.call(to, key) && key !== except)
20
- __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
21
- }
22
- return to;
23
- };
24
- var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
25
- // If the importer is in node compatibility mode or this is not an ESM
26
- // file that has been converted to a CommonJS file using a Babel-
27
- // compatible transform (i.e. "__esModule" has not been set), then set
28
- // "default" to the CommonJS "module.exports" for node compatibility.
29
- isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
30
- mod
31
- ));
32
-
33
- // ../../packages/parquet-reader/dist/index.js
34
- var require_dist = __commonJS({
35
- "../../packages/parquet-reader/dist/index.js"(exports, module) {
36
- "use strict";
37
- var __create2 = Object.create;
38
- var __defProp2 = Object.defineProperty;
39
- var __getOwnPropDesc2 = Object.getOwnPropertyDescriptor;
40
- var __getOwnPropNames2 = Object.getOwnPropertyNames;
41
- var __getProtoOf2 = Object.getPrototypeOf;
42
- var __hasOwnProp2 = Object.prototype.hasOwnProperty;
43
- var __export = (target, all) => {
44
- for (var name in all)
45
- __defProp2(target, name, { get: all[name], enumerable: true });
46
- };
47
- var __copyProps2 = (to, from, except, desc) => {
48
- if (from && typeof from === "object" || typeof from === "function") {
49
- for (let key of __getOwnPropNames2(from))
50
- if (!__hasOwnProp2.call(to, key) && key !== except)
51
- __defProp2(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc2(from, key)) || desc.enumerable });
52
- }
53
- return to;
54
- };
55
- var __toESM2 = (mod, isNodeMode, target) => (target = mod != null ? __create2(__getProtoOf2(mod)) : {}, __copyProps2(
56
- // If the importer is in node compatibility mode or this is not an ESM
57
- // file that has been converted to a CommonJS file using a Babel-
58
- // compatible transform (i.e. "__esModule" has not been set), then set
59
- // "default" to the CommonJS "module.exports" for node compatibility.
60
- isNodeMode || !mod || !mod.__esModule ? __defProp2(target, "default", { value: mod, enumerable: true }) : target,
61
- mod
62
- ));
63
- var __toCommonJS = (mod) => __copyProps2(__defProp2({}, "__esModule", { value: true }), mod);
64
- var index_exports = {};
65
- __export(index_exports, {
66
- bufferStdinToTempFile: () => bufferStdinToTempFile,
67
- createParquetBufferSource: () => createParquetBufferSource,
68
- openParquetBufferFromPath: () => openParquetBufferFromPath,
69
- readParquetMetadataFromBuffer: () => readParquetMetadataFromBuffer,
70
- readParquetTableFromBuffer: () => readParquetTableFromBuffer,
71
- readParquetTableFromPath: () => readParquetTableFromPath,
72
- readParquetTableFromStdin: () => readParquetTableFromStdin
73
- });
74
- module.exports = __toCommonJS(index_exports);
75
- var import_node_buffer = __require("buffer");
76
- var import_node_fs = __require("fs");
77
- var import_node_fs2 = __require("fs");
78
- var import_node_crypto = __require("crypto");
79
- var import_node_os = __require("os");
80
- var import_node_path = __toESM2(__require("path"));
81
- var import_promises = __require("stream/promises");
82
- var import_apache_arrow = __require("apache-arrow");
83
- var import_parquet_wasm = __require("parquet-wasm");
84
- var BlobCtor = typeof Blob === "undefined" ? import_node_buffer.Blob : Blob;
85
- function readParquetTableFromBuffer(buffer, options) {
86
- const wasmTable = (0, import_parquet_wasm.readParquet)(buffer, options ?? void 0);
87
- const ipcStream = wasmTable.intoIPCStream();
88
- return (0, import_apache_arrow.tableFromIPC)(ipcStream);
89
- }
90
- function createParquetBufferSource(buffer) {
91
- let metadataPromise = null;
92
- return {
93
- buffer,
94
- byteLength: buffer.byteLength,
95
- readTable: (options) => readParquetTableFromBuffer(buffer, options),
96
- readMetadata: () => {
97
- if (!metadataPromise) {
98
- metadataPromise = readParquetMetadataFromBuffer(buffer);
99
- }
100
- return metadataPromise;
101
- }
102
- };
103
- }
104
- async function openParquetBufferFromPath(filePath) {
105
- const buffer = await import_node_fs2.promises.readFile(filePath);
106
- return createParquetBufferSource(buffer);
107
- }
108
- async function readParquetTableFromPath(filePath, options) {
109
- const buffer = await import_node_fs2.promises.readFile(filePath);
110
- return readParquetTableFromBuffer(buffer, options);
111
- }
112
- async function readParquetMetadataFromBuffer(buffer) {
113
- const blobInput = new Uint8Array(buffer).buffer;
114
- const file = await import_parquet_wasm.ParquetFile.fromFile(new BlobCtor([blobInput]));
115
- const meta = file.metadata();
116
- const fileMeta = meta.fileMetadata();
117
- const createdBy = fileMeta.createdBy();
118
- const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());
119
- fileMeta.free();
120
- meta.free();
121
- file.free();
122
- return {
123
- createdBy: createdBy ?? void 0,
124
- keyValueMetadata: normalizeMetadataValues(keyValueMetadata)
125
- };
126
- }
127
- function normalizeMetadataValues(input) {
128
- const normalized = {};
129
- for (const [key, value] of Object.entries(input)) {
130
- if (value === null || value === void 0) {
131
- normalized[key] = "";
132
- continue;
133
- }
134
- normalized[key] = typeof value === "string" ? value : String(value);
135
- }
136
- return normalized;
137
- }
138
- async function bufferStdinToTempFile(filenameHint = "stdin.parquet") {
139
- const tempDir = await import_node_fs2.promises.mkdtemp(import_node_path.default.join((0, import_node_os.tmpdir)(), "parquetlens-"));
140
- const safeName = filenameHint.replace(/[\\/]/g, "_");
141
- const filePath = import_node_path.default.join(tempDir, `${(0, import_node_crypto.randomUUID)()}-${safeName}`);
142
- const writeStream = (0, import_node_fs.createWriteStream)(filePath);
143
- await (0, import_promises.pipeline)(process.stdin, writeStream);
144
- return {
145
- path: filePath,
146
- cleanup: async () => {
147
- await import_node_fs2.promises.rm(tempDir, { recursive: true, force: true });
148
- }
149
- };
150
- }
151
- async function readParquetTableFromStdin(filenameHint = "stdin.parquet", options) {
152
- const temp = await bufferStdinToTempFile(filenameHint);
153
- try {
154
- return await readParquetTableFromPath(temp.path, options);
155
- } finally {
156
- await temp.cleanup();
157
- }
158
- }
159
- }
160
- });
161
-
162
- export {
163
- __toESM,
164
- require_dist
165
- };
166
- //# sourceMappingURL=chunk-E6TEBKS4.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../packages/parquet-reader/src/index.ts"],"sourcesContent":["import { Blob as NodeBlob } from \"node:buffer\";\nimport { createWriteStream } from \"node:fs\";\nimport { promises as fs } from \"node:fs\";\nimport { randomUUID } from \"node:crypto\";\nimport { tmpdir } from \"node:os\";\nimport path from \"node:path\";\nimport { pipeline } from \"node:stream/promises\";\n\nimport { tableFromIPC, Table } from \"apache-arrow\";\nimport { ParquetFile, readParquet, ReaderOptions } from \"parquet-wasm\";\n\nconst BlobCtor: typeof Blob =\n typeof Blob === \"undefined\" ? (NodeBlob as unknown as typeof Blob) : Blob;\n\nexport type TempParquetFile = {\n path: string;\n cleanup: () => Promise<void>;\n};\n\nexport type ParquetReadOptions = Pick<\n ReaderOptions,\n \"batchSize\" | \"columns\" | \"limit\" | \"offset\" | \"rowGroups\"\n>;\n\nexport type ParquetFileMetadata = {\n createdBy?: string;\n keyValueMetadata: Record<string, string>;\n};\n\nexport type ParquetBufferSource = {\n buffer: Uint8Array;\n byteLength: number;\n readTable: (options?: ParquetReadOptions) => Table;\n readMetadata: () => Promise<ParquetFileMetadata>;\n};\n\nexport function readParquetTableFromBuffer(\n buffer: Uint8Array,\n options?: ParquetReadOptions,\n): Table {\n const wasmTable = readParquet(buffer, options ?? undefined);\n const ipcStream = wasmTable.intoIPCStream();\n return tableFromIPC(ipcStream);\n}\n\nexport function createParquetBufferSource(buffer: Uint8Array): ParquetBufferSource {\n let metadataPromise: Promise<ParquetFileMetadata> | null = null;\n\n return {\n buffer,\n byteLength: buffer.byteLength,\n readTable: (options?: ParquetReadOptions) => readParquetTableFromBuffer(buffer, options),\n readMetadata: () => {\n if (!metadataPromise) {\n metadataPromise = readParquetMetadataFromBuffer(buffer);\n }\n return metadataPromise;\n },\n };\n}\n\nexport async function openParquetBufferFromPath(filePath: string): Promise<ParquetBufferSource> {\n const buffer = await fs.readFile(filePath);\n return createParquetBufferSource(buffer);\n}\n\nexport async function readParquetTableFromPath(\n filePath: string,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const buffer = await fs.readFile(filePath);\n return readParquetTableFromBuffer(buffer, options);\n}\n\nexport async function readParquetMetadataFromBuffer(\n buffer: Uint8Array,\n): Promise<ParquetFileMetadata> {\n const blobInput = new Uint8Array(buffer).buffer as ArrayBuffer;\n const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));\n const meta = file.metadata();\n const fileMeta = meta.fileMetadata();\n const createdBy = fileMeta.createdBy();\n const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());\n\n fileMeta.free();\n meta.free();\n file.free();\n\n return {\n createdBy: createdBy ?? undefined,\n keyValueMetadata: normalizeMetadataValues(keyValueMetadata),\n };\n}\n\nfunction normalizeMetadataValues(input: Record<string, unknown>): Record<string, string> {\n const normalized: Record<string, string> = {};\n\n for (const [key, value] of Object.entries(input)) {\n if (value === null || value === undefined) {\n normalized[key] = \"\";\n continue;\n }\n normalized[key] = typeof value === \"string\" ? value : String(value);\n }\n\n return normalized;\n}\n\nexport async function bufferStdinToTempFile(\n filenameHint = \"stdin.parquet\",\n): Promise<TempParquetFile> {\n const tempDir = await fs.mkdtemp(path.join(tmpdir(), \"parquetlens-\"));\n const safeName = filenameHint.replace(/[\\\\/]/g, \"_\");\n const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);\n const writeStream = createWriteStream(filePath);\n\n await pipeline(process.stdin, writeStream);\n\n return {\n path: filePath,\n cleanup: async () => {\n await fs.rm(tempDir, { recursive: true, force: true });\n },\n };\n}\n\nexport async function readParquetTableFromStdin(\n filenameHint = \"stdin.parquet\",\n options?: ParquetReadOptions,\n): Promise<Table> {\n const temp = await bufferStdinToTempFile(filenameHint);\n\n try {\n return await readParquetTableFromPath(temp.path, options);\n } finally {\n await temp.cleanup();\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,QAAA,gBAAA,CAAA;AAAA,aAAA,eAAA;MAAA,uBAAA,MAAA;MAAA,2BAAA,MAAA;MAAA,2BAAA,MAAA;MAAA,+BAAA,MAAA;MAAA,4BAAA,MAAA;MAAA,0BAAA,MAAA;MAAA,2BAAA,MAAA;IAAA,CAAA;AAAA,WAAA,UAAA,aAAA,aAAA;AAAA,QAAA,qBAAiC,UAAA,QAAA;AACjC,QAAA,iBAAkC,UAAA,IAAA;AAClC,QAAAA,kBAA+B,UAAA,IAAA;AAC/B,QAAA,qBAA2B,UAAA,QAAA;AAC3B,QAAA,iBAAuB,UAAA,IAAA;AACvB,QAAA,mBAAiBC,SAAA,UAAA,MAAA,CAAA;AACjB,QAAA,kBAAyB,UAAA,iBAAA;AAEzB,QAAA,sBAAoC,UAAA,cAAA;AACpC,QAAA,sBAAwD,UAAA,cAAA;AAExD,QAAM,WACJ,OAAO,SAAS,cAAe,mBAAAC,OAAsC;AAwBhE,aAAS,2BACd,QACA,SACO;AACP,YAAM,aAAA,GAAY,oBAAA,aAAY,QAAQ,WAAW,MAAS;AAC1D,YAAM,YAAY,UAAU,cAAc;AAC1C,cAAA,GAAO,oBAAA,cAAa,SAAS;IAC/B;AAEO,aAAS,0BAA0B,QAAyC;AACjF,UAAI,kBAAuD;AAE3D,aAAO;QACL;QACA,YAAY,OAAO;QACnB,WAAW,CAAC,YAAiC,2BAA2B,QAAQ,OAAO;QACvF,cAAc,MAAM;AAClB,cAAI,CAAC,iBAAiB;AACpB,8BAAkB,8BAA8B,MAAM;UACxD;AACA,iBAAO;QACT;MACF;IACF;AAEA,mBAAsB,0BAA0B,UAAgD;AAC9F,YAAM,SAAS,MAAM,gBAAAC,SAAG,SAAS,QAAQ;AACzC,aAAO,0BAA0B,MAAM;IACzC;AAEA,mBAAsB,yBACpB,UACA,SACgB;AAChB,YAAM,SAAS,MAAM,gBAAAA,SAAG,SAAS,QAAQ;AACzC,aAAO,2BAA2B,QAAQ,OAAO;IACnD;AAEA,mBAAsB,8BACpB,QAC8B;AAC9B,YAAM,YAAY,IAAI,WAAW,MAAM,EAAE;AACzC,YAAM,OAAO,MAAM,oBAAA,YAAY,SAAS,IAAI,SAAS,CAAC,SAAS,CAAC,CAAC;AACjE,YAAM,OAAO,KAAK,SAAS;AAC3B,YAAM,WAAW,KAAK,aAAa;AACnC,YAAM,YAAY,SAAS,UAAU;AACrC,YAAM,mBAAmB,OAAO,YAAY,SAAS,iBAAiB,CAAC;AAEvE,eAAS,KAAK;AACd,WAAK,KAAK;AACV,WAAK,KAAK;AAEV,aAAO;QACL,WAAW,aAAa;QACxB,kBAAkB,wBAAwB,gBAAgB;MAC5D;IACF;AAEA,aAAS,wBAAwB,OAAwD;AACvF,YAAM,aAAqC,CAAC;AAE5C,iBAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,KAAK,GAAG;AAChD,YAAI,UAAU,QAAQ,UAAU,QAAW;AACzC,qBAAW,GAAG,IAAI;AAClB;QACF;AACA,mBAAW,GAAG,IAAI,OAAO,UAAU,WAAW,QAAQ,OAAO,KAAK;MACpE;AAEA,aAAO;IACT;AAEA,mBAAsB,sBACpB,eAAe,iBACW;AAC1B,YAAM,UAAU,MAAM,gBAAAA,SAAG,QAAQ,iBAAAC,QAAK,MAAA,GAAK,eAAA,QAAO,GAAG,cAAc,CAAC;AACpE,YAAM,WAAW,aAAa,QAAQ,UAAU,GAAG;AACnD,YAAM,WAAW,iBAAAA,QAAK,KAAK,SAAS,IAAA,GAAG,mBAAA,YAAW,CAAC,IAAI,QAAQ,EAAE;AACjE,YAAM,eAAA,GAAc,eAAA,mBAAkB,QAAQ;AAE9C,aAAA,GAAM,gBAAA,UAAS,QAAQ,OAAO,WAAW;AAEzC,aAAO;QACL,MAAM;QACN,SAAS,YAAY;AACnB,gBAAM,gBAAAD,SAAG,GAAG,SAAS,EAAE,WAAW,MAAM,OAAO,KAAK,CAAC;QACvD;MACF;IACF;AAEA,mBAAsB,0BACpB,eAAe,iBACf,SACgB;AAChB,YAAM,OAAO,MAAM,sBAAsB,YAAY;AAErD,UAAI;AACF,eAAO,MAAM,yBAAyB,KAAK,MAAM,OAAO;MAC1D,UAAA;AACE,cAAM,KAAK,QAAQ;MACrB;IACF;;;","names":["import_node_fs","__toESM","NodeBlob","fs","path"]}
@@ -1,112 +0,0 @@
1
- import { createRequire } from 'module';
2
- import { fileURLToPath } from 'url';
3
- import { dirname } from 'path';
4
- const require = createRequire(import.meta.url);
5
- const __filename = fileURLToPath(import.meta.url);
6
- const __dirname = dirname(__filename);
7
-
8
- // ../../packages/parquet-reader/dist/index.js
9
- import { Blob as NodeBlob } from "buffer";
10
- import { createWriteStream, readFileSync } from "fs";
11
- import { promises as fs } from "fs";
12
- import { randomUUID } from "crypto";
13
- import { createRequire } from "module";
14
- import { tmpdir } from "os";
15
- import path from "path";
16
- import { pipeline } from "stream/promises";
17
- import { tableFromIPC } from "apache-arrow";
18
- import { initSync, ParquetFile, readParquet } from "parquet-wasm/esm";
19
- var BlobCtor = typeof Blob === "undefined" ? NodeBlob : Blob;
20
- var wasmInitialized = false;
21
- function ensureWasmInitialized() {
22
- if (wasmInitialized) return;
23
- const require2 = createRequire(import.meta.url);
24
- const wasmPath = require2.resolve("parquet-wasm/esm/parquet_wasm_bg.wasm");
25
- const wasmBytes = readFileSync(wasmPath);
26
- initSync({ module: wasmBytes });
27
- wasmInitialized = true;
28
- }
29
- async function readParquetTableFromBuffer(buffer, options) {
30
- ensureWasmInitialized();
31
- const wasmTable = readParquet(buffer, options ?? void 0);
32
- const ipcStream = wasmTable.intoIPCStream();
33
- return tableFromIPC(ipcStream);
34
- }
35
- function createParquetBufferSource(buffer) {
36
- let metadataPromise = null;
37
- return {
38
- buffer,
39
- byteLength: buffer.byteLength,
40
- readTable: (options) => readParquetTableFromBuffer(buffer, options),
41
- readMetadata: () => {
42
- if (!metadataPromise) {
43
- metadataPromise = readParquetMetadataFromBuffer(buffer);
44
- }
45
- return metadataPromise;
46
- }
47
- };
48
- }
49
- async function openParquetBufferFromPath(filePath) {
50
- const buffer = await fs.readFile(filePath);
51
- return createParquetBufferSource(buffer);
52
- }
53
- async function readParquetTableFromPath(filePath, options) {
54
- const buffer = await fs.readFile(filePath);
55
- return readParquetTableFromBuffer(buffer, options);
56
- }
57
- async function readParquetMetadataFromBuffer(buffer) {
58
- ensureWasmInitialized();
59
- const blobInput = new Uint8Array(buffer).buffer;
60
- const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));
61
- const meta = file.metadata();
62
- const fileMeta = meta.fileMetadata();
63
- const createdBy = fileMeta.createdBy();
64
- const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());
65
- fileMeta.free();
66
- meta.free();
67
- file.free();
68
- return {
69
- createdBy: createdBy ?? void 0,
70
- keyValueMetadata: normalizeMetadataValues(keyValueMetadata)
71
- };
72
- }
73
- function normalizeMetadataValues(input) {
74
- const normalized = {};
75
- for (const [key, value] of Object.entries(input)) {
76
- if (value === null || value === void 0) {
77
- normalized[key] = "";
78
- continue;
79
- }
80
- normalized[key] = typeof value === "string" ? value : String(value);
81
- }
82
- return normalized;
83
- }
84
- async function bufferStdinToTempFile(filenameHint = "stdin.parquet") {
85
- const tempDir = await fs.mkdtemp(path.join(tmpdir(), "parquetlens-"));
86
- const safeName = filenameHint.replace(/[\\/]/g, "_");
87
- const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);
88
- const writeStream = createWriteStream(filePath);
89
- await pipeline(process.stdin, writeStream);
90
- return {
91
- path: filePath,
92
- cleanup: async () => {
93
- await fs.rm(tempDir, { recursive: true, force: true });
94
- }
95
- };
96
- }
97
- async function readParquetTableFromStdin(filenameHint = "stdin.parquet", options) {
98
- const temp = await bufferStdinToTempFile(filenameHint);
99
- try {
100
- return await readParquetTableFromPath(temp.path, options);
101
- } finally {
102
- await temp.cleanup();
103
- }
104
- }
105
-
106
- export {
107
- readParquetTableFromBuffer,
108
- openParquetBufferFromPath,
109
- readParquetTableFromPath,
110
- readParquetTableFromStdin
111
- };
112
- //# sourceMappingURL=chunk-IMVXDI4K.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../packages/parquet-reader/src/index.ts"],"sourcesContent":["import { Blob as NodeBlob } from \"node:buffer\";\nimport { createWriteStream, readFileSync } from \"node:fs\";\nimport { promises as fs } from \"node:fs\";\nimport { randomUUID } from \"node:crypto\";\nimport { createRequire } from \"node:module\";\nimport { tmpdir } from \"node:os\";\nimport path from \"node:path\";\nimport { pipeline } from \"node:stream/promises\";\n\nimport { tableFromIPC, Table } from \"apache-arrow\";\nimport { initSync, ParquetFile, readParquet, type ReaderOptions } from \"parquet-wasm/esm\";\n\nconst BlobCtor: typeof Blob =\n typeof Blob === \"undefined\" ? (NodeBlob as unknown as typeof Blob) : Blob;\n\nlet wasmInitialized = false;\n\nfunction ensureWasmInitialized(): void {\n if (wasmInitialized) return;\n\n const require = createRequire(import.meta.url);\n const wasmPath = require.resolve(\"parquet-wasm/esm/parquet_wasm_bg.wasm\");\n const wasmBytes = readFileSync(wasmPath);\n initSync({ module: wasmBytes });\n wasmInitialized = true;\n}\n\nexport type TempParquetFile = {\n path: string;\n cleanup: () => Promise<void>;\n};\n\nexport type ParquetReadOptions = Pick<\n ReaderOptions,\n \"batchSize\" | \"columns\" | \"limit\" | \"offset\" | \"rowGroups\"\n>;\n\nexport type ParquetFileMetadata = {\n createdBy?: string;\n keyValueMetadata: Record<string, string>;\n};\n\nexport type ParquetBufferSource = {\n buffer: Uint8Array;\n byteLength: number;\n readTable: (options?: ParquetReadOptions) => Promise<Table>;\n readMetadata: () => Promise<ParquetFileMetadata>;\n};\n\nexport async function readParquetTableFromBuffer(\n buffer: Uint8Array,\n options?: ParquetReadOptions,\n): Promise<Table> {\n ensureWasmInitialized();\n const wasmTable = readParquet(buffer, options ?? undefined);\n const ipcStream = wasmTable.intoIPCStream();\n return tableFromIPC(ipcStream);\n}\n\nexport function createParquetBufferSource(buffer: Uint8Array): ParquetBufferSource {\n let metadataPromise: Promise<ParquetFileMetadata> | null = null;\n\n return {\n buffer,\n byteLength: buffer.byteLength,\n readTable: (options?: ParquetReadOptions) => readParquetTableFromBuffer(buffer, options),\n readMetadata: () => {\n if (!metadataPromise) {\n metadataPromise = readParquetMetadataFromBuffer(buffer);\n }\n return metadataPromise;\n },\n };\n}\n\nexport async function openParquetBufferFromPath(filePath: string): Promise<ParquetBufferSource> {\n const buffer = await fs.readFile(filePath);\n return createParquetBufferSource(buffer);\n}\n\nexport async function readParquetTableFromPath(\n filePath: string,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const buffer = await fs.readFile(filePath);\n return readParquetTableFromBuffer(buffer, options);\n}\n\nexport async function readParquetMetadataFromBuffer(\n buffer: Uint8Array,\n): Promise<ParquetFileMetadata> {\n ensureWasmInitialized();\n const blobInput = new Uint8Array(buffer).buffer as ArrayBuffer;\n const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));\n const meta = file.metadata();\n const fileMeta = meta.fileMetadata();\n const createdBy = fileMeta.createdBy();\n const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());\n\n fileMeta.free();\n meta.free();\n file.free();\n\n return {\n createdBy: createdBy ?? undefined,\n keyValueMetadata: normalizeMetadataValues(keyValueMetadata),\n };\n}\n\nfunction normalizeMetadataValues(input: Record<string, unknown>): Record<string, string> {\n const normalized: Record<string, string> = {};\n\n for (const [key, value] of Object.entries(input)) {\n if (value === null || value === undefined) {\n normalized[key] = \"\";\n continue;\n }\n normalized[key] = typeof value === \"string\" ? value : String(value);\n }\n\n return normalized;\n}\n\nexport async function bufferStdinToTempFile(\n filenameHint = \"stdin.parquet\",\n): Promise<TempParquetFile> {\n const tempDir = await fs.mkdtemp(path.join(tmpdir(), \"parquetlens-\"));\n const safeName = filenameHint.replace(/[\\\\/]/g, \"_\");\n const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);\n const writeStream = createWriteStream(filePath);\n\n await pipeline(process.stdin, writeStream);\n\n return {\n path: filePath,\n cleanup: async () => {\n await fs.rm(tempDir, { recursive: true, force: true });\n },\n };\n}\n\nexport async function readParquetTableFromStdin(\n filenameHint = \"stdin.parquet\",\n options?: ParquetReadOptions,\n): Promise<Table> {\n const temp = await bufferStdinToTempFile(filenameHint);\n\n try {\n return await readParquetTableFromPath(temp.path, options);\n } finally {\n await temp.cleanup();\n }\n}\n"],"mappings":";;;;;;;;AAAA,SAAS,QAAQ,gBAAgB;AACjC,SAAS,mBAAmB,oBAAoB;AAChD,SAAS,YAAY,UAAU;AAC/B,SAAS,kBAAkB;AAC3B,SAAS,qBAAqB;AAC9B,SAAS,cAAc;AACvB,OAAO,UAAU;AACjB,SAAS,gBAAgB;AAEzB,SAAS,oBAA2B;AACpC,SAAS,UAAU,aAAa,mBAAuC;AAEvE,IAAM,WACJ,OAAO,SAAS,cAAe,WAAsC;AAEvE,IAAI,kBAAkB;AAEtB,SAAS,wBAA8B;AACrC,MAAI,gBAAiB;AAErB,QAAMA,WAAU,cAAc,YAAY,GAAG;AAC7C,QAAM,WAAWA,SAAQ,QAAQ,uCAAuC;AACxE,QAAM,YAAY,aAAa,QAAQ;AACvC,WAAS,EAAE,QAAQ,UAAU,CAAC;AAC9B,oBAAkB;AACpB;AAwBA,eAAsB,2BACpB,QACA,SACgB;AAChB,wBAAsB;AACtB,QAAM,YAAY,YAAY,QAAQ,WAAW,MAAS;AAC1D,QAAM,YAAY,UAAU,cAAc;AAC1C,SAAO,aAAa,SAAS;AAC/B;AAEO,SAAS,0BAA0B,QAAyC;AACjF,MAAI,kBAAuD;AAE3D,SAAO;IACL;IACA,YAAY,OAAO;IACnB,WAAW,CAAC,YAAiC,2BAA2B,QAAQ,OAAO;IACvF,cAAc,MAAM;AAClB,UAAI,CAAC,iBAAiB;AACpB,0BAAkB,8BAA8B,MAAM;MACxD;AACA,aAAO;IACT;EACF;AACF;AAEA,eAAsB,0BAA0B,UAAgD;AAC9F,QAAM,SAAS,MAAM,GAAG,SAAS,QAAQ;AACzC,SAAO,0BAA0B,MAAM;AACzC;AAEA,eAAsB,yBACpB,UACA,SACgB;AAChB,QAAM,SAAS,MAAM,GAAG,SAAS,QAAQ;AACzC,SAAO,2BAA2B,QAAQ,OAAO;AACnD;AAEA,eAAsB,8BACpB,QAC8B;AAC9B,wBAAsB;AACtB,QAAM,YAAY,IAAI,WAAW,MAAM,EAAE;AACzC,QAAM,OAAO,MAAM,YAAY,SAAS,IAAI,SAAS,CAAC,SAAS,CAAC,CAAC;AACjE,QAAM,OAAO,KAAK,SAAS;AAC3B,QAAM,WAAW,KAAK,aAAa;AACnC,QAAM,YAAY,SAAS,UAAU;AACrC,QAAM,mBAAmB,OAAO,YAAY,SAAS,iBAAiB,CAAC;AAEvE,WAAS,KAAK;AACd,OAAK,KAAK;AACV,OAAK,KAAK;AAEV,SAAO;IACL,WAAW,aAAa;IACxB,kBAAkB,wBAAwB,gBAAgB;EAC5D;AACF;AAEA,SAAS,wBAAwB,OAAwD;AACvF,QAAM,aAAqC,CAAC;AAE5C,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,KAAK,GAAG;AAChD,QAAI,UAAU,QAAQ,UAAU,QAAW;AACzC,iBAAW,GAAG,IAAI;AAClB;IACF;AACA,eAAW,GAAG,IAAI,OAAO,UAAU,WAAW,QAAQ,OAAO,KAAK;EACpE;AAEA,SAAO;AACT;AAEA,eAAsB,sBACpB,eAAe,iBACW;AAC1B,QAAM,UAAU,MAAM,GAAG,QAAQ,KAAK,KAAK,OAAO,GAAG,cAAc,CAAC;AACpE,QAAM,WAAW,aAAa,QAAQ,UAAU,GAAG;AACnD,QAAM,WAAW,KAAK,KAAK,SAAS,GAAG,WAAW,CAAC,IAAI,QAAQ,EAAE;AACjE,QAAM,cAAc,kBAAkB,QAAQ;AAE9C,QAAM,SAAS,QAAQ,OAAO,WAAW;AAEzC,SAAO;IACL,MAAM;IACN,SAAS,YAAY;AACnB,YAAM,GAAG,GAAG,SAAS,EAAE,WAAW,MAAM,OAAO,KAAK,CAAC;IACvD;EACF;AACF;AAEA,eAAsB,0BACpB,eAAe,iBACf,SACgB;AAChB,QAAM,OAAO,MAAM,sBAAsB,YAAY;AAErD,MAAI;AACF,WAAO,MAAM,yBAAyB,KAAK,MAAM,OAAO;EAC1D,UAAA;AACE,UAAM,KAAK,QAAQ;EACrB;AACF;","names":["require"]}
@@ -1,99 +0,0 @@
1
- import { createRequire } from 'module';
2
- import { fileURLToPath } from 'url';
3
- import { dirname } from 'path';
4
- const require = createRequire(import.meta.url);
5
- const __filename = fileURLToPath(import.meta.url);
6
- const __dirname = dirname(__filename);
7
-
8
- // ../../packages/parquet-reader/dist/index.js
9
- import { Blob as NodeBlob } from "buffer";
10
- import { createWriteStream } from "fs";
11
- import { promises as fs } from "fs";
12
- import { randomUUID } from "crypto";
13
- import { tmpdir } from "os";
14
- import path from "path";
15
- import { pipeline } from "stream/promises";
16
- import { tableFromIPC } from "apache-arrow";
17
- import { ParquetFile, readParquet } from "parquet-wasm/esm";
18
- var BlobCtor = typeof Blob === "undefined" ? NodeBlob : Blob;
19
- function readParquetTableFromBuffer(buffer, options) {
20
- const wasmTable = readParquet(buffer, options ?? void 0);
21
- const ipcStream = wasmTable.intoIPCStream();
22
- return tableFromIPC(ipcStream);
23
- }
24
- function createParquetBufferSource(buffer) {
25
- let metadataPromise = null;
26
- return {
27
- buffer,
28
- byteLength: buffer.byteLength,
29
- readTable: (options) => readParquetTableFromBuffer(buffer, options),
30
- readMetadata: () => {
31
- if (!metadataPromise) {
32
- metadataPromise = readParquetMetadataFromBuffer(buffer);
33
- }
34
- return metadataPromise;
35
- }
36
- };
37
- }
38
- async function openParquetBufferFromPath(filePath) {
39
- const buffer = await fs.readFile(filePath);
40
- return createParquetBufferSource(buffer);
41
- }
42
- async function readParquetTableFromPath(filePath, options) {
43
- const buffer = await fs.readFile(filePath);
44
- return readParquetTableFromBuffer(buffer, options);
45
- }
46
- async function readParquetMetadataFromBuffer(buffer) {
47
- const blobInput = new Uint8Array(buffer).buffer;
48
- const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));
49
- const meta = file.metadata();
50
- const fileMeta = meta.fileMetadata();
51
- const createdBy = fileMeta.createdBy();
52
- const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());
53
- fileMeta.free();
54
- meta.free();
55
- file.free();
56
- return {
57
- createdBy: createdBy ?? void 0,
58
- keyValueMetadata: normalizeMetadataValues(keyValueMetadata)
59
- };
60
- }
61
- function normalizeMetadataValues(input) {
62
- const normalized = {};
63
- for (const [key, value] of Object.entries(input)) {
64
- if (value === null || value === void 0) {
65
- normalized[key] = "";
66
- continue;
67
- }
68
- normalized[key] = typeof value === "string" ? value : String(value);
69
- }
70
- return normalized;
71
- }
72
- async function bufferStdinToTempFile(filenameHint = "stdin.parquet") {
73
- const tempDir = await fs.mkdtemp(path.join(tmpdir(), "parquetlens-"));
74
- const safeName = filenameHint.replace(/[\\/]/g, "_");
75
- const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);
76
- const writeStream = createWriteStream(filePath);
77
- await pipeline(process.stdin, writeStream);
78
- return {
79
- path: filePath,
80
- cleanup: async () => {
81
- await fs.rm(tempDir, { recursive: true, force: true });
82
- }
83
- };
84
- }
85
- async function readParquetTableFromStdin(filenameHint = "stdin.parquet", options) {
86
- const temp = await bufferStdinToTempFile(filenameHint);
87
- try {
88
- return await readParquetTableFromPath(temp.path, options);
89
- } finally {
90
- await temp.cleanup();
91
- }
92
- }
93
-
94
- export {
95
- openParquetBufferFromPath,
96
- readParquetTableFromPath,
97
- readParquetTableFromStdin
98
- };
99
- //# sourceMappingURL=chunk-JOHKCQYH.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../packages/parquet-reader/src/index.ts"],"sourcesContent":["import { Blob as NodeBlob } from \"node:buffer\";\nimport { createWriteStream } from \"node:fs\";\nimport { promises as fs } from \"node:fs\";\nimport { randomUUID } from \"node:crypto\";\nimport { tmpdir } from \"node:os\";\nimport path from \"node:path\";\nimport { pipeline } from \"node:stream/promises\";\n\nimport { tableFromIPC, Table } from \"apache-arrow\";\nimport { ParquetFile, readParquet, type ReaderOptions } from \"parquet-wasm/esm\";\n\nconst BlobCtor: typeof Blob =\n typeof Blob === \"undefined\" ? (NodeBlob as unknown as typeof Blob) : Blob;\n\nexport type TempParquetFile = {\n path: string;\n cleanup: () => Promise<void>;\n};\n\nexport type ParquetReadOptions = Pick<\n ReaderOptions,\n \"batchSize\" | \"columns\" | \"limit\" | \"offset\" | \"rowGroups\"\n>;\n\nexport type ParquetFileMetadata = {\n createdBy?: string;\n keyValueMetadata: Record<string, string>;\n};\n\nexport type ParquetBufferSource = {\n buffer: Uint8Array;\n byteLength: number;\n readTable: (options?: ParquetReadOptions) => Table;\n readMetadata: () => Promise<ParquetFileMetadata>;\n};\n\nexport function readParquetTableFromBuffer(\n buffer: Uint8Array,\n options?: ParquetReadOptions,\n): Table {\n const wasmTable = readParquet(buffer, options ?? undefined);\n const ipcStream = wasmTable.intoIPCStream();\n return tableFromIPC(ipcStream);\n}\n\nexport function createParquetBufferSource(buffer: Uint8Array): ParquetBufferSource {\n let metadataPromise: Promise<ParquetFileMetadata> | null = null;\n\n return {\n buffer,\n byteLength: buffer.byteLength,\n readTable: (options?: ParquetReadOptions) => readParquetTableFromBuffer(buffer, options),\n readMetadata: () => {\n if (!metadataPromise) {\n metadataPromise = readParquetMetadataFromBuffer(buffer);\n }\n return metadataPromise;\n },\n };\n}\n\nexport async function openParquetBufferFromPath(filePath: string): Promise<ParquetBufferSource> {\n const buffer = await fs.readFile(filePath);\n return createParquetBufferSource(buffer);\n}\n\nexport async function readParquetTableFromPath(\n filePath: string,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const buffer = await fs.readFile(filePath);\n return readParquetTableFromBuffer(buffer, options);\n}\n\nexport async function readParquetMetadataFromBuffer(\n buffer: Uint8Array,\n): Promise<ParquetFileMetadata> {\n const blobInput = new Uint8Array(buffer).buffer as ArrayBuffer;\n const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));\n const meta = file.metadata();\n const fileMeta = meta.fileMetadata();\n const createdBy = fileMeta.createdBy();\n const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());\n\n fileMeta.free();\n meta.free();\n file.free();\n\n return {\n createdBy: createdBy ?? undefined,\n keyValueMetadata: normalizeMetadataValues(keyValueMetadata),\n };\n}\n\nfunction normalizeMetadataValues(input: Record<string, unknown>): Record<string, string> {\n const normalized: Record<string, string> = {};\n\n for (const [key, value] of Object.entries(input)) {\n if (value === null || value === undefined) {\n normalized[key] = \"\";\n continue;\n }\n normalized[key] = typeof value === \"string\" ? value : String(value);\n }\n\n return normalized;\n}\n\nexport async function bufferStdinToTempFile(\n filenameHint = \"stdin.parquet\",\n): Promise<TempParquetFile> {\n const tempDir = await fs.mkdtemp(path.join(tmpdir(), \"parquetlens-\"));\n const safeName = filenameHint.replace(/[\\\\/]/g, \"_\");\n const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);\n const writeStream = createWriteStream(filePath);\n\n await pipeline(process.stdin, writeStream);\n\n return {\n path: filePath,\n cleanup: async () => {\n await fs.rm(tempDir, { recursive: true, force: true });\n },\n };\n}\n\nexport async function readParquetTableFromStdin(\n filenameHint = \"stdin.parquet\",\n options?: ParquetReadOptions,\n): Promise<Table> {\n const temp = await bufferStdinToTempFile(filenameHint);\n\n try {\n return await readParquetTableFromPath(temp.path, options);\n } finally {\n await temp.cleanup();\n }\n}\n"],"mappings":";;;;;;;;AAAA,SAAS,QAAQ,gBAAgB;AACjC,SAAS,yBAAyB;AAClC,SAAS,YAAY,UAAU;AAC/B,SAAS,kBAAkB;AAC3B,SAAS,cAAc;AACvB,OAAO,UAAU;AACjB,SAAS,gBAAgB;AAEzB,SAAS,oBAA2B;AACpC,SAAS,aAAa,mBAAuC;AAE7D,IAAM,WACJ,OAAO,SAAS,cAAe,WAAsC;AAwBhE,SAAS,2BACd,QACA,SACO;AACP,QAAM,YAAY,YAAY,QAAQ,WAAW,MAAS;AAC1D,QAAM,YAAY,UAAU,cAAc;AAC1C,SAAO,aAAa,SAAS;AAC/B;AAEO,SAAS,0BAA0B,QAAyC;AACjF,MAAI,kBAAuD;AAE3D,SAAO;IACL;IACA,YAAY,OAAO;IACnB,WAAW,CAAC,YAAiC,2BAA2B,QAAQ,OAAO;IACvF,cAAc,MAAM;AAClB,UAAI,CAAC,iBAAiB;AACpB,0BAAkB,8BAA8B,MAAM;MACxD;AACA,aAAO;IACT;EACF;AACF;AAEA,eAAsB,0BAA0B,UAAgD;AAC9F,QAAM,SAAS,MAAM,GAAG,SAAS,QAAQ;AACzC,SAAO,0BAA0B,MAAM;AACzC;AAEA,eAAsB,yBACpB,UACA,SACgB;AAChB,QAAM,SAAS,MAAM,GAAG,SAAS,QAAQ;AACzC,SAAO,2BAA2B,QAAQ,OAAO;AACnD;AAEA,eAAsB,8BACpB,QAC8B;AAC9B,QAAM,YAAY,IAAI,WAAW,MAAM,EAAE;AACzC,QAAM,OAAO,MAAM,YAAY,SAAS,IAAI,SAAS,CAAC,SAAS,CAAC,CAAC;AACjE,QAAM,OAAO,KAAK,SAAS;AAC3B,QAAM,WAAW,KAAK,aAAa;AACnC,QAAM,YAAY,SAAS,UAAU;AACrC,QAAM,mBAAmB,OAAO,YAAY,SAAS,iBAAiB,CAAC;AAEvE,WAAS,KAAK;AACd,OAAK,KAAK;AACV,OAAK,KAAK;AAEV,SAAO;IACL,WAAW,aAAa;IACxB,kBAAkB,wBAAwB,gBAAgB;EAC5D;AACF;AAEA,SAAS,wBAAwB,OAAwD;AACvF,QAAM,aAAqC,CAAC;AAE5C,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,KAAK,GAAG;AAChD,QAAI,UAAU,QAAQ,UAAU,QAAW;AACzC,iBAAW,GAAG,IAAI;AAClB;IACF;AACA,eAAW,GAAG,IAAI,OAAO,UAAU,WAAW,QAAQ,OAAO,KAAK;EACpE;AAEA,SAAO;AACT;AAEA,eAAsB,sBACpB,eAAe,iBACW;AAC1B,QAAM,UAAU,MAAM,GAAG,QAAQ,KAAK,KAAK,OAAO,GAAG,cAAc,CAAC;AACpE,QAAM,WAAW,aAAa,QAAQ,UAAU,GAAG;AACnD,QAAM,WAAW,KAAK,KAAK,SAAS,GAAG,WAAW,CAAC,IAAI,QAAQ,EAAE;AACjE,QAAM,cAAc,kBAAkB,QAAQ;AAE9C,QAAM,SAAS,QAAQ,OAAO,WAAW;AAEzC,SAAO;IACL,MAAM;IACN,SAAS,YAAY;AACnB,YAAM,GAAG,GAAG,SAAS,EAAE,WAAW,MAAM,OAAO,KAAK,CAAC;IACvD;EACF;AACF;AAEA,eAAsB,0BACpB,eAAe,iBACf,SACgB;AAChB,QAAM,OAAO,MAAM,sBAAsB,YAAY;AAErD,MAAI;AACF,WAAO,MAAM,yBAAyB,KAAK,MAAM,OAAO;EAC1D,UAAA;AACE,UAAM,KAAK,QAAQ;EACrB;AACF;","names":[]}
@@ -1,99 +0,0 @@
1
- import { createRequire } from 'module';
2
- import { fileURLToPath } from 'url';
3
- import { dirname } from 'path';
4
- const require = createRequire(import.meta.url);
5
- const __filename = fileURLToPath(import.meta.url);
6
- const __dirname = dirname(__filename);
7
-
8
- // ../../packages/parquet-reader/dist/index.js
9
- import { Blob as NodeBlob } from "buffer";
10
- import { createWriteStream } from "fs";
11
- import { promises as fs } from "fs";
12
- import { randomUUID } from "crypto";
13
- import { tmpdir } from "os";
14
- import path from "path";
15
- import { pipeline } from "stream/promises";
16
- import { tableFromIPC } from "apache-arrow";
17
- import { ParquetFile, readParquet } from "parquet-wasm";
18
- var BlobCtor = typeof Blob === "undefined" ? NodeBlob : Blob;
19
- function readParquetTableFromBuffer(buffer, options) {
20
- const wasmTable = readParquet(buffer, options ?? void 0);
21
- const ipcStream = wasmTable.intoIPCStream();
22
- return tableFromIPC(ipcStream);
23
- }
24
- function createParquetBufferSource(buffer) {
25
- let metadataPromise = null;
26
- return {
27
- buffer,
28
- byteLength: buffer.byteLength,
29
- readTable: (options) => readParquetTableFromBuffer(buffer, options),
30
- readMetadata: () => {
31
- if (!metadataPromise) {
32
- metadataPromise = readParquetMetadataFromBuffer(buffer);
33
- }
34
- return metadataPromise;
35
- }
36
- };
37
- }
38
- async function openParquetBufferFromPath(filePath) {
39
- const buffer = await fs.readFile(filePath);
40
- return createParquetBufferSource(buffer);
41
- }
42
- async function readParquetTableFromPath(filePath, options) {
43
- const buffer = await fs.readFile(filePath);
44
- return readParquetTableFromBuffer(buffer, options);
45
- }
46
- async function readParquetMetadataFromBuffer(buffer) {
47
- const blobInput = new Uint8Array(buffer).buffer;
48
- const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));
49
- const meta = file.metadata();
50
- const fileMeta = meta.fileMetadata();
51
- const createdBy = fileMeta.createdBy();
52
- const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());
53
- fileMeta.free();
54
- meta.free();
55
- file.free();
56
- return {
57
- createdBy: createdBy ?? void 0,
58
- keyValueMetadata: normalizeMetadataValues(keyValueMetadata)
59
- };
60
- }
61
- function normalizeMetadataValues(input) {
62
- const normalized = {};
63
- for (const [key, value] of Object.entries(input)) {
64
- if (value === null || value === void 0) {
65
- normalized[key] = "";
66
- continue;
67
- }
68
- normalized[key] = typeof value === "string" ? value : String(value);
69
- }
70
- return normalized;
71
- }
72
- async function bufferStdinToTempFile(filenameHint = "stdin.parquet") {
73
- const tempDir = await fs.mkdtemp(path.join(tmpdir(), "parquetlens-"));
74
- const safeName = filenameHint.replace(/[\\/]/g, "_");
75
- const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);
76
- const writeStream = createWriteStream(filePath);
77
- await pipeline(process.stdin, writeStream);
78
- return {
79
- path: filePath,
80
- cleanup: async () => {
81
- await fs.rm(tempDir, { recursive: true, force: true });
82
- }
83
- };
84
- }
85
- async function readParquetTableFromStdin(filenameHint = "stdin.parquet", options) {
86
- const temp = await bufferStdinToTempFile(filenameHint);
87
- try {
88
- return await readParquetTableFromPath(temp.path, options);
89
- } finally {
90
- await temp.cleanup();
91
- }
92
- }
93
-
94
- export {
95
- openParquetBufferFromPath,
96
- readParquetTableFromPath,
97
- readParquetTableFromStdin
98
- };
99
- //# sourceMappingURL=chunk-LHMHT2IQ.js.map
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../packages/parquet-reader/src/index.ts"],"sourcesContent":["import { Blob as NodeBlob } from \"node:buffer\";\nimport { createWriteStream } from \"node:fs\";\nimport { promises as fs } from \"node:fs\";\nimport { randomUUID } from \"node:crypto\";\nimport { tmpdir } from \"node:os\";\nimport path from \"node:path\";\nimport { pipeline } from \"node:stream/promises\";\n\nimport { tableFromIPC, Table } from \"apache-arrow\";\nimport { ParquetFile, readParquet, ReaderOptions } from \"parquet-wasm\";\n\nconst BlobCtor: typeof Blob =\n typeof Blob === \"undefined\" ? (NodeBlob as unknown as typeof Blob) : Blob;\n\nexport type TempParquetFile = {\n path: string;\n cleanup: () => Promise<void>;\n};\n\nexport type ParquetReadOptions = Pick<\n ReaderOptions,\n \"batchSize\" | \"columns\" | \"limit\" | \"offset\" | \"rowGroups\"\n>;\n\nexport type ParquetFileMetadata = {\n createdBy?: string;\n keyValueMetadata: Record<string, string>;\n};\n\nexport type ParquetBufferSource = {\n buffer: Uint8Array;\n byteLength: number;\n readTable: (options?: ParquetReadOptions) => Table;\n readMetadata: () => Promise<ParquetFileMetadata>;\n};\n\nexport function readParquetTableFromBuffer(\n buffer: Uint8Array,\n options?: ParquetReadOptions,\n): Table {\n const wasmTable = readParquet(buffer, options ?? undefined);\n const ipcStream = wasmTable.intoIPCStream();\n return tableFromIPC(ipcStream);\n}\n\nexport function createParquetBufferSource(buffer: Uint8Array): ParquetBufferSource {\n let metadataPromise: Promise<ParquetFileMetadata> | null = null;\n\n return {\n buffer,\n byteLength: buffer.byteLength,\n readTable: (options?: ParquetReadOptions) => readParquetTableFromBuffer(buffer, options),\n readMetadata: () => {\n if (!metadataPromise) {\n metadataPromise = readParquetMetadataFromBuffer(buffer);\n }\n return metadataPromise;\n },\n };\n}\n\nexport async function openParquetBufferFromPath(filePath: string): Promise<ParquetBufferSource> {\n const buffer = await fs.readFile(filePath);\n return createParquetBufferSource(buffer);\n}\n\nexport async function readParquetTableFromPath(\n filePath: string,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const buffer = await fs.readFile(filePath);\n return readParquetTableFromBuffer(buffer, options);\n}\n\nexport async function readParquetMetadataFromBuffer(\n buffer: Uint8Array,\n): Promise<ParquetFileMetadata> {\n const blobInput = new Uint8Array(buffer).buffer as ArrayBuffer;\n const file = await ParquetFile.fromFile(new BlobCtor([blobInput]));\n const meta = file.metadata();\n const fileMeta = meta.fileMetadata();\n const createdBy = fileMeta.createdBy();\n const keyValueMetadata = Object.fromEntries(fileMeta.keyValueMetadata());\n\n fileMeta.free();\n meta.free();\n file.free();\n\n return {\n createdBy: createdBy ?? undefined,\n keyValueMetadata: normalizeMetadataValues(keyValueMetadata),\n };\n}\n\nfunction normalizeMetadataValues(input: Record<string, unknown>): Record<string, string> {\n const normalized: Record<string, string> = {};\n\n for (const [key, value] of Object.entries(input)) {\n if (value === null || value === undefined) {\n normalized[key] = \"\";\n continue;\n }\n normalized[key] = typeof value === \"string\" ? value : String(value);\n }\n\n return normalized;\n}\n\nexport async function bufferStdinToTempFile(\n filenameHint = \"stdin.parquet\",\n): Promise<TempParquetFile> {\n const tempDir = await fs.mkdtemp(path.join(tmpdir(), \"parquetlens-\"));\n const safeName = filenameHint.replace(/[\\\\/]/g, \"_\");\n const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);\n const writeStream = createWriteStream(filePath);\n\n await pipeline(process.stdin, writeStream);\n\n return {\n path: filePath,\n cleanup: async () => {\n await fs.rm(tempDir, { recursive: true, force: true });\n },\n };\n}\n\nexport async function readParquetTableFromStdin(\n filenameHint = \"stdin.parquet\",\n options?: ParquetReadOptions,\n): Promise<Table> {\n const temp = await bufferStdinToTempFile(filenameHint);\n\n try {\n return await readParquetTableFromPath(temp.path, options);\n } finally {\n await temp.cleanup();\n }\n}\n"],"mappings":";;;;;;;;AAAA,SAAS,QAAQ,gBAAgB;AACjC,SAAS,yBAAyB;AAClC,SAAS,YAAY,UAAU;AAC/B,SAAS,kBAAkB;AAC3B,SAAS,cAAc;AACvB,OAAO,UAAU;AACjB,SAAS,gBAAgB;AAEzB,SAAS,oBAA2B;AACpC,SAAS,aAAa,mBAAkC;AAExD,IAAM,WACJ,OAAO,SAAS,cAAe,WAAsC;AAwBhE,SAAS,2BACd,QACA,SACO;AACP,QAAM,YAAY,YAAY,QAAQ,WAAW,MAAS;AAC1D,QAAM,YAAY,UAAU,cAAc;AAC1C,SAAO,aAAa,SAAS;AAC/B;AAEO,SAAS,0BAA0B,QAAyC;AACjF,MAAI,kBAAuD;AAE3D,SAAO;IACL;IACA,YAAY,OAAO;IACnB,WAAW,CAAC,YAAiC,2BAA2B,QAAQ,OAAO;IACvF,cAAc,MAAM;AAClB,UAAI,CAAC,iBAAiB;AACpB,0BAAkB,8BAA8B,MAAM;MACxD;AACA,aAAO;IACT;EACF;AACF;AAEA,eAAsB,0BAA0B,UAAgD;AAC9F,QAAM,SAAS,MAAM,GAAG,SAAS,QAAQ;AACzC,SAAO,0BAA0B,MAAM;AACzC;AAEA,eAAsB,yBACpB,UACA,SACgB;AAChB,QAAM,SAAS,MAAM,GAAG,SAAS,QAAQ;AACzC,SAAO,2BAA2B,QAAQ,OAAO;AACnD;AAEA,eAAsB,8BACpB,QAC8B;AAC9B,QAAM,YAAY,IAAI,WAAW,MAAM,EAAE;AACzC,QAAM,OAAO,MAAM,YAAY,SAAS,IAAI,SAAS,CAAC,SAAS,CAAC,CAAC;AACjE,QAAM,OAAO,KAAK,SAAS;AAC3B,QAAM,WAAW,KAAK,aAAa;AACnC,QAAM,YAAY,SAAS,UAAU;AACrC,QAAM,mBAAmB,OAAO,YAAY,SAAS,iBAAiB,CAAC;AAEvE,WAAS,KAAK;AACd,OAAK,KAAK;AACV,OAAK,KAAK;AAEV,SAAO;IACL,WAAW,aAAa;IACxB,kBAAkB,wBAAwB,gBAAgB;EAC5D;AACF;AAEA,SAAS,wBAAwB,OAAwD;AACvF,QAAM,aAAqC,CAAC;AAE5C,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,KAAK,GAAG;AAChD,QAAI,UAAU,QAAQ,UAAU,QAAW;AACzC,iBAAW,GAAG,IAAI;AAClB;IACF;AACA,eAAW,GAAG,IAAI,OAAO,UAAU,WAAW,QAAQ,OAAO,KAAK;EACpE;AAEA,SAAO;AACT;AAEA,eAAsB,sBACpB,eAAe,iBACW;AAC1B,QAAM,UAAU,MAAM,GAAG,QAAQ,KAAK,KAAK,OAAO,GAAG,cAAc,CAAC;AACpE,QAAM,WAAW,aAAa,QAAQ,UAAU,GAAG;AACnD,QAAM,WAAW,KAAK,KAAK,SAAS,GAAG,WAAW,CAAC,IAAI,QAAQ,EAAE;AACjE,QAAM,cAAc,kBAAkB,QAAQ;AAE9C,QAAM,SAAS,QAAQ,OAAO,WAAW;AAEzC,SAAO;IACL,MAAM;IACN,SAAS,YAAY;AACnB,YAAM,GAAG,GAAG,SAAS,EAAE,WAAW,MAAM,OAAO,KAAK,CAAC;IACvD;EACF;AACF;AAEA,eAAsB,0BACpB,eAAe,iBACf,SACgB;AAChB,QAAM,OAAO,MAAM,sBAAsB,YAAY;AAErD,MAAI;AACF,WAAO,MAAM,yBAAyB,KAAK,MAAM,OAAO;EAC1D,UAAA;AACE,UAAM,KAAK,QAAQ;EACrB;AACF;","names":[]}
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../../../packages/parquet-reader/src/index.ts","../../../packages/parquet-reader/src/urls.ts"],"sourcesContent":["import { createWriteStream, readFileSync, rmSync } from \"node:fs\";\nimport { promises as fs } from \"node:fs\";\nimport { randomUUID } from \"node:crypto\";\nimport { spawnSync } from \"node:child_process\";\nimport { Buffer } from \"node:buffer\";\nimport { createRequire as nodeCreateRequire } from \"node:module\";\nimport { tmpdir } from \"node:os\";\nimport path from \"node:path\";\nimport { pipeline } from \"node:stream/promises\";\n\nimport type { Table } from \"apache-arrow\";\nimport {\n DuckDBAccessMode,\n DuckDBBundles,\n DuckDBConnection,\n DuckDBDataProtocol,\n DuckDBBindings,\n FileFlags,\n NODE_RUNTIME,\n VoidLogger,\n createDuckDB,\n failWith,\n readString,\n} from \"@duckdb/duckdb-wasm/blocking\";\nimport type { DuckDBModule } from \"@duckdb/duckdb-wasm/blocking\";\n\nimport { resolveParquetUrl } from \"./urls.js\";\n\nexport type TempParquetFile = {\n path: string;\n cleanup: () => Promise<void>;\n};\n\nexport type ParquetReadOptions = {\n batchSize?: number;\n columns?: string[];\n limit?: number;\n offset?: number;\n rowGroups?: number[];\n};\n\nexport type ParquetFileMetadata = {\n createdBy?: string;\n keyValueMetadata: Record<string, string>;\n};\n\nexport type ParquetSource = {\n readTable: (options?: ParquetReadOptions) => Promise<Table>;\n readMetadata: () => Promise<ParquetFileMetadata>;\n close: () => Promise<void>;\n};\n\nlet duckDbPromise: Promise<DuckDBBindings> | null = null;\nlet httpRuntimePatched = false;\n\ntype HttpBuffer = {\n dataPtr: number;\n size: number;\n};\n\nconst httpBuffers = new Map<number, HttpBuffer>();\n\ntype WasmModule = DuckDBModule & {\n HEAPU8: Uint8Array;\n HEAPF64: Float64Array;\n _malloc: (size: number) => number;\n _free: (ptr: number) => void;\n};\n\nasync function getDuckDb(): Promise<DuckDBBindings> {\n if (!duckDbPromise) {\n duckDbPromise = (async () => {\n ensureHttpRuntimeSupport();\n const bundles = getDuckDbBundles();\n const db = await createDuckDB(bundles, new VoidLogger(), NODE_RUNTIME);\n await db.instantiate();\n db.open({\n accessMode: DuckDBAccessMode.READ_WRITE,\n filesystem: {\n allowFullHTTPReads: true,\n },\n });\n return db;\n })();\n }\n\n return duckDbPromise;\n}\n\nfunction getDuckDbBundles(): DuckDBBundles {\n const localRequire = nodeCreateRequire(import.meta.url);\n const mvpModule = localRequire.resolve(\"@duckdb/duckdb-wasm/dist/duckdb-mvp.wasm\");\n const mvpWorker = localRequire.resolve(\"@duckdb/duckdb-wasm/dist/duckdb-node-mvp.worker.cjs\");\n const ehModule = localRequire.resolve(\"@duckdb/duckdb-wasm/dist/duckdb-eh.wasm\");\n const ehWorker = localRequire.resolve(\"@duckdb/duckdb-wasm/dist/duckdb-node-eh.worker.cjs\");\n\n return {\n mvp: {\n mainModule: mvpModule,\n mainWorker: mvpWorker,\n },\n eh: {\n mainModule: ehModule,\n mainWorker: ehWorker,\n },\n };\n}\n\nfunction ensureHttpRuntimeSupport(): void {\n if (httpRuntimePatched) {\n return;\n }\n httpRuntimePatched = true;\n\n const nodeOpenFile = NODE_RUNTIME.openFile as unknown as (\n mod: WasmModule,\n fileId: number,\n flags: FileFlags,\n ) => number;\n const nodeReadFile = NODE_RUNTIME.readFile as unknown as (\n mod: WasmModule,\n fileId: number,\n buffer: number,\n bytes: number,\n location: number,\n ) => number;\n const nodeCheckFile = NODE_RUNTIME.checkFile.bind(NODE_RUNTIME);\n const nodeGlob = NODE_RUNTIME.glob.bind(NODE_RUNTIME);\n const nodeCloseFile = NODE_RUNTIME.closeFile.bind(NODE_RUNTIME);\n const nodeGetLastModified = NODE_RUNTIME.getLastFileModificationTime.bind(NODE_RUNTIME);\n\n NODE_RUNTIME.openFile = (mod: WasmModule, fileId: number, flags: FileFlags): number => {\n const file = NODE_RUNTIME.resolveFileInfo(mod, fileId);\n if (!file || file.dataProtocol !== DuckDBDataProtocol.HTTP) {\n return nodeOpenFile(mod, fileId, flags);\n }\n\n if (flags & FileFlags.FILE_FLAGS_WRITE || flags & FileFlags.FILE_FLAGS_APPEND) {\n failWith(mod, `Opening file ${file.fileName} failed: HTTP writes are not supported`);\n return 0;\n }\n\n if (!(flags & FileFlags.FILE_FLAGS_READ)) {\n failWith(mod, `Opening file ${file.fileName} failed: unsupported file flags: ${flags}`);\n return 0;\n }\n\n if (!file.dataUrl) {\n failWith(mod, `Opening file ${file.fileName} failed: missing data URL`);\n return 0;\n }\n\n const allowFull = file.allowFullHttpReads ?? true;\n const forceFull = file.forceFullHttpReads ?? false;\n\n if (!forceFull) {\n try {\n const probe = requestHttpRange(file.dataUrl, 0, 0);\n if (probe.status === 206) {\n const total =\n parseContentRangeTotal(probe.headers[\"content-range\"]) ??\n parseContentLength(probe.headers[\"content-length\"]);\n if (total !== null) {\n return buildOpenResult(mod, total, 0);\n }\n }\n\n if (probe.status === 200 && allowFull) {\n const dataPtr = writeResponseToHeap(mod, probe.bytes);\n httpBuffers.set(fileId, { dataPtr, size: probe.bytes.length });\n return buildOpenResult(mod, probe.bytes.length, dataPtr);\n }\n } catch (error) {\n if (!allowFull) {\n failWith(mod, `Opening file ${file.fileName} failed: ${String(error)}`);\n return 0;\n }\n }\n }\n\n if (allowFull) {\n try {\n const full = requestHttp(file.dataUrl);\n if (full.status === 200) {\n const dataPtr = writeResponseToHeap(mod, full.bytes);\n httpBuffers.set(fileId, { dataPtr, size: full.bytes.length });\n return buildOpenResult(mod, full.bytes.length, dataPtr);\n }\n } catch (error) {\n failWith(mod, `Opening file ${file.fileName} failed: ${String(error)}`);\n return 0;\n }\n }\n\n failWith(mod, `Opening file ${file.fileName} failed: HTTP range requests unavailable`);\n return 0;\n };\n\n NODE_RUNTIME.readFile = (\n mod: WasmModule,\n fileId: number,\n buffer: number,\n bytes: number,\n location: number,\n ): number => {\n if (bytes === 0) {\n return 0;\n }\n\n const file = NODE_RUNTIME.resolveFileInfo(mod, fileId);\n if (!file || file.dataProtocol !== DuckDBDataProtocol.HTTP) {\n return nodeReadFile(mod, fileId, buffer, bytes, location);\n }\n\n const cached = httpBuffers.get(fileId);\n if (cached) {\n const sliceStart = Math.max(0, location);\n const sliceEnd = Math.min(cached.size, location + bytes);\n const length = Math.max(0, sliceEnd - sliceStart);\n if (length > 0) {\n const src = mod.HEAPU8.subarray(cached.dataPtr + sliceStart, cached.dataPtr + sliceEnd);\n mod.HEAPU8.set(src, buffer);\n }\n return length;\n }\n\n if (!file.dataUrl) {\n failWith(mod, `Reading file ${file.fileName} failed: missing data URL`);\n return 0;\n }\n\n try {\n const response = requestHttpRange(file.dataUrl, location, location + bytes - 1);\n if (response.status === 206 || (response.status === 200 && location === 0)) {\n const length = Math.min(bytes, response.bytes.length);\n if (length > 0) {\n mod.HEAPU8.set(response.bytes.subarray(0, length), buffer);\n }\n return length;\n }\n\n failWith(mod, `Reading file ${file.fileName} failed with HTTP ${response.status}`);\n return 0;\n } catch (error) {\n failWith(mod, `Reading file ${file.fileName} failed: ${String(error)}`);\n return 0;\n }\n };\n\n NODE_RUNTIME.checkFile = (mod: DuckDBModule, pathPtr: number, pathLen: number): boolean => {\n const path = readString(mod, pathPtr, pathLen);\n if (isHttpUrl(path)) {\n const response = requestHttpHead(path);\n return response.status === 200 || response.status === 206;\n }\n return nodeCheckFile(mod, pathPtr, pathLen);\n };\n\n NODE_RUNTIME.glob = (mod: DuckDBModule, pathPtr: number, pathLen: number): void => {\n const path = readString(mod, pathPtr, pathLen);\n if (isHttpUrl(path)) {\n const response = requestHttpHead(path);\n if (response.status === 200 || response.status === 206) {\n mod.ccall(\"duckdb_web_fs_glob_add_path\", null, [\"string\"], [path]);\n }\n return;\n }\n\n return nodeGlob(mod, pathPtr, pathLen);\n };\n\n NODE_RUNTIME.closeFile = (mod: DuckDBModule, fileId: number): void => {\n const cached = httpBuffers.get(fileId);\n if (cached) {\n if (cached.dataPtr) {\n (mod as WasmModule)._free(cached.dataPtr);\n }\n httpBuffers.delete(fileId);\n }\n nodeCloseFile(mod, fileId);\n };\n\n NODE_RUNTIME.getLastFileModificationTime = (mod: DuckDBModule, fileId: number): number => {\n const file = NODE_RUNTIME.resolveFileInfo(mod, fileId);\n if (file?.dataProtocol === DuckDBDataProtocol.HTTP) {\n return Date.now() / 1000;\n }\n return nodeGetLastModified(mod, fileId);\n };\n}\n\nfunction isHttpUrl(value: string): boolean {\n return value.startsWith(\"http://\") || value.startsWith(\"https://\");\n}\n\nfunction buildOpenResult(mod: WasmModule, size: number, dataPtr: number): number {\n const result = mod._malloc(2 * 8);\n mod.HEAPF64[(result >> 3) + 0] = +size;\n mod.HEAPF64[(result >> 3) + 1] = dataPtr;\n return result;\n}\n\nfunction writeResponseToHeap(mod: WasmModule, bytes: Uint8Array): number {\n const dataPtr = mod._malloc(bytes.byteLength);\n mod.HEAPU8.set(bytes, dataPtr);\n return dataPtr;\n}\n\nfunction parseContentRangeTotal(contentRange: string | null): number | null {\n if (!contentRange) {\n return null;\n }\n const [, total] = contentRange.split(\"/\");\n if (!total) {\n return null;\n }\n const parsed = Number.parseInt(total, 10);\n return Number.isFinite(parsed) ? parsed : null;\n}\n\nfunction parseContentLength(contentLength: string | null): number | null {\n if (!contentLength) {\n return null;\n }\n const parsed = Number.parseInt(contentLength, 10);\n return Number.isFinite(parsed) ? parsed : null;\n}\n\ntype HttpResponse = {\n status: number;\n bytes: Uint8Array;\n headers: Record<string, string>;\n};\n\nfunction requestHttp(url: string): HttpResponse {\n return requestCurl([url]);\n}\n\nfunction requestHttpHead(url: string): HttpResponse {\n return requestCurl([\"-I\", url]);\n}\n\nfunction requestHttpRange(url: string, start: number, end: number): HttpResponse {\n return requestCurl([\"-r\", `${start}-${end}`, url]);\n}\n\nfunction requestCurl(args: string[]): HttpResponse {\n const tempPath = path.join(tmpdir(), `parquetlens-http-${randomUUID()}`);\n try {\n const result = spawnSync(\"curl\", [\"-sS\", \"-L\", \"-D\", \"-\", \"-o\", tempPath, ...args], {\n encoding: \"buffer\",\n maxBuffer: 4 * 1024 * 1024,\n });\n\n if (result.error) {\n if ((result.error as NodeJS.ErrnoException).code === \"ENOENT\") {\n throw new Error(\"curl not found (required for HTTP range reads)\");\n }\n throw result.error;\n }\n if (result.status !== 0) {\n const stderr = result.stderr?.toString(\"utf8\").trim();\n throw new Error(stderr || \"curl failed\");\n }\n\n const body = readFileSync(tempPath);\n return parseCurlResponse(Buffer.from(result.stdout ?? []), body);\n } finally {\n rmSync(tempPath, { force: true });\n }\n}\n\nfunction parseCurlResponse(headersBuffer: Buffer, body: Buffer): HttpResponse {\n const headerBlob = headersBuffer.toString(\"latin1\");\n const blocks = headerBlob.split(/\\r\\n\\r\\n/).filter(Boolean);\n const lastBlock = blocks[blocks.length - 1] ?? \"\";\n const lines = lastBlock.split(/\\r\\n/).filter(Boolean);\n const statusLine = lines.shift() ?? \"\";\n const statusToken = statusLine.split(\" \")[1] ?? \"\";\n const status = Number.parseInt(statusToken, 10);\n const headers: Record<string, string> = {};\n\n for (const line of lines) {\n const index = line.indexOf(\":\");\n if (index === -1) {\n continue;\n }\n const key = line.slice(0, index).trim().toLowerCase();\n const value = line.slice(index + 1).trim();\n headers[key] = value;\n }\n\n return {\n status: Number.isFinite(status) ? status : 0,\n bytes: new Uint8Array(body),\n headers,\n };\n}\n\nexport async function openParquetSourceFromPath(filePath: string): Promise<ParquetSource> {\n const db = await getDuckDb();\n const conn = db.connect();\n const fileName = buildDuckDbFileName(filePath);\n\n db.registerFileURL(fileName, filePath, DuckDBDataProtocol.NODE_FS, true);\n\n return createParquetSource(db, conn, fileName);\n}\n\nexport async function openParquetSourceFromUrl(input: string): Promise<ParquetSource> {\n const resolved = resolveParquetUrl(input);\n if (!resolved) {\n throw new Error(\"Not a URL\");\n }\n\n const db = await getDuckDb();\n const conn = db.connect();\n const fileName = buildDuckDbFileName(resolved.url);\n\n db.registerFileURL(fileName, resolved.url, DuckDBDataProtocol.HTTP, true);\n\n return createParquetSource(db, conn, fileName);\n}\n\nexport async function openParquetSourceFromBuffer(buffer: Uint8Array): Promise<ParquetSource> {\n const db = await getDuckDb();\n const conn = db.connect();\n const fileName = buildDuckDbFileName(\"buffer\");\n\n db.registerFileBuffer(fileName, buffer);\n\n return createParquetSource(db, conn, fileName);\n}\n\nexport async function openParquetSource(input: string): Promise<ParquetSource> {\n const resolved = resolveParquetUrl(input);\n if (resolved) {\n return openParquetSourceFromUrl(input);\n }\n\n return openParquetSourceFromPath(input);\n}\n\nexport async function readParquetTableFromBuffer(\n buffer: Uint8Array,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const source = await openParquetSourceFromBuffer(buffer);\n\n try {\n return await source.readTable(options);\n } finally {\n await source.close();\n }\n}\n\nexport async function readParquetTableFromPath(\n filePath: string,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const source = await openParquetSourceFromPath(filePath);\n\n try {\n return await source.readTable(options);\n } finally {\n await source.close();\n }\n}\n\nexport async function readParquetTableFromUrl(\n input: string,\n options?: ParquetReadOptions,\n): Promise<Table> {\n const source = await openParquetSourceFromUrl(input);\n\n try {\n return await source.readTable(options);\n } finally {\n await source.close();\n }\n}\n\nexport async function readParquetMetadataFromBuffer(\n buffer: Uint8Array,\n): Promise<ParquetFileMetadata> {\n const source = await openParquetSourceFromBuffer(buffer);\n\n try {\n return await source.readMetadata();\n } finally {\n await source.close();\n }\n}\n\nexport async function bufferStdinToTempFile(\n filenameHint = \"stdin.parquet\",\n): Promise<TempParquetFile> {\n const tempDir = await fs.mkdtemp(path.join(tmpdir(), \"parquetlens-\"));\n const safeName = filenameHint.replace(/[\\\\/]/g, \"_\");\n const filePath = path.join(tempDir, `${randomUUID()}-${safeName}`);\n const writeStream = createWriteStream(filePath);\n\n await pipeline(process.stdin, writeStream);\n\n return {\n path: filePath,\n cleanup: async () => {\n await fs.rm(tempDir, { recursive: true, force: true });\n },\n };\n}\n\nexport async function readParquetTableFromStdin(\n filenameHint = \"stdin.parquet\",\n options?: ParquetReadOptions,\n): Promise<Table> {\n const temp = await bufferStdinToTempFile(filenameHint);\n\n try {\n return await readParquetTableFromPath(temp.path, options);\n } finally {\n await temp.cleanup();\n }\n}\n\nfunction createParquetSource(\n db: DuckDBBindings,\n conn: DuckDBConnection,\n fileName: string,\n): ParquetSource {\n let metadataPromise: Promise<ParquetFileMetadata> | null = null;\n\n return {\n readTable: async (options?: ParquetReadOptions) => {\n const query = buildSelectQuery(fileName, options);\n return conn.query(query);\n },\n readMetadata: () => {\n if (!metadataPromise) {\n metadataPromise = readParquetMetadata(conn, fileName);\n }\n return metadataPromise;\n },\n close: async () => {\n conn.close();\n db.dropFile(fileName);\n },\n };\n}\n\nfunction buildDuckDbFileName(input: string): string {\n const suffix = path.extname(input) || \".parquet\";\n return `parquetlens-${randomUUID()}${suffix}`;\n}\n\nfunction buildSelectQuery(fileName: string, options?: ParquetReadOptions): string {\n const columns = options?.columns && options.columns.length > 0 ? options.columns : null;\n const selectList = columns ? columns.map(quoteIdentifier).join(\", \") : \"*\";\n const limit = options?.limit;\n const offset = options?.offset;\n\n let query = `select ${selectList} from read_parquet(${quoteLiteral(fileName)})`;\n\n if (typeof limit === \"number\") {\n query += ` limit ${Math.max(0, limit)}`;\n }\n\n if (typeof offset === \"number\" && offset > 0) {\n query += ` offset ${Math.max(0, offset)}`;\n }\n\n return query;\n}\n\nfunction quoteIdentifier(value: string): string {\n return `\"${value.replace(/\"/g, '\"\"')}\"`;\n}\n\nfunction quoteLiteral(value: string): string {\n return `'${value.replace(/'/g, \"''\")}'`;\n}\n\nasync function readParquetMetadata(\n conn: DuckDBConnection,\n fileName: string,\n): Promise<ParquetFileMetadata> {\n const metadataRows = tableToObjects(\n conn.query(`select * from parquet_file_metadata(${quoteLiteral(fileName)})`),\n );\n const kvRows = tableToObjects(\n conn.query(`select * from parquet_kv_metadata(${quoteLiteral(fileName)})`),\n );\n\n const createdByRaw = metadataRows[0]?.created_by ?? metadataRows[0]?.createdBy ?? null;\n const keyValueMetadata: Record<string, unknown> = {};\n\n for (const row of kvRows) {\n const key = row.key ?? row.key_name ?? row.name;\n if (typeof key !== \"string\" || key.length === 0) {\n continue;\n }\n keyValueMetadata[key] = row.value ?? row.val ?? \"\";\n }\n\n return {\n createdBy: normalizeMetadataValue(createdByRaw),\n keyValueMetadata: normalizeMetadataValues(keyValueMetadata),\n };\n}\n\nfunction tableToObjects(table: Table): Record<string, unknown>[] {\n const fields = table.schema.fields.map((field) => field.name);\n const rows: Record<string, unknown>[] = [];\n\n for (const batch of table.batches) {\n const vectors = fields.map((_, index) => batch.getChildAt(index));\n\n for (let rowIndex = 0; rowIndex < batch.numRows; rowIndex += 1) {\n const row: Record<string, unknown> = {};\n\n for (let colIndex = 0; colIndex < fields.length; colIndex += 1) {\n row[fields[colIndex]] = vectors[colIndex]?.get(rowIndex);\n }\n\n rows.push(row);\n }\n }\n\n return rows;\n}\n\nfunction normalizeMetadataValue(value: unknown): string | undefined {\n if (value === null || value === undefined) {\n return undefined;\n }\n\n if (value instanceof Uint8Array) {\n return Buffer.from(value).toString(\"utf8\");\n }\n\n if (typeof value === \"string\") {\n return value;\n }\n\n return String(value);\n}\n\nfunction normalizeMetadataValues(input: Record<string, unknown>): Record<string, string> {\n const normalized: Record<string, string> = {};\n\n for (const [key, value] of Object.entries(input)) {\n const normalizedValue = normalizeMetadataValue(value);\n normalized[key] = normalizedValue ?? \"\";\n }\n\n return normalized;\n}\n\nexport { resolveParquetUrl } from \"./urls.js\";\nexport type { ResolvedParquetUrl } from \"./urls.js\";\n","export type ResolvedParquetUrl = {\n url: string;\n};\n\nexport function resolveParquetUrl(input: string): ResolvedParquetUrl | null {\n if (input.startsWith(\"hf://\")) {\n return resolveHuggingFaceUrl(input);\n }\n\n if (input.startsWith(\"http://\") || input.startsWith(\"https://\")) {\n return { url: input };\n }\n\n return null;\n}\n\nfunction resolveHuggingFaceUrl(input: string): ResolvedParquetUrl {\n const match = input.match(/^hf:\\/\\/(datasets|models)\\/([^@\\/]+)\\/([^@\\/]+)(?:@([^\\/]+))?\\/(.+)$/);\n\n if (!match) {\n throw new Error(`Invalid hf:// URL: ${input}`);\n }\n\n const [, type, user, repo, branch = \"main\", filePath] = match;\n\n return {\n url: `https://huggingface.co/${type}/${user}/${repo}/resolve/${branch}/${filePath}`,\n };\n}\n"],"mappings":";;;;;;;;AAAA,SAAS,mBAAmB,cAAc,cAAc;AACxD,SAAS,YAAY,UAAU;AAC/B,SAAS,kBAAkB;AAC3B,SAAS,iBAAiB;AAC1B,SAAS,cAAc;AACvB,SAAS,iBAAiB,yBAAyB;AACnD,SAAS,cAAc;AACvB,OAAO,UAAU;AACjB,SAAS,gBAAgB;AAGzB;EACE;EAGA;EAEA;EACA;EACA;EACA;EACA;EACA;OACK;ACnBA,SAAS,kBAAkB,OAA0C;AAC1E,MAAI,MAAM,WAAW,OAAO,GAAG;AAC7B,WAAO,sBAAsB,KAAK;EACpC;AAEA,MAAI,MAAM,WAAW,SAAS,KAAK,MAAM,WAAW,UAAU,GAAG;AAC/D,WAAO,EAAE,KAAK,MAAM;EACtB;AAEA,SAAO;AACT;AAEA,SAAS,sBAAsB,OAAmC;AAChE,QAAM,QAAQ,MAAM,MAAM,sEAAsE;AAEhG,MAAI,CAAC,OAAO;AACV,UAAM,IAAI,MAAM,sBAAsB,KAAK,EAAE;EAC/C;AAEA,QAAM,CAAC,EAAE,MAAM,MAAM,MAAM,SAAS,QAAQ,QAAQ,IAAI;AAExD,SAAO;IACL,KAAK,0BAA0B,IAAI,IAAI,IAAI,IAAI,IAAI,YAAY,MAAM,IAAI,QAAQ;EACnF;AACF;ADwBA,IAAI,gBAAgD;AACpD,IAAI,qBAAqB;AAOzB,IAAM,cAAc,oBAAI,IAAwB;AAShD,eAAe,YAAqC;AAClD,MAAI,CAAC,eAAe;AAClB,qBAAiB,YAAY;AAC3B,+BAAyB;AACzB,YAAM,UAAU,iBAAiB;AACjC,YAAM,KAAK,MAAM,aAAa,SAAS,IAAI,WAAW,GAAG,YAAY;AACrE,YAAM,GAAG,YAAY;AACrB,SAAG,KAAK;QACN,YAAY,iBAAiB;QAC7B,YAAY;UACV,oBAAoB;QACtB;MACF,CAAC;AACD,aAAO;IACT,GAAG;EACL;AAEA,SAAO;AACT;AAEA,SAAS,mBAAkC;AACzC,QAAM,eAAe,kBAAkB,YAAY,GAAG;AACtD,QAAM,YAAY,aAAa,QAAQ,0CAA0C;AACjF,QAAM,YAAY,aAAa,QAAQ,qDAAqD;AAC5F,QAAM,WAAW,aAAa,QAAQ,yCAAyC;AAC/E,QAAM,WAAW,aAAa,QAAQ,oDAAoD;AAE1F,SAAO;IACL,KAAK;MACH,YAAY;MACZ,YAAY;IACd;IACA,IAAI;MACF,YAAY;MACZ,YAAY;IACd;EACF;AACF;AAEA,SAAS,2BAAiC;AACxC,MAAI,oBAAoB;AACtB;EACF;AACA,uBAAqB;AAErB,QAAM,eAAe,aAAa;AAKlC,QAAM,eAAe,aAAa;AAOlC,QAAM,gBAAgB,aAAa,UAAU,KAAK,YAAY;AAC9D,QAAM,WAAW,aAAa,KAAK,KAAK,YAAY;AACpD,QAAM,gBAAgB,aAAa,UAAU,KAAK,YAAY;AAC9D,QAAM,sBAAsB,aAAa,4BAA4B,KAAK,YAAY;AAEtF,eAAa,WAAW,CAAC,KAAiB,QAAgB,UAA6B;AACrF,UAAM,OAAO,aAAa,gBAAgB,KAAK,MAAM;AACrD,QAAI,CAAC,QAAQ,KAAK,iBAAiB,mBAAmB,MAAM;AAC1D,aAAO,aAAa,KAAK,QAAQ,KAAK;IACxC;AAEA,QAAI,QAAQ,UAAU,oBAAoB,QAAQ,UAAU,mBAAmB;AAC7E,eAAS,KAAK,gBAAgB,KAAK,QAAQ,wCAAwC;AACnF,aAAO;IACT;AAEA,QAAI,EAAE,QAAQ,UAAU,kBAAkB;AACxC,eAAS,KAAK,gBAAgB,KAAK,QAAQ,oCAAoC,KAAK,EAAE;AACtF,aAAO;IACT;AAEA,QAAI,CAAC,KAAK,SAAS;AACjB,eAAS,KAAK,gBAAgB,KAAK,QAAQ,2BAA2B;AACtE,aAAO;IACT;AAEA,UAAM,YAAY,KAAK,sBAAsB;AAC7C,UAAM,YAAY,KAAK,sBAAsB;AAE7C,QAAI,CAAC,WAAW;AACd,UAAI;AACF,cAAM,QAAQ,iBAAiB,KAAK,SAAS,GAAG,CAAC;AACjD,YAAI,MAAM,WAAW,KAAK;AACxB,gBAAM,QACJ,uBAAuB,MAAM,QAAQ,eAAe,CAAC,KACrD,mBAAmB,MAAM,QAAQ,gBAAgB,CAAC;AACpD,cAAI,UAAU,MAAM;AAClB,mBAAO,gBAAgB,KAAK,OAAO,CAAC;UACtC;QACF;AAEA,YAAI,MAAM,WAAW,OAAO,WAAW;AACrC,gBAAM,UAAU,oBAAoB,KAAK,MAAM,KAAK;AACpD,sBAAY,IAAI,QAAQ,EAAE,SAAS,MAAM,MAAM,MAAM,OAAO,CAAC;AAC7D,iBAAO,gBAAgB,KAAK,MAAM,MAAM,QAAQ,OAAO;QACzD;MACF,SAAS,OAAO;AACd,YAAI,CAAC,WAAW;AACd,mBAAS,KAAK,gBAAgB,KAAK,QAAQ,YAAY,OAAO,KAAK,CAAC,EAAE;AACtE,iBAAO;QACT;MACF;IACF;AAEA,QAAI,WAAW;AACb,UAAI;AACF,cAAM,OAAO,YAAY,KAAK,OAAO;AACrC,YAAI,KAAK,WAAW,KAAK;AACvB,gBAAM,UAAU,oBAAoB,KAAK,KAAK,KAAK;AACnD,sBAAY,IAAI,QAAQ,EAAE,SAAS,MAAM,KAAK,MAAM,OAAO,CAAC;AAC5D,iBAAO,gBAAgB,KAAK,KAAK,MAAM,QAAQ,OAAO;QACxD;MACF,SAAS,OAAO;AACd,iBAAS,KAAK,gBAAgB,KAAK,QAAQ,YAAY,OAAO,KAAK,CAAC,EAAE;AACtE,eAAO;MACT;IACF;AAEA,aAAS,KAAK,gBAAgB,KAAK,QAAQ,0CAA0C;AACrF,WAAO;EACT;AAEA,eAAa,WAAW,CACtB,KACA,QACA,QACA,OACA,aACW;AACX,QAAI,UAAU,GAAG;AACf,aAAO;IACT;AAEA,UAAM,OAAO,aAAa,gBAAgB,KAAK,MAAM;AACrD,QAAI,CAAC,QAAQ,KAAK,iBAAiB,mBAAmB,MAAM;AAC1D,aAAO,aAAa,KAAK,QAAQ,QAAQ,OAAO,QAAQ;IAC1D;AAEA,UAAM,SAAS,YAAY,IAAI,MAAM;AACrC,QAAI,QAAQ;AACV,YAAM,aAAa,KAAK,IAAI,GAAG,QAAQ;AACvC,YAAM,WAAW,KAAK,IAAI,OAAO,MAAM,WAAW,KAAK;AACvD,YAAM,SAAS,KAAK,IAAI,GAAG,WAAW,UAAU;AAChD,UAAI,SAAS,GAAG;AACd,cAAM,MAAM,IAAI,OAAO,SAAS,OAAO,UAAU,YAAY,OAAO,UAAU,QAAQ;AACtF,YAAI,OAAO,IAAI,KAAK,MAAM;MAC5B;AACA,aAAO;IACT;AAEA,QAAI,CAAC,KAAK,SAAS;AACjB,eAAS,KAAK,gBAAgB,KAAK,QAAQ,2BAA2B;AACtE,aAAO;IACT;AAEA,QAAI;AACF,YAAM,WAAW,iBAAiB,KAAK,SAAS,UAAU,WAAW,QAAQ,CAAC;AAC9E,UAAI,SAAS,WAAW,OAAQ,SAAS,WAAW,OAAO,aAAa,GAAI;AAC1E,cAAM,SAAS,KAAK,IAAI,OAAO,SAAS,MAAM,MAAM;AACpD,YAAI,SAAS,GAAG;AACd,cAAI,OAAO,IAAI,SAAS,MAAM,SAAS,GAAG,MAAM,GAAG,MAAM;QAC3D;AACA,eAAO;MACT;AAEA,eAAS,KAAK,gBAAgB,KAAK,QAAQ,qBAAqB,SAAS,MAAM,EAAE;AACjF,aAAO;IACT,SAAS,OAAO;AACd,eAAS,KAAK,gBAAgB,KAAK,QAAQ,YAAY,OAAO,KAAK,CAAC,EAAE;AACtE,aAAO;IACT;EACF;AAEA,eAAa,YAAY,CAAC,KAAmB,SAAiB,YAA6B;AACzF,UAAMA,QAAO,WAAW,KAAK,SAAS,OAAO;AAC7C,QAAI,UAAUA,KAAI,GAAG;AACnB,YAAM,WAAW,gBAAgBA,KAAI;AACrC,aAAO,SAAS,WAAW,OAAO,SAAS,WAAW;IACxD;AACA,WAAO,cAAc,KAAK,SAAS,OAAO;EAC5C;AAEA,eAAa,OAAO,CAAC,KAAmB,SAAiB,YAA0B;AACjF,UAAMA,QAAO,WAAW,KAAK,SAAS,OAAO;AAC7C,QAAI,UAAUA,KAAI,GAAG;AACnB,YAAM,WAAW,gBAAgBA,KAAI;AACrC,UAAI,SAAS,WAAW,OAAO,SAAS,WAAW,KAAK;AACtD,YAAI,MAAM,+BAA+B,MAAM,CAAC,QAAQ,GAAG,CAACA,KAAI,CAAC;MACnE;AACA;IACF;AAEA,WAAO,SAAS,KAAK,SAAS,OAAO;EACvC;AAEA,eAAa,YAAY,CAAC,KAAmB,WAAyB;AACpE,UAAM,SAAS,YAAY,IAAI,MAAM;AACrC,QAAI,QAAQ;AACV,UAAI,OAAO,SAAS;AACjB,YAAmB,MAAM,OAAO,OAAO;MAC1C;AACA,kBAAY,OAAO,MAAM;IAC3B;AACA,kBAAc,KAAK,MAAM;EAC3B;AAEA,eAAa,8BAA8B,CAAC,KAAmB,WAA2B;AACxF,UAAM,OAAO,aAAa,gBAAgB,KAAK,MAAM;AACrD,QAAI,MAAM,iBAAiB,mBAAmB,MAAM;AAClD,aAAO,KAAK,IAAI,IAAI;IACtB;AACA,WAAO,oBAAoB,KAAK,MAAM;EACxC;AACF;AAEA,SAAS,UAAU,OAAwB;AACzC,SAAO,MAAM,WAAW,SAAS,KAAK,MAAM,WAAW,UAAU;AACnE;AAEA,SAAS,gBAAgB,KAAiB,MAAc,SAAyB;AAC/E,QAAM,SAAS,IAAI,QAAQ,IAAI,CAAC;AAChC,MAAI,SAAS,UAAU,KAAK,CAAC,IAAI,CAAC;AAClC,MAAI,SAAS,UAAU,KAAK,CAAC,IAAI;AACjC,SAAO;AACT;AAEA,SAAS,oBAAoB,KAAiB,OAA2B;AACvE,QAAM,UAAU,IAAI,QAAQ,MAAM,UAAU;AAC5C,MAAI,OAAO,IAAI,OAAO,OAAO;AAC7B,SAAO;AACT;AAEA,SAAS,uBAAuB,cAA4C;AAC1E,MAAI,CAAC,cAAc;AACjB,WAAO;EACT;AACA,QAAM,CAAC,EAAE,KAAK,IAAI,aAAa,MAAM,GAAG;AACxC,MAAI,CAAC,OAAO;AACV,WAAO;EACT;AACA,QAAM,SAAS,OAAO,SAAS,OAAO,EAAE;AACxC,SAAO,OAAO,SAAS,MAAM,IAAI,SAAS;AAC5C;AAEA,SAAS,mBAAmB,eAA6C;AACvE,MAAI,CAAC,eAAe;AAClB,WAAO;EACT;AACA,QAAM,SAAS,OAAO,SAAS,eAAe,EAAE;AAChD,SAAO,OAAO,SAAS,MAAM,IAAI,SAAS;AAC5C;AAQA,SAAS,YAAY,KAA2B;AAC9C,SAAO,YAAY,CAAC,GAAG,CAAC;AAC1B;AAEA,SAAS,gBAAgB,KAA2B;AAClD,SAAO,YAAY,CAAC,MAAM,GAAG,CAAC;AAChC;AAEA,SAAS,iBAAiB,KAAa,OAAe,KAA2B;AAC/E,SAAO,YAAY,CAAC,MAAM,GAAG,KAAK,IAAI,GAAG,IAAI,GAAG,CAAC;AACnD;AAEA,SAAS,YAAY,MAA8B;AACjD,QAAM,WAAW,KAAK,KAAK,OAAO,GAAG,oBAAoB,WAAW,CAAC,EAAE;AACvE,MAAI;AACF,UAAM,SAAS,UAAU,QAAQ,CAAC,OAAO,MAAM,MAAM,KAAK,MAAM,UAAU,GAAG,IAAI,GAAG;MAClF,UAAU;MACV,WAAW,IAAI,OAAO;IACxB,CAAC;AAED,QAAI,OAAO,OAAO;AAChB,UAAK,OAAO,MAAgC,SAAS,UAAU;AAC7D,cAAM,IAAI,MAAM,gDAAgD;MAClE;AACA,YAAM,OAAO;IACf;AACA,QAAI,OAAO,WAAW,GAAG;AACvB,YAAM,SAAS,OAAO,QAAQ,SAAS,MAAM,EAAE,KAAK;AACpD,YAAM,IAAI,MAAM,UAAU,aAAa;IACzC;AAEA,UAAM,OAAO,aAAa,QAAQ;AAClC,WAAO,kBAAkB,OAAO,KAAK,OAAO,UAAU,CAAC,CAAC,GAAG,IAAI;EACjE,UAAA;AACE,WAAO,UAAU,EAAE,OAAO,KAAK,CAAC;EAClC;AACF;AAEA,SAAS,kBAAkB,eAAuB,MAA4B;AAC5E,QAAM,aAAa,cAAc,SAAS,QAAQ;AAClD,QAAM,SAAS,WAAW,MAAM,UAAU,EAAE,OAAO,OAAO;AAC1D,QAAM,YAAY,OAAO,OAAO,SAAS,CAAC,KAAK;AAC/C,QAAM,QAAQ,UAAU,MAAM,MAAM,EAAE,OAAO,OAAO;AACpD,QAAM,aAAa,MAAM,MAAM,KAAK;AACpC,QAAM,cAAc,WAAW,MAAM,GAAG,EAAE,CAAC,KAAK;AAChD,QAAM,SAAS,OAAO,SAAS,aAAa,EAAE;AAC9C,QAAM,UAAkC,CAAC;AAEzC,aAAW,QAAQ,OAAO;AACxB,UAAM,QAAQ,KAAK,QAAQ,GAAG;AAC9B,QAAI,UAAU,IAAI;AAChB;IACF;AACA,UAAM,MAAM,KAAK,MAAM,GAAG,KAAK,EAAE,KAAK,EAAE,YAAY;AACpD,UAAM,QAAQ,KAAK,MAAM,QAAQ,CAAC,EAAE,KAAK;AACzC,YAAQ,GAAG,IAAI;EACjB;AAEA,SAAO;IACL,QAAQ,OAAO,SAAS,MAAM,IAAI,SAAS;IAC3C,OAAO,IAAI,WAAW,IAAI;IAC1B;EACF;AACF;AAEA,eAAsB,0BAA0B,UAA0C;AACxF,QAAM,KAAK,MAAM,UAAU;AAC3B,QAAM,OAAO,GAAG,QAAQ;AACxB,QAAM,WAAW,oBAAoB,QAAQ;AAE7C,KAAG,gBAAgB,UAAU,UAAU,mBAAmB,SAAS,IAAI;AAEvE,SAAO,oBAAoB,IAAI,MAAM,QAAQ;AAC/C;AAEA,eAAsB,yBAAyB,OAAuC;AACpF,QAAM,WAAW,kBAAkB,KAAK;AACxC,MAAI,CAAC,UAAU;AACb,UAAM,IAAI,MAAM,WAAW;EAC7B;AAEA,QAAM,KAAK,MAAM,UAAU;AAC3B,QAAM,OAAO,GAAG,QAAQ;AACxB,QAAM,WAAW,oBAAoB,SAAS,GAAG;AAEjD,KAAG,gBAAgB,UAAU,SAAS,KAAK,mBAAmB,MAAM,IAAI;AAExE,SAAO,oBAAoB,IAAI,MAAM,QAAQ;AAC/C;AAYA,eAAsB,kBAAkB,OAAuC;AAC7E,QAAM,WAAW,kBAAkB,KAAK;AACxC,MAAI,UAAU;AACZ,WAAO,yBAAyB,KAAK;EACvC;AAEA,SAAO,0BAA0B,KAAK;AACxC;AAeA,eAAsB,yBACpB,UACA,SACgB;AAChB,QAAM,SAAS,MAAM,0BAA0B,QAAQ;AAEvD,MAAI;AACF,WAAO,MAAM,OAAO,UAAU,OAAO;EACvC,UAAA;AACE,UAAM,OAAO,MAAM;EACrB;AACF;AAEA,eAAsB,wBACpB,OACA,SACgB;AAChB,QAAM,SAAS,MAAM,yBAAyB,KAAK;AAEnD,MAAI;AACF,WAAO,MAAM,OAAO,UAAU,OAAO;EACvC,UAAA;AACE,UAAM,OAAO,MAAM;EACrB;AACF;AAcA,eAAsB,sBACpB,eAAe,iBACW;AAC1B,QAAM,UAAU,MAAM,GAAG,QAAQ,KAAK,KAAK,OAAO,GAAG,cAAc,CAAC;AACpE,QAAM,WAAW,aAAa,QAAQ,UAAU,GAAG;AACnD,QAAM,WAAW,KAAK,KAAK,SAAS,GAAG,WAAW,CAAC,IAAI,QAAQ,EAAE;AACjE,QAAM,cAAc,kBAAkB,QAAQ;AAE9C,QAAM,SAAS,QAAQ,OAAO,WAAW;AAEzC,SAAO;IACL,MAAM;IACN,SAAS,YAAY;AACnB,YAAM,GAAG,GAAG,SAAS,EAAE,WAAW,MAAM,OAAO,KAAK,CAAC;IACvD;EACF;AACF;AAEA,eAAsB,0BACpB,eAAe,iBACf,SACgB;AAChB,QAAM,OAAO,MAAM,sBAAsB,YAAY;AAErD,MAAI;AACF,WAAO,MAAM,yBAAyB,KAAK,MAAM,OAAO;EAC1D,UAAA;AACE,UAAM,KAAK,QAAQ;EACrB;AACF;AAEA,SAAS,oBACP,IACA,MACA,UACe;AACf,MAAI,kBAAuD;AAE3D,SAAO;IACL,WAAW,OAAO,YAAiC;AACjD,YAAM,QAAQ,iBAAiB,UAAU,OAAO;AAChD,aAAO,KAAK,MAAM,KAAK;IACzB;IACA,cAAc,MAAM;AAClB,UAAI,CAAC,iBAAiB;AACpB,0BAAkB,oBAAoB,MAAM,QAAQ;MACtD;AACA,aAAO;IACT;IACA,OAAO,YAAY;AACjB,WAAK,MAAM;AACX,SAAG,SAAS,QAAQ;IACtB;EACF;AACF;AAEA,SAAS,oBAAoB,OAAuB;AAClD,QAAM,SAAS,KAAK,QAAQ,KAAK,KAAK;AACtC,SAAO,eAAe,WAAW,CAAC,GAAG,MAAM;AAC7C;AAEA,SAAS,iBAAiB,UAAkB,SAAsC;AAChF,QAAM,UAAU,SAAS,WAAW,QAAQ,QAAQ,SAAS,IAAI,QAAQ,UAAU;AACnF,QAAM,aAAa,UAAU,QAAQ,IAAI,eAAe,EAAE,KAAK,IAAI,IAAI;AACvE,QAAM,QAAQ,SAAS;AACvB,QAAM,SAAS,SAAS;AAExB,MAAI,QAAQ,UAAU,UAAU,sBAAsB,aAAa,QAAQ,CAAC;AAE5E,MAAI,OAAO,UAAU,UAAU;AAC7B,aAAS,UAAU,KAAK,IAAI,GAAG,KAAK,CAAC;EACvC;AAEA,MAAI,OAAO,WAAW,YAAY,SAAS,GAAG;AAC5C,aAAS,WAAW,KAAK,IAAI,GAAG,MAAM,CAAC;EACzC;AAEA,SAAO;AACT;AAEA,SAAS,gBAAgB,OAAuB;AAC9C,SAAO,IAAI,MAAM,QAAQ,MAAM,IAAI,CAAC;AACtC;AAEA,SAAS,aAAa,OAAuB;AAC3C,SAAO,IAAI,MAAM,QAAQ,MAAM,IAAI,CAAC;AACtC;AAEA,eAAe,oBACb,MACA,UAC8B;AAC9B,QAAM,eAAe;IACnB,KAAK,MAAM,uCAAuC,aAAa,QAAQ,CAAC,GAAG;EAC7E;AACA,QAAM,SAAS;IACb,KAAK,MAAM,qCAAqC,aAAa,QAAQ,CAAC,GAAG;EAC3E;AAEA,QAAM,eAAe,aAAa,CAAC,GAAG,cAAc,aAAa,CAAC,GAAG,aAAa;AAClF,QAAM,mBAA4C,CAAC;AAEnD,aAAW,OAAO,QAAQ;AACxB,UAAM,MAAM,IAAI,OAAO,IAAI,YAAY,IAAI;AAC3C,QAAI,OAAO,QAAQ,YAAY,IAAI,WAAW,GAAG;AAC/C;IACF;AACA,qBAAiB,GAAG,IAAI,IAAI,SAAS,IAAI,OAAO;EAClD;AAEA,SAAO;IACL,WAAW,uBAAuB,YAAY;IAC9C,kBAAkB,wBAAwB,gBAAgB;EAC5D;AACF;AAEA,SAAS,eAAe,OAAyC;AAC/D,QAAM,SAAS,MAAM,OAAO,OAAO,IAAI,CAAC,UAAU,MAAM,IAAI;AAC5D,QAAM,OAAkC,CAAC;AAEzC,aAAW,SAAS,MAAM,SAAS;AACjC,UAAM,UAAU,OAAO,IAAI,CAAC,GAAG,UAAU,MAAM,WAAW,KAAK,CAAC;AAEhE,aAAS,WAAW,GAAG,WAAW,MAAM,SAAS,YAAY,GAAG;AAC9D,YAAM,MAA+B,CAAC;AAEtC,eAAS,WAAW,GAAG,WAAW,OAAO,QAAQ,YAAY,GAAG;AAC9D,YAAI,OAAO,QAAQ,CAAC,IAAI,QAAQ,QAAQ,GAAG,IAAI,QAAQ;MACzD;AAEA,WAAK,KAAK,GAAG;IACf;EACF;AAEA,SAAO;AACT;AAEA,SAAS,uBAAuB,OAAoC;AAClE,MAAI,UAAU,QAAQ,UAAU,QAAW;AACzC,WAAO;EACT;AAEA,MAAI,iBAAiB,YAAY;AAC/B,WAAO,OAAO,KAAK,KAAK,EAAE,SAAS,MAAM;EAC3C;AAEA,MAAI,OAAO,UAAU,UAAU;AAC7B,WAAO;EACT;AAEA,SAAO,OAAO,KAAK;AACrB;AAEA,SAAS,wBAAwB,OAAwD;AACvF,QAAM,aAAqC,CAAC;AAE5C,aAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,KAAK,GAAG;AAChD,UAAM,kBAAkB,uBAAuB,KAAK;AACpD,eAAW,GAAG,IAAI,mBAAmB;EACvC;AAEA,SAAO;AACT;","names":["path"]}