@milaboratories/pl-model-common 1.19.7 → 1.19.9

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (238) hide show
  1. package/dist/author_marker.d.ts +0 -1
  2. package/dist/base32_encode.cjs +56 -0
  3. package/dist/base32_encode.cjs.map +1 -0
  4. package/dist/base32_encode.d.ts +0 -1
  5. package/dist/base32_encode.js +54 -0
  6. package/dist/base32_encode.js.map +1 -0
  7. package/dist/block_state.d.ts +0 -1
  8. package/dist/bmodel/block_config.cjs +15 -0
  9. package/dist/bmodel/block_config.cjs.map +1 -0
  10. package/dist/bmodel/block_config.d.ts +0 -1
  11. package/dist/bmodel/block_config.js +13 -0
  12. package/dist/bmodel/block_config.js.map +1 -0
  13. package/dist/bmodel/code.d.ts +0 -1
  14. package/dist/bmodel/container.d.ts +0 -1
  15. package/dist/bmodel/index.d.ts +0 -1
  16. package/dist/bmodel/normalization.cjs +86 -0
  17. package/dist/bmodel/normalization.cjs.map +1 -0
  18. package/dist/bmodel/normalization.d.ts +0 -1
  19. package/dist/bmodel/normalization.js +84 -0
  20. package/dist/bmodel/normalization.js.map +1 -0
  21. package/dist/bmodel/types.d.ts +0 -1
  22. package/dist/branding.d.ts +0 -1
  23. package/dist/common_types.d.ts +0 -1
  24. package/dist/driver_kit.d.ts +0 -1
  25. package/dist/drivers/blob.cjs +27 -0
  26. package/dist/drivers/blob.cjs.map +1 -0
  27. package/dist/drivers/blob.d.ts +0 -1
  28. package/dist/drivers/blob.js +23 -0
  29. package/dist/drivers/blob.js.map +1 -0
  30. package/dist/drivers/index.d.ts +0 -1
  31. package/dist/drivers/interfaces.d.ts +0 -1
  32. package/dist/drivers/log.cjs +9 -0
  33. package/dist/drivers/log.cjs.map +1 -0
  34. package/dist/drivers/log.d.ts +0 -1
  35. package/dist/drivers/log.js +7 -0
  36. package/dist/drivers/log.js.map +1 -0
  37. package/dist/drivers/ls.cjs +39 -0
  38. package/dist/drivers/ls.cjs.map +1 -0
  39. package/dist/drivers/ls.d.ts +0 -1
  40. package/dist/drivers/ls.js +34 -0
  41. package/dist/drivers/ls.js.map +1 -0
  42. package/dist/drivers/pframe/column_filter.d.ts +0 -1
  43. package/dist/drivers/pframe/data_info.cjs +275 -0
  44. package/dist/drivers/pframe/data_info.cjs.map +1 -0
  45. package/dist/drivers/pframe/data_info.d.ts +0 -1
  46. package/dist/drivers/pframe/data_info.js +266 -0
  47. package/dist/drivers/pframe/data_info.js.map +1 -0
  48. package/dist/drivers/pframe/data_types.cjs +91 -0
  49. package/dist/drivers/pframe/data_types.cjs.map +1 -0
  50. package/dist/drivers/pframe/data_types.d.ts +0 -1
  51. package/dist/drivers/pframe/data_types.js +83 -0
  52. package/dist/drivers/pframe/data_types.js.map +1 -0
  53. package/dist/drivers/pframe/driver.d.ts +0 -1
  54. package/dist/drivers/pframe/find_columns.d.ts +0 -1
  55. package/dist/drivers/pframe/index.d.ts +0 -1
  56. package/dist/drivers/pframe/linker_columns.cjs +218 -0
  57. package/dist/drivers/pframe/linker_columns.cjs.map +1 -0
  58. package/dist/drivers/pframe/linker_columns.d.ts +0 -1
  59. package/dist/drivers/pframe/linker_columns.js +216 -0
  60. package/dist/drivers/pframe/linker_columns.js.map +1 -0
  61. package/dist/drivers/pframe/pframe.d.ts +0 -1
  62. package/dist/drivers/pframe/spec/anchored.cjs +234 -0
  63. package/dist/drivers/pframe/spec/anchored.cjs.map +1 -0
  64. package/dist/drivers/pframe/spec/anchored.d.ts +0 -1
  65. package/dist/drivers/pframe/spec/anchored.js +231 -0
  66. package/dist/drivers/pframe/spec/anchored.js.map +1 -0
  67. package/dist/drivers/pframe/spec/filtered_column.cjs +13 -0
  68. package/dist/drivers/pframe/spec/filtered_column.cjs.map +1 -0
  69. package/dist/drivers/pframe/spec/filtered_column.d.ts +0 -1
  70. package/dist/drivers/pframe/spec/filtered_column.js +11 -0
  71. package/dist/drivers/pframe/spec/filtered_column.js.map +1 -0
  72. package/dist/drivers/pframe/spec/ids.cjs +24 -0
  73. package/dist/drivers/pframe/spec/ids.cjs.map +1 -0
  74. package/dist/drivers/pframe/spec/ids.d.ts +0 -1
  75. package/dist/drivers/pframe/spec/ids.js +21 -0
  76. package/dist/drivers/pframe/spec/ids.js.map +1 -0
  77. package/dist/drivers/pframe/spec/index.d.ts +0 -1
  78. package/dist/drivers/pframe/spec/native_id.cjs +20 -0
  79. package/dist/drivers/pframe/spec/native_id.cjs.map +1 -0
  80. package/dist/drivers/pframe/spec/native_id.d.ts +0 -1
  81. package/dist/drivers/pframe/spec/native_id.js +18 -0
  82. package/dist/drivers/pframe/spec/native_id.js.map +1 -0
  83. package/dist/drivers/pframe/spec/selectors.cjs +120 -0
  84. package/dist/drivers/pframe/spec/selectors.cjs.map +1 -0
  85. package/dist/drivers/pframe/spec/selectors.d.ts +0 -1
  86. package/dist/drivers/pframe/spec/selectors.js +116 -0
  87. package/dist/drivers/pframe/spec/selectors.js.map +1 -0
  88. package/dist/drivers/pframe/spec/spec.cjs +361 -0
  89. package/dist/drivers/pframe/spec/spec.cjs.map +1 -0
  90. package/dist/drivers/pframe/spec/spec.d.ts +0 -1
  91. package/dist/drivers/pframe/spec/spec.js +332 -0
  92. package/dist/drivers/pframe/spec/spec.js.map +1 -0
  93. package/dist/drivers/pframe/table.d.ts +0 -1
  94. package/dist/drivers/pframe/table_calculate.cjs +43 -0
  95. package/dist/drivers/pframe/table_calculate.cjs.map +1 -0
  96. package/dist/drivers/pframe/table_calculate.d.ts +0 -1
  97. package/dist/drivers/pframe/table_calculate.js +40 -0
  98. package/dist/drivers/pframe/table_calculate.js.map +1 -0
  99. package/dist/drivers/pframe/table_common.cjs +19 -0
  100. package/dist/drivers/pframe/table_common.cjs.map +1 -0
  101. package/dist/drivers/pframe/table_common.d.ts +0 -1
  102. package/dist/drivers/pframe/table_common.js +17 -0
  103. package/dist/drivers/pframe/table_common.js.map +1 -0
  104. package/dist/drivers/pframe/type_util.d.ts +0 -1
  105. package/dist/drivers/pframe/unique_values.d.ts +0 -1
  106. package/dist/drivers/upload.d.ts +0 -1
  107. package/dist/drivers/urls.cjs +14 -0
  108. package/dist/drivers/urls.cjs.map +1 -0
  109. package/dist/drivers/urls.d.ts +0 -1
  110. package/dist/drivers/urls.js +11 -0
  111. package/dist/drivers/urls.js.map +1 -0
  112. package/dist/errors.cjs +141 -0
  113. package/dist/errors.cjs.map +1 -0
  114. package/dist/errors.d.ts +0 -1
  115. package/dist/errors.js +121 -0
  116. package/dist/errors.js.map +1 -0
  117. package/dist/flags/block_flags.cjs +8 -0
  118. package/dist/flags/block_flags.cjs.map +1 -0
  119. package/dist/flags/block_flags.d.ts +0 -1
  120. package/dist/flags/block_flags.js +5 -0
  121. package/dist/flags/block_flags.js.map +1 -0
  122. package/dist/flags/flag_utils.cjs +100 -0
  123. package/dist/flags/flag_utils.cjs.map +1 -0
  124. package/dist/flags/flag_utils.d.ts +0 -1
  125. package/dist/flags/flag_utils.js +94 -0
  126. package/dist/flags/flag_utils.js.map +1 -0
  127. package/dist/flags/index.d.ts +0 -1
  128. package/dist/flags/type_utils.d.ts +0 -1
  129. package/dist/index.cjs +151 -0
  130. package/dist/index.cjs.map +1 -0
  131. package/dist/index.d.ts +10 -12
  132. package/dist/index.js +27 -1
  133. package/dist/index.js.map +1 -1
  134. package/dist/json.cjs +18 -0
  135. package/dist/json.cjs.map +1 -0
  136. package/dist/json.d.ts +0 -1
  137. package/dist/json.js +14 -0
  138. package/dist/json.js.map +1 -0
  139. package/dist/navigation.cjs +6 -0
  140. package/dist/navigation.cjs.map +1 -0
  141. package/dist/navigation.d.ts +0 -1
  142. package/dist/navigation.js +4 -0
  143. package/dist/navigation.js.map +1 -0
  144. package/dist/plid.cjs +37 -0
  145. package/dist/plid.cjs.map +1 -0
  146. package/dist/plid.d.ts +0 -1
  147. package/dist/plid.js +30 -0
  148. package/dist/plid.js.map +1 -0
  149. package/dist/pool/entry.d.ts +0 -1
  150. package/dist/pool/index.d.ts +0 -1
  151. package/dist/pool/query.cjs +49 -0
  152. package/dist/pool/query.cjs.map +1 -0
  153. package/dist/pool/query.d.ts +0 -1
  154. package/dist/pool/query.js +47 -0
  155. package/dist/pool/query.js.map +1 -0
  156. package/dist/pool/spec.cjs +67 -0
  157. package/dist/pool/spec.cjs.map +1 -0
  158. package/dist/pool/spec.d.ts +0 -1
  159. package/dist/pool/spec.js +59 -0
  160. package/dist/pool/spec.js.map +1 -0
  161. package/dist/ref.cjs +88 -0
  162. package/dist/ref.cjs.map +1 -0
  163. package/dist/ref.d.ts +0 -1
  164. package/dist/ref.js +82 -0
  165. package/dist/ref.js.map +1 -0
  166. package/dist/utag.d.ts +0 -1
  167. package/dist/util.cjs +8 -0
  168. package/dist/util.cjs.map +1 -0
  169. package/dist/util.d.ts +0 -1
  170. package/dist/util.js +6 -0
  171. package/dist/util.js.map +1 -0
  172. package/dist/value_or_error.cjs +8 -0
  173. package/dist/value_or_error.cjs.map +1 -0
  174. package/dist/value_or_error.d.ts +0 -1
  175. package/dist/value_or_error.js +6 -0
  176. package/dist/value_or_error.js.map +1 -0
  177. package/package.json +13 -11
  178. package/src/drivers/pframe/type_util.ts +1 -0
  179. package/src/flags/type_utils.ts +1 -0
  180. package/src/index.ts +10 -11
  181. package/dist/author_marker.d.ts.map +0 -1
  182. package/dist/base32_encode.d.ts.map +0 -1
  183. package/dist/block_state.d.ts.map +0 -1
  184. package/dist/bmodel/block_config.d.ts.map +0 -1
  185. package/dist/bmodel/code.d.ts.map +0 -1
  186. package/dist/bmodel/container.d.ts.map +0 -1
  187. package/dist/bmodel/index.d.ts.map +0 -1
  188. package/dist/bmodel/normalization.d.ts.map +0 -1
  189. package/dist/bmodel/types.d.ts.map +0 -1
  190. package/dist/branding.d.ts.map +0 -1
  191. package/dist/common_types.d.ts.map +0 -1
  192. package/dist/driver_kit.d.ts.map +0 -1
  193. package/dist/drivers/blob.d.ts.map +0 -1
  194. package/dist/drivers/index.d.ts.map +0 -1
  195. package/dist/drivers/interfaces.d.ts.map +0 -1
  196. package/dist/drivers/log.d.ts.map +0 -1
  197. package/dist/drivers/ls.d.ts.map +0 -1
  198. package/dist/drivers/pframe/column_filter.d.ts.map +0 -1
  199. package/dist/drivers/pframe/data_info.d.ts.map +0 -1
  200. package/dist/drivers/pframe/data_types.d.ts.map +0 -1
  201. package/dist/drivers/pframe/driver.d.ts.map +0 -1
  202. package/dist/drivers/pframe/find_columns.d.ts.map +0 -1
  203. package/dist/drivers/pframe/index.d.ts.map +0 -1
  204. package/dist/drivers/pframe/linker_columns.d.ts.map +0 -1
  205. package/dist/drivers/pframe/pframe.d.ts.map +0 -1
  206. package/dist/drivers/pframe/spec/anchored.d.ts.map +0 -1
  207. package/dist/drivers/pframe/spec/filtered_column.d.ts.map +0 -1
  208. package/dist/drivers/pframe/spec/ids.d.ts.map +0 -1
  209. package/dist/drivers/pframe/spec/index.d.ts.map +0 -1
  210. package/dist/drivers/pframe/spec/native_id.d.ts.map +0 -1
  211. package/dist/drivers/pframe/spec/selectors.d.ts.map +0 -1
  212. package/dist/drivers/pframe/spec/spec.d.ts.map +0 -1
  213. package/dist/drivers/pframe/table.d.ts.map +0 -1
  214. package/dist/drivers/pframe/table_calculate.d.ts.map +0 -1
  215. package/dist/drivers/pframe/table_common.d.ts.map +0 -1
  216. package/dist/drivers/pframe/type_util.d.ts.map +0 -1
  217. package/dist/drivers/pframe/unique_values.d.ts.map +0 -1
  218. package/dist/drivers/upload.d.ts.map +0 -1
  219. package/dist/drivers/urls.d.ts.map +0 -1
  220. package/dist/errors.d.ts.map +0 -1
  221. package/dist/flags/block_flags.d.ts.map +0 -1
  222. package/dist/flags/flag_utils.d.ts.map +0 -1
  223. package/dist/flags/index.d.ts.map +0 -1
  224. package/dist/flags/type_utils.d.ts.map +0 -1
  225. package/dist/index.d.ts.map +0 -1
  226. package/dist/index.mjs +0 -1587
  227. package/dist/index.mjs.map +0 -1
  228. package/dist/json.d.ts.map +0 -1
  229. package/dist/navigation.d.ts.map +0 -1
  230. package/dist/plid.d.ts.map +0 -1
  231. package/dist/pool/entry.d.ts.map +0 -1
  232. package/dist/pool/index.d.ts.map +0 -1
  233. package/dist/pool/query.d.ts.map +0 -1
  234. package/dist/pool/spec.d.ts.map +0 -1
  235. package/dist/ref.d.ts.map +0 -1
  236. package/dist/utag.d.ts.map +0 -1
  237. package/dist/util.d.ts.map +0 -1
  238. package/dist/value_or_error.d.ts.map +0 -1
@@ -0,0 +1 @@
1
+ {"version":3,"file":"ls.cjs","sources":["../../src/drivers/ls.ts"],"sourcesContent":["import { assertNever } from '../util';\nimport type { Branded } from '../branding';\nimport type { TableRange } from './pframe';\nimport type { FileLike } from './interfaces';\n\nconst uploadPrefix = 'upload://upload/';\nconst indexPrefix = 'index://index/';\n\nexport type ImportFileHandleUpload = `upload://upload/${string}`;\nexport type ImportFileHandleIndex = `index://index/${string}`;\n\nexport type ImportFileHandle = ImportFileHandleUpload | ImportFileHandleIndex;\n\nexport type LocalImportFileHandle = Branded<ImportFileHandle, 'Local'>;\n\nexport function isImportFileHandleUpload(\n handle: ImportFileHandle,\n): handle is ImportFileHandleUpload {\n return handle.startsWith(uploadPrefix);\n}\n\nexport function isImportFileHandleIndex(handle: ImportFileHandle): handle is ImportFileHandleIndex {\n return handle.startsWith(indexPrefix);\n}\n\n/** Results in upload */\nexport type StorageHandleLocal = `local://${string}`;\n\n/** Results in index */\nexport type StorageHandleRemote = `remote://${string}`;\n\nexport type StorageHandle = StorageHandleLocal | StorageHandleRemote;\n\nexport type StorageEntry = {\n name: string;\n handle: StorageHandle;\n initialFullPath: string;\n\n // TODO\n // pathStartsWithDisk\n};\n\nexport type ListFilesResult = {\n parent?: string;\n entries: LsEntry[];\n};\n\nexport type LsEntry =\n | {\n type: 'dir';\n name: string;\n fullPath: string;\n }\n | {\n type: 'file';\n name: string;\n fullPath: string;\n\n /** This handle should be set to args... */\n handle: ImportFileHandle;\n };\n\nexport type OpenDialogFilter = {\n /** Human-readable file type name */\n readonly name: string;\n /** File extensions */\n readonly extensions: string[];\n};\n\nexport type OpenDialogOps = {\n /** Open dialog window title */\n readonly title?: string;\n /** Custom label for the confirmation button, when left empty the default label will be used. */\n readonly buttonLabel?: string;\n /** Limits of file types user can select */\n readonly filters?: OpenDialogFilter[];\n};\n\nexport type OpenSingleFileResponse = {\n /** Contains local file handle, allowing file importing or content reading. If user canceled\n * the dialog, field will be undefined. */\n readonly file?: LocalImportFileHandle;\n};\n\nexport type OpenMultipleFilesResponse = {\n /** Contains local file handles, allowing file importing or content reading. If user canceled\n * the dialog, field will be undefined. */\n readonly files?: LocalImportFileHandle[];\n};\n\n/** Can be used to limit request for local file content to a certain bytes range */\nexport type FileRange = {\n /** From byte index (inclusive) */\n readonly from: number;\n /** To byte index (exclusive) */\n readonly to: number;\n};\n\nexport interface LsDriver {\n /** remote and local storages */\n getStorageList(): Promise<StorageEntry[]>;\n\n listFiles(storage: StorageHandle, fullPath: string): Promise<ListFilesResult>;\n\n /** Opens system file open dialog allowing to select single file and awaits user action */\n showOpenSingleFileDialog(ops: OpenDialogOps): Promise<OpenSingleFileResponse>;\n\n /** Opens system file open dialog allowing to multiple files and awaits user action */\n showOpenMultipleFilesDialog(ops: OpenDialogOps): Promise<OpenMultipleFilesResponse>;\n\n /** Given a handle to a local file, allows to get file size */\n getLocalFileSize(file: LocalImportFileHandle): Promise<number>;\n\n /** Given a handle to a local file, allows to get its content */\n getLocalFileContent(file: LocalImportFileHandle, range?: TableRange): Promise<Uint8Array>;\n\n /**\n * Resolves browser's File object into platforma's import file handle.\n *\n * This method is useful among other things for implementation of UI\n * components, that handle file Drag&Drop.\n * */\n fileToImportHandle(file: FileLike): Promise<ImportFileHandle>;\n\n /** Saves currently opened block webview as a PDF. */\n exportToPdf?(): Promise<void>;\n}\n\n/** Gets a file path from an import handle. */\nexport function getFilePathFromHandle(handle: ImportFileHandle): string {\n if (isImportFileHandleIndex(handle)) {\n const trimmed = handle.slice(indexPrefix.length);\n const data = JSON.parse(decodeURIComponent(trimmed)) as { path: string };\n return data.path;\n } else if (isImportFileHandleUpload(handle)) {\n const trimmed = handle.slice(uploadPrefix.length);\n const data = JSON.parse(decodeURIComponent(trimmed)) as { localPath: string };\n return data.localPath;\n }\n\n assertNever(handle);\n}\n\nfunction extractFileName(filePath: string) {\n return filePath.replace(/^.*[\\\\/]/, '');\n}\n\n/** Gets a file name from an import handle. */\nexport function getFileNameFromHandle(handle: ImportFileHandle): string {\n return extractFileName(getFilePathFromHandle(handle));\n}\n"],"names":["assertNever"],"mappings":";;;;AAKA,MAAM,YAAY,GAAG,kBAAkB;AACvC,MAAM,WAAW,GAAG,gBAAgB;AAS9B,SAAU,wBAAwB,CACtC,MAAwB,EAAA;AAExB,IAAA,OAAO,MAAM,CAAC,UAAU,CAAC,YAAY,CAAC;AACxC;AAEM,SAAU,uBAAuB,CAAC,MAAwB,EAAA;AAC9D,IAAA,OAAO,MAAM,CAAC,UAAU,CAAC,WAAW,CAAC;AACvC;AAyGA;AACM,SAAU,qBAAqB,CAAC,MAAwB,EAAA;AAC5D,IAAA,IAAI,uBAAuB,CAAC,MAAM,CAAC,EAAE;QACnC,MAAM,OAAO,GAAG,MAAM,CAAC,KAAK,CAAC,WAAW,CAAC,MAAM,CAAC;QAChD,MAAM,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,kBAAkB,CAAC,OAAO,CAAC,CAAqB;QACxE,OAAO,IAAI,CAAC,IAAI;IAClB;AAAO,SAAA,IAAI,wBAAwB,CAAC,MAAM,CAAC,EAAE;QAC3C,MAAM,OAAO,GAAG,MAAM,CAAC,KAAK,CAAC,YAAY,CAAC,MAAM,CAAC;QACjD,MAAM,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,kBAAkB,CAAC,OAAO,CAAC,CAA0B;QAC7E,OAAO,IAAI,CAAC,SAAS;IACvB;IAEAA,gBAAW,CAAC,MAAM,CAAC;AACrB;AAEA,SAAS,eAAe,CAAC,QAAgB,EAAA;IACvC,OAAO,QAAQ,CAAC,OAAO,CAAC,UAAU,EAAE,EAAE,CAAC;AACzC;AAEA;AACM,SAAU,qBAAqB,CAAC,MAAwB,EAAA;AAC5D,IAAA,OAAO,eAAe,CAAC,qBAAqB,CAAC,MAAM,CAAC,CAAC;AACvD;;;;;;;"}
@@ -89,4 +89,3 @@ export interface LsDriver {
89
89
  export declare function getFilePathFromHandle(handle: ImportFileHandle): string;
90
90
  /** Gets a file name from an import handle. */
91
91
  export declare function getFileNameFromHandle(handle: ImportFileHandle): string;
92
- //# sourceMappingURL=ls.d.ts.map
@@ -0,0 +1,34 @@
1
+ import { assertNever } from '../util.js';
2
+
3
+ const uploadPrefix = 'upload://upload/';
4
+ const indexPrefix = 'index://index/';
5
+ function isImportFileHandleUpload(handle) {
6
+ return handle.startsWith(uploadPrefix);
7
+ }
8
+ function isImportFileHandleIndex(handle) {
9
+ return handle.startsWith(indexPrefix);
10
+ }
11
+ /** Gets a file path from an import handle. */
12
+ function getFilePathFromHandle(handle) {
13
+ if (isImportFileHandleIndex(handle)) {
14
+ const trimmed = handle.slice(indexPrefix.length);
15
+ const data = JSON.parse(decodeURIComponent(trimmed));
16
+ return data.path;
17
+ }
18
+ else if (isImportFileHandleUpload(handle)) {
19
+ const trimmed = handle.slice(uploadPrefix.length);
20
+ const data = JSON.parse(decodeURIComponent(trimmed));
21
+ return data.localPath;
22
+ }
23
+ assertNever(handle);
24
+ }
25
+ function extractFileName(filePath) {
26
+ return filePath.replace(/^.*[\\/]/, '');
27
+ }
28
+ /** Gets a file name from an import handle. */
29
+ function getFileNameFromHandle(handle) {
30
+ return extractFileName(getFilePathFromHandle(handle));
31
+ }
32
+
33
+ export { getFileNameFromHandle, getFilePathFromHandle, isImportFileHandleIndex, isImportFileHandleUpload };
34
+ //# sourceMappingURL=ls.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"ls.js","sources":["../../src/drivers/ls.ts"],"sourcesContent":["import { assertNever } from '../util';\nimport type { Branded } from '../branding';\nimport type { TableRange } from './pframe';\nimport type { FileLike } from './interfaces';\n\nconst uploadPrefix = 'upload://upload/';\nconst indexPrefix = 'index://index/';\n\nexport type ImportFileHandleUpload = `upload://upload/${string}`;\nexport type ImportFileHandleIndex = `index://index/${string}`;\n\nexport type ImportFileHandle = ImportFileHandleUpload | ImportFileHandleIndex;\n\nexport type LocalImportFileHandle = Branded<ImportFileHandle, 'Local'>;\n\nexport function isImportFileHandleUpload(\n handle: ImportFileHandle,\n): handle is ImportFileHandleUpload {\n return handle.startsWith(uploadPrefix);\n}\n\nexport function isImportFileHandleIndex(handle: ImportFileHandle): handle is ImportFileHandleIndex {\n return handle.startsWith(indexPrefix);\n}\n\n/** Results in upload */\nexport type StorageHandleLocal = `local://${string}`;\n\n/** Results in index */\nexport type StorageHandleRemote = `remote://${string}`;\n\nexport type StorageHandle = StorageHandleLocal | StorageHandleRemote;\n\nexport type StorageEntry = {\n name: string;\n handle: StorageHandle;\n initialFullPath: string;\n\n // TODO\n // pathStartsWithDisk\n};\n\nexport type ListFilesResult = {\n parent?: string;\n entries: LsEntry[];\n};\n\nexport type LsEntry =\n | {\n type: 'dir';\n name: string;\n fullPath: string;\n }\n | {\n type: 'file';\n name: string;\n fullPath: string;\n\n /** This handle should be set to args... */\n handle: ImportFileHandle;\n };\n\nexport type OpenDialogFilter = {\n /** Human-readable file type name */\n readonly name: string;\n /** File extensions */\n readonly extensions: string[];\n};\n\nexport type OpenDialogOps = {\n /** Open dialog window title */\n readonly title?: string;\n /** Custom label for the confirmation button, when left empty the default label will be used. */\n readonly buttonLabel?: string;\n /** Limits of file types user can select */\n readonly filters?: OpenDialogFilter[];\n};\n\nexport type OpenSingleFileResponse = {\n /** Contains local file handle, allowing file importing or content reading. If user canceled\n * the dialog, field will be undefined. */\n readonly file?: LocalImportFileHandle;\n};\n\nexport type OpenMultipleFilesResponse = {\n /** Contains local file handles, allowing file importing or content reading. If user canceled\n * the dialog, field will be undefined. */\n readonly files?: LocalImportFileHandle[];\n};\n\n/** Can be used to limit request for local file content to a certain bytes range */\nexport type FileRange = {\n /** From byte index (inclusive) */\n readonly from: number;\n /** To byte index (exclusive) */\n readonly to: number;\n};\n\nexport interface LsDriver {\n /** remote and local storages */\n getStorageList(): Promise<StorageEntry[]>;\n\n listFiles(storage: StorageHandle, fullPath: string): Promise<ListFilesResult>;\n\n /** Opens system file open dialog allowing to select single file and awaits user action */\n showOpenSingleFileDialog(ops: OpenDialogOps): Promise<OpenSingleFileResponse>;\n\n /** Opens system file open dialog allowing to multiple files and awaits user action */\n showOpenMultipleFilesDialog(ops: OpenDialogOps): Promise<OpenMultipleFilesResponse>;\n\n /** Given a handle to a local file, allows to get file size */\n getLocalFileSize(file: LocalImportFileHandle): Promise<number>;\n\n /** Given a handle to a local file, allows to get its content */\n getLocalFileContent(file: LocalImportFileHandle, range?: TableRange): Promise<Uint8Array>;\n\n /**\n * Resolves browser's File object into platforma's import file handle.\n *\n * This method is useful among other things for implementation of UI\n * components, that handle file Drag&Drop.\n * */\n fileToImportHandle(file: FileLike): Promise<ImportFileHandle>;\n\n /** Saves currently opened block webview as a PDF. */\n exportToPdf?(): Promise<void>;\n}\n\n/** Gets a file path from an import handle. */\nexport function getFilePathFromHandle(handle: ImportFileHandle): string {\n if (isImportFileHandleIndex(handle)) {\n const trimmed = handle.slice(indexPrefix.length);\n const data = JSON.parse(decodeURIComponent(trimmed)) as { path: string };\n return data.path;\n } else if (isImportFileHandleUpload(handle)) {\n const trimmed = handle.slice(uploadPrefix.length);\n const data = JSON.parse(decodeURIComponent(trimmed)) as { localPath: string };\n return data.localPath;\n }\n\n assertNever(handle);\n}\n\nfunction extractFileName(filePath: string) {\n return filePath.replace(/^.*[\\\\/]/, '');\n}\n\n/** Gets a file name from an import handle. */\nexport function getFileNameFromHandle(handle: ImportFileHandle): string {\n return extractFileName(getFilePathFromHandle(handle));\n}\n"],"names":[],"mappings":";;AAKA,MAAM,YAAY,GAAG,kBAAkB;AACvC,MAAM,WAAW,GAAG,gBAAgB;AAS9B,SAAU,wBAAwB,CACtC,MAAwB,EAAA;AAExB,IAAA,OAAO,MAAM,CAAC,UAAU,CAAC,YAAY,CAAC;AACxC;AAEM,SAAU,uBAAuB,CAAC,MAAwB,EAAA;AAC9D,IAAA,OAAO,MAAM,CAAC,UAAU,CAAC,WAAW,CAAC;AACvC;AAyGA;AACM,SAAU,qBAAqB,CAAC,MAAwB,EAAA;AAC5D,IAAA,IAAI,uBAAuB,CAAC,MAAM,CAAC,EAAE;QACnC,MAAM,OAAO,GAAG,MAAM,CAAC,KAAK,CAAC,WAAW,CAAC,MAAM,CAAC;QAChD,MAAM,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,kBAAkB,CAAC,OAAO,CAAC,CAAqB;QACxE,OAAO,IAAI,CAAC,IAAI;IAClB;AAAO,SAAA,IAAI,wBAAwB,CAAC,MAAM,CAAC,EAAE;QAC3C,MAAM,OAAO,GAAG,MAAM,CAAC,KAAK,CAAC,YAAY,CAAC,MAAM,CAAC;QACjD,MAAM,IAAI,GAAG,IAAI,CAAC,KAAK,CAAC,kBAAkB,CAAC,OAAO,CAAC,CAA0B;QAC7E,OAAO,IAAI,CAAC,SAAS;IACvB;IAEA,WAAW,CAAC,MAAM,CAAC;AACrB;AAEA,SAAS,eAAe,CAAC,QAAgB,EAAA;IACvC,OAAO,QAAQ,CAAC,OAAO,CAAC,UAAU,EAAE,EAAE,CAAC;AACzC;AAEA;AACM,SAAU,qBAAqB,CAAC,MAAwB,EAAA;AAC5D,IAAA,OAAO,eAAe,CAAC,qBAAqB,CAAC,MAAM,CAAC,CAAC;AACvD;;;;"}
@@ -15,4 +15,3 @@ export interface ColumnFilter {
15
15
  * pattern. */
16
16
  readonly annotationPattern?: Record<string, string>;
17
17
  }
18
- //# sourceMappingURL=column_filter.d.ts.map
@@ -0,0 +1,275 @@
1
+ 'use strict';
2
+
3
+ var util = require('../../util.cjs');
4
+
5
+ /**
6
+ * Type guard function that checks if the given value is a valid DataInfo.
7
+ *
8
+ * @param value - The value to check
9
+ * @returns True if the value is a valid DataInfo, false otherwise
10
+ */
11
+ function isDataInfo(value) {
12
+ if (!value || typeof value !== 'object') {
13
+ return false;
14
+ }
15
+ const data = value;
16
+ if (!('type' in data)) {
17
+ return false;
18
+ }
19
+ switch (data.type) {
20
+ case 'Json':
21
+ return (typeof data.keyLength === 'number'
22
+ && data.data !== undefined
23
+ && typeof data.data === 'object');
24
+ case 'JsonPartitioned':
25
+ case 'BinaryPartitioned':
26
+ case 'ParquetPartitioned':
27
+ return (typeof data.partitionKeyLength === 'number'
28
+ && data.parts !== undefined
29
+ && typeof data.parts === 'object');
30
+ default:
31
+ return false;
32
+ }
33
+ }
34
+ function mapDataInfo(dataInfo, mapFn) {
35
+ if (dataInfo === undefined) {
36
+ return undefined;
37
+ }
38
+ switch (dataInfo.type) {
39
+ case 'Json':
40
+ // Json type doesn't contain blobs, so return as is
41
+ return dataInfo;
42
+ case 'JsonPartitioned': {
43
+ // Map each blob in parts
44
+ const newParts = {};
45
+ for (const [key, blob] of Object.entries(dataInfo.parts)) {
46
+ newParts[key] = mapFn(blob);
47
+ }
48
+ return {
49
+ ...dataInfo,
50
+ parts: newParts,
51
+ };
52
+ }
53
+ case 'BinaryPartitioned': {
54
+ // Map each index and values blob in parts
55
+ const newParts = {};
56
+ for (const [key, chunk] of Object.entries(dataInfo.parts)) {
57
+ newParts[key] = {
58
+ index: mapFn(chunk.index),
59
+ values: mapFn(chunk.values),
60
+ };
61
+ }
62
+ return {
63
+ ...dataInfo,
64
+ parts: newParts,
65
+ };
66
+ }
67
+ case 'ParquetPartitioned': {
68
+ // Map each blob in parts
69
+ const newParts = {};
70
+ for (const [key, blob] of Object.entries(dataInfo.parts)) {
71
+ newParts[key] = mapFn(blob);
72
+ }
73
+ return {
74
+ ...dataInfo,
75
+ parts: newParts,
76
+ };
77
+ }
78
+ }
79
+ }
80
+ /**
81
+ * @param dataInfo - The source DataInfo object
82
+ * @param cb - Callback, function that have access to every blob to visit them all
83
+ * @returns Nothing
84
+ */
85
+ function visitDataInfo(dataInfo, cb) {
86
+ switch (dataInfo.type) {
87
+ case 'Json':
88
+ // Json type doesn't contain blobs, so return as is
89
+ break;
90
+ case 'JsonPartitioned': {
91
+ // Visit each blob in parts
92
+ Object.values(dataInfo.parts).forEach(cb);
93
+ break;
94
+ }
95
+ case 'BinaryPartitioned': {
96
+ // Visit each index and values blob in parts
97
+ Object.values(dataInfo.parts).forEach((chunk) => {
98
+ cb(chunk.index);
99
+ cb(chunk.values);
100
+ });
101
+ break;
102
+ }
103
+ case 'ParquetPartitioned': {
104
+ // Visit each blob in parts
105
+ Object.values(dataInfo.parts).forEach(cb);
106
+ break;
107
+ }
108
+ }
109
+ }
110
+ /**
111
+ * Type guard function that checks if the given value is a valid DataInfoEntries.
112
+ *
113
+ * @param value - The value to check
114
+ * @returns True if the value is a valid DataInfoEntries, false otherwise
115
+ */
116
+ function isDataInfoEntries(value) {
117
+ if (!value || typeof value !== 'object') {
118
+ return false;
119
+ }
120
+ const data = value;
121
+ if (!('type' in data)) {
122
+ return false;
123
+ }
124
+ switch (data.type) {
125
+ case 'Json':
126
+ return (typeof data.keyLength === 'number'
127
+ && Array.isArray(data.data));
128
+ case 'JsonPartitioned':
129
+ case 'BinaryPartitioned':
130
+ case 'ParquetPartitioned':
131
+ return (typeof data.partitionKeyLength === 'number'
132
+ && Array.isArray(data.parts));
133
+ default:
134
+ return false;
135
+ }
136
+ }
137
+ /**
138
+ * Type guard function that checks if the given value is a valid PartitionedDataInfoEntries.
139
+ *
140
+ * @template Blob - Type parameter representing the storage reference type
141
+ * @param value - The value to check
142
+ * @returns True if the value is a valid PartitionedDataInfoEntries, false otherwise
143
+ */
144
+ function isPartitionedDataInfoEntries(value) {
145
+ if (!isDataInfoEntries(value))
146
+ return false;
147
+ switch (value.type) {
148
+ case 'JsonPartitioned':
149
+ case 'BinaryPartitioned':
150
+ case 'ParquetPartitioned':
151
+ return true;
152
+ default:
153
+ return false;
154
+ }
155
+ }
156
+ /**
157
+ * Converts DataInfo to DataInfoEntries
158
+ *
159
+ * @param dataInfo - The record-based DataInfo object
160
+ * @returns The equivalent entry-based DataInfoEntries object
161
+ */
162
+ function dataInfoToEntries(dataInfo) {
163
+ switch (dataInfo.type) {
164
+ case 'Json': return {
165
+ type: 'Json',
166
+ keyLength: dataInfo.keyLength,
167
+ data: Object.entries(dataInfo.data).map(([keyStr, value]) => {
168
+ const key = JSON.parse(keyStr);
169
+ return { key, value };
170
+ }),
171
+ };
172
+ case 'JsonPartitioned': return {
173
+ type: 'JsonPartitioned',
174
+ partitionKeyLength: dataInfo.partitionKeyLength,
175
+ parts: Object.entries(dataInfo.parts).map(([keyStr, blob]) => {
176
+ const key = JSON.parse(keyStr);
177
+ return { key, value: blob };
178
+ }),
179
+ };
180
+ case 'BinaryPartitioned': return {
181
+ type: 'BinaryPartitioned',
182
+ partitionKeyLength: dataInfo.partitionKeyLength,
183
+ parts: Object.entries(dataInfo.parts).map(([keyStr, chunk]) => {
184
+ const key = JSON.parse(keyStr);
185
+ return { key, value: chunk };
186
+ }),
187
+ };
188
+ case 'ParquetPartitioned': return {
189
+ type: 'ParquetPartitioned',
190
+ partitionKeyLength: dataInfo.partitionKeyLength,
191
+ parts: Object.entries(dataInfo.parts).map(([keyStr, blob]) => {
192
+ const key = JSON.parse(keyStr);
193
+ return { key, value: blob };
194
+ }),
195
+ };
196
+ default:
197
+ util.assertNever(dataInfo);
198
+ }
199
+ }
200
+ /**
201
+ * Converts DataInfoEntries to DataInfo
202
+ *
203
+ * @param dataInfoEntries - The entry-based DataInfoEntries object
204
+ * @returns The equivalent record-based DataInfo object
205
+ */
206
+ function entriesToDataInfo(dataInfoEntries) {
207
+ switch (dataInfoEntries.type) {
208
+ case 'Json': return {
209
+ type: 'Json',
210
+ keyLength: dataInfoEntries.keyLength,
211
+ data: Object.fromEntries(dataInfoEntries.data.map(({ key, value }) => [JSON.stringify(key), value])),
212
+ };
213
+ case 'JsonPartitioned': return {
214
+ type: 'JsonPartitioned',
215
+ partitionKeyLength: dataInfoEntries.partitionKeyLength,
216
+ parts: Object.fromEntries(dataInfoEntries.parts.map(({ key, value }) => [JSON.stringify(key), value])),
217
+ };
218
+ case 'BinaryPartitioned': return {
219
+ type: 'BinaryPartitioned',
220
+ partitionKeyLength: dataInfoEntries.partitionKeyLength,
221
+ parts: Object.fromEntries(dataInfoEntries.parts.map(({ key, value }) => [JSON.stringify(key), value])),
222
+ };
223
+ case 'ParquetPartitioned': return {
224
+ type: 'ParquetPartitioned',
225
+ partitionKeyLength: dataInfoEntries.partitionKeyLength,
226
+ parts: Object.fromEntries(dataInfoEntries.parts.map(({ key, value }) => [JSON.stringify(key), value])),
227
+ };
228
+ default:
229
+ util.assertNever(dataInfoEntries);
230
+ }
231
+ }
232
+ function mapDataInfoEntries(dataInfoEntries, mapFn) {
233
+ if (dataInfoEntries === undefined) {
234
+ return undefined;
235
+ }
236
+ switch (dataInfoEntries.type) {
237
+ case 'Json':
238
+ // Json type doesn't contain blobs, so return as is
239
+ return dataInfoEntries;
240
+ case 'JsonPartitioned': return {
241
+ ...dataInfoEntries,
242
+ parts: dataInfoEntries.parts.map((entry) => ({
243
+ key: entry.key,
244
+ value: mapFn(entry.value),
245
+ })),
246
+ };
247
+ case 'BinaryPartitioned': return {
248
+ ...dataInfoEntries,
249
+ parts: dataInfoEntries.parts.map((entry) => ({
250
+ key: entry.key,
251
+ value: {
252
+ index: mapFn(entry.value.index),
253
+ values: mapFn(entry.value.values),
254
+ },
255
+ })),
256
+ };
257
+ case 'ParquetPartitioned': return {
258
+ ...dataInfoEntries,
259
+ parts: dataInfoEntries.parts.map((entry) => ({
260
+ key: entry.key,
261
+ value: mapFn(entry.value),
262
+ })),
263
+ };
264
+ }
265
+ }
266
+
267
+ exports.dataInfoToEntries = dataInfoToEntries;
268
+ exports.entriesToDataInfo = entriesToDataInfo;
269
+ exports.isDataInfo = isDataInfo;
270
+ exports.isDataInfoEntries = isDataInfoEntries;
271
+ exports.isPartitionedDataInfoEntries = isPartitionedDataInfoEntries;
272
+ exports.mapDataInfo = mapDataInfo;
273
+ exports.mapDataInfoEntries = mapDataInfoEntries;
274
+ exports.visitDataInfo = visitDataInfo;
275
+ //# sourceMappingURL=data_info.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"data_info.cjs","sources":["../../../src/drivers/pframe/data_info.ts"],"sourcesContent":["import { assertNever } from '../../util';\n\n/**\n * Represents a JavaScript representation of a value in a PColumn. Can be null, a number, or a string.\n * These are the primitive types that can be stored directly in PColumns.\n *\n * Note: Actual columns can hold more value types, which are converted to these JavaScript types\n * once they enter the JavaScript runtime.\n */\nexport type PColumnValue = null | number | string;\n\n/**\n * Represents a key for a PColumn value.\n * Can be an array of strings or numbers.\n */\nexport type PColumnKey = (number | string)[];\n\n/**\n * Represents a single entry in a PColumn's data structure.\n * Contains a key and a value.\n */\nexport type PColumnDataEntry<T> = {\n /** Key for the value */\n key: PColumnKey;\n\n /** Value / blob at the given key */\n value: T;\n};\n\n/**\n * Represents column data stored as a simple JSON structure.\n * Used for small datasets that can be efficiently stored directly in memory.\n */\nexport type JsonDataInfo = {\n /** Identifier for this data format ('Json') */\n type: 'Json';\n\n /** Number of axes that make up the complete key (tuple length) */\n keyLength: number;\n\n /**\n * Key-value pairs where keys are stringified tuples of axis values\n * and values are the column values for those coordinates\n */\n data: Record<string, PColumnValue>;\n};\n\n/**\n * Represents column data partitioned across multiple JSON blobs.\n * Used for larger datasets that need to be split into manageable chunks.\n */\nexport type JsonPartitionedDataInfo<Blob> = {\n /** Identifier for this data format ('JsonPartitioned') */\n type: 'JsonPartitioned';\n\n /** Number of leading axes used for partitioning */\n partitionKeyLength: number;\n\n /** Map of stringified partition keys to blob references */\n parts: Record<string, Blob>;\n};\n\n/**\n * Represents a binary format chunk containing index and values as separate blobs.\n * Used for efficient storage and retrieval of column data in binary format.\n */\nexport type BinaryChunk<Blob> = {\n /** Binary blob containing structured index information */\n index: Blob;\n\n /** Binary blob containing the actual values */\n values: Blob;\n};\n\n/**\n * Represents column data partitioned across multiple binary chunks.\n * Optimized for efficient storage and retrieval of large datasets.\n */\nexport type BinaryPartitionedDataInfo<Blob> = {\n /** Identifier for this data format ('BinaryPartitioned') */\n type: 'BinaryPartitioned';\n\n /** Number of leading axes used for partitioning */\n partitionKeyLength: number;\n\n /** Map of stringified partition keys to binary chunks */\n parts: Record<string, BinaryChunk<Blob>>;\n};\n\nexport type ParquetChunkMappingAxis = {\n /** Data type (matches PColumn axis types) */\n type: 'Int' | 'Long' | 'String';\n\n /** Field name in the Parquet file */\n id: string;\n};\n\nexport type ParquetChunkMappingColumn = {\n /** Data type (matches PColumn value type) */\n type: 'Int' | 'Long' | 'Float' | 'Double' | 'String';\n\n /** Field name in the Parquet file */\n id: string;\n};\n\nexport type ParquetChunkMapping = {\n /** Axes mappings - Parquet file is sorted by these fields in this order */\n axes: ParquetChunkMappingAxis[];\n\n /** Column mapping */\n column: ParquetChunkMappingColumn;\n};\n\nexport type ParquetChunkStats = {\n /** Number of rows in the chunk */\n numberOfRows: number;\n /** Byte size information for storage optimization and query planning */\n size: {\n /** Byte sizes for each axis column in the same order as axes mapping */\n axes: number[];\n /** Byte size for the data column */\n column: number;\n };\n};\n\nexport type ParquetChunkMetadata = {\n /** Content hash calculated for the specific axes and data this chunk represents */\n dataDigest: string;\n\n /** Pre-computed statistics for optimization without blob download */\n stats: Partial<ParquetChunkStats>;\n};\n\nexport type ParquetChunk<Blob> = {\n /** Parquet file (PTable) containing column data */\n data: Blob;\n} & ParquetChunkMapping & Partial<ParquetChunkMetadata>;\n\nexport type ParquetPartitionedDataInfo<Blob> = {\n /** Identifier for this data format ('ParquetPartitioned') */\n type: 'ParquetPartitioned';\n\n /** Number of leading axes used for partitioning */\n partitionKeyLength: number;\n\n /** Map of stringified partition keys to parquet files */\n parts: Record<string, Blob>;\n};\n\n/**\n * Union type representing all possible data storage formats for PColumn data.\n * The specific format used depends on data size, access patterns, and performance requirements.\n *\n * @template Blob - Type parameter representing the storage reference type (could be ResourceInfo, PFrameBlobId, etc.)\n */\nexport type DataInfo<Blob> =\n | JsonDataInfo\n | JsonPartitionedDataInfo<Blob>\n | BinaryPartitionedDataInfo<Blob>\n | ParquetPartitionedDataInfo<Blob>;\n\n/**\n * Type guard function that checks if the given value is a valid DataInfo.\n *\n * @param value - The value to check\n * @returns True if the value is a valid DataInfo, false otherwise\n */\nexport function isDataInfo<Blob>(value: unknown): value is DataInfo<Blob> {\n if (!value || typeof value !== 'object') {\n return false;\n }\n\n const data = value as Record<string, unknown>;\n if (!('type' in data)) {\n return false;\n }\n\n switch (data.type) {\n case 'Json':\n return (\n typeof data.keyLength === 'number'\n && data.data !== undefined\n && typeof data.data === 'object'\n );\n case 'JsonPartitioned':\n case 'BinaryPartitioned':\n case 'ParquetPartitioned':\n return (\n typeof data.partitionKeyLength === 'number'\n && data.parts !== undefined\n && typeof data.parts === 'object'\n );\n default:\n return false;\n }\n}\n\n/**\n * Maps blob references in a DataInfo object from one type to another using a mapping function.\n *\n * @template B1 - Source blob type\n * @template B2 - Target blob type\n * @param dataInfo - The source DataInfo object\n * @param mapFn - Function to transform blobs from type B1 to type B2\n * @returns A new DataInfo object with transformed blob references\n */\nexport function mapDataInfo<B1, B2>(\n dataInfo: ParquetPartitionedDataInfo<B1>,\n mapFn: (blob: B1) => B2,\n): ParquetPartitionedDataInfo<B2>;\nexport function mapDataInfo<B1, B2>(\n dataInfo: Exclude<DataInfo<B1>, ParquetPartitionedDataInfo<B1>>,\n mapFn: (blob: B1) => B2,\n): Exclude<DataInfo<B2>, ParquetPartitionedDataInfo<B2>>;\nexport function mapDataInfo<B1, B2>(\n dataInfo: DataInfo<B1>,\n mapFn: (blob: B1) => B2,\n): DataInfo<B2>;\nexport function mapDataInfo<B1, B2>(\n dataInfo: DataInfo<B1> | undefined,\n mapFn: (blob: B1) => B2,\n): DataInfo<B2> | undefined {\n if (dataInfo === undefined) {\n return undefined;\n }\n\n switch (dataInfo.type) {\n case 'Json':\n // Json type doesn't contain blobs, so return as is\n return dataInfo;\n case 'JsonPartitioned': {\n // Map each blob in parts\n const newParts: Record<string, B2> = {};\n for (const [key, blob] of Object.entries(dataInfo.parts)) {\n newParts[key] = mapFn(blob);\n }\n return {\n ...dataInfo,\n parts: newParts,\n };\n }\n case 'BinaryPartitioned': {\n // Map each index and values blob in parts\n const newParts: Record<string, BinaryChunk<B2>> = {};\n for (const [key, chunk] of Object.entries(dataInfo.parts)) {\n newParts[key] = {\n index: mapFn(chunk.index),\n values: mapFn(chunk.values),\n };\n }\n return {\n ...dataInfo,\n parts: newParts,\n };\n }\n case 'ParquetPartitioned': {\n // Map each blob in parts\n const newParts: Record<string, B2> = {};\n for (const [key, blob] of Object.entries(dataInfo.parts)) {\n newParts[key] = mapFn(blob);\n }\n return {\n ...dataInfo,\n parts: newParts,\n };\n }\n }\n}\n\n/**\n * @param dataInfo - The source DataInfo object\n * @param cb - Callback, function that have access to every blob to visit them all\n * @returns Nothing\n */\nexport function visitDataInfo<B>(\n dataInfo: DataInfo<B>,\n cb: (blob: B) => void,\n): void {\n switch (dataInfo.type) {\n case 'Json':\n // Json type doesn't contain blobs, so return as is\n break;\n case 'JsonPartitioned': {\n // Visit each blob in parts\n Object.values(dataInfo.parts).forEach(cb);\n break;\n }\n case 'BinaryPartitioned': {\n // Visit each index and values blob in parts\n Object.values(dataInfo.parts).forEach((chunk) => {\n cb(chunk.index);\n cb(chunk.values);\n });\n break;\n }\n case 'ParquetPartitioned': {\n // Visit each blob in parts\n Object.values(dataInfo.parts).forEach(cb);\n break;\n }\n }\n}\n\n//\n// Lightway representation for ExplicitJsonData\n//\n\n/**\n * Represents a single key-value entry in a column's explicit data structure.\n * Used when directly instantiating PColumns with explicit data.\n */\nexport type PColumnValuesEntry = {\n key: PColumnKey;\n val: PColumnValue;\n};\n\n/**\n * Array of key-value entries representing explicit column data.\n * Used for lightweight explicit instantiation of PColumns.\n */\nexport type PColumnValues = PColumnValuesEntry[];\n\n/**\n * Entry-based representation of JsonDataInfo\n */\nexport interface JsonDataInfoEntries {\n type: 'Json';\n keyLength: number;\n data: PColumnDataEntry<PColumnValue>[];\n}\n\n/**\n * Entry-based representation of JsonPartitionedDataInfo\n */\nexport interface JsonPartitionedDataInfoEntries<Blob> {\n type: 'JsonPartitioned';\n partitionKeyLength: number;\n parts: PColumnDataEntry<Blob>[];\n}\n\n/**\n * Entry-based representation of BinaryPartitionedDataInfo\n */\nexport interface BinaryPartitionedDataInfoEntries<Blob> {\n type: 'BinaryPartitioned';\n partitionKeyLength: number;\n parts: PColumnDataEntry<BinaryChunk<Blob>>[];\n}\n\n/**\n * Entry-based representation of ParquetPartitionedDataInfo\n */\nexport interface ParquetPartitionedDataInfoEntries<Blob> {\n type: 'ParquetPartitioned';\n partitionKeyLength: number;\n parts: PColumnDataEntry<Blob>[];\n}\n/**\n * Union type representing all possible entry-based partitioned data storage formats\n */\nexport type PartitionedDataInfoEntries<Blob> =\n | JsonPartitionedDataInfoEntries<Blob>\n | BinaryPartitionedDataInfoEntries<Blob>\n | ParquetPartitionedDataInfoEntries<Blob>;\n\n/**\n * Union type representing all possible entry-based data storage formats\n */\nexport type DataInfoEntries<Blob> =\n | JsonDataInfoEntries\n | PartitionedDataInfoEntries<Blob>;\n\n/**\n * Type guard function that checks if the given value is a valid DataInfoEntries.\n *\n * @param value - The value to check\n * @returns True if the value is a valid DataInfoEntries, false otherwise\n */\nexport function isDataInfoEntries<Blob>(value: unknown): value is DataInfoEntries<Blob> {\n if (!value || typeof value !== 'object') {\n return false;\n }\n\n const data = value as Record<string, unknown>;\n if (!('type' in data)) {\n return false;\n }\n\n switch (data.type) {\n case 'Json':\n return (\n typeof data.keyLength === 'number'\n && Array.isArray(data.data)\n );\n case 'JsonPartitioned':\n case 'BinaryPartitioned':\n case 'ParquetPartitioned':\n return (\n typeof data.partitionKeyLength === 'number'\n && Array.isArray(data.parts)\n );\n default:\n return false;\n }\n}\n\n/**\n * Type guard function that checks if the given value is a valid PartitionedDataInfoEntries.\n *\n * @template Blob - Type parameter representing the storage reference type\n * @param value - The value to check\n * @returns True if the value is a valid PartitionedDataInfoEntries, false otherwise\n */\nexport function isPartitionedDataInfoEntries<Blob>(value: unknown): value is PartitionedDataInfoEntries<Blob> {\n if (!isDataInfoEntries(value)) return false;\n switch (value.type) {\n case 'JsonPartitioned':\n case 'BinaryPartitioned':\n case 'ParquetPartitioned':\n return true;\n default:\n return false;\n }\n}\n\n/**\n * Converts DataInfo to DataInfoEntries\n *\n * @param dataInfo - The record-based DataInfo object\n * @returns The equivalent entry-based DataInfoEntries object\n */\nexport function dataInfoToEntries<Blob>(dataInfo: DataInfo<Blob>): DataInfoEntries<Blob> {\n switch (dataInfo.type) {\n case 'Json': return {\n type: 'Json',\n keyLength: dataInfo.keyLength,\n data: Object.entries(dataInfo.data).map(([keyStr, value]) => {\n const key = JSON.parse(keyStr) as PColumnKey;\n return { key, value } as PColumnDataEntry<PColumnValue>;\n }),\n };\n case 'JsonPartitioned': return {\n type: 'JsonPartitioned',\n partitionKeyLength: dataInfo.partitionKeyLength,\n parts: Object.entries(dataInfo.parts).map(([keyStr, blob]) => {\n const key = JSON.parse(keyStr) as PColumnKey;\n return { key, value: blob } as PColumnDataEntry<Blob>;\n }),\n };\n case 'BinaryPartitioned': return {\n type: 'BinaryPartitioned',\n partitionKeyLength: dataInfo.partitionKeyLength,\n parts: Object.entries(dataInfo.parts).map(([keyStr, chunk]) => {\n const key = JSON.parse(keyStr) as PColumnKey;\n return { key, value: chunk } as PColumnDataEntry<BinaryChunk<Blob>>;\n }),\n };\n case 'ParquetPartitioned': return {\n type: 'ParquetPartitioned',\n partitionKeyLength: dataInfo.partitionKeyLength,\n parts: Object.entries(dataInfo.parts).map(([keyStr, blob]) => {\n const key = JSON.parse(keyStr) as PColumnKey;\n return { key, value: blob } as PColumnDataEntry<Blob>;\n }),\n };\n default:\n assertNever(dataInfo);\n }\n}\n\n/**\n * Converts DataInfoEntries to DataInfo\n *\n * @param dataInfoEntries - The entry-based DataInfoEntries object\n * @returns The equivalent record-based DataInfo object\n */\nexport function entriesToDataInfo<Blob>(dataInfoEntries: DataInfoEntries<Blob>): DataInfo<Blob> {\n switch (dataInfoEntries.type) {\n case 'Json': return {\n type: 'Json',\n keyLength: dataInfoEntries.keyLength,\n data: Object.fromEntries(\n dataInfoEntries.data.map(({ key, value }) => [JSON.stringify(key), value]),\n ),\n };\n case 'JsonPartitioned': return {\n type: 'JsonPartitioned',\n partitionKeyLength: dataInfoEntries.partitionKeyLength,\n parts: Object.fromEntries(\n dataInfoEntries.parts.map(({ key, value }) => [JSON.stringify(key), value]),\n ),\n };\n case 'BinaryPartitioned': return {\n type: 'BinaryPartitioned',\n partitionKeyLength: dataInfoEntries.partitionKeyLength,\n parts: Object.fromEntries(\n dataInfoEntries.parts.map(({ key, value }) => [JSON.stringify(key), value]),\n ),\n };\n case 'ParquetPartitioned': return {\n type: 'ParquetPartitioned',\n partitionKeyLength: dataInfoEntries.partitionKeyLength,\n parts: Object.fromEntries(\n dataInfoEntries.parts.map(({ key, value }) => [JSON.stringify(key), value]),\n ),\n };\n default:\n assertNever(dataInfoEntries);\n }\n}\n\n/**\n * Maps blob references in a DataInfoEntries object from one type to another using a mapping function.\n *\n * @template B1 - Source blob type\n * @template B2 - Target blob type\n * @param dataInfoEntries - The source DataInfoEntries object\n * @param mapFn - Function to transform blobs from type B1 to type B2\n * @returns A new DataInfoEntries object with transformed blob references\n */\nexport function mapDataInfoEntries<B1, B2>(\n dataInfoEntries: DataInfoEntries<B1>,\n mapFn: (blob: B1) => B2,\n): DataInfoEntries<B2>;\nexport function mapDataInfoEntries<B1, B2>(\n dataInfoEntries: DataInfoEntries<B1> | undefined,\n mapFn: (blob: B1) => B2,\n): DataInfoEntries<B2> | undefined {\n if (dataInfoEntries === undefined) {\n return undefined;\n }\n\n switch (dataInfoEntries.type) {\n case 'Json':\n // Json type doesn't contain blobs, so return as is\n return dataInfoEntries;\n case 'JsonPartitioned': return {\n ...dataInfoEntries,\n parts: dataInfoEntries.parts.map((entry) => ({\n key: entry.key,\n value: mapFn(entry.value),\n })),\n };\n case 'BinaryPartitioned': return {\n ...dataInfoEntries,\n parts: dataInfoEntries.parts.map((entry) => ({\n key: entry.key,\n value: {\n index: mapFn(entry.value.index),\n values: mapFn(entry.value.values),\n },\n })),\n };\n case 'ParquetPartitioned': return {\n ...dataInfoEntries,\n parts: dataInfoEntries.parts.map((entry) => ({\n key: entry.key,\n value: mapFn(entry.value),\n })),\n };\n }\n}\n"],"names":["assertNever"],"mappings":";;;;AAiKA;;;;;AAKG;AACG,SAAU,UAAU,CAAO,KAAc,EAAA;IAC7C,IAAI,CAAC,KAAK,IAAI,OAAO,KAAK,KAAK,QAAQ,EAAE;AACvC,QAAA,OAAO,KAAK;IACd;IAEA,MAAM,IAAI,GAAG,KAAgC;AAC7C,IAAA,IAAI,EAAE,MAAM,IAAI,IAAI,CAAC,EAAE;AACrB,QAAA,OAAO,KAAK;IACd;AAEA,IAAA,QAAQ,IAAI,CAAC,IAAI;AACf,QAAA,KAAK,MAAM;AACT,YAAA,QACE,OAAO,IAAI,CAAC,SAAS,KAAK;mBACvB,IAAI,CAAC,IAAI,KAAK;AACd,mBAAA,OAAO,IAAI,CAAC,IAAI,KAAK,QAAQ;AAEpC,QAAA,KAAK,iBAAiB;AACtB,QAAA,KAAK,mBAAmB;AACxB,QAAA,KAAK,oBAAoB;AACvB,YAAA,QACE,OAAO,IAAI,CAAC,kBAAkB,KAAK;mBAChC,IAAI,CAAC,KAAK,KAAK;AACf,mBAAA,OAAO,IAAI,CAAC,KAAK,KAAK,QAAQ;AAErC,QAAA;AACE,YAAA,OAAO,KAAK;;AAElB;AAuBM,SAAU,WAAW,CACzB,QAAkC,EAClC,KAAuB,EAAA;AAEvB,IAAA,IAAI,QAAQ,KAAK,SAAS,EAAE;AAC1B,QAAA,OAAO,SAAS;IAClB;AAEA,IAAA,QAAQ,QAAQ,CAAC,IAAI;AACnB,QAAA,KAAK,MAAM;;AAET,YAAA,OAAO,QAAQ;QACjB,KAAK,iBAAiB,EAAE;;YAEtB,MAAM,QAAQ,GAAuB,EAAE;AACvC,YAAA,KAAK,MAAM,CAAC,GAAG,EAAE,IAAI,CAAC,IAAI,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,KAAK,CAAC,EAAE;gBACxD,QAAQ,CAAC,GAAG,CAAC,GAAG,KAAK,CAAC,IAAI,CAAC;YAC7B;YACA,OAAO;AACL,gBAAA,GAAG,QAAQ;AACX,gBAAA,KAAK,EAAE,QAAQ;aAChB;QACH;QACA,KAAK,mBAAmB,EAAE;;YAExB,MAAM,QAAQ,GAAoC,EAAE;AACpD,YAAA,KAAK,MAAM,CAAC,GAAG,EAAE,KAAK,CAAC,IAAI,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,KAAK,CAAC,EAAE;gBACzD,QAAQ,CAAC,GAAG,CAAC,GAAG;AACd,oBAAA,KAAK,EAAE,KAAK,CAAC,KAAK,CAAC,KAAK,CAAC;AACzB,oBAAA,MAAM,EAAE,KAAK,CAAC,KAAK,CAAC,MAAM,CAAC;iBAC5B;YACH;YACA,OAAO;AACL,gBAAA,GAAG,QAAQ;AACX,gBAAA,KAAK,EAAE,QAAQ;aAChB;QACH;QACA,KAAK,oBAAoB,EAAE;;YAEzB,MAAM,QAAQ,GAAuB,EAAE;AACvC,YAAA,KAAK,MAAM,CAAC,GAAG,EAAE,IAAI,CAAC,IAAI,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,KAAK,CAAC,EAAE;gBACxD,QAAQ,CAAC,GAAG,CAAC,GAAG,KAAK,CAAC,IAAI,CAAC;YAC7B;YACA,OAAO;AACL,gBAAA,GAAG,QAAQ;AACX,gBAAA,KAAK,EAAE,QAAQ;aAChB;QACH;;AAEJ;AAEA;;;;AAIG;AACG,SAAU,aAAa,CAC3B,QAAqB,EACrB,EAAqB,EAAA;AAErB,IAAA,QAAQ,QAAQ,CAAC,IAAI;AACnB,QAAA,KAAK,MAAM;;YAET;QACF,KAAK,iBAAiB,EAAE;;AAEtB,YAAA,MAAM,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,OAAO,CAAC,EAAE,CAAC;YACzC;QACF;QACA,KAAK,mBAAmB,EAAE;;AAExB,YAAA,MAAM,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,OAAO,CAAC,CAAC,KAAK,KAAI;AAC9C,gBAAA,EAAE,CAAC,KAAK,CAAC,KAAK,CAAC;AACf,gBAAA,EAAE,CAAC,KAAK,CAAC,MAAM,CAAC;AAClB,YAAA,CAAC,CAAC;YACF;QACF;QACA,KAAK,oBAAoB,EAAE;;AAEzB,YAAA,MAAM,CAAC,MAAM,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,OAAO,CAAC,EAAE,CAAC;YACzC;QACF;;AAEJ;AAuEA;;;;;AAKG;AACG,SAAU,iBAAiB,CAAO,KAAc,EAAA;IACpD,IAAI,CAAC,KAAK,IAAI,OAAO,KAAK,KAAK,QAAQ,EAAE;AACvC,QAAA,OAAO,KAAK;IACd;IAEA,MAAM,IAAI,GAAG,KAAgC;AAC7C,IAAA,IAAI,EAAE,MAAM,IAAI,IAAI,CAAC,EAAE;AACrB,QAAA,OAAO,KAAK;IACd;AAEA,IAAA,QAAQ,IAAI,CAAC,IAAI;AACf,QAAA,KAAK,MAAM;AACT,YAAA,QACE,OAAO,IAAI,CAAC,SAAS,KAAK;mBACvB,KAAK,CAAC,OAAO,CAAC,IAAI,CAAC,IAAI,CAAC;AAE/B,QAAA,KAAK,iBAAiB;AACtB,QAAA,KAAK,mBAAmB;AACxB,QAAA,KAAK,oBAAoB;AACvB,YAAA,QACE,OAAO,IAAI,CAAC,kBAAkB,KAAK;mBAChC,KAAK,CAAC,OAAO,CAAC,IAAI,CAAC,KAAK,CAAC;AAEhC,QAAA;AACE,YAAA,OAAO,KAAK;;AAElB;AAEA;;;;;;AAMG;AACG,SAAU,4BAA4B,CAAO,KAAc,EAAA;AAC/D,IAAA,IAAI,CAAC,iBAAiB,CAAC,KAAK,CAAC;AAAE,QAAA,OAAO,KAAK;AAC3C,IAAA,QAAQ,KAAK,CAAC,IAAI;AAChB,QAAA,KAAK,iBAAiB;AACtB,QAAA,KAAK,mBAAmB;AACxB,QAAA,KAAK,oBAAoB;AACvB,YAAA,OAAO,IAAI;AACb,QAAA;AACE,YAAA,OAAO,KAAK;;AAElB;AAEA;;;;;AAKG;AACG,SAAU,iBAAiB,CAAO,QAAwB,EAAA;AAC9D,IAAA,QAAQ,QAAQ,CAAC,IAAI;QACnB,KAAK,MAAM,EAAE,OAAO;AAClB,YAAA,IAAI,EAAE,MAAM;YACZ,SAAS,EAAE,QAAQ,CAAC,SAAS;AAC7B,YAAA,IAAI,EAAE,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,MAAM,EAAE,KAAK,CAAC,KAAI;gBAC1D,MAAM,GAAG,GAAG,IAAI,CAAC,KAAK,CAAC,MAAM,CAAe;AAC5C,gBAAA,OAAO,EAAE,GAAG,EAAE,KAAK,EAAoC;AACzD,YAAA,CAAC,CAAC;SACH;QACD,KAAK,iBAAiB,EAAE,OAAO;AAC7B,YAAA,IAAI,EAAE,iBAAiB;YACvB,kBAAkB,EAAE,QAAQ,CAAC,kBAAkB;AAC/C,YAAA,KAAK,EAAE,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,MAAM,EAAE,IAAI,CAAC,KAAI;gBAC3D,MAAM,GAAG,GAAG,IAAI,CAAC,KAAK,CAAC,MAAM,CAAe;AAC5C,gBAAA,OAAO,EAAE,GAAG,EAAE,KAAK,EAAE,IAAI,EAA4B;AACvD,YAAA,CAAC,CAAC;SACH;QACD,KAAK,mBAAmB,EAAE,OAAO;AAC/B,YAAA,IAAI,EAAE,mBAAmB;YACzB,kBAAkB,EAAE,QAAQ,CAAC,kBAAkB;AAC/C,YAAA,KAAK,EAAE,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,MAAM,EAAE,KAAK,CAAC,KAAI;gBAC5D,MAAM,GAAG,GAAG,IAAI,CAAC,KAAK,CAAC,MAAM,CAAe;AAC5C,gBAAA,OAAO,EAAE,GAAG,EAAE,KAAK,EAAE,KAAK,EAAyC;AACrE,YAAA,CAAC,CAAC;SACH;QACD,KAAK,oBAAoB,EAAE,OAAO;AAChC,YAAA,IAAI,EAAE,oBAAoB;YAC1B,kBAAkB,EAAE,QAAQ,CAAC,kBAAkB;AAC/C,YAAA,KAAK,EAAE,MAAM,CAAC,OAAO,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,MAAM,EAAE,IAAI,CAAC,KAAI;gBAC3D,MAAM,GAAG,GAAG,IAAI,CAAC,KAAK,CAAC,MAAM,CAAe;AAC5C,gBAAA,OAAO,EAAE,GAAG,EAAE,KAAK,EAAE,IAAI,EAA4B;AACvD,YAAA,CAAC,CAAC;SACH;AACD,QAAA;YACEA,gBAAW,CAAC,QAAQ,CAAC;;AAE3B;AAEA;;;;;AAKG;AACG,SAAU,iBAAiB,CAAO,eAAsC,EAAA;AAC5E,IAAA,QAAQ,eAAe,CAAC,IAAI;QAC1B,KAAK,MAAM,EAAE,OAAO;AAClB,YAAA,IAAI,EAAE,MAAM;YACZ,SAAS,EAAE,eAAe,CAAC,SAAS;AACpC,YAAA,IAAI,EAAE,MAAM,CAAC,WAAW,CACtB,eAAe,CAAC,IAAI,CAAC,GAAG,CAAC,CAAC,EAAE,GAAG,EAAE,KAAK,EAAE,KAAK,CAAC,IAAI,CAAC,SAAS,CAAC,GAAG,CAAC,EAAE,KAAK,CAAC,CAAC,CAC3E;SACF;QACD,KAAK,iBAAiB,EAAE,OAAO;AAC7B,YAAA,IAAI,EAAE,iBAAiB;YACvB,kBAAkB,EAAE,eAAe,CAAC,kBAAkB;AACtD,YAAA,KAAK,EAAE,MAAM,CAAC,WAAW,CACvB,eAAe,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,EAAE,GAAG,EAAE,KAAK,EAAE,KAAK,CAAC,IAAI,CAAC,SAAS,CAAC,GAAG,CAAC,EAAE,KAAK,CAAC,CAAC,CAC5E;SACF;QACD,KAAK,mBAAmB,EAAE,OAAO;AAC/B,YAAA,IAAI,EAAE,mBAAmB;YACzB,kBAAkB,EAAE,eAAe,CAAC,kBAAkB;AACtD,YAAA,KAAK,EAAE,MAAM,CAAC,WAAW,CACvB,eAAe,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,EAAE,GAAG,EAAE,KAAK,EAAE,KAAK,CAAC,IAAI,CAAC,SAAS,CAAC,GAAG,CAAC,EAAE,KAAK,CAAC,CAAC,CAC5E;SACF;QACD,KAAK,oBAAoB,EAAE,OAAO;AAChC,YAAA,IAAI,EAAE,oBAAoB;YAC1B,kBAAkB,EAAE,eAAe,CAAC,kBAAkB;AACtD,YAAA,KAAK,EAAE,MAAM,CAAC,WAAW,CACvB,eAAe,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,EAAE,GAAG,EAAE,KAAK,EAAE,KAAK,CAAC,IAAI,CAAC,SAAS,CAAC,GAAG,CAAC,EAAE,KAAK,CAAC,CAAC,CAC5E;SACF;AACD,QAAA;YACEA,gBAAW,CAAC,eAAe,CAAC;;AAElC;AAeM,SAAU,kBAAkB,CAChC,eAAgD,EAChD,KAAuB,EAAA;AAEvB,IAAA,IAAI,eAAe,KAAK,SAAS,EAAE;AACjC,QAAA,OAAO,SAAS;IAClB;AAEA,IAAA,QAAQ,eAAe,CAAC,IAAI;AAC1B,QAAA,KAAK,MAAM;;AAET,YAAA,OAAO,eAAe;QACxB,KAAK,iBAAiB,EAAE,OAAO;AAC7B,YAAA,GAAG,eAAe;AAClB,YAAA,KAAK,EAAE,eAAe,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,KAAK,MAAM;gBAC3C,GAAG,EAAE,KAAK,CAAC,GAAG;AACd,gBAAA,KAAK,EAAE,KAAK,CAAC,KAAK,CAAC,KAAK,CAAC;AAC1B,aAAA,CAAC,CAAC;SACJ;QACD,KAAK,mBAAmB,EAAE,OAAO;AAC/B,YAAA,GAAG,eAAe;AAClB,YAAA,KAAK,EAAE,eAAe,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,KAAK,MAAM;gBAC3C,GAAG,EAAE,KAAK,CAAC,GAAG;AACd,gBAAA,KAAK,EAAE;oBACL,KAAK,EAAE,KAAK,CAAC,KAAK,CAAC,KAAK,CAAC,KAAK,CAAC;oBAC/B,MAAM,EAAE,KAAK,CAAC,KAAK,CAAC,KAAK,CAAC,MAAM,CAAC;AAClC,iBAAA;AACF,aAAA,CAAC,CAAC;SACJ;QACD,KAAK,oBAAoB,EAAE,OAAO;AAChC,YAAA,GAAG,eAAe;AAClB,YAAA,KAAK,EAAE,eAAe,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,KAAK,MAAM;gBAC3C,GAAG,EAAE,KAAK,CAAC,GAAG;AACd,gBAAA,KAAK,EAAE,KAAK,CAAC,KAAK,CAAC,KAAK,CAAC;AAC1B,aAAA,CAAC,CAAC;SACJ;;AAEL;;;;;;;;;;;"}
@@ -241,4 +241,3 @@ export declare function entriesToDataInfo<Blob>(dataInfoEntries: DataInfoEntries
241
241
  * @returns A new DataInfoEntries object with transformed blob references
242
242
  */
243
243
  export declare function mapDataInfoEntries<B1, B2>(dataInfoEntries: DataInfoEntries<B1>, mapFn: (blob: B1) => B2): DataInfoEntries<B2>;
244
- //# sourceMappingURL=data_info.d.ts.map