@etohq/workflows-flows 0.0.1-next-20260318155517

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (63) hide show
  1. package/README.md +14 -0
  2. package/dist/dataset/index.d.ts +3 -0
  3. package/dist/dataset/index.d.ts.map +1 -0
  4. package/dist/dataset/index.js +19 -0
  5. package/dist/dataset/index.js.map +1 -0
  6. package/dist/dataset/steps/convert-csv-to-records.d.ts +6 -0
  7. package/dist/dataset/steps/convert-csv-to-records.d.ts.map +1 -0
  8. package/dist/dataset/steps/convert-csv-to-records.js +11 -0
  9. package/dist/dataset/steps/convert-csv-to-records.js.map +1 -0
  10. package/dist/dataset/steps/create-dataset-import-run.d.ts +5 -0
  11. package/dist/dataset/steps/create-dataset-import-run.d.ts.map +1 -0
  12. package/dist/dataset/steps/create-dataset-import-run.js +12 -0
  13. package/dist/dataset/steps/create-dataset-import-run.js.map +1 -0
  14. package/dist/dataset/steps/import-csv-records.d.ts +11 -0
  15. package/dist/dataset/steps/import-csv-records.d.ts.map +1 -0
  16. package/dist/dataset/steps/import-csv-records.js +15 -0
  17. package/dist/dataset/steps/import-csv-records.js.map +1 -0
  18. package/dist/dataset/steps/index.d.ts +8 -0
  19. package/dist/dataset/steps/index.d.ts.map +1 -0
  20. package/dist/dataset/steps/index.js +24 -0
  21. package/dist/dataset/steps/index.js.map +1 -0
  22. package/dist/dataset/steps/pdf-to-text.d.ts +6 -0
  23. package/dist/dataset/steps/pdf-to-text.d.ts.map +1 -0
  24. package/dist/dataset/steps/pdf-to-text.js +15 -0
  25. package/dist/dataset/steps/pdf-to-text.js.map +1 -0
  26. package/dist/dataset/steps/preview-structure-text.d.ts +8 -0
  27. package/dist/dataset/steps/preview-structure-text.d.ts.map +1 -0
  28. package/dist/dataset/steps/preview-structure-text.js +11 -0
  29. package/dist/dataset/steps/preview-structure-text.js.map +1 -0
  30. package/dist/dataset/steps/retrieve-dataset-draft.d.ts +6 -0
  31. package/dist/dataset/steps/retrieve-dataset-draft.d.ts.map +1 -0
  32. package/dist/dataset/steps/retrieve-dataset-draft.js +12 -0
  33. package/dist/dataset/steps/retrieve-dataset-draft.js.map +1 -0
  34. package/dist/dataset/steps/update-dataset-import-run.d.ts +5 -0
  35. package/dist/dataset/steps/update-dataset-import-run.d.ts.map +1 -0
  36. package/dist/dataset/steps/update-dataset-import-run.js +12 -0
  37. package/dist/dataset/steps/update-dataset-import-run.js.map +1 -0
  38. package/dist/dataset/workflows/index.d.ts +4 -0
  39. package/dist/dataset/workflows/index.d.ts.map +1 -0
  40. package/dist/dataset/workflows/index.js +20 -0
  41. package/dist/dataset/workflows/index.js.map +1 -0
  42. package/dist/dataset/workflows/preview-dataset-from-csv.d.ts +12 -0
  43. package/dist/dataset/workflows/preview-dataset-from-csv.d.ts.map +1 -0
  44. package/dist/dataset/workflows/preview-dataset-from-csv.js +16 -0
  45. package/dist/dataset/workflows/preview-dataset-from-csv.js.map +1 -0
  46. package/dist/dataset/workflows/preview-dataset-from-pdf.d.ts +18 -0
  47. package/dist/dataset/workflows/preview-dataset-from-pdf.d.ts.map +1 -0
  48. package/dist/dataset/workflows/preview-dataset-from-pdf.js +27 -0
  49. package/dist/dataset/workflows/preview-dataset-from-pdf.js.map +1 -0
  50. package/dist/dataset/workflows/run-dataset-import-from-csv.d.ts +10 -0
  51. package/dist/dataset/workflows/run-dataset-import-from-csv.d.ts.map +1 -0
  52. package/dist/dataset/workflows/run-dataset-import-from-csv.js +63 -0
  53. package/dist/dataset/workflows/run-dataset-import-from-csv.js.map +1 -0
  54. package/dist/index.d.ts +3 -0
  55. package/dist/index.d.ts.map +1 -0
  56. package/dist/index.js +19 -0
  57. package/dist/index.js.map +1 -0
  58. package/dist/templates/index.d.ts +2 -0
  59. package/dist/templates/index.d.ts.map +1 -0
  60. package/dist/templates/index.js +5 -0
  61. package/dist/templates/index.js.map +1 -0
  62. package/dist/tsconfig.tsbuildinfo +1 -0
  63. package/package.json +36 -0
package/README.md ADDED
@@ -0,0 +1,14 @@
1
+ # `eto-workflows` flows
2
+
3
+ Workflow orchestration layer.
4
+
5
+ This package is where cross-module behavior should live:
6
+ - parsing/preview steps for CSV/PDF imports
7
+ - rule evaluation and validation orchestration
8
+ - multi-step import workflows (draft -> preview -> confirm -> execute)
9
+
10
+ Module services should remain persistence-focused; avoid embedding parsing/transformation logic there.
11
+
12
+ ## Dataset Preview Workflows
13
+ - CSV preview: `dataset.preview-from-csv`
14
+ - PDF preview (Node-only): `dataset.preview-from-pdf`
@@ -0,0 +1,3 @@
1
+ export * from "./steps";
2
+ export * from "./workflows";
3
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/dataset/index.ts"],"names":[],"mappings":"AAAA,cAAc,SAAS,CAAA;AACvB,cAAc,aAAa,CAAA"}
@@ -0,0 +1,19 @@
1
+ "use strict";
2
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
3
+ if (k2 === undefined) k2 = k;
4
+ var desc = Object.getOwnPropertyDescriptor(m, k);
5
+ if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
6
+ desc = { enumerable: true, get: function() { return m[k]; } };
7
+ }
8
+ Object.defineProperty(o, k2, desc);
9
+ }) : (function(o, m, k, k2) {
10
+ if (k2 === undefined) k2 = k;
11
+ o[k2] = m[k];
12
+ }));
13
+ var __exportStar = (this && this.__exportStar) || function(m, exports) {
14
+ for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
15
+ };
16
+ Object.defineProperty(exports, "__esModule", { value: true });
17
+ __exportStar(require("./steps"), exports);
18
+ __exportStar(require("./workflows"), exports);
19
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/dataset/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,0CAAuB;AACvB,8CAA2B"}
@@ -0,0 +1,6 @@
1
+ export type ConvertCsvToRecordsStepInput = {
2
+ csv: string;
3
+ };
4
+ export declare const convertCsvToRecordsStepId = "dataset.convert-csv-to-records";
5
+ export declare const convertCsvToRecordsStep: import("@etohq/framework/workflows-sdk").StepFunction<ConvertCsvToRecordsStepInput, import("@etohq/workflows-input-dataset-runtime").CsvRecord[]>;
6
+ //# sourceMappingURL=convert-csv-to-records.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"convert-csv-to-records.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/convert-csv-to-records.ts"],"names":[],"mappings":"AAGA,MAAM,MAAM,4BAA4B,GAAG;IACzC,GAAG,EAAE,MAAM,CAAA;CACZ,CAAA;AAED,eAAO,MAAM,yBAAyB,mCAAmC,CAAA;AAEzE,eAAO,MAAM,uBAAuB,mJAMnC,CAAA"}
@@ -0,0 +1,11 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.convertCsvToRecordsStep = exports.convertCsvToRecordsStepId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const workflows_input_dataset_runtime_1 = require("@etohq/workflows-input-dataset-runtime");
6
+ exports.convertCsvToRecordsStepId = "dataset.convert-csv-to-records";
7
+ exports.convertCsvToRecordsStep = (0, workflows_sdk_1.createStep)(exports.convertCsvToRecordsStepId, async (input) => {
8
+ const records = (0, workflows_input_dataset_runtime_1.convertCsvToRecords)(input.csv);
9
+ return new workflows_sdk_1.StepResponse(records);
10
+ });
11
+ //# sourceMappingURL=convert-csv-to-records.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"convert-csv-to-records.js","sourceRoot":"","sources":["../../../src/dataset/steps/convert-csv-to-records.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AACzE,4FAA4E;AAM/D,QAAA,yBAAyB,GAAG,gCAAgC,CAAA;AAE5D,QAAA,uBAAuB,GAAG,IAAA,0BAAU,EAC/C,iCAAyB,EACzB,KAAK,EAAE,KAAmC,EAAE,EAAE;IAC5C,MAAM,OAAO,GAAG,IAAA,qDAAmB,EAAC,KAAK,CAAC,GAAG,CAAC,CAAA;IAC9C,OAAO,IAAI,4BAAY,CAAC,OAAO,CAAC,CAAA;AAClC,CAAC,CACF,CAAA"}
@@ -0,0 +1,5 @@
1
+ import type { CreateDatasetImportRunDTO } from "@etohq/workflows-types";
2
+ export type CreateDatasetImportRunStepInput = CreateDatasetImportRunDTO;
3
+ export declare const createDatasetImportRunStepId = "dataset.create-import-run";
4
+ export declare const createDatasetImportRunStep: import("@etohq/framework/workflows-sdk").StepFunction<CreateDatasetImportRunDTO, import("@etohq/workflows-types").DatasetImportRunDTO>;
5
+ //# sourceMappingURL=create-dataset-import-run.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"create-dataset-import-run.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/create-dataset-import-run.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,yBAAyB,EAAyB,MAAM,wBAAwB,CAAA;AAG9F,MAAM,MAAM,+BAA+B,GAAG,yBAAyB,CAAA;AAEvE,eAAO,MAAM,4BAA4B,8BAA8B,CAAA;AAEvE,eAAO,MAAM,0BAA0B,wIAUtC,CAAA"}
@@ -0,0 +1,12 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.createDatasetImportRunStep = exports.createDatasetImportRunStepId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const workflows_types_1 = require("@etohq/workflows-types");
6
+ exports.createDatasetImportRunStepId = "dataset.create-import-run";
7
+ exports.createDatasetImportRunStep = (0, workflows_sdk_1.createStep)(exports.createDatasetImportRunStepId, async (input, { container }) => {
8
+ const datasetService = container.resolve(workflows_types_1.WORKFLOWS_MODULE_KEYS.DATASET);
9
+ const run = await datasetService.createDatasetImportRuns(input);
10
+ return new workflows_sdk_1.StepResponse(run);
11
+ });
12
+ //# sourceMappingURL=create-dataset-import-run.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"create-dataset-import-run.js","sourceRoot":"","sources":["../../../src/dataset/steps/create-dataset-import-run.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,4DAA8D;AAIjD,QAAA,4BAA4B,GAAG,2BAA2B,CAAA;AAE1D,QAAA,0BAA0B,GAAG,IAAA,0BAAU,EAClD,oCAA4B,EAC5B,KAAK,EAAE,KAAsC,EAAE,EAAE,SAAS,EAAE,EAAE,EAAE;IAC9D,MAAM,cAAc,GAAG,SAAS,CAAC,OAAO,CACtC,uCAAqB,CAAC,OAAO,CAC9B,CAAA;IAED,MAAM,GAAG,GAAG,MAAM,cAAc,CAAC,uBAAuB,CAAC,KAAK,CAAC,CAAA;IAC/D,OAAO,IAAI,4BAAY,CAAC,GAAG,CAAC,CAAA;AAC9B,CAAC,CACF,CAAA"}
@@ -0,0 +1,11 @@
1
+ import type { InputSchemaSpec } from "@etohq/workflows-input-schema";
2
+ import type { DatasetImportSpec } from "@etohq/workflows-input-dataset";
3
+ import type { CsvRecord } from "@etohq/workflows-input-dataset-runtime";
4
+ export type ImportCsvRecordsStepInput = {
5
+ schema: InputSchemaSpec;
6
+ spec: DatasetImportSpec;
7
+ records: CsvRecord[];
8
+ };
9
+ export declare const importCsvRecordsStepId = "dataset.import-csv-records";
10
+ export declare const importCsvRecordsStep: import("@etohq/framework/workflows-sdk").StepFunction<ImportCsvRecordsStepInput, import("@etohq/workflows-input-dataset-runtime").DatasetImportResult>;
11
+ //# sourceMappingURL=import-csv-records.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"import-csv-records.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/import-csv-records.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,+BAA+B,CAAA;AACpE,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,gCAAgC,CAAA;AACvE,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,wCAAwC,CAAA;AAGvE,MAAM,MAAM,yBAAyB,GAAG;IACtC,MAAM,EAAE,eAAe,CAAA;IACvB,IAAI,EAAE,iBAAiB,CAAA;IACvB,OAAO,EAAE,SAAS,EAAE,CAAA;CACrB,CAAA;AAED,eAAO,MAAM,sBAAsB,+BAA+B,CAAA;AAElE,eAAO,MAAM,oBAAoB,wJAUhC,CAAA"}
@@ -0,0 +1,15 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.importCsvRecordsStep = exports.importCsvRecordsStepId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const workflows_input_dataset_runtime_1 = require("@etohq/workflows-input-dataset-runtime");
6
+ exports.importCsvRecordsStepId = "dataset.import-csv-records";
7
+ exports.importCsvRecordsStep = (0, workflows_sdk_1.createStep)(exports.importCsvRecordsStepId, async (input) => {
8
+ const result = (0, workflows_input_dataset_runtime_1.importCsvRecords)({
9
+ schema: input.schema,
10
+ spec: input.spec,
11
+ records: input.records,
12
+ });
13
+ return new workflows_sdk_1.StepResponse(result);
14
+ });
15
+ //# sourceMappingURL=import-csv-records.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"import-csv-records.js","sourceRoot":"","sources":["../../../src/dataset/steps/import-csv-records.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAIzE,4FAAyE;AAQ5D,QAAA,sBAAsB,GAAG,4BAA4B,CAAA;AAErD,QAAA,oBAAoB,GAAG,IAAA,0BAAU,EAC5C,8BAAsB,EACtB,KAAK,EAAE,KAAgC,EAAE,EAAE;IACzC,MAAM,MAAM,GAAG,IAAA,kDAAgB,EAAC;QAC9B,MAAM,EAAE,KAAK,CAAC,MAAM;QACpB,IAAI,EAAE,KAAK,CAAC,IAAI;QAChB,OAAO,EAAE,KAAK,CAAC,OAAO;KACvB,CAAC,CAAA;IACF,OAAO,IAAI,4BAAY,CAAC,MAAM,CAAC,CAAA;AACjC,CAAC,CACF,CAAA"}
@@ -0,0 +1,8 @@
1
+ export * from "./convert-csv-to-records";
2
+ export * from "./import-csv-records";
3
+ export * from "./pdf-to-text";
4
+ export * from "./preview-structure-text";
5
+ export * from "./retrieve-dataset-draft";
6
+ export * from "./create-dataset-import-run";
7
+ export * from "./update-dataset-import-run";
8
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/index.ts"],"names":[],"mappings":"AAAA,cAAc,0BAA0B,CAAA;AACxC,cAAc,sBAAsB,CAAA;AACpC,cAAc,eAAe,CAAA;AAC7B,cAAc,0BAA0B,CAAA;AACxC,cAAc,0BAA0B,CAAA;AACxC,cAAc,6BAA6B,CAAA;AAC3C,cAAc,6BAA6B,CAAA"}
@@ -0,0 +1,24 @@
1
+ "use strict";
2
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
3
+ if (k2 === undefined) k2 = k;
4
+ var desc = Object.getOwnPropertyDescriptor(m, k);
5
+ if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
6
+ desc = { enumerable: true, get: function() { return m[k]; } };
7
+ }
8
+ Object.defineProperty(o, k2, desc);
9
+ }) : (function(o, m, k, k2) {
10
+ if (k2 === undefined) k2 = k;
11
+ o[k2] = m[k];
12
+ }));
13
+ var __exportStar = (this && this.__exportStar) || function(m, exports) {
14
+ for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
15
+ };
16
+ Object.defineProperty(exports, "__esModule", { value: true });
17
+ __exportStar(require("./convert-csv-to-records"), exports);
18
+ __exportStar(require("./import-csv-records"), exports);
19
+ __exportStar(require("./pdf-to-text"), exports);
20
+ __exportStar(require("./preview-structure-text"), exports);
21
+ __exportStar(require("./retrieve-dataset-draft"), exports);
22
+ __exportStar(require("./create-dataset-import-run"), exports);
23
+ __exportStar(require("./update-dataset-import-run"), exports);
24
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../../src/dataset/steps/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,2DAAwC;AACxC,uDAAoC;AACpC,gDAA6B;AAC7B,2DAAwC;AACxC,2DAAwC;AACxC,8DAA2C;AAC3C,8DAA2C"}
@@ -0,0 +1,6 @@
1
+ export type PdfToTextStepInput = {
2
+ pdf_bytes: Uint8Array;
3
+ };
4
+ export declare const pdfToTextStepId = "dataset.pdf-to-text";
5
+ export declare const pdfToTextStep: import("@etohq/framework/workflows-sdk").StepFunction<PdfToTextStepInput, string>;
6
+ //# sourceMappingURL=pdf-to-text.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"pdf-to-text.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/pdf-to-text.ts"],"names":[],"mappings":"AAIA,MAAM,MAAM,kBAAkB,GAAG;IAC/B,SAAS,EAAE,UAAU,CAAA;CACtB,CAAA;AAED,eAAO,MAAM,eAAe,wBAAwB,CAAA;AAEpD,eAAO,MAAM,aAAa,mFASzB,CAAA"}
@@ -0,0 +1,15 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.pdfToTextStep = exports.pdfToTextStepId = void 0;
4
+ const utils_1 = require("@etohq/framework/utils");
5
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
6
+ const workflows_input_dataset_pdf_runtime_1 = require("@etohq/workflows-input-dataset-pdf-runtime");
7
+ exports.pdfToTextStepId = "dataset.pdf-to-text";
8
+ exports.pdfToTextStep = (0, workflows_sdk_1.createStep)(exports.pdfToTextStepId, async (input) => {
9
+ const res = await (0, workflows_input_dataset_pdf_runtime_1.pdfToText)(input.pdf_bytes);
10
+ if (!res.ok) {
11
+ throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, res.error);
12
+ }
13
+ return new workflows_sdk_1.StepResponse(res.text);
14
+ });
15
+ //# sourceMappingURL=pdf-to-text.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"pdf-to-text.js","sourceRoot":"","sources":["../../../src/dataset/steps/pdf-to-text.ts"],"names":[],"mappings":";;;AAAA,kDAAiD;AACjD,kEAAyE;AACzE,oGAAsE;AAMzD,QAAA,eAAe,GAAG,qBAAqB,CAAA;AAEvC,QAAA,aAAa,GAAG,IAAA,0BAAU,EACrC,uBAAe,EACf,KAAK,EAAE,KAAyB,EAAE,EAAE;IAClC,MAAM,GAAG,GAAG,MAAM,IAAA,+CAAS,EAAC,KAAK,CAAC,SAAS,CAAC,CAAA;IAC5C,IAAI,CAAC,GAAG,CAAC,EAAE,EAAE,CAAC;QACZ,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,GAAG,CAAC,KAAK,CAAC,CAAA;IAC5D,CAAC;IACD,OAAO,IAAI,4BAAY,CAAC,GAAG,CAAC,IAAI,CAAC,CAAA;AACnC,CAAC,CACF,CAAA"}
@@ -0,0 +1,8 @@
1
+ import type { PdfExtractSpec } from "@etohq/workflows-input-dataset";
2
+ export type PreviewStructureTextStepInput = {
3
+ text: string;
4
+ spec: PdfExtractSpec;
5
+ };
6
+ export declare const previewStructureTextStepId = "dataset.preview-structure-text";
7
+ export declare const previewStructureTextStep: import("@etohq/framework/workflows-sdk").StepFunction<PreviewStructureTextStepInput, import("@etohq/workflows-input-dataset-pdf-runtime").StructureTextPreviewResult>;
8
+ //# sourceMappingURL=preview-structure-text.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"preview-structure-text.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/preview-structure-text.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAGpE,MAAM,MAAM,6BAA6B,GAAG;IAC1C,IAAI,EAAE,MAAM,CAAA;IACZ,IAAI,EAAE,cAAc,CAAA;CACrB,CAAA;AAED,eAAO,MAAM,0BAA0B,mCAAmC,CAAA;AAE1E,eAAO,MAAM,wBAAwB,uKAMpC,CAAA"}
@@ -0,0 +1,11 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.previewStructureTextStep = exports.previewStructureTextStepId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const workflows_input_dataset_pdf_runtime_1 = require("@etohq/workflows-input-dataset-pdf-runtime");
6
+ exports.previewStructureTextStepId = "dataset.preview-structure-text";
7
+ exports.previewStructureTextStep = (0, workflows_sdk_1.createStep)(exports.previewStructureTextStepId, async (input) => {
8
+ const result = (0, workflows_input_dataset_pdf_runtime_1.previewStructureRecordsFromText)(input.text, input.spec);
9
+ return new workflows_sdk_1.StepResponse(result);
10
+ });
11
+ //# sourceMappingURL=preview-structure-text.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"preview-structure-text.js","sourceRoot":"","sources":["../../../src/dataset/steps/preview-structure-text.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,oGAA4F;AAO/E,QAAA,0BAA0B,GAAG,gCAAgC,CAAA;AAE7D,QAAA,wBAAwB,GAAG,IAAA,0BAAU,EAChD,kCAA0B,EAC1B,KAAK,EAAE,KAAoC,EAAE,EAAE;IAC7C,MAAM,MAAM,GAAG,IAAA,qEAA+B,EAAC,KAAK,CAAC,IAAI,EAAE,KAAK,CAAC,IAAI,CAAC,CAAA;IACtE,OAAO,IAAI,4BAAY,CAAC,MAAM,CAAC,CAAA;AACjC,CAAC,CACF,CAAA"}
@@ -0,0 +1,6 @@
1
+ export type RetrieveDatasetDraftStepInput = {
2
+ id: string;
3
+ };
4
+ export declare const retrieveDatasetDraftStepId = "dataset.retrieve-dataset-draft";
5
+ export declare const retrieveDatasetDraftStep: import("@etohq/framework/workflows-sdk").StepFunction<RetrieveDatasetDraftStepInput, import("@etohq/workflows-types").DatasetDraftDTO>;
6
+ //# sourceMappingURL=retrieve-dataset-draft.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"retrieve-dataset-draft.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/retrieve-dataset-draft.ts"],"names":[],"mappings":"AAIA,MAAM,MAAM,6BAA6B,GAAG;IAC1C,EAAE,EAAE,MAAM,CAAA;CACX,CAAA;AAED,eAAO,MAAM,0BAA0B,mCAAmC,CAAA;AAE1E,eAAO,MAAM,wBAAwB,wIAUpC,CAAA"}
@@ -0,0 +1,12 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.retrieveDatasetDraftStep = exports.retrieveDatasetDraftStepId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const workflows_types_1 = require("@etohq/workflows-types");
6
+ exports.retrieveDatasetDraftStepId = "dataset.retrieve-dataset-draft";
7
+ exports.retrieveDatasetDraftStep = (0, workflows_sdk_1.createStep)(exports.retrieveDatasetDraftStepId, async (input, { container }) => {
8
+ const datasetService = container.resolve(workflows_types_1.WORKFLOWS_MODULE_KEYS.DATASET);
9
+ const draft = await datasetService.retrieveDatasetDraft(input.id);
10
+ return new workflows_sdk_1.StepResponse(draft);
11
+ });
12
+ //# sourceMappingURL=retrieve-dataset-draft.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"retrieve-dataset-draft.js","sourceRoot":"","sources":["../../../src/dataset/steps/retrieve-dataset-draft.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,4DAA8D;AAMjD,QAAA,0BAA0B,GAAG,gCAAgC,CAAA;AAE7D,QAAA,wBAAwB,GAAG,IAAA,0BAAU,EAChD,kCAA0B,EAC1B,KAAK,EAAE,KAAoC,EAAE,EAAE,SAAS,EAAE,EAAE,EAAE;IAC5D,MAAM,cAAc,GAAG,SAAS,CAAC,OAAO,CACtC,uCAAqB,CAAC,OAAO,CAC9B,CAAA;IAED,MAAM,KAAK,GAAG,MAAM,cAAc,CAAC,oBAAoB,CAAC,KAAK,CAAC,EAAE,CAAC,CAAA;IACjE,OAAO,IAAI,4BAAY,CAAC,KAAK,CAAC,CAAA;AAChC,CAAC,CACF,CAAA"}
@@ -0,0 +1,5 @@
1
+ import type { UpdateDatasetImportRunRequestDTO } from "@etohq/workflows-types";
2
+ export type UpdateDatasetImportRunStepInput = UpdateDatasetImportRunRequestDTO;
3
+ export declare const updateDatasetImportRunStepId = "dataset.update-import-run";
4
+ export declare const updateDatasetImportRunStep: import("@etohq/framework/workflows-sdk").StepFunction<UpdateDatasetImportRunRequestDTO, import("@etohq/workflows-types").DatasetImportRunDTO>;
5
+ //# sourceMappingURL=update-dataset-import-run.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"update-dataset-import-run.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/update-dataset-import-run.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAyB,gCAAgC,EAAE,MAAM,wBAAwB,CAAA;AAGrG,MAAM,MAAM,+BAA+B,GAAG,gCAAgC,CAAA;AAE9E,eAAO,MAAM,4BAA4B,8BAA8B,CAAA;AAEvE,eAAO,MAAM,0BAA0B,+IAUtC,CAAA"}
@@ -0,0 +1,12 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.updateDatasetImportRunStep = exports.updateDatasetImportRunStepId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const workflows_types_1 = require("@etohq/workflows-types");
6
+ exports.updateDatasetImportRunStepId = "dataset.update-import-run";
7
+ exports.updateDatasetImportRunStep = (0, workflows_sdk_1.createStep)(exports.updateDatasetImportRunStepId, async (input, { container }) => {
8
+ const datasetService = container.resolve(workflows_types_1.WORKFLOWS_MODULE_KEYS.DATASET);
9
+ const updated = await datasetService.updateDatasetImportRuns(input);
10
+ return new workflows_sdk_1.StepResponse(updated);
11
+ });
12
+ //# sourceMappingURL=update-dataset-import-run.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"update-dataset-import-run.js","sourceRoot":"","sources":["../../../src/dataset/steps/update-dataset-import-run.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,4DAA8D;AAIjD,QAAA,4BAA4B,GAAG,2BAA2B,CAAA;AAE1D,QAAA,0BAA0B,GAAG,IAAA,0BAAU,EAClD,oCAA4B,EAC5B,KAAK,EAAE,KAAsC,EAAE,EAAE,SAAS,EAAE,EAAE,EAAE;IAC9D,MAAM,cAAc,GAAG,SAAS,CAAC,OAAO,CACtC,uCAAqB,CAAC,OAAO,CAC9B,CAAA;IAED,MAAM,OAAO,GAAG,MAAM,cAAc,CAAC,uBAAuB,CAAC,KAAK,CAAC,CAAA;IACnE,OAAO,IAAI,4BAAY,CAAC,OAAO,CAAC,CAAA;AAClC,CAAC,CACF,CAAA"}
@@ -0,0 +1,4 @@
1
+ export * from "./preview-dataset-from-csv";
2
+ export * from "./preview-dataset-from-pdf";
3
+ export * from "./run-dataset-import-from-csv";
4
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/index.ts"],"names":[],"mappings":"AAAA,cAAc,4BAA4B,CAAA;AAC1C,cAAc,4BAA4B,CAAA;AAC1C,cAAc,+BAA+B,CAAA"}
@@ -0,0 +1,20 @@
1
+ "use strict";
2
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
3
+ if (k2 === undefined) k2 = k;
4
+ var desc = Object.getOwnPropertyDescriptor(m, k);
5
+ if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
6
+ desc = { enumerable: true, get: function() { return m[k]; } };
7
+ }
8
+ Object.defineProperty(o, k2, desc);
9
+ }) : (function(o, m, k, k2) {
10
+ if (k2 === undefined) k2 = k;
11
+ o[k2] = m[k];
12
+ }));
13
+ var __exportStar = (this && this.__exportStar) || function(m, exports) {
14
+ for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
15
+ };
16
+ Object.defineProperty(exports, "__esModule", { value: true });
17
+ __exportStar(require("./preview-dataset-from-csv"), exports);
18
+ __exportStar(require("./preview-dataset-from-pdf"), exports);
19
+ __exportStar(require("./run-dataset-import-from-csv"), exports);
20
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../../src/dataset/workflows/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,6DAA0C;AAC1C,6DAA0C;AAC1C,gEAA6C"}
@@ -0,0 +1,12 @@
1
+ import type { ReturnWorkflow } from "@etohq/framework/workflows-sdk";
2
+ import type { InputSchemaSpec } from "@etohq/workflows-input-schema";
3
+ import type { DatasetImportSpec } from "@etohq/workflows-input-dataset";
4
+ import type { DatasetImportResult } from "@etohq/workflows-input-dataset-runtime";
5
+ export type PreviewDatasetFromCsvWorkflowInput = {
6
+ schema: InputSchemaSpec;
7
+ spec: DatasetImportSpec;
8
+ csv: string;
9
+ };
10
+ export declare const previewDatasetFromCsvWorkflowId = "dataset.preview-from-csv";
11
+ export declare const previewDatasetFromCsvWorkflow: ReturnWorkflow<PreviewDatasetFromCsvWorkflowInput, DatasetImportResult, unknown[]>;
12
+ //# sourceMappingURL=preview-dataset-from-csv.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"preview-dataset-from-csv.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-csv.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAEpE,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,+BAA+B,CAAA;AACpE,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,gCAAgC,CAAA;AAEvE,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AAEjF,MAAM,MAAM,kCAAkC,GAAG;IAC/C,MAAM,EAAE,eAAe,CAAA;IACvB,IAAI,EAAE,iBAAiB,CAAA;IACvB,GAAG,EAAE,MAAM,CAAA;CACZ,CAAA;AAED,eAAO,MAAM,+BAA+B,6BAA6B,CAAA;AAEzE,eAAO,MAAM,6BAA6B,EAAE,cAAc,CACxD,kCAAkC,EAClC,mBAAmB,EACnB,OAAO,EAAE,CAkBV,CAAA"}
@@ -0,0 +1,16 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.previewDatasetFromCsvWorkflow = exports.previewDatasetFromCsvWorkflowId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const steps_1 = require("../steps");
6
+ exports.previewDatasetFromCsvWorkflowId = "dataset.preview-from-csv";
7
+ exports.previewDatasetFromCsvWorkflow = (0, workflows_sdk_1.createWorkflow)(exports.previewDatasetFromCsvWorkflowId, function (input) {
8
+ const records = (0, steps_1.convertCsvToRecordsStep)({ csv: input.csv });
9
+ const imported = (0, steps_1.importCsvRecordsStep)((0, workflows_sdk_1.transform)({ input, records }, ({ input, records }) => ({
10
+ schema: input.schema,
11
+ spec: input.spec,
12
+ records,
13
+ })));
14
+ return new workflows_sdk_1.WorkflowResponse(imported);
15
+ });
16
+ //# sourceMappingURL=preview-dataset-from-csv.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"preview-dataset-from-csv.js","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-csv.ts"],"names":[],"mappings":";;;AACA,kEAA4F;AAG5F,oCAAwE;AAS3D,QAAA,+BAA+B,GAAG,0BAA0B,CAAA;AAE5D,QAAA,6BAA6B,GAItC,IAAA,8BAAc,EAChB,uCAA+B,EAC/B,UACE,KAAyC;IAEzC,MAAM,OAAO,GAAG,IAAA,+BAAuB,EAAC,EAAE,GAAG,EAAE,KAAK,CAAC,GAAG,EAAE,CAAC,CAAA;IAE3D,MAAM,QAAQ,GAAG,IAAA,4BAAoB,EACnC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,OAAO,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC,CAAC;QACrD,MAAM,EAAE,KAAK,CAAC,MAAM;QACpB,IAAI,EAAE,KAAK,CAAC,IAAI;QAChB,OAAO;KACR,CAAC,CAAC,CACJ,CAAA;IAED,OAAO,IAAI,gCAAgB,CAAC,QAAQ,CAAC,CAAA;AACvC,CAAC,CACF,CAAA"}
@@ -0,0 +1,18 @@
1
+ import type { ReturnWorkflow } from "@etohq/framework/workflows-sdk";
2
+ import type { InputSchemaSpec } from "@etohq/workflows-input-schema";
3
+ import type { DatasetImportSpec, PdfExtractSpec } from "@etohq/workflows-input-dataset";
4
+ export type PreviewDatasetFromPdfWorkflowInput = {
5
+ schema: InputSchemaSpec;
6
+ import_spec: DatasetImportSpec;
7
+ extract_spec: PdfExtractSpec;
8
+ pdf_bytes: Uint8Array;
9
+ };
10
+ export declare const previewDatasetFromPdfWorkflowId = "dataset.preview-from-pdf";
11
+ /**
12
+ * Node-only workflow: uses `pdf-parse` under the hood.
13
+ *
14
+ * Output is a preview-friendly structure result + an import result that maps extracted records
15
+ * into schema submissions and validation issues.
16
+ */
17
+ export declare const previewDatasetFromPdfWorkflow: ReturnWorkflow<PreviewDatasetFromPdfWorkflowInput, unknown, unknown[]>;
18
+ //# sourceMappingURL=preview-dataset-from-pdf.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"preview-dataset-from-pdf.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-pdf.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAEpE,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,+BAA+B,CAAA;AACpE,OAAO,KAAK,EAAE,iBAAiB,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAGvF,MAAM,MAAM,kCAAkC,GAAG;IAC/C,MAAM,EAAE,eAAe,CAAA;IACvB,WAAW,EAAE,iBAAiB,CAAA;IAC9B,YAAY,EAAE,cAAc,CAAA;IAC5B,SAAS,EAAE,UAAU,CAAA;CACtB,CAAA;AAED,eAAO,MAAM,+BAA+B,6BAA6B,CAAA;AAEzE;;;;;GAKG;AACH,eAAO,MAAM,6BAA6B,EAAE,cAAc,CACxD,kCAAkC,EAClC,OAAO,EACP,OAAO,EAAE,CA0BV,CAAA"}
@@ -0,0 +1,27 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.previewDatasetFromPdfWorkflow = exports.previewDatasetFromPdfWorkflowId = void 0;
4
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
5
+ const steps_1 = require("../steps");
6
+ exports.previewDatasetFromPdfWorkflowId = "dataset.preview-from-pdf";
7
+ /**
8
+ * Node-only workflow: uses `pdf-parse` under the hood.
9
+ *
10
+ * Output is a preview-friendly structure result + an import result that maps extracted records
11
+ * into schema submissions and validation issues.
12
+ */
13
+ exports.previewDatasetFromPdfWorkflow = (0, workflows_sdk_1.createWorkflow)(exports.previewDatasetFromPdfWorkflowId, function (input) {
14
+ const text = (0, steps_1.pdfToTextStep)({ pdf_bytes: input.pdf_bytes });
15
+ const structured = (0, steps_1.previewStructureTextStep)((0, workflows_sdk_1.transform)({ input, text }, ({ input, text }) => ({
16
+ text,
17
+ spec: input.extract_spec,
18
+ })));
19
+ const imported = (0, steps_1.importCsvRecordsStep)((0, workflows_sdk_1.transform)({ input, structured }, ({ input, structured }) => ({
20
+ schema: input.schema,
21
+ spec: input.import_spec,
22
+ records: structured.records,
23
+ })));
24
+ // Intentionally keep result opaque at this layer to avoid leaking complex WorkflowData unions.
25
+ return new workflows_sdk_1.WorkflowResponse({ structured, imported });
26
+ });
27
+ //# sourceMappingURL=preview-dataset-from-pdf.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"preview-dataset-from-pdf.js","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-pdf.ts"],"names":[],"mappings":";;;AACA,kEAA4F;AAG5F,oCAAwF;AAS3E,QAAA,+BAA+B,GAAG,0BAA0B,CAAA;AAEzE;;;;;GAKG;AACU,QAAA,6BAA6B,GAItC,IAAA,8BAAc,EAChB,uCAA+B,EAC/B,UACE,KAAyC;IAEzC,MAAM,IAAI,GAAG,IAAA,qBAAa,EAAC,EAAE,SAAS,EAAE,KAAK,CAAC,SAAS,EAAE,CAAC,CAAA;IAE1D,MAAM,UAAU,GAAG,IAAA,gCAAwB,EACzC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,IAAI,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,IAAI,EAAE,EAAE,EAAE,CAAC,CAAC;QAC/C,IAAI;QACJ,IAAI,EAAE,KAAK,CAAC,YAAY;KACzB,CAAC,CAAC,CACJ,CAAA;IAED,MAAM,QAAQ,GAAG,IAAA,4BAAoB,EACnC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,UAAU,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,UAAU,EAAE,EAAE,EAAE,CAAC,CAAC;QAC3D,MAAM,EAAE,KAAK,CAAC,MAAM;QACpB,IAAI,EAAE,KAAK,CAAC,WAAW;QACvB,OAAO,EAAE,UAAU,CAAC,OAAO;KAC5B,CAAC,CAAC,CACJ,CAAA;IAED,+FAA+F;IAC/F,OAAO,IAAI,gCAAgB,CAAC,EAAE,UAAU,EAAE,QAAQ,EAAQ,CAAC,CAAA;AAC7D,CAAC,CACF,CAAA"}
@@ -0,0 +1,10 @@
1
+ import type { ReturnWorkflow } from "@etohq/framework/workflows-sdk";
2
+ import type { DatasetImportResult } from "@etohq/workflows-input-dataset-runtime";
3
+ export type RunDatasetImportFromCsvWorkflowInput = {
4
+ tenant_id: string;
5
+ draft_id: string;
6
+ csv: string;
7
+ };
8
+ export declare const runDatasetImportFromCsvWorkflowId = "dataset.run-import-from-csv";
9
+ export declare const runDatasetImportFromCsvWorkflow: ReturnWorkflow<RunDatasetImportFromCsvWorkflowInput, DatasetImportResult, unknown[]>;
10
+ //# sourceMappingURL=run-dataset-import-from-csv.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"run-dataset-import-from-csv.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/run-dataset-import-from-csv.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAEpE,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AAQjF,MAAM,MAAM,oCAAoC,GAAG;IACjD,SAAS,EAAE,MAAM,CAAA;IACjB,QAAQ,EAAE,MAAM,CAAA;IAChB,GAAG,EAAE,MAAM,CAAA;CACZ,CAAA;AAED,eAAO,MAAM,iCAAiC,gCAAgC,CAAA;AAE9E,eAAO,MAAM,+BAA+B,EAAE,cAAc,CAC1D,oCAAoC,EACpC,mBAAmB,EACnB,OAAO,EAAE,CAyEV,CAAA"}
@@ -0,0 +1,63 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.runDatasetImportFromCsvWorkflow = exports.runDatasetImportFromCsvWorkflowId = void 0;
4
+ const utils_1 = require("@etohq/framework/utils");
5
+ const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
6
+ const steps_1 = require("../steps");
7
+ const steps_2 = require("../steps");
8
+ exports.runDatasetImportFromCsvWorkflowId = "dataset.run-import-from-csv";
9
+ exports.runDatasetImportFromCsvWorkflow = (0, workflows_sdk_1.createWorkflow)(exports.runDatasetImportFromCsvWorkflowId, function (input) {
10
+ const draft = (0, steps_2.retrieveDatasetDraftStep)({ id: input.draft_id });
11
+ const schema = (0, workflows_sdk_1.transform)({ draft }, ({ draft }) => {
12
+ if (!draft.schema_spec) {
13
+ throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.schema_spec is required");
14
+ }
15
+ if (!draft.dataset_import_spec) {
16
+ throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.dataset_import_spec is required");
17
+ }
18
+ if (draft.kind !== "csv") {
19
+ throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.kind must be csv");
20
+ }
21
+ if (draft.status !== "ready") {
22
+ throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.status must be ready");
23
+ }
24
+ return { schema: draft.schema_spec, spec: draft.dataset_import_spec, revision: draft.revision };
25
+ });
26
+ const run = (0, steps_2.createDatasetImportRunStep)((0, workflows_sdk_1.transform)({ input, draft, schema }, ({ input, draft, schema }) => ({
27
+ tenant_id: input.tenant_id,
28
+ status: "running",
29
+ source_id: draft.source_id,
30
+ draft_id: draft.id,
31
+ draft_revision: schema.revision,
32
+ records_total: null,
33
+ records_succeeded: null,
34
+ records_failed: null,
35
+ started_at: null,
36
+ finished_at: null,
37
+ error_summary: null,
38
+ metadata: {},
39
+ })));
40
+ const records = (0, steps_1.convertCsvToRecordsStep)({ csv: input.csv });
41
+ const imported = (0, steps_1.importCsvRecordsStep)((0, workflows_sdk_1.transform)({ schema, records }, ({ schema, records }) => ({
42
+ schema: schema.schema,
43
+ spec: schema.spec,
44
+ records,
45
+ })));
46
+ const summary = (0, workflows_sdk_1.transform)({ imported }, ({ imported }) => {
47
+ const total = imported.rows.length;
48
+ const failed = imported.rows.filter((r) => r.validation.status === "invalid").length;
49
+ const succeeded = total - failed;
50
+ return { total, succeeded, failed, status: imported.status };
51
+ });
52
+ const errorSummary = (0, workflows_sdk_1.transform)({ imported, summary }, ({ imported, summary }) => summary.status === "invalid" ? { issues: imported.issues } : null);
53
+ (0, steps_2.updateDatasetImportRunStep)((0, workflows_sdk_1.transform)({ run, summary, errorSummary }, ({ run, summary, errorSummary }) => ({
54
+ id: run.id,
55
+ status: summary.status === "invalid" ? "failed" : "succeeded",
56
+ records_total: summary.total,
57
+ records_succeeded: summary.succeeded,
58
+ records_failed: summary.failed,
59
+ error_summary: errorSummary,
60
+ })));
61
+ return new workflows_sdk_1.WorkflowResponse(imported);
62
+ });
63
+ //# sourceMappingURL=run-dataset-import-from-csv.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"run-dataset-import-from-csv.js","sourceRoot":"","sources":["../../../src/dataset/workflows/run-dataset-import-from-csv.ts"],"names":[],"mappings":";;;AAAA,kDAAiD;AAEjD,kEAA4F;AAE5F,oCAAwE;AACxE,oCAIiB;AAQJ,QAAA,iCAAiC,GAAG,6BAA6B,CAAA;AAEjE,QAAA,+BAA+B,GAIxC,IAAA,8BAAc,EAChB,yCAAiC,EACjC,UAAU,KAA2C;IACnD,MAAM,KAAK,GAAG,IAAA,gCAAwB,EAAC,EAAE,EAAE,EAAE,KAAK,CAAC,QAAQ,EAAE,CAAC,CAAA;IAE9D,MAAM,MAAM,GAAG,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,EAAE,EAAE;QAChD,IAAI,CAAC,KAAK,CAAC,WAAW,EAAE,CAAC;YACvB,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,+BAA+B,CAAC,CAAA;QAClF,CAAC;QACD,IAAI,CAAC,KAAK,CAAC,mBAAmB,EAAE,CAAC;YAC/B,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,uCAAuC,CAAC,CAAA;QAC1F,CAAC;QACD,IAAI,KAAK,CAAC,IAAI,KAAK,KAAK,EAAE,CAAC;YACzB,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,wBAAwB,CAAC,CAAA;QAC3E,CAAC;QACD,IAAI,KAAK,CAAC,MAAM,KAAK,OAAO,EAAE,CAAC;YAC7B,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,4BAA4B,CAAC,CAAA;QAC/E,CAAC;QACD,OAAO,EAAE,MAAM,EAAE,KAAK,CAAC,WAAW,EAAE,IAAI,EAAE,KAAK,CAAC,mBAAmB,EAAE,QAAQ,EAAE,KAAK,CAAC,QAAQ,EAAE,CAAA;IACjG,CAAC,CAAC,CAAA;IAEF,MAAM,GAAG,GAAG,IAAA,kCAA0B,EACpC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,EAAE,CAAC,CAAC;QACjE,SAAS,EAAE,KAAK,CAAC,SAAS;QAC1B,MAAM,EAAE,SAAS;QACjB,SAAS,EAAE,KAAK,CAAC,SAAS;QAC1B,QAAQ,EAAE,KAAK,CAAC,EAAE;QAClB,cAAc,EAAE,MAAM,CAAC,QAAQ;QAC/B,aAAa,EAAE,IAAI;QACnB,iBAAiB,EAAE,IAAI;QACvB,cAAc,EAAE,IAAI;QACpB,UAAU,EAAE,IAAI;QAChB,WAAW,EAAE,IAAI;QACjB,aAAa,EAAE,IAAI;QACnB,QAAQ,EAAE,EAAE;KACb,CAAC,CAAC,CACJ,CAAA;IAED,MAAM,OAAO,GAAG,IAAA,+BAAuB,EAAC,EAAE,GAAG,EAAE,KAAK,CAAC,GAAG,EAAE,CAAC,CAAA;IAE3D,MAAM,QAAQ,GAAG,IAAA,4BAAoB,EACnC,IAAA,yBAAS,EAAC,EAAE,MAAM,EAAE,OAAO,EAAE,EAAE,CAAC,EAAE,MAAM,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC,CAAC;QACvD,MAAM,EAAE,MAAM,CAAC,MAAM;QACrB,IAAI,EAAE,MAAM,CAAC,IAAI;QACjB,OAAO;KACR,CAAC,CAAC,CACJ,CAAA;IAED,MAAM,OAAO,GAAG,IAAA,yBAAS,EAAC,EAAE,QAAQ,EAAE,EAAE,CAAC,EAAE,QAAQ,EAAE,EAAE,EAAE;QACvD,MAAM,KAAK,GAAG,QAAQ,CAAC,IAAI,CAAC,MAAM,CAAA;QAClC,MAAM,MAAM,GAAG,QAAQ,CAAC,IAAI,CAAC,MAAM,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,UAAU,CAAC,MAAM,KAAK,SAAS,CAAC,CAAC,MAAM,CAAA;QACpF,MAAM,SAAS,GAAG,KAAK,GAAG,MAAM,CAAA;QAChC,OAAO,EAAE,KAAK,EAAE,SAAS,EAAE,MAAM,EAAE,MAAM,EAAE,QAAQ,CAAC,MAAM,EAAE,CAAA;IAC9D,CAAC,CAAC,CAAA;IAEF,MAAM,YAAY,GAAG,IAAA,yBAAS,EAAC,EAAE,QAAQ,EAAE,OAAO,EAAE,EAAE,CAAC,EAAE,QAAQ,EAAE,OAAO,EAAE,EAAE,EAAE,CAC9E,OAAO,CAAC,MAAM,KAAK,SAAS,CAAC,CAAC,CAAE,EAAE,MAAM,EAAE,QAAQ,CAAC,MAAM,EAAS,CAAC,CAAC,CAAC,IAAI,CAC1E,CAAA;IAED,IAAA,kCAA0B,EACxB,IAAA,yBAAS,EAAC,EAAE,GAAG,EAAE,OAAO,EAAE,YAAY,EAAE,EAAE,CAAC,EAAE,GAAG,EAAE,OAAO,EAAE,YAAY,EAAE,EAAE,EAAE,CAAC,CAAC;QAC7E,EAAE,EAAE,GAAG,CAAC,EAAE;QACV,MAAM,EAAE,OAAO,CAAC,MAAM,KAAK,SAAS,CAAC,CAAC,CAAC,QAAQ,CAAC,CAAC,CAAC,WAAW;QAC7D,aAAa,EAAE,OAAO,CAAC,KAAK;QAC5B,iBAAiB,EAAE,OAAO,CAAC,SAAS;QACpC,cAAc,EAAE,OAAO,CAAC,MAAM;QAC9B,aAAa,EAAE,YAAY;KAC5B,CAAC,CAAC,CACJ,CAAA;IAED,OAAO,IAAI,gCAAgB,CAAC,QAAQ,CAAC,CAAA;AACvC,CAAC,CACF,CAAA"}
@@ -0,0 +1,3 @@
1
+ export * from "./templates";
2
+ export * from "./dataset";
3
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,cAAc,aAAa,CAAA;AAC3B,cAAc,WAAW,CAAA"}
package/dist/index.js ADDED
@@ -0,0 +1,19 @@
1
+ "use strict";
2
+ var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
3
+ if (k2 === undefined) k2 = k;
4
+ var desc = Object.getOwnPropertyDescriptor(m, k);
5
+ if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
6
+ desc = { enumerable: true, get: function() { return m[k]; } };
7
+ }
8
+ Object.defineProperty(o, k2, desc);
9
+ }) : (function(o, m, k, k2) {
10
+ if (k2 === undefined) k2 = k;
11
+ o[k2] = m[k];
12
+ }));
13
+ var __exportStar = (this && this.__exportStar) || function(m, exports) {
14
+ for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
15
+ };
16
+ Object.defineProperty(exports, "__esModule", { value: true });
17
+ __exportStar(require("./templates"), exports);
18
+ __exportStar(require("./dataset"), exports);
19
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,8CAA2B;AAC3B,4CAAyB"}
@@ -0,0 +1,2 @@
1
+ export declare const WORKFLOWS_TEMPLATE_LIBRARY_VERSION = 1;
2
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/templates/index.ts"],"names":[],"mappings":"AAAA,eAAO,MAAM,kCAAkC,IAAI,CAAA"}
@@ -0,0 +1,5 @@
1
+ "use strict";
2
+ Object.defineProperty(exports, "__esModule", { value: true });
3
+ exports.WORKFLOWS_TEMPLATE_LIBRARY_VERSION = void 0;
4
+ exports.WORKFLOWS_TEMPLATE_LIBRARY_VERSION = 1;
5
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/templates/index.ts"],"names":[],"mappings":";;;AAAa,QAAA,kCAAkC,GAAG,CAAC,CAAA"}
@@ -0,0 +1 @@
1
+ {"root":["../src/index.ts","../src/dataset/index.ts","../src/dataset/steps/convert-csv-to-records.ts","../src/dataset/steps/create-dataset-import-run.ts","../src/dataset/steps/import-csv-records.ts","../src/dataset/steps/index.ts","../src/dataset/steps/pdf-to-text.ts","../src/dataset/steps/preview-structure-text.ts","../src/dataset/steps/retrieve-dataset-draft.ts","../src/dataset/steps/update-dataset-import-run.ts","../src/dataset/workflows/index.ts","../src/dataset/workflows/preview-dataset-from-csv.ts","../src/dataset/workflows/preview-dataset-from-pdf.ts","../src/dataset/workflows/run-dataset-import-from-csv.ts","../src/templates/index.ts"],"version":"5.8.3"}
package/package.json ADDED
@@ -0,0 +1,36 @@
1
+ {
2
+ "name": "@etohq/workflows-flows",
3
+ "version": "0.0.1-next-20260318155517",
4
+ "description": "Reusable workflow template library (seed definitions + schemas)",
5
+ "main": "dist/index.js",
6
+ "types": "dist/index.d.ts",
7
+ "exports": {
8
+ ".": {
9
+ "types": "./dist/index.d.ts",
10
+ "import": "./dist/index.js",
11
+ "require": "./dist/index.js"
12
+ }
13
+ },
14
+ "files": [
15
+ "dist"
16
+ ],
17
+ "dependencies": {
18
+ "@etohq/framework": "1.5.5-next-20260317140633",
19
+ "@etohq/workflows-input-dataset": "0.0.1-next-20260318155517",
20
+ "@etohq/workflows-input-dataset-pdf-runtime": "0.0.1-next-20260318155517",
21
+ "@etohq/workflows-input-dataset-runtime": "0.0.1-next-20260318155517",
22
+ "@etohq/workflows-input-schema": "0.0.1-next-20260318155517",
23
+ "@etohq/workflows-input-schema-runtime": "0.0.1-next-20260318155517",
24
+ "@etohq/workflows-types": "0.0.1-next-20260318155517"
25
+ },
26
+ "devDependencies": {
27
+ "@types/node": "22.10.5",
28
+ "rimraf": "5.0.2",
29
+ "typescript": "5.8.3"
30
+ },
31
+ "scripts": {
32
+ "build": "rimraf dist && tsc --build",
33
+ "watch": "tsc --build --watch",
34
+ "test": "exit 0"
35
+ }
36
+ }