@etohq/workflows-flows 0.0.1-next-20260318155517
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +14 -0
- package/dist/dataset/index.d.ts +3 -0
- package/dist/dataset/index.d.ts.map +1 -0
- package/dist/dataset/index.js +19 -0
- package/dist/dataset/index.js.map +1 -0
- package/dist/dataset/steps/convert-csv-to-records.d.ts +6 -0
- package/dist/dataset/steps/convert-csv-to-records.d.ts.map +1 -0
- package/dist/dataset/steps/convert-csv-to-records.js +11 -0
- package/dist/dataset/steps/convert-csv-to-records.js.map +1 -0
- package/dist/dataset/steps/create-dataset-import-run.d.ts +5 -0
- package/dist/dataset/steps/create-dataset-import-run.d.ts.map +1 -0
- package/dist/dataset/steps/create-dataset-import-run.js +12 -0
- package/dist/dataset/steps/create-dataset-import-run.js.map +1 -0
- package/dist/dataset/steps/import-csv-records.d.ts +11 -0
- package/dist/dataset/steps/import-csv-records.d.ts.map +1 -0
- package/dist/dataset/steps/import-csv-records.js +15 -0
- package/dist/dataset/steps/import-csv-records.js.map +1 -0
- package/dist/dataset/steps/index.d.ts +8 -0
- package/dist/dataset/steps/index.d.ts.map +1 -0
- package/dist/dataset/steps/index.js +24 -0
- package/dist/dataset/steps/index.js.map +1 -0
- package/dist/dataset/steps/pdf-to-text.d.ts +6 -0
- package/dist/dataset/steps/pdf-to-text.d.ts.map +1 -0
- package/dist/dataset/steps/pdf-to-text.js +15 -0
- package/dist/dataset/steps/pdf-to-text.js.map +1 -0
- package/dist/dataset/steps/preview-structure-text.d.ts +8 -0
- package/dist/dataset/steps/preview-structure-text.d.ts.map +1 -0
- package/dist/dataset/steps/preview-structure-text.js +11 -0
- package/dist/dataset/steps/preview-structure-text.js.map +1 -0
- package/dist/dataset/steps/retrieve-dataset-draft.d.ts +6 -0
- package/dist/dataset/steps/retrieve-dataset-draft.d.ts.map +1 -0
- package/dist/dataset/steps/retrieve-dataset-draft.js +12 -0
- package/dist/dataset/steps/retrieve-dataset-draft.js.map +1 -0
- package/dist/dataset/steps/update-dataset-import-run.d.ts +5 -0
- package/dist/dataset/steps/update-dataset-import-run.d.ts.map +1 -0
- package/dist/dataset/steps/update-dataset-import-run.js +12 -0
- package/dist/dataset/steps/update-dataset-import-run.js.map +1 -0
- package/dist/dataset/workflows/index.d.ts +4 -0
- package/dist/dataset/workflows/index.d.ts.map +1 -0
- package/dist/dataset/workflows/index.js +20 -0
- package/dist/dataset/workflows/index.js.map +1 -0
- package/dist/dataset/workflows/preview-dataset-from-csv.d.ts +12 -0
- package/dist/dataset/workflows/preview-dataset-from-csv.d.ts.map +1 -0
- package/dist/dataset/workflows/preview-dataset-from-csv.js +16 -0
- package/dist/dataset/workflows/preview-dataset-from-csv.js.map +1 -0
- package/dist/dataset/workflows/preview-dataset-from-pdf.d.ts +18 -0
- package/dist/dataset/workflows/preview-dataset-from-pdf.d.ts.map +1 -0
- package/dist/dataset/workflows/preview-dataset-from-pdf.js +27 -0
- package/dist/dataset/workflows/preview-dataset-from-pdf.js.map +1 -0
- package/dist/dataset/workflows/run-dataset-import-from-csv.d.ts +10 -0
- package/dist/dataset/workflows/run-dataset-import-from-csv.d.ts.map +1 -0
- package/dist/dataset/workflows/run-dataset-import-from-csv.js +63 -0
- package/dist/dataset/workflows/run-dataset-import-from-csv.js.map +1 -0
- package/dist/index.d.ts +3 -0
- package/dist/index.d.ts.map +1 -0
- package/dist/index.js +19 -0
- package/dist/index.js.map +1 -0
- package/dist/templates/index.d.ts +2 -0
- package/dist/templates/index.d.ts.map +1 -0
- package/dist/templates/index.js +5 -0
- package/dist/templates/index.js.map +1 -0
- package/dist/tsconfig.tsbuildinfo +1 -0
- package/package.json +36 -0
package/README.md
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
1
|
+
# `eto-workflows` flows
|
|
2
|
+
|
|
3
|
+
Workflow orchestration layer.
|
|
4
|
+
|
|
5
|
+
This package is where cross-module behavior should live:
|
|
6
|
+
- parsing/preview steps for CSV/PDF imports
|
|
7
|
+
- rule evaluation and validation orchestration
|
|
8
|
+
- multi-step import workflows (draft -> preview -> confirm -> execute)
|
|
9
|
+
|
|
10
|
+
Module services should remain persistence-focused; avoid embedding parsing/transformation logic there.
|
|
11
|
+
|
|
12
|
+
## Dataset Preview Workflows
|
|
13
|
+
- CSV preview: `dataset.preview-from-csv`
|
|
14
|
+
- PDF preview (Node-only): `dataset.preview-from-pdf`
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/dataset/index.ts"],"names":[],"mappings":"AAAA,cAAc,SAAS,CAAA;AACvB,cAAc,aAAa,CAAA"}
|
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
3
|
+
if (k2 === undefined) k2 = k;
|
|
4
|
+
var desc = Object.getOwnPropertyDescriptor(m, k);
|
|
5
|
+
if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
|
|
6
|
+
desc = { enumerable: true, get: function() { return m[k]; } };
|
|
7
|
+
}
|
|
8
|
+
Object.defineProperty(o, k2, desc);
|
|
9
|
+
}) : (function(o, m, k, k2) {
|
|
10
|
+
if (k2 === undefined) k2 = k;
|
|
11
|
+
o[k2] = m[k];
|
|
12
|
+
}));
|
|
13
|
+
var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
14
|
+
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
15
|
+
};
|
|
16
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
|
+
__exportStar(require("./steps"), exports);
|
|
18
|
+
__exportStar(require("./workflows"), exports);
|
|
19
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/dataset/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,0CAAuB;AACvB,8CAA2B"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export type ConvertCsvToRecordsStepInput = {
|
|
2
|
+
csv: string;
|
|
3
|
+
};
|
|
4
|
+
export declare const convertCsvToRecordsStepId = "dataset.convert-csv-to-records";
|
|
5
|
+
export declare const convertCsvToRecordsStep: import("@etohq/framework/workflows-sdk").StepFunction<ConvertCsvToRecordsStepInput, import("@etohq/workflows-input-dataset-runtime").CsvRecord[]>;
|
|
6
|
+
//# sourceMappingURL=convert-csv-to-records.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"convert-csv-to-records.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/convert-csv-to-records.ts"],"names":[],"mappings":"AAGA,MAAM,MAAM,4BAA4B,GAAG;IACzC,GAAG,EAAE,MAAM,CAAA;CACZ,CAAA;AAED,eAAO,MAAM,yBAAyB,mCAAmC,CAAA;AAEzE,eAAO,MAAM,uBAAuB,mJAMnC,CAAA"}
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.convertCsvToRecordsStep = exports.convertCsvToRecordsStepId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const workflows_input_dataset_runtime_1 = require("@etohq/workflows-input-dataset-runtime");
|
|
6
|
+
exports.convertCsvToRecordsStepId = "dataset.convert-csv-to-records";
|
|
7
|
+
exports.convertCsvToRecordsStep = (0, workflows_sdk_1.createStep)(exports.convertCsvToRecordsStepId, async (input) => {
|
|
8
|
+
const records = (0, workflows_input_dataset_runtime_1.convertCsvToRecords)(input.csv);
|
|
9
|
+
return new workflows_sdk_1.StepResponse(records);
|
|
10
|
+
});
|
|
11
|
+
//# sourceMappingURL=convert-csv-to-records.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"convert-csv-to-records.js","sourceRoot":"","sources":["../../../src/dataset/steps/convert-csv-to-records.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AACzE,4FAA4E;AAM/D,QAAA,yBAAyB,GAAG,gCAAgC,CAAA;AAE5D,QAAA,uBAAuB,GAAG,IAAA,0BAAU,EAC/C,iCAAyB,EACzB,KAAK,EAAE,KAAmC,EAAE,EAAE;IAC5C,MAAM,OAAO,GAAG,IAAA,qDAAmB,EAAC,KAAK,CAAC,GAAG,CAAC,CAAA;IAC9C,OAAO,IAAI,4BAAY,CAAC,OAAO,CAAC,CAAA;AAClC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,5 @@
|
|
|
1
|
+
import type { CreateDatasetImportRunDTO } from "@etohq/workflows-types";
|
|
2
|
+
export type CreateDatasetImportRunStepInput = CreateDatasetImportRunDTO;
|
|
3
|
+
export declare const createDatasetImportRunStepId = "dataset.create-import-run";
|
|
4
|
+
export declare const createDatasetImportRunStep: import("@etohq/framework/workflows-sdk").StepFunction<CreateDatasetImportRunDTO, import("@etohq/workflows-types").DatasetImportRunDTO>;
|
|
5
|
+
//# sourceMappingURL=create-dataset-import-run.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"create-dataset-import-run.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/create-dataset-import-run.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,yBAAyB,EAAyB,MAAM,wBAAwB,CAAA;AAG9F,MAAM,MAAM,+BAA+B,GAAG,yBAAyB,CAAA;AAEvE,eAAO,MAAM,4BAA4B,8BAA8B,CAAA;AAEvE,eAAO,MAAM,0BAA0B,wIAUtC,CAAA"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.createDatasetImportRunStep = exports.createDatasetImportRunStepId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const workflows_types_1 = require("@etohq/workflows-types");
|
|
6
|
+
exports.createDatasetImportRunStepId = "dataset.create-import-run";
|
|
7
|
+
exports.createDatasetImportRunStep = (0, workflows_sdk_1.createStep)(exports.createDatasetImportRunStepId, async (input, { container }) => {
|
|
8
|
+
const datasetService = container.resolve(workflows_types_1.WORKFLOWS_MODULE_KEYS.DATASET);
|
|
9
|
+
const run = await datasetService.createDatasetImportRuns(input);
|
|
10
|
+
return new workflows_sdk_1.StepResponse(run);
|
|
11
|
+
});
|
|
12
|
+
//# sourceMappingURL=create-dataset-import-run.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"create-dataset-import-run.js","sourceRoot":"","sources":["../../../src/dataset/steps/create-dataset-import-run.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,4DAA8D;AAIjD,QAAA,4BAA4B,GAAG,2BAA2B,CAAA;AAE1D,QAAA,0BAA0B,GAAG,IAAA,0BAAU,EAClD,oCAA4B,EAC5B,KAAK,EAAE,KAAsC,EAAE,EAAE,SAAS,EAAE,EAAE,EAAE;IAC9D,MAAM,cAAc,GAAG,SAAS,CAAC,OAAO,CACtC,uCAAqB,CAAC,OAAO,CAC9B,CAAA;IAED,MAAM,GAAG,GAAG,MAAM,cAAc,CAAC,uBAAuB,CAAC,KAAK,CAAC,CAAA;IAC/D,OAAO,IAAI,4BAAY,CAAC,GAAG,CAAC,CAAA;AAC9B,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
import type { InputSchemaSpec } from "@etohq/workflows-input-schema";
|
|
2
|
+
import type { DatasetImportSpec } from "@etohq/workflows-input-dataset";
|
|
3
|
+
import type { CsvRecord } from "@etohq/workflows-input-dataset-runtime";
|
|
4
|
+
export type ImportCsvRecordsStepInput = {
|
|
5
|
+
schema: InputSchemaSpec;
|
|
6
|
+
spec: DatasetImportSpec;
|
|
7
|
+
records: CsvRecord[];
|
|
8
|
+
};
|
|
9
|
+
export declare const importCsvRecordsStepId = "dataset.import-csv-records";
|
|
10
|
+
export declare const importCsvRecordsStep: import("@etohq/framework/workflows-sdk").StepFunction<ImportCsvRecordsStepInput, import("@etohq/workflows-input-dataset-runtime").DatasetImportResult>;
|
|
11
|
+
//# sourceMappingURL=import-csv-records.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"import-csv-records.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/import-csv-records.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,+BAA+B,CAAA;AACpE,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,gCAAgC,CAAA;AACvE,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,wCAAwC,CAAA;AAGvE,MAAM,MAAM,yBAAyB,GAAG;IACtC,MAAM,EAAE,eAAe,CAAA;IACvB,IAAI,EAAE,iBAAiB,CAAA;IACvB,OAAO,EAAE,SAAS,EAAE,CAAA;CACrB,CAAA;AAED,eAAO,MAAM,sBAAsB,+BAA+B,CAAA;AAElE,eAAO,MAAM,oBAAoB,wJAUhC,CAAA"}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.importCsvRecordsStep = exports.importCsvRecordsStepId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const workflows_input_dataset_runtime_1 = require("@etohq/workflows-input-dataset-runtime");
|
|
6
|
+
exports.importCsvRecordsStepId = "dataset.import-csv-records";
|
|
7
|
+
exports.importCsvRecordsStep = (0, workflows_sdk_1.createStep)(exports.importCsvRecordsStepId, async (input) => {
|
|
8
|
+
const result = (0, workflows_input_dataset_runtime_1.importCsvRecords)({
|
|
9
|
+
schema: input.schema,
|
|
10
|
+
spec: input.spec,
|
|
11
|
+
records: input.records,
|
|
12
|
+
});
|
|
13
|
+
return new workflows_sdk_1.StepResponse(result);
|
|
14
|
+
});
|
|
15
|
+
//# sourceMappingURL=import-csv-records.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"import-csv-records.js","sourceRoot":"","sources":["../../../src/dataset/steps/import-csv-records.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAIzE,4FAAyE;AAQ5D,QAAA,sBAAsB,GAAG,4BAA4B,CAAA;AAErD,QAAA,oBAAoB,GAAG,IAAA,0BAAU,EAC5C,8BAAsB,EACtB,KAAK,EAAE,KAAgC,EAAE,EAAE;IACzC,MAAM,MAAM,GAAG,IAAA,kDAAgB,EAAC;QAC9B,MAAM,EAAE,KAAK,CAAC,MAAM;QACpB,IAAI,EAAE,KAAK,CAAC,IAAI;QAChB,OAAO,EAAE,KAAK,CAAC,OAAO;KACvB,CAAC,CAAA;IACF,OAAO,IAAI,4BAAY,CAAC,MAAM,CAAC,CAAA;AACjC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
export * from "./convert-csv-to-records";
|
|
2
|
+
export * from "./import-csv-records";
|
|
3
|
+
export * from "./pdf-to-text";
|
|
4
|
+
export * from "./preview-structure-text";
|
|
5
|
+
export * from "./retrieve-dataset-draft";
|
|
6
|
+
export * from "./create-dataset-import-run";
|
|
7
|
+
export * from "./update-dataset-import-run";
|
|
8
|
+
//# sourceMappingURL=index.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/index.ts"],"names":[],"mappings":"AAAA,cAAc,0BAA0B,CAAA;AACxC,cAAc,sBAAsB,CAAA;AACpC,cAAc,eAAe,CAAA;AAC7B,cAAc,0BAA0B,CAAA;AACxC,cAAc,0BAA0B,CAAA;AACxC,cAAc,6BAA6B,CAAA;AAC3C,cAAc,6BAA6B,CAAA"}
|
|
@@ -0,0 +1,24 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
3
|
+
if (k2 === undefined) k2 = k;
|
|
4
|
+
var desc = Object.getOwnPropertyDescriptor(m, k);
|
|
5
|
+
if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
|
|
6
|
+
desc = { enumerable: true, get: function() { return m[k]; } };
|
|
7
|
+
}
|
|
8
|
+
Object.defineProperty(o, k2, desc);
|
|
9
|
+
}) : (function(o, m, k, k2) {
|
|
10
|
+
if (k2 === undefined) k2 = k;
|
|
11
|
+
o[k2] = m[k];
|
|
12
|
+
}));
|
|
13
|
+
var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
14
|
+
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
15
|
+
};
|
|
16
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
|
+
__exportStar(require("./convert-csv-to-records"), exports);
|
|
18
|
+
__exportStar(require("./import-csv-records"), exports);
|
|
19
|
+
__exportStar(require("./pdf-to-text"), exports);
|
|
20
|
+
__exportStar(require("./preview-structure-text"), exports);
|
|
21
|
+
__exportStar(require("./retrieve-dataset-draft"), exports);
|
|
22
|
+
__exportStar(require("./create-dataset-import-run"), exports);
|
|
23
|
+
__exportStar(require("./update-dataset-import-run"), exports);
|
|
24
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../src/dataset/steps/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,2DAAwC;AACxC,uDAAoC;AACpC,gDAA6B;AAC7B,2DAAwC;AACxC,2DAAwC;AACxC,8DAA2C;AAC3C,8DAA2C"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export type PdfToTextStepInput = {
|
|
2
|
+
pdf_bytes: Uint8Array;
|
|
3
|
+
};
|
|
4
|
+
export declare const pdfToTextStepId = "dataset.pdf-to-text";
|
|
5
|
+
export declare const pdfToTextStep: import("@etohq/framework/workflows-sdk").StepFunction<PdfToTextStepInput, string>;
|
|
6
|
+
//# sourceMappingURL=pdf-to-text.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"pdf-to-text.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/pdf-to-text.ts"],"names":[],"mappings":"AAIA,MAAM,MAAM,kBAAkB,GAAG;IAC/B,SAAS,EAAE,UAAU,CAAA;CACtB,CAAA;AAED,eAAO,MAAM,eAAe,wBAAwB,CAAA;AAEpD,eAAO,MAAM,aAAa,mFASzB,CAAA"}
|
|
@@ -0,0 +1,15 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.pdfToTextStep = exports.pdfToTextStepId = void 0;
|
|
4
|
+
const utils_1 = require("@etohq/framework/utils");
|
|
5
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
6
|
+
const workflows_input_dataset_pdf_runtime_1 = require("@etohq/workflows-input-dataset-pdf-runtime");
|
|
7
|
+
exports.pdfToTextStepId = "dataset.pdf-to-text";
|
|
8
|
+
exports.pdfToTextStep = (0, workflows_sdk_1.createStep)(exports.pdfToTextStepId, async (input) => {
|
|
9
|
+
const res = await (0, workflows_input_dataset_pdf_runtime_1.pdfToText)(input.pdf_bytes);
|
|
10
|
+
if (!res.ok) {
|
|
11
|
+
throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, res.error);
|
|
12
|
+
}
|
|
13
|
+
return new workflows_sdk_1.StepResponse(res.text);
|
|
14
|
+
});
|
|
15
|
+
//# sourceMappingURL=pdf-to-text.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"pdf-to-text.js","sourceRoot":"","sources":["../../../src/dataset/steps/pdf-to-text.ts"],"names":[],"mappings":";;;AAAA,kDAAiD;AACjD,kEAAyE;AACzE,oGAAsE;AAMzD,QAAA,eAAe,GAAG,qBAAqB,CAAA;AAEvC,QAAA,aAAa,GAAG,IAAA,0BAAU,EACrC,uBAAe,EACf,KAAK,EAAE,KAAyB,EAAE,EAAE;IAClC,MAAM,GAAG,GAAG,MAAM,IAAA,+CAAS,EAAC,KAAK,CAAC,SAAS,CAAC,CAAA;IAC5C,IAAI,CAAC,GAAG,CAAC,EAAE,EAAE,CAAC;QACZ,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,GAAG,CAAC,KAAK,CAAC,CAAA;IAC5D,CAAC;IACD,OAAO,IAAI,4BAAY,CAAC,GAAG,CAAC,IAAI,CAAC,CAAA;AACnC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
import type { PdfExtractSpec } from "@etohq/workflows-input-dataset";
|
|
2
|
+
export type PreviewStructureTextStepInput = {
|
|
3
|
+
text: string;
|
|
4
|
+
spec: PdfExtractSpec;
|
|
5
|
+
};
|
|
6
|
+
export declare const previewStructureTextStepId = "dataset.preview-structure-text";
|
|
7
|
+
export declare const previewStructureTextStep: import("@etohq/framework/workflows-sdk").StepFunction<PreviewStructureTextStepInput, import("@etohq/workflows-input-dataset-pdf-runtime").StructureTextPreviewResult>;
|
|
8
|
+
//# sourceMappingURL=preview-structure-text.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"preview-structure-text.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/preview-structure-text.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAGpE,MAAM,MAAM,6BAA6B,GAAG;IAC1C,IAAI,EAAE,MAAM,CAAA;IACZ,IAAI,EAAE,cAAc,CAAA;CACrB,CAAA;AAED,eAAO,MAAM,0BAA0B,mCAAmC,CAAA;AAE1E,eAAO,MAAM,wBAAwB,uKAMpC,CAAA"}
|
|
@@ -0,0 +1,11 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.previewStructureTextStep = exports.previewStructureTextStepId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const workflows_input_dataset_pdf_runtime_1 = require("@etohq/workflows-input-dataset-pdf-runtime");
|
|
6
|
+
exports.previewStructureTextStepId = "dataset.preview-structure-text";
|
|
7
|
+
exports.previewStructureTextStep = (0, workflows_sdk_1.createStep)(exports.previewStructureTextStepId, async (input) => {
|
|
8
|
+
const result = (0, workflows_input_dataset_pdf_runtime_1.previewStructureRecordsFromText)(input.text, input.spec);
|
|
9
|
+
return new workflows_sdk_1.StepResponse(result);
|
|
10
|
+
});
|
|
11
|
+
//# sourceMappingURL=preview-structure-text.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"preview-structure-text.js","sourceRoot":"","sources":["../../../src/dataset/steps/preview-structure-text.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,oGAA4F;AAO/E,QAAA,0BAA0B,GAAG,gCAAgC,CAAA;AAE7D,QAAA,wBAAwB,GAAG,IAAA,0BAAU,EAChD,kCAA0B,EAC1B,KAAK,EAAE,KAAoC,EAAE,EAAE;IAC7C,MAAM,MAAM,GAAG,IAAA,qEAA+B,EAAC,KAAK,CAAC,IAAI,EAAE,KAAK,CAAC,IAAI,CAAC,CAAA;IACtE,OAAO,IAAI,4BAAY,CAAC,MAAM,CAAC,CAAA;AACjC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,6 @@
|
|
|
1
|
+
export type RetrieveDatasetDraftStepInput = {
|
|
2
|
+
id: string;
|
|
3
|
+
};
|
|
4
|
+
export declare const retrieveDatasetDraftStepId = "dataset.retrieve-dataset-draft";
|
|
5
|
+
export declare const retrieveDatasetDraftStep: import("@etohq/framework/workflows-sdk").StepFunction<RetrieveDatasetDraftStepInput, import("@etohq/workflows-types").DatasetDraftDTO>;
|
|
6
|
+
//# sourceMappingURL=retrieve-dataset-draft.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"retrieve-dataset-draft.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/retrieve-dataset-draft.ts"],"names":[],"mappings":"AAIA,MAAM,MAAM,6BAA6B,GAAG;IAC1C,EAAE,EAAE,MAAM,CAAA;CACX,CAAA;AAED,eAAO,MAAM,0BAA0B,mCAAmC,CAAA;AAE1E,eAAO,MAAM,wBAAwB,wIAUpC,CAAA"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.retrieveDatasetDraftStep = exports.retrieveDatasetDraftStepId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const workflows_types_1 = require("@etohq/workflows-types");
|
|
6
|
+
exports.retrieveDatasetDraftStepId = "dataset.retrieve-dataset-draft";
|
|
7
|
+
exports.retrieveDatasetDraftStep = (0, workflows_sdk_1.createStep)(exports.retrieveDatasetDraftStepId, async (input, { container }) => {
|
|
8
|
+
const datasetService = container.resolve(workflows_types_1.WORKFLOWS_MODULE_KEYS.DATASET);
|
|
9
|
+
const draft = await datasetService.retrieveDatasetDraft(input.id);
|
|
10
|
+
return new workflows_sdk_1.StepResponse(draft);
|
|
11
|
+
});
|
|
12
|
+
//# sourceMappingURL=retrieve-dataset-draft.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"retrieve-dataset-draft.js","sourceRoot":"","sources":["../../../src/dataset/steps/retrieve-dataset-draft.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,4DAA8D;AAMjD,QAAA,0BAA0B,GAAG,gCAAgC,CAAA;AAE7D,QAAA,wBAAwB,GAAG,IAAA,0BAAU,EAChD,kCAA0B,EAC1B,KAAK,EAAE,KAAoC,EAAE,EAAE,SAAS,EAAE,EAAE,EAAE;IAC5D,MAAM,cAAc,GAAG,SAAS,CAAC,OAAO,CACtC,uCAAqB,CAAC,OAAO,CAC9B,CAAA;IAED,MAAM,KAAK,GAAG,MAAM,cAAc,CAAC,oBAAoB,CAAC,KAAK,CAAC,EAAE,CAAC,CAAA;IACjE,OAAO,IAAI,4BAAY,CAAC,KAAK,CAAC,CAAA;AAChC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,5 @@
|
|
|
1
|
+
import type { UpdateDatasetImportRunRequestDTO } from "@etohq/workflows-types";
|
|
2
|
+
export type UpdateDatasetImportRunStepInput = UpdateDatasetImportRunRequestDTO;
|
|
3
|
+
export declare const updateDatasetImportRunStepId = "dataset.update-import-run";
|
|
4
|
+
export declare const updateDatasetImportRunStep: import("@etohq/framework/workflows-sdk").StepFunction<UpdateDatasetImportRunRequestDTO, import("@etohq/workflows-types").DatasetImportRunDTO>;
|
|
5
|
+
//# sourceMappingURL=update-dataset-import-run.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"update-dataset-import-run.d.ts","sourceRoot":"","sources":["../../../src/dataset/steps/update-dataset-import-run.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAyB,gCAAgC,EAAE,MAAM,wBAAwB,CAAA;AAGrG,MAAM,MAAM,+BAA+B,GAAG,gCAAgC,CAAA;AAE9E,eAAO,MAAM,4BAA4B,8BAA8B,CAAA;AAEvE,eAAO,MAAM,0BAA0B,+IAUtC,CAAA"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.updateDatasetImportRunStep = exports.updateDatasetImportRunStepId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const workflows_types_1 = require("@etohq/workflows-types");
|
|
6
|
+
exports.updateDatasetImportRunStepId = "dataset.update-import-run";
|
|
7
|
+
exports.updateDatasetImportRunStep = (0, workflows_sdk_1.createStep)(exports.updateDatasetImportRunStepId, async (input, { container }) => {
|
|
8
|
+
const datasetService = container.resolve(workflows_types_1.WORKFLOWS_MODULE_KEYS.DATASET);
|
|
9
|
+
const updated = await datasetService.updateDatasetImportRuns(input);
|
|
10
|
+
return new workflows_sdk_1.StepResponse(updated);
|
|
11
|
+
});
|
|
12
|
+
//# sourceMappingURL=update-dataset-import-run.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"update-dataset-import-run.js","sourceRoot":"","sources":["../../../src/dataset/steps/update-dataset-import-run.ts"],"names":[],"mappings":";;;AAAA,kEAAyE;AAEzE,4DAA8D;AAIjD,QAAA,4BAA4B,GAAG,2BAA2B,CAAA;AAE1D,QAAA,0BAA0B,GAAG,IAAA,0BAAU,EAClD,oCAA4B,EAC5B,KAAK,EAAE,KAAsC,EAAE,EAAE,SAAS,EAAE,EAAE,EAAE;IAC9D,MAAM,cAAc,GAAG,SAAS,CAAC,OAAO,CACtC,uCAAqB,CAAC,OAAO,CAC9B,CAAA;IAED,MAAM,OAAO,GAAG,MAAM,cAAc,CAAC,uBAAuB,CAAC,KAAK,CAAC,CAAA;IACnE,OAAO,IAAI,4BAAY,CAAC,OAAO,CAAC,CAAA;AAClC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/index.ts"],"names":[],"mappings":"AAAA,cAAc,4BAA4B,CAAA;AAC1C,cAAc,4BAA4B,CAAA;AAC1C,cAAc,+BAA+B,CAAA"}
|
|
@@ -0,0 +1,20 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
3
|
+
if (k2 === undefined) k2 = k;
|
|
4
|
+
var desc = Object.getOwnPropertyDescriptor(m, k);
|
|
5
|
+
if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
|
|
6
|
+
desc = { enumerable: true, get: function() { return m[k]; } };
|
|
7
|
+
}
|
|
8
|
+
Object.defineProperty(o, k2, desc);
|
|
9
|
+
}) : (function(o, m, k, k2) {
|
|
10
|
+
if (k2 === undefined) k2 = k;
|
|
11
|
+
o[k2] = m[k];
|
|
12
|
+
}));
|
|
13
|
+
var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
14
|
+
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
15
|
+
};
|
|
16
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
|
+
__exportStar(require("./preview-dataset-from-csv"), exports);
|
|
18
|
+
__exportStar(require("./preview-dataset-from-pdf"), exports);
|
|
19
|
+
__exportStar(require("./run-dataset-import-from-csv"), exports);
|
|
20
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../src/dataset/workflows/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,6DAA0C;AAC1C,6DAA0C;AAC1C,gEAA6C"}
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
import type { ReturnWorkflow } from "@etohq/framework/workflows-sdk";
|
|
2
|
+
import type { InputSchemaSpec } from "@etohq/workflows-input-schema";
|
|
3
|
+
import type { DatasetImportSpec } from "@etohq/workflows-input-dataset";
|
|
4
|
+
import type { DatasetImportResult } from "@etohq/workflows-input-dataset-runtime";
|
|
5
|
+
export type PreviewDatasetFromCsvWorkflowInput = {
|
|
6
|
+
schema: InputSchemaSpec;
|
|
7
|
+
spec: DatasetImportSpec;
|
|
8
|
+
csv: string;
|
|
9
|
+
};
|
|
10
|
+
export declare const previewDatasetFromCsvWorkflowId = "dataset.preview-from-csv";
|
|
11
|
+
export declare const previewDatasetFromCsvWorkflow: ReturnWorkflow<PreviewDatasetFromCsvWorkflowInput, DatasetImportResult, unknown[]>;
|
|
12
|
+
//# sourceMappingURL=preview-dataset-from-csv.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"preview-dataset-from-csv.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-csv.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAEpE,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,+BAA+B,CAAA;AACpE,OAAO,KAAK,EAAE,iBAAiB,EAAE,MAAM,gCAAgC,CAAA;AAEvE,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AAEjF,MAAM,MAAM,kCAAkC,GAAG;IAC/C,MAAM,EAAE,eAAe,CAAA;IACvB,IAAI,EAAE,iBAAiB,CAAA;IACvB,GAAG,EAAE,MAAM,CAAA;CACZ,CAAA;AAED,eAAO,MAAM,+BAA+B,6BAA6B,CAAA;AAEzE,eAAO,MAAM,6BAA6B,EAAE,cAAc,CACxD,kCAAkC,EAClC,mBAAmB,EACnB,OAAO,EAAE,CAkBV,CAAA"}
|
|
@@ -0,0 +1,16 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.previewDatasetFromCsvWorkflow = exports.previewDatasetFromCsvWorkflowId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const steps_1 = require("../steps");
|
|
6
|
+
exports.previewDatasetFromCsvWorkflowId = "dataset.preview-from-csv";
|
|
7
|
+
exports.previewDatasetFromCsvWorkflow = (0, workflows_sdk_1.createWorkflow)(exports.previewDatasetFromCsvWorkflowId, function (input) {
|
|
8
|
+
const records = (0, steps_1.convertCsvToRecordsStep)({ csv: input.csv });
|
|
9
|
+
const imported = (0, steps_1.importCsvRecordsStep)((0, workflows_sdk_1.transform)({ input, records }, ({ input, records }) => ({
|
|
10
|
+
schema: input.schema,
|
|
11
|
+
spec: input.spec,
|
|
12
|
+
records,
|
|
13
|
+
})));
|
|
14
|
+
return new workflows_sdk_1.WorkflowResponse(imported);
|
|
15
|
+
});
|
|
16
|
+
//# sourceMappingURL=preview-dataset-from-csv.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"preview-dataset-from-csv.js","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-csv.ts"],"names":[],"mappings":";;;AACA,kEAA4F;AAG5F,oCAAwE;AAS3D,QAAA,+BAA+B,GAAG,0BAA0B,CAAA;AAE5D,QAAA,6BAA6B,GAItC,IAAA,8BAAc,EAChB,uCAA+B,EAC/B,UACE,KAAyC;IAEzC,MAAM,OAAO,GAAG,IAAA,+BAAuB,EAAC,EAAE,GAAG,EAAE,KAAK,CAAC,GAAG,EAAE,CAAC,CAAA;IAE3D,MAAM,QAAQ,GAAG,IAAA,4BAAoB,EACnC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,OAAO,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC,CAAC;QACrD,MAAM,EAAE,KAAK,CAAC,MAAM;QACpB,IAAI,EAAE,KAAK,CAAC,IAAI;QAChB,OAAO;KACR,CAAC,CAAC,CACJ,CAAA;IAED,OAAO,IAAI,gCAAgB,CAAC,QAAQ,CAAC,CAAA;AACvC,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,18 @@
|
|
|
1
|
+
import type { ReturnWorkflow } from "@etohq/framework/workflows-sdk";
|
|
2
|
+
import type { InputSchemaSpec } from "@etohq/workflows-input-schema";
|
|
3
|
+
import type { DatasetImportSpec, PdfExtractSpec } from "@etohq/workflows-input-dataset";
|
|
4
|
+
export type PreviewDatasetFromPdfWorkflowInput = {
|
|
5
|
+
schema: InputSchemaSpec;
|
|
6
|
+
import_spec: DatasetImportSpec;
|
|
7
|
+
extract_spec: PdfExtractSpec;
|
|
8
|
+
pdf_bytes: Uint8Array;
|
|
9
|
+
};
|
|
10
|
+
export declare const previewDatasetFromPdfWorkflowId = "dataset.preview-from-pdf";
|
|
11
|
+
/**
|
|
12
|
+
* Node-only workflow: uses `pdf-parse` under the hood.
|
|
13
|
+
*
|
|
14
|
+
* Output is a preview-friendly structure result + an import result that maps extracted records
|
|
15
|
+
* into schema submissions and validation issues.
|
|
16
|
+
*/
|
|
17
|
+
export declare const previewDatasetFromPdfWorkflow: ReturnWorkflow<PreviewDatasetFromPdfWorkflowInput, unknown, unknown[]>;
|
|
18
|
+
//# sourceMappingURL=preview-dataset-from-pdf.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"preview-dataset-from-pdf.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-pdf.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAEpE,OAAO,KAAK,EAAE,eAAe,EAAE,MAAM,+BAA+B,CAAA;AACpE,OAAO,KAAK,EAAE,iBAAiB,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAGvF,MAAM,MAAM,kCAAkC,GAAG;IAC/C,MAAM,EAAE,eAAe,CAAA;IACvB,WAAW,EAAE,iBAAiB,CAAA;IAC9B,YAAY,EAAE,cAAc,CAAA;IAC5B,SAAS,EAAE,UAAU,CAAA;CACtB,CAAA;AAED,eAAO,MAAM,+BAA+B,6BAA6B,CAAA;AAEzE;;;;;GAKG;AACH,eAAO,MAAM,6BAA6B,EAAE,cAAc,CACxD,kCAAkC,EAClC,OAAO,EACP,OAAO,EAAE,CA0BV,CAAA"}
|
|
@@ -0,0 +1,27 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.previewDatasetFromPdfWorkflow = exports.previewDatasetFromPdfWorkflowId = void 0;
|
|
4
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
5
|
+
const steps_1 = require("../steps");
|
|
6
|
+
exports.previewDatasetFromPdfWorkflowId = "dataset.preview-from-pdf";
|
|
7
|
+
/**
|
|
8
|
+
* Node-only workflow: uses `pdf-parse` under the hood.
|
|
9
|
+
*
|
|
10
|
+
* Output is a preview-friendly structure result + an import result that maps extracted records
|
|
11
|
+
* into schema submissions and validation issues.
|
|
12
|
+
*/
|
|
13
|
+
exports.previewDatasetFromPdfWorkflow = (0, workflows_sdk_1.createWorkflow)(exports.previewDatasetFromPdfWorkflowId, function (input) {
|
|
14
|
+
const text = (0, steps_1.pdfToTextStep)({ pdf_bytes: input.pdf_bytes });
|
|
15
|
+
const structured = (0, steps_1.previewStructureTextStep)((0, workflows_sdk_1.transform)({ input, text }, ({ input, text }) => ({
|
|
16
|
+
text,
|
|
17
|
+
spec: input.extract_spec,
|
|
18
|
+
})));
|
|
19
|
+
const imported = (0, steps_1.importCsvRecordsStep)((0, workflows_sdk_1.transform)({ input, structured }, ({ input, structured }) => ({
|
|
20
|
+
schema: input.schema,
|
|
21
|
+
spec: input.import_spec,
|
|
22
|
+
records: structured.records,
|
|
23
|
+
})));
|
|
24
|
+
// Intentionally keep result opaque at this layer to avoid leaking complex WorkflowData unions.
|
|
25
|
+
return new workflows_sdk_1.WorkflowResponse({ structured, imported });
|
|
26
|
+
});
|
|
27
|
+
//# sourceMappingURL=preview-dataset-from-pdf.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"preview-dataset-from-pdf.js","sourceRoot":"","sources":["../../../src/dataset/workflows/preview-dataset-from-pdf.ts"],"names":[],"mappings":";;;AACA,kEAA4F;AAG5F,oCAAwF;AAS3E,QAAA,+BAA+B,GAAG,0BAA0B,CAAA;AAEzE;;;;;GAKG;AACU,QAAA,6BAA6B,GAItC,IAAA,8BAAc,EAChB,uCAA+B,EAC/B,UACE,KAAyC;IAEzC,MAAM,IAAI,GAAG,IAAA,qBAAa,EAAC,EAAE,SAAS,EAAE,KAAK,CAAC,SAAS,EAAE,CAAC,CAAA;IAE1D,MAAM,UAAU,GAAG,IAAA,gCAAwB,EACzC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,IAAI,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,IAAI,EAAE,EAAE,EAAE,CAAC,CAAC;QAC/C,IAAI;QACJ,IAAI,EAAE,KAAK,CAAC,YAAY;KACzB,CAAC,CAAC,CACJ,CAAA;IAED,MAAM,QAAQ,GAAG,IAAA,4BAAoB,EACnC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,UAAU,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,UAAU,EAAE,EAAE,EAAE,CAAC,CAAC;QAC3D,MAAM,EAAE,KAAK,CAAC,MAAM;QACpB,IAAI,EAAE,KAAK,CAAC,WAAW;QACvB,OAAO,EAAE,UAAU,CAAC,OAAO;KAC5B,CAAC,CAAC,CACJ,CAAA;IAED,+FAA+F;IAC/F,OAAO,IAAI,gCAAgB,CAAC,EAAE,UAAU,EAAE,QAAQ,EAAQ,CAAC,CAAA;AAC7D,CAAC,CACF,CAAA"}
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
import type { ReturnWorkflow } from "@etohq/framework/workflows-sdk";
|
|
2
|
+
import type { DatasetImportResult } from "@etohq/workflows-input-dataset-runtime";
|
|
3
|
+
export type RunDatasetImportFromCsvWorkflowInput = {
|
|
4
|
+
tenant_id: string;
|
|
5
|
+
draft_id: string;
|
|
6
|
+
csv: string;
|
|
7
|
+
};
|
|
8
|
+
export declare const runDatasetImportFromCsvWorkflowId = "dataset.run-import-from-csv";
|
|
9
|
+
export declare const runDatasetImportFromCsvWorkflow: ReturnWorkflow<RunDatasetImportFromCsvWorkflowInput, DatasetImportResult, unknown[]>;
|
|
10
|
+
//# sourceMappingURL=run-dataset-import-from-csv.d.ts.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"run-dataset-import-from-csv.d.ts","sourceRoot":"","sources":["../../../src/dataset/workflows/run-dataset-import-from-csv.ts"],"names":[],"mappings":"AACA,OAAO,KAAK,EAAE,cAAc,EAAE,MAAM,gCAAgC,CAAA;AAEpE,OAAO,KAAK,EAAE,mBAAmB,EAAE,MAAM,wCAAwC,CAAA;AAQjF,MAAM,MAAM,oCAAoC,GAAG;IACjD,SAAS,EAAE,MAAM,CAAA;IACjB,QAAQ,EAAE,MAAM,CAAA;IAChB,GAAG,EAAE,MAAM,CAAA;CACZ,CAAA;AAED,eAAO,MAAM,iCAAiC,gCAAgC,CAAA;AAE9E,eAAO,MAAM,+BAA+B,EAAE,cAAc,CAC1D,oCAAoC,EACpC,mBAAmB,EACnB,OAAO,EAAE,CAyEV,CAAA"}
|
|
@@ -0,0 +1,63 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.runDatasetImportFromCsvWorkflow = exports.runDatasetImportFromCsvWorkflowId = void 0;
|
|
4
|
+
const utils_1 = require("@etohq/framework/utils");
|
|
5
|
+
const workflows_sdk_1 = require("@etohq/framework/workflows-sdk");
|
|
6
|
+
const steps_1 = require("../steps");
|
|
7
|
+
const steps_2 = require("../steps");
|
|
8
|
+
exports.runDatasetImportFromCsvWorkflowId = "dataset.run-import-from-csv";
|
|
9
|
+
exports.runDatasetImportFromCsvWorkflow = (0, workflows_sdk_1.createWorkflow)(exports.runDatasetImportFromCsvWorkflowId, function (input) {
|
|
10
|
+
const draft = (0, steps_2.retrieveDatasetDraftStep)({ id: input.draft_id });
|
|
11
|
+
const schema = (0, workflows_sdk_1.transform)({ draft }, ({ draft }) => {
|
|
12
|
+
if (!draft.schema_spec) {
|
|
13
|
+
throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.schema_spec is required");
|
|
14
|
+
}
|
|
15
|
+
if (!draft.dataset_import_spec) {
|
|
16
|
+
throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.dataset_import_spec is required");
|
|
17
|
+
}
|
|
18
|
+
if (draft.kind !== "csv") {
|
|
19
|
+
throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.kind must be csv");
|
|
20
|
+
}
|
|
21
|
+
if (draft.status !== "ready") {
|
|
22
|
+
throw new utils_1.EtoError(utils_1.EtoError.Types.INVALID_DATA, "draft.status must be ready");
|
|
23
|
+
}
|
|
24
|
+
return { schema: draft.schema_spec, spec: draft.dataset_import_spec, revision: draft.revision };
|
|
25
|
+
});
|
|
26
|
+
const run = (0, steps_2.createDatasetImportRunStep)((0, workflows_sdk_1.transform)({ input, draft, schema }, ({ input, draft, schema }) => ({
|
|
27
|
+
tenant_id: input.tenant_id,
|
|
28
|
+
status: "running",
|
|
29
|
+
source_id: draft.source_id,
|
|
30
|
+
draft_id: draft.id,
|
|
31
|
+
draft_revision: schema.revision,
|
|
32
|
+
records_total: null,
|
|
33
|
+
records_succeeded: null,
|
|
34
|
+
records_failed: null,
|
|
35
|
+
started_at: null,
|
|
36
|
+
finished_at: null,
|
|
37
|
+
error_summary: null,
|
|
38
|
+
metadata: {},
|
|
39
|
+
})));
|
|
40
|
+
const records = (0, steps_1.convertCsvToRecordsStep)({ csv: input.csv });
|
|
41
|
+
const imported = (0, steps_1.importCsvRecordsStep)((0, workflows_sdk_1.transform)({ schema, records }, ({ schema, records }) => ({
|
|
42
|
+
schema: schema.schema,
|
|
43
|
+
spec: schema.spec,
|
|
44
|
+
records,
|
|
45
|
+
})));
|
|
46
|
+
const summary = (0, workflows_sdk_1.transform)({ imported }, ({ imported }) => {
|
|
47
|
+
const total = imported.rows.length;
|
|
48
|
+
const failed = imported.rows.filter((r) => r.validation.status === "invalid").length;
|
|
49
|
+
const succeeded = total - failed;
|
|
50
|
+
return { total, succeeded, failed, status: imported.status };
|
|
51
|
+
});
|
|
52
|
+
const errorSummary = (0, workflows_sdk_1.transform)({ imported, summary }, ({ imported, summary }) => summary.status === "invalid" ? { issues: imported.issues } : null);
|
|
53
|
+
(0, steps_2.updateDatasetImportRunStep)((0, workflows_sdk_1.transform)({ run, summary, errorSummary }, ({ run, summary, errorSummary }) => ({
|
|
54
|
+
id: run.id,
|
|
55
|
+
status: summary.status === "invalid" ? "failed" : "succeeded",
|
|
56
|
+
records_total: summary.total,
|
|
57
|
+
records_succeeded: summary.succeeded,
|
|
58
|
+
records_failed: summary.failed,
|
|
59
|
+
error_summary: errorSummary,
|
|
60
|
+
})));
|
|
61
|
+
return new workflows_sdk_1.WorkflowResponse(imported);
|
|
62
|
+
});
|
|
63
|
+
//# sourceMappingURL=run-dataset-import-from-csv.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"run-dataset-import-from-csv.js","sourceRoot":"","sources":["../../../src/dataset/workflows/run-dataset-import-from-csv.ts"],"names":[],"mappings":";;;AAAA,kDAAiD;AAEjD,kEAA4F;AAE5F,oCAAwE;AACxE,oCAIiB;AAQJ,QAAA,iCAAiC,GAAG,6BAA6B,CAAA;AAEjE,QAAA,+BAA+B,GAIxC,IAAA,8BAAc,EAChB,yCAAiC,EACjC,UAAU,KAA2C;IACnD,MAAM,KAAK,GAAG,IAAA,gCAAwB,EAAC,EAAE,EAAE,EAAE,KAAK,CAAC,QAAQ,EAAE,CAAC,CAAA;IAE9D,MAAM,MAAM,GAAG,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,EAAE,EAAE;QAChD,IAAI,CAAC,KAAK,CAAC,WAAW,EAAE,CAAC;YACvB,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,+BAA+B,CAAC,CAAA;QAClF,CAAC;QACD,IAAI,CAAC,KAAK,CAAC,mBAAmB,EAAE,CAAC;YAC/B,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,uCAAuC,CAAC,CAAA;QAC1F,CAAC;QACD,IAAI,KAAK,CAAC,IAAI,KAAK,KAAK,EAAE,CAAC;YACzB,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,wBAAwB,CAAC,CAAA;QAC3E,CAAC;QACD,IAAI,KAAK,CAAC,MAAM,KAAK,OAAO,EAAE,CAAC;YAC7B,MAAM,IAAI,gBAAQ,CAAC,gBAAQ,CAAC,KAAK,CAAC,YAAY,EAAE,4BAA4B,CAAC,CAAA;QAC/E,CAAC;QACD,OAAO,EAAE,MAAM,EAAE,KAAK,CAAC,WAAW,EAAE,IAAI,EAAE,KAAK,CAAC,mBAAmB,EAAE,QAAQ,EAAE,KAAK,CAAC,QAAQ,EAAE,CAAA;IACjG,CAAC,CAAC,CAAA;IAEF,MAAM,GAAG,GAAG,IAAA,kCAA0B,EACpC,IAAA,yBAAS,EAAC,EAAE,KAAK,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,CAAC,EAAE,KAAK,EAAE,KAAK,EAAE,MAAM,EAAE,EAAE,EAAE,CAAC,CAAC;QACjE,SAAS,EAAE,KAAK,CAAC,SAAS;QAC1B,MAAM,EAAE,SAAS;QACjB,SAAS,EAAE,KAAK,CAAC,SAAS;QAC1B,QAAQ,EAAE,KAAK,CAAC,EAAE;QAClB,cAAc,EAAE,MAAM,CAAC,QAAQ;QAC/B,aAAa,EAAE,IAAI;QACnB,iBAAiB,EAAE,IAAI;QACvB,cAAc,EAAE,IAAI;QACpB,UAAU,EAAE,IAAI;QAChB,WAAW,EAAE,IAAI;QACjB,aAAa,EAAE,IAAI;QACnB,QAAQ,EAAE,EAAE;KACb,CAAC,CAAC,CACJ,CAAA;IAED,MAAM,OAAO,GAAG,IAAA,+BAAuB,EAAC,EAAE,GAAG,EAAE,KAAK,CAAC,GAAG,EAAE,CAAC,CAAA;IAE3D,MAAM,QAAQ,GAAG,IAAA,4BAAoB,EACnC,IAAA,yBAAS,EAAC,EAAE,MAAM,EAAE,OAAO,EAAE,EAAE,CAAC,EAAE,MAAM,EAAE,OAAO,EAAE,EAAE,EAAE,CAAC,CAAC;QACvD,MAAM,EAAE,MAAM,CAAC,MAAM;QACrB,IAAI,EAAE,MAAM,CAAC,IAAI;QACjB,OAAO;KACR,CAAC,CAAC,CACJ,CAAA;IAED,MAAM,OAAO,GAAG,IAAA,yBAAS,EAAC,EAAE,QAAQ,EAAE,EAAE,CAAC,EAAE,QAAQ,EAAE,EAAE,EAAE;QACvD,MAAM,KAAK,GAAG,QAAQ,CAAC,IAAI,CAAC,MAAM,CAAA;QAClC,MAAM,MAAM,GAAG,QAAQ,CAAC,IAAI,CAAC,MAAM,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,UAAU,CAAC,MAAM,KAAK,SAAS,CAAC,CAAC,MAAM,CAAA;QACpF,MAAM,SAAS,GAAG,KAAK,GAAG,MAAM,CAAA;QAChC,OAAO,EAAE,KAAK,EAAE,SAAS,EAAE,MAAM,EAAE,MAAM,EAAE,QAAQ,CAAC,MAAM,EAAE,CAAA;IAC9D,CAAC,CAAC,CAAA;IAEF,MAAM,YAAY,GAAG,IAAA,yBAAS,EAAC,EAAE,QAAQ,EAAE,OAAO,EAAE,EAAE,CAAC,EAAE,QAAQ,EAAE,OAAO,EAAE,EAAE,EAAE,CAC9E,OAAO,CAAC,MAAM,KAAK,SAAS,CAAC,CAAC,CAAE,EAAE,MAAM,EAAE,QAAQ,CAAC,MAAM,EAAS,CAAC,CAAC,CAAC,IAAI,CAC1E,CAAA;IAED,IAAA,kCAA0B,EACxB,IAAA,yBAAS,EAAC,EAAE,GAAG,EAAE,OAAO,EAAE,YAAY,EAAE,EAAE,CAAC,EAAE,GAAG,EAAE,OAAO,EAAE,YAAY,EAAE,EAAE,EAAE,CAAC,CAAC;QAC7E,EAAE,EAAE,GAAG,CAAC,EAAE;QACV,MAAM,EAAE,OAAO,CAAC,MAAM,KAAK,SAAS,CAAC,CAAC,CAAC,QAAQ,CAAC,CAAC,CAAC,WAAW;QAC7D,aAAa,EAAE,OAAO,CAAC,KAAK;QAC5B,iBAAiB,EAAE,OAAO,CAAC,SAAS;QACpC,cAAc,EAAE,OAAO,CAAC,MAAM;QAC9B,aAAa,EAAE,YAAY;KAC5B,CAAC,CAAC,CACJ,CAAA;IAED,OAAO,IAAI,gCAAgB,CAAC,QAAQ,CAAC,CAAA;AACvC,CAAC,CACF,CAAA"}
|
package/dist/index.d.ts
ADDED
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAAA,cAAc,aAAa,CAAA;AAC3B,cAAc,WAAW,CAAA"}
|
package/dist/index.js
ADDED
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
3
|
+
if (k2 === undefined) k2 = k;
|
|
4
|
+
var desc = Object.getOwnPropertyDescriptor(m, k);
|
|
5
|
+
if (!desc || ("get" in desc ? !m.__esModule : desc.writable || desc.configurable)) {
|
|
6
|
+
desc = { enumerable: true, get: function() { return m[k]; } };
|
|
7
|
+
}
|
|
8
|
+
Object.defineProperty(o, k2, desc);
|
|
9
|
+
}) : (function(o, m, k, k2) {
|
|
10
|
+
if (k2 === undefined) k2 = k;
|
|
11
|
+
o[k2] = m[k];
|
|
12
|
+
}));
|
|
13
|
+
var __exportStar = (this && this.__exportStar) || function(m, exports) {
|
|
14
|
+
for (var p in m) if (p !== "default" && !Object.prototype.hasOwnProperty.call(exports, p)) __createBinding(exports, m, p);
|
|
15
|
+
};
|
|
16
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
17
|
+
__exportStar(require("./templates"), exports);
|
|
18
|
+
__exportStar(require("./dataset"), exports);
|
|
19
|
+
//# sourceMappingURL=index.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,8CAA2B;AAC3B,4CAAyB"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/templates/index.ts"],"names":[],"mappings":"AAAA,eAAO,MAAM,kCAAkC,IAAI,CAAA"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../src/templates/index.ts"],"names":[],"mappings":";;;AAAa,QAAA,kCAAkC,GAAG,CAAC,CAAA"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"root":["../src/index.ts","../src/dataset/index.ts","../src/dataset/steps/convert-csv-to-records.ts","../src/dataset/steps/create-dataset-import-run.ts","../src/dataset/steps/import-csv-records.ts","../src/dataset/steps/index.ts","../src/dataset/steps/pdf-to-text.ts","../src/dataset/steps/preview-structure-text.ts","../src/dataset/steps/retrieve-dataset-draft.ts","../src/dataset/steps/update-dataset-import-run.ts","../src/dataset/workflows/index.ts","../src/dataset/workflows/preview-dataset-from-csv.ts","../src/dataset/workflows/preview-dataset-from-pdf.ts","../src/dataset/workflows/run-dataset-import-from-csv.ts","../src/templates/index.ts"],"version":"5.8.3"}
|
package/package.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
{
|
|
2
|
+
"name": "@etohq/workflows-flows",
|
|
3
|
+
"version": "0.0.1-next-20260318155517",
|
|
4
|
+
"description": "Reusable workflow template library (seed definitions + schemas)",
|
|
5
|
+
"main": "dist/index.js",
|
|
6
|
+
"types": "dist/index.d.ts",
|
|
7
|
+
"exports": {
|
|
8
|
+
".": {
|
|
9
|
+
"types": "./dist/index.d.ts",
|
|
10
|
+
"import": "./dist/index.js",
|
|
11
|
+
"require": "./dist/index.js"
|
|
12
|
+
}
|
|
13
|
+
},
|
|
14
|
+
"files": [
|
|
15
|
+
"dist"
|
|
16
|
+
],
|
|
17
|
+
"dependencies": {
|
|
18
|
+
"@etohq/framework": "1.5.5-next-20260317140633",
|
|
19
|
+
"@etohq/workflows-input-dataset": "0.0.1-next-20260318155517",
|
|
20
|
+
"@etohq/workflows-input-dataset-pdf-runtime": "0.0.1-next-20260318155517",
|
|
21
|
+
"@etohq/workflows-input-dataset-runtime": "0.0.1-next-20260318155517",
|
|
22
|
+
"@etohq/workflows-input-schema": "0.0.1-next-20260318155517",
|
|
23
|
+
"@etohq/workflows-input-schema-runtime": "0.0.1-next-20260318155517",
|
|
24
|
+
"@etohq/workflows-types": "0.0.1-next-20260318155517"
|
|
25
|
+
},
|
|
26
|
+
"devDependencies": {
|
|
27
|
+
"@types/node": "22.10.5",
|
|
28
|
+
"rimraf": "5.0.2",
|
|
29
|
+
"typescript": "5.8.3"
|
|
30
|
+
},
|
|
31
|
+
"scripts": {
|
|
32
|
+
"build": "rimraf dist && tsc --build",
|
|
33
|
+
"watch": "tsc --build --watch",
|
|
34
|
+
"test": "exit 0"
|
|
35
|
+
}
|
|
36
|
+
}
|