@uploadista/flow-utility-nodes 0.1.4-beta.1 → 0.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1 +1 @@
1
- let e=require(`@uploadista/core/flow`),t=require(`@uploadista/core/types`),n=require(`effect`),r=require(`@uploadista/core/errors`),i=require(`@uploadista/core/upload`),a=require(`zod`);function o(r,{field:i,operator:a,value:o}){return(0,e.createFlowNode)({id:r,name:`Conditional Router`,description:`Routes flow based on ${i} ${a} ${o}`,type:e.NodeType.conditional,nodeTypeId:`conditional`,inputSchema:t.uploadFileSchema,outputSchema:t.uploadFileSchema,condition:{field:i,operator:a,value:o},run:({data:t})=>n.Effect.succeed((0,e.completeNodeExecution)(t))})}const s=a.z.record(a.z.string(),t.uploadFileSchema),c=t.uploadFileSchema;function l(t,{strategy:a,separator:o}){return n.Effect.gen(function*(){let o=yield*i.UploadEngine;return yield*(0,e.createFlowNode)({id:t,name:`Merge Files`,description:`Merges multiple files using ${a} strategy`,type:e.NodeType.merge,nodeTypeId:`merge`,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,inputSchema:s,outputSchema:c,multiInput:!0,run:({data:t,storageId:i,clientId:s})=>n.Effect.gen(function*(){if(!t||Object.keys(t).length===0)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`No inputs provided to merge node`}));let c=Object.values(t);if(c.length===0)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`No files to merge`}));switch(a){case`concat`:{let t=[],n=0;for(let e of c){let r=yield*o.read(e.id,s);t.push(r),n+=r.byteLength}let r=new Uint8Array(n),a=0;for(let e of t)r.set(e,a),a+=e.byteLength;let l=new ReadableStream({start(e){e.enqueue(r),e.close()}});return(0,e.completeNodeExecution)(yield*o.upload({storageId:i,size:r.byteLength,type:`application/octet-stream`,fileName:`merged_${c.length}_files.bin`,lastModified:0,metadata:JSON.stringify({mimeType:`application/octet-stream`,originalName:`merged_${c.length}_files`,extension:`bin`})},s,l))}default:return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`Unknown merge strategy: ${a}`}))}})})})}function u(a,{outputCount:o,strategy:s}){return n.Effect.gen(function*(){let o=yield*i.UploadEngine;return yield*(0,e.createFlowNode)({id:a,name:`Multiplex`,description:`Multiplexes input using ${s} strategy`,type:e.NodeType.multiplex,nodeTypeId:`multiplex`,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,inputSchema:t.uploadFileSchema,outputSchema:t.uploadFileSchema,multiOutput:!0,run:({data:t,storageId:i,clientId:a})=>n.Effect.gen(function*(){let{type:c,fileName:l,metadata:u,metadataJson:d}=(0,e.resolveUploadMetadata)(t.metadata),f=u?{...t,metadata:u}:t;if(s===`copy`){let t=yield*o.read(f.id,a),n=new ReadableStream({start(e){e.enqueue(t),e.close()}}),r=yield*o.upload({storageId:i,size:t.byteLength,type:c,fileName:l,lastModified:0,metadata:d},a,n),s=(0,e.resolveUploadMetadata)(r.metadata);return(0,e.completeNodeExecution)(s.metadata?{...r,metadata:s.metadata}:r)}else if(s===`split`)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`Split strategy is not supported with UploadFile pattern`}));return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`Unknown multiplex strategy: ${s}`}))})})})}function d(r,i){let a=i?.keepOutput??!0;return(0,e.createFlowNode)({id:r,name:`Passthrough`,description:`Passes file through without transformation, acting as an output sink`,type:e.NodeType.process,nodeTypeId:`passthrough`,inputSchema:t.uploadFileSchema,outputSchema:t.uploadFileSchema,keepOutput:a,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,run:({data:t})=>n.Effect.succeed((0,e.completeNodeExecution)(t))})}const f=a.z.record(a.z.string(),t.uploadFileSchema),p=t.uploadFileSchema;function m(t,{zipName:a,includeMetadata:o}){return n.Effect.gen(function*(){let s=yield*i.UploadEngine,c=yield*e.ZipPlugin;return yield*(0,e.createFlowNode)({id:t,name:`Zip Files`,description:`Combines multiple files into a zip archive`,type:e.NodeType.process,nodeTypeId:`zip`,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,inputSchema:f,outputSchema:p,multiInput:!0,run:({data:t,storageId:i,clientId:l})=>n.Effect.gen(function*(){if(!t||Object.keys(t).length===0)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`No inputs provided to zip node`}));let u=yield*n.Effect.forEach(Object.values(t),e=>n.Effect.gen(function*(){let t=yield*s.read(e.id,l);return{id:e.id,data:t,metadata:e.metadata}}),{concurrency:`unbounded`}),d=yield*c.zip(u,{zipName:a,includeMetadata:o}),f=new ReadableStream({start(e){e.enqueue(d),e.close()}});return(0,e.completeNodeExecution)(yield*s.upload({storageId:i,size:d.byteLength,type:`application/zip`,fileName:a,lastModified:0,metadata:JSON.stringify({mimeType:`application/zip`,type:`application/zip`,originalName:a,fileName:a,extension:`zip`})},l,f))})})})}exports.createConditionalNode=o,exports.createMergeNode=l,exports.createMultiplexNode=u,exports.createPassthroughNode=d,exports.createZipNode=m;
1
+ Object.defineProperty(exports,Symbol.toStringTag,{value:`Module`});let e=require(`@uploadista/core/flow`),t=require(`@uploadista/core/types`),n=require(`effect`),r=require(`@uploadista/core/errors`),i=require(`@uploadista/core/upload`),a=require(`zod`);function o(r,{field:i,operator:a,value:o}){return(0,e.createFlowNode)({id:r,name:`Conditional Router`,description:`Routes flow based on ${i} ${a} ${o}`,type:e.NodeType.conditional,nodeTypeId:`conditional`,inputSchema:t.uploadFileSchema,outputSchema:t.uploadFileSchema,condition:{field:i,operator:a,value:o},run:({data:t})=>n.Effect.succeed((0,e.completeNodeExecution)(t))})}const s=a.z.record(a.z.string(),t.uploadFileSchema),c=t.uploadFileSchema;function l(t,{strategy:a,separator:o}){return n.Effect.gen(function*(){let o=yield*i.UploadEngine;return yield*(0,e.createFlowNode)({id:t,name:`Merge Files`,description:`Merges multiple files using ${a} strategy`,type:e.NodeType.merge,nodeTypeId:`merge`,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,inputSchema:s,outputSchema:c,multiInput:!0,run:({data:t,storageId:i,clientId:s})=>n.Effect.gen(function*(){if(!t||Object.keys(t).length===0)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`No inputs provided to merge node`}));let c=Object.values(t);if(c.length===0)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`No files to merge`}));switch(a){case`concat`:{let t=[],n=0;for(let e of c){let r=yield*o.read(e.id,s);t.push(r),n+=r.byteLength}let r=new Uint8Array(n),a=0;for(let e of t)r.set(e,a),a+=e.byteLength;let l=new ReadableStream({start(e){e.enqueue(r),e.close()}});return(0,e.completeNodeExecution)(yield*o.upload({storageId:i,size:r.byteLength,type:`application/octet-stream`,fileName:`merged_${c.length}_files.bin`,lastModified:0,metadata:JSON.stringify({mimeType:`application/octet-stream`,originalName:`merged_${c.length}_files`,extension:`bin`})},s,l))}default:return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`Unknown merge strategy: ${a}`}))}})})})}function u(a,{outputCount:o,strategy:s}){return n.Effect.gen(function*(){let o=yield*i.UploadEngine;return yield*(0,e.createFlowNode)({id:a,name:`Multiplex`,description:`Multiplexes input using ${s} strategy`,type:e.NodeType.multiplex,nodeTypeId:`multiplex`,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,inputSchema:t.uploadFileSchema,outputSchema:t.uploadFileSchema,multiOutput:!0,run:({data:t,storageId:i,clientId:a})=>n.Effect.gen(function*(){let{type:c,fileName:l,metadata:u,metadataJson:d}=(0,e.resolveUploadMetadata)(t.metadata),f=u?{...t,metadata:u}:t;if(s===`copy`){let t=yield*o.read(f.id,a),n=new ReadableStream({start(e){e.enqueue(t),e.close()}}),r=yield*o.upload({storageId:i,size:t.byteLength,type:c,fileName:l,lastModified:0,metadata:d},a,n),s=(0,e.resolveUploadMetadata)(r.metadata);return(0,e.completeNodeExecution)(s.metadata?{...r,metadata:s.metadata}:r)}else if(s===`split`)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`Split strategy is not supported with UploadFile pattern`}));return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`Unknown multiplex strategy: ${s}`}))})})})}function d(r,i){let a=i?.keepOutput??!0;return(0,e.createFlowNode)({id:r,name:`Passthrough`,description:`Passes file through without transformation, acting as an output sink`,type:e.NodeType.process,nodeTypeId:`passthrough`,inputSchema:t.uploadFileSchema,outputSchema:t.uploadFileSchema,keepOutput:a,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,run:({data:t})=>n.Effect.succeed((0,e.completeNodeExecution)(t))})}const f=a.z.record(a.z.string(),t.uploadFileSchema),p=t.uploadFileSchema;function m(t,{zipName:a,includeMetadata:o}){return n.Effect.gen(function*(){let s=yield*i.UploadEngine,c=yield*e.ZipPlugin;return yield*(0,e.createFlowNode)({id:t,name:`Zip Files`,description:`Combines multiple files into a zip archive`,type:e.NodeType.process,nodeTypeId:`zip`,outputTypeId:e.STORAGE_OUTPUT_TYPE_ID,inputSchema:f,outputSchema:p,multiInput:!0,run:({data:t,storageId:i,clientId:l})=>n.Effect.gen(function*(){if(!t||Object.keys(t).length===0)return yield*n.Effect.fail(r.UploadistaError.fromCode(`VALIDATION_ERROR`,{body:`No inputs provided to zip node`}));let u=yield*n.Effect.forEach(Object.values(t),e=>n.Effect.gen(function*(){let t=yield*s.read(e.id,l);return{id:e.id,data:t,metadata:e.metadata}}),{concurrency:`unbounded`}),d=yield*c.zip(u,{zipName:a,includeMetadata:o}),f=new ReadableStream({start(e){e.enqueue(d),e.close()}});return(0,e.completeNodeExecution)(yield*s.upload({storageId:i,size:d.byteLength,type:`application/zip`,fileName:a,lastModified:0,metadata:JSON.stringify({mimeType:`application/zip`,type:`application/zip`,originalName:a,fileName:a,extension:`zip`})},l,f))})})})}exports.createConditionalNode=o,exports.createMergeNode=l,exports.createMultiplexNode=u,exports.createPassthroughNode=d,exports.createZipNode=m;
@@ -1,8 +1,8 @@
1
- import { a as MergeParams, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-BbFUu_Oc.cjs";
2
- import * as _uploadista_core_flow2 from "@uploadista/core/flow";
1
+ import { a as MergeParams, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-D-UE9ml-.cjs";
2
+ import * as _uploadista_core_flow0 from "@uploadista/core/flow";
3
3
  import { NodeType, ZipPlugin } from "@uploadista/core/flow";
4
4
  import * as zod_v4_core0 from "zod/v4/core";
5
- import * as zod0 from "zod";
5
+ import * as zod from "zod";
6
6
  import { z } from "zod";
7
7
  import * as _uploadista_core_errors0 from "@uploadista/core/errors";
8
8
  import { UploadistaError } from "@uploadista/core/errors";
@@ -15,9 +15,9 @@ declare function createConditionalNode(id: string, {
15
15
  field,
16
16
  operator,
17
17
  value
18
- }: ConditionalParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
19
- inputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
20
- outputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
18
+ }: ConditionalParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
19
+ inputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
20
+ outputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
21
21
  run: (args: {
22
22
  data: UploadFile;
23
23
  jobId: string;
@@ -25,7 +25,7 @@ declare function createConditionalNode(id: string, {
25
25
  flowId: string;
26
26
  inputs?: Record<string, unknown>;
27
27
  clientId: string | null;
28
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
28
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
29
29
  condition?: {
30
30
  field: string;
31
31
  operator: string;
@@ -39,7 +39,7 @@ declare function createConditionalNode(id: string, {
39
39
  retryDelay?: number;
40
40
  exponentialBackoff?: boolean;
41
41
  };
42
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
42
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
43
43
  } & {
44
44
  type: NodeType;
45
45
  }, _uploadista_core_errors0.UploadistaError, never>;
@@ -48,7 +48,7 @@ declare function createConditionalNode(id: string, {
48
48
  declare function createMergeNode(id: string, {
49
49
  strategy,
50
50
  separator: _separator
51
- }: MergeParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
51
+ }: MergeParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
52
52
  inputSchema: z.ZodType<Record<string, UploadFile>, unknown, z.core.$ZodTypeInternals<Record<string, UploadFile>, unknown>>;
53
53
  outputSchema: z.ZodType<UploadFile, unknown, z.core.$ZodTypeInternals<UploadFile, unknown>>;
54
54
  run: (args: {
@@ -58,7 +58,7 @@ declare function createMergeNode(id: string, {
58
58
  flowId: string;
59
59
  inputs?: Record<string, unknown>;
60
60
  clientId: string | null;
61
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, UploadistaError, never>;
61
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, UploadistaError, never>;
62
62
  condition?: {
63
63
  field: string;
64
64
  operator: string;
@@ -72,7 +72,7 @@ declare function createMergeNode(id: string, {
72
72
  retryDelay?: number;
73
73
  exponentialBackoff?: boolean;
74
74
  };
75
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
75
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
76
76
  } & {
77
77
  type: NodeType;
78
78
  }, UploadistaError, UploadEngine>;
@@ -81,9 +81,9 @@ declare function createMergeNode(id: string, {
81
81
  declare function createMultiplexNode(id: string, {
82
82
  outputCount: _outputCount,
83
83
  strategy
84
- }: MultiplexParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
85
- inputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
86
- outputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
84
+ }: MultiplexParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
85
+ inputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
86
+ outputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
87
87
  run: (args: {
88
88
  data: UploadFile;
89
89
  jobId: string;
@@ -91,7 +91,7 @@ declare function createMultiplexNode(id: string, {
91
91
  flowId: string;
92
92
  inputs?: Record<string, unknown>;
93
93
  clientId: string | null;
94
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, UploadistaError, never>;
94
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, UploadistaError, never>;
95
95
  condition?: {
96
96
  field: string;
97
97
  operator: string;
@@ -105,7 +105,7 @@ declare function createMultiplexNode(id: string, {
105
105
  retryDelay?: number;
106
106
  exponentialBackoff?: boolean;
107
107
  };
108
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
108
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
109
109
  } & {
110
110
  type: NodeType;
111
111
  }, UploadistaError, UploadEngine>;
@@ -143,9 +143,9 @@ declare function createMultiplexNode(id: string, {
143
143
  */
144
144
  declare function createPassthroughNode(id: string, options?: {
145
145
  keepOutput?: boolean;
146
- }): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
147
- inputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
148
- outputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
146
+ }): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
147
+ inputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
148
+ outputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
149
149
  run: (args: {
150
150
  data: UploadFile;
151
151
  jobId: string;
@@ -153,7 +153,7 @@ declare function createPassthroughNode(id: string, options?: {
153
153
  flowId: string;
154
154
  inputs?: Record<string, unknown>;
155
155
  clientId: string | null;
156
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
156
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
157
157
  condition?: {
158
158
  field: string;
159
159
  operator: string;
@@ -167,7 +167,7 @@ declare function createPassthroughNode(id: string, options?: {
167
167
  retryDelay?: number;
168
168
  exponentialBackoff?: boolean;
169
169
  };
170
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
170
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
171
171
  } & {
172
172
  type: NodeType;
173
173
  }, _uploadista_core_errors0.UploadistaError, never>;
@@ -176,7 +176,7 @@ declare function createPassthroughNode(id: string, options?: {
176
176
  declare function createZipNode(id: string, {
177
177
  zipName,
178
178
  includeMetadata
179
- }: ZipParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
179
+ }: ZipParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
180
180
  inputSchema: z.ZodType<Record<string, UploadFile>, unknown, z.core.$ZodTypeInternals<Record<string, UploadFile>, unknown>>;
181
181
  outputSchema: z.ZodType<UploadFile, unknown, z.core.$ZodTypeInternals<UploadFile, unknown>>;
182
182
  run: (args: {
@@ -186,7 +186,7 @@ declare function createZipNode(id: string, {
186
186
  flowId: string;
187
187
  inputs?: Record<string, unknown>;
188
188
  clientId: string | null;
189
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, UploadistaError, never>;
189
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, UploadistaError, never>;
190
190
  condition?: {
191
191
  field: string;
192
192
  operator: string;
@@ -200,7 +200,7 @@ declare function createZipNode(id: string, {
200
200
  retryDelay?: number;
201
201
  exponentialBackoff?: boolean;
202
202
  };
203
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
203
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
204
204
  } & {
205
205
  type: NodeType;
206
206
  }, UploadistaError, UploadEngine | ZipPlugin>;
@@ -1,12 +1,12 @@
1
- import { a as MergeParams, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-bdgJgJHI.mjs";
2
- import * as _uploadista_core_flow2 from "@uploadista/core/flow";
1
+ import { a as MergeParams, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-BBj1lGeU.mjs";
2
+ import * as _uploadista_core_flow0 from "@uploadista/core/flow";
3
3
  import { NodeType, ZipPlugin } from "@uploadista/core/flow";
4
4
  import { UploadFile } from "@uploadista/core/types";
5
5
  import { Effect } from "effect";
6
6
  import * as _uploadista_core_errors0 from "@uploadista/core/errors";
7
7
  import { UploadistaError } from "@uploadista/core/errors";
8
8
  import { UploadEngine } from "@uploadista/core/upload";
9
- import * as zod0 from "zod";
9
+ import * as zod from "zod";
10
10
  import { z } from "zod";
11
11
  import * as zod_v4_core0 from "zod/v4/core";
12
12
 
@@ -15,9 +15,9 @@ declare function createConditionalNode(id: string, {
15
15
  field,
16
16
  operator,
17
17
  value
18
- }: ConditionalParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
19
- inputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
20
- outputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
18
+ }: ConditionalParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
19
+ inputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
20
+ outputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
21
21
  run: (args: {
22
22
  data: UploadFile;
23
23
  jobId: string;
@@ -25,7 +25,7 @@ declare function createConditionalNode(id: string, {
25
25
  flowId: string;
26
26
  inputs?: Record<string, unknown>;
27
27
  clientId: string | null;
28
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
28
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
29
29
  condition?: {
30
30
  field: string;
31
31
  operator: string;
@@ -39,7 +39,7 @@ declare function createConditionalNode(id: string, {
39
39
  retryDelay?: number;
40
40
  exponentialBackoff?: boolean;
41
41
  };
42
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
42
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
43
43
  } & {
44
44
  type: NodeType;
45
45
  }, _uploadista_core_errors0.UploadistaError, never>;
@@ -48,7 +48,7 @@ declare function createConditionalNode(id: string, {
48
48
  declare function createMergeNode(id: string, {
49
49
  strategy,
50
50
  separator: _separator
51
- }: MergeParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
51
+ }: MergeParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
52
52
  inputSchema: z.ZodType<Record<string, UploadFile>, unknown, z.core.$ZodTypeInternals<Record<string, UploadFile>, unknown>>;
53
53
  outputSchema: z.ZodType<UploadFile, unknown, z.core.$ZodTypeInternals<UploadFile, unknown>>;
54
54
  run: (args: {
@@ -58,7 +58,7 @@ declare function createMergeNode(id: string, {
58
58
  flowId: string;
59
59
  inputs?: Record<string, unknown>;
60
60
  clientId: string | null;
61
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, UploadistaError, never>;
61
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, UploadistaError, never>;
62
62
  condition?: {
63
63
  field: string;
64
64
  operator: string;
@@ -72,7 +72,7 @@ declare function createMergeNode(id: string, {
72
72
  retryDelay?: number;
73
73
  exponentialBackoff?: boolean;
74
74
  };
75
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
75
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
76
76
  } & {
77
77
  type: NodeType;
78
78
  }, UploadistaError, UploadEngine>;
@@ -81,9 +81,9 @@ declare function createMergeNode(id: string, {
81
81
  declare function createMultiplexNode(id: string, {
82
82
  outputCount: _outputCount,
83
83
  strategy
84
- }: MultiplexParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
85
- inputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
86
- outputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
84
+ }: MultiplexParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
85
+ inputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
86
+ outputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
87
87
  run: (args: {
88
88
  data: UploadFile;
89
89
  jobId: string;
@@ -91,7 +91,7 @@ declare function createMultiplexNode(id: string, {
91
91
  flowId: string;
92
92
  inputs?: Record<string, unknown>;
93
93
  clientId: string | null;
94
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, UploadistaError, never>;
94
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, UploadistaError, never>;
95
95
  condition?: {
96
96
  field: string;
97
97
  operator: string;
@@ -105,7 +105,7 @@ declare function createMultiplexNode(id: string, {
105
105
  retryDelay?: number;
106
106
  exponentialBackoff?: boolean;
107
107
  };
108
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
108
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
109
109
  } & {
110
110
  type: NodeType;
111
111
  }, UploadistaError, UploadEngine>;
@@ -143,9 +143,9 @@ declare function createMultiplexNode(id: string, {
143
143
  */
144
144
  declare function createPassthroughNode(id: string, options?: {
145
145
  keepOutput?: boolean;
146
- }): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
147
- inputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
148
- outputSchema: zod0.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
146
+ }): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
147
+ inputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
148
+ outputSchema: zod.ZodType<UploadFile, unknown, zod_v4_core0.$ZodTypeInternals<UploadFile, unknown>>;
149
149
  run: (args: {
150
150
  data: UploadFile;
151
151
  jobId: string;
@@ -153,7 +153,7 @@ declare function createPassthroughNode(id: string, options?: {
153
153
  flowId: string;
154
154
  inputs?: Record<string, unknown>;
155
155
  clientId: string | null;
156
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
156
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, _uploadista_core_errors0.UploadistaError, never>;
157
157
  condition?: {
158
158
  field: string;
159
159
  operator: string;
@@ -167,7 +167,7 @@ declare function createPassthroughNode(id: string, options?: {
167
167
  retryDelay?: number;
168
168
  exponentialBackoff?: boolean;
169
169
  };
170
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
170
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
171
171
  } & {
172
172
  type: NodeType;
173
173
  }, _uploadista_core_errors0.UploadistaError, never>;
@@ -176,7 +176,7 @@ declare function createPassthroughNode(id: string, options?: {
176
176
  declare function createZipNode(id: string, {
177
177
  zipName,
178
178
  includeMetadata
179
- }: ZipParams): Effect.Effect<_uploadista_core_flow2.FlowNodeData & {
179
+ }: ZipParams): Effect.Effect<_uploadista_core_flow0.FlowNodeData & {
180
180
  inputSchema: z.ZodType<Record<string, UploadFile>, unknown, z.core.$ZodTypeInternals<Record<string, UploadFile>, unknown>>;
181
181
  outputSchema: z.ZodType<UploadFile, unknown, z.core.$ZodTypeInternals<UploadFile, unknown>>;
182
182
  run: (args: {
@@ -186,7 +186,7 @@ declare function createZipNode(id: string, {
186
186
  flowId: string;
187
187
  inputs?: Record<string, unknown>;
188
188
  clientId: string | null;
189
- }) => Effect.Effect<_uploadista_core_flow2.NodeExecutionResult<UploadFile>, UploadistaError, never>;
189
+ }) => Effect.Effect<_uploadista_core_flow0.NodeExecutionResult<UploadFile>, UploadistaError, never>;
190
190
  condition?: {
191
191
  field: string;
192
192
  operator: string;
@@ -200,7 +200,7 @@ declare function createZipNode(id: string, {
200
200
  retryDelay?: number;
201
201
  exponentialBackoff?: boolean;
202
202
  };
203
- circuitBreaker?: _uploadista_core_flow2.FlowCircuitBreakerConfig;
203
+ circuitBreaker?: _uploadista_core_flow0.FlowCircuitBreakerConfig;
204
204
  } & {
205
205
  type: NodeType;
206
206
  }, UploadistaError, UploadEngine | ZipPlugin>;
@@ -1,2 +1,2 @@
1
- let e=require(`zod`);const t=e.z.object({field:e.z.enum([`mimeType`,`size`,`width`,`height`,`extension`]),operator:e.z.enum([`equals`,`notEquals`,`greaterThan`,`lessThan`,`contains`,`startsWith`]),value:e.z.union([e.z.string(),e.z.number()])}),n=e.z.object({strategy:e.z.enum([`concat`,`batch`]).default(`batch`),separator:e.z.string().default(`
1
+ Object.defineProperty(exports,Symbol.toStringTag,{value:`Module`});let e=require(`zod`);const t=e.z.object({field:e.z.enum([`mimeType`,`size`,`width`,`height`,`extension`]),operator:e.z.enum([`equals`,`notEquals`,`greaterThan`,`lessThan`,`contains`,`startsWith`]),value:e.z.union([e.z.string(),e.z.number()])}),n=e.z.object({strategy:e.z.enum([`concat`,`batch`]).default(`batch`),separator:e.z.string().default(`
2
2
  `).optional(),inputCount:e.z.number().min(2).max(10).default(2)}),r=e.z.object({outputCount:e.z.number().min(1).max(10),strategy:e.z.enum([`copy`,`split`]).default(`copy`)}),i=e.z.object({zipName:e.z.string().default(`archive.zip`),includeMetadata:e.z.boolean().default(!1),inputCount:e.z.number().min(2).max(10).default(2)});exports.conditionalParamsSchema=t,exports.mergeParamsSchema=n,exports.multiplexParamsSchema=r,exports.zipParamsSchema=i;
@@ -1,2 +1,2 @@
1
- import { a as MergeParams, c as conditionalParamsSchema, i as multiplexParamsSchema, n as zipParamsSchema, o as mergeParamsSchema, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-BbFUu_Oc.cjs";
1
+ import { a as MergeParams, c as conditionalParamsSchema, i as multiplexParamsSchema, n as zipParamsSchema, o as mergeParamsSchema, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-D-UE9ml-.cjs";
2
2
  export { ConditionalParams, MergeParams, MultiplexParams, ZipParams, conditionalParamsSchema, mergeParamsSchema, multiplexParamsSchema, zipParamsSchema };
@@ -1,2 +1,2 @@
1
- import { a as MergeParams, c as conditionalParamsSchema, i as multiplexParamsSchema, n as zipParamsSchema, o as mergeParamsSchema, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-bdgJgJHI.mjs";
1
+ import { a as MergeParams, c as conditionalParamsSchema, i as multiplexParamsSchema, n as zipParamsSchema, o as mergeParamsSchema, r as MultiplexParams, s as ConditionalParams, t as ZipParams } from "../zip-node-BBj1lGeU.mjs";
2
2
  export { ConditionalParams, MergeParams, MultiplexParams, ZipParams, conditionalParamsSchema, mergeParamsSchema, multiplexParamsSchema, zipParamsSchema };
@@ -3,19 +3,19 @@ import { z } from "zod";
3
3
  //#region src/types/conditional-node.d.ts
4
4
  declare const conditionalParamsSchema: z.ZodObject<{
5
5
  field: z.ZodEnum<{
6
- size: "size";
7
6
  mimeType: "mimeType";
7
+ size: "size";
8
8
  width: "width";
9
9
  height: "height";
10
10
  extension: "extension";
11
11
  }>;
12
12
  operator: z.ZodEnum<{
13
- startsWith: "startsWith";
14
13
  equals: "equals";
15
14
  notEquals: "notEquals";
16
15
  greaterThan: "greaterThan";
17
16
  lessThan: "lessThan";
18
17
  contains: "contains";
18
+ startsWith: "startsWith";
19
19
  }>;
20
20
  value: z.ZodUnion<readonly [z.ZodString, z.ZodNumber]>;
21
21
  }, z.core.$strip>;
@@ -36,8 +36,8 @@ type MergeParams = z.infer<typeof mergeParamsSchema>;
36
36
  declare const multiplexParamsSchema: z.ZodObject<{
37
37
  outputCount: z.ZodNumber;
38
38
  strategy: z.ZodDefault<z.ZodEnum<{
39
- split: "split";
40
39
  copy: "copy";
40
+ split: "split";
41
41
  }>>;
42
42
  }, z.core.$strip>;
43
43
  type MultiplexParams = z.infer<typeof multiplexParamsSchema>;
@@ -51,4 +51,4 @@ declare const zipParamsSchema: z.ZodObject<{
51
51
  type ZipParams = z.infer<typeof zipParamsSchema>;
52
52
  //#endregion
53
53
  export { MergeParams as a, conditionalParamsSchema as c, multiplexParamsSchema as i, zipParamsSchema as n, mergeParamsSchema as o, MultiplexParams as r, ConditionalParams as s, ZipParams as t };
54
- //# sourceMappingURL=zip-node-BbFUu_Oc.d.cts.map
54
+ //# sourceMappingURL=zip-node-BBj1lGeU.d.mts.map
@@ -1 +1 @@
1
- {"version":3,"file":"zip-node-BbFUu_Oc.d.cts","names":[],"sources":["../src/types/conditional-node.ts","../src/types/merge-node.ts","../src/types/multiplex-node.ts","../src/types/zip-node.ts"],"mappings":";;;cAEa,uBAAA,EAAuB,CAAA,CAAA,SAAA;;;;;;;;;;;;;;;;;;KAaxB,iBAAA,GAAoB,CAAA,CAAE,KAAA,QAAa,uBAAA;;;cCblC,iBAAA,EAAiB,CAAA,CAAA,SAAA;;;;;;;;KAMlB,WAAA,GAAc,CAAA,CAAE,KAAA,QAAa,iBAAA;;;cCN5B,qBAAA,EAAqB,CAAA,CAAA,SAAA;;;;;;;KAKtB,eAAA,GAAkB,CAAA,CAAE,KAAA,QAAa,qBAAA;;;cCLhC,eAAA,EAAe,CAAA,CAAA,SAAA;;;;;KAMhB,SAAA,GAAY,CAAA,CAAE,KAAA,QAAa,eAAA"}
1
+ {"version":3,"file":"zip-node-BBj1lGeU.d.mts","names":[],"sources":["../src/types/conditional-node.ts","../src/types/merge-node.ts","../src/types/multiplex-node.ts","../src/types/zip-node.ts"],"mappings":";;;cAEa,uBAAA,EAAuB,CAAA,CAAA,SAAA;;;;;;;;;;;;;;;;;;KAaxB,iBAAA,GAAoB,CAAA,CAAE,KAAA,QAAa,uBAAA;;;cCblC,iBAAA,EAAiB,CAAA,CAAA,SAAA;;;;;;;;KAMlB,WAAA,GAAc,CAAA,CAAE,KAAA,QAAa,iBAAA;;;cCN5B,qBAAA,EAAqB,CAAA,CAAA,SAAA;;;;;;;KAKtB,eAAA,GAAkB,CAAA,CAAE,KAAA,QAAa,qBAAA;;;cCLhC,eAAA,EAAe,CAAA,CAAA,SAAA;;;;;KAMhB,SAAA,GAAY,CAAA,CAAE,KAAA,QAAa,eAAA"}
@@ -3,19 +3,19 @@ import { z } from "zod";
3
3
  //#region src/types/conditional-node.d.ts
4
4
  declare const conditionalParamsSchema: z.ZodObject<{
5
5
  field: z.ZodEnum<{
6
- size: "size";
7
6
  mimeType: "mimeType";
7
+ size: "size";
8
8
  width: "width";
9
9
  height: "height";
10
10
  extension: "extension";
11
11
  }>;
12
12
  operator: z.ZodEnum<{
13
- startsWith: "startsWith";
14
13
  equals: "equals";
15
14
  notEquals: "notEquals";
16
15
  greaterThan: "greaterThan";
17
16
  lessThan: "lessThan";
18
17
  contains: "contains";
18
+ startsWith: "startsWith";
19
19
  }>;
20
20
  value: z.ZodUnion<readonly [z.ZodString, z.ZodNumber]>;
21
21
  }, z.core.$strip>;
@@ -36,8 +36,8 @@ type MergeParams = z.infer<typeof mergeParamsSchema>;
36
36
  declare const multiplexParamsSchema: z.ZodObject<{
37
37
  outputCount: z.ZodNumber;
38
38
  strategy: z.ZodDefault<z.ZodEnum<{
39
- split: "split";
40
39
  copy: "copy";
40
+ split: "split";
41
41
  }>>;
42
42
  }, z.core.$strip>;
43
43
  type MultiplexParams = z.infer<typeof multiplexParamsSchema>;
@@ -51,4 +51,4 @@ declare const zipParamsSchema: z.ZodObject<{
51
51
  type ZipParams = z.infer<typeof zipParamsSchema>;
52
52
  //#endregion
53
53
  export { MergeParams as a, conditionalParamsSchema as c, multiplexParamsSchema as i, zipParamsSchema as n, mergeParamsSchema as o, MultiplexParams as r, ConditionalParams as s, ZipParams as t };
54
- //# sourceMappingURL=zip-node-bdgJgJHI.d.mts.map
54
+ //# sourceMappingURL=zip-node-D-UE9ml-.d.cts.map
@@ -1 +1 @@
1
- {"version":3,"file":"zip-node-bdgJgJHI.d.mts","names":[],"sources":["../src/types/conditional-node.ts","../src/types/merge-node.ts","../src/types/multiplex-node.ts","../src/types/zip-node.ts"],"mappings":";;;cAEa,uBAAA,EAAuB,CAAA,CAAA,SAAA;;;;;;;;;;;;;;;;;;KAaxB,iBAAA,GAAoB,CAAA,CAAE,KAAA,QAAa,uBAAA;;;cCblC,iBAAA,EAAiB,CAAA,CAAA,SAAA;;;;;;;;KAMlB,WAAA,GAAc,CAAA,CAAE,KAAA,QAAa,iBAAA;;;cCN5B,qBAAA,EAAqB,CAAA,CAAA,SAAA;;;;;;;KAKtB,eAAA,GAAkB,CAAA,CAAE,KAAA,QAAa,qBAAA;;;cCLhC,eAAA,EAAe,CAAA,CAAA,SAAA;;;;;KAMhB,SAAA,GAAY,CAAA,CAAE,KAAA,QAAa,eAAA"}
1
+ {"version":3,"file":"zip-node-D-UE9ml-.d.cts","names":[],"sources":["../src/types/conditional-node.ts","../src/types/merge-node.ts","../src/types/multiplex-node.ts","../src/types/zip-node.ts"],"mappings":";;;cAEa,uBAAA,EAAuB,CAAA,CAAA,SAAA;;;;;;;;;;;;;;;;;;KAaxB,iBAAA,GAAoB,CAAA,CAAE,KAAA,QAAa,uBAAA;;;cCblC,iBAAA,EAAiB,CAAA,CAAA,SAAA;;;;;;;;KAMlB,WAAA,GAAc,CAAA,CAAE,KAAA,QAAa,iBAAA;;;cCN5B,qBAAA,EAAqB,CAAA,CAAA,SAAA;;;;;;;KAKtB,eAAA,GAAkB,CAAA,CAAE,KAAA,QAAa,qBAAA;;;cCLhC,eAAA,EAAe,CAAA,CAAA,SAAA;;;;;KAMhB,SAAA,GAAY,CAAA,CAAE,KAAA,QAAa,eAAA"}
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@uploadista/flow-utility-nodes",
3
3
  "type": "module",
4
- "version": "0.1.4-beta.1",
4
+ "version": "0.2.0",
5
5
  "description": "Utility nodes for Uploadista Flow",
6
6
  "license": "MIT",
7
7
  "author": "Uploadista",
@@ -20,7 +20,7 @@
20
20
  }
21
21
  },
22
22
  "dependencies": {
23
- "@uploadista/core": "0.1.4-beta.1"
23
+ "@uploadista/core": "0.2.0"
24
24
  },
25
25
  "peerDependencies": {
26
26
  "effect": "^3.0.0",
@@ -29,11 +29,11 @@
29
29
  "devDependencies": {
30
30
  "@effect/vitest": "0.27.0",
31
31
  "@types/node": "24.10.9",
32
- "effect": "3.19.15",
33
- "tsdown": "0.20.1",
32
+ "effect": "3.19.17",
33
+ "tsdown": "0.20.3",
34
34
  "vitest": "4.0.18",
35
35
  "zod": "4.3.6",
36
- "@uploadista/typescript-config": "0.1.4-beta.1"
36
+ "@uploadista/typescript-config": "0.2.0"
37
37
  },
38
38
  "scripts": {
39
39
  "build": "tsc --noEmit && tsdown",