@civitai/client 0.2.0-beta.52 → 0.2.0-beta.54
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/generated/index.d.ts +38 -31
- package/dist/generated/index.js +2 -2
- package/dist/generated/sdk.gen.d.ts +47 -0
- package/dist/generated/sdk.gen.js +43 -0
- package/dist/generated/types.gen.d.ts +456 -329
- package/dist/generated/types.gen.js +36 -0
- package/package.json +1 -1
|
@@ -151,10 +151,32 @@ export type AceStepAudioInput = {
|
|
|
151
151
|
*/
|
|
152
152
|
vocalWeight?: number;
|
|
153
153
|
/**
|
|
154
|
-
* Optional model override
|
|
154
|
+
* Optional diffusion model (unet) override. Defaults to ACE-Step 1.5 turbo 2B.
|
|
155
155
|
*/
|
|
156
|
-
|
|
156
|
+
diffusionModel?: null | string;
|
|
157
157
|
cover?: AceStepAudioCover;
|
|
158
|
+
/**
|
|
159
|
+
* Optional text encoder (CLIP) override. Defaults to qwen_4b_ace15.
|
|
160
|
+
*/
|
|
161
|
+
clipModel?: null | string;
|
|
162
|
+
/**
|
|
163
|
+
* Optional VAE override. Defaults to ace_1.5_vae.
|
|
164
|
+
*/
|
|
165
|
+
vaeModel?: null | string;
|
|
166
|
+
/**
|
|
167
|
+
* Optional language model override for audio code generation. Defaults to qwen_0.6b_ace15.
|
|
168
|
+
*/
|
|
169
|
+
languageModel?: null | string;
|
|
170
|
+
/**
|
|
171
|
+
* Number of sampling steps. Turbo variants converge at 8; non-turbo base / sft
|
|
172
|
+
* variants expect the full 50-step schedule.
|
|
173
|
+
*/
|
|
174
|
+
steps?: number;
|
|
175
|
+
/**
|
|
176
|
+
* Classifier-free guidance scale. Turbo variants run CFG-off at 1.0;
|
|
177
|
+
* non-turbo base / sft variants expect CFG on (around 4).
|
|
178
|
+
*/
|
|
179
|
+
cfg?: number;
|
|
158
180
|
};
|
|
159
181
|
/**
|
|
160
182
|
* Output from ACE Step 1.5 audio generation workflow step.
|
|
@@ -360,6 +382,17 @@ export declare const AnylineMergeWith: {
|
|
|
360
382
|
readonly MANGA_LINE: 'manga_line';
|
|
361
383
|
};
|
|
362
384
|
export type AnylineMergeWith = (typeof AnylineMergeWith)[keyof typeof AnylineMergeWith];
|
|
385
|
+
/**
|
|
386
|
+
* Container format for a Civitai.Orchestration.Grains.Workflows.Steps.BlobArchive.BlobArchiveStep output.
|
|
387
|
+
*/
|
|
388
|
+
export declare const ArchiveFormat: {
|
|
389
|
+
readonly ZIP: 'zip';
|
|
390
|
+
readonly TAR: 'tar';
|
|
391
|
+
};
|
|
392
|
+
/**
|
|
393
|
+
* Container format for a Civitai.Orchestration.Grains.Workflows.Steps.BlobArchive.BlobArchiveStep output.
|
|
394
|
+
*/
|
|
395
|
+
export type ArchiveFormat = (typeof ArchiveFormat)[keyof typeof ArchiveFormat];
|
|
363
396
|
/**
|
|
364
397
|
* An assistant message representing a prior response.
|
|
365
398
|
*/
|
|
@@ -431,6 +464,72 @@ export type Blob = {
|
|
|
431
464
|
*/
|
|
432
465
|
blockedReason?: null | string;
|
|
433
466
|
};
|
|
467
|
+
/**
|
|
468
|
+
* A single entry within a Civitai.Orchestration.Grains.Workflows.Steps.BlobArchive.BlobArchiveStep.
|
|
469
|
+
*/
|
|
470
|
+
export type BlobArchiveEntry = {
|
|
471
|
+
/**
|
|
472
|
+
* The blob ID to include in the archive.
|
|
473
|
+
*/
|
|
474
|
+
blobId: string;
|
|
475
|
+
/**
|
|
476
|
+
* Optional filename to use inside the archive. When omitted, the blob ID is used.
|
|
477
|
+
* Path components and control characters are stripped.
|
|
478
|
+
*/
|
|
479
|
+
fileName?: null | string;
|
|
480
|
+
};
|
|
481
|
+
/**
|
|
482
|
+
* Input configuration for the BlobArchive workflow step.
|
|
483
|
+
*/
|
|
484
|
+
export type BlobArchiveInput = {
|
|
485
|
+
/**
|
|
486
|
+
* The blobs to include in the archive. Must contain at least 1 and at most 1000 entries.
|
|
487
|
+
*/
|
|
488
|
+
entries: Array<BlobArchiveEntry>;
|
|
489
|
+
/**
|
|
490
|
+
* Optional filename for the archive itself (used in the Content-Disposition header).
|
|
491
|
+
* Defaults to "archive.zip" or "archive.tar" based on Civitai.Orchestration.Grains.Workflows.Steps.BlobArchive.BlobArchiveInput.Format.
|
|
492
|
+
*/
|
|
493
|
+
archiveName?: null | string;
|
|
494
|
+
format?: ArchiveFormat;
|
|
495
|
+
};
|
|
496
|
+
/**
|
|
497
|
+
* Output produced by the BlobArchive workflow step.
|
|
498
|
+
*/
|
|
499
|
+
export type BlobArchiveOutput = {
|
|
500
|
+
/**
|
|
501
|
+
* The signed URL that streams the archive when requested.
|
|
502
|
+
*/
|
|
503
|
+
url: string;
|
|
504
|
+
/**
|
|
505
|
+
* The number of entries included in the archive.
|
|
506
|
+
*/
|
|
507
|
+
entryCount: number;
|
|
508
|
+
format: ArchiveFormat;
|
|
509
|
+
/**
|
|
510
|
+
* The UTC time at which this URL expires.
|
|
511
|
+
*/
|
|
512
|
+
expiresAt: string;
|
|
513
|
+
};
|
|
514
|
+
/**
|
|
515
|
+
* Bundles a set of blobs into a single archive (zip or tar) that callers can
|
|
516
|
+
* download from a signed streaming URL. Runs in-process in the orchestrator;
|
|
517
|
+
* no worker job is dispatched.
|
|
518
|
+
*/
|
|
519
|
+
export type BlobArchiveStep = Omit<WorkflowStep, '$type'> & {
|
|
520
|
+
input: BlobArchiveInput;
|
|
521
|
+
output?: BlobArchiveOutput;
|
|
522
|
+
$type: 'blobArchive';
|
|
523
|
+
};
|
|
524
|
+
/**
|
|
525
|
+
* Bundles a set of blobs into a single archive (zip or tar) that callers can
|
|
526
|
+
* download from a signed streaming URL. Runs in-process in the orchestrator;
|
|
527
|
+
* no worker job is dispatched.
|
|
528
|
+
*/
|
|
529
|
+
export type BlobArchiveStepTemplate = Omit<WorkflowStepTemplate, '$type'> & {
|
|
530
|
+
input: BlobArchiveInput;
|
|
531
|
+
$type: 'blobArchive';
|
|
532
|
+
};
|
|
434
533
|
/**
|
|
435
534
|
* A rectangular region defined by pixel coordinates.
|
|
436
535
|
*/
|
|
@@ -725,6 +824,73 @@ export declare const CoarseMode: {
|
|
|
725
824
|
readonly ENABLE: 'enable';
|
|
726
825
|
};
|
|
727
826
|
export type CoarseMode = (typeof CoarseMode)[keyof typeof CoarseMode];
|
|
827
|
+
export type ComfyErnieImageGenInput = Omit<ComfyImageGenInput, 'engine' | 'ecosystem'> & {
|
|
828
|
+
model: string;
|
|
829
|
+
ecosystem: 'ernie';
|
|
830
|
+
engine: 'comfy';
|
|
831
|
+
};
|
|
832
|
+
export type ComfyErnieStandardCreateImageGenInput = Omit<
|
|
833
|
+
ComfyErnieStandardImageGenInput,
|
|
834
|
+
'engine' | 'ecosystem' | 'model' | 'operation'
|
|
835
|
+
> & {
|
|
836
|
+
width?: number;
|
|
837
|
+
height?: number;
|
|
838
|
+
operation: 'createImage';
|
|
839
|
+
model: 'ernie';
|
|
840
|
+
ecosystem: 'ernie';
|
|
841
|
+
engine: 'comfy';
|
|
842
|
+
};
|
|
843
|
+
export type ComfyErnieStandardImageGenInput = Omit<
|
|
844
|
+
ComfyErnieImageGenInput,
|
|
845
|
+
'engine' | 'ecosystem' | 'model'
|
|
846
|
+
> & {
|
|
847
|
+
operation: string;
|
|
848
|
+
prompt: string;
|
|
849
|
+
negativePrompt?: null | string;
|
|
850
|
+
sampler?: ComfySampler;
|
|
851
|
+
scheduler?: ComfyScheduler;
|
|
852
|
+
steps?: number;
|
|
853
|
+
cfgScale?: number;
|
|
854
|
+
seed?: null | number;
|
|
855
|
+
quantity?: number;
|
|
856
|
+
loras?: {
|
|
857
|
+
[key: string]: number;
|
|
858
|
+
};
|
|
859
|
+
model: 'ernie';
|
|
860
|
+
ecosystem: 'ernie';
|
|
861
|
+
engine: 'comfy';
|
|
862
|
+
};
|
|
863
|
+
export type ComfyErnieTurboCreateImageGenInput = Omit<
|
|
864
|
+
ComfyErnieTurboImageGenInput,
|
|
865
|
+
'engine' | 'ecosystem' | 'model' | 'operation'
|
|
866
|
+
> & {
|
|
867
|
+
width?: number;
|
|
868
|
+
height?: number;
|
|
869
|
+
operation: 'createImage';
|
|
870
|
+
model: 'turbo';
|
|
871
|
+
ecosystem: 'ernie';
|
|
872
|
+
engine: 'comfy';
|
|
873
|
+
};
|
|
874
|
+
export type ComfyErnieTurboImageGenInput = Omit<
|
|
875
|
+
ComfyErnieImageGenInput,
|
|
876
|
+
'engine' | 'ecosystem' | 'model'
|
|
877
|
+
> & {
|
|
878
|
+
operation: string;
|
|
879
|
+
prompt: string;
|
|
880
|
+
negativePrompt?: null | string;
|
|
881
|
+
sampler?: ComfySampler;
|
|
882
|
+
scheduler?: ComfyScheduler;
|
|
883
|
+
steps?: number;
|
|
884
|
+
cfgScale?: number;
|
|
885
|
+
seed?: null | number;
|
|
886
|
+
quantity?: number;
|
|
887
|
+
loras?: {
|
|
888
|
+
[key: string]: number;
|
|
889
|
+
};
|
|
890
|
+
model: 'turbo';
|
|
891
|
+
ecosystem: 'ernie';
|
|
892
|
+
engine: 'comfy';
|
|
893
|
+
};
|
|
728
894
|
export type ComfyFlux1CreateImageGenInput = Omit<
|
|
729
895
|
ComfyFlux1ImageGenInput,
|
|
730
896
|
'engine' | 'ecosystem' | 'operation'
|
|
@@ -1315,6 +1481,13 @@ export type EpochResult = {
|
|
|
1315
1481
|
*/
|
|
1316
1482
|
blobUrl: string;
|
|
1317
1483
|
};
|
|
1484
|
+
/**
|
|
1485
|
+
* AI Toolkit training for ERNIE-Image models
|
|
1486
|
+
*/
|
|
1487
|
+
export type ErnieAiToolkitTrainingInput = Omit<AiToolkitTrainingInput, 'engine' | 'ecosystem'> & {
|
|
1488
|
+
ecosystem: 'ernie';
|
|
1489
|
+
engine: 'ai-toolkit';
|
|
1490
|
+
};
|
|
1318
1491
|
/**
|
|
1319
1492
|
* Represents a single face detection with embeddings and landmarks.
|
|
1320
1493
|
*/
|
|
@@ -2060,6 +2233,11 @@ export type ImageJobControlNet = {
|
|
|
2060
2233
|
* A value representing the end step selected for the ControlNet.
|
|
2061
2234
|
*/
|
|
2062
2235
|
endStep: number;
|
|
2236
|
+
/**
|
|
2237
|
+
* The preprocessed control image. Accepts an AIR URN, a URL, or a base64 data URL; the
|
|
2238
|
+
* orchestrator processes it into a stored blob before the job is dispatched.
|
|
2239
|
+
*/
|
|
2240
|
+
image?: null | string;
|
|
2063
2241
|
};
|
|
2064
2242
|
export type ImageJobNetworkParams = {
|
|
2065
2243
|
/**
|
|
@@ -2175,8 +2353,39 @@ export type ImageTransform = {
|
|
|
2175
2353
|
*/
|
|
2176
2354
|
export declare const ImageTransformer: {
|
|
2177
2355
|
readonly CANNY: 'canny';
|
|
2356
|
+
readonly MLSD: 'mlsd';
|
|
2178
2357
|
readonly DEPTH_ZOE: 'depthZoe';
|
|
2358
|
+
readonly DEPTH_ANYTHING: 'depthAnything';
|
|
2359
|
+
readonly DEPTH_ANYTHING_V2: 'depthAnythingV2';
|
|
2360
|
+
readonly ZOE_DEPTH_ANYTHING: 'zoeDepthAnything';
|
|
2361
|
+
readonly ZOE_DEPTH: 'zoeDepth';
|
|
2362
|
+
readonly MIDAS_DEPTH: 'midasDepth';
|
|
2363
|
+
readonly LERES_DEPTH: 'leresDepth';
|
|
2364
|
+
readonly METRIC3D_DEPTH: 'metric3dDepth';
|
|
2179
2365
|
readonly SOFTEDGE_PIDINET: 'softedgePidinet';
|
|
2366
|
+
readonly HED: 'hed';
|
|
2367
|
+
readonly TEED: 'teed';
|
|
2368
|
+
readonly MIDAS_NORMAL: 'midasNormal';
|
|
2369
|
+
readonly BAE_NORMAL: 'baeNormal';
|
|
2370
|
+
readonly DSINE_NORMAL: 'dsineNormal';
|
|
2371
|
+
readonly METRIC3D_NORMAL: 'metric3dNormal';
|
|
2372
|
+
readonly LINEART_REALISTIC: 'lineartRealistic';
|
|
2373
|
+
readonly LINEART_STANDARD: 'lineartStandard';
|
|
2374
|
+
readonly LINEART_ANIME: 'lineartAnime';
|
|
2375
|
+
readonly LINEART_MANGA: 'lineartManga';
|
|
2376
|
+
readonly ANYLINE: 'anyline';
|
|
2377
|
+
readonly SCRIBBLE: 'scribble';
|
|
2378
|
+
readonly SCRIBBLE_XDOG: 'scribbleXdog';
|
|
2379
|
+
readonly SCRIBBLE_PIDINET: 'scribblePidinet';
|
|
2380
|
+
readonly FAKE_SCRIBBLE: 'fakeScribble';
|
|
2381
|
+
readonly OPENPOSE: 'openpose';
|
|
2382
|
+
readonly DWPOSE: 'dwpose';
|
|
2383
|
+
readonly ONEFORMER_COCO: 'oneformerCoco';
|
|
2384
|
+
readonly ONEFORMER_ADE20K: 'oneformerAde20k';
|
|
2385
|
+
readonly UNIFORMER: 'uniformer';
|
|
2386
|
+
readonly SHUFFLE: 'shuffle';
|
|
2387
|
+
readonly TILE: 'tile';
|
|
2388
|
+
readonly GRAY: 'gray';
|
|
2180
2389
|
readonly REMBG: 'rembg';
|
|
2181
2390
|
};
|
|
2182
2391
|
/**
|
|
@@ -2636,9 +2845,10 @@ export type MediaHashType = (typeof MediaHashType)[keyof typeof MediaHashType];
|
|
|
2636
2845
|
*/
|
|
2637
2846
|
export type MediaRatingInput = {
|
|
2638
2847
|
/**
|
|
2639
|
-
* The
|
|
2848
|
+
* The image to rate. Pre-processed in the API layer — imported into blob storage so the
|
|
2849
|
+
* worker can pre-fetch it as a declared resource. After processing, contains the blob AIR URL.
|
|
2640
2850
|
*/
|
|
2641
|
-
|
|
2851
|
+
image?: null | string;
|
|
2642
2852
|
/**
|
|
2643
2853
|
* The engine to use for media rating. Valid values: "default" (HiveVLM) or "civitai".
|
|
2644
2854
|
*/
|
|
@@ -2774,7 +2984,7 @@ export type ModelHashOutput = {
|
|
|
2774
2984
|
/**
|
|
2775
2985
|
* SHA256 hash of the full file.
|
|
2776
2986
|
*/
|
|
2777
|
-
|
|
2987
|
+
sha256?: null | string;
|
|
2778
2988
|
/**
|
|
2779
2989
|
* AutoV1 short hash (8 chars of SHA256 over a 64 KB block starting at 1 MB).
|
|
2780
2990
|
*/
|
|
@@ -2794,7 +3004,7 @@ export type ModelHashOutput = {
|
|
|
2794
3004
|
/**
|
|
2795
3005
|
* CRC32 of the full file.
|
|
2796
3006
|
*/
|
|
2797
|
-
|
|
3007
|
+
crc32?: null | string;
|
|
2798
3008
|
};
|
|
2799
3009
|
/**
|
|
2800
3010
|
* ModelHash
|
|
@@ -2811,6 +3021,40 @@ export type ModelHashStepTemplate = Omit<WorkflowStepTemplate, '$type'> & {
|
|
|
2811
3021
|
input: ModelHashInput;
|
|
2812
3022
|
$type: 'modelHash';
|
|
2813
3023
|
};
|
|
3024
|
+
/**
|
|
3025
|
+
* Represents the input information needed for the ModelParseMetadata workflow step.
|
|
3026
|
+
*/
|
|
3027
|
+
export type ModelParseMetadataInput = {
|
|
3028
|
+
/**
|
|
3029
|
+
* The AIR of the model file to read metadata from.
|
|
3030
|
+
*/
|
|
3031
|
+
model: string;
|
|
3032
|
+
};
|
|
3033
|
+
/**
|
|
3034
|
+
* Represents the output information returned from the ModelParseMetadata workflow step.
|
|
3035
|
+
*/
|
|
3036
|
+
export type ModelParseMetadataOutput = {
|
|
3037
|
+
/**
|
|
3038
|
+
* The raw JSON metadata header as stored in the safetensors file, or null when
|
|
3039
|
+
* the model is not a safetensors file or the header could not be parsed.
|
|
3040
|
+
*/
|
|
3041
|
+
metadata?: null | string;
|
|
3042
|
+
};
|
|
3043
|
+
/**
|
|
3044
|
+
* ModelParseMetadata
|
|
3045
|
+
*/
|
|
3046
|
+
export type ModelParseMetadataStep = Omit<WorkflowStep, '$type'> & {
|
|
3047
|
+
input: ModelParseMetadataInput;
|
|
3048
|
+
output?: ModelParseMetadataOutput;
|
|
3049
|
+
$type: 'modelParseMetadata';
|
|
3050
|
+
};
|
|
3051
|
+
/**
|
|
3052
|
+
* ModelParseMetadata
|
|
3053
|
+
*/
|
|
3054
|
+
export type ModelParseMetadataStepTemplate = Omit<WorkflowStepTemplate, '$type'> & {
|
|
3055
|
+
input: ModelParseMetadataInput;
|
|
3056
|
+
$type: 'modelParseMetadata';
|
|
3057
|
+
};
|
|
2814
3058
|
/**
|
|
2815
3059
|
* Represents the input information needed for the ModelPickleScan workflow step.
|
|
2816
3060
|
*/
|
|
@@ -3050,6 +3294,49 @@ export type OpenAiGpt1ImageGenInput = Omit<OpenApiImageGenInput, 'engine' | 'mod
|
|
|
3050
3294
|
model: 'gpt-image-1';
|
|
3051
3295
|
engine: 'openai';
|
|
3052
3296
|
};
|
|
3297
|
+
export type OpenAiGpt2CreateImageInput = Omit<
|
|
3298
|
+
OpenAiGpt2ImageGenInput,
|
|
3299
|
+
'engine' | 'model' | 'operation'
|
|
3300
|
+
> & {
|
|
3301
|
+
operation: 'createImage';
|
|
3302
|
+
model: 'gpt-image-2';
|
|
3303
|
+
engine: 'openai';
|
|
3304
|
+
};
|
|
3305
|
+
export type OpenAiGpt2EditImageInput = Omit<
|
|
3306
|
+
OpenAiGpt2ImageGenInput,
|
|
3307
|
+
'engine' | 'model' | 'operation'
|
|
3308
|
+
> & {
|
|
3309
|
+
images: Array<string>;
|
|
3310
|
+
/**
|
|
3311
|
+
* Either A URL, A DataURL or a Base64 string
|
|
3312
|
+
*/
|
|
3313
|
+
maskImage?: null | string;
|
|
3314
|
+
/**
|
|
3315
|
+
* When null, fal infers output size from the input images (image_size: "auto").
|
|
3316
|
+
* When set, the requested width is sent to fal as image_size.width.
|
|
3317
|
+
* Both Width and Height must be set together, or both null.
|
|
3318
|
+
*/
|
|
3319
|
+
width?: null | number;
|
|
3320
|
+
/**
|
|
3321
|
+
* When null, fal infers output size from the input images (image_size: "auto").
|
|
3322
|
+
* When set, the requested height is sent to fal as image_size.height.
|
|
3323
|
+
* Both Width and Height must be set together, or both null.
|
|
3324
|
+
*/
|
|
3325
|
+
height?: null | number;
|
|
3326
|
+
operation: 'editImage';
|
|
3327
|
+
model: 'gpt-image-2';
|
|
3328
|
+
engine: 'openai';
|
|
3329
|
+
};
|
|
3330
|
+
export type OpenAiGpt2ImageGenInput = Omit<OpenApiImageGenInput, 'engine' | 'model'> & {
|
|
3331
|
+
operation: string;
|
|
3332
|
+
prompt: string;
|
|
3333
|
+
width?: null | number;
|
|
3334
|
+
height?: null | number;
|
|
3335
|
+
quantity?: number;
|
|
3336
|
+
quality?: 'low' | 'medium' | 'high';
|
|
3337
|
+
model: 'gpt-image-2';
|
|
3338
|
+
engine: 'openai';
|
|
3339
|
+
};
|
|
3053
3340
|
export type OpenApiImageGenInput = Omit<ImageGenInput, 'engine'> & {
|
|
3054
3341
|
model: string;
|
|
3055
3342
|
prompt: string;
|
|
@@ -3859,6 +4146,21 @@ export type SdxlVariantImageGenInput = Omit<
|
|
|
3859
4146
|
ecosystem: 'sdxl';
|
|
3860
4147
|
engine: 'sdcpp';
|
|
3861
4148
|
};
|
|
4149
|
+
export declare const SeedanceModel: {
|
|
4150
|
+
readonly V2: 'v2';
|
|
4151
|
+
readonly V2_FAST: 'v2-fast';
|
|
4152
|
+
};
|
|
4153
|
+
export type SeedanceModel = (typeof SeedanceModel)[keyof typeof SeedanceModel];
|
|
4154
|
+
export type SeedanceVideoGenInput = Omit<VideoGenInput, 'engine'> & {
|
|
4155
|
+
model?: SeedanceModel;
|
|
4156
|
+
aspectRatio: '16:9' | '9:16' | '1:1' | '4:3' | '3:4' | '21:9' | 'adaptive';
|
|
4157
|
+
duration: 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15;
|
|
4158
|
+
generateAudio?: boolean;
|
|
4159
|
+
seed?: null | number;
|
|
4160
|
+
resolution: '480p' | '720p';
|
|
4161
|
+
images?: Array<string>;
|
|
4162
|
+
engine: 'seedance';
|
|
4163
|
+
};
|
|
3862
4164
|
export type SeedreamImageGenInput = Omit<ImageGenInput, 'engine'> & {
|
|
3863
4165
|
prompt: string;
|
|
3864
4166
|
quantity?: number;
|
|
@@ -5477,9 +5779,12 @@ export type WorkflowStep = {
|
|
|
5477
5779
|
[key: string]: unknown;
|
|
5478
5780
|
};
|
|
5479
5781
|
/**
|
|
5480
|
-
* An estimation on the current progression of this step, or null if there is no estimation
|
|
5782
|
+
* An estimation on the current progression of this step, or null if there is no estimation.
|
|
5783
|
+
* Computed by the step's handler (see `WorkflowStepHandler.GetEstimatedProgressRate`)
|
|
5784
|
+
* and refreshed by `WorkflowStepManager` on each job event. Cleared to null once the
|
|
5785
|
+
* step reaches a final status.
|
|
5481
5786
|
*/
|
|
5482
|
-
|
|
5787
|
+
estimatedProgressRate?: null | number;
|
|
5483
5788
|
};
|
|
5484
5789
|
/**
|
|
5485
5790
|
* Details of a workflow step event.
|
|
@@ -5939,23 +6244,6 @@ export declare const ZoeDepthEnvironment: {
|
|
|
5939
6244
|
readonly OUTDOOR: 'outdoor';
|
|
5940
6245
|
};
|
|
5941
6246
|
export type ZoeDepthEnvironment = (typeof ZoeDepthEnvironment)[keyof typeof ZoeDepthEnvironment];
|
|
5942
|
-
/**
|
|
5943
|
-
* Workflow step for generating music using ACE Step 1.5.
|
|
5944
|
-
* Produces full songs from text descriptions and structured lyrics.
|
|
5945
|
-
*/
|
|
5946
|
-
export type AceStepAudioStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
5947
|
-
input: AceStepAudioInput;
|
|
5948
|
-
output?: AceStepAudioOutput;
|
|
5949
|
-
$type: 'aceStepAudio';
|
|
5950
|
-
};
|
|
5951
|
-
/**
|
|
5952
|
-
* Age classification
|
|
5953
|
-
*/
|
|
5954
|
-
export type AgeClassificationStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
5955
|
-
input: AgeClassificationInput;
|
|
5956
|
-
output?: AgeClassificationOutput;
|
|
5957
|
-
$type: 'ageClassification';
|
|
5958
|
-
};
|
|
5959
6247
|
/**
|
|
5960
6248
|
* Base type for message content parts.
|
|
5961
6249
|
* Supports both camelCase (imageUrl) and snake_case (image_url) type discriminators via ContentPartJsonConverter.
|
|
@@ -5967,42 +6255,10 @@ export type ChatCompletionContentPartWritable = {
|
|
|
5967
6255
|
text?: null | string;
|
|
5968
6256
|
imageUrl?: ChatCompletionImageUrl;
|
|
5969
6257
|
};
|
|
5970
|
-
/**
|
|
5971
|
-
* ChatCompletion
|
|
5972
|
-
*/
|
|
5973
|
-
export type ChatCompletionStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
5974
|
-
input: ChatCompletionInput;
|
|
5975
|
-
output?: ChatCompletionOutput;
|
|
5976
|
-
$type: 'chatCompletion';
|
|
5977
|
-
};
|
|
5978
|
-
/**
|
|
5979
|
-
* Comfy workflows
|
|
5980
|
-
*/
|
|
5981
|
-
export type ComfyStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
5982
|
-
input: ComfyInput;
|
|
5983
|
-
output?: ComfyOutput;
|
|
5984
|
-
$type: 'comfy';
|
|
5985
|
-
};
|
|
5986
|
-
/**
|
|
5987
|
-
* A workflow step that converts images to different formats and applies optional transforms.
|
|
5988
|
-
*/
|
|
5989
|
-
export type ConvertImageStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
5990
|
-
input: ConvertImageInput;
|
|
5991
|
-
output?: ConvertImageOutput;
|
|
5992
|
-
$type: 'convertImage';
|
|
5993
|
-
};
|
|
5994
6258
|
export type CursedArrayOfTelemetryCursorAndWorkflowWritable = {
|
|
5995
6259
|
next: string;
|
|
5996
6260
|
items: Array<WorkflowWritable>;
|
|
5997
6261
|
};
|
|
5998
|
-
/**
|
|
5999
|
-
* Echo
|
|
6000
|
-
*/
|
|
6001
|
-
export type EchoStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6002
|
-
input: EchoInput;
|
|
6003
|
-
output?: EchoOutput;
|
|
6004
|
-
$type: 'echo';
|
|
6005
|
-
};
|
|
6006
6262
|
export type Flux1KontextDevImageGenInputWritable = Omit<
|
|
6007
6263
|
Flux1KontextImageGenInputWritable,
|
|
6008
6264
|
'engine'
|
|
@@ -6033,38 +6289,6 @@ export type Flux1KontextProImageGenInputWritable = Omit<
|
|
|
6033
6289
|
model: 'pro';
|
|
6034
6290
|
engine: 'flux1-kontext';
|
|
6035
6291
|
};
|
|
6036
|
-
/**
|
|
6037
|
-
* Image Generation
|
|
6038
|
-
*/
|
|
6039
|
-
export type ImageGenStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6040
|
-
input: ImageGenInput;
|
|
6041
|
-
output?: ImageGenOutput;
|
|
6042
|
-
$type: 'imageGen';
|
|
6043
|
-
};
|
|
6044
|
-
/**
|
|
6045
|
-
* LORA Training
|
|
6046
|
-
*/
|
|
6047
|
-
export type ImageResourceTrainingStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6048
|
-
input: ImageResourceTrainingInput;
|
|
6049
|
-
output?: ImageResourceTrainingOutput;
|
|
6050
|
-
$type: 'imageResourceTraining';
|
|
6051
|
-
};
|
|
6052
|
-
/**
|
|
6053
|
-
* Image upload
|
|
6054
|
-
*/
|
|
6055
|
-
export type ImageUploadStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6056
|
-
/**
|
|
6057
|
-
* The workflow's input.
|
|
6058
|
-
*/
|
|
6059
|
-
input: string;
|
|
6060
|
-
output?: ImageUploadOutput;
|
|
6061
|
-
$type: 'imageUpload';
|
|
6062
|
-
};
|
|
6063
|
-
export type ImageUpscalerStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6064
|
-
input: ImageUpscalerInput;
|
|
6065
|
-
output?: ImageUpscalerOutput;
|
|
6066
|
-
$type: 'imageUpscaler';
|
|
6067
|
-
};
|
|
6068
6292
|
export type KohyaImageResourceTrainingInputWritable = Omit<
|
|
6069
6293
|
ImageResourceTrainingInputWritable,
|
|
6070
6294
|
'engine'
|
|
@@ -6156,54 +6380,6 @@ export type KohyaImageResourceTrainingInputWritable = Omit<
|
|
|
6156
6380
|
optimizerType?: null | string;
|
|
6157
6381
|
engine: 'kohya';
|
|
6158
6382
|
};
|
|
6159
|
-
/**
|
|
6160
|
-
* Media Captioning
|
|
6161
|
-
*/
|
|
6162
|
-
export type MediaCaptioningStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6163
|
-
input: MediaCaptioningInput;
|
|
6164
|
-
output?: MediaCaptioningOutput;
|
|
6165
|
-
$type: 'mediaCaptioning';
|
|
6166
|
-
};
|
|
6167
|
-
/**
|
|
6168
|
-
* MediaHash
|
|
6169
|
-
*/
|
|
6170
|
-
export type MediaHashStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6171
|
-
input: MediaHashInput;
|
|
6172
|
-
output?: MediaHashOutput;
|
|
6173
|
-
$type: 'mediaHash';
|
|
6174
|
-
};
|
|
6175
|
-
/**
|
|
6176
|
-
* MediaRating
|
|
6177
|
-
*/
|
|
6178
|
-
export type MediaRatingStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6179
|
-
input: MediaRatingInput;
|
|
6180
|
-
output?: MediaRatingOutput;
|
|
6181
|
-
$type: 'mediaRating';
|
|
6182
|
-
};
|
|
6183
|
-
/**
|
|
6184
|
-
* ModelClamScan
|
|
6185
|
-
*/
|
|
6186
|
-
export type ModelClamScanStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6187
|
-
input: ModelClamScanInput;
|
|
6188
|
-
output?: ModelClamScanOutput;
|
|
6189
|
-
$type: 'modelClamScan';
|
|
6190
|
-
};
|
|
6191
|
-
/**
|
|
6192
|
-
* ModelHash
|
|
6193
|
-
*/
|
|
6194
|
-
export type ModelHashStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6195
|
-
input: ModelHashInput;
|
|
6196
|
-
output?: ModelHashOutput;
|
|
6197
|
-
$type: 'modelHash';
|
|
6198
|
-
};
|
|
6199
|
-
/**
|
|
6200
|
-
* ModelPickleScan
|
|
6201
|
-
*/
|
|
6202
|
-
export type ModelPickleScanStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6203
|
-
input: ModelPickleScanInput;
|
|
6204
|
-
output?: ModelPickleScanOutput;
|
|
6205
|
-
$type: 'modelPickleScan';
|
|
6206
|
-
};
|
|
6207
6383
|
export type MusubiImageResourceTrainingInputWritable = Omit<
|
|
6208
6384
|
ImageResourceTrainingInputWritable,
|
|
6209
6385
|
'engine'
|
|
@@ -6455,7 +6631,7 @@ export type PreprocessImageStandardLineartInputWritable = Omit<
|
|
|
6455
6631
|
intensityThreshold?: number;
|
|
6456
6632
|
kind: 'lineart-standard';
|
|
6457
6633
|
};
|
|
6458
|
-
export type PreprocessImageStepWritable = Omit<
|
|
6634
|
+
export type PreprocessImageStepWritable = Omit<WorkflowStepWritable, '$type'> & {
|
|
6459
6635
|
input: PreprocessImageInputWritable;
|
|
6460
6636
|
output?: PreprocessImageOutput;
|
|
6461
6637
|
$type: 'preprocessImage';
|
|
@@ -6485,14 +6661,6 @@ export type PreprocessImageZoeDepthAnythingInputWritable = Omit<
|
|
|
6485
6661
|
export type PreprocessImageZoeDepthInputWritable = Omit<PreprocessImageInputWritable2, 'kind'> & {
|
|
6486
6662
|
kind: 'zoe-depth';
|
|
6487
6663
|
};
|
|
6488
|
-
/**
|
|
6489
|
-
* PromptEnhancement
|
|
6490
|
-
*/
|
|
6491
|
-
export type PromptEnhancementStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6492
|
-
input: PromptEnhancementInput;
|
|
6493
|
-
output?: PromptEnhancementOutput;
|
|
6494
|
-
$type: 'promptEnhancement';
|
|
6495
|
-
};
|
|
6496
6664
|
export type Qwen20bEditImageGenInputWritable = Omit<
|
|
6497
6665
|
Qwen20bImageGenInput,
|
|
6498
6666
|
'engine' | 'ecosystem' | 'model' | 'operation'
|
|
@@ -6519,52 +6687,6 @@ export type Qwen20bVariantImageGenInputWritable = Omit<
|
|
|
6519
6687
|
ecosystem: 'qwen';
|
|
6520
6688
|
engine: 'sdcpp';
|
|
6521
6689
|
};
|
|
6522
|
-
/**
|
|
6523
|
-
* Represents the output information returned from the Repeat workflow step.
|
|
6524
|
-
*/
|
|
6525
|
-
export type RepeatOutputWritable = {
|
|
6526
|
-
steps: Array<WorkflowStepWritable>;
|
|
6527
|
-
};
|
|
6528
|
-
/**
|
|
6529
|
-
* TextToImage
|
|
6530
|
-
*/
|
|
6531
|
-
export type TextToImageStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6532
|
-
input: TextToImageInput;
|
|
6533
|
-
output?: TextToImageOutput;
|
|
6534
|
-
$type: 'textToImage';
|
|
6535
|
-
};
|
|
6536
|
-
/**
|
|
6537
|
-
* Text-to-Speech
|
|
6538
|
-
*/
|
|
6539
|
-
export type TextToSpeechStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6540
|
-
input: TextToSpeechInput;
|
|
6541
|
-
output?: TextToSpeechOutput;
|
|
6542
|
-
$type: 'textToSpeech';
|
|
6543
|
-
};
|
|
6544
|
-
/**
|
|
6545
|
-
* Training
|
|
6546
|
-
*/
|
|
6547
|
-
export type TrainingStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6548
|
-
input: TrainingInput;
|
|
6549
|
-
output?: TrainingOutput;
|
|
6550
|
-
$type: 'training';
|
|
6551
|
-
};
|
|
6552
|
-
/**
|
|
6553
|
-
* Transcoding
|
|
6554
|
-
*/
|
|
6555
|
-
export type TranscodeStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6556
|
-
input: TranscodeInput;
|
|
6557
|
-
output?: TranscodeOutput;
|
|
6558
|
-
$type: 'transcode';
|
|
6559
|
-
};
|
|
6560
|
-
/**
|
|
6561
|
-
* Transcription
|
|
6562
|
-
*/
|
|
6563
|
-
export type TranscriptionStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6564
|
-
input: TranscriptionInput;
|
|
6565
|
-
output?: TranscriptionOutput;
|
|
6566
|
-
$type: 'transcription';
|
|
6567
|
-
};
|
|
6568
6690
|
/**
|
|
6569
6691
|
* A user message that can contain text and/or images.
|
|
6570
6692
|
*/
|
|
@@ -6580,62 +6702,6 @@ export type UserMessageWritable = Omit<ChatCompletionMessageWritable, 'role'> &
|
|
|
6580
6702
|
name?: null | string;
|
|
6581
6703
|
role: 'user';
|
|
6582
6704
|
};
|
|
6583
|
-
/**
|
|
6584
|
-
* Upscale videos and/or interpolate frames
|
|
6585
|
-
*/
|
|
6586
|
-
export type VideoEnhancementStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6587
|
-
input: VideoEnhancementInput;
|
|
6588
|
-
output?: VideoEnhancementOutput;
|
|
6589
|
-
$type: 'videoEnhancement';
|
|
6590
|
-
};
|
|
6591
|
-
/**
|
|
6592
|
-
* Video Frame Extraction
|
|
6593
|
-
*/
|
|
6594
|
-
export type VideoFrameExtractionStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6595
|
-
input: VideoFrameExtractionInput;
|
|
6596
|
-
output?: VideoFrameExtractionOutput;
|
|
6597
|
-
$type: 'videoFrameExtraction';
|
|
6598
|
-
};
|
|
6599
|
-
/**
|
|
6600
|
-
* Video generation
|
|
6601
|
-
*/
|
|
6602
|
-
export type VideoGenStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6603
|
-
input: VideoGenInput;
|
|
6604
|
-
output?: VideoGenOutput;
|
|
6605
|
-
$type: 'videoGen';
|
|
6606
|
-
};
|
|
6607
|
-
/**
|
|
6608
|
-
* Interpolate videos using VFI Mamba
|
|
6609
|
-
*/
|
|
6610
|
-
export type VideoInterpolationStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6611
|
-
input: VideoInterpolationInput;
|
|
6612
|
-
output?: VideoInterpolationOutput;
|
|
6613
|
-
$type: 'videoInterpolation';
|
|
6614
|
-
};
|
|
6615
|
-
/**
|
|
6616
|
-
* Extract metadata from videos including width, height, FPS, and duration
|
|
6617
|
-
*/
|
|
6618
|
-
export type VideoMetadataStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6619
|
-
input: VideoMetadataInput;
|
|
6620
|
-
output?: VideoMetadataOutput;
|
|
6621
|
-
$type: 'videoMetadata';
|
|
6622
|
-
};
|
|
6623
|
-
/**
|
|
6624
|
-
* Upscale videos using FlashVSR
|
|
6625
|
-
*/
|
|
6626
|
-
export type VideoUpscalerStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6627
|
-
input: VideoUpscalerInput;
|
|
6628
|
-
output?: VideoUpscalerOutput;
|
|
6629
|
-
$type: 'videoUpscaler';
|
|
6630
|
-
};
|
|
6631
|
-
/**
|
|
6632
|
-
* WDTagging
|
|
6633
|
-
*/
|
|
6634
|
-
export type WdTaggingStepWritable = Omit<WorkflowStepWritable2, '$type'> & {
|
|
6635
|
-
input: WdTaggingInput;
|
|
6636
|
-
output?: WdTaggingOutput;
|
|
6637
|
-
$type: 'wdTagging';
|
|
6638
|
-
};
|
|
6639
6705
|
/**
|
|
6640
6706
|
* Details of a workflow.
|
|
6641
6707
|
*/
|
|
@@ -6734,44 +6800,6 @@ export type WorkflowEventWritable = {
|
|
|
6734
6800
|
timestamp: string;
|
|
6735
6801
|
details?: WorkflowEventDetails;
|
|
6736
6802
|
};
|
|
6737
|
-
/**
|
|
6738
|
-
* Details of a workflow step.
|
|
6739
|
-
*/
|
|
6740
|
-
export type WorkflowStepWritable = {
|
|
6741
|
-
$type: string;
|
|
6742
|
-
/**
|
|
6743
|
-
* The name of the workflow step. Used to allow steps to refer to one another.
|
|
6744
|
-
*/
|
|
6745
|
-
name: string;
|
|
6746
|
-
priority: Priority;
|
|
6747
|
-
/**
|
|
6748
|
-
* The maximum time to wait for this step to complete.
|
|
6749
|
-
*/
|
|
6750
|
-
timeout?: null | string;
|
|
6751
|
-
/**
|
|
6752
|
-
* The maximum number of times this step should be retried.
|
|
6753
|
-
*/
|
|
6754
|
-
retries?: null | number;
|
|
6755
|
-
/**
|
|
6756
|
-
* The jobs generated by this step.
|
|
6757
|
-
*/
|
|
6758
|
-
jobs?: null | Array<WorkflowStepJob>;
|
|
6759
|
-
status: WorkflowStatus;
|
|
6760
|
-
/**
|
|
6761
|
-
* The date / time the step was started. Null if not yet started.
|
|
6762
|
-
*/
|
|
6763
|
-
startedAt?: null | string;
|
|
6764
|
-
/**
|
|
6765
|
-
* The date / time the step was completed. Null if not yet completed.
|
|
6766
|
-
*/
|
|
6767
|
-
completedAt?: null | string;
|
|
6768
|
-
/**
|
|
6769
|
-
* A collection of user defined metadata for the workflow step.
|
|
6770
|
-
*/
|
|
6771
|
-
metadata: {
|
|
6772
|
-
[key: string]: unknown;
|
|
6773
|
-
};
|
|
6774
|
-
};
|
|
6775
6803
|
/**
|
|
6776
6804
|
* Details of a workflow step event.
|
|
6777
6805
|
*/
|
|
@@ -6830,7 +6858,7 @@ export type XGuardModerationOutputWritable = {
|
|
|
6830
6858
|
/**
|
|
6831
6859
|
* XGuardModeration
|
|
6832
6860
|
*/
|
|
6833
|
-
export type XGuardModerationStepWritable = Omit<
|
|
6861
|
+
export type XGuardModerationStepWritable = Omit<WorkflowStepWritable, '$type'> & {
|
|
6834
6862
|
input: XGuardModerationInput;
|
|
6835
6863
|
output?: XGuardModerationOutputWritable;
|
|
6836
6864
|
$type: 'xGuardModeration';
|
|
@@ -6856,48 +6884,6 @@ export type XGuardSignalMetadataWritable = {
|
|
|
6856
6884
|
};
|
|
6857
6885
|
customSignals: Array<XGuardCustomSignalMetadataWritable>;
|
|
6858
6886
|
};
|
|
6859
|
-
/**
|
|
6860
|
-
* Details of a workflow step.
|
|
6861
|
-
*/
|
|
6862
|
-
export type WorkflowStepWritable2 = {
|
|
6863
|
-
$type: string;
|
|
6864
|
-
/**
|
|
6865
|
-
* The name of the workflow step. Used to allow steps to refer to one another.
|
|
6866
|
-
*/
|
|
6867
|
-
name: string;
|
|
6868
|
-
priority: Priority;
|
|
6869
|
-
/**
|
|
6870
|
-
* The maximum time to wait for this step to complete.
|
|
6871
|
-
*/
|
|
6872
|
-
timeout?: null | string;
|
|
6873
|
-
/**
|
|
6874
|
-
* The maximum number of times this step should be retried.
|
|
6875
|
-
*/
|
|
6876
|
-
retries?: null | number;
|
|
6877
|
-
/**
|
|
6878
|
-
* The jobs generated by this step.
|
|
6879
|
-
*/
|
|
6880
|
-
jobs?: null | Array<WorkflowStepJob>;
|
|
6881
|
-
status: WorkflowStatus;
|
|
6882
|
-
/**
|
|
6883
|
-
* The date / time the step was started. Null if not yet started.
|
|
6884
|
-
*/
|
|
6885
|
-
startedAt?: null | string;
|
|
6886
|
-
/**
|
|
6887
|
-
* The date / time the step was completed. Null if not yet completed.
|
|
6888
|
-
*/
|
|
6889
|
-
completedAt?: null | string;
|
|
6890
|
-
/**
|
|
6891
|
-
* A collection of user defined metadata for the workflow step.
|
|
6892
|
-
*/
|
|
6893
|
-
metadata: {
|
|
6894
|
-
[key: string]: unknown;
|
|
6895
|
-
};
|
|
6896
|
-
/**
|
|
6897
|
-
* An estimation on the current progression of this step, or null if there is no estimation
|
|
6898
|
-
*/
|
|
6899
|
-
readonly estimatedProgressRate?: null | number;
|
|
6900
|
-
};
|
|
6901
6887
|
export type ImageGenInputWritable = {
|
|
6902
6888
|
engine: string;
|
|
6903
6889
|
outputFormat?: ImageGenOutputFormat;
|
|
@@ -6949,6 +6935,51 @@ export type PreprocessImageInputWritable2 = {
|
|
|
6949
6935
|
*/
|
|
6950
6936
|
readonly preprocessorType: string;
|
|
6951
6937
|
};
|
|
6938
|
+
/**
|
|
6939
|
+
* Details of a workflow step.
|
|
6940
|
+
*/
|
|
6941
|
+
export type WorkflowStepWritable = {
|
|
6942
|
+
$type: string;
|
|
6943
|
+
/**
|
|
6944
|
+
* The name of the workflow step. Used to allow steps to refer to one another.
|
|
6945
|
+
*/
|
|
6946
|
+
name: string;
|
|
6947
|
+
priority: Priority;
|
|
6948
|
+
/**
|
|
6949
|
+
* The maximum time to wait for this step to complete.
|
|
6950
|
+
*/
|
|
6951
|
+
timeout?: null | string;
|
|
6952
|
+
/**
|
|
6953
|
+
* The maximum number of times this step should be retried.
|
|
6954
|
+
*/
|
|
6955
|
+
retries?: null | number;
|
|
6956
|
+
/**
|
|
6957
|
+
* The jobs generated by this step.
|
|
6958
|
+
*/
|
|
6959
|
+
jobs?: null | Array<WorkflowStepJob>;
|
|
6960
|
+
status: WorkflowStatus;
|
|
6961
|
+
/**
|
|
6962
|
+
* The date / time the step was started. Null if not yet started.
|
|
6963
|
+
*/
|
|
6964
|
+
startedAt?: null | string;
|
|
6965
|
+
/**
|
|
6966
|
+
* The date / time the step was completed. Null if not yet completed.
|
|
6967
|
+
*/
|
|
6968
|
+
completedAt?: null | string;
|
|
6969
|
+
/**
|
|
6970
|
+
* A collection of user defined metadata for the workflow step.
|
|
6971
|
+
*/
|
|
6972
|
+
metadata: {
|
|
6973
|
+
[key: string]: unknown;
|
|
6974
|
+
};
|
|
6975
|
+
/**
|
|
6976
|
+
* An estimation on the current progression of this step, or null if there is no estimation.
|
|
6977
|
+
* Computed by the step's handler (see `WorkflowStepHandler.GetEstimatedProgressRate`)
|
|
6978
|
+
* and refreshed by `WorkflowStepManager` on each job event. Cleared to null once the
|
|
6979
|
+
* step reaches a final status.
|
|
6980
|
+
*/
|
|
6981
|
+
estimatedProgressRate?: null | number;
|
|
6982
|
+
};
|
|
6952
6983
|
/**
|
|
6953
6984
|
* Details of a workflow step template.
|
|
6954
6985
|
*/
|
|
@@ -6998,6 +7029,14 @@ export type GetBlobData = {
|
|
|
6998
7029
|
* A maximum nsfw level. If this is specified and the blob does not have a NSFW level specified or the NSFW level exceeds our max then we'll return an error
|
|
6999
7030
|
*/
|
|
7000
7031
|
nsfwLevel?: NsfwLevel;
|
|
7032
|
+
/**
|
|
7033
|
+
* Optional filename to use in the Content-Disposition header when the blob is served. Path components and control characters are stripped.
|
|
7034
|
+
*/
|
|
7035
|
+
filename?: string;
|
|
7036
|
+
/**
|
|
7037
|
+
* When true, the blob is served with Content-Disposition: attachment to prompt a download. Defaults to inline.
|
|
7038
|
+
*/
|
|
7039
|
+
download?: boolean;
|
|
7001
7040
|
};
|
|
7002
7041
|
url: '/v2/consumer/blobs/{blobId}';
|
|
7003
7042
|
};
|
|
@@ -7162,6 +7201,34 @@ export type GetBlockedContentErrors = {
|
|
|
7162
7201
|
403: ProblemDetails;
|
|
7163
7202
|
};
|
|
7164
7203
|
export type GetBlockedContentError = GetBlockedContentErrors[keyof GetBlockedContentErrors];
|
|
7204
|
+
export type GetBlobArchiveData = {
|
|
7205
|
+
body?: never;
|
|
7206
|
+
path: {
|
|
7207
|
+
/**
|
|
7208
|
+
* The signed token containing the archive manifest.
|
|
7209
|
+
*/
|
|
7210
|
+
encryptedToken: string;
|
|
7211
|
+
};
|
|
7212
|
+
query?: never;
|
|
7213
|
+
url: '/v2/consumer/blobs/archive/{encryptedToken}';
|
|
7214
|
+
};
|
|
7215
|
+
export type GetBlobArchiveErrors = {
|
|
7216
|
+
/**
|
|
7217
|
+
* Unauthorized
|
|
7218
|
+
*/
|
|
7219
|
+
401: ProblemDetails;
|
|
7220
|
+
/**
|
|
7221
|
+
* Gone
|
|
7222
|
+
*/
|
|
7223
|
+
410: ProblemDetails;
|
|
7224
|
+
};
|
|
7225
|
+
export type GetBlobArchiveError = GetBlobArchiveErrors[keyof GetBlobArchiveErrors];
|
|
7226
|
+
export type GetBlobArchiveResponses = {
|
|
7227
|
+
/**
|
|
7228
|
+
* OK
|
|
7229
|
+
*/
|
|
7230
|
+
200: unknown;
|
|
7231
|
+
};
|
|
7165
7232
|
export type RefreshBlobData = {
|
|
7166
7233
|
body?: never;
|
|
7167
7234
|
path: {
|
|
@@ -7308,6 +7375,36 @@ export type InvokeBatchOcrSafetyClassificationStepTemplateResponses = {
|
|
|
7308
7375
|
};
|
|
7309
7376
|
export type InvokeBatchOcrSafetyClassificationStepTemplateResponse =
|
|
7310
7377
|
InvokeBatchOcrSafetyClassificationStepTemplateResponses[keyof InvokeBatchOcrSafetyClassificationStepTemplateResponses];
|
|
7378
|
+
export type InvokeBlobArchiveStepTemplateData = {
|
|
7379
|
+
body?: BlobArchiveInput;
|
|
7380
|
+
path?: never;
|
|
7381
|
+
query?: {
|
|
7382
|
+
experimental?: boolean;
|
|
7383
|
+
allowMatureContent?: boolean;
|
|
7384
|
+
whatif?: boolean;
|
|
7385
|
+
};
|
|
7386
|
+
url: '/v2/consumer/recipes/blobArchive';
|
|
7387
|
+
};
|
|
7388
|
+
export type InvokeBlobArchiveStepTemplateErrors = {
|
|
7389
|
+
/**
|
|
7390
|
+
* Bad Request
|
|
7391
|
+
*/
|
|
7392
|
+
400: ProblemDetails;
|
|
7393
|
+
/**
|
|
7394
|
+
* Unauthorized
|
|
7395
|
+
*/
|
|
7396
|
+
401: ProblemDetails;
|
|
7397
|
+
};
|
|
7398
|
+
export type InvokeBlobArchiveStepTemplateError =
|
|
7399
|
+
InvokeBlobArchiveStepTemplateErrors[keyof InvokeBlobArchiveStepTemplateErrors];
|
|
7400
|
+
export type InvokeBlobArchiveStepTemplateResponses = {
|
|
7401
|
+
/**
|
|
7402
|
+
* OK
|
|
7403
|
+
*/
|
|
7404
|
+
200: BlobArchiveOutput;
|
|
7405
|
+
};
|
|
7406
|
+
export type InvokeBlobArchiveStepTemplateResponse =
|
|
7407
|
+
InvokeBlobArchiveStepTemplateResponses[keyof InvokeBlobArchiveStepTemplateResponses];
|
|
7311
7408
|
export type InvokeChatCompletionStepTemplateData = {
|
|
7312
7409
|
body?: ChatCompletionInput;
|
|
7313
7410
|
path?: never;
|
|
@@ -7728,6 +7825,36 @@ export type InvokeModelHashStepTemplateResponses = {
|
|
|
7728
7825
|
};
|
|
7729
7826
|
export type InvokeModelHashStepTemplateResponse =
|
|
7730
7827
|
InvokeModelHashStepTemplateResponses[keyof InvokeModelHashStepTemplateResponses];
|
|
7828
|
+
export type InvokeModelParseMetadataStepTemplateData = {
|
|
7829
|
+
body?: ModelParseMetadataInput;
|
|
7830
|
+
path?: never;
|
|
7831
|
+
query?: {
|
|
7832
|
+
experimental?: boolean;
|
|
7833
|
+
allowMatureContent?: boolean;
|
|
7834
|
+
whatif?: boolean;
|
|
7835
|
+
};
|
|
7836
|
+
url: '/v2/consumer/recipes/modelParseMetadata';
|
|
7837
|
+
};
|
|
7838
|
+
export type InvokeModelParseMetadataStepTemplateErrors = {
|
|
7839
|
+
/**
|
|
7840
|
+
* Bad Request
|
|
7841
|
+
*/
|
|
7842
|
+
400: ProblemDetails;
|
|
7843
|
+
/**
|
|
7844
|
+
* Unauthorized
|
|
7845
|
+
*/
|
|
7846
|
+
401: ProblemDetails;
|
|
7847
|
+
};
|
|
7848
|
+
export type InvokeModelParseMetadataStepTemplateError =
|
|
7849
|
+
InvokeModelParseMetadataStepTemplateErrors[keyof InvokeModelParseMetadataStepTemplateErrors];
|
|
7850
|
+
export type InvokeModelParseMetadataStepTemplateResponses = {
|
|
7851
|
+
/**
|
|
7852
|
+
* OK
|
|
7853
|
+
*/
|
|
7854
|
+
200: ModelParseMetadataOutput;
|
|
7855
|
+
};
|
|
7856
|
+
export type InvokeModelParseMetadataStepTemplateResponse =
|
|
7857
|
+
InvokeModelParseMetadataStepTemplateResponses[keyof InvokeModelParseMetadataStepTemplateResponses];
|
|
7731
7858
|
export type InvokeModelPickleScanStepTemplateData = {
|
|
7732
7859
|
body?: ModelPickleScanInput;
|
|
7733
7860
|
path?: never;
|