@civitai/client 0.2.0-beta.56 → 0.2.0-beta.58

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -96,6 +96,49 @@ export type AiToolkitTrainingInput = Omit<TrainingInput, 'engine'> & {
96
96
  triggerWord?: null | string;
97
97
  engine: 'ai-toolkit';
98
98
  };
99
+ /**
100
+ * AI Toolkit training for ACE-Step 1.5 base models.
101
+ */
102
+ export type AceStep15AiToolkitTrainingInput = Omit<
103
+ AiToolkitTrainingInput,
104
+ 'engine' | 'ecosystem'
105
+ > & {
106
+ ecosystem: 'ace_step_15';
107
+ engine: 'ai-toolkit';
108
+ };
109
+ /**
110
+ * AI Toolkit training for ACE-Step 1.5 XL models.
111
+ */
112
+ export type AceStep15XlAiToolkitTrainingInput = Omit<
113
+ AiToolkitTrainingInput,
114
+ 'engine' | 'ecosystem'
115
+ > & {
116
+ modelVariant: string;
117
+ ecosystem: 'ace_step_15_xl';
118
+ engine: 'ai-toolkit';
119
+ };
120
+ /**
121
+ * AI Toolkit training for ACE-Step 1.5 XL base models.
122
+ */
123
+ export type AceStep15XlBaseAiToolkitTrainingInput = Omit<
124
+ AceStep15XlAiToolkitTrainingInput,
125
+ 'engine' | 'ecosystem' | 'modelVariant'
126
+ > & {
127
+ modelVariant: 'base';
128
+ ecosystem: 'ace_step_15_xl';
129
+ engine: 'ai-toolkit';
130
+ };
131
+ /**
132
+ * AI Toolkit training for ACE-Step 1.5 XL SFT models.
133
+ */
134
+ export type AceStep15XlSftAiToolkitTrainingInput = Omit<
135
+ AceStep15XlAiToolkitTrainingInput,
136
+ 'engine' | 'ecosystem' | 'modelVariant'
137
+ > & {
138
+ modelVariant: 'sft';
139
+ ecosystem: 'ace_step_15_xl';
140
+ engine: 'ai-toolkit';
141
+ };
99
142
  /**
100
143
  * Cover image configuration for ACE Step audio output.
101
144
  * When present, the output is a WebM video with this image as the visual.
@@ -133,15 +176,73 @@ export type AceStepAudioInput = {
133
176
  /**
134
177
  * Time signature (e.g., "4" for 4/4 time)
135
178
  */
136
- timeSignature?: string;
179
+ timeSignature?: '2' | '3' | '4' | '6';
137
180
  /**
138
181
  * Language code (e.g., "en", "zh", "ja", "ko")
139
182
  */
140
- language?: string;
141
- /**
142
- * Musical key (e.g., "C major", "E minor")
143
- */
144
- key?: string;
183
+ language?:
184
+ | 'en'
185
+ | 'ja'
186
+ | 'zh'
187
+ | 'es'
188
+ | 'de'
189
+ | 'fr'
190
+ | 'pt'
191
+ | 'ru'
192
+ | 'it'
193
+ | 'nl'
194
+ | 'pl'
195
+ | 'tr'
196
+ | 'vi'
197
+ | 'cs'
198
+ | 'fa'
199
+ | 'id'
200
+ | 'ko'
201
+ | 'uk'
202
+ | 'hu'
203
+ | 'ar'
204
+ | 'sv'
205
+ | 'ro'
206
+ | 'el';
207
+ /**
208
+ * Musical key (e.g., "C major", "E minor"). Mirrors ComfyUI's
209
+ * TextEncodeAceStepAudio1.5 keyscale combo: 17 roots × {major, minor}.
210
+ */
211
+ key?:
212
+ | 'C major'
213
+ | 'C# major'
214
+ | 'Db major'
215
+ | 'D major'
216
+ | 'D# major'
217
+ | 'Eb major'
218
+ | 'E major'
219
+ | 'F major'
220
+ | 'F# major'
221
+ | 'Gb major'
222
+ | 'G major'
223
+ | 'G# major'
224
+ | 'Ab major'
225
+ | 'A major'
226
+ | 'A# major'
227
+ | 'Bb major'
228
+ | 'B major'
229
+ | 'C minor'
230
+ | 'C# minor'
231
+ | 'Db minor'
232
+ | 'D minor'
233
+ | 'D# minor'
234
+ | 'Eb minor'
235
+ | 'E minor'
236
+ | 'F minor'
237
+ | 'F# minor'
238
+ | 'Gb minor'
239
+ | 'G minor'
240
+ | 'G# minor'
241
+ | 'Ab minor'
242
+ | 'A minor'
243
+ | 'A# minor'
244
+ | 'Bb minor'
245
+ | 'B minor';
145
246
  /**
146
247
  * Weight for instrumental elements (0.0-1.0)
147
248
  */
@@ -177,6 +278,14 @@ export type AceStepAudioInput = {
177
278
  * non-turbo base / sft variants expect CFG on (around 4).
178
279
  */
179
280
  cfg?: number;
281
+ /**
282
+ * Optional LoRAs to apply. Each entry's strength is applied to both the
283
+ * diffusion model (UNET) and the dual CLIP via ComfyUI's LoraLoader.
284
+ * Compatibility with non-default base models is the caller's responsibility.
285
+ */
286
+ loras: {
287
+ [key: string]: number;
288
+ };
180
289
  };
181
290
  /**
182
291
  * Output from ACE Step 1.5 audio generation workflow step.
@@ -413,6 +522,11 @@ export type AssistantMessage = Omit<ChatCompletionMessage, 'role'> & {
413
522
  * Tool calls requested by the model.
414
523
  */
415
524
  tool_calls?: null | Array<ChatCompletionToolCall>;
525
+ /**
526
+ * Generated images attached to this assistant message, populated when the request
527
+ * included "image" in its modalities. Each entry is a base64 data URI.
528
+ */
529
+ images?: null | Array<ChatCompletionGeneratedImage>;
416
530
  role: 'assistant';
417
531
  };
418
532
  export type AudioBlob = Omit<Blob, 'type'> & {
@@ -422,6 +536,57 @@ export type AudioBlob = Omit<Blob, 'type'> & {
422
536
  duration?: null | number;
423
537
  type: 'audio';
424
538
  };
539
+ export type AudioCaptioningInput = {
540
+ /**
541
+ * The URL of the audio file or zip archive to caption.
542
+ */
543
+ mediaUrl: string;
544
+ /**
545
+ * Sampling temperature for both ACE-Step requests.
546
+ */
547
+ temperature: number;
548
+ /**
549
+ * Maximum number of tokens to generate for each ACE-Step request.
550
+ */
551
+ maxNewTokens: number;
552
+ };
553
+ export type AudioCaptioningOutput = {
554
+ /**
555
+ * Combined caption, transcription, and music metadata for the first or only audio item.
556
+ */
557
+ text?: null | string;
558
+ /**
559
+ * Per-file results when the input is an archive or contains multiple audio items.
560
+ */
561
+ results: {
562
+ [key: string]: AudioCaptioningOutputItem;
563
+ };
564
+ };
565
+ export type AudioCaptioningOutputItem = {
566
+ text: string;
567
+ caption?: null | string;
568
+ lyrics?: null | string;
569
+ bpm?: null | string;
570
+ keyScale?: null | string;
571
+ timeSignature?: null | string;
572
+ duration?: null | string;
573
+ language?: null | string;
574
+ };
575
+ /**
576
+ * Audio Captioning
577
+ */
578
+ export type AudioCaptioningStep = Omit<WorkflowStep, '$type'> & {
579
+ input: AudioCaptioningInput;
580
+ output?: AudioCaptioningOutput;
581
+ $type: 'audioCaptioning';
582
+ };
583
+ /**
584
+ * Audio Captioning
585
+ */
586
+ export type AudioCaptioningStepTemplate = Omit<WorkflowStepTemplate, '$type'> & {
587
+ input: AudioCaptioningInput;
588
+ $type: 'audioCaptioning';
589
+ };
425
590
  export type BatchOcrSafetyClassificationInput = {
426
591
  mediaUrls: Array<string>;
427
592
  };
@@ -639,6 +804,44 @@ export type ChatCompletionFunctionCall = {
639
804
  name: string;
640
805
  arguments: string;
641
806
  };
807
+ /**
808
+ * A generated image attached to an assistant message when "image" is included in
809
+ * Civitai.Orchestration.Grains.Workflows.Steps.ChatCompletion.ChatCompletionInput.Modalities. Matches the OpenRouter wire shape.
810
+ */
811
+ export type ChatCompletionGeneratedImage = {
812
+ /**
813
+ * Always "image_url".
814
+ */
815
+ type?: string;
816
+ image_url: ChatCompletionGeneratedImageUrl;
817
+ };
818
+ /**
819
+ * The image_url payload on a generated image. Currently always a base64 data URI.
820
+ */
821
+ export type ChatCompletionGeneratedImageUrl = {
822
+ /**
823
+ * Base64 data URI (e.g. "data:image/png;base64,...").
824
+ */
825
+ url: string;
826
+ };
827
+ /**
828
+ * Image generation parameters used when Civitai.Orchestration.Grains.Workflows.Steps.ChatCompletion.ChatCompletionInput.Modalities contains "image".
829
+ * Mirrors OpenRouter's `image_config` shape on chat-completion requests.
830
+ */
831
+ export type ChatCompletionImageConfig = {
832
+ /**
833
+ * Width:height aspect ratio. Examples: "1:1", "16:9", "9:16", "4:3", "3:4", "21:9".
834
+ */
835
+ aspect_ratio?: '1:1' | '16:9' | '9:16' | '4:3' | '3:4' | '21:9';
836
+ /**
837
+ * Approximate output resolution. "1K" ≈ 1MP, "2K" ≈ 2MP, etc. Engines clamp to their supported range.
838
+ */
839
+ image_size?: '0.5K' | '1K' | '2K' | '4K';
840
+ /**
841
+ * Number of images to generate. Engines may clamp to their supported maximum.
842
+ */
843
+ n?: null | number;
844
+ };
642
845
  /**
643
846
  * Image URL details matching OpenAI API spec.
644
847
  */
@@ -727,6 +930,23 @@ export type ChatCompletionInput = {
727
930
  * Can be "auto", "none", "required", or an object specifying a particular function.
728
931
  */
729
932
  tool_choice?: null;
933
+ /**
934
+ * Output modalities the model should produce. Defaults to text-only when omitted.
935
+ * Supported values: "text", "image". When "image" is included, the request is routed
936
+ * to the image generation pipeline and returns generated images on the assistant message.
937
+ */
938
+ modalities?: null | Array<string>;
939
+ image_config?: ChatCompletionImageConfig;
940
+ responseFormat?: ChatCompletionResponseFormat;
941
+ };
942
+ export type ChatCompletionJsonSchema = {
943
+ name: string;
944
+ description?: null | string;
945
+ schema: unknown;
946
+ /**
947
+ * OpenAI strict mode. Forces additionalProperties=false and all fields required.
948
+ */
949
+ strict?: null | boolean;
730
950
  };
731
951
  /**
732
952
  * Base type for chat messages, discriminated by the "role" property.
@@ -764,6 +984,27 @@ export type ChatCompletionOutput = {
764
984
  * System fingerprint for the model configuration.
765
985
  */
766
986
  systemFingerprint?: null | string;
987
+ /**
988
+ * Parsed JSON content of `Choices[0].Message.Content`. Populated when the request
989
+ * specified a JSON-flavored `response_format` and the content was parseable.
990
+ * Reachable from downstream workflow steps via `$ref` paths like
991
+ * `output.parsed.<field>` — DynamicAssignmentEvaluator walks JsonElement trees
992
+ * the same way it walks the rest of this output.
993
+ */
994
+ parsed?: null;
995
+ };
996
+ /**
997
+ * OpenAI-compatible response_format. When Civitai.Orchestration.Grains.Workflows.Steps.ChatCompletion.ChatCompletionResponseFormat.Type is `json_object` or
998
+ * `json_schema`, the LLM-emitted content is parsed as JSON server-side and exposed on
999
+ * Civitai.Orchestration.Grains.Workflows.Steps.ChatCompletion.ChatCompletionOutput.Parsed, allowing downstream workflow steps to reference
1000
+ * individual fields via `$ref` paths like `output.parsed.<field>`.
1001
+ */
1002
+ export type ChatCompletionResponseFormat = {
1003
+ /**
1004
+ * OpenAI-compatible value: `text`, `json_object`, or `json_schema`.
1005
+ */
1006
+ type: string;
1007
+ jsonSchema?: ChatCompletionJsonSchema;
767
1008
  };
768
1009
  /**
769
1010
  * ChatCompletion
@@ -785,7 +1026,11 @@ export type ChatCompletionStepTemplate = Omit<WorkflowStepTemplate, '$type'> & {
785
1026
  */
786
1027
  export type ChatCompletionTool = {
787
1028
  type: string;
788
- function: ChatCompletionFunction;
1029
+ function?: ChatCompletionFunction;
1030
+ /**
1031
+ * Server-tool parameters for providers such as OpenRouter.
1032
+ */
1033
+ parameters?: null;
789
1034
  };
790
1035
  /**
791
1036
  * A tool call returned in an assistant message response.
@@ -824,6 +1069,33 @@ export declare const CoarseMode: {
824
1069
  readonly ENABLE: 'enable';
825
1070
  };
826
1071
  export type CoarseMode = (typeof CoarseMode)[keyof typeof CoarseMode];
1072
+ export type ComfyAnimaCreateImageGenInput = Omit<
1073
+ ComfyAnimaImageGenInput,
1074
+ 'engine' | 'ecosystem' | 'operation'
1075
+ > & {
1076
+ width?: number;
1077
+ height?: number;
1078
+ operation: 'createImage';
1079
+ ecosystem: 'anima';
1080
+ engine: 'comfy';
1081
+ };
1082
+ export type ComfyAnimaImageGenInput = Omit<ComfyImageGenInput, 'engine' | 'ecosystem'> & {
1083
+ operation: string;
1084
+ prompt: string;
1085
+ negativePrompt?: null | string;
1086
+ sampler?: ComfySampler;
1087
+ scheduler?: ComfyScheduler;
1088
+ steps?: number;
1089
+ cfgScale?: number;
1090
+ seed?: null | number;
1091
+ quantity?: number;
1092
+ loras?: {
1093
+ [key: string]: number;
1094
+ };
1095
+ diffuserModel?: string;
1096
+ ecosystem: 'anima';
1097
+ engine: 'comfy';
1098
+ };
827
1099
  export type ComfyErnieImageGenInput = Omit<ComfyImageGenInput, 'engine' | 'ecosystem'> & {
828
1100
  model: string;
829
1101
  ecosystem: 'ernie';
@@ -935,6 +1207,41 @@ export type ComfyFlux1VariantImageGenInput = Omit<
935
1207
  ecosystem: 'flux1';
936
1208
  engine: 'comfy';
937
1209
  };
1210
+ export type ComfyFlux2DevCreateImageInput = Omit<
1211
+ ComfyFlux2DevImageGenInput,
1212
+ 'engine' | 'ecosystem' | 'operation'
1213
+ > & {
1214
+ operation: 'createImage';
1215
+ ecosystem: 'flux2Dev';
1216
+ engine: 'comfy';
1217
+ };
1218
+ export type ComfyFlux2DevEditImageInput = Omit<
1219
+ ComfyFlux2DevImageGenInput,
1220
+ 'engine' | 'ecosystem' | 'operation'
1221
+ > & {
1222
+ images?: Array<string>;
1223
+ operation: 'editImage';
1224
+ ecosystem: 'flux2Dev';
1225
+ engine: 'comfy';
1226
+ };
1227
+ export type ComfyFlux2DevImageGenInput = Omit<ComfyImageGenInput, 'engine' | 'ecosystem'> & {
1228
+ operation: string;
1229
+ prompt: string;
1230
+ width?: number;
1231
+ height?: number;
1232
+ seed?: null | number;
1233
+ quantity?: number;
1234
+ cfgScale?: number;
1235
+ steps?: number;
1236
+ sampler?: ComfySampler;
1237
+ scheduler?: ComfyScheduler;
1238
+ negativePrompt?: null | string;
1239
+ loras?: {
1240
+ [key: string]: number;
1241
+ };
1242
+ ecosystem: 'flux2Dev';
1243
+ engine: 'comfy';
1244
+ };
938
1245
  export type ComfyImageGenInput = Omit<ImageGenInput, 'engine'> & {
939
1246
  ecosystem: string;
940
1247
  engine: 'comfy';
@@ -1055,6 +1362,12 @@ export type ComfyLtx23VideoGenInput = Omit<VideoGenInput, 'engine'> & {
1055
1362
  loras?: {
1056
1363
  [key: string]: number;
1057
1364
  };
1365
+ /**
1366
+ * Optional override for the LTX 2.3 diffusion-model checkpoint. When set, replaces the
1367
+ * transformer file selected by Civitai.Orchestration.Grains.Workflows.Steps.VideoGen.ComfyLtx23VideoGenInput.Model while leaving the CLIPs, VAEs, and
1368
+ * upscale-LoRA behavior unchanged. Use to point at a community fine-tune (e.g. SulphurAI/Sulphur-2-base).
1369
+ */
1370
+ diffusionModel?: null | string;
1058
1371
  engine: 'ltx2.3';
1059
1372
  };
1060
1373
  /**
@@ -1188,6 +1501,7 @@ export declare const ComfySampler: {
1188
1501
  readonly DDIM: 'ddim';
1189
1502
  readonly UNI_PC: 'uni_pc';
1190
1503
  readonly UNI_PC_BH2: 'uni_pc_bh2';
1504
+ readonly ER_SDE: 'er_sde';
1191
1505
  };
1192
1506
  export type ComfySampler = (typeof ComfySampler)[keyof typeof ComfySampler];
1193
1507
  export declare const ComfyScheduler: {
@@ -3030,7 +3344,27 @@ export type ModelClamScanOutput = {
3030
3344
  * The raw ClamAV scan output.
3031
3345
  */
3032
3346
  output?: null | string;
3347
+ status?: ModelClamScanStatus;
3348
+ /**
3349
+ * True when ClamAV reported one or more infected files (exit code 1).
3350
+ */
3351
+ infected?: null | boolean;
3352
+ /**
3353
+ * Number of infected files parsed from the ClamAV scan summary.
3354
+ */
3355
+ infectedFileCount?: null | number;
3356
+ /**
3357
+ * Number of files scanned, parsed from the ClamAV scan summary.
3358
+ */
3359
+ scannedFileCount?: null | number;
3033
3360
  };
3361
+ export declare const ModelClamScanStatus: {
3362
+ readonly CLEAN: 'clean';
3363
+ readonly INFECTED: 'infected';
3364
+ readonly TIMEOUT: 'timeout';
3365
+ readonly ERROR: 'error';
3366
+ };
3367
+ export type ModelClamScanStatus = (typeof ModelClamScanStatus)[keyof typeof ModelClamScanStatus];
3034
3368
  /**
3035
3369
  * ModelClamScan
3036
3370
  */
@@ -3113,8 +3447,9 @@ export type ModelParseMetadataInput = {
3113
3447
  */
3114
3448
  export type ModelParseMetadataOutput = {
3115
3449
  /**
3116
- * The raw JSON metadata header as stored in the safetensors file, or null when
3117
- * the model is not a safetensors file or the header could not be parsed.
3450
+ * The `__metadata__` object from the safetensors header as a JSON string,
3451
+ * or null when the model is not a safetensors file, the header could not be
3452
+ * parsed, or no `__metadata__` object is present.
3118
3453
  */
3119
3454
  metadata?: null | string;
3120
3455
  };
@@ -3162,7 +3497,43 @@ export type ModelPickleScanOutput = {
3162
3497
  * Dangerous imports discovered during pickle scanning.
3163
3498
  */
3164
3499
  dangerousImports?: null | Array<string>;
3500
+ status?: ModelPickleScanStatus;
3501
+ /**
3502
+ * True when one or more dangerous imports were detected.
3503
+ */
3504
+ dangerousImportsFound?: null | boolean;
3505
+ /**
3506
+ * True when picklescan was skipped (e.g. file is safetensors and cannot contain pickled code).
3507
+ */
3508
+ skipped?: null | boolean;
3509
+ /**
3510
+ * Reason picklescan was skipped, if applicable. Examples: "safetensors", "safetensors-extension".
3511
+ */
3512
+ skipReason?: null | string;
3513
+ /**
3514
+ * Number of files scanned, parsed from the picklescan summary.
3515
+ */
3516
+ scannedFileCount?: null | number;
3517
+ /**
3518
+ * Number of infected files reported by picklescan.
3519
+ */
3520
+ infectedFileCount?: null | number;
3521
+ /**
3522
+ * Number of dangerous globals reported by picklescan.
3523
+ */
3524
+ dangerousGlobalCount?: null | number;
3165
3525
  };
3526
+ export declare const ModelPickleScanStatus: {
3527
+ readonly CLEAN: 'clean';
3528
+ readonly DANGEROUS_IMPORTS_FOUND: 'dangerousImportsFound';
3529
+ readonly SKIPPED_SAFETENSORS: 'skippedSafetensors';
3530
+ readonly SKIPPED_GGUF: 'skippedGguf';
3531
+ readonly SKIPPED: 'skipped';
3532
+ readonly PARSE_ERROR: 'parseError';
3533
+ readonly ERROR: 'error';
3534
+ };
3535
+ export type ModelPickleScanStatus =
3536
+ (typeof ModelPickleScanStatus)[keyof typeof ModelPickleScanStatus];
3166
3537
  /**
3167
3538
  * ModelPickleScan
3168
3539
  */
@@ -3665,6 +4036,13 @@ export type PromptEnhancementInput = {
3665
4036
  * Optional instruction to guide how the prompt is enhanced (e.g., "expand to 77 tokens", "keep it under 20 words").
3666
4037
  */
3667
4038
  instruction?: null | string;
4039
+ /**
4040
+ * Optional reference images for the prompt enhancement model to consider when enhancing the prompt
4041
+ * (subject, style, lighting, composition, color palette). Accepts URLs, data URIs, raw base64, or AIR strings.
4042
+ * Requires the per-ecosystem prompt-analysis model to be a vision-capable LLM (configured via
4043
+ * IPromptAnalysisGrain.SetConfigurationAsync) — non-VLM models will silently ignore the images.
4044
+ */
4045
+ images?: null | Array<string>;
3668
4046
  };
3669
4047
  /**
3670
4048
  * Output from prompt enhancement analysis.
@@ -3882,6 +4260,19 @@ export type Qwen3TextToSpeechInput = Omit<VllmOmniTextToSpeechInput, 'engine' |
3882
4260
  ecosystem: 'qwen3';
3883
4261
  engine: 'vllm-omni';
3884
4262
  };
4263
+ export type Qwen3VoiceDesignTtsInput = Omit<
4264
+ Qwen3TextToSpeechInput,
4265
+ 'engine' | 'ecosystem' | 'operation'
4266
+ > & {
4267
+ /**
4268
+ * Natural-language description of the desired voice
4269
+ * (e.g., "a calm middle-aged male narrator with a slight British accent").
4270
+ */
4271
+ instruct: string;
4272
+ operation: 'voiceDesign';
4273
+ ecosystem: 'qwen3';
4274
+ engine: 'vllm-omni';
4275
+ };
3885
4276
  /**
3886
4277
  * AI Toolkit training for Qwen Image models
3887
4278
  */
@@ -4009,7 +4400,7 @@ export type ResourceInfo = {
4009
4400
  publishedAt?: null | string;
4010
4401
  /**
4011
4402
  * A boolean indicating whether this resource restricts to SFW content generation.
4012
- * NSFWContent covers X and AA whereas MatureContent includes R rated content.
4403
+ * NSFWContent covers X and XXX whereas MatureContent includes R rated content.
4013
4404
  */
4014
4405
  hasNSFWContentRestriction: boolean;
4015
4406
  };
@@ -4235,7 +4626,7 @@ export type SeedanceVideoGenInput = Omit<VideoGenInput, 'engine'> & {
4235
4626
  duration: 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15;
4236
4627
  generateAudio?: boolean;
4237
4628
  seed?: null | number;
4238
- resolution: '480p' | '720p';
4629
+ resolution: '480p' | '720p' | '1080p';
4239
4630
  images?: Array<string>;
4240
4631
  engine: 'seedance';
4241
4632
  };
@@ -6259,6 +6650,7 @@ export type ZImageBaseImageGenInput = Omit<
6259
6650
  };
6260
6651
  export type ZImageImageGenInput = Omit<SdCppImageGenInput, 'engine' | 'ecosystem'> & {
6261
6652
  model: string;
6653
+ diffuserModel?: string;
6262
6654
  ecosystem: 'zImage';
6263
6655
  engine: 'sdcpp';
6264
6656
  };
@@ -7423,6 +7815,36 @@ export type InvokeAgeClassificationStepTemplateResponses = {
7423
7815
  };
7424
7816
  export type InvokeAgeClassificationStepTemplateResponse =
7425
7817
  InvokeAgeClassificationStepTemplateResponses[keyof InvokeAgeClassificationStepTemplateResponses];
7818
+ export type InvokeAudioCaptioningStepTemplateData = {
7819
+ body?: AudioCaptioningInput;
7820
+ path?: never;
7821
+ query?: {
7822
+ experimental?: boolean;
7823
+ allowMatureContent?: boolean;
7824
+ whatif?: boolean;
7825
+ };
7826
+ url: '/v2/consumer/recipes/audioCaptioning';
7827
+ };
7828
+ export type InvokeAudioCaptioningStepTemplateErrors = {
7829
+ /**
7830
+ * Bad Request
7831
+ */
7832
+ 400: ProblemDetails;
7833
+ /**
7834
+ * Unauthorized
7835
+ */
7836
+ 401: ProblemDetails;
7837
+ };
7838
+ export type InvokeAudioCaptioningStepTemplateError =
7839
+ InvokeAudioCaptioningStepTemplateErrors[keyof InvokeAudioCaptioningStepTemplateErrors];
7840
+ export type InvokeAudioCaptioningStepTemplateResponses = {
7841
+ /**
7842
+ * OK
7843
+ */
7844
+ 200: AudioCaptioningOutput;
7845
+ };
7846
+ export type InvokeAudioCaptioningStepTemplateResponse =
7847
+ InvokeAudioCaptioningStepTemplateResponses[keyof InvokeAudioCaptioningStepTemplateResponses];
7426
7848
  export type InvokeBatchOcrSafetyClassificationStepTemplateData = {
7427
7849
  body?: BatchOcrSafetyClassificationInput;
7428
7850
  path?: never;
@@ -61,6 +61,7 @@ export const ComfySampler = {
61
61
  DDIM: 'ddim',
62
62
  UNI_PC: 'uni_pc',
63
63
  UNI_PC_BH2: 'uni_pc_bh2',
64
+ ER_SDE: 'er_sde',
64
65
  };
65
66
  export const ComfyScheduler = {
66
67
  NORMAL: 'normal',
@@ -232,6 +233,21 @@ export const Metric3dBackbone = {
232
233
  VIT_GIANT2: 'vit-giant2',
233
234
  };
234
235
  export const MiniMaxVideoGenModel = { HAILOU: 'hailou' };
236
+ export const ModelClamScanStatus = {
237
+ CLEAN: 'clean',
238
+ INFECTED: 'infected',
239
+ TIMEOUT: 'timeout',
240
+ ERROR: 'error',
241
+ };
242
+ export const ModelPickleScanStatus = {
243
+ CLEAN: 'clean',
244
+ DANGEROUS_IMPORTS_FOUND: 'dangerousImportsFound',
245
+ SKIPPED_SAFETENSORS: 'skippedSafetensors',
246
+ SKIPPED_GGUF: 'skippedGguf',
247
+ SKIPPED: 'skipped',
248
+ PARSE_ERROR: 'parseError',
249
+ ERROR: 'error',
250
+ };
235
251
  export const NsfwLevel = {
236
252
  PG: 'pg',
237
253
  PG13: 'pg13',
@@ -0,0 +1,3 @@
1
+ type Prettify<T> = {
2
+ [K in keyof T]: T[K];
3
+ } & NonNullable<unknown>;
@@ -0,0 +1 @@
1
+ "use strict";
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@civitai/client",
3
- "version": "0.2.0-beta.56",
3
+ "version": "0.2.0-beta.58",
4
4
  "description": "Civitai's javascript client for generating ai content",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
@@ -54,4 +54,4 @@
54
54
  "publishConfig": {
55
55
  "access": "public"
56
56
  }
57
- }
57
+ }