fal-endpoint-types 1.3.33 → 1.3.35

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,5 +1,60 @@
1
1
  import type * as Components from './components.js';
2
2
 
3
+ export interface XaiTtsV1Input {
4
+ /**
5
+ * Language
6
+ * @description BCP-47 language code or 'auto' for automatic detection. Supported: en, zh, fr, de, hi, id, it, ja, ko, pt-BR, pt-PT, ru, es-MX, es-ES, tr, vi, bn, ar-EG, ar-SA, ar-AE.
7
+ * @default auto
8
+ * @enum {string}
9
+ */
10
+ language?:
11
+ | 'auto'
12
+ | 'en'
13
+ | 'ar-EG'
14
+ | 'ar-SA'
15
+ | 'ar-AE'
16
+ | 'bn'
17
+ | 'zh'
18
+ | 'fr'
19
+ | 'de'
20
+ | 'hi'
21
+ | 'id'
22
+ | 'it'
23
+ | 'ja'
24
+ | 'ko'
25
+ | 'pt-BR'
26
+ | 'pt-PT'
27
+ | 'ru'
28
+ | 'es-MX'
29
+ | 'es-ES'
30
+ | 'tr'
31
+ | 'vi';
32
+ output_format?: Components.OutputFormat;
33
+ /**
34
+ * Text
35
+ * @description The text to convert to speech. Maximum 15,000 characters. Supports speech tags for expressive delivery: inline tags like [laugh], [pause], [sigh] and wrapping tags like <whisper>text</whisper>, <slow>text</slow>.
36
+ * @example Hello! This is xAI text to speech, brought to you by Fal AI.
37
+ */
38
+ text: string;
39
+ /**
40
+ * Voice
41
+ * @description Voice to use for synthesis. eve: energetic, upbeat. ara: warm, friendly. rex: confident, clear. sal: smooth, balanced. leo: authoritative, strong.
42
+ * @default eve
43
+ * @enum {string}
44
+ */
45
+ voice?: 'eve' | 'ara' | 'rex' | 'sal' | 'leo';
46
+ }
47
+
48
+ export interface XaiTtsV1Output {
49
+ /**
50
+ * @description The generated audio file.
51
+ * @example {
52
+ * "url": "https://v3b.fal.media/files/b/0a92750b/exZJCm6TDejS5xIulJs2r_xai_tts_output.mp3"
53
+ * }
54
+ */
55
+ audio: Components.File;
56
+ }
57
+
3
58
  export interface XaiGrokImagineVideoTextToVideoInput {
4
59
  /**
5
60
  * Aspect Ratio
@@ -39,8 +94,8 @@ export interface XaiGrokImagineVideoTextToVideoOutput {
39
94
  * "fps": 24,
40
95
  * "width": 1280,
41
96
  * "file_name": "RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
42
- * "content_type": "video/mp4",
43
- * "num_frames": 145
97
+ * "num_frames": 145,
98
+ * "content_type": "video/mp4"
44
99
  * }
45
100
  */
46
101
  video: Components.VideoFile;
@@ -90,8 +145,8 @@ export interface XaiGrokImagineVideoImageToVideoOutput {
90
145
  * "fps": 24,
91
146
  * "width": 1280,
92
147
  * "file_name": "0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
93
- * "content_type": "video/mp4",
94
- * "num_frames": 145
148
+ * "num_frames": 145,
149
+ * "content_type": "video/mp4"
95
150
  * }
96
151
  */
97
152
  video: Components.VideoFile;
@@ -129,8 +184,8 @@ export interface XaiGrokImagineVideoEditVideoOutput {
129
184
  * "fps": 24,
130
185
  * "width": 1280,
131
186
  * "file_name": "EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
132
- * "content_type": "video/mp4",
133
- * "num_frames": 121
187
+ * "num_frames": 121,
188
+ * "content_type": "video/mp4"
134
189
  * }
135
190
  */
136
191
  video: Components.VideoFile;
@@ -1527,38 +1582,6 @@ export interface SharedType_fe5 {
1527
1582
  };
1528
1583
  }
1529
1584
 
1530
- export interface SharedType_fda {
1531
- /**
1532
- * Masks
1533
- * @description Dictionary of label: mask image
1534
- * @example [
1535
- * {
1536
- * "height": 1200,
1537
- * "file_size": 15724,
1538
- * "file_name": "019c3c1e3c50446e9996f709d36debb4.png",
1539
- * "content_type": "image/png",
1540
- * "url": "https://v3.fal.media/files/monkey/6ITmhHQJ-69s-UxajrY5T_019c3c1e3c50446e9996f709d36debb4.png",
1541
- * "width": 1800
1542
- * },
1543
- * {
1544
- * "height": 1200,
1545
- * "file_size": 14905,
1546
- * "file_name": "0a1522ca410942c7ad6c73efa15b3549.png",
1547
- * "content_type": "image/png",
1548
- * "url": "https://v3.fal.media/files/monkey/IljtMxahoo9-7SUpx0fth_0a1522ca410942c7ad6c73efa15b3549.png",
1549
- * "width": 1800
1550
- * }
1551
- * ]
1552
- */
1553
- masks: Components.Image[];
1554
- /**
1555
- * Output
1556
- * @description Generated output
1557
- * @example <p> A white pickup truck </p> [SEG] is parked on the side of <p> the red building </p> [SEG] , creating a unique and eye-catching contrast.<|im_end|>
1558
- */
1559
- output: string;
1560
- }
1561
-
1562
1585
  export interface SharedType_fd1 {
1563
1586
  /**
1564
1587
  * @description The generated video
@@ -1886,6 +1909,102 @@ export interface SharedType_f67 {
1886
1909
  video_url: string;
1887
1910
  }
1888
1911
 
1912
+ export interface SharedType_f66 {
1913
+ /**
1914
+ * Acceleration
1915
+ * @description The acceleration level to use for image generation.
1916
+ * @default regular
1917
+ * @enum {string}
1918
+ */
1919
+ acceleration?: 'none' | 'regular' | 'high';
1920
+ /**
1921
+ * Enable Safety Checker
1922
+ * @description If set to true, the safety checker will be enabled.
1923
+ * @default true
1924
+ */
1925
+ enable_safety_checker?: boolean;
1926
+ /**
1927
+ * Guidance Scale
1928
+ * @description Guidance scale for classifier-free guidance.
1929
+ * @default 5
1930
+ */
1931
+ guidance_scale?: number;
1932
+ /**
1933
+ * Image Size
1934
+ * @description The size of the generated image. If not provided, uses the input image size.
1935
+ * @example {
1936
+ * "height": 1152,
1937
+ * "width": 2016
1938
+ * }
1939
+ */
1940
+ image_size?:
1941
+ | Components.ImageSize
1942
+ | (
1943
+ | 'square_hd'
1944
+ | 'square'
1945
+ | 'portrait_4_3'
1946
+ | 'portrait_16_9'
1947
+ | 'landscape_4_3'
1948
+ | 'landscape_16_9'
1949
+ );
1950
+ /**
1951
+ * Image URLs
1952
+ * @description The URLs of the images for editing. A maximum of 4 images are allowed.
1953
+ * @example [
1954
+ * "https://storage.googleapis.com/falserverless/example_inputs/flux2_dev_edit_input.png"
1955
+ * ]
1956
+ */
1957
+ image_urls: string[];
1958
+ /**
1959
+ * Loras
1960
+ * @description List of LoRA weights to apply (maximum 3).
1961
+ * @default []
1962
+ */
1963
+ loras?: Components.LoRAInput[];
1964
+ /**
1965
+ * Negative Prompt
1966
+ * @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
1967
+ * @default
1968
+ */
1969
+ negative_prompt?: string;
1970
+ /**
1971
+ * Number of Images
1972
+ * @description The number of images to generate.
1973
+ * @default 1
1974
+ */
1975
+ num_images?: number;
1976
+ /**
1977
+ * Number of Inference Steps
1978
+ * @description The number of inference steps to perform.
1979
+ * @default 28
1980
+ */
1981
+ num_inference_steps?: number;
1982
+ /**
1983
+ * Output Format
1984
+ * @description The format of the generated image.
1985
+ * @default png
1986
+ * @enum {string}
1987
+ */
1988
+ output_format?: 'jpeg' | 'png' | 'webp';
1989
+ /**
1990
+ * Prompt
1991
+ * @description The prompt to edit the image.
1992
+ * @example Change his clothes to casual suit and tie
1993
+ */
1994
+ prompt: string;
1995
+ /**
1996
+ * Seed
1997
+ * @description The seed to use for the generation. If not provided, a random seed will be used.
1998
+ */
1999
+ seed?: number;
2000
+ /**
2001
+ * Sync Mode
2002
+ * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
2003
+ * @default false
2004
+ */
2005
+ sync_mode?: boolean;
2006
+ }
2007
+
1889
2008
  export interface SharedType_f63 {
1890
2009
  /**
1891
2010
  * Acceleration
@@ -2505,91 +2624,6 @@ export interface SharedType_e43 {
2505
2624
  image: Components.Image_2;
2506
2625
  }
2507
2626
 
2508
- export interface SharedType_e34 {
2509
- /**
2510
- * Acceleration
2511
- * @description The acceleration level to use for image generation.
2512
- * @default regular
2513
- * @enum {string}
2514
- */
2515
- acceleration?: 'none' | 'regular' | 'high';
2516
- /**
2517
- * Enable Safety Checker
2518
- * @description If set to true, the safety checker will be enabled.
2519
- * @default true
2520
- */
2521
- enable_safety_checker?: boolean;
2522
- /**
2523
- * Guidance Scale
2524
- * @description Guidance scale for classifier-free guidance.
2525
- * @default 5
2526
- */
2527
- guidance_scale?: number;
2528
- /**
2529
- * Image Size
2530
- * @description The size of the image to generate.
2531
- * @default landscape_4_3
2532
- */
2533
- image_size?:
2534
- | Components.ImageSize
2535
- | (
2536
- | 'square_hd'
2537
- | 'square'
2538
- | 'portrait_4_3'
2539
- | 'portrait_16_9'
2540
- | 'landscape_4_3'
2541
- | 'landscape_16_9'
2542
- );
2543
- /**
2544
- * Loras
2545
- * @description List of LoRA weights to apply (maximum 3).
2546
- * @default []
2547
- */
2548
- loras?: Components.LoRAInput_1[];
2549
- /**
2550
- * Negative Prompt
2551
- * @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
2552
- * @default
2553
- */
2554
- negative_prompt?: string;
2555
- /**
2556
- * Number of Images
2557
- * @description The number of images to generate.
2558
- * @default 1
2559
- */
2560
- num_images?: number;
2561
- /**
2562
- * Number of Inference Steps
2563
- * @description The number of inference steps to perform.
2564
- * @default 28
2565
- */
2566
- num_inference_steps?: number;
2567
- /**
2568
- * Output Format
2569
- * @description The format of the generated image.
2570
- * @default png
2571
- * @enum {string}
2572
- */
2573
- output_format?: 'jpeg' | 'png' | 'webp';
2574
- /**
2575
- * Prompt
2576
- * @description The prompt to generate an image from.
2577
- * @example A serene Japanese garden with cherry blossoms, koi pond, and traditional wooden bridge at golden hour
2578
- */
2579
- prompt: string;
2580
- /**
2581
- * Seed
2582
- * @description The seed to use for the generation. If not provided, a random seed will be used.
2583
- */
2584
- seed?: number;
2585
- /**
2586
- * Sync Mode
2587
- * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
2588
- * @default false
2589
- */
2590
- sync_mode?: boolean;
2591
- }
2592
-
2593
2627
  export interface SharedType_e33 {
2594
2628
  /**
2595
2629
  * Has Nsfw Concepts
@@ -2625,29 +2659,6 @@ export interface SharedType_e33 {
2625
2659
  };
2626
2660
  }
2627
2661
 
2628
- export interface SharedType_e19 {
2629
- /**
2630
- * Images
2631
- * @description The generated images
2632
- * @example [
2633
- * {
2634
- * "height": 1536,
2635
- * "file_size": 3731290,
2636
- * "file_name": "257cf8e7bd3a47c2959396343d5b38cf.png",
2637
- * "content_type": "image/png",
2638
- * "url": "https://v3.fal.media/files/tiger/48e63e0K6C9XQYBuomoU-_257cf8e7bd3a47c2959396343d5b38cf.png",
2639
- * "width": 1536
2640
- * }
2641
- * ]
2642
- */
2643
- images: Components.Image[];
2644
- /**
2645
- * Seed
2646
- * @description Seed value used for generation.
2647
- */
2648
- seed: number;
2649
- }
2650
-
2651
2662
  export interface SharedType_e18 {
2652
2663
  /**
2653
2664
  * Default Caption
@@ -3234,24 +3245,36 @@ export interface SharedType_d94 {
3234
3245
  video: Components.File;
3235
3246
  }
3236
3247
 
3237
- export interface SharedType_d661 {
3248
+ export interface SharedType_d72 {
3238
3249
  /**
3239
- * Image Url
3240
- * @description The URL of the image to be processed.
3241
- * @example https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/car.jpg
3242
- * @example http://ecx.images-amazon.com/images/I/51UUzBDAMsL.jpg
3250
+ * Has Nsfw Concepts
3251
+ * @description Whether the generated images contain NSFW concepts.
3243
3252
  */
3244
- image_url: string;
3253
+ has_nsfw_concepts: boolean[];
3245
3254
  /**
3246
- * @description The user input coordinates
3247
- * @example {
3248
- * "x1": 100,
3249
- * "x2": 200,
3250
- * "y1": 100,
3251
- * "y2": 200
3252
- * }
3255
+ * Images
3256
+ * @description The generated images
3257
+ * @example [
3258
+ * {
3259
+ * "url": "https://v3b.fal.media/files/b/0a8a69c7/gU9ZgfFC9oAZjpIoveAac.png"
3260
+ * }
3261
+ * ]
3253
3262
  */
3254
- region: Components.Region;
3263
+ images: Components.ImageFile[];
3264
+ /**
3265
+ * Prompt
3266
+ * @description The prompt used for generating the image.
3267
+ */
3268
+ prompt: string;
3269
+ /**
3270
+ * Seed
3271
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
3272
+ * input or the randomly generated that was used in case none was passed.
3273
+ */
3274
+ seed: number;
3275
+ timings: {
3276
+ [key: string]: number;
3277
+ };
3255
3278
  }
3256
3279
 
3257
3280
  export interface SharedType_d66 {
@@ -3494,20 +3517,14 @@ export interface SharedType_d35 {
3494
3517
  seed?: number;
3495
3518
  }
3496
3519
 
3497
- export interface SharedType_d26 {
3498
- /**
3499
- * Seed
3500
- * @description The seed used for generation.
3501
- */
3502
- seed: number;
3520
+ export interface SharedType_d2a {
3503
3521
  /**
3504
- * Video
3505
- * @description The generated video file.
3522
+ * @description The generated speech audio
3506
3523
  * @example {
3507
- * "url": "https://storage.googleapis.com/falserverless/vace/out_video_vace.mp4"
3524
+ * "url": "https://v3.fal.media/files/kangaroo/RQ_pxc7oPdueYqWUqEbPE_tmpjnzvvzx_.wav"
3508
3525
  * }
3509
3526
  */
3510
- video: Components.File_1;
3527
+ audio: Components.File;
3511
3528
  }
3512
3529
 
3513
3530
  export interface SharedType_d1a {
@@ -3863,6 +3880,35 @@ export interface SharedType_cbd {
3863
3880
  sync_mode?: boolean;
3864
3881
  }
3865
3882
 
3883
+ export interface SharedType_cb6 {
3884
+ /**
3885
+ * Prompt
3886
+ * @description The prompt used for the generation.
3887
+ * @example A woman stands still amid a busy neon-lit street at night. The camera slowly dollies in toward her face as people blur past, their motion emphasizing her calm presence. City lights flicker and reflections shift across her denim jacket.
3888
+ */
3889
+ prompt: string;
3890
+ /**
3891
+ * Seed
3892
+ * @description The seed used for the random number generator.
3893
+ * @example 2078003885
3894
+ */
3895
+ seed: number;
3896
+ /**
3897
+ * @description The generated video.
3898
+ * @example {
3899
+ * "height": 704,
3900
+ * "duration": 6.44,
3901
+ * "url": "https://v3b.fal.media/files/b/0a894013/N9lnMTq7W3uMC0lOQg845_BknRPV8I.mp4",
3902
+ * "fps": 25,
3903
+ * "width": 1248,
3904
+ * "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
3905
+ * "num_frames": 161,
3906
+ * "content_type": "video/mp4"
3907
+ * }
3908
+ */
3909
+ video: Components.VideoFile;
3910
+ }
3911
+
3866
3912
  export interface SharedType_cb4 {
3867
3913
  /**
3868
3914
  * Acceleration
@@ -4466,6 +4512,91 @@ export interface SharedType_bb3 {
4466
4512
  resolution?: '540p' | '720p' | '1080p';
4467
4513
  }
4468
4514
 
4515
+ export interface SharedType_bb2 {
4516
+ /**
4517
+ * Acceleration
4518
+ * @description The acceleration level to use for image generation.
4519
+ * @default regular
4520
+ * @enum {string}
4521
+ */
4522
+ acceleration?: 'none' | 'regular' | 'high';
4523
+ /**
4524
+ * Enable Safety Checker
4525
+ * @description If set to true, the safety checker will be enabled.
4526
+ * @default true
4527
+ */
4528
+ enable_safety_checker?: boolean;
4529
+ /**
4530
+ * Guidance Scale
4531
+ * @description Guidance scale for classifier-free guidance.
4532
+ * @default 5
4533
+ */
4534
+ guidance_scale?: number;
4535
+ /**
4536
+ * Image Size
4537
+ * @description The size of the image to generate.
4538
+ * @default landscape_4_3
4539
+ */
4540
+ image_size?:
4541
+ | Components.ImageSize
4542
+ | (
4543
+ | 'square_hd'
4544
+ | 'square'
4545
+ | 'portrait_4_3'
4546
+ | 'portrait_16_9'
4547
+ | 'landscape_4_3'
4548
+ | 'landscape_16_9'
4549
+ );
4550
+ /**
4551
+ * Loras
4552
+ * @description List of LoRA weights to apply (maximum 3).
4553
+ * @default []
4554
+ */
4555
+ loras?: Components.LoRAInput[];
4556
+ /**
4557
+ * Negative Prompt
4558
+ * @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
4559
+ * @default
4560
+ */
4561
+ negative_prompt?: string;
4562
+ /**
4563
+ * Number of Images
4564
+ * @description The number of images to generate.
4565
+ * @default 1
4566
+ */
4567
+ num_images?: number;
4568
+ /**
4569
+ * Number of Inference Steps
4570
+ * @description The number of inference steps to perform.
4571
+ * @default 28
4572
+ */
4573
+ num_inference_steps?: number;
4574
+ /**
4575
+ * Output Format
4576
+ * @description The format of the generated image.
4577
+ * @default png
4578
+ * @enum {string}
4579
+ */
4580
+ output_format?: 'jpeg' | 'png' | 'webp';
4581
+ /**
4582
+ * Prompt
4583
+ * @description The prompt to generate an image from.
4584
+ * @example A serene Japanese garden with cherry blossoms, koi pond, and traditional wooden bridge at golden hour
4585
+ */
4586
+ prompt: string;
4587
+ /**
4588
+ * Seed
4589
+ * @description The seed to use for the generation. If not provided, a random seed will be used.
4590
+ */
4591
+ seed?: number;
4592
+ /**
4593
+ * Sync Mode
4594
+ * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
4595
+ * @default false
4596
+ */
4597
+ sync_mode?: boolean;
4598
+ }
4599
+
4469
4600
  export interface SharedType_bb0 {
4470
4601
  /**
4471
4602
  * Image URL
@@ -4612,76 +4743,6 @@ export interface SharedType_b89 {
4612
4743
  image: Components.Image;
4613
4744
  }
4614
4745
 
4615
- export interface SharedType_b7d {
4616
- /**
4617
- * Enable Safety Checker
4618
- * @description If set to true, the safety checker will be enabled.
4619
- * @default true
4620
- */
4621
- enable_safety_checker?: boolean;
4622
- /**
4623
- * Guidance scale (CFG)
4624
- * @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
4625
- * the model to stick to your prompt when looking for a related image to show you.
4626
- * @default 3.5
4627
- */
4628
- guidance_scale?: number;
4629
- /**
4630
- * Image Size
4631
- * @description The size of the generated image.
4632
- * @default landscape_4_3
4633
- */
4634
- image_size?:
4635
- | Components.ImageSize
4636
- | (
4637
- | 'square_hd'
4638
- | 'square'
4639
- | 'portrait_4_3'
4640
- | 'portrait_16_9'
4641
- | 'landscape_4_3'
4642
- | 'landscape_16_9'
4643
- );
4644
- /**
4645
- * Negative prompt
4646
- * @description Negative Prompt for generation.
4647
- * @default
4648
- * @example Blurry, out of focus, low resolution, bad anatomy, ugly, deformed, poorly drawn, extra limbs
4649
- */
4650
- negative_prompt?: string;
4651
- /**
4652
- * Num Images
4653
- * @description The number of images to generate.
4654
- * @default 1
4655
- */
4656
- num_images?: number;
4657
- /**
4658
- * Num Inference Steps
4659
- * @description The number of inference steps to perform.
4660
- * @default 28
4661
- */
4662
- num_inference_steps?: number;
4663
- /**
4664
- * Prompt
4665
- * @description The prompt to generate an image from.
4666
- * @example Mount Fuji at sunset, with the iconic snow-capped peak silhouetted against a vibrant orange and purple sky. A tranquil lake in the foreground perfectly reflects the mountain and colorful sky. A few traditional Japanese cherry blossom trees frame the scene, with their delicate pink petals visible in the foreground.
4667
- */
4668
- prompt: string;
4669
- /**
4670
- * Seed
4671
- * @description The same seed and the same prompt given to the same version of the model
4672
- * will output the same image every time.
4673
- */
4674
- seed?: number;
4675
- /**
4676
- * Sync Mode
4677
- * @description If set to true, the function will wait for the image to be generated and uploaded
4678
- * before returning the response. This will increase the latency of the function but
4679
- * it allows you to get the image directly in the response without going through the CDN.
4680
- * @default false
4681
- */
4682
- sync_mode?: boolean;
4683
- }
4684
-
4685
4746
  export interface SharedType_b71 {
4686
4747
  model_mesh: Components.File;
4687
4748
  /**
@@ -4831,6 +4892,16 @@ export interface SharedType_b35 {
4831
4892
  sync_mode?: boolean;
4832
4893
  }
4833
4894
 
4895
+ export interface SharedType_b23 {
4896
+ /**
4897
+ * Seed
4898
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
4899
+ * input or the randomly generated that was used in case none was passed.
4900
+ */
4901
+ seed: number;
4902
+ video: Components.File;
4903
+ }
4904
+
4834
4905
  export interface SharedType_b08 {
4835
4906
  /**
4836
4907
  * Acceleration
@@ -5159,6 +5230,29 @@ export interface SharedType_a9b {
5159
5230
  seed?: number;
5160
5231
  }
5161
5232
 
5233
+ export interface SharedType_a97 {
5234
+ /**
5235
+ * Images
5236
+ * @description The generated images
5237
+ * @example [
5238
+ * {
5239
+ * "file_size": 3731290,
5240
+ * "height": 1536,
5241
+ * "file_name": "257cf8e7bd3a47c2959396343d5b38cf.png",
5242
+ * "content_type": "image/png",
5243
+ * "url": "https://v3.fal.media/files/tiger/48e63e0K6C9XQYBuomoU-_257cf8e7bd3a47c2959396343d5b38cf.png",
5244
+ * "width": 1536
5245
+ * }
5246
+ * ]
5247
+ */
5248
+ images: Components.Image[];
5249
+ /**
5250
+ * Seed
5251
+ * @description Seed value used for generation.
5252
+ */
5253
+ seed: number;
5254
+ }
5255
+
5162
5256
  export interface SharedType_a8f {
5163
5257
  /**
5164
5258
  * Image Url
@@ -5224,6 +5318,38 @@ export interface SharedType_a77 {
5224
5318
  audio_url: string;
5225
5319
  }
5226
5320
 
5321
+ export interface SharedType_a75 {
5322
+ /**
5323
+ * Has Nsfw Concepts
5324
+ * @description Whether the generated images contain NSFW concepts.
5325
+ */
5326
+ has_nsfw_concepts: boolean[];
5327
+ /**
5328
+ * Images
5329
+ * @description The generated images
5330
+ * @example [
5331
+ * {
5332
+ * "url": "https://v3b.fal.media/files/b/0a8a6905/DuNRy1OODaGrlUGEfl9SX.png"
5333
+ * }
5334
+ * ]
5335
+ */
5336
+ images: Components.ImageFile[];
5337
+ /**
5338
+ * Prompt
5339
+ * @description The prompt used for generating the image.
5340
+ */
5341
+ prompt: string;
5342
+ /**
5343
+ * Seed
5344
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
5345
+ * input or the randomly generated that was used in case none was passed.
5346
+ */
5347
+ seed: number;
5348
+ timings: {
5349
+ [key: string]: number;
5350
+ };
5351
+ }
5352
+
5227
5353
  export interface SharedType_a58 {
5228
5354
  /**
5229
5355
  * Duration
@@ -5427,35 +5553,6 @@ export interface SharedType_a1f {
5427
5553
  };
5428
5554
  }
5429
5555
 
5430
- export interface SharedType_a00 {
5431
- /**
5432
- * Prompt
5433
- * @description The prompt used for the generation.
5434
- * @example Continue the scene naturally, maintaining the same style and motion.
5435
- */
5436
- prompt: string;
5437
- /**
5438
- * Seed
5439
- * @description The seed used for the random number generator.
5440
- * @example 866232447
5441
- */
5442
- seed: number;
5443
- /**
5444
- * @description The generated video.
5445
- * @example {
5446
- * "height": 704,
5447
- * "duration": 10.28,
5448
- * "url": "https://v3b.fal.media/files/b/0a88289e/CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
5449
- * "fps": 25,
5450
- * "width": 1248,
5451
- * "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
5452
- * "content_type": "video/mp4",
5453
- * "num_frames": 257
5454
- * }
5455
- */
5456
- video: Components.VideoFile;
5457
- }
5458
-
5459
5556
  export interface SharedType_9f4 {
5460
5557
  /**
5461
5558
  * Audio Url
@@ -5858,12 +5955,81 @@ export interface SharedType_97e {
5858
5955
  voice_setting?: Components.VoiceSetting;
5859
5956
  }
5860
5957
 
5861
- export interface SharedType_973 {
5958
+ export interface SharedType_974 {
5862
5959
  /**
5863
- * Image
5864
- * @description The depth map.
5960
+ * Enable Safety Checker
5961
+ * @description If set to true, the safety checker will be enabled.
5962
+ * @default true
5865
5963
  */
5866
- image: Components.Image_2;
5964
+ enable_safety_checker?: boolean;
5965
+ /**
5966
+ * Image Size
5967
+ * @description The size of the generated image. If not provided, uses the input image size.
5968
+ * @example {
5969
+ * "height": 1152,
5970
+ * "width": 2016
5971
+ * }
5972
+ */
5973
+ image_size?:
5974
+ | Components.ImageSize
5975
+ | (
5976
+ | 'square_hd'
5977
+ | 'square'
5978
+ | 'portrait_4_3'
5979
+ | 'portrait_16_9'
5980
+ | 'landscape_4_3'
5981
+ | 'landscape_16_9'
5982
+ );
5983
+ /**
5984
+ * Image URLs
5985
+ * @description The URLs of the images for editing. A maximum of 4 images are allowed.
5986
+ * @example [
5987
+ * "https://v3b.fal.media/files/b/0a8a69d5/kkXxFfj1QeVtw35kxy5Py_1a7e3511-bd2c-46be-923a-8e6be2496f12.png"
5988
+ * ]
5989
+ */
5990
+ image_urls: string[];
5991
+ /**
5992
+ * Loras
5993
+ * @description List of LoRA weights to apply (maximum 3).
5994
+ * @default []
5995
+ */
5996
+ loras?: Components.LoRAInput[];
5997
+ /**
5998
+ * Number of Images
5999
+ * @description The number of images to generate.
6000
+ * @default 1
6001
+ */
6002
+ num_images?: number;
6003
+ /**
6004
+ * Number of Inference Steps
6005
+ * @description The number of inference steps to perform.
6006
+ * @default 4
6007
+ */
6008
+ num_inference_steps?: number;
6009
+ /**
6010
+ * Output Format
6011
+ * @description The format of the generated image.
6012
+ * @default png
6013
+ * @enum {string}
6014
+ */
6015
+ output_format?: 'jpeg' | 'png' | 'webp';
6016
+ /**
6017
+ * Prompt
6018
+ * @description The prompt to edit the image.
6019
+ * @example Turn this into a realistic image
6020
+ */
6021
+ prompt: string;
6022
+ /**
6023
+ * Seed
6024
+ * @description The seed to use for the generation. If not provided, a random seed will be used.
6025
+ */
6026
+ seed?: number;
6027
+ /**
6028
+ * Sync Mode
6029
+ * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
6030
+ * @default false
6031
+ */
6032
+ sync_mode?: boolean;
5867
6033
  }
5868
6034
 
5869
6035
  export interface SharedType_95d {
@@ -5979,6 +6145,38 @@ export interface SharedType_94e {
5979
6145
  video: Components.File;
5980
6146
  }
5981
6147
 
6148
+ export interface SharedType_938 {
6149
+ /**
6150
+ * Has Nsfw Concepts
6151
+ * @description Whether the generated images contain NSFW concepts.
6152
+ */
6153
+ has_nsfw_concepts: boolean[];
6154
+ /**
6155
+ * Images
6156
+ * @description The edited images
6157
+ * @example [
6158
+ * {
6159
+ * "url": "https://v3b.fal.media/files/b/0a8a69d6/M73KvDgfEgIM77t4mFsS2.png"
6160
+ * }
6161
+ * ]
6162
+ */
6163
+ images: Components.ImageFile[];
6164
+ /**
6165
+ * Prompt
6166
+ * @description The prompt used for generating the image.
6167
+ */
6168
+ prompt: string;
6169
+ /**
6170
+ * Seed
6171
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
6172
+ * input or the randomly generated that was used in case none was passed.
6173
+ */
6174
+ seed: number;
6175
+ timings: {
6176
+ [key: string]: number;
6177
+ };
6178
+ }
6179
+
5982
6180
  export interface SharedType_92f {
5983
6181
  /**
5984
6182
  * Has Nsfw Concepts
@@ -6825,6 +7023,35 @@ export interface SharedType_85d {
6825
7023
  sync_mode?: boolean;
6826
7024
  }
6827
7025
 
7026
+ export interface SharedType_857 {
7027
+ /**
7028
+ * Prompt
7029
+ * @description The prompt used for the generation.
7030
+ * @example A cowboy walking through a dusty town at high noon, camera following from behind, cinematic depth, realistic lighting, western mood, 4K film grain.
7031
+ */
7032
+ prompt: string;
7033
+ /**
7034
+ * Seed
7035
+ * @description The seed used for the random number generator.
7036
+ * @example 149063119
7037
+ */
7038
+ seed: number;
7039
+ /**
7040
+ * @description The generated video.
7041
+ * @example {
7042
+ * "height": 704,
7043
+ * "duration": 6.44,
7044
+ * "url": "https://v3b.fal.media/files/b/0a8824b1/sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
7045
+ * "fps": 25,
7046
+ * "width": 1248,
7047
+ * "file_name": "sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
7048
+ * "num_frames": 161,
7049
+ * "content_type": "video/mp4"
7050
+ * }
7051
+ */
7052
+ video: Components.VideoFile;
7053
+ }
7054
+
6828
7055
  export interface SharedType_844 {
6829
7056
  /**
6830
7057
  * Prompt
@@ -7780,6 +8007,10 @@ export interface SharedType_7c6 {
7780
8007
  };
7781
8008
  }
7782
8009
 
8010
+ export interface SharedType_7c2 {
8011
+ lottie_file: Components.File;
8012
+ }
8013
+
7783
8014
  export interface SharedType_7c0 {
7784
8015
  /**
7785
8016
  * @description Generated 3D object file.
@@ -8124,102 +8355,6 @@ export interface SharedType_709 {
8124
8355
  video_url: string;
8125
8356
  }
8126
8357
 
8127
- export interface SharedType_6f9 {
8128
- /**
8129
- * Acceleration
8130
- * @description The acceleration level to use for image generation.
8131
- * @default regular
8132
- * @enum {string}
8133
- */
8134
- acceleration?: 'none' | 'regular' | 'high';
8135
- /**
8136
- * Enable Safety Checker
8137
- * @description If set to true, the safety checker will be enabled.
8138
- * @default true
8139
- */
8140
- enable_safety_checker?: boolean;
8141
- /**
8142
- * Guidance Scale
8143
- * @description Guidance scale for classifier-free guidance.
8144
- * @default 5
8145
- */
8146
- guidance_scale?: number;
8147
- /**
8148
- * Image Size
8149
- * @description The size of the generated image. If not provided, uses the input image size.
8150
- * @example {
8151
- * "height": 1152,
8152
- * "width": 2016
8153
- * }
8154
- */
8155
- image_size?:
8156
- | Components.ImageSize
8157
- | (
8158
- | 'square_hd'
8159
- | 'square'
8160
- | 'portrait_4_3'
8161
- | 'portrait_16_9'
8162
- | 'landscape_4_3'
8163
- | 'landscape_16_9'
8164
- );
8165
- /**
8166
- * Image URLs
8167
- * @description The URLs of the images for editing. A maximum of 4 images are allowed.
8168
- * @example [
8169
- * "https://storage.googleapis.com/falserverless/example_inputs/flux2_dev_edit_input.png"
8170
- * ]
8171
- */
8172
- image_urls: string[];
8173
- /**
8174
- * Loras
8175
- * @description List of LoRA weights to apply (maximum 3).
8176
- * @default []
8177
- */
8178
- loras?: Components.LoRAInput_1[];
8179
- /**
8180
- * Negative Prompt
8181
- * @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
8182
- * @default
8183
- */
8184
- negative_prompt?: string;
8185
- /**
8186
- * Number of Images
8187
- * @description The number of images to generate.
8188
- * @default 1
8189
- */
8190
- num_images?: number;
8191
- /**
8192
- * Number of Inference Steps
8193
- * @description The number of inference steps to perform.
8194
- * @default 28
8195
- */
8196
- num_inference_steps?: number;
8197
- /**
8198
- * Output Format
8199
- * @description The format of the generated image.
8200
- * @default png
8201
- * @enum {string}
8202
- */
8203
- output_format?: 'jpeg' | 'png' | 'webp';
8204
- /**
8205
- * Prompt
8206
- * @description The prompt to edit the image.
8207
- * @example Change his clothes to casual suit and tie
8208
- */
8209
- prompt: string;
8210
- /**
8211
- * Seed
8212
- * @description The seed to use for the generation. If not provided, a random seed will be used.
8213
- */
8214
- seed?: number;
8215
- /**
8216
- * Sync Mode
8217
- * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
8218
- * @default false
8219
- */
8220
- sync_mode?: boolean;
8221
- }
8222
-
8223
8358
  export interface SharedType_6cd {
8224
8359
  /**
8225
8360
  * Aspect Ratio
@@ -8821,35 +8956,6 @@ export interface SharedType_62e {
8821
8956
  video: Components.VideoFile;
8822
8957
  }
8823
8958
 
8824
- export interface SharedType_61b {
8825
- /**
8826
- * Prompt
8827
- * @description The prompt used for the generation.
8828
- * @example A cowboy walking through a dusty town at high noon, camera following from behind, cinematic depth, realistic lighting, western mood, 4K film grain.
8829
- */
8830
- prompt: string;
8831
- /**
8832
- * Seed
8833
- * @description The seed used for the random number generator.
8834
- * @example 149063119
8835
- */
8836
- seed: number;
8837
- /**
8838
- * @description The generated video.
8839
- * @example {
8840
- * "height": 704,
8841
- * "duration": 6.44,
8842
- * "url": "https://v3b.fal.media/files/b/0a8824b1/sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
8843
- * "fps": 25,
8844
- * "width": 1248,
8845
- * "file_name": "sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
8846
- * "content_type": "video/mp4",
8847
- * "num_frames": 161
8848
- * }
8849
- */
8850
- video: Components.VideoFile;
8851
- }
8852
-
8853
8959
  export interface SharedType_618 {
8854
8960
  /**
8855
8961
  * Video
@@ -9264,6 +9370,24 @@ export interface SharedType_595 {
9264
9370
  video: Components.File_1;
9265
9371
  }
9266
9372
 
9373
+ export interface SharedType_58d {
9374
+ /**
9375
+ * Seed
9376
+ * @description The seed used for generation.
9377
+ */
9378
+ seed: number;
9379
+ /**
9380
+ * @description The generated video file.
9381
+ * @example {
9382
+ * "file_size": 797478,
9383
+ * "file_name": "6c9dd31e1d9a4482877747a52a661a0a.mp4",
9384
+ * "content_type": "application/octet-stream",
9385
+ * "url": "https://v3.fal.media/files/elephant/-huMN0zTaXmBr2CqzCMps_6c9dd31e1d9a4482877747a52a661a0a.mp4"
9386
+ * }
9387
+ */
9388
+ video: Components.File;
9389
+ }
9390
+
9267
9391
  export interface SharedType_57e {
9268
9392
  /**
9269
9393
  * Acceleration
@@ -9528,35 +9652,6 @@ export interface SharedType_576 {
9528
9652
  };
9529
9653
  }
9530
9654
 
9531
- export interface SharedType_567 {
9532
- /**
9533
- * Prompt
9534
- * @description The prompt used for the generation.
9535
- * @example A woman stands still amid a busy neon-lit street at night. The camera slowly dollies in toward her face as people blur past, their motion emphasizing her calm presence. City lights flicker and reflections shift across her denim jacket.
9536
- */
9537
- prompt: string;
9538
- /**
9539
- * Seed
9540
- * @description The seed used for the random number generator.
9541
- * @example 2078003885
9542
- */
9543
- seed: number;
9544
- /**
9545
- * @description The generated video.
9546
- * @example {
9547
- * "height": 704,
9548
- * "duration": 6.44,
9549
- * "url": "https://v3b.fal.media/files/b/0a894013/N9lnMTq7W3uMC0lOQg845_BknRPV8I.mp4",
9550
- * "fps": 25,
9551
- * "width": 1248,
9552
- * "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
9553
- * "content_type": "video/mp4",
9554
- * "num_frames": 161
9555
- * }
9556
- */
9557
- video: Components.VideoFile;
9558
- }
9559
-
9560
9655
  export interface SharedType_554 {
9561
9656
  boundingbox_frames_zip?: Components.File;
9562
9657
  /**
@@ -10031,35 +10126,6 @@ export interface SharedType_503 {
10031
10126
  seed?: number;
10032
10127
  }
10033
10128
 
10034
- export interface SharedType_4ed {
10035
- /**
10036
- * Prompt
10037
- * @description The prompt used for the generation.
10038
- * @example black-and-white video, a cowboy walks through a dusty town, film grain
10039
- */
10040
- prompt: string;
10041
- /**
10042
- * Seed
10043
- * @description The seed used for the random number generator.
10044
- * @example 1490631192028410600
10045
- */
10046
- seed: number;
10047
- /**
10048
- * @description The generated video.
10049
- * @example {
10050
- * "height": 704,
10051
- * "duration": 6.44,
10052
- * "url": "https://v3b.fal.media/files/b/0a895ed5/SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
10053
- * "fps": 25,
10054
- * "width": 1248,
10055
- * "file_name": "SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
10056
- * "content_type": "video/mp4",
10057
- * "num_frames": 161
10058
- * }
10059
- */
10060
- video: Components.VideoFile;
10061
- }
10062
-
10063
10129
  export interface SharedType_4d6 {
10064
10130
  /**
10065
10131
  * Has Nsfw Concepts
@@ -10523,20 +10589,6 @@ export interface SharedType_49b {
10523
10589
  sync_mode?: boolean;
10524
10590
  }
10525
10591
 
10526
- export interface SharedType_484 {
10527
- /**
10528
- * Seed
10529
- * @description Seed of the generated Image. It will be the same value of the one passed in the
10530
- * input or the randomly generated that was used in case none was passed.
10531
- */
10532
- seed: number;
10533
- /**
10534
- * Video
10535
- * @description The generated video file.
10536
- */
10537
- video: Components.File_1;
10538
- }
10539
-
10540
10592
  export interface SharedType_480 {
10541
10593
  /**
10542
10594
  * @description The generated video from image using the Q2 model
@@ -11098,6 +11150,35 @@ export interface SharedType_3be {
11098
11150
  video: Components.File;
11099
11151
  }
11100
11152
 
11153
+ export interface SharedType_3b0 {
11154
+ /**
11155
+ * Prompt
11156
+ * @description The prompt used for the generation.
11157
+ * @example black-and-white video, a cowboy walks through a dusty town, film grain
11158
+ */
11159
+ prompt: string;
11160
+ /**
11161
+ * Seed
11162
+ * @description The seed used for the random number generator.
11163
+ * @example 1490631192028410600
11164
+ */
11165
+ seed: number;
11166
+ /**
11167
+ * @description The generated video.
11168
+ * @example {
11169
+ * "height": 704,
11170
+ * "duration": 6.44,
11171
+ * "url": "https://v3b.fal.media/files/b/0a895ed5/SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
11172
+ * "fps": 25,
11173
+ * "width": 1248,
11174
+ * "file_name": "SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
11175
+ * "num_frames": 161,
11176
+ * "content_type": "video/mp4"
11177
+ * }
11178
+ */
11179
+ video: Components.VideoFile;
11180
+ }
11181
+
11101
11182
  export interface SharedType_397 {
11102
11183
  /**
11103
11184
  * Acceleration
@@ -11309,6 +11390,44 @@ export interface SharedType_397 {
11309
11390
  video_write_mode?: 'fast' | 'balanced' | 'small';
11310
11391
  }
11311
11392
 
11393
+ export interface SharedType_390 {
11394
+ /**
11395
+ * Has Nsfw Concepts
11396
+ * @description Whether the generated images contain NSFW concepts.
11397
+ */
11398
+ has_nsfw_concepts: boolean[];
11399
+ /**
11400
+ * Images
11401
+ * @description The generated image files info.
11402
+ * @example [
11403
+ * {
11404
+ * "height": 2048,
11405
+ * "content_type": "image/png",
11406
+ * "url": "https://v3b.fal.media/files/b/0a9276c3/6KxxMb0bPmfNLH3DqcuSf_qRUe8cIL.png",
11407
+ * "width": 2048
11408
+ * }
11409
+ * ]
11410
+ */
11411
+ images: Components.ImageFile[];
11412
+ /**
11413
+ * Prompt
11414
+ * @description The prompt used for generating the image.
11415
+ */
11416
+ prompt: string;
11417
+ /**
11418
+ * Seed
11419
+ * @description Seed of the generated Image. It will be the same value of the one passed in the input or the randomly generated that was used in case none was passed.
11420
+ */
11421
+ seed: number;
11422
+ /**
11423
+ * Timings
11424
+ * @description The timings of the generation process.
11425
+ */
11426
+ timings: {
11427
+ [key: string]: number;
11428
+ };
11429
+ }
11430
+
11312
11431
  export interface SharedType_38d {
11313
11432
  /**
11314
11433
  * Image Url
@@ -11380,6 +11499,35 @@ export interface SharedType_371 {
11380
11499
  prompt_optimizer?: boolean;
11381
11500
  }
11382
11501
 
11502
+ export interface SharedType_369 {
11503
+ /**
11504
+ * Prompt
11505
+ * @description The prompt used for the generation.
11506
+ * @example Continue the scene naturally, maintaining the same style and motion.
11507
+ */
11508
+ prompt: string;
11509
+ /**
11510
+ * Seed
11511
+ * @description The seed used for the random number generator.
11512
+ * @example 866232447
11513
+ */
11514
+ seed: number;
11515
+ /**
11516
+ * @description The generated video.
11517
+ * @example {
11518
+ * "height": 704,
11519
+ * "duration": 10.28,
11520
+ * "url": "https://v3b.fal.media/files/b/0a88289e/CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
11521
+ * "fps": 25,
11522
+ * "width": 1248,
11523
+ * "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
11524
+ * "num_frames": 257,
11525
+ * "content_type": "video/mp4"
11526
+ * }
11527
+ */
11528
+ video: Components.VideoFile;
11529
+ }
11530
+
11383
11531
  export interface SharedType_367 {
11384
11532
  /**
11385
11533
  * Aspect Ratio
@@ -11589,6 +11737,26 @@ export interface SharedType_304 {
11589
11737
  voice: Components.TextVoice;
11590
11738
  }
11591
11739
 
11740
+ export interface SharedType_2ff {
11741
+ /**
11742
+ * Image Url
11743
+ * @description The URL of the image to be processed.
11744
+ * @example https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/car.jpg
11745
+ * @example http://ecx.images-amazon.com/images/I/51UUzBDAMsL.jpg
11746
+ */
11747
+ image_url: string;
11748
+ /**
11749
+ * @description The user input coordinates
11750
+ * @example {
11751
+ * "y2": 200,
11752
+ * "x2": 200,
11753
+ * "x1": 100,
11754
+ * "y1": 100
11755
+ * }
11756
+ */
11757
+ region: Components.Region;
11758
+ }
11759
+
11592
11760
  export interface SharedType_2e6 {
11593
11761
  /**
11594
11762
  * Aspect Ratio
@@ -12184,6 +12352,74 @@ export interface SharedType_21d {
12184
12352
  turn_off_cfg_start_si?: number;
12185
12353
  }
12186
12354
 
12355
+ export interface SharedType_219 {
12356
+ /**
12357
+ * Enable Safety Checker
12358
+ * @description If set to true, the safety checker will be enabled.
12359
+ * @default true
12360
+ */
12361
+ enable_safety_checker?: boolean;
12362
+ /**
12363
+ * Guidance scale (CFG)
12364
+ * @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
12365
+ * the model to stick to your prompt when looking for a related image to show you.
12366
+ * @default 3.5
12367
+ */
12368
+ guidance_scale?: number;
12369
+ /**
12370
+ * Image Size
12371
+ * @description The size of the generated image.
12372
+ * @default landscape_4_3
12373
+ */
12374
+ image_size?:
12375
+ | Components.ImageSize
12376
+ | (
12377
+ | 'square_hd'
12378
+ | 'square'
12379
+ | 'portrait_4_3'
12380
+ | 'portrait_16_9'
12381
+ | 'landscape_4_3'
12382
+ | 'landscape_16_9'
12383
+ );
12384
+ /**
12385
+ * Negative prompt
12386
+ * @description Negative Prompt for generation.
12387
+ * @default
12388
+ * @example Blurry, out of focus, low resolution, bad anatomy, ugly, deformed, poorly drawn, extra limbs
12389
+ */
12390
+ negative_prompt?: string;
12391
+ /**
12392
+ * Num Images
12393
+ * @description The number of images to generate.
12394
+ * @default 1
12395
+ */
12396
+ num_images?: number;
12397
+ /**
12398
+ * Num Inference Steps
12399
+ * @description The number of inference steps to perform.
12400
+ * @default 28
12401
+ */
12402
+ num_inference_steps?: number;
12403
+ /**
12404
+ * Prompt
12405
+ * @description The prompt to generate an image from.
12406
+ * @example Mount Fuji at sunset, with the iconic snow-capped peak silhouetted against a vibrant orange and purple sky. A tranquil lake in the foreground perfectly reflects the mountain and colorful sky. A few traditional Japanese cherry blossom trees frame the scene, with their delicate pink petals visible in the foreground.
12407
+ */
12408
+ prompt: string;
12409
+ /**
12410
+ * Seed
12411
+ * @description The same seed and the same prompt given to the same version of the model
12412
+ * will output the same image every time.
12413
+ */
12414
+ seed?: number;
12415
+ /**
12416
+ * Sync Mode
12417
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
12418
+ * @default false
12419
+ */
12420
+ sync_mode?: boolean;
12421
+ }
12422
+
12187
12423
  export interface SharedType_215 {
12188
12424
  /**
12189
12425
  * Acceleration
@@ -12778,7 +13014,7 @@ export interface SharedType_129 {
12778
13014
  results: string;
12779
13015
  }
12780
13016
 
12781
- export interface SharedType_100 {
13017
+ export interface SharedType_1001 {
12782
13018
  /**
12783
13019
  * Prompt
12784
13020
  * @description The prompt used for generating the video.
@@ -12796,6 +13032,38 @@ export interface SharedType_100 {
12796
13032
  video: Components.File;
12797
13033
  }
12798
13034
 
13035
+ export interface SharedType_100 {
13036
+ /**
13037
+ * Masks
13038
+ * @description Dictionary of label: mask image
13039
+ * @example [
13040
+ * {
13041
+ * "file_size": 15724,
13042
+ * "height": 1200,
13043
+ * "file_name": "019c3c1e3c50446e9996f709d36debb4.png",
13044
+ * "content_type": "image/png",
13045
+ * "url": "https://v3.fal.media/files/monkey/6ITmhHQJ-69s-UxajrY5T_019c3c1e3c50446e9996f709d36debb4.png",
13046
+ * "width": 1800
13047
+ * },
13048
+ * {
13049
+ * "file_size": 14905,
13050
+ * "height": 1200,
13051
+ * "file_name": "0a1522ca410942c7ad6c73efa15b3549.png",
13052
+ * "content_type": "image/png",
13053
+ * "url": "https://v3.fal.media/files/monkey/IljtMxahoo9-7SUpx0fth_0a1522ca410942c7ad6c73efa15b3549.png",
13054
+ * "width": 1800
13055
+ * }
13056
+ * ]
13057
+ */
13058
+ masks: Components.Image[];
13059
+ /**
13060
+ * Output
13061
+ * @description Generated output
13062
+ * @example <p> A white pickup truck </p> [SEG] is parked on the side of <p> the red building </p> [SEG] , creating a unique and eye-catching contrast.<|im_end|>
13063
+ */
13064
+ output: string;
13065
+ }
13066
+
12799
13067
  export interface SharedType_0ff {
12800
13068
  /**
12801
13069
  * @description The generated speaker embedding file in safetensors format.
@@ -12990,6 +13258,24 @@ export interface SharedType_0cd {
12990
13258
  sync_mode?: boolean;
12991
13259
  }
12992
13260
 
13261
+ export interface SharedType_0c6 {
13262
+ /**
13263
+ * Seed
13264
+ * @description The seed used for generation.
13265
+ */
13266
+ seed: number;
13267
+ /**
13268
+ * @description The generated video file.
13269
+ * @example {
13270
+ * "file_size": 515275,
13271
+ * "file_name": "74af6c0bdd6041c3b1130d54885e3eee.mp4",
13272
+ * "content_type": "application/octet-stream",
13273
+ * "url": "https://v3.fal.media/files/kangaroo/z6VqUwNTwzuWa6YE1g7In_74af6c0bdd6041c3b1130d54885e3eee.mp4"
13274
+ * }
13275
+ */
13276
+ video: Components.File;
13277
+ }
13278
+
12993
13279
  export interface SharedType_0c5 {
12994
13280
  /**
12995
13281
  * Images
@@ -14907,7 +15193,7 @@ export interface ZonosOutput {
14907
15193
  audio: Components.File_1;
14908
15194
  }
14909
15195
 
14910
- export interface ZImageTurboLoraInput {
15196
+ export interface ZImageTurboTilingLoraInput {
14911
15197
  /**
14912
15198
  * Acceleration
14913
15199
  * @description The acceleration level to use.
@@ -14929,8 +15215,8 @@ export interface ZImageTurboLoraInput {
14929
15215
  enable_safety_checker?: boolean;
14930
15216
  /**
14931
15217
  * Image Size
14932
- * @description The size of the generated image.
14933
- * @default landscape_4_3
15218
+ * @description The size of the generated image. Use 'auto' to match the input image size (or 1024x1024 if no image).
15219
+ * @default square_hd
14934
15220
  */
14935
15221
  image_size?:
14936
15222
  | Components.ImageSize
@@ -14941,13 +15227,24 @@ export interface ZImageTurboLoraInput {
14941
15227
  | 'portrait_16_9'
14942
15228
  | 'landscape_4_3'
14943
15229
  | 'landscape_16_9'
15230
+ | 'auto'
14944
15231
  );
15232
+ /**
15233
+ * Image URL
15234
+ * @description URL of an image for image-to-image or inpainting. When provided without mask_image_url, performs image-to-image; with mask_image_url, performs inpainting.
15235
+ */
15236
+ image_url?: string;
14945
15237
  /**
14946
15238
  * Loras
14947
15239
  * @description List of LoRA weights to apply (maximum 3).
14948
15240
  * @default []
14949
15241
  */
14950
15242
  loras?: Components.LoRAInput_2[];
15243
+ /**
15244
+ * Mask Image URL
15245
+ * @description URL of a mask image for inpainting. White regions are regenerated, black regions are preserved. Requires image_url.
15246
+ */
15247
+ mask_image_url?: string;
14951
15248
  /**
14952
15249
  * Number of Images
14953
15250
  * @description The number of images to generate.
@@ -14970,7 +15267,7 @@ export interface ZImageTurboLoraInput {
14970
15267
  /**
14971
15268
  * Prompt
14972
15269
  * @description The prompt to generate an image from.
14973
- * @example A hyper-realistic, close-up portrait of a tribal elder from the Omo Valley, painted with intricate white chalk patterns and adorned with a headdress made of dried flowers, seed pods, and rusted bottle caps. The focus is razor-sharp on the texture of the skin, showing every pore, wrinkle, and scar that tells a story of survival. The background is a blurred, smoky hut interior, with the warm glow of a cooking fire reflecting in the subject's dark, soulful eyes. Shot on a Leica M6 with Kodak Portra 400 film grain aesthetic.
15270
+ * @example A hyper-realistic, high-resolution 4k texture of an ancient weathered brick wall heavily overgrown with lush green moss and soft lichens. The bricks are aged, featuring deep earthy tones, natural cracks, and gritty textures. Vibrant emerald moss fills the mortar lines and spills over the rough surfaces of the stones. Uniform, flat cinematic lighting ensures no harsh shadows, highlighting the intricate organic details and damp stone surfaces. The composition is a perfectly balanced overhead view, showcasing a rich tapestry of botanical growth and masonry craftsmanship with professional clarity and hyper-detailed grit.
14974
15271
  */
14975
15272
  prompt: string;
14976
15273
  /**
@@ -14979,17 +15276,42 @@ export interface ZImageTurboLoraInput {
14979
15276
  * will output the same image every time.
14980
15277
  */
14981
15278
  seed?: number;
15279
+ /**
15280
+ * Strength
15281
+ * @description How much to transform the input image. Only used when image_url is provided.
15282
+ * @default 0.6
15283
+ */
15284
+ strength?: number;
14982
15285
  /**
14983
15286
  * Sync Mode
14984
15287
  * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
14985
15288
  * @default false
14986
15289
  */
14987
15290
  sync_mode?: boolean;
15291
+ /**
15292
+ * Tile Size
15293
+ * @description Tile size in latent space (64 = 512px, 128 = 1024px, 256 = 2048px).
15294
+ * @default 128
15295
+ */
15296
+ tile_size?: number;
15297
+ /**
15298
+ * Tile Stride
15299
+ * @description Tile stride in latent space. (32 = 256px, 64 = 512px, 128 = 1024px).
15300
+ * @default 64
15301
+ */
15302
+ tile_stride?: number;
15303
+ /**
15304
+ * Tiling Mode
15305
+ * @description Tiling direction: 'both' (omnidirectional), 'horizontal' (left-right only), 'vertical' (top-bottom only).
15306
+ * @default both
15307
+ * @enum {string}
15308
+ */
15309
+ tiling_mode?: 'both' | 'horizontal' | 'vertical';
14988
15310
  }
14989
15311
 
14990
- export interface ZImageTurboLoraOutput extends SharedType_352 {}
15312
+ export interface ZImageTurboTilingLoraOutput extends SharedType_390 {}
14991
15313
 
14992
- export interface ZImageTurboInpaintLoraInput {
15314
+ export interface ZImageTurboTilingInput {
14993
15315
  /**
14994
15316
  * Acceleration
14995
15317
  * @description The acceleration level to use.
@@ -14998,23 +15320,117 @@ export interface ZImageTurboInpaintLoraInput {
14998
15320
  */
14999
15321
  acceleration?: 'none' | 'regular' | 'high';
15000
15322
  /**
15001
- * Control End
15002
- * @description The end of the controlnet conditioning.
15003
- * @default 0.8
15323
+ * Enable Prompt Expansion
15324
+ * @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
15325
+ * @default false
15004
15326
  */
15005
- control_end?: number;
15327
+ enable_prompt_expansion?: boolean;
15006
15328
  /**
15007
- * Control Scale
15008
- * @description The scale of the controlnet conditioning.
15009
- * @default 0.75
15329
+ * Enable Safety Checker
15330
+ * @description If set to true, the safety checker will be enabled.
15331
+ * @default true
15010
15332
  */
15011
- control_scale?: number;
15333
+ enable_safety_checker?: boolean;
15012
15334
  /**
15013
- * Control Start
15014
- * @description The start of the controlnet conditioning.
15015
- * @default 0
15335
+ * Image Size
15336
+ * @description The size of the generated image. Use 'auto' to match the input image size (or 1024x1024 if no image).
15337
+ * @default square_hd
15016
15338
  */
15017
- control_start?: number;
15339
+ image_size?:
15340
+ | Components.ImageSize
15341
+ | (
15342
+ | 'square_hd'
15343
+ | 'square'
15344
+ | 'portrait_4_3'
15345
+ | 'portrait_16_9'
15346
+ | 'landscape_4_3'
15347
+ | 'landscape_16_9'
15348
+ | 'auto'
15349
+ );
15350
+ /**
15351
+ * Image URL
15352
+ * @description URL of an image for image-to-image or inpainting. When provided without mask_image_url, performs image-to-image; with mask_image_url, performs inpainting.
15353
+ */
15354
+ image_url?: string;
15355
+ /**
15356
+ * Mask Image URL
15357
+ * @description URL of a mask image for inpainting. White regions are regenerated, black regions are preserved. Requires image_url.
15358
+ */
15359
+ mask_image_url?: string;
15360
+ /**
15361
+ * Number of Images
15362
+ * @description The number of images to generate.
15363
+ * @default 1
15364
+ */
15365
+ num_images?: number;
15366
+ /**
15367
+ * Number of Inference Steps
15368
+ * @description The number of inference steps to perform.
15369
+ * @default 8
15370
+ */
15371
+ num_inference_steps?: number;
15372
+ /**
15373
+ * Output Format
15374
+ * @description The format of the generated image.
15375
+ * @default png
15376
+ * @enum {string}
15377
+ */
15378
+ output_format?: 'jpeg' | 'png' | 'webp';
15379
+ /**
15380
+ * Prompt
15381
+ * @description The prompt to generate an image from.
15382
+ * @example A hyper-realistic, high-resolution 4k texture of an ancient weathered brick wall heavily overgrown with lush green moss and soft lichens. The bricks are aged, featuring deep earthy tones, natural cracks, and gritty textures. Vibrant emerald moss fills the mortar lines and spills over the rough surfaces of the stones. Uniform, flat cinematic lighting ensures no harsh shadows, highlighting the intricate organic details and damp stone surfaces. The composition is a perfectly balanced overhead view, showcasing a rich tapestry of botanical growth and masonry craftsmanship with professional clarity and hyper-detailed grit.
15383
+ */
15384
+ prompt: string;
15385
+ /**
15386
+ * Seed
15387
+ * @description The same seed and the same prompt given to the same version of the model
15388
+ * will output the same image every time.
15389
+ */
15390
+ seed?: number;
15391
+ /**
15392
+ * Strength
15393
+ * @description How much to transform the input image. Only used when image_url is provided.
15394
+ * @default 0.6
15395
+ */
15396
+ strength?: number;
15397
+ /**
15398
+ * Sync Mode
15399
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
15400
+ * @default false
15401
+ */
15402
+ sync_mode?: boolean;
15403
+ /**
15404
+ * Tile Size
15405
+ * @description Tile size in latent space (64 = 512px, 128 = 1024px, 256 = 2048px).
15406
+ * @default 128
15407
+ */
15408
+ tile_size?: number;
15409
+ /**
15410
+ * Tile Stride
15411
+ * @description Tile stride in latent space. (32 = 256px, 64 = 512px, 128 = 1024px).
15412
+ * @default 64
15413
+ */
15414
+ tile_stride?: number;
15415
+ /**
15416
+ * Tiling Mode
15417
+ * @description Tiling direction: 'both' (omnidirectional), 'horizontal' (left-right only), 'vertical' (top-bottom only).
15418
+ * @default both
15419
+ * @enum {string}
15420
+ */
15421
+ tiling_mode?: 'both' | 'horizontal' | 'vertical';
15422
+ }
15423
+
15424
+ export interface ZImageTurboTilingOutput extends SharedType_390 {}
15425
+
15426
+ export interface ZImageTurboLoraInput {
15427
+ /**
15428
+ * Acceleration
15429
+ * @description The acceleration level to use.
15430
+ * @default regular
15431
+ * @enum {string}
15432
+ */
15433
+ acceleration?: 'none' | 'regular' | 'high';
15018
15434
  /**
15019
15435
  * Enable Prompt Expansion
15020
15436
  * @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
@@ -15030,7 +15446,7 @@ export interface ZImageTurboInpaintLoraInput {
15030
15446
  /**
15031
15447
  * Image Size
15032
15448
  * @description The size of the generated image.
15033
- * @default auto
15449
+ * @default landscape_4_3
15034
15450
  */
15035
15451
  image_size?:
15036
15452
  | Components.ImageSize
@@ -15041,26 +15457,13 @@ export interface ZImageTurboInpaintLoraInput {
15041
15457
  | 'portrait_16_9'
15042
15458
  | 'landscape_4_3'
15043
15459
  | 'landscape_16_9'
15044
- | 'auto'
15045
15460
  );
15046
- /**
15047
- * Image URL
15048
- * @description URL of Image for Inpaint generation.
15049
- * @example https://storage.googleapis.com/falserverless/example_inputs/inpaint-input.jpg
15050
- */
15051
- image_url: string;
15052
15461
  /**
15053
15462
  * Loras
15054
15463
  * @description List of LoRA weights to apply (maximum 3).
15055
15464
  * @default []
15056
15465
  */
15057
15466
  loras?: Components.LoRAInput_2[];
15058
- /**
15059
- * Mask Image URL
15060
- * @description URL of Mask for Inpaint generation.
15061
- * @example https://storage.googleapis.com/falserverless/whls/z-image-inpaint-mask.jpg
15062
- */
15063
- mask_image_url: string;
15064
15467
  /**
15065
15468
  * Number of Images
15066
15469
  * @description The number of images to generate.
@@ -15083,7 +15486,7 @@ export interface ZImageTurboInpaintLoraInput {
15083
15486
  /**
15084
15487
  * Prompt
15085
15488
  * @description The prompt to generate an image from.
15086
- * @example A young Asian woman with long, vibrant purple hair stands on a sunlit sandy beach, posing confidently with her left hand resting on her hip. She gazes directly at the camera with a neutral expression. A sleek black ribbon bow is tied neatly on the right side of her head, just above her ear. She wears a flowing white cotton dress with a fitted bodice and a flared skirt that reaches mid-calf, slightly lifted by a gentle sea breeze. The beach behind her features fine, pale golden sand with subtle footprints, leading to calm turquoise waves under a clear blue sky with soft, wispy clouds. The lighting is natural daylight, casting soft shadows to her left, indicating late afternoon sun. The horizon line is visible in the background, with a faint silhouette of distant dunes. Her skin tone is fair with a natural glow, and her facial features are delicately defined. The composition is centered on her figure, framed from mid-thigh up, with shallow depth of field blurring the distant waves slightly.
15489
+ * @example A hyper-realistic, close-up portrait of a tribal elder from the Omo Valley, painted with intricate white chalk patterns and adorned with a headdress made of dried flowers, seed pods, and rusted bottle caps. The focus is razor-sharp on the texture of the skin, showing every pore, wrinkle, and scar that tells a story of survival. The background is a blurred, smoky hut interior, with the warm glow of a cooking fire reflecting in the subject's dark, soulful eyes. Shot on a Leica M6 with Kodak Portra 400 film grain aesthetic.
15087
15490
  */
15088
15491
  prompt: string;
15089
15492
  /**
@@ -15092,12 +15495,6 @@ export interface ZImageTurboInpaintLoraInput {
15092
15495
  * will output the same image every time.
15093
15496
  */
15094
15497
  seed?: number;
15095
- /**
15096
- * Strength
15097
- * @description The strength of the inpaint conditioning.
15098
- * @default 1
15099
- */
15100
- strength?: number;
15101
15498
  /**
15102
15499
  * Sync Mode
15103
15500
  * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
@@ -15106,9 +15503,9 @@ export interface ZImageTurboInpaintLoraInput {
15106
15503
  sync_mode?: boolean;
15107
15504
  }
15108
15505
 
15109
- export interface ZImageTurboInpaintLoraOutput extends SharedType_18d {}
15506
+ export interface ZImageTurboLoraOutput extends SharedType_352 {}
15110
15507
 
15111
- export interface ZImageTurboInpaintInput {
15508
+ export interface ZImageTurboInpaintLoraInput {
15112
15509
  /**
15113
15510
  * Acceleration
15114
15511
  * @description The acceleration level to use.
@@ -15168,6 +15565,12 @@ export interface ZImageTurboInpaintInput {
15168
15565
  * @example https://storage.googleapis.com/falserverless/example_inputs/inpaint-input.jpg
15169
15566
  */
15170
15567
  image_url: string;
15568
+ /**
15569
+ * Loras
15570
+ * @description List of LoRA weights to apply (maximum 3).
15571
+ * @default []
15572
+ */
15573
+ loras?: Components.LoRAInput_2[];
15171
15574
  /**
15172
15575
  * Mask Image URL
15173
15576
  * @description URL of Mask for Inpaint generation.
@@ -15219,9 +15622,9 @@ export interface ZImageTurboInpaintInput {
15219
15622
  sync_mode?: boolean;
15220
15623
  }
15221
15624
 
15222
- export interface ZImageTurboInpaintOutput extends SharedType_18d {}
15625
+ export interface ZImageTurboInpaintLoraOutput extends SharedType_18d {}
15223
15626
 
15224
- export interface ZImageTurboImageToImageLoraInput {
15627
+ export interface ZImageTurboInpaintInput {
15225
15628
  /**
15226
15629
  * Acceleration
15227
15630
  * @description The acceleration level to use.
@@ -15229,6 +15632,24 @@ export interface ZImageTurboImageToImageLoraInput {
15229
15632
  * @enum {string}
15230
15633
  */
15231
15634
  acceleration?: 'none' | 'regular' | 'high';
15635
+ /**
15636
+ * Control End
15637
+ * @description The end of the controlnet conditioning.
15638
+ * @default 0.8
15639
+ */
15640
+ control_end?: number;
15641
+ /**
15642
+ * Control Scale
15643
+ * @description The scale of the controlnet conditioning.
15644
+ * @default 0.75
15645
+ */
15646
+ control_scale?: number;
15647
+ /**
15648
+ * Control Start
15649
+ * @description The start of the controlnet conditioning.
15650
+ * @default 0
15651
+ */
15652
+ control_start?: number;
15232
15653
  /**
15233
15654
  * Enable Prompt Expansion
15234
15655
  * @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
@@ -15259,16 +15680,16 @@ export interface ZImageTurboImageToImageLoraInput {
15259
15680
  );
15260
15681
  /**
15261
15682
  * Image URL
15262
- * @description URL of Image for Image-to-Image generation.
15263
- * @example https://storage.googleapis.com/falserverless/example_inputs/z-image-turbo-i2i-input.png
15683
+ * @description URL of Image for Inpaint generation.
15684
+ * @example https://storage.googleapis.com/falserverless/example_inputs/inpaint-input.jpg
15264
15685
  */
15265
15686
  image_url: string;
15266
15687
  /**
15267
- * Loras
15268
- * @description List of LoRA weights to apply (maximum 3).
15269
- * @default []
15688
+ * Mask Image URL
15689
+ * @description URL of Mask for Inpaint generation.
15690
+ * @example https://storage.googleapis.com/falserverless/whls/z-image-inpaint-mask.jpg
15270
15691
  */
15271
- loras?: Components.LoRAInput_2[];
15692
+ mask_image_url: string;
15272
15693
  /**
15273
15694
  * Number of Images
15274
15695
  * @description The number of images to generate.
@@ -15302,8 +15723,8 @@ export interface ZImageTurboImageToImageLoraInput {
15302
15723
  seed?: number;
15303
15724
  /**
15304
15725
  * Strength
15305
- * @description The strength of the image-to-image conditioning.
15306
- * @default 0.6
15726
+ * @description The strength of the inpaint conditioning.
15727
+ * @default 1
15307
15728
  */
15308
15729
  strength?: number;
15309
15730
  /**
@@ -15314,9 +15735,104 @@ export interface ZImageTurboImageToImageLoraInput {
15314
15735
  sync_mode?: boolean;
15315
15736
  }
15316
15737
 
15317
- export interface ZImageTurboImageToImageLoraOutput extends SharedType_6b3 {}
15738
+ export interface ZImageTurboInpaintOutput extends SharedType_18d {}
15318
15739
 
15319
- export interface ZImageTurboImageToImageInput {
15740
+ export interface ZImageTurboImageToImageLoraInput {
15741
+ /**
15742
+ * Acceleration
15743
+ * @description The acceleration level to use.
15744
+ * @default regular
15745
+ * @enum {string}
15746
+ */
15747
+ acceleration?: 'none' | 'regular' | 'high';
15748
+ /**
15749
+ * Enable Prompt Expansion
15750
+ * @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
15751
+ * @default false
15752
+ */
15753
+ enable_prompt_expansion?: boolean;
15754
+ /**
15755
+ * Enable Safety Checker
15756
+ * @description If set to true, the safety checker will be enabled.
15757
+ * @default true
15758
+ */
15759
+ enable_safety_checker?: boolean;
15760
+ /**
15761
+ * Image Size
15762
+ * @description The size of the generated image.
15763
+ * @default auto
15764
+ */
15765
+ image_size?:
15766
+ | Components.ImageSize
15767
+ | (
15768
+ | 'square_hd'
15769
+ | 'square'
15770
+ | 'portrait_4_3'
15771
+ | 'portrait_16_9'
15772
+ | 'landscape_4_3'
15773
+ | 'landscape_16_9'
15774
+ | 'auto'
15775
+ );
15776
+ /**
15777
+ * Image URL
15778
+ * @description URL of Image for Image-to-Image generation.
15779
+ * @example https://storage.googleapis.com/falserverless/example_inputs/z-image-turbo-i2i-input.png
15780
+ */
15781
+ image_url: string;
15782
+ /**
15783
+ * Loras
15784
+ * @description List of LoRA weights to apply (maximum 3).
15785
+ * @default []
15786
+ */
15787
+ loras?: Components.LoRAInput_2[];
15788
+ /**
15789
+ * Number of Images
15790
+ * @description The number of images to generate.
15791
+ * @default 1
15792
+ */
15793
+ num_images?: number;
15794
+ /**
15795
+ * Number of Inference Steps
15796
+ * @description The number of inference steps to perform.
15797
+ * @default 8
15798
+ */
15799
+ num_inference_steps?: number;
15800
+ /**
15801
+ * Output Format
15802
+ * @description The format of the generated image.
15803
+ * @default png
15804
+ * @enum {string}
15805
+ */
15806
+ output_format?: 'jpeg' | 'png' | 'webp';
15807
+ /**
15808
+ * Prompt
15809
+ * @description The prompt to generate an image from.
15810
+ * @example A young Asian woman with long, vibrant purple hair stands on a sunlit sandy beach, posing confidently with her left hand resting on her hip. She gazes directly at the camera with a neutral expression. A sleek black ribbon bow is tied neatly on the right side of her head, just above her ear. She wears a flowing white cotton dress with a fitted bodice and a flared skirt that reaches mid-calf, slightly lifted by a gentle sea breeze. The beach behind her features fine, pale golden sand with subtle footprints, leading to calm turquoise waves under a clear blue sky with soft, wispy clouds. The lighting is natural daylight, casting soft shadows to her left, indicating late afternoon sun. The horizon line is visible in the background, with a faint silhouette of distant dunes. Her skin tone is fair with a natural glow, and her facial features are delicately defined. The composition is centered on her figure, framed from mid-thigh up, with shallow depth of field blurring the distant waves slightly.
15811
+ */
15812
+ prompt: string;
15813
+ /**
15814
+ * Seed
15815
+ * @description The same seed and the same prompt given to the same version of the model
15816
+ * will output the same image every time.
15817
+ */
15818
+ seed?: number;
15819
+ /**
15820
+ * Strength
15821
+ * @description The strength of the image-to-image conditioning.
15822
+ * @default 0.6
15823
+ */
15824
+ strength?: number;
15825
+ /**
15826
+ * Sync Mode
15827
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
15828
+ * @default false
15829
+ */
15830
+ sync_mode?: boolean;
15831
+ }
15832
+
15833
+ export interface ZImageTurboImageToImageLoraOutput extends SharedType_6b3 {}
15834
+
15835
+ export interface ZImageTurboImageToImageInput {
15320
15836
  /**
15321
15837
  * Acceleration
15322
15838
  * @description The acceleration level to use.
@@ -20089,7 +20605,21 @@ export interface WanVace13bInput {
20089
20605
  video_url?: string;
20090
20606
  }
20091
20607
 
20092
- export interface WanVace13bOutput extends SharedType_d26 {}
20608
+ export interface WanVace13bOutput {
20609
+ /**
20610
+ * Seed
20611
+ * @description The seed used for generation.
20612
+ */
20613
+ seed: number;
20614
+ /**
20615
+ * Video
20616
+ * @description The generated video file.
20617
+ * @example {
20618
+ * "url": "https://storage.googleapis.com/falserverless/vace/out_video_vace.mp4"
20619
+ * }
20620
+ */
20621
+ video: Components.File_1;
20622
+ }
20093
20623
 
20094
20624
  export interface WanVaceInput {
20095
20625
  /**
@@ -20202,7 +20732,20 @@ export interface WanVaceInput {
20202
20732
  video_url?: string;
20203
20733
  }
20204
20734
 
20205
- export interface WanVaceOutput extends SharedType_d26 {}
20735
+ export interface WanVaceOutput {
20736
+ /**
20737
+ * Seed
20738
+ * @description The seed used for generation.
20739
+ */
20740
+ seed: number;
20741
+ /**
20742
+ * @description The generated video file.
20743
+ * @example {
20744
+ * "url": "https://storage.googleapis.com/falserverless/vace/out_video_vace.mp4"
20745
+ * }
20746
+ */
20747
+ video: Components.File;
20748
+ }
20206
20749
 
20207
20750
  export interface WanTrainerT2v14bInput extends SharedType_5951 {}
20208
20751
 
@@ -20257,7 +20800,7 @@ export interface WanT2vLoraInput {
20257
20800
  * @description LoRA weights to be used in the inference.
20258
20801
  * @default []
20259
20802
  */
20260
- loras?: Components.LoraWeight_3[];
20803
+ loras?: Components.LoraWeight_5[];
20261
20804
  /**
20262
20805
  * Negative Prompt
20263
20806
  * @description Negative prompt for video generation.
@@ -20572,7 +21115,7 @@ export interface WanI2vLoraInput {
20572
21115
  * @description LoRA weights to be used in the inference.
20573
21116
  * @default []
20574
21117
  */
20575
- loras?: Components.LoraWeight_3[];
21118
+ loras?: Components.LoraWeight_5[];
20576
21119
  /**
20577
21120
  * Negative Prompt
20578
21121
  * @description Negative prompt for video generation.
@@ -24153,13 +24696,12 @@ export interface ThinksoundAudioInput extends SharedType_4bc {}
24153
24696
 
24154
24697
  export interface ThinksoundAudioOutput {
24155
24698
  /**
24156
- * Audio
24157
24699
  * @description The generated audio file.
24158
24700
  * @example {
24159
24701
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/thinksound-audio.wav"
24160
24702
  * }
24161
24703
  */
24162
- audio: Components.File_1;
24704
+ audio: Components.File;
24163
24705
  /**
24164
24706
  * Prompt
24165
24707
  * @description The prompt used to generate the audio.
@@ -24178,13 +24720,12 @@ export interface ThinksoundOutput {
24178
24720
  */
24179
24721
  prompt: string;
24180
24722
  /**
24181
- * Video
24182
24723
  * @description The generated video with audio.
24183
24724
  * @example {
24184
24725
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/thinksound-output.mp4"
24185
24726
  * }
24186
24727
  */
24187
- video: Components.File_1;
24728
+ video: Components.File;
24188
24729
  }
24189
24730
 
24190
24731
  export interface TheraInput {
@@ -24834,7 +25375,6 @@ export interface SteadyDancerOutput {
24834
25375
  */
24835
25376
  seed: number;
24836
25377
  /**
24837
- * Video
24838
25378
  * @description The generated dance animation video.
24839
25379
  * @example {
24840
25380
  * "file_size": 7772111,
@@ -24843,7 +25383,7 @@ export interface SteadyDancerOutput {
24843
25383
  * "url": "https://v3b.fal.media/files/b/0a87871b/ll5ps0ZyBgxBkuWz-fHcT_output_with_audio.mp4"
24844
25384
  * }
24845
25385
  */
24846
- video: Components.File_1;
25386
+ video: Components.File;
24847
25387
  }
24848
25388
 
24849
25389
  export interface StarVectorInput {
@@ -25360,6 +25900,7 @@ export interface StableDiffusionV15Input {
25360
25900
  * Seed
25361
25901
  * @description The same seed and the same prompt given to the same version of Stable Diffusion
25362
25902
  * will output the same image every time.
25903
+ * @default null
25363
25904
  */
25364
25905
  seed?: number;
25365
25906
  /**
@@ -25604,11 +26145,10 @@ export interface StableAvatarInput {
25604
26145
 
25605
26146
  export interface StableAvatarOutput {
25606
26147
  /**
25607
- * Video
25608
26148
  * @description The generated video file.
25609
26149
  * @example https://storage.googleapis.com/falserverless/example_outputs/stable-avatar-output.mp4
25610
26150
  */
25611
- video: Components.File_1;
26151
+ video: Components.File;
25612
26152
  }
25613
26153
 
25614
26154
  export interface StableAudio25TextToAudioInput {
@@ -26275,92 +26815,6 @@ export interface SkyreelsI2vOutput {
26275
26815
  video: Components.File;
26276
26816
  }
26277
26817
 
26278
- export interface SkyRaccoonInput {
26279
- /**
26280
- * Enable Prompt Expansion
26281
- * @description Whether to enable prompt expansion.
26282
- * @default false
26283
- * @example false
26284
- */
26285
- enable_prompt_expansion?: boolean;
26286
- /**
26287
- * Enable Safety Checker
26288
- * @description If set to true, the safety checker will be enabled.
26289
- * @default false
26290
- * @example true
26291
- */
26292
- enable_safety_checker?: boolean;
26293
- /**
26294
- * Image Size
26295
- * @description The size of the generated image.
26296
- * @default {
26297
- * "height": 1024,
26298
- * "width": 1024
26299
- * }
26300
- */
26301
- image_size?:
26302
- | Components.ImageSize
26303
- | (
26304
- | 'square_hd'
26305
- | 'square'
26306
- | 'portrait_4_3'
26307
- | 'portrait_16_9'
26308
- | 'landscape_4_3'
26309
- | 'landscape_16_9'
26310
- );
26311
- /**
26312
- * Loras
26313
- * @description LoRA weights to be used in the inference.
26314
- * @default []
26315
- */
26316
- loras?: Components.LoraWeight_3[];
26317
- /**
26318
- * Negative Prompt
26319
- * @description Negative prompt for video generation.
26320
- * @default bright colors, overexposed, static, blurred details, subtitles, style, artwork, painting, picture, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, malformed limbs, fused fingers, still picture, cluttered background, three legs, many people in the background, walking backwards
26321
- * @example bright colors, overexposed, static, blurred details, subtitles, style, artwork, painting, picture, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, malformed limbs, fused fingers, still picture, cluttered background, three legs, many people in the background, walking backwards
26322
- */
26323
- negative_prompt?: string;
26324
- /**
26325
- * Num Inference Steps
26326
- * @description Number of inference steps for sampling. Higher values give better quality but take longer.
26327
- * @default 30
26328
- */
26329
- num_inference_steps?: number;
26330
- /**
26331
- * Prompt
26332
- * @description The text prompt to guide video generation.
26333
- * @example A stylish woman walks down a Tokyo street filled with warm glowing neon and animated city signage. She wears a black leather jacket, a long red dress, and black boots, and carries a black purse.
26334
- */
26335
- prompt: string;
26336
- /**
26337
- * Seed
26338
- * @description Random seed for reproducibility. If None, a random seed is chosen.
26339
- */
26340
- seed?: number;
26341
- /**
26342
- * Turbo Mode
26343
- * @description If true, the image will be generated faster with no noticeable degradation in the visual quality.
26344
- * @default true
26345
- */
26346
- turbo_mode?: boolean;
26347
- }
26348
-
26349
- export interface SkyRaccoonOutput {
26350
- /**
26351
- * @description The generated image file.
26352
- * @example {
26353
- * "url": "https://v3.fal.media/files/kangaroo/xP17uR1w0uHA1T2W2k6Gi_video-1752605675.png"
26354
- * }
26355
- */
26356
- image: Components.File;
26357
- /**
26358
- * Seed
26359
- * @description The seed used for generation.
26360
- */
26361
- seed: number;
26362
- }
26363
-
26364
26818
  export interface SileroVadInput {
26365
26819
  /**
26366
26820
  * Audio URL
@@ -27219,7 +27673,7 @@ export interface Sd15DepthControlnetInput {
27219
27673
  * @description The list of LoRA weights to use.
27220
27674
  * @default []
27221
27675
  */
27222
- loras?: Components.LoraWeight_4[];
27676
+ loras?: Components.LoraWeight_3[];
27223
27677
  /**
27224
27678
  * Negative Prompt
27225
27679
  * @description The negative prompt to use. Use it to address details that you don't want
@@ -27544,15 +27998,15 @@ export interface Sam2VideoInput {
27544
27998
  * @example [
27545
27999
  * {
27546
28000
  * "y": 350,
27547
- * "label": 1,
28001
+ * "x": 210,
27548
28002
  * "frame_index": 0,
27549
- * "x": 210
28003
+ * "label": 1
27550
28004
  * },
27551
28005
  * {
27552
28006
  * "y": 220,
27553
- * "label": 1,
28007
+ * "x": 250,
27554
28008
  * "frame_index": 0,
27555
- * "x": 250
28009
+ * "label": 1
27556
28010
  * }
27557
28011
  * ]
27558
28012
  */
@@ -27611,8 +28065,8 @@ export interface Sam2ImageInput {
27611
28065
  * @example [
27612
28066
  * {
27613
28067
  * "y": 375,
27614
- * "label": 1,
27615
- * "x": 500
28068
+ * "x": 500,
28069
+ * "label": 1
27616
28070
  * }
27617
28071
  * ]
27618
28072
  */
@@ -28052,6 +28506,13 @@ export interface Sam3VideoInput {
28052
28506
  * @description [DEPRECATED] Use 'prompt' instead. Kept for backward compatibility.
28053
28507
  */
28054
28508
  text_prompt?: string;
28509
+ /**
28510
+ * Video Output Type
28511
+ * @description The output type of the generated video.
28512
+ * @default X264 (.mp4)
28513
+ * @enum {string}
28514
+ */
28515
+ video_output_type?: 'X264 (.mp4)' | 'VP9 (.webm)';
28055
28516
  /**
28056
28517
  * Video Url
28057
28518
  * @description The URL of the video to be segmented.
@@ -28409,7 +28870,7 @@ export interface Sa2va8bVideoOutput extends SharedType_800 {}
28409
28870
 
28410
28871
  export interface Sa2va8bImageInput extends SharedType_f51 {}
28411
28872
 
28412
- export interface Sa2va8bImageOutput extends SharedType_fda {}
28873
+ export interface Sa2va8bImageOutput extends SharedType_100 {}
28413
28874
 
28414
28875
  export interface Sa2va4bVideoInput extends SharedType_5f5 {}
28415
28876
 
@@ -28417,7 +28878,7 @@ export interface Sa2va4bVideoOutput extends SharedType_800 {}
28417
28878
 
28418
28879
  export interface Sa2va4bImageInput extends SharedType_f51 {}
28419
28880
 
28420
- export interface Sa2va4bImageOutput extends SharedType_fda {}
28881
+ export interface Sa2va4bImageOutput extends SharedType_100 {}
28421
28882
 
28422
28883
  export interface RifeVideoInput {
28423
28884
  /**
@@ -28558,13 +29019,12 @@ export interface RetoucherInput {
28558
29019
 
28559
29020
  export interface RetoucherOutput {
28560
29021
  /**
28561
- * Image
28562
29022
  * @description The generated image file info.
28563
29023
  * @example {
28564
29024
  * "url": "https://storage.googleapis.com/falserverless/model_tests/retoucher/retoucher_example_output.png"
28565
29025
  * }
28566
29026
  */
28567
- image: Components.Image_2;
29027
+ image: Components.Image;
28568
29028
  /**
28569
29029
  * Seed
28570
29030
  * @description The seed used for the generation.
@@ -36059,15 +36519,7 @@ export interface OrpheusTtsInput {
36059
36519
  voice?: 'tara' | 'leah' | 'jess' | 'leo' | 'dan' | 'mia' | 'zac' | 'zoe';
36060
36520
  }
36061
36521
 
36062
- export interface OrpheusTtsOutput {
36063
- /**
36064
- * @description The generated speech audio
36065
- * @example {
36066
- * "url": "https://v3.fal.media/files/kangaroo/RQ_pxc7oPdueYqWUqEbPE_tmpjnzvvzx_.wav"
36067
- * }
36068
- */
36069
- audio: Components.File;
36070
- }
36522
+ export interface OrpheusTtsOutput extends SharedType_d2a {}
36071
36523
 
36072
36524
  export interface OnerewardInput {
36073
36525
  /**
@@ -36266,6 +36718,122 @@ export interface OmnipartOutput {
36266
36718
  seed: number;
36267
36719
  }
36268
36720
 
36721
+ export interface OmnilottieVideoToLottieInput {
36722
+ /**
36723
+ * Max Tokens
36724
+ * @description Maximum number of Lottie tokens to generate.
36725
+ * @default 4096
36726
+ */
36727
+ max_tokens?: number;
36728
+ /**
36729
+ * Prompt
36730
+ * @description Optional text description guiding the conversion.
36731
+ */
36732
+ prompt?: string;
36733
+ /**
36734
+ * Temperature
36735
+ * @description Sampling temperature for generation.
36736
+ * @default 0.9
36737
+ */
36738
+ temperature?: number;
36739
+ /**
36740
+ * Top K
36741
+ * @description Top-k sampling parameter.
36742
+ * @default 5
36743
+ */
36744
+ top_k?: number;
36745
+ /**
36746
+ * Top P
36747
+ * @description Nucleus sampling probability threshold.
36748
+ * @default 0.25
36749
+ */
36750
+ top_p?: number;
36751
+ /**
36752
+ * Video Url
36753
+ * @description URL of the video to convert into a Lottie animation.
36754
+ * @example https://storage.googleapis.com/falserverless/gallery/0dd62535b46d6f8ecf42f22dc0e148cd.mp4
36755
+ */
36756
+ video_url: string;
36757
+ }
36758
+
36759
+ export interface OmnilottieVideoToLottieOutput extends SharedType_7c2 {}
36760
+
36761
+ export interface OmnilottieImageToLottieInput {
36762
+ /**
36763
+ * Image Url
36764
+ * @description URL of the reference image to animate.
36765
+ * @example https://storage.googleapis.com/falserverless/gallery/00de75e2c031cb3fc3f472e356aba5b6.png
36766
+ */
36767
+ image_url: string;
36768
+ /**
36769
+ * Max Tokens
36770
+ * @description Maximum number of Lottie tokens to generate.
36771
+ * @default 4096
36772
+ */
36773
+ max_tokens?: number;
36774
+ /**
36775
+ * Prompt
36776
+ * @description Text description guiding the animation of the image.
36777
+ * @example A cyan-colored, elongated shape resembling a smiley face, consisting of two pink oval eyes with navy blue circular pupils positioned symmetrically. The shape rotates gradually around its horizontal axis, revealing small animated sparkles—pink and teal—that appear sporadically near the eyes, enhancing the dynamic visual effect.
36778
+ */
36779
+ prompt: string;
36780
+ /**
36781
+ * Temperature
36782
+ * @description Sampling temperature for generation.
36783
+ * @default 0.9
36784
+ */
36785
+ temperature?: number;
36786
+ /**
36787
+ * Top K
36788
+ * @description Top-k sampling parameter.
36789
+ * @default 5
36790
+ */
36791
+ top_k?: number;
36792
+ /**
36793
+ * Top P
36794
+ * @description Nucleus sampling probability threshold.
36795
+ * @default 0.25
36796
+ */
36797
+ top_p?: number;
36798
+ }
36799
+
36800
+ export interface OmnilottieImageToLottieOutput extends SharedType_7c2 {}
36801
+
36802
+ export interface OmnilottieInput {
36803
+ /**
36804
+ * Max Tokens
36805
+ * @description Maximum number of Lottie tokens to generate.
36806
+ * @default 4096
36807
+ */
36808
+ max_tokens?: number;
36809
+ /**
36810
+ * Prompt
36811
+ * @description Text description of the Lottie animation to generate.
36812
+ * @example A red ball bouncing up and down
36813
+ */
36814
+ prompt: string;
36815
+ /**
36816
+ * Temperature
36817
+ * @description Sampling temperature for generation.
36818
+ * @default 0.9
36819
+ */
36820
+ temperature?: number;
36821
+ /**
36822
+ * Top K
36823
+ * @description Top-k sampling parameter.
36824
+ * @default 5
36825
+ */
36826
+ top_k?: number;
36827
+ /**
36828
+ * Top P
36829
+ * @description Nucleus sampling probability threshold.
36830
+ * @default 0.25
36831
+ */
36832
+ top_p?: number;
36833
+ }
36834
+
36835
+ export interface OmnilottieOutput extends SharedType_7c2 {}
36836
+
36269
36837
  export interface OmnigenV2Input {
36270
36838
  /**
36271
36839
  * Cfg Range End
@@ -37043,10 +37611,10 @@ export interface Moondream3PreviewQueryOutput {
37043
37611
  * @description Usage information for the request
37044
37612
  * @example {
37045
37613
  * "output_tokens": 23,
37046
- * "decode_time_ms": 811.5944429300725,
37614
+ * "prefill_time_ms": 54.45315001998097,
37047
37615
  * "input_tokens": 737,
37048
37616
  * "ttft_ms": 91.87838807702065,
37049
- * "prefill_time_ms": 54.45315001998097
37617
+ * "decode_time_ms": 811.5944429300725
37050
37618
  * }
37051
37619
  */
37052
37620
  usage_info: Components.UsageInfo_1;
@@ -37143,10 +37711,10 @@ export interface Moondream3PreviewPointOutput {
37143
37711
  * @description Usage information for the request
37144
37712
  * @example {
37145
37713
  * "output_tokens": 23,
37146
- * "decode_time_ms": 811.5944429300725,
37714
+ * "prefill_time_ms": 54.45315001998097,
37147
37715
  * "input_tokens": 737,
37148
37716
  * "ttft_ms": 91.87838807702065,
37149
- * "prefill_time_ms": 54.45315001998097
37717
+ * "decode_time_ms": 811.5944429300725
37150
37718
  * }
37151
37719
  */
37152
37720
  usage_info: Components.UsageInfo_1;
@@ -37211,10 +37779,10 @@ export interface Moondream3PreviewDetectOutput {
37211
37779
  * @description Usage information for the request
37212
37780
  * @example {
37213
37781
  * "output_tokens": 23,
37214
- * "decode_time_ms": 811.5944429300725,
37782
+ * "prefill_time_ms": 54.45315001998097,
37215
37783
  * "input_tokens": 737,
37216
37784
  * "ttft_ms": 91.87838807702065,
37217
- * "prefill_time_ms": 54.45315001998097
37785
+ * "decode_time_ms": 811.5944429300725
37218
37786
  * }
37219
37787
  */
37220
37788
  usage_info: Components.UsageInfo_1;
@@ -37263,10 +37831,10 @@ export interface Moondream3PreviewCaptionOutput {
37263
37831
  * @description Usage information for the request
37264
37832
  * @example {
37265
37833
  * "output_tokens": 23,
37266
- * "decode_time_ms": 811.5944429300725,
37834
+ * "prefill_time_ms": 54.45315001998097,
37267
37835
  * "input_tokens": 737,
37268
37836
  * "ttft_ms": 91.87838807702065,
37269
- * "prefill_time_ms": 54.45315001998097
37837
+ * "decode_time_ms": 811.5944429300725
37270
37838
  * }
37271
37839
  */
37272
37840
  usage_info: Components.UsageInfo_1;
@@ -40350,6 +40918,12 @@ export interface Ltx2VideoTrainerInput {
40350
40918
  * @example false
40351
40919
  */
40352
40920
  auto_scale_input?: boolean;
40921
+ /**
40922
+ * Debug Dataset
40923
+ * @description When enabled, the trainer returns a downloadable archive of your preprocessed training data for manual inspection. Use this to verify that your videos, images, and captions were processed correctly before committing to a full training run.
40924
+ * @default false
40925
+ */
40926
+ debug_dataset?: boolean;
40353
40927
  /**
40354
40928
  * First Frame Conditioning P
40355
40929
  * @description Probability of conditioning on the first frame during training. Higher values improve image-to-video performance.
@@ -40512,6 +41086,12 @@ export interface Ltx2V2vTrainerInput {
40512
41086
  * @example false
40513
41087
  */
40514
41088
  auto_scale_input?: boolean;
41089
+ /**
41090
+ * Debug Dataset
41091
+ * @description When enabled, the trainer returns a downloadable archive of your preprocessed training data for manual inspection. Use this to verify that your videos, images, and captions were processed correctly before committing to a full training run.
41092
+ * @default false
41093
+ */
41094
+ debug_dataset?: boolean;
40515
41095
  /**
40516
41096
  * First Frame Conditioning P
40517
41097
  * @description Probability of conditioning on the first frame during training. Lower values work better for video-to-video transformation.
@@ -42797,7 +43377,7 @@ export interface Ltx219bVideoToVideoLoraInput {
42797
43377
  * LoRAs
42798
43378
  * @description The LoRAs to use for the generation.
42799
43379
  */
42800
- loras: Components.LoRAInput[];
43380
+ loras: Components.LoRAInput_1[];
42801
43381
  /**
42802
43382
  * Match Input FPS
42803
43383
  * @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
@@ -42910,7 +43490,7 @@ export interface Ltx219bVideoToVideoLoraInput {
42910
43490
  video_write_mode?: 'fast' | 'balanced' | 'small';
42911
43491
  }
42912
43492
 
42913
- export interface Ltx219bVideoToVideoLoraOutput extends SharedType_4ed {}
43493
+ export interface Ltx219bVideoToVideoLoraOutput extends SharedType_3b0 {}
42914
43494
 
42915
43495
  export interface Ltx219bVideoToVideoInput {
42916
43496
  /**
@@ -43132,7 +43712,7 @@ export interface Ltx219bVideoToVideoInput {
43132
43712
  video_write_mode?: 'fast' | 'balanced' | 'small';
43133
43713
  }
43134
43714
 
43135
- export interface Ltx219bVideoToVideoOutput extends SharedType_4ed {}
43715
+ export interface Ltx219bVideoToVideoOutput extends SharedType_3b0 {}
43136
43716
 
43137
43717
  export interface Ltx219bTextToVideoLoraInput {
43138
43718
  /**
@@ -43199,7 +43779,7 @@ export interface Ltx219bTextToVideoLoraInput {
43199
43779
  * LoRAs
43200
43780
  * @description The LoRAs to use for the generation.
43201
43781
  */
43202
- loras: Components.LoRAInput[];
43782
+ loras: Components.LoRAInput_1[];
43203
43783
  /**
43204
43784
  * Negative Prompt
43205
43785
  * @description The negative prompt to generate the video from.
@@ -43279,7 +43859,7 @@ export interface Ltx219bTextToVideoLoraInput {
43279
43859
  video_write_mode?: 'fast' | 'balanced' | 'small';
43280
43860
  }
43281
43861
 
43282
- export interface Ltx219bTextToVideoLoraOutput extends SharedType_61b {}
43862
+ export interface Ltx219bTextToVideoLoraOutput extends SharedType_857 {}
43283
43863
 
43284
43864
  export interface Ltx219bTextToVideoInput {
43285
43865
  /**
@@ -43421,7 +44001,7 @@ export interface Ltx219bTextToVideoInput {
43421
44001
  video_write_mode?: 'fast' | 'balanced' | 'small';
43422
44002
  }
43423
44003
 
43424
- export interface Ltx219bTextToVideoOutput extends SharedType_61b {}
44004
+ export interface Ltx219bTextToVideoOutput extends SharedType_857 {}
43425
44005
 
43426
44006
  export interface Ltx219bImageToVideoLoraInput {
43427
44007
  /**
@@ -43518,7 +44098,7 @@ export interface Ltx219bImageToVideoLoraInput {
43518
44098
  * LoRAs
43519
44099
  * @description The LoRAs to use for the generation.
43520
44100
  */
43521
- loras: Components.LoRAInput[];
44101
+ loras: Components.LoRAInput_1[];
43522
44102
  /**
43523
44103
  * Negative Prompt
43524
44104
  * @description The negative prompt to generate the video from.
@@ -43599,7 +44179,7 @@ export interface Ltx219bImageToVideoLoraInput {
43599
44179
  video_write_mode?: 'fast' | 'balanced' | 'small';
43600
44180
  }
43601
44181
 
43602
- export interface Ltx219bImageToVideoLoraOutput extends SharedType_a00 {}
44182
+ export interface Ltx219bImageToVideoLoraOutput extends SharedType_369 {}
43603
44183
 
43604
44184
  export interface Ltx219bImageToVideoInput {
43605
44185
  /**
@@ -43772,7 +44352,7 @@ export interface Ltx219bImageToVideoInput {
43772
44352
  video_write_mode?: 'fast' | 'balanced' | 'small';
43773
44353
  }
43774
44354
 
43775
- export interface Ltx219bImageToVideoOutput extends SharedType_a00 {}
44355
+ export interface Ltx219bImageToVideoOutput extends SharedType_369 {}
43776
44356
 
43777
44357
  export interface Ltx219bExtendVideoLoraInput {
43778
44358
  /**
@@ -43863,7 +44443,7 @@ export interface Ltx219bExtendVideoLoraInput {
43863
44443
  * LoRAs
43864
44444
  * @description The LoRAs to use for the generation.
43865
44445
  */
43866
- loras: Components.LoRAInput[];
44446
+ loras: Components.LoRAInput_1[];
43867
44447
  /**
43868
44448
  * Match Input FPS
43869
44449
  * @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
@@ -43968,7 +44548,7 @@ export interface Ltx219bExtendVideoLoraInput {
43968
44548
  video_write_mode?: 'fast' | 'balanced' | 'small';
43969
44549
  }
43970
44550
 
43971
- export interface Ltx219bExtendVideoLoraOutput extends SharedType_567 {}
44551
+ export interface Ltx219bExtendVideoLoraOutput extends SharedType_cb6 {}
43972
44552
 
43973
44553
  export interface Ltx219bExtendVideoInput {
43974
44554
  /**
@@ -44159,7 +44739,7 @@ export interface Ltx219bExtendVideoInput {
44159
44739
  video_write_mode?: 'fast' | 'balanced' | 'small';
44160
44740
  }
44161
44741
 
44162
- export interface Ltx219bExtendVideoOutput extends SharedType_567 {}
44742
+ export interface Ltx219bExtendVideoOutput extends SharedType_cb6 {}
44163
44743
 
44164
44744
  export interface Ltx219bDistilledVideoToVideoLoraInput {
44165
44745
  /**
@@ -44267,7 +44847,7 @@ export interface Ltx219bDistilledVideoToVideoLoraInput {
44267
44847
  * LoRAs
44268
44848
  * @description The LoRAs to use for the generation.
44269
44849
  */
44270
- loras: Components.LoRAInput[];
44850
+ loras: Components.LoRAInput_1[];
44271
44851
  /**
44272
44852
  * Match Input FPS
44273
44853
  * @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
@@ -44374,7 +44954,7 @@ export interface Ltx219bDistilledVideoToVideoLoraInput {
44374
44954
  video_write_mode?: 'fast' | 'balanced' | 'small';
44375
44955
  }
44376
44956
 
44377
- export interface Ltx219bDistilledVideoToVideoLoraOutput extends SharedType_4ed {}
44957
+ export interface Ltx219bDistilledVideoToVideoLoraOutput extends SharedType_3b0 {}
44378
44958
 
44379
44959
  export interface Ltx219bDistilledVideoToVideoInput {
44380
44960
  /**
@@ -44584,7 +45164,7 @@ export interface Ltx219bDistilledVideoToVideoInput {
44584
45164
  video_write_mode?: 'fast' | 'balanced' | 'small';
44585
45165
  }
44586
45166
 
44587
- export interface Ltx219bDistilledVideoToVideoOutput extends SharedType_4ed {}
45167
+ export interface Ltx219bDistilledVideoToVideoOutput extends SharedType_3b0 {}
44588
45168
 
44589
45169
  export interface Ltx219bDistilledTextToVideoLoraInput {
44590
45170
  /**
@@ -44645,7 +45225,7 @@ export interface Ltx219bDistilledTextToVideoLoraInput {
44645
45225
  * LoRAs
44646
45226
  * @description The LoRAs to use for the generation.
44647
45227
  */
44648
- loras: Components.LoRAInput[];
45228
+ loras: Components.LoRAInput_1[];
44649
45229
  /**
44650
45230
  * Negative Prompt
44651
45231
  * @description The negative prompt to generate the video from.
@@ -44719,7 +45299,7 @@ export interface Ltx219bDistilledTextToVideoLoraInput {
44719
45299
  video_write_mode?: 'fast' | 'balanced' | 'small';
44720
45300
  }
44721
45301
 
44722
- export interface Ltx219bDistilledTextToVideoLoraOutput extends SharedType_61b {}
45302
+ export interface Ltx219bDistilledTextToVideoLoraOutput extends SharedType_857 {}
44723
45303
 
44724
45304
  export interface Ltx219bDistilledTextToVideoInput {
44725
45305
  /**
@@ -44849,7 +45429,7 @@ export interface Ltx219bDistilledTextToVideoInput {
44849
45429
  video_write_mode?: 'fast' | 'balanced' | 'small';
44850
45430
  }
44851
45431
 
44852
- export interface Ltx219bDistilledTextToVideoOutput extends SharedType_61b {}
45432
+ export interface Ltx219bDistilledTextToVideoOutput extends SharedType_857 {}
44853
45433
 
44854
45434
  export interface Ltx219bDistilledImageToVideoLoraInput {
44855
45435
  /**
@@ -44940,7 +45520,7 @@ export interface Ltx219bDistilledImageToVideoLoraInput {
44940
45520
  * LoRAs
44941
45521
  * @description The LoRAs to use for the generation.
44942
45522
  */
44943
- loras: Components.LoRAInput[];
45523
+ loras: Components.LoRAInput_1[];
44944
45524
  /**
44945
45525
  * Negative Prompt
44946
45526
  * @description The negative prompt to generate the video from.
@@ -45015,7 +45595,7 @@ export interface Ltx219bDistilledImageToVideoLoraInput {
45015
45595
  video_write_mode?: 'fast' | 'balanced' | 'small';
45016
45596
  }
45017
45597
 
45018
- export interface Ltx219bDistilledImageToVideoLoraOutput extends SharedType_a00 {}
45598
+ export interface Ltx219bDistilledImageToVideoLoraOutput extends SharedType_369 {}
45019
45599
 
45020
45600
  export interface Ltx219bDistilledImageToVideoInput {
45021
45601
  /**
@@ -45176,7 +45756,7 @@ export interface Ltx219bDistilledImageToVideoInput {
45176
45756
  video_write_mode?: 'fast' | 'balanced' | 'small';
45177
45757
  }
45178
45758
 
45179
- export interface Ltx219bDistilledImageToVideoOutput extends SharedType_a00 {}
45759
+ export interface Ltx219bDistilledImageToVideoOutput extends SharedType_369 {}
45180
45760
 
45181
45761
  export interface Ltx219bDistilledExtendVideoLoraInput {
45182
45762
  /**
@@ -45261,7 +45841,7 @@ export interface Ltx219bDistilledExtendVideoLoraInput {
45261
45841
  * LoRAs
45262
45842
  * @description The LoRAs to use for the generation.
45263
45843
  */
45264
- loras: Components.LoRAInput[];
45844
+ loras: Components.LoRAInput_1[];
45265
45845
  /**
45266
45846
  * Match Input FPS
45267
45847
  * @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
@@ -45360,7 +45940,7 @@ export interface Ltx219bDistilledExtendVideoLoraInput {
45360
45940
  video_write_mode?: 'fast' | 'balanced' | 'small';
45361
45941
  }
45362
45942
 
45363
- export interface Ltx219bDistilledExtendVideoLoraOutput extends SharedType_567 {}
45943
+ export interface Ltx219bDistilledExtendVideoLoraOutput extends SharedType_cb6 {}
45364
45944
 
45365
45945
  export interface Ltx219bDistilledExtendVideoInput {
45366
45946
  /**
@@ -45539,7 +46119,7 @@ export interface Ltx219bDistilledExtendVideoInput {
45539
46119
  video_write_mode?: 'fast' | 'balanced' | 'small';
45540
46120
  }
45541
46121
 
45542
- export interface Ltx219bDistilledExtendVideoOutput extends SharedType_567 {}
46122
+ export interface Ltx219bDistilledExtendVideoOutput extends SharedType_cb6 {}
45543
46123
 
45544
46124
  export interface Ltx219bDistilledAudioToVideoLoraInput {
45545
46125
  /**
@@ -45629,7 +46209,7 @@ export interface Ltx219bDistilledAudioToVideoLoraInput {
45629
46209
  * LoRAs
45630
46210
  * @description The LoRAs to use for the generation.
45631
46211
  */
45632
- loras: Components.LoRAInput[];
46212
+ loras: Components.LoRAInput_1[];
45633
46213
  /**
45634
46214
  * Match Audio Length
45635
46215
  * @description When enabled, the number of frames will be calculated based on the audio duration and FPS. When disabled, use the specified num_frames.
@@ -45984,7 +46564,7 @@ export interface Ltx219bAudioToVideoLoraInput {
45984
46564
  * LoRAs
45985
46565
  * @description The LoRAs to use for the generation.
45986
46566
  */
45987
- loras: Components.LoRAInput[];
46567
+ loras: Components.LoRAInput_1[];
45988
46568
  /**
45989
46569
  * Match Audio Length
45990
46570
  * @description When enabled, the number of frames will be calculated based on the audio duration and FPS. When disabled, use the specified num_frames.
@@ -53389,23 +53969,7 @@ export interface InfinitalkSingleTextInput {
53389
53969
  | 'Bill';
53390
53970
  }
53391
53971
 
53392
- export interface InfinitalkSingleTextOutput {
53393
- /**
53394
- * Seed
53395
- * @description The seed used for generation.
53396
- */
53397
- seed: number;
53398
- /**
53399
- * @description The generated video file.
53400
- * @example {
53401
- * "file_size": 797478,
53402
- * "file_name": "6c9dd31e1d9a4482877747a52a661a0a.mp4",
53403
- * "content_type": "application/octet-stream",
53404
- * "url": "https://v3.fal.media/files/elephant/-huMN0zTaXmBr2CqzCMps_6c9dd31e1d9a4482877747a52a661a0a.mp4"
53405
- * }
53406
- */
53407
- video: Components.File;
53408
- }
53972
+ export interface InfinitalkSingleTextOutput extends SharedType_58d {}
53409
53973
 
53410
53974
  export interface InfinitalkInput {
53411
53975
  /**
@@ -53454,23 +54018,7 @@ export interface InfinitalkInput {
53454
54018
  seed?: number;
53455
54019
  }
53456
54020
 
53457
- export interface InfinitalkOutput {
53458
- /**
53459
- * Seed
53460
- * @description The seed used for generation.
53461
- */
53462
- seed: number;
53463
- /**
53464
- * @description The generated video file.
53465
- * @example {
53466
- * "file_size": 515275,
53467
- * "file_name": "74af6c0bdd6041c3b1130d54885e3eee.mp4",
53468
- * "content_type": "application/octet-stream",
53469
- * "url": "https://v3.fal.media/files/kangaroo/z6VqUwNTwzuWa6YE1g7In_74af6c0bdd6041c3b1130d54885e3eee.mp4"
53470
- * }
53471
- */
53472
- video: Components.File;
53473
- }
54021
+ export interface InfinitalkOutput extends SharedType_0c6 {}
53474
54022
 
53475
54023
  export interface IndexTts2TextToSpeechInput {
53476
54024
  /**
@@ -53545,13 +54093,7 @@ export interface ImageutilsRembgInput {
53545
54093
  sync_mode?: boolean;
53546
54094
  }
53547
54095
 
53548
- export interface ImageutilsRembgOutput {
53549
- /**
53550
- * Image
53551
- * @description Background removed image.
53552
- */
53553
- image: Components.Image_2;
53554
- }
54096
+ export interface ImageutilsRembgOutput extends SharedType_357 {}
53555
54097
 
53556
54098
  export interface ImageutilsNsfwInput {
53557
54099
  /**
@@ -53600,7 +54142,7 @@ export interface ImageutilsMarigoldDepthInput {
53600
54142
  processing_res?: number;
53601
54143
  }
53602
54144
 
53603
- export interface ImageutilsMarigoldDepthOutput extends SharedType_973 {}
54145
+ export interface ImageutilsMarigoldDepthOutput extends SharedType_357 {}
53604
54146
 
53605
54147
  export interface ImageutilsDepthInput {
53606
54148
  /**
@@ -53629,7 +54171,7 @@ export interface ImageutilsDepthInput {
53629
54171
  image_url: string;
53630
54172
  }
53631
54173
 
53632
- export interface ImageutilsDepthOutput extends SharedType_973 {}
54174
+ export interface ImageutilsDepthOutput extends SharedType_357 {}
53633
54175
 
53634
54176
  export interface Imagen4PreviewUltraInput {
53635
54177
  /**
@@ -58006,17 +58548,17 @@ export interface Hunyuan3dV3TextTo3dOutput {
58006
58548
  /**
58007
58549
  * @description URLs for different 3D model formats
58008
58550
  * @example {
58009
- * "glb": {
58010
- * "file_size": 64724836,
58011
- * "file_name": "model.glb",
58012
- * "content_type": "model/gltf-binary",
58013
- * "url": "https://v3b.fal.media/files/b/0a8686a8/1hPquv3AqqkfnqSM9fpmB_model.glb"
58014
- * },
58015
58551
  * "obj": {
58016
58552
  * "file_size": 44084728,
58017
58553
  * "file_name": "model.obj",
58018
58554
  * "content_type": "text/plain",
58019
58555
  * "url": "https://v3b.fal.media/files/b/0a8686a8/AVgdsVFrGAKGAFr4e2g56_model.obj"
58556
+ * },
58557
+ * "glb": {
58558
+ * "file_size": 64724836,
58559
+ * "file_name": "model.glb",
58560
+ * "content_type": "model/gltf-binary",
58561
+ * "url": "https://v3b.fal.media/files/b/0a8686a8/1hPquv3AqqkfnqSM9fpmB_model.glb"
58020
58562
  * }
58021
58563
  * }
58022
58564
  */
@@ -58079,17 +58621,17 @@ export interface Hunyuan3dV3SketchTo3dOutput {
58079
58621
  /**
58080
58622
  * @description URLs for different 3D model formats
58081
58623
  * @example {
58082
- * "glb": {
58083
- * "file_size": 30655724,
58084
- * "file_name": "model.glb",
58085
- * "content_type": "model/gltf-binary",
58086
- * "url": "https://v3b.fal.media/files/b/0a8688bb/vd2SlBP92cZls3zG5EPbg_model.glb"
58087
- * },
58088
58624
  * "obj": {
58089
58625
  * "file_size": 23418473,
58090
58626
  * "file_name": "model.obj",
58091
58627
  * "content_type": "text/plain",
58092
58628
  * "url": "https://v3b.fal.media/files/b/0a8688bb/QNik1DVxzvj23YEF3vhs__model.obj"
58629
+ * },
58630
+ * "glb": {
58631
+ * "file_size": 30655724,
58632
+ * "file_name": "model.glb",
58633
+ * "content_type": "model/gltf-binary",
58634
+ * "url": "https://v3b.fal.media/files/b/0a8688bb/vd2SlBP92cZls3zG5EPbg_model.glb"
58093
58635
  * }
58094
58636
  * }
58095
58637
  */
@@ -58175,17 +58717,17 @@ export interface Hunyuan3dV3ImageTo3dOutput {
58175
58717
  /**
58176
58718
  * @description URLs for different 3D model formats
58177
58719
  * @example {
58178
- * "glb": {
58179
- * "file_size": 64122888,
58180
- * "file_name": "model.glb",
58181
- * "content_type": "model/gltf-binary",
58182
- * "url": "https://v3b.fal.media/files/b/0a8686ae/MQN_KtP32PbqtPr_VLcyp_model.glb"
58183
- * },
58184
58720
  * "obj": {
58185
58721
  * "file_size": 42886419,
58186
58722
  * "file_name": "model.obj",
58187
58723
  * "content_type": "text/plain",
58188
58724
  * "url": "https://v3b.fal.media/files/b/0a8686ad/ifdJskhUfQysq-NN20iQR_model.obj"
58725
+ * },
58726
+ * "glb": {
58727
+ * "file_size": 64122888,
58728
+ * "file_name": "model.glb",
58729
+ * "content_type": "model/gltf-binary",
58730
+ * "url": "https://v3b.fal.media/files/b/0a8686ae/MQN_KtP32PbqtPr_VLcyp_model.glb"
58189
58731
  * }
58190
58732
  * }
58191
58733
  */
@@ -59472,17 +60014,17 @@ export interface Hunyuan3dV31SmartTopologyOutput {
59472
60014
  /**
59473
60015
  * @description URLs for different 3D model formats
59474
60016
  * @example {
59475
- * "glb": {
59476
- * "file_size": 206004,
59477
- * "file_name": "model.glb",
59478
- * "content_type": "model/gltf-binary",
59479
- * "url": "https://v3b.fal.media/files/b/0a8c0ab4/eX-_x0Wv8fZL05l9CGp6Y_model.glb"
59480
- * },
59481
60017
  * "obj": {
59482
60018
  * "file_size": 394409,
59483
60019
  * "file_name": "model.obj",
59484
60020
  * "content_type": "model/obj",
59485
60021
  * "url": "https://v3b.fal.media/files/b/0a8c0ab4/tqMY5NJLnHjpwN8rQ15dj_model.obj"
60022
+ * },
60023
+ * "glb": {
60024
+ * "file_size": 206004,
60025
+ * "file_name": "model.glb",
60026
+ * "content_type": "model/gltf-binary",
60027
+ * "url": "https://v3b.fal.media/files/b/0a8c0ab4/eX-_x0Wv8fZL05l9CGp6Y_model.glb"
59486
60028
  * }
59487
60029
  * }
59488
60030
  */
@@ -59719,17 +60261,17 @@ export interface Hunyuan3dV31ProTextTo3dOutput {
59719
60261
  * "content_type": "text/plain",
59720
60262
  * "url": "https://v3b.fal.media/files/b/0a8c5482/ZxJepsEkhM67VSugmZ7QT_material.mtl"
59721
60263
  * },
59722
- * "glb": {
59723
- * "file_size": 35833072,
59724
- * "file_name": "model.glb",
59725
- * "content_type": "model/gltf-binary",
59726
- * "url": "https://v3b.fal.media/files/b/0a8c5483/z6sbpr5wBRjqgnQlJM2Ot_model.glb"
59727
- * },
59728
60264
  * "obj": {
59729
60265
  * "file_size": 34755929,
59730
60266
  * "file_name": "6168030a8817075aaa55c94cc5145000.obj",
59731
60267
  * "content_type": "model/obj",
59732
60268
  * "url": "https://v3b.fal.media/files/b/0a8c5482/ZzC1xlOftyGQxhDkbZzVW_6168030a8817075aaa55c94cc5145000.obj"
60269
+ * },
60270
+ * "glb": {
60271
+ * "file_size": 35833072,
60272
+ * "file_name": "model.glb",
60273
+ * "content_type": "model/gltf-binary",
60274
+ * "url": "https://v3b.fal.media/files/b/0a8c5483/z6sbpr5wBRjqgnQlJM2Ot_model.glb"
59733
60275
  * }
59734
60276
  * }
59735
60277
  */
@@ -59828,17 +60370,17 @@ export interface Hunyuan3dV31ProImageTo3dOutput {
59828
60370
  /**
59829
60371
  * @description URLs for different 3D model formats
59830
60372
  * @example {
59831
- * "glb": {
59832
- * "file_size": 38554640,
59833
- * "file_name": "model.glb",
59834
- * "content_type": "model/gltf-binary",
59835
- * "url": "https://v3b.fal.media/files/b/0a8c3187/jOeZmtBuhdQMkDu65AkdT_model.glb"
59836
- * },
59837
60373
  * "obj": {
59838
60374
  * "file_size": 31447160,
59839
60375
  * "file_name": "model.obj",
59840
60376
  * "content_type": "model/obj",
59841
60377
  * "url": "https://v3b.fal.media/files/b/0a8c3186/no-aBFEDnOuthILfv-wzs_model.obj"
60378
+ * },
60379
+ * "glb": {
60380
+ * "file_size": 38554640,
60381
+ * "file_name": "model.glb",
60382
+ * "content_type": "model/gltf-binary",
60383
+ * "url": "https://v3b.fal.media/files/b/0a8c3187/jOeZmtBuhdQMkDu65AkdT_model.glb"
59842
60384
  * }
59843
60385
  * }
59844
60386
  */
@@ -59941,8 +60483,8 @@ export interface Hunyuan_worldOutput {
59941
60483
  /**
59942
60484
  * @description The generated panorama image.
59943
60485
  * @example {
59944
- * "height": 960,
59945
60486
  * "file_size": 2738127,
60487
+ * "height": 960,
59946
60488
  * "file_name": "5db7925423b44f2a98098cd8f7cad7ec.png",
59947
60489
  * "content_type": "image/png",
59948
60490
  * "url": "https://v3.fal.media/files/kangaroo/P2AmXuLlyDIsivqjV_rAr_5db7925423b44f2a98098cd8f7cad7ec.png",
@@ -59991,7 +60533,7 @@ export interface HidreamI1FullImageToImageInput {
59991
60533
  * @description A list of LoRAs to apply to the model. Each LoRA specifies its path, scale, and optional weight name.
59992
60534
  * @default []
59993
60535
  */
59994
- loras?: Components.LoraWeight_3[];
60536
+ loras?: Components.LoraWeight_5[];
59995
60537
  /**
59996
60538
  * Negative Prompt
59997
60539
  * @description The negative prompt to use. Use it to address details that you don't want
@@ -60118,7 +60660,7 @@ export interface HidreamI1FullInput {
60118
60660
  * @description A list of LoRAs to apply to the model. Each LoRA specifies its path, scale, and optional weight name.
60119
60661
  * @default []
60120
60662
  */
60121
- loras?: Components.LoraWeight_3[];
60663
+ loras?: Components.LoraWeight_5[];
60122
60664
  /**
60123
60665
  * Negative Prompt
60124
60666
  * @description The negative prompt to use. Use it to address details that you don't want
@@ -60167,7 +60709,7 @@ export interface HidreamI1FullInput {
60167
60709
  sync_mode?: boolean;
60168
60710
  }
60169
60711
 
60170
- export interface HidreamI1FullOutput extends SharedType_4411 {}
60712
+ export interface HidreamI1FullOutput extends SharedType_7c6 {}
60171
60713
 
60172
60714
  export interface HidreamI1FastInput {
60173
60715
  /**
@@ -61106,6 +61648,115 @@ export interface GotOcrV2Output {
61106
61648
  outputs: string[];
61107
61649
  }
61108
61650
 
61651
+ export interface GoalForceInput {
61652
+ /**
61653
+ * Enable Safety Checker
61654
+ * @description Enable safety checker for generated content.
61655
+ * @default true
61656
+ */
61657
+ enable_safety_checker?: boolean;
61658
+ /**
61659
+ * @description Indirect goal force applied to a target object. The model generates physics-plausible interactions to move the target in the specified direction. Specify either this, projectile_force, or both.
61660
+ * @example {
61661
+ * "y": 0.18,
61662
+ * "mass": -1,
61663
+ * "angle": 0,
61664
+ * "magnitude": 350,
61665
+ * "x": 0.22
61666
+ * }
61667
+ * @example {
61668
+ * "y": 0.49,
61669
+ * "mass": 3.5,
61670
+ * "angle": 180,
61671
+ * "magnitude": 350,
61672
+ * "x": 0.57
61673
+ * }
61674
+ */
61675
+ goal_force?: Components.ForceParams;
61676
+ /**
61677
+ * Guidance Scale
61678
+ * @description Classifier-free guidance scale.
61679
+ * @default 5
61680
+ * @example 5
61681
+ */
61682
+ guidance_scale?: number;
61683
+ /**
61684
+ * Image Url
61685
+ * @description URL of the input image (first frame). The image will be resized to 832x480.
61686
+ * @default https://raw.githubusercontent.com/brown-palm/goal-force/2d0209dc27333bffdb3f06df657ae05b11407615/datasets/examples/human-object-interaction/images/_bulb_tool.png
61687
+ * @example https://raw.githubusercontent.com/brown-palm/goal-force/2d0209dc27333bffdb3f06df657ae05b11407615/datasets/examples/human-object-interaction/images/_bulb_tool.png
61688
+ */
61689
+ image_url?: string;
61690
+ /**
61691
+ * Negative Prompt
61692
+ * @description Negative prompt for generation.
61693
+ * @default 色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走
61694
+ */
61695
+ negative_prompt?: string;
61696
+ /**
61697
+ * Num Inference Steps
61698
+ * @description Number of denoising steps. Higher values produce better quality but take longer.
61699
+ * @default 50
61700
+ * @example 50
61701
+ */
61702
+ num_inference_steps?: number;
61703
+ /**
61704
+ * @description Direct force applied to a projectile object, causing it to move along the force direction. Specify either this, goal_force, or both.
61705
+ * @example {
61706
+ * "y": 0.18,
61707
+ * "mass": -1,
61708
+ * "angle": 0,
61709
+ * "magnitude": 350,
61710
+ * "x": 0.22
61711
+ * }
61712
+ */
61713
+ projectile_force?: Components.ForceParams;
61714
+ /**
61715
+ * Prompt
61716
+ * @description Text description of the scene and the desired physics interaction.
61717
+ * @default A hand pushes the hanging bulb, causing it to sway.
61718
+ * @example A hand pushes the hanging bulb, causing it to sway.
61719
+ * @example A soccer ball is kicked and rolls across the field.
61720
+ * @example A cue stick strikes a pool ball, causing it to roll across the table.
61721
+ * @example A golf club hits a golf ball.
61722
+ */
61723
+ prompt?: string;
61724
+ /**
61725
+ * Seed
61726
+ * @description Random seed for reproducibility.
61727
+ */
61728
+ seed?: number;
61729
+ }
61730
+
61731
+ export interface GoalForceOutput {
61732
+ /**
61733
+ * Seed
61734
+ * @description The seed used for generation.
61735
+ * @example 5
61736
+ */
61737
+ seed: number;
61738
+ /**
61739
+ * Timings
61740
+ * @description Timing information.
61741
+ * @example {
61742
+ * "inference": 234.79
61743
+ * }
61744
+ */
61745
+ timings: {
61746
+ [key: string]: number;
61747
+ };
61748
+ /**
61749
+ * @description The generated video.
61750
+ * @example {
61751
+ * "file_size": 502438,
61752
+ * "file_name": "tmpbrjvor4y.mp4",
61753
+ * "content_type": "application/octet-stream",
61754
+ * "url": "https://v3b.fal.media/files/b/0a91e900/S4Q1EK9vvZW_HATdXnbBo_tmpbrjvor4y.mp4"
61755
+ * }
61756
+ */
61757
+ video: Components.VideoFile;
61758
+ }
61759
+
61109
61760
  export interface GlmImageImageToImageInput {
61110
61761
  /**
61111
61762
  * Enable Prompt Expansion
@@ -61684,7 +62335,7 @@ export interface FooocusUpscaleOrVaryInput {
61684
62335
  * }
61685
62336
  * ]
61686
62337
  */
61687
- loras?: Components.LoraWeight_5[];
62338
+ loras?: Components.LoraWeight_4[];
61688
62339
  /**
61689
62340
  * Mixing Image Prompt and Vary/Upscale
61690
62341
  * @description Mixing Image Prompt and Vary/Upscale
@@ -61759,9 +62410,9 @@ export interface FooocusUpscaleOrVaryInput {
61759
62410
  * Styles
61760
62411
  * @description The style to use.
61761
62412
  * @default [
62413
+ * "Fooocus Enhance",
61762
62414
  * "Fooocus V2",
61763
- * "Fooocus Sharp",
61764
- * "Fooocus Enhance"
62415
+ * "Fooocus Sharp"
61765
62416
  * ]
61766
62417
  */
61767
62418
  styles?: (
@@ -62172,7 +62823,7 @@ export interface FooocusInpaintInput {
62172
62823
  * }
62173
62824
  * ]
62174
62825
  */
62175
- loras?: Components.LoraWeight_5[];
62826
+ loras?: Components.LoraWeight_4[];
62176
62827
  /**
62177
62828
  * Mask Image Url
62178
62829
  * @description The image to use as a mask for the generated image.
@@ -62268,9 +62919,9 @@ export interface FooocusInpaintInput {
62268
62919
  * Styles
62269
62920
  * @description The style to use.
62270
62921
  * @default [
62922
+ * "Fooocus Enhance",
62271
62923
  * "Fooocus V2",
62272
- * "Fooocus Sharp",
62273
- * "Fooocus Enhance"
62924
+ * "Fooocus Sharp"
62274
62925
  * ]
62275
62926
  */
62276
62927
  styles?: (
@@ -62585,10 +63236,10 @@ export interface FooocusImagePromptInput {
62585
63236
  guidance_scale?: number;
62586
63237
  /**
62587
63238
  * @example {
62588
- * "weight": 1,
63239
+ * "image_url": "https://storage.googleapis.com/falserverless/model_tests/fooocus/Pikachu.webp",
62589
63240
  * "stop_at": 1,
62590
63241
  * "type": "PyraCanny",
62591
- * "image_url": "https://storage.googleapis.com/falserverless/model_tests/fooocus/Pikachu.webp"
63242
+ * "weight": 1
62592
63243
  * }
62593
63244
  */
62594
63245
  image_prompt_1: Components.ImagePrompt;
@@ -62627,7 +63278,7 @@ export interface FooocusImagePromptInput {
62627
63278
  * }
62628
63279
  * ]
62629
63280
  */
62630
- loras?: Components.LoraWeight_5[];
63281
+ loras?: Components.LoraWeight_4[];
62631
63282
  /**
62632
63283
  * Mask Image URL
62633
63284
  * @description The image to use as a mask for the generated image.
@@ -62719,9 +63370,9 @@ export interface FooocusImagePromptInput {
62719
63370
  * Styles
62720
63371
  * @description The style to use.
62721
63372
  * @default [
63373
+ * "Fooocus Enhance",
62722
63374
  * "Fooocus V2",
62723
- * "Fooocus Sharp",
62724
- * "Fooocus Enhance"
63375
+ * "Fooocus Sharp"
62725
63376
  * ]
62726
63377
  */
62727
63378
  styles?: (
@@ -63098,7 +63749,7 @@ export interface FooocusInput {
63098
63749
  * }
63099
63750
  * ]
63100
63751
  */
63101
- loras?: Components.LoraWeight_5[];
63752
+ loras?: Components.LoraWeight_4[];
63102
63753
  /**
63103
63754
  * Mask Image Url
63104
63755
  * @description The image to use as a mask for the generated image.
@@ -63177,9 +63828,9 @@ export interface FooocusInput {
63177
63828
  * Styles
63178
63829
  * @description The style to use.
63179
63830
  * @default [
63831
+ * "Fooocus Enhance",
63180
63832
  * "Fooocus V2",
63181
- * "Fooocus Sharp",
63182
- * "Fooocus Enhance"
63833
+ * "Fooocus Sharp"
63183
63834
  * ]
63184
63835
  */
63185
63836
  styles?: (
@@ -64139,8 +64790,8 @@ export interface FluxVisionUpscalerOutput {
64139
64790
  /**
64140
64791
  * @description The URL of the generated image.
64141
64792
  * @example {
64142
- * "height": 2048,
64143
64793
  * "file_size": 8842156,
64794
+ * "height": 2048,
64144
64795
  * "file_name": "20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
64145
64796
  * "content_type": "image/png",
64146
64797
  * "url": "https://v3b.fal.media/files/b/panda/20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
@@ -67276,7 +67927,7 @@ export interface Flux2LoraEditInput {
67276
67927
  * @description List of LoRA weights to apply (maximum 3). Each LoRA can be a URL, HuggingFace repo ID, or local path.
67277
67928
  * @default []
67278
67929
  */
67279
- loras?: Components.LoRAInput_1[];
67930
+ loras?: Components.LoRAInput[];
67280
67931
  /**
67281
67932
  * Number of Images
67282
67933
  * @description The number of images to generate.
@@ -67394,7 +68045,7 @@ export interface Flux2LoraInput {
67394
68045
  * @description List of LoRA weights to apply (maximum 3). Each LoRA can be a URL, HuggingFace repo ID, or local path.
67395
68046
  * @default []
67396
68047
  */
67397
- loras?: Components.LoRAInput_1[];
68048
+ loras?: Components.LoRAInput[];
67398
68049
  /**
67399
68050
  * Number of Images
67400
68051
  * @description The number of images to generate.
@@ -67529,15 +68180,87 @@ export interface Flux2KleinRealtimeOutput {
67529
68180
  seed: number;
67530
68181
  }
67531
68182
 
68183
+ export interface Flux2Klein9bLoraInput {
68184
+ /**
68185
+ * Enable Safety Checker
68186
+ * @description If set to true, the safety checker will be enabled.
68187
+ * @default true
68188
+ */
68189
+ enable_safety_checker?: boolean;
68190
+ /**
68191
+ * Image Size
68192
+ * @description The size of the image to generate.
68193
+ * @default landscape_4_3
68194
+ */
68195
+ image_size?:
68196
+ | Components.ImageSize
68197
+ | (
68198
+ | 'square_hd'
68199
+ | 'square'
68200
+ | 'portrait_4_3'
68201
+ | 'portrait_16_9'
68202
+ | 'landscape_4_3'
68203
+ | 'landscape_16_9'
68204
+ );
68205
+ /**
68206
+ * Loras
68207
+ * @description List of LoRA weights to apply (maximum 3).
68208
+ * @default []
68209
+ */
68210
+ loras?: Components.LoRAInput[];
68211
+ /**
68212
+ * Number of Images
68213
+ * @description The number of images to generate.
68214
+ * @default 1
68215
+ */
68216
+ num_images?: number;
68217
+ /**
68218
+ * Number of Inference Steps
68219
+ * @description The number of inference steps to perform.
68220
+ * @default 4
68221
+ */
68222
+ num_inference_steps?: number;
68223
+ /**
68224
+ * Output Format
68225
+ * @description The format of the generated image.
68226
+ * @default png
68227
+ * @enum {string}
68228
+ */
68229
+ output_format?: 'jpeg' | 'png' | 'webp';
68230
+ /**
68231
+ * Prompt
68232
+ * @description The prompt to generate an image from.
68233
+ * @example A bright, surreal scene on a vast white salt flat under a clear blue sky, featuring two fluffy white alpacas standing in the foreground. Behind them sits a sleek white supercar with its scissor doors open, creating a dramatic futuristic silhouette. The sunlight is strong and crisp, casting sharp shadows on the ground and giving the image a clean, high-contrast cinematic look. Minimalist composition, playful luxury vibe, ultra sharp detail, wide-angle perspective, high resolution.
68234
+ */
68235
+ prompt: string;
68236
+ /**
68237
+ * Seed
68238
+ * @description The seed to use for the generation. If not provided, a random seed will be used.
68239
+ */
68240
+ seed?: number;
68241
+ /**
68242
+ * Sync Mode
68243
+ * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
68244
+ * @default false
68245
+ */
68246
+ sync_mode?: boolean;
68247
+ }
68248
+
68249
+ export interface Flux2Klein9bLoraOutput extends SharedType_a75 {}
68250
+
68251
+ export interface Flux2Klein9bEditLoraInput extends SharedType_974 {}
68252
+
68253
+ export interface Flux2Klein9bEditLoraOutput extends SharedType_1651 {}
68254
+
67532
68255
  export interface Flux2Klein9bEditInput extends SharedType_2db {}
67533
68256
 
67534
68257
  export interface Flux2Klein9bEditOutput extends SharedType_1651 {}
67535
68258
 
67536
- export interface Flux2Klein9bBaseLoraInput extends SharedType_e34 {}
68259
+ export interface Flux2Klein9bBaseLoraInput extends SharedType_bb2 {}
67537
68260
 
67538
68261
  export interface Flux2Klein9bBaseLoraOutput extends SharedType_1651 {}
67539
68262
 
67540
- export interface Flux2Klein9bBaseEditLoraInput extends SharedType_6f9 {}
68263
+ export interface Flux2Klein9bBaseEditLoraInput extends SharedType_f66 {}
67541
68264
 
67542
68265
  export interface Flux2Klein9bBaseEditLoraOutput extends SharedType_1651 {}
67543
68266
 
@@ -67614,8 +68337,119 @@ export interface Flux2Klein9bBaseEditInput {
67614
68337
  output_format?: 'jpeg' | 'png' | 'webp';
67615
68338
  /**
67616
68339
  * Prompt
67617
- * @description The prompt to edit the image.
67618
- * @example Imagine a young woman. Use the Style from the Reference Image.
68340
+ * @description The prompt to edit the image.
68341
+ * @example Imagine a young woman. Use the Style from the Reference Image.
68342
+ */
68343
+ prompt: string;
68344
+ /**
68345
+ * Seed
68346
+ * @description The seed to use for the generation. If not provided, a random seed will be used.
68347
+ */
68348
+ seed?: number;
68349
+ /**
68350
+ * Sync Mode
68351
+ * @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
68352
+ * @default false
68353
+ */
68354
+ sync_mode?: boolean;
68355
+ }
68356
+
68357
+ export interface Flux2Klein9bBaseEditOutput {
68358
+ /**
68359
+ * Has Nsfw Concepts
68360
+ * @description Whether the generated images contain NSFW concepts.
68361
+ */
68362
+ has_nsfw_concepts: boolean[];
68363
+ /**
68364
+ * Images
68365
+ * @description The edited images
68366
+ * @example [
68367
+ * {
68368
+ * "url": "https://v3b.fal.media/files/b/0a8a69f1/HpSn20CQnEgVbFRD5E-Eh.png"
68369
+ * }
68370
+ * ]
68371
+ */
68372
+ images: Components.ImageFile[];
68373
+ /**
68374
+ * Prompt
68375
+ * @description The prompt used for generating the image.
68376
+ */
68377
+ prompt: string;
68378
+ /**
68379
+ * Seed
68380
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
68381
+ * input or the randomly generated that was used in case none was passed.
68382
+ */
68383
+ seed: number;
68384
+ timings: {
68385
+ [key: string]: number;
68386
+ };
68387
+ }
68388
+
68389
+ export interface Flux2Klein9bBaseInput {
68390
+ /**
68391
+ * Acceleration
68392
+ * @description The acceleration level to use for image generation.
68393
+ * @default regular
68394
+ * @enum {string}
68395
+ */
68396
+ acceleration?: 'none' | 'regular' | 'high';
68397
+ /**
68398
+ * Enable Safety Checker
68399
+ * @description If set to true, the safety checker will be enabled.
68400
+ * @default true
68401
+ */
68402
+ enable_safety_checker?: boolean;
68403
+ /**
68404
+ * Guidance Scale
68405
+ * @description Guidance scale for classifier-free guidance.
68406
+ * @default 5
68407
+ */
68408
+ guidance_scale?: number;
68409
+ /**
68410
+ * Image Size
68411
+ * @description The size of the image to generate.
68412
+ * @default landscape_4_3
68413
+ */
68414
+ image_size?:
68415
+ | Components.ImageSize
68416
+ | (
68417
+ | 'square_hd'
68418
+ | 'square'
68419
+ | 'portrait_4_3'
68420
+ | 'portrait_16_9'
68421
+ | 'landscape_4_3'
68422
+ | 'landscape_16_9'
68423
+ );
68424
+ /**
68425
+ * Negative Prompt
68426
+ * @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
68427
+ * @default
68428
+ */
68429
+ negative_prompt?: string;
68430
+ /**
68431
+ * Number of Images
68432
+ * @description The number of images to generate.
68433
+ * @default 1
68434
+ */
68435
+ num_images?: number;
68436
+ /**
68437
+ * Number of Inference Steps
68438
+ * @description The number of inference steps to perform.
68439
+ * @default 28
68440
+ */
68441
+ num_inference_steps?: number;
68442
+ /**
68443
+ * Output Format
68444
+ * @description The format of the generated image.
68445
+ * @default png
68446
+ * @enum {string}
68447
+ */
68448
+ output_format?: 'jpeg' | 'png' | 'webp';
68449
+ /**
68450
+ * Prompt
68451
+ * @description The prompt to generate an image from.
68452
+ * @example A cyberpunk samurai standing in neon-lit Tokyo street at night, rain-soaked pavement reflecting holographic advertisements in pink and blue, steam rising from street vents, wearing futuristic armor with glowing accents, cinematic composition, Unreal Engine 5 quality, ray-traced reflections, shallow depth of field, hyper-detailed
67619
68453
  */
67620
68454
  prompt: string;
67621
68455
  /**
@@ -67631,7 +68465,7 @@ export interface Flux2Klein9bBaseEditInput {
67631
68465
  sync_mode?: boolean;
67632
68466
  }
67633
68467
 
67634
- export interface Flux2Klein9bBaseEditOutput {
68468
+ export interface Flux2Klein9bBaseOutput {
67635
68469
  /**
67636
68470
  * Has Nsfw Concepts
67637
68471
  * @description Whether the generated images contain NSFW concepts.
@@ -67639,10 +68473,10 @@ export interface Flux2Klein9bBaseEditOutput {
67639
68473
  has_nsfw_concepts: boolean[];
67640
68474
  /**
67641
68475
  * Images
67642
- * @description The edited images
68476
+ * @description The generated images
67643
68477
  * @example [
67644
68478
  * {
67645
- * "url": "https://v3b.fal.media/files/b/0a8a69f1/HpSn20CQnEgVbFRD5E-Eh.png"
68479
+ * "url": "https://v3b.fal.media/files/b/0a8a69bc/EaxKO6wroq3eaDb2Znfpo.png"
67646
68480
  * }
67647
68481
  * ]
67648
68482
  */
@@ -67663,26 +68497,13 @@ export interface Flux2Klein9bBaseEditOutput {
67663
68497
  };
67664
68498
  }
67665
68499
 
67666
- export interface Flux2Klein9bBaseInput {
67667
- /**
67668
- * Acceleration
67669
- * @description The acceleration level to use for image generation.
67670
- * @default regular
67671
- * @enum {string}
67672
- */
67673
- acceleration?: 'none' | 'regular' | 'high';
68500
+ export interface Flux2Klein9bInput {
67674
68501
  /**
67675
68502
  * Enable Safety Checker
67676
68503
  * @description If set to true, the safety checker will be enabled.
67677
68504
  * @default true
67678
68505
  */
67679
68506
  enable_safety_checker?: boolean;
67680
- /**
67681
- * Guidance Scale
67682
- * @description Guidance scale for classifier-free guidance.
67683
- * @default 5
67684
- */
67685
- guidance_scale?: number;
67686
68507
  /**
67687
68508
  * Image Size
67688
68509
  * @description The size of the image to generate.
@@ -67698,12 +68519,6 @@ export interface Flux2Klein9bBaseInput {
67698
68519
  | 'landscape_4_3'
67699
68520
  | 'landscape_16_9'
67700
68521
  );
67701
- /**
67702
- * Negative Prompt
67703
- * @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
67704
- * @default
67705
- */
67706
- negative_prompt?: string;
67707
68522
  /**
67708
68523
  * Number of Images
67709
68524
  * @description The number of images to generate.
@@ -67713,7 +68528,7 @@ export interface Flux2Klein9bBaseInput {
67713
68528
  /**
67714
68529
  * Number of Inference Steps
67715
68530
  * @description The number of inference steps to perform.
67716
- * @default 28
68531
+ * @default 4
67717
68532
  */
67718
68533
  num_inference_steps?: number;
67719
68534
  /**
@@ -67726,7 +68541,7 @@ export interface Flux2Klein9bBaseInput {
67726
68541
  /**
67727
68542
  * Prompt
67728
68543
  * @description The prompt to generate an image from.
67729
- * @example A cyberpunk samurai standing in neon-lit Tokyo street at night, rain-soaked pavement reflecting holographic advertisements in pink and blue, steam rising from street vents, wearing futuristic armor with glowing accents, cinematic composition, Unreal Engine 5 quality, ray-traced reflections, shallow depth of field, hyper-detailed
68544
+ * @example A bright, surreal scene on a vast white salt flat under a clear blue sky, featuring two fluffy white alpacas standing in the foreground. Behind them sits a sleek white supercar with its scissor doors open, creating a dramatic futuristic silhouette. The sunlight is strong and crisp, casting sharp shadows on the ground and giving the image a clean, high-contrast cinematic look. Minimalist composition, playful luxury vibe, ultra sharp detail, wide-angle perspective, high resolution.
67730
68545
  */
67731
68546
  prompt: string;
67732
68547
  /**
@@ -67742,39 +68557,9 @@ export interface Flux2Klein9bBaseInput {
67742
68557
  sync_mode?: boolean;
67743
68558
  }
67744
68559
 
67745
- export interface Flux2Klein9bBaseOutput {
67746
- /**
67747
- * Has Nsfw Concepts
67748
- * @description Whether the generated images contain NSFW concepts.
67749
- */
67750
- has_nsfw_concepts: boolean[];
67751
- /**
67752
- * Images
67753
- * @description The generated images
67754
- * @example [
67755
- * {
67756
- * "url": "https://v3b.fal.media/files/b/0a8a69bc/EaxKO6wroq3eaDb2Znfpo.png"
67757
- * }
67758
- * ]
67759
- */
67760
- images: Components.ImageFile[];
67761
- /**
67762
- * Prompt
67763
- * @description The prompt used for generating the image.
67764
- */
67765
- prompt: string;
67766
- /**
67767
- * Seed
67768
- * @description Seed of the generated Image. It will be the same value of the one passed in the
67769
- * input or the randomly generated that was used in case none was passed.
67770
- */
67771
- seed: number;
67772
- timings: {
67773
- [key: string]: number;
67774
- };
67775
- }
68560
+ export interface Flux2Klein9bOutput extends SharedType_a75 {}
67776
68561
 
67777
- export interface Flux2Klein9bInput {
68562
+ export interface Flux2Klein4bLoraInput {
67778
68563
  /**
67779
68564
  * Enable Safety Checker
67780
68565
  * @description If set to true, the safety checker will be enabled.
@@ -67796,6 +68581,12 @@ export interface Flux2Klein9bInput {
67796
68581
  | 'landscape_4_3'
67797
68582
  | 'landscape_16_9'
67798
68583
  );
68584
+ /**
68585
+ * Loras
68586
+ * @description List of LoRA weights to apply (maximum 3).
68587
+ * @default []
68588
+ */
68589
+ loras?: Components.LoRAInput[];
67799
68590
  /**
67800
68591
  * Number of Images
67801
68592
  * @description The number of images to generate.
@@ -67818,7 +68609,7 @@ export interface Flux2Klein9bInput {
67818
68609
  /**
67819
68610
  * Prompt
67820
68611
  * @description The prompt to generate an image from.
67821
- * @example A bright, surreal scene on a vast white salt flat under a clear blue sky, featuring two fluffy white alpacas standing in the foreground. Behind them sits a sleek white supercar with its scissor doors open, creating a dramatic futuristic silhouette. The sunlight is strong and crisp, casting sharp shadows on the ground and giving the image a clean, high-contrast cinematic look. Minimalist composition, playful luxury vibe, ultra sharp detail, wide-angle perspective, high resolution.
68612
+ * @example Cheetah in pre-sprint crouch, muscles tensed, gazelle visible in distance, African savanna golden grass, dust particles in air, National Geographic wildlife photography
67822
68613
  */
67823
68614
  prompt: string;
67824
68615
  /**
@@ -67834,77 +68625,21 @@ export interface Flux2Klein9bInput {
67834
68625
  sync_mode?: boolean;
67835
68626
  }
67836
68627
 
67837
- export interface Flux2Klein9bOutput {
67838
- /**
67839
- * Has Nsfw Concepts
67840
- * @description Whether the generated images contain NSFW concepts.
67841
- */
67842
- has_nsfw_concepts: boolean[];
67843
- /**
67844
- * Images
67845
- * @description The generated images
67846
- * @example [
67847
- * {
67848
- * "url": "https://v3b.fal.media/files/b/0a8a6905/DuNRy1OODaGrlUGEfl9SX.png"
67849
- * }
67850
- * ]
67851
- */
67852
- images: Components.ImageFile[];
67853
- /**
67854
- * Prompt
67855
- * @description The prompt used for generating the image.
67856
- */
67857
- prompt: string;
67858
- /**
67859
- * Seed
67860
- * @description Seed of the generated Image. It will be the same value of the one passed in the
67861
- * input or the randomly generated that was used in case none was passed.
67862
- */
67863
- seed: number;
67864
- timings: {
67865
- [key: string]: number;
67866
- };
67867
- }
68628
+ export interface Flux2Klein4bLoraOutput extends SharedType_d72 {}
68629
+
68630
+ export interface Flux2Klein4bEditLoraInput extends SharedType_974 {}
68631
+
68632
+ export interface Flux2Klein4bEditLoraOutput extends SharedType_938 {}
67868
68633
 
67869
68634
  export interface Flux2Klein4bEditInput extends SharedType_2db {}
67870
68635
 
67871
- export interface Flux2Klein4bEditOutput {
67872
- /**
67873
- * Has Nsfw Concepts
67874
- * @description Whether the generated images contain NSFW concepts.
67875
- */
67876
- has_nsfw_concepts: boolean[];
67877
- /**
67878
- * Images
67879
- * @description The edited images
67880
- * @example [
67881
- * {
67882
- * "url": "https://v3b.fal.media/files/b/0a8a69d6/M73KvDgfEgIM77t4mFsS2.png"
67883
- * }
67884
- * ]
67885
- */
67886
- images: Components.ImageFile[];
67887
- /**
67888
- * Prompt
67889
- * @description The prompt used for generating the image.
67890
- */
67891
- prompt: string;
67892
- /**
67893
- * Seed
67894
- * @description Seed of the generated Image. It will be the same value of the one passed in the
67895
- * input or the randomly generated that was used in case none was passed.
67896
- */
67897
- seed: number;
67898
- timings: {
67899
- [key: string]: number;
67900
- };
67901
- }
68636
+ export interface Flux2Klein4bEditOutput extends SharedType_938 {}
67902
68637
 
67903
- export interface Flux2Klein4bBaseLoraInput extends SharedType_e34 {}
68638
+ export interface Flux2Klein4bBaseLoraInput extends SharedType_bb2 {}
67904
68639
 
67905
68640
  export interface Flux2Klein4bBaseLoraOutput extends SharedType_1651 {}
67906
68641
 
67907
- export interface Flux2Klein4bBaseEditLoraInput extends SharedType_6f9 {}
68642
+ export interface Flux2Klein4bBaseEditLoraInput extends SharedType_f66 {}
67908
68643
 
67909
68644
  export interface Flux2Klein4bBaseEditLoraOutput extends SharedType_1651 {}
67910
68645
 
@@ -68201,37 +68936,7 @@ export interface Flux2Klein4bInput {
68201
68936
  sync_mode?: boolean;
68202
68937
  }
68203
68938
 
68204
- export interface Flux2Klein4bOutput {
68205
- /**
68206
- * Has Nsfw Concepts
68207
- * @description Whether the generated images contain NSFW concepts.
68208
- */
68209
- has_nsfw_concepts: boolean[];
68210
- /**
68211
- * Images
68212
- * @description The generated images
68213
- * @example [
68214
- * {
68215
- * "url": "https://v3b.fal.media/files/b/0a8a69c7/gU9ZgfFC9oAZjpIoveAac.png"
68216
- * }
68217
- * ]
68218
- */
68219
- images: Components.ImageFile[];
68220
- /**
68221
- * Prompt
68222
- * @description The prompt used for generating the image.
68223
- */
68224
- prompt: string;
68225
- /**
68226
- * Seed
68227
- * @description Seed of the generated Image. It will be the same value of the one passed in the
68228
- * input or the randomly generated that was used in case none was passed.
68229
- */
68230
- seed: number;
68231
- timings: {
68232
- [key: string]: number;
68233
- };
68234
- }
68939
+ export interface Flux2Klein4bOutput extends SharedType_d72 {}
68235
68940
 
68236
68941
  export interface Flux2FlashEditInput {
68237
68942
  /**
@@ -71090,15 +71795,15 @@ export interface FloweditOutput {
71090
71795
  seed: number;
71091
71796
  }
71092
71797
 
71093
- export interface Florence2LargeRegionToSegmentationInput extends SharedType_d661 {}
71798
+ export interface Florence2LargeRegionToSegmentationInput extends SharedType_2ff {}
71094
71799
 
71095
71800
  export interface Florence2LargeRegionToSegmentationOutput extends SharedType_4aa {}
71096
71801
 
71097
- export interface Florence2LargeRegionToDescriptionInput extends SharedType_d661 {}
71802
+ export interface Florence2LargeRegionToDescriptionInput extends SharedType_2ff {}
71098
71803
 
71099
71804
  export interface Florence2LargeRegionToDescriptionOutput extends SharedType_129 {}
71100
71805
 
71101
- export interface Florence2LargeRegionToCategoryInput extends SharedType_d661 {}
71806
+ export interface Florence2LargeRegionToCategoryInput extends SharedType_2ff {}
71102
71807
 
71103
71808
  export interface Florence2LargeRegionToCategoryOutput extends SharedType_129 {}
71104
71809
 
@@ -72105,7 +72810,19 @@ export interface FastSvdTextToVideoInput {
72105
72810
  );
72106
72811
  }
72107
72812
 
72108
- export interface FastSvdTextToVideoOutput extends SharedType_484 {}
72813
+ export interface FastSvdTextToVideoOutput {
72814
+ /**
72815
+ * Seed
72816
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
72817
+ * input or the randomly generated that was used in case none was passed.
72818
+ */
72819
+ seed: number;
72820
+ /**
72821
+ * Video
72822
+ * @description The generated video file.
72823
+ */
72824
+ video: Components.File_1;
72825
+ }
72109
72826
 
72110
72827
  export interface FastSvdLcmTextToVideoInput {
72111
72828
  /**
@@ -72167,7 +72884,7 @@ export interface FastSvdLcmTextToVideoInput {
72167
72884
  );
72168
72885
  }
72169
72886
 
72170
- export interface FastSvdLcmTextToVideoOutput extends SharedType_484 {}
72887
+ export interface FastSvdLcmTextToVideoOutput extends SharedType_b23 {}
72171
72888
 
72172
72889
  export interface FastSvdLcmInput {
72173
72890
  /**
@@ -72217,7 +72934,7 @@ export interface FastSvdLcmInput {
72217
72934
  steps?: number;
72218
72935
  }
72219
72936
 
72220
- export interface FastSvdLcmOutput extends SharedType_484 {}
72937
+ export interface FastSvdLcmOutput extends SharedType_b23 {}
72221
72938
 
72222
72939
  export interface FastSdxlInpaintingInput {
72223
72940
  /**
@@ -72547,7 +73264,7 @@ export interface FastSdxlControlnetCannyInpaintingInput {
72547
73264
  * @description The list of LoRA weights to use.
72548
73265
  * @default []
72549
73266
  */
72550
- loras?: Components.LoraWeight_4[];
73267
+ loras?: Components.LoraWeight_3[];
72551
73268
  /**
72552
73269
  * Mask Url
72553
73270
  * @description The URL of the mask to use for inpainting.
@@ -72662,7 +73379,7 @@ export interface FastSdxlControlnetCannyImageToImageInput {
72662
73379
  * @description The list of LoRA weights to use.
72663
73380
  * @default []
72664
73381
  */
72665
- loras?: Components.LoraWeight_4[];
73382
+ loras?: Components.LoraWeight_3[];
72666
73383
  /**
72667
73384
  * Negative Prompt
72668
73385
  * @description The negative prompt to use.Use it to address details that you don't want
@@ -72771,7 +73488,7 @@ export interface FastSdxlControlnetCannyInput {
72771
73488
  * @description The list of LoRA weights to use.
72772
73489
  * @default []
72773
73490
  */
72774
- loras?: Components.LoraWeight_4[];
73491
+ loras?: Components.LoraWeight_3[];
72775
73492
  /**
72776
73493
  * Negative Prompt
72777
73494
  * @description The negative prompt to use. Use it to address details that you don't want
@@ -74505,11 +75222,11 @@ export interface F5TtsOutput {
74505
75222
  audio_url: Components.AudioFile_2;
74506
75223
  }
74507
75224
 
74508
- export interface FLiteTextureInput extends SharedType_b7d {}
75225
+ export interface FLiteTextureInput extends SharedType_219 {}
74509
75226
 
74510
75227
  export interface FLiteTextureOutput extends SharedType_7c6 {}
74511
75228
 
74512
- export interface FLiteStandardInput extends SharedType_b7d {}
75229
+ export interface FLiteStandardInput extends SharedType_219 {}
74513
75230
 
74514
75231
  export interface FLiteStandardOutput extends SharedType_7c6 {}
74515
75232
 
@@ -77474,7 +78191,7 @@ export interface Cogvideox5bVideoToVideoInput {
77474
78191
  video_url: string;
77475
78192
  }
77476
78193
 
77477
- export interface Cogvideox5bVideoToVideoOutput extends SharedType_100 {}
78194
+ export interface Cogvideox5bVideoToVideoOutput extends SharedType_1001 {}
77478
78195
 
77479
78196
  export interface Cogvideox5bImageToVideoInput {
77480
78197
  /**
@@ -77553,7 +78270,7 @@ export interface Cogvideox5bImageToVideoInput {
77553
78270
  );
77554
78271
  }
77555
78272
 
77556
- export interface Cogvideox5bImageToVideoOutput extends SharedType_100 {}
78273
+ export interface Cogvideox5bImageToVideoOutput extends SharedType_1001 {}
77557
78274
 
77558
78275
  export interface Cogvideox5bInput {
77559
78276
  /**
@@ -77626,7 +78343,7 @@ export interface Cogvideox5bInput {
77626
78343
  );
77627
78344
  }
77628
78345
 
77629
- export interface Cogvideox5bOutput extends SharedType_100 {}
78346
+ export interface Cogvideox5bOutput extends SharedType_1001 {}
77630
78347
 
77631
78348
  export interface CodeformerInput {
77632
78349
  /**
@@ -78108,7 +78825,6 @@ export interface ChatterboxTextToSpeechMultilingualInput {
78108
78825
  /**
78109
78826
  * Custom Audio Language
78110
78827
  * @description If using a custom audio URL, specify the language of the audio here. Ignored if voice is not a custom url.
78111
- * @enum {string}
78112
78828
  */
78113
78829
  custom_audio_language?:
78114
78830
  | 'english'
@@ -78191,13 +78907,12 @@ export interface ChatterboxTextToSpeechMultilingualInput {
78191
78907
 
78192
78908
  export interface ChatterboxTextToSpeechMultilingualOutput {
78193
78909
  /**
78194
- * Audio
78195
78910
  * @description The generated multilingual speech audio file
78196
78911
  * @example {
78197
78912
  * "url": "https://v3.fal.media/files/example/multilingual_speech_output.wav"
78198
78913
  * }
78199
78914
  */
78200
- audio: Components.File_1;
78915
+ audio: Components.File;
78201
78916
  }
78202
78917
 
78203
78918
  export interface ChatterboxTextToSpeechInput {
@@ -78237,7 +78952,7 @@ export interface ChatterboxTextToSpeechInput {
78237
78952
  text: string;
78238
78953
  }
78239
78954
 
78240
- export interface ChatterboxTextToSpeechOutput extends SharedType_4411 {}
78955
+ export interface ChatterboxTextToSpeechOutput extends SharedType_d2a {}
78241
78956
 
78242
78957
  export interface ChatterboxSpeechToSpeechInput {
78243
78958
  /**
@@ -78253,16 +78968,7 @@ export interface ChatterboxSpeechToSpeechInput {
78253
78968
  target_voice_audio_url?: string;
78254
78969
  }
78255
78970
 
78256
- export interface ChatterboxSpeechToSpeechOutput {
78257
- /**
78258
- * Audio
78259
- * @description The generated speech audio
78260
- * @example {
78261
- * "url": "https://v3.fal.media/files/kangaroo/RQ_pxc7oPdueYqWUqEbPE_tmpjnzvvzx_.wav"
78262
- * }
78263
- */
78264
- audio: Components.File_1;
78265
- }
78971
+ export interface ChatterboxSpeechToSpeechOutput extends SharedType_d2a {}
78266
78972
 
78267
78973
  export interface ChainOfZoomInput {
78268
78974
  /**
@@ -79868,6 +80574,84 @@ export interface BytedanceSeedanceV15ProImageToVideoOutput {
79868
80574
  video: Components.File;
79869
80575
  }
79870
80576
 
80577
+ export interface BytedanceSeedV2MiniInput {
80578
+ /**
80579
+ * Image Urls
80580
+ * @description URLs of images for visual understanding. Supported formats: JPEG, PNG, WebP. A maximum of 6 images is supported. Any additional images will be ignored.
80581
+ */
80582
+ image_urls?: string[];
80583
+ /**
80584
+ * Max Completion Tokens
80585
+ * @description Controls the maximum length of the model's output, including both the model's response and its chain-of-thought content, measured in tokens.
80586
+ * @default 4096
80587
+ */
80588
+ max_completion_tokens?: number;
80589
+ /**
80590
+ * Messages
80591
+ * @description Optional prior conversation history for multi-turn conversations. Pass back the `messages` field from a previous response to provide context. The current `prompt`, `image_urls`, `video_urls`, and `system_prompt` are always appended as the latest user turn.
80592
+ * @example []
80593
+ */
80594
+ messages?: Components.Seed2MiniMessage[];
80595
+ /**
80596
+ * Prompt
80597
+ * @description The text prompt or question for the model.
80598
+ * @example What can you do?
80599
+ */
80600
+ prompt: string;
80601
+ /**
80602
+ * Reasoning Effort
80603
+ * @description Controls the depth of reasoning before the model responds. Only applicable when `thinking` is `enabled` or `auto`. `minimal` for immediate response, `low` for faster response with light reasoning, `medium` for balanced speed and depth, `high` for deep analysis of complex issues.
80604
+ */
80605
+ reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
80606
+ /**
80607
+ * System Prompt
80608
+ * @description Optional system prompt to guide the model's behavior.
80609
+ */
80610
+ system_prompt?: string;
80611
+ /**
80612
+ * Temperature
80613
+ * @description Controls randomness in the response. Lower values make output more focused and deterministic, higher values make it more creative.
80614
+ * @default 1
80615
+ */
80616
+ temperature?: number;
80617
+ /**
80618
+ * Thinking
80619
+ * @description Controls the model's chain-of-thought reasoning. `enabled` always includes reasoning, `disabled` never includes reasoning, `auto` lets the model decide based on the query.
80620
+ * @default enabled
80621
+ * @enum {string}
80622
+ */
80623
+ thinking?: 'enabled' | 'disabled' | 'auto';
80624
+ /**
80625
+ * Top P
80626
+ * @description Nucleus sampling parameter. The model considers tokens with top_p cumulative probability mass. Lower values narrow the token selection.
80627
+ * @default 0.7
80628
+ */
80629
+ top_p?: number;
80630
+ /**
80631
+ * Video Urls
80632
+ * @description URLs of videos for video understanding. Supported formats: MP4, MOV. Audio comprehension is not supported. A maximum of 3 videos is supported. Any additional videos will be ignored.
80633
+ */
80634
+ video_urls?: string[];
80635
+ }
80636
+
80637
+ export interface BytedanceSeedV2MiniOutput {
80638
+ /**
80639
+ * Messages
80640
+ * @description The full conversation history including the model's response. Pass this back as the `messages` input field to continue the conversation.
80641
+ */
80642
+ messages: Components.Seed2MiniMessage[];
80643
+ /**
80644
+ * Output
80645
+ * @description The model's text response.
80646
+ */
80647
+ output: string;
80648
+ /**
80649
+ * Reasoning Content
80650
+ * @description The model's chain-of-thought reasoning content. Only present when `thinking` is `enabled` or `auto`.
80651
+ */
80652
+ reasoning_content?: string;
80653
+ }
80654
+
79871
80655
  export interface BytedanceOmnihumanV15Input {
79872
80656
  /**
79873
80657
  * Audio Url
@@ -80086,7 +80870,7 @@ export interface BytedanceUpscalerUpscaleVideoOutput {
80086
80870
 
80087
80871
  export interface BriaTextToImageHdInput extends SharedType_411 {}
80088
80872
 
80089
- export interface BriaTextToImageHdOutput extends SharedType_e19 {}
80873
+ export interface BriaTextToImageHdOutput extends SharedType_a97 {}
80090
80874
 
80091
80875
  export interface BriaTextToImageFastInput {
80092
80876
  /**
@@ -80159,11 +80943,11 @@ export interface BriaTextToImageFastInput {
80159
80943
  sync_mode?: boolean;
80160
80944
  }
80161
80945
 
80162
- export interface BriaTextToImageFastOutput extends SharedType_e19 {}
80946
+ export interface BriaTextToImageFastOutput extends SharedType_a97 {}
80163
80947
 
80164
80948
  export interface BriaTextToImageBaseInput extends SharedType_411 {}
80165
80949
 
80166
- export interface BriaTextToImageBaseOutput extends SharedType_e19 {}
80950
+ export interface BriaTextToImageBaseOutput extends SharedType_a97 {}
80167
80951
 
80168
80952
  export interface BriaReimagineInput {
80169
80953
  /**
@@ -80392,8 +81176,8 @@ export interface BriaGenfillOutput {
80392
81176
  * @description Generated Images
80393
81177
  * @example [
80394
81178
  * {
80395
- * "height": 768,
80396
81179
  * "file_size": 1064550,
81180
+ * "height": 768,
80397
81181
  * "file_name": "a0d138e6820c4ad58f1fd3c758f16047.png",
80398
81182
  * "content_type": "image/png",
80399
81183
  * "url": "https://storage.googleapis.com/falserverless/bria/bria_genfill_res.png",
@@ -80477,8 +81261,8 @@ export interface BriaExpandOutput {
80477
81261
  /**
80478
81262
  * @description The generated image
80479
81263
  * @example {
80480
- * "height": 674,
80481
81264
  * "file_size": 1471342,
81265
+ * "height": 674,
80482
81266
  * "file_name": "afa402a35ea742cdb5c3e219b2b19bfb.png",
80483
81267
  * "content_type": "image/png",
80484
81268
  * "url": "https://v3.fal.media/files/koala/8np-spgxxG-I1r3cjthRV_afa402a35ea742cdb5c3e219b2b19bfb.png",
@@ -80634,8 +81418,8 @@ export interface BriaBackgroundRemoveOutput {
80634
81418
  /**
80635
81419
  * @description The generated image
80636
81420
  * @example {
80637
- * "height": 1024,
80638
81421
  * "file_size": 1076276,
81422
+ * "height": 1024,
80639
81423
  * "file_name": "070c731993e949d993c10ef6283d335d.png",
80640
81424
  * "content_type": "image/png",
80641
81425
  * "url": "https://v3.fal.media/files/tiger/GQEMNjRyxSoza7N8LPPqb_070c731993e949d993c10ef6283d335d.png",
@@ -80837,8 +81621,8 @@ export interface BirefnetV2VideoOutput {
80837
81621
  * "width": 1920,
80838
81622
  * "fps": 24,
80839
81623
  * "file_name": "birefnet-video-output.webm",
80840
- * "content_type": "video/webm",
80841
- * "num_frames": 192
81624
+ * "num_frames": 192,
81625
+ * "content_type": "video/webm"
80842
81626
  * }
80843
81627
  */
80844
81628
  video: Components.VideoFile;
@@ -81685,24 +82469,7 @@ export interface AiAvatarSingleTextInput {
81685
82469
  | 'Bill';
81686
82470
  }
81687
82471
 
81688
- export interface AiAvatarSingleTextOutput {
81689
- /**
81690
- * Seed
81691
- * @description The seed used for generation.
81692
- */
81693
- seed: number;
81694
- /**
81695
- * Video
81696
- * @description The generated video file.
81697
- * @example {
81698
- * "file_size": 797478,
81699
- * "file_name": "6c9dd31e1d9a4482877747a52a661a0a.mp4",
81700
- * "content_type": "application/octet-stream",
81701
- * "url": "https://v3.fal.media/files/elephant/-huMN0zTaXmBr2CqzCMps_6c9dd31e1d9a4482877747a52a661a0a.mp4"
81702
- * }
81703
- */
81704
- video: Components.File_1;
81705
- }
82472
+ export interface AiAvatarSingleTextOutput extends SharedType_58d {}
81706
82473
 
81707
82474
  export interface AiAvatarMultiTextInput {
81708
82475
  /**
@@ -81818,7 +82585,6 @@ export interface AiAvatarMultiTextOutput {
81818
82585
  */
81819
82586
  seed: number;
81820
82587
  /**
81821
- * Video
81822
82588
  * @description The generated video file.
81823
82589
  * @example {
81824
82590
  * "file_size": 352679,
@@ -81827,7 +82593,7 @@ export interface AiAvatarMultiTextOutput {
81827
82593
  * "url": "https://v3.fal.media/files/zebra/lKMkUvzCqKn-gHC0vyUPP_30b76b90c2164f9a926527497c20832b.mp4"
81828
82594
  * }
81829
82595
  */
81830
- video: Components.File_1;
82596
+ video: Components.File;
81831
82597
  }
81832
82598
 
81833
82599
  export interface AiAvatarMultiInput {
@@ -81896,7 +82662,6 @@ export interface AiAvatarMultiOutput {
81896
82662
  */
81897
82663
  seed: number;
81898
82664
  /**
81899
- * Video
81900
82665
  * @description The generated video file.
81901
82666
  * @example {
81902
82667
  * "file_size": 704757,
@@ -81905,7 +82670,7 @@ export interface AiAvatarMultiOutput {
81905
82670
  * "url": "https://v3.fal.media/files/kangaroo/uAF7N-Ow8WwuvbFw8J4Br_ab27ac57e9464dbea1ef78f7a25469d2.mp4"
81906
82671
  * }
81907
82672
  */
81908
- video: Components.File_1;
82673
+ video: Components.File;
81909
82674
  }
81910
82675
 
81911
82676
  export interface AiAvatarInput {
@@ -81955,24 +82720,7 @@ export interface AiAvatarInput {
81955
82720
  seed?: number;
81956
82721
  }
81957
82722
 
81958
- export interface AiAvatarOutput {
81959
- /**
81960
- * Seed
81961
- * @description The seed used for generation.
81962
- */
81963
- seed: number;
81964
- /**
81965
- * Video
81966
- * @description The generated video file.
81967
- * @example {
81968
- * "file_size": 515275,
81969
- * "file_name": "74af6c0bdd6041c3b1130d54885e3eee.mp4",
81970
- * "content_type": "application/octet-stream",
81971
- * "url": "https://v3.fal.media/files/kangaroo/z6VqUwNTwzuWa6YE1g7In_74af6c0bdd6041c3b1130d54885e3eee.mp4"
81972
- * }
81973
- */
81974
- video: Components.File_1;
81975
- }
82723
+ export interface AiAvatarOutput extends SharedType_0c6 {}
81976
82724
 
81977
82725
  export interface AceStepPromptToAudioInput {
81978
82726
  /**
@@ -82837,8 +83585,8 @@ export interface ClarityaiCrystalVideoUpscalerOutput {
82837
83585
  * "fps": 23.130193905817176,
82838
83586
  * "width": 4096,
82839
83587
  * "file_name": "w0VQQvPdwvV2GSCtRTMzh_hDH8SPrB.mp4",
82840
- * "num_frames": 302,
82841
- * "content_type": "video/mp4"
83588
+ * "content_type": "video/mp4",
83589
+ * "num_frames": 302
82842
83590
  * }
82843
83591
  */
82844
83592
  video: Components.VideoFile;