fal-endpoint-types 1.3.20 → 1.3.22

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -31,20 +31,19 @@ export interface XaiGrokImagineVideoTextToVideoInput {
31
31
 
32
32
  export interface XaiGrokImagineVideoTextToVideoOutput {
33
33
  /**
34
- * Video
35
34
  * @description The generated video.
36
35
  * @example {
37
36
  * "height": 720,
38
37
  * "duration": 6.041667,
39
38
  * "url": "https://v3b.fal.media/files/b/0a8b90e4/RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
40
- * "width": 1280,
41
39
  * "fps": 24,
40
+ * "width": 1280,
42
41
  * "file_name": "RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
43
42
  * "content_type": "video/mp4",
44
43
  * "num_frames": 145
45
44
  * }
46
45
  */
47
- video: Components.VideoFile;
46
+ video: Components.VideoFile_1;
48
47
  }
49
48
 
50
49
  export interface XaiGrokImagineVideoImageToVideoInput {
@@ -52,7 +51,6 @@ export interface XaiGrokImagineVideoImageToVideoInput {
52
51
  * Aspect Ratio
53
52
  * @description Aspect ratio of the generated video.
54
53
  * @default auto
55
- * @enum {string}
56
54
  */
57
55
  aspect_ratio?: 'auto' | '16:9' | '4:3' | '3:2' | '1:1' | '2:3' | '3:4' | '9:16';
58
56
  /**
@@ -84,20 +82,19 @@ export interface XaiGrokImagineVideoImageToVideoInput {
84
82
 
85
83
  export interface XaiGrokImagineVideoImageToVideoOutput {
86
84
  /**
87
- * Video
88
85
  * @description The generated video.
89
86
  * @example {
90
87
  * "height": 720,
91
88
  * "duration": 6.041667,
92
89
  * "url": "https://v3b.fal.media/files/b/0a8b90e0/0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
93
- * "width": 1280,
94
90
  * "fps": 24,
91
+ * "width": 1280,
95
92
  * "file_name": "0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
96
93
  * "content_type": "video/mp4",
97
94
  * "num_frames": 145
98
95
  * }
99
96
  */
100
- video: Components.VideoFile;
97
+ video: Components.VideoFile_1;
101
98
  }
102
99
 
103
100
  export interface XaiGrokImagineVideoEditVideoInput {
@@ -124,20 +121,19 @@ export interface XaiGrokImagineVideoEditVideoInput {
124
121
 
125
122
  export interface XaiGrokImagineVideoEditVideoOutput {
126
123
  /**
127
- * Video
128
124
  * @description The generated video.
129
125
  * @example {
130
126
  * "height": 720,
131
127
  * "duration": 5.041667,
132
128
  * "url": "https://v3b.fal.media/files/b/0a8b9113/EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
133
- * "width": 1280,
134
129
  * "fps": 24,
130
+ * "width": 1280,
135
131
  * "file_name": "EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
136
132
  * "content_type": "video/mp4",
137
133
  * "num_frames": 121
138
134
  * }
139
135
  */
140
- video: Components.VideoFile;
136
+ video: Components.VideoFile_1;
141
137
  }
142
138
 
143
139
  export interface XaiGrokImagineImageEditInput {
@@ -184,7 +180,7 @@ export interface XaiGrokImagineImageEditOutput {
184
180
  * }
185
181
  * ]
186
182
  */
187
- images: Components.ImageFile[];
183
+ images: Components.ImageFile_1[];
188
184
  /**
189
185
  * Revised Prompt
190
186
  * @description The enhanced prompt that was used to generate the image.
@@ -250,7 +246,7 @@ export interface XaiGrokImagineImageOutput {
250
246
  * }
251
247
  * ]
252
248
  */
253
- images: Components.ImageFile[];
249
+ images: Components.ImageFile_1[];
254
250
  /**
255
251
  * Revised Prompt
256
252
  * @description The enhanced prompt that was used to generate the image.
@@ -941,7 +937,7 @@ export interface VeedAvatarsTextToVideoInput {
941
937
  text: string;
942
938
  }
943
939
 
944
- export interface VeedAvatarsTextToVideoOutput extends SharedType_78a {}
940
+ export interface VeedAvatarsTextToVideoOutput extends SharedType_78a1 {}
945
941
 
946
942
  export interface VeedAvatarsAudioToVideoInput {
947
943
  /**
@@ -986,7 +982,7 @@ export interface VeedAvatarsAudioToVideoInput {
986
982
  | 'any_female_side';
987
983
  }
988
984
 
989
- export interface VeedAvatarsAudioToVideoOutput extends SharedType_78a {}
985
+ export interface VeedAvatarsAudioToVideoOutput extends SharedType_78a1 {}
990
986
 
991
987
  export interface Tripo3dTripoV25MultiviewTo3dInput {
992
988
  /**
@@ -2104,6 +2100,59 @@ export interface SharedType_f5e {
2104
2100
  seed?: number;
2105
2101
  }
2106
2102
 
2103
+ export interface SharedType_f57 {
2104
+ /**
2105
+ * Aspect Ratio
2106
+ * @description The aspect ratio of the generated image.
2107
+ * @default 1:1
2108
+ * @enum {string}
2109
+ */
2110
+ aspect_ratio?: '21:9' | '16:9' | '3:2' | '4:3' | '5:4' | '1:1' | '4:5' | '3:4' | '2:3' | '9:16';
2111
+ /**
2112
+ * Limit Generations
2113
+ * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
2114
+ * @default false
2115
+ */
2116
+ limit_generations?: boolean;
2117
+ /**
2118
+ * Number of Images
2119
+ * @description The number of images to generate.
2120
+ * @default 1
2121
+ */
2122
+ num_images?: number;
2123
+ /**
2124
+ * Output Format
2125
+ * @description The format of the generated image.
2126
+ * @default png
2127
+ * @enum {string}
2128
+ */
2129
+ output_format?: 'jpeg' | 'png' | 'webp';
2130
+ /**
2131
+ * Prompt
2132
+ * @description The text prompt to generate an image from.
2133
+ * @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
2134
+ */
2135
+ prompt: string;
2136
+ /**
2137
+ * Safety Tolerance
2138
+ * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
2139
+ * @default 4
2140
+ * @enum {string}
2141
+ */
2142
+ safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
2143
+ /**
2144
+ * Seed
2145
+ * @description The seed for the random number generator.
2146
+ */
2147
+ seed?: number;
2148
+ /**
2149
+ * Sync Mode
2150
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
2151
+ * @default false
2152
+ */
2153
+ sync_mode?: boolean;
2154
+ }
2155
+
2107
2156
  export interface SharedType_f51 {
2108
2157
  /**
2109
2158
  * Image Url
@@ -2296,29 +2345,6 @@ export interface SharedType_ee0 {
2296
2345
  speaker_embedding: Components.File;
2297
2346
  }
2298
2347
 
2299
- export interface SharedType_ec2 {
2300
- /**
2301
- * Prompt
2302
- * @description The text prompt used for video generation.
2303
- * @default
2304
- * @example A close-up of a young woman smiling gently in the rain, raindrops glistening on her face and eyelashes. The video captures the delicate details of her expression and the water droplets, with soft light reflecting off her skin in the rainy atmosphere.
2305
- */
2306
- prompt?: string;
2307
- /**
2308
- * Seed
2309
- * @description The seed used for generation.
2310
- */
2311
- seed: number;
2312
- /**
2313
- * Video
2314
- * @description The generated video file.
2315
- * @example {
2316
- * "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-woman-output.mp4"
2317
- * }
2318
- */
2319
- video: Components.File;
2320
- }
2321
-
2322
2348
  export interface SharedType_ec0 {
2323
2349
  /**
2324
2350
  * Art Style
@@ -2564,6 +2590,94 @@ export interface SharedType_e4b {
2564
2590
  steps?: number;
2565
2591
  }
2566
2592
 
2593
+ export interface SharedType_e49 {
2594
+ /**
2595
+ * Actual Prompt
2596
+ * @description The actual prompt used if prompt expansion was enabled
2597
+ */
2598
+ actual_prompt?: string;
2599
+ /**
2600
+ * Model Glb
2601
+ * @description Generated 3D object in GLB format.
2602
+ * @example {
2603
+ * "file_size": 9314028,
2604
+ * "file_name": "model.glb",
2605
+ * "content_type": "model/gltf-binary",
2606
+ * "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
2607
+ * }
2608
+ */
2609
+ model_glb: Components.File;
2610
+ /**
2611
+ * Model Urls
2612
+ * @description URLs for different 3D model formats
2613
+ * @example {
2614
+ * "fbx": {
2615
+ * "file_size": 5444380,
2616
+ * "file_name": "model.fbx",
2617
+ * "content_type": "application/octet-stream",
2618
+ * "url": "https://v3b.fal.media/files/b/kangaroo/7nUUw5dHN9a0DKlOpAKbP_model.fbx"
2619
+ * },
2620
+ * "usdz": {
2621
+ * "file_size": 9834246,
2622
+ * "file_name": "model.usdz",
2623
+ * "content_type": "model/vnd.usdz+zip",
2624
+ * "url": "https://v3b.fal.media/files/b/panda/XcC-mIJywUvH7coyrzENU_model.usdz"
2625
+ * },
2626
+ * "obj": {
2627
+ * "file_size": 2755145,
2628
+ * "file_name": "model.obj",
2629
+ * "content_type": "text/plain",
2630
+ * "url": "https://v3b.fal.media/files/b/monkey/cCNMHqUbKSNtDN1iGmiYm_model.obj"
2631
+ * },
2632
+ * "glb": {
2633
+ * "file_size": 9314028,
2634
+ * "file_name": "model.glb",
2635
+ * "content_type": "model/gltf-binary",
2636
+ * "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
2637
+ * }
2638
+ * }
2639
+ */
2640
+ model_urls: Components.ModelUrls;
2641
+ /**
2642
+ * Prompt
2643
+ * @description The text prompt used for generation
2644
+ * @example A rustic, antique wooden treasure chest with a curved, domed lid, constructed from weathered, dark brown planks exhibiting prominent wood grain and subtle distress. It's heavily reinforced with broad, dark grey, oxidized metal bands secured by numerous circular rivets. Ornate, dark iron decorative elements featuring swirling foliate patterns and dragon motifs adorn the corners and lid. A prominent, circular, intricately carved metal lock plate with a central keyhole dominates the front, flanked by two large, dark metallic pull rings.
2645
+ */
2646
+ prompt: string;
2647
+ /**
2648
+ * Seed
2649
+ * @description The seed used for generation
2650
+ * @example 4002110719
2651
+ */
2652
+ seed?: number;
2653
+ /**
2654
+ * Texture Urls
2655
+ * @description Array of texture file objects
2656
+ * @example [
2657
+ * {
2658
+ * "base_color": {
2659
+ * "file_size": 4254502,
2660
+ * "file_name": "texture_0.png",
2661
+ * "content_type": "image/png",
2662
+ * "url": "https://v3b.fal.media/files/b/panda/DoPKAuZY0tTjnr6C9ee-Q_texture_0.png"
2663
+ * }
2664
+ * }
2665
+ * ]
2666
+ */
2667
+ texture_urls?: Components.TextureFiles[];
2668
+ /**
2669
+ * Thumbnail
2670
+ * @description Preview thumbnail of the generated model
2671
+ * @example {
2672
+ * "file_size": 173792,
2673
+ * "file_name": "preview.png",
2674
+ * "content_type": "image/png",
2675
+ * "url": "https://v3b.fal.media/files/b/koala/6LJISu4ilkZXcdOETwl_d_preview.png"
2676
+ * }
2677
+ */
2678
+ thumbnail?: Components.File;
2679
+ }
2680
+
2567
2681
  export interface SharedType_e3b {
2568
2682
  /**
2569
2683
  * Video
@@ -4060,6 +4174,17 @@ export interface SharedType_c7e {
4060
4174
  audio: Components.File;
4061
4175
  }
4062
4176
 
4177
+ export interface SharedType_c61 {
4178
+ /**
4179
+ * Video
4180
+ * @description The generated video from image using the Q3 model
4181
+ * @example {
4182
+ * "url": "https://v3b.fal.media/files/b/0a8c9189/n9z3uUDPqmU2msAtqr25-_output.mp4"
4183
+ * }
4184
+ */
4185
+ video: Components.File;
4186
+ }
4187
+
4063
4188
  export interface SharedType_c5e {
4064
4189
  /**
4065
4190
  * Acceleration
@@ -4256,6 +4381,21 @@ export interface SharedType_c3a {
4256
4381
  speakers: Components.VibeVoiceSpeaker[];
4257
4382
  }
4258
4383
 
4384
+ export interface SharedType_c0f {
4385
+ /**
4386
+ * @description The generated image file.
4387
+ * @example {
4388
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/wan/t2i-output.png"
4389
+ * }
4390
+ */
4391
+ image: Components.File_1;
4392
+ /**
4393
+ * Seed
4394
+ * @description The seed used for generation.
4395
+ */
4396
+ seed: number;
4397
+ }
4398
+
4259
4399
  export interface SharedType_bf2 {
4260
4400
  /**
4261
4401
  * Image
@@ -4314,29 +4454,6 @@ export interface SharedType_bda {
4314
4454
  video_url: string;
4315
4455
  }
4316
4456
 
4317
- export interface SharedType_bd3 {
4318
- /**
4319
- * Prompt
4320
- * @description The text prompt used for video generation.
4321
- * @default
4322
- * @example The white dragon warrior stands still, eyes full of determination and strength. The camera slowly moves closer or circles around the warrior, highlighting the powerful presence and heroic spirit of the character.
4323
- */
4324
- prompt?: string;
4325
- /**
4326
- * Seed
4327
- * @description The seed used for generation.
4328
- */
4329
- seed: number;
4330
- /**
4331
- * Video
4332
- * @description The generated video file.
4333
- * @example {
4334
- * "url": "https://v3.fal.media/files/lion/Fbuh3lO_HMT-pS0DATbio_tmp08c3v477.mp4"
4335
- * }
4336
- */
4337
- video: Components.File;
4338
- }
4339
-
4340
4457
  export interface SharedType_bc9 {
4341
4458
  /**
4342
4459
  * Has Nsfw Concepts
@@ -4729,29 +4846,6 @@ export interface SharedType_b61 {
4729
4846
  }[];
4730
4847
  }
4731
4848
 
4732
- export interface SharedType_b5a {
4733
- /**
4734
- * Prompt
4735
- * @description The text prompt used for video generation.
4736
- * @default
4737
- * @example A medium shot establishes a modern, minimalist office setting: clean lines, muted grey walls, and polished wood surfaces. The focus shifts to a close-up on a woman in sharp, navy blue business attire. Her crisp white blouse contrasts with the deep blue of her tailored suit jacket. The subtle texture of the fabric is visible—a fine weave with a slight sheen. Her expression is serious, yet engaging, as she speaks to someone unseen just beyond the frame. Close-up on her eyes, showing the intensity of her gaze and the fine lines around them that hint at experience and focus. Her lips are slightly parted, as if mid-sentence. The light catches the subtle highlights in her auburn hair, meticulously styled. Note the slight catch of light on the silver band of her watch. High resolution 4k
4738
- */
4739
- prompt?: string;
4740
- /**
4741
- * Seed
4742
- * @description The seed used for generation.
4743
- */
4744
- seed: number;
4745
- /**
4746
- * Video
4747
- * @description The generated video file.
4748
- * @example {
4749
- * "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-small-output.mp4"
4750
- * }
4751
- */
4752
- video: Components.File;
4753
- }
4754
-
4755
4849
  export interface SharedType_b59 {
4756
4850
  /**
4757
4851
  * Aspect Ratio
@@ -5056,94 +5150,6 @@ export interface SharedType_aed {
5056
5150
  sync_mode?: boolean;
5057
5151
  }
5058
5152
 
5059
- export interface SharedType_ae3 {
5060
- /**
5061
- * Actual Prompt
5062
- * @description The actual prompt used if prompt expansion was enabled
5063
- */
5064
- actual_prompt?: string;
5065
- /**
5066
- * Model Glb
5067
- * @description Generated 3D object in GLB format.
5068
- * @example {
5069
- * "file_size": 9314028,
5070
- * "file_name": "model.glb",
5071
- * "content_type": "model/gltf-binary",
5072
- * "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
5073
- * }
5074
- */
5075
- model_glb: Components.File;
5076
- /**
5077
- * Model Urls
5078
- * @description URLs for different 3D model formats
5079
- * @example {
5080
- * "fbx": {
5081
- * "file_size": 5444380,
5082
- * "file_name": "model.fbx",
5083
- * "content_type": "application/octet-stream",
5084
- * "url": "https://v3b.fal.media/files/b/kangaroo/7nUUw5dHN9a0DKlOpAKbP_model.fbx"
5085
- * },
5086
- * "usdz": {
5087
- * "file_size": 9834246,
5088
- * "file_name": "model.usdz",
5089
- * "content_type": "model/vnd.usdz+zip",
5090
- * "url": "https://v3b.fal.media/files/b/panda/XcC-mIJywUvH7coyrzENU_model.usdz"
5091
- * },
5092
- * "glb": {
5093
- * "file_size": 9314028,
5094
- * "file_name": "model.glb",
5095
- * "content_type": "model/gltf-binary",
5096
- * "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
5097
- * },
5098
- * "obj": {
5099
- * "file_size": 2755145,
5100
- * "file_name": "model.obj",
5101
- * "content_type": "text/plain",
5102
- * "url": "https://v3b.fal.media/files/b/monkey/cCNMHqUbKSNtDN1iGmiYm_model.obj"
5103
- * }
5104
- * }
5105
- */
5106
- model_urls: Components.ModelUrls;
5107
- /**
5108
- * Prompt
5109
- * @description The text prompt used for generation
5110
- * @example A rustic, antique wooden treasure chest with a curved, domed lid, constructed from weathered, dark brown planks exhibiting prominent wood grain and subtle distress. It's heavily reinforced with broad, dark grey, oxidized metal bands secured by numerous circular rivets. Ornate, dark iron decorative elements featuring swirling foliate patterns and dragon motifs adorn the corners and lid. A prominent, circular, intricately carved metal lock plate with a central keyhole dominates the front, flanked by two large, dark metallic pull rings.
5111
- */
5112
- prompt: string;
5113
- /**
5114
- * Seed
5115
- * @description The seed used for generation
5116
- * @example 4002110719
5117
- */
5118
- seed?: number;
5119
- /**
5120
- * Texture Urls
5121
- * @description Array of texture file objects
5122
- * @example [
5123
- * {
5124
- * "base_color": {
5125
- * "file_size": 4254502,
5126
- * "file_name": "texture_0.png",
5127
- * "content_type": "image/png",
5128
- * "url": "https://v3b.fal.media/files/b/panda/DoPKAuZY0tTjnr6C9ee-Q_texture_0.png"
5129
- * }
5130
- * }
5131
- * ]
5132
- */
5133
- texture_urls?: Components.TextureFiles[];
5134
- /**
5135
- * Thumbnail
5136
- * @description Preview thumbnail of the generated model
5137
- * @example {
5138
- * "file_size": 173792,
5139
- * "file_name": "preview.png",
5140
- * "content_type": "image/png",
5141
- * "url": "https://v3b.fal.media/files/b/koala/6LJISu4ilkZXcdOETwl_d_preview.png"
5142
- * }
5143
- */
5144
- thumbnail?: Components.File;
5145
- }
5146
-
5147
5153
  export interface SharedType_ade {
5148
5154
  /**
5149
5155
  * Video
@@ -6064,26 +6070,6 @@ export interface SharedType_9af {
6064
6070
  topology?: 'quad' | 'triangle';
6065
6071
  }
6066
6072
 
6067
- export interface SharedType_98c {
6068
- /**
6069
- * Description
6070
- * @description The description of the generated images.
6071
- */
6072
- description: string;
6073
- /**
6074
- * Images
6075
- * @description The edited images.
6076
- * @example [
6077
- * {
6078
- * "file_name": "nano-banana-multi-edit-output.png",
6079
- * "content_type": "image/png",
6080
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-multi-edit-output.png"
6081
- * }
6082
- * ]
6083
- */
6084
- images: Components.ImageFile[];
6085
- }
6086
-
6087
6073
  export interface SharedType_97f {
6088
6074
  /**
6089
6075
  * Values
@@ -6098,47 +6084,6 @@ export interface SharedType_97f {
6098
6084
  }[];
6099
6085
  }
6100
6086
 
6101
- export interface SharedType_97e {
6102
- /**
6103
- * Aspect Ratio
6104
- * @description The aspect ratio of the generated image.
6105
- * @default 1:1
6106
- * @enum {string}
6107
- */
6108
- aspect_ratio?: '21:9' | '16:9' | '3:2' | '4:3' | '5:4' | '1:1' | '4:5' | '3:4' | '2:3' | '9:16';
6109
- /**
6110
- * Limit Generations
6111
- * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
6112
- * @default false
6113
- */
6114
- limit_generations?: boolean;
6115
- /**
6116
- * Number of Images
6117
- * @description The number of images to generate.
6118
- * @default 1
6119
- */
6120
- num_images?: number;
6121
- /**
6122
- * Output Format
6123
- * @description The format of the generated image.
6124
- * @default png
6125
- * @enum {string}
6126
- */
6127
- output_format?: 'jpeg' | 'png' | 'webp';
6128
- /**
6129
- * Prompt
6130
- * @description The text prompt to generate an image from.
6131
- * @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
6132
- */
6133
- prompt: string;
6134
- /**
6135
- * Sync Mode
6136
- * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
6137
- * @default false
6138
- */
6139
- sync_mode?: boolean;
6140
- }
6141
-
6142
6087
  export interface SharedType_977 {
6143
6088
  /**
6144
6089
  * Aspect Ratio
@@ -6269,7 +6214,7 @@ export interface SharedType_962 {
6269
6214
  };
6270
6215
  }
6271
6216
 
6272
- export interface SharedType_95d {
6217
+ export interface SharedType_95d1 {
6273
6218
  /**
6274
6219
  * Video
6275
6220
  * @description The generated video
@@ -6282,6 +6227,46 @@ export interface SharedType_95d {
6282
6227
  video: Components.File;
6283
6228
  }
6284
6229
 
6230
+ export interface SharedType_95d {
6231
+ /**
6232
+ * Aspect Ratio
6233
+ * @description The aspect ratio of the output video
6234
+ * @default 16:9
6235
+ * @enum {string}
6236
+ */
6237
+ aspect_ratio?: '16:9' | '9:16' | '4:3' | '3:4' | '1:1';
6238
+ /**
6239
+ * Audio
6240
+ * @description Whether to use direct audio-video generation. When true, outputs video with sound.
6241
+ * @default true
6242
+ */
6243
+ audio?: boolean;
6244
+ /**
6245
+ * Duration
6246
+ * @description Duration of the video in seconds
6247
+ * @default 5
6248
+ */
6249
+ duration?: number;
6250
+ /**
6251
+ * Prompt
6252
+ * @description Text prompt for video generation, max 2000 characters
6253
+ * @example In an ultra-realistic fashion photography style featuring light blue and pale amber tones, an astronaut in a spacesuit walks through the fog.
6254
+ */
6255
+ prompt: string;
6256
+ /**
6257
+ * Resolution
6258
+ * @description Output video resolution
6259
+ * @default 720p
6260
+ * @enum {string}
6261
+ */
6262
+ resolution?: '360p' | '540p' | '720p' | '1080p';
6263
+ /**
6264
+ * Seed
6265
+ * @description Random seed for reproducibility. If None, a random seed is chosen.
6266
+ */
6267
+ seed?: number;
6268
+ }
6269
+
6285
6270
  export interface SharedType_95c {
6286
6271
  /**
6287
6272
  * Images
@@ -6591,6 +6576,83 @@ export interface SharedType_905 {
6591
6576
  seed?: number;
6592
6577
  }
6593
6578
 
6579
+ export interface SharedType_8f1 {
6580
+ /**
6581
+ * Model Glb
6582
+ * @description Generated 3D object in GLB format.
6583
+ * @example {
6584
+ * "file_size": 9242744,
6585
+ * "file_name": "model.glb",
6586
+ * "content_type": "model/gltf-binary",
6587
+ * "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
6588
+ * }
6589
+ */
6590
+ model_glb: Components.File;
6591
+ /**
6592
+ * Model Urls
6593
+ * @description URLs for different 3D model formats
6594
+ * @example {
6595
+ * "fbx": {
6596
+ * "file_size": 5427052,
6597
+ * "file_name": "model.fbx",
6598
+ * "content_type": "application/octet-stream",
6599
+ * "url": "https://v3b.fal.media/files/b/kangaroo/4Q2qdpTvfLVdzAKH1-72v_model.fbx"
6600
+ * },
6601
+ * "usdz": {
6602
+ * "file_size": 9991969,
6603
+ * "file_name": "model.usdz",
6604
+ * "content_type": "model/vnd.usdz+zip",
6605
+ * "url": "https://v3b.fal.media/files/b/lion/RgJG9EBQ_GAHMVWV3wCis_model.usdz"
6606
+ * },
6607
+ * "obj": {
6608
+ * "file_size": 2744413,
6609
+ * "file_name": "model.obj",
6610
+ * "content_type": "text/plain",
6611
+ * "url": "https://v3b.fal.media/files/b/koala/_Vg0d084-hd3EdpIJDf7U_model.obj"
6612
+ * },
6613
+ * "glb": {
6614
+ * "file_size": 9242744,
6615
+ * "file_name": "model.glb",
6616
+ * "content_type": "model/gltf-binary",
6617
+ * "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
6618
+ * }
6619
+ * }
6620
+ */
6621
+ model_urls: Components.ModelUrls;
6622
+ /**
6623
+ * Seed
6624
+ * @description The seed used for generation (if available)
6625
+ * @example 2009275957
6626
+ */
6627
+ seed?: number;
6628
+ /**
6629
+ * Texture Urls
6630
+ * @description Array of texture file objects, matching Meshy API structure
6631
+ * @example [
6632
+ * {
6633
+ * "base_color": {
6634
+ * "file_size": 4328755,
6635
+ * "file_name": "texture_0.png",
6636
+ * "content_type": "image/png",
6637
+ * "url": "https://v3b.fal.media/files/b/tiger/NkgxcEom_42V4_8UUXiRR_texture_0.png"
6638
+ * }
6639
+ * }
6640
+ * ]
6641
+ */
6642
+ texture_urls?: Components.TextureFiles[];
6643
+ /**
6644
+ * Thumbnail
6645
+ * @description Preview thumbnail of the generated model
6646
+ * @example {
6647
+ * "file_size": 54279,
6648
+ * "file_name": "preview.png",
6649
+ * "content_type": "image/png",
6650
+ * "url": "https://v3b.fal.media/files/b/penguin/rfnS6ClmeEWgDXp_oD5tN_preview.png"
6651
+ * }
6652
+ */
6653
+ thumbnail?: Components.File;
6654
+ }
6655
+
6594
6656
  export interface SharedType_8df {
6595
6657
  /**
6596
6658
  * Aspect Ratio
@@ -6735,6 +6797,82 @@ export interface SharedType_8c5 {
6735
6797
  seed: number;
6736
6798
  }
6737
6799
 
6800
+ export interface SharedType_8c2 {
6801
+ /**
6802
+ * Aspect Ratio
6803
+ * @description The aspect ratio of the generated image. Use "auto" to let the model decide based on the prompt.
6804
+ * @default 1:1
6805
+ */
6806
+ aspect_ratio?:
6807
+ | 'auto'
6808
+ | '21:9'
6809
+ | '16:9'
6810
+ | '3:2'
6811
+ | '4:3'
6812
+ | '5:4'
6813
+ | '1:1'
6814
+ | '4:5'
6815
+ | '3:4'
6816
+ | '2:3'
6817
+ | '9:16';
6818
+ /**
6819
+ * Enable Web Search
6820
+ * @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
6821
+ * @default false
6822
+ */
6823
+ enable_web_search?: boolean;
6824
+ /**
6825
+ * Limit Generations
6826
+ * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
6827
+ * @default false
6828
+ */
6829
+ limit_generations?: boolean;
6830
+ /**
6831
+ * Number of Images
6832
+ * @description The number of images to generate.
6833
+ * @default 1
6834
+ */
6835
+ num_images?: number;
6836
+ /**
6837
+ * Output Format
6838
+ * @description The format of the generated image.
6839
+ * @default png
6840
+ * @enum {string}
6841
+ */
6842
+ output_format?: 'jpeg' | 'png' | 'webp';
6843
+ /**
6844
+ * Prompt
6845
+ * @description The text prompt to generate an image from.
6846
+ * @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
6847
+ */
6848
+ prompt: string;
6849
+ /**
6850
+ * Resolution
6851
+ * @description The resolution of the image to generate.
6852
+ * @default 1K
6853
+ * @enum {string}
6854
+ */
6855
+ resolution?: '1K' | '2K' | '4K';
6856
+ /**
6857
+ * Safety Tolerance
6858
+ * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
6859
+ * @default 4
6860
+ * @enum {string}
6861
+ */
6862
+ safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
6863
+ /**
6864
+ * Seed
6865
+ * @description The seed for the random number generator.
6866
+ */
6867
+ seed?: number;
6868
+ /**
6869
+ * Sync Mode
6870
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
6871
+ * @default false
6872
+ */
6873
+ sync_mode?: boolean;
6874
+ }
6875
+
6738
6876
  export interface SharedType_8c0 {
6739
6877
  /**
6740
6878
  * Seed
@@ -6803,6 +6941,26 @@ export interface SharedType_87d {
6803
6941
  video: Components.File;
6804
6942
  }
6805
6943
 
6944
+ export interface SharedType_876 {
6945
+ /**
6946
+ * Description
6947
+ * @description The description of the generated images.
6948
+ */
6949
+ description: string;
6950
+ /**
6951
+ * Images
6952
+ * @description The edited images.
6953
+ * @example [
6954
+ * {
6955
+ * "file_name": "nano-banana-multi-edit-output.png",
6956
+ * "content_type": "image/png",
6957
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-multi-edit-output.png"
6958
+ * }
6959
+ * ]
6960
+ */
6961
+ images: Components.ImageFile_1[];
6962
+ }
6963
+
6806
6964
  export interface SharedType_86b {
6807
6965
  /**
6808
6966
  * Audio Path
@@ -6870,6 +7028,78 @@ export interface SharedType_868 {
6870
7028
  sync_mode?: boolean;
6871
7029
  }
6872
7030
 
7031
+ export interface SharedType_85d {
7032
+ /**
7033
+ * Aspect Ratio
7034
+ * @description The aspect ratio of the generated image.
7035
+ * @default auto
7036
+ */
7037
+ aspect_ratio?:
7038
+ | 'auto'
7039
+ | '21:9'
7040
+ | '16:9'
7041
+ | '3:2'
7042
+ | '4:3'
7043
+ | '5:4'
7044
+ | '1:1'
7045
+ | '4:5'
7046
+ | '3:4'
7047
+ | '2:3'
7048
+ | '9:16';
7049
+ /**
7050
+ * Image URLs
7051
+ * @description The URLs of the images to use for image-to-image generation or image editing.
7052
+ * @example [
7053
+ * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
7054
+ * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
7055
+ * ]
7056
+ */
7057
+ image_urls: string[];
7058
+ /**
7059
+ * Limit Generations
7060
+ * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
7061
+ * @default false
7062
+ */
7063
+ limit_generations?: boolean;
7064
+ /**
7065
+ * Number of Images
7066
+ * @description The number of images to generate.
7067
+ * @default 1
7068
+ */
7069
+ num_images?: number;
7070
+ /**
7071
+ * Output Format
7072
+ * @description The format of the generated image.
7073
+ * @default png
7074
+ * @enum {string}
7075
+ */
7076
+ output_format?: 'jpeg' | 'png' | 'webp';
7077
+ /**
7078
+ * Prompt
7079
+ * @description The prompt for image editing.
7080
+ * @example make a photo of the man driving the car down the california coastline
7081
+ */
7082
+ prompt: string;
7083
+ /**
7084
+ * Safety Tolerance
7085
+ * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
7086
+ * @default 4
7087
+ * @enum {string}
7088
+ */
7089
+ safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
7090
+ /**
7091
+ * Seed
7092
+ * @description The seed for the random number generator.
7093
+ */
7094
+ seed?: number;
7095
+ /**
7096
+ * Sync Mode
7097
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
7098
+ * @default false
7099
+ */
7100
+ sync_mode?: boolean;
7101
+ }
7102
+
6873
7103
  export interface SharedType_82a {
6874
7104
  /**
6875
7105
  * Aspect Ratio (W:H)
@@ -7110,67 +7340,6 @@ export interface SharedType_815 {
7110
7340
  video: Components.File;
7111
7341
  }
7112
7342
 
7113
- export interface SharedType_813 {
7114
- /**
7115
- * Aspect Ratio
7116
- * @description The aspect ratio of the generated image.
7117
- * @default auto
7118
- * @enum {string}
7119
- */
7120
- aspect_ratio?:
7121
- | 'auto'
7122
- | '21:9'
7123
- | '16:9'
7124
- | '3:2'
7125
- | '4:3'
7126
- | '5:4'
7127
- | '1:1'
7128
- | '4:5'
7129
- | '3:4'
7130
- | '2:3'
7131
- | '9:16';
7132
- /**
7133
- * Image URLs
7134
- * @description The URLs of the images to use for image-to-image generation or image editing.
7135
- * @example [
7136
- * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
7137
- * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
7138
- * ]
7139
- */
7140
- image_urls: string[];
7141
- /**
7142
- * Limit Generations
7143
- * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
7144
- * @default false
7145
- */
7146
- limit_generations?: boolean;
7147
- /**
7148
- * Number of Images
7149
- * @description The number of images to generate.
7150
- * @default 1
7151
- */
7152
- num_images?: number;
7153
- /**
7154
- * Output Format
7155
- * @description The format of the generated image.
7156
- * @default png
7157
- * @enum {string}
7158
- */
7159
- output_format?: 'jpeg' | 'png' | 'webp';
7160
- /**
7161
- * Prompt
7162
- * @description The prompt for image editing.
7163
- * @example make a photo of the man driving the car down the california coastline
7164
- */
7165
- prompt: string;
7166
- /**
7167
- * Sync Mode
7168
- * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
7169
- * @default false
7170
- */
7171
- sync_mode?: boolean;
7172
- }
7173
-
7174
7343
  export interface SharedType_812 {
7175
7344
  /**
7176
7345
  * Video
@@ -7380,6 +7549,26 @@ export interface SharedType_7e9 {
7380
7549
  video?: Components.File;
7381
7550
  }
7382
7551
 
7552
+ export interface SharedType_7b9 {
7553
+ /**
7554
+ * Description
7555
+ * @description The description of the generated images.
7556
+ */
7557
+ description: string;
7558
+ /**
7559
+ * Images
7560
+ * @description The generated images.
7561
+ * @example [
7562
+ * {
7563
+ * "file_name": "nano-banana-t2i-output.png",
7564
+ * "content_type": "image/png",
7565
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-t2i-output.png"
7566
+ * }
7567
+ * ]
7568
+ */
7569
+ images: Components.ImageFile_1[];
7570
+ }
7571
+
7383
7572
  export interface SharedType_7a3 {
7384
7573
  /**
7385
7574
  * Aspect Ratio
@@ -7458,7 +7647,7 @@ export interface SharedType_79c {
7458
7647
  seed?: number;
7459
7648
  }
7460
7649
 
7461
- export interface SharedType_78a {
7650
+ export interface SharedType_78a1 {
7462
7651
  /**
7463
7652
  * @example {
7464
7653
  * "content_type": "video/mp4",
@@ -7468,6 +7657,28 @@ export interface SharedType_78a {
7468
7657
  video: Components.File_1;
7469
7658
  }
7470
7659
 
7660
+ export interface SharedType_78a {
7661
+ /**
7662
+ * Prompt
7663
+ * @description The text prompt used for video generation.
7664
+ * @default
7665
+ * @example The white dragon warrior stands still, eyes full of determination and strength. The camera slowly moves closer or circles around the warrior, highlighting the powerful presence and heroic spirit of the character.
7666
+ */
7667
+ prompt?: string;
7668
+ /**
7669
+ * Seed
7670
+ * @description The seed used for generation.
7671
+ */
7672
+ seed: number;
7673
+ /**
7674
+ * @description The generated video file.
7675
+ * @example {
7676
+ * "url": "https://v3.fal.media/files/lion/Fbuh3lO_HMT-pS0DATbio_tmp08c3v477.mp4"
7677
+ * }
7678
+ */
7679
+ video: Components.File_1;
7680
+ }
7681
+
7471
7682
  export interface SharedType_778 {
7472
7683
  /**
7473
7684
  * Character Orientation
@@ -7620,83 +7831,6 @@ export interface SharedType_74f {
7620
7831
  video: Components.File;
7621
7832
  }
7622
7833
 
7623
- export interface SharedType_748 {
7624
- /**
7625
- * Model Glb
7626
- * @description Generated 3D object in GLB format.
7627
- * @example {
7628
- * "file_size": 9242744,
7629
- * "file_name": "model.glb",
7630
- * "content_type": "model/gltf-binary",
7631
- * "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
7632
- * }
7633
- */
7634
- model_glb: Components.File;
7635
- /**
7636
- * Model Urls
7637
- * @description URLs for different 3D model formats
7638
- * @example {
7639
- * "fbx": {
7640
- * "file_size": 5427052,
7641
- * "file_name": "model.fbx",
7642
- * "content_type": "application/octet-stream",
7643
- * "url": "https://v3b.fal.media/files/b/kangaroo/4Q2qdpTvfLVdzAKH1-72v_model.fbx"
7644
- * },
7645
- * "usdz": {
7646
- * "file_size": 9991969,
7647
- * "file_name": "model.usdz",
7648
- * "content_type": "model/vnd.usdz+zip",
7649
- * "url": "https://v3b.fal.media/files/b/lion/RgJG9EBQ_GAHMVWV3wCis_model.usdz"
7650
- * },
7651
- * "glb": {
7652
- * "file_size": 9242744,
7653
- * "file_name": "model.glb",
7654
- * "content_type": "model/gltf-binary",
7655
- * "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
7656
- * },
7657
- * "obj": {
7658
- * "file_size": 2744413,
7659
- * "file_name": "model.obj",
7660
- * "content_type": "text/plain",
7661
- * "url": "https://v3b.fal.media/files/b/koala/_Vg0d084-hd3EdpIJDf7U_model.obj"
7662
- * }
7663
- * }
7664
- */
7665
- model_urls: Components.ModelUrls;
7666
- /**
7667
- * Seed
7668
- * @description The seed used for generation (if available)
7669
- * @example 2009275957
7670
- */
7671
- seed?: number;
7672
- /**
7673
- * Texture Urls
7674
- * @description Array of texture file objects, matching Meshy API structure
7675
- * @example [
7676
- * {
7677
- * "base_color": {
7678
- * "file_size": 4328755,
7679
- * "file_name": "texture_0.png",
7680
- * "content_type": "image/png",
7681
- * "url": "https://v3b.fal.media/files/b/tiger/NkgxcEom_42V4_8UUXiRR_texture_0.png"
7682
- * }
7683
- * }
7684
- * ]
7685
- */
7686
- texture_urls?: Components.TextureFiles[];
7687
- /**
7688
- * Thumbnail
7689
- * @description Preview thumbnail of the generated model
7690
- * @example {
7691
- * "file_size": 54279,
7692
- * "file_name": "preview.png",
7693
- * "content_type": "image/png",
7694
- * "url": "https://v3b.fal.media/files/b/penguin/rfnS6ClmeEWgDXp_oD5tN_preview.png"
7695
- * }
7696
- */
7697
- thumbnail?: Components.File;
7698
- }
7699
-
7700
7834
  export interface SharedType_744 {
7701
7835
  /**
7702
7836
  * Image
@@ -8252,26 +8386,6 @@ export interface SharedType_663 {
8252
8386
  sync_mode?: boolean;
8253
8387
  }
8254
8388
 
8255
- export interface SharedType_662 {
8256
- /**
8257
- * Description
8258
- * @description The description of the generated images.
8259
- */
8260
- description: string;
8261
- /**
8262
- * Images
8263
- * @description The generated images.
8264
- * @example [
8265
- * {
8266
- * "file_name": "nano-banana-t2i-output.png",
8267
- * "content_type": "image/png",
8268
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-t2i-output.png"
8269
- * }
8270
- * ]
8271
- */
8272
- images: Components.ImageFile[];
8273
- }
8274
-
8275
8389
  export interface SharedType_659 {
8276
8390
  /** @description Image with depth map */
8277
8391
  image: Components.Image_2;
@@ -8496,6 +8610,64 @@ export interface SharedType_5c6 {
8496
8610
  results: Components.BoundingBoxes;
8497
8611
  }
8498
8612
 
8613
+ export interface SharedType_5c5 {
8614
+ /**
8615
+ * Prompt
8616
+ * @description The text prompt used for video generation.
8617
+ * @default
8618
+ * @example A medium shot establishes a modern, minimalist office setting: clean lines, muted grey walls, and polished wood surfaces. The focus shifts to a close-up on a woman in sharp, navy blue business attire. Her crisp white blouse contrasts with the deep blue of her tailored suit jacket. The subtle texture of the fabric is visible—a fine weave with a slight sheen. Her expression is serious, yet engaging, as she speaks to someone unseen just beyond the frame. Close-up on her eyes, showing the intensity of her gaze and the fine lines around them that hint at experience and focus. Her lips are slightly parted, as if mid-sentence. The light catches the subtle highlights in her auburn hair, meticulously styled. Note the slight catch of light on the silver band of her watch. High resolution 4k
8619
+ */
8620
+ prompt?: string;
8621
+ /**
8622
+ * Seed
8623
+ * @description The seed used for generation.
8624
+ */
8625
+ seed: number;
8626
+ /**
8627
+ * @description The generated video file.
8628
+ * @example {
8629
+ * "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-small-output.mp4"
8630
+ * }
8631
+ */
8632
+ video: Components.File_1;
8633
+ }
8634
+
8635
+ export interface SharedType_5c2 {
8636
+ /**
8637
+ * Has Nsfw Concepts
8638
+ * @description Whether the generated images contain NSFW concepts.
8639
+ */
8640
+ has_nsfw_concepts: boolean[];
8641
+ /**
8642
+ * Images
8643
+ * @description The generated image(s).
8644
+ * @example [
8645
+ * {
8646
+ * "height": 768,
8647
+ * "content_type": "image/jpeg",
8648
+ * "url": "https://v3b.fal.media/files/b/0a8e76a4/n8AJdfVTm-9msRC-sVQFn.jpg",
8649
+ * "width": 1024
8650
+ * }
8651
+ * ]
8652
+ */
8653
+ images: Components.Image_1[];
8654
+ /**
8655
+ * Prompt
8656
+ * @description The prompt used for generating the image.
8657
+ */
8658
+ prompt: string;
8659
+ /**
8660
+ * Seed
8661
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
8662
+ * input or the randomly generated that was used in case none was passed.
8663
+ */
8664
+ seed: number;
8665
+ /** Timings */
8666
+ timings: {
8667
+ [key: string]: number;
8668
+ };
8669
+ }
8670
+
8499
8671
  export interface SharedType_5bd {
8500
8672
  /**
8501
8673
  * Model Mesh
@@ -9184,22 +9356,6 @@ export interface SharedType_524 {
9184
9356
  video: Components.VideoFile_1;
9185
9357
  }
9186
9358
 
9187
- export interface SharedType_4f3 {
9188
- /**
9189
- * Image
9190
- * @description The generated image file.
9191
- * @example {
9192
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/wan/t2i-output.png"
9193
- * }
9194
- */
9195
- image: Components.File;
9196
- /**
9197
- * Seed
9198
- * @description The seed used for generation.
9199
- */
9200
- seed: number;
9201
- }
9202
-
9203
9359
  export interface SharedType_4c2 {
9204
9360
  /**
9205
9361
  * Video
@@ -10101,6 +10257,28 @@ export interface SharedType_386 {
10101
10257
  images: Components.Image[];
10102
10258
  }
10103
10259
 
10260
+ export interface SharedType_382 {
10261
+ /**
10262
+ * Prompt
10263
+ * @description The text prompt used for video generation.
10264
+ * @default
10265
+ * @example A close-up of a young woman smiling gently in the rain, raindrops glistening on her face and eyelashes. The video captures the delicate details of her expression and the water droplets, with soft light reflecting off her skin in the rainy atmosphere.
10266
+ */
10267
+ prompt?: string;
10268
+ /**
10269
+ * Seed
10270
+ * @description The seed used for generation.
10271
+ */
10272
+ seed: number;
10273
+ /**
10274
+ * @description The generated video file.
10275
+ * @example {
10276
+ * "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-woman-output.mp4"
10277
+ * }
10278
+ */
10279
+ video: Components.File_1;
10280
+ }
10281
+
10104
10282
  export interface SharedType_377 {
10105
10283
  /**
10106
10284
  * Output
@@ -10192,6 +10370,91 @@ export interface SharedType_368 {
10192
10370
  };
10193
10371
  }
10194
10372
 
10373
+ export interface SharedType_367 {
10374
+ /**
10375
+ * Aspect Ratio
10376
+ * @description The aspect ratio of the generated image.
10377
+ * @default auto
10378
+ */
10379
+ aspect_ratio?:
10380
+ | 'auto'
10381
+ | '21:9'
10382
+ | '16:9'
10383
+ | '3:2'
10384
+ | '4:3'
10385
+ | '5:4'
10386
+ | '1:1'
10387
+ | '4:5'
10388
+ | '3:4'
10389
+ | '2:3'
10390
+ | '9:16';
10391
+ /**
10392
+ * Enable Web Search
10393
+ * @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
10394
+ * @default false
10395
+ */
10396
+ enable_web_search?: boolean;
10397
+ /**
10398
+ * Image URLs
10399
+ * @description The URLs of the images to use for image-to-image generation or image editing.
10400
+ * @example [
10401
+ * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
10402
+ * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
10403
+ * ]
10404
+ */
10405
+ image_urls: string[];
10406
+ /**
10407
+ * Limit Generations
10408
+ * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
10409
+ * @default false
10410
+ */
10411
+ limit_generations?: boolean;
10412
+ /**
10413
+ * Number of Images
10414
+ * @description The number of images to generate.
10415
+ * @default 1
10416
+ */
10417
+ num_images?: number;
10418
+ /**
10419
+ * Output Format
10420
+ * @description The format of the generated image.
10421
+ * @default png
10422
+ * @enum {string}
10423
+ */
10424
+ output_format?: 'jpeg' | 'png' | 'webp';
10425
+ /**
10426
+ * Prompt
10427
+ * @description The prompt for image editing.
10428
+ * @example make a photo of the man driving the car down the california coastline
10429
+ */
10430
+ prompt: string;
10431
+ /**
10432
+ * Resolution
10433
+ * @description The resolution of the image to generate.
10434
+ * @default 1K
10435
+ * @enum {string}
10436
+ */
10437
+ resolution?: '1K' | '2K' | '4K';
10438
+ /**
10439
+ * Safety Tolerance
10440
+ * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
10441
+ * @default 4
10442
+ * @enum {string}
10443
+ */
10444
+ safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
10445
+ /**
10446
+ * Seed
10447
+ * @description The seed for the random number generator.
10448
+ */
10449
+ seed?: number;
10450
+ /**
10451
+ * Sync Mode
10452
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
10453
+ * @default false
10454
+ */
10455
+ sync_mode?: boolean;
10456
+ }
10457
+
10195
10458
  export interface SharedType_35f {
10196
10459
  /**
10197
10460
  * Has Nsfw Concepts
@@ -11670,6 +11933,17 @@ export interface SharedType_0ab {
11670
11933
  video_write_mode?: 'fast' | 'balanced' | 'small';
11671
11934
  }
11672
11935
 
11936
+ export interface SharedType_0a0 {
11937
+ /**
11938
+ * Video
11939
+ * @description The generated video from text using the Q3 model
11940
+ * @example {
11941
+ * "url": "https://v3b.fal.media/files/b/0a8c915a/8dtkty-vyhotySs--cTKS_output.mp4"
11942
+ * }
11943
+ */
11944
+ video: Components.File;
11945
+ }
11946
+
11673
11947
  export interface SharedType_096 {
11674
11948
  /**
11675
11949
  * Images
@@ -12069,6 +12343,51 @@ export interface SharedType_02c {
12069
12343
  start_image_url: string;
12070
12344
  }
12071
12345
 
12346
+ export interface SharedType_015 {
12347
+ /**
12348
+ * Audio
12349
+ * @description Whether to use direct audio-video generation. When true, outputs video with sound (including dialogue and sound effects).
12350
+ * @default true
12351
+ */
12352
+ audio?: boolean;
12353
+ /**
12354
+ * Duration
12355
+ * @description Duration of the video in seconds (1-16 for Q3 models)
12356
+ * @default 5
12357
+ */
12358
+ duration?: number;
12359
+ /**
12360
+ * End Image Url
12361
+ * @description URL of the image to use as the ending frame. When provided, generates a transition video between start and end frames.
12362
+ */
12363
+ end_image_url?: string;
12364
+ /**
12365
+ * Image Url
12366
+ * @description URL or base64 image to use as the starting frame
12367
+ * @example https://prod-ss-images.s3.cn-northwest-1.amazonaws.com.cn/vidu-maas/template/image2video.png
12368
+ */
12369
+ image_url: string;
12370
+ /**
12371
+ * Prompt
12372
+ * @description Text prompt for video generation, max 2000 characters
12373
+ * @default
12374
+ * @example The astronaut waved and the camera moved up.
12375
+ */
12376
+ prompt?: string;
12377
+ /**
12378
+ * Resolution
12379
+ * @description Output video resolution. Note: 360p is not available when end_image_url is provided.
12380
+ * @default 720p
12381
+ * @enum {string}
12382
+ */
12383
+ resolution?: '360p' | '540p' | '720p' | '1080p';
12384
+ /**
12385
+ * Seed
12386
+ * @description Random seed for reproducibility. If None, a random seed is chosen.
12387
+ */
12388
+ seed?: number;
12389
+ }
12390
+
12072
12391
  export interface SharedType_000 {
12073
12392
  /**
12074
12393
  * Aspect Ratio
@@ -13804,66 +14123,6 @@ export interface HalfmoonaiAiHomeEditOutput {
13804
14123
  status: string;
13805
14124
  }
13806
14125
 
13807
- export interface HalfmoonaiAiFaceSwapFaceswapvideomultiInput {
13808
- /**
13809
- * Enable Occlusion Prevention
13810
- * @description Enable occlusion prevention for handling faces covered by hands/objects. Warning: Enabling this runs an occlusion-aware model which costs 2x more.
13811
- * @default false
13812
- */
13813
- enable_occlusion_prevention?: boolean;
13814
- /**
13815
- * Source Face Url 1
13816
- * @description First source face image (will be mapped to faces matching source_gender_1). Allowed items: bmp, jpeg, png, tiff, webp
13817
- * @example https://images.pexels.com/photos/1642228/pexels-photo-1642228.jpeg
13818
- */
13819
- source_face_url_1: string;
13820
- /**
13821
- * Source Face Url 2
13822
- * @description Second source face image (optional, will be mapped to faces matching source_gender_2). If provided, source_gender_2 must also be specified. Allowed items: bmp, jpeg, png, tiff, webp
13823
- * @example https://images.pexels.com/photos/6066877/pexels-photo-6066877.jpeg
13824
- */
13825
- source_face_url_2?: string;
13826
- /**
13827
- * Source Gender 1
13828
- * @description Gender of source 1 face - faces matching this gender in target will be swapped with source_face_url_1
13829
- * @enum {string}
13830
- */
13831
- source_gender_1: 'female' | 'male';
13832
- /**
13833
- * Source Gender 2
13834
- * @description Gender of source 2 face - required when source_face_url_2 is provided. Faces matching this gender in target will be swapped with source_face_url_2
13835
- * @default male
13836
- */
13837
- source_gender_2?: 'female' | 'male';
13838
- /**
13839
- * Target Video Url
13840
- * @description Target video URL (max 25 minutes, will be truncated if longer; FPS capped at 25). Allowed items: avi, m4v, mkv, mp4, mpeg, mov, mxf, webm, wmv
13841
- * @example https://www.pexels.com/download/video/5686255/
13842
- */
13843
- target_video_url: string;
13844
- }
13845
-
13846
- export interface HalfmoonaiAiFaceSwapFaceswapvideomultiOutput {
13847
- /**
13848
- * Processing Time Ms
13849
- * @description Optional processing duration in milliseconds
13850
- */
13851
- processing_time_ms?: number;
13852
- /**
13853
- * @description Generated video result
13854
- * @example {
13855
- * "content_type": "video/mp4",
13856
- * "url": "https://v3b.fal.media/files/b/0a8c13e7/TI9qpBqF8L_9nuOajc55C_swapped_with_audio.mp4"
13857
- * }
13858
- */
13859
- video: Components.File_1;
13860
- /**
13861
- * Warning
13862
- * @description Warning message if video was modified (e.g., truncated or FPS reduced)
13863
- */
13864
- warning?: string;
13865
- }
13866
-
13867
14126
  export interface HalfmoonaiAiFaceSwapFaceswapvideoInput {
13868
14127
  /**
13869
14128
  * Enable Occlusion Prevention
@@ -13906,63 +14165,6 @@ export interface HalfmoonaiAiFaceSwapFaceswapvideoOutput {
13906
14165
  warning?: string;
13907
14166
  }
13908
14167
 
13909
- export interface HalfmoonaiAiFaceSwapFaceswapimagemultiInput {
13910
- /**
13911
- * Enable Occlusion Prevention
13912
- * @description Enable occlusion prevention for handling faces covered by hands/objects. Warning: Enabling this runs an occlusion-aware model which costs 2x more.
13913
- * @default false
13914
- */
13915
- enable_occlusion_prevention?: boolean;
13916
- /**
13917
- * Source Face Url 1
13918
- * @description First source face image (will be mapped to faces matching source_gender_1). Allowed items: bmp, jpeg, png, tiff, webp
13919
- * @example https://images.pexels.com/photos/1642228/pexels-photo-1642228.jpeg
13920
- */
13921
- source_face_url_1: string;
13922
- /**
13923
- * Source Face Url 2
13924
- * @description Second source face image (optional, will be mapped to faces matching source_gender_2). If provided, source_gender_2 must also be specified. Allowed items: bmp, jpeg, png, tiff, webp
13925
- * @example https://images.pexels.com/photos/6066877/pexels-photo-6066877.jpeg
13926
- */
13927
- source_face_url_2?: string;
13928
- /**
13929
- * Source Gender 1
13930
- * @description Gender of source 1 face - faces matching this gender in target will be swapped with source_face_url_1
13931
- * @enum {string}
13932
- */
13933
- source_gender_1: 'female' | 'male';
13934
- /**
13935
- * Source Gender 2
13936
- * @description Gender of source 2 face - required when source_face_url_2 is provided. Faces matching this gender in target will be swapped with source_face_url_2
13937
- * @default male
13938
- */
13939
- source_gender_2?: 'female' | 'male';
13940
- /**
13941
- * Target Image Url
13942
- * @description Target image URL. Allowed items: bmp, jpeg, png, tiff, webp
13943
- * @example https://images.pexels.com/photos/16280336/pexels-photo-16280336.jpeg
13944
- */
13945
- target_image_url: string;
13946
- }
13947
-
13948
- export interface HalfmoonaiAiFaceSwapFaceswapimagemultiOutput {
13949
- /**
13950
- * @description Generated image result
13951
- * @example {
13952
- * "height": 742,
13953
- * "content_type": "image/jpeg",
13954
- * "url": "https://v3b.fal.media/files/b/0a8c136c/B7HMnRi-GNwIhlRBwaHrU_b41c39d877904258903121452d9668ab.jpg",
13955
- * "width": 520
13956
- * }
13957
- */
13958
- image: Components.Image_2;
13959
- /**
13960
- * Processing Time Ms
13961
- * @description Optional processing duration in milliseconds
13962
- */
13963
- processing_time_ms?: number;
13964
- }
13965
-
13966
14168
  export interface HalfmoonaiAiFaceSwapFaceswapimageInput {
13967
14169
  /**
13968
14170
  * Enable Occlusion Prevention
@@ -15393,6 +15595,61 @@ export interface XAilabNsfwOutput {
15393
15595
  has_nsfw_concepts: boolean[];
15394
15596
  }
15395
15597
 
15598
+ export interface WorkflowUtilitiesTrimVideoInput {
15599
+ /**
15600
+ * Duration
15601
+ * @description Duration in seconds from start_time. Ignored if end_time is provided.
15602
+ * @default 2
15603
+ */
15604
+ duration?: number;
15605
+ /**
15606
+ * End Time
15607
+ * @description End time in seconds. If not provided, uses duration instead.
15608
+ * @example 3
15609
+ */
15610
+ end_time?: number;
15611
+ /**
15612
+ * Start Time
15613
+ * @description Start time in seconds
15614
+ * @default 1
15615
+ */
15616
+ start_time?: number;
15617
+ /**
15618
+ * Video Url
15619
+ * @description URL of the video file to trim
15620
+ *
15621
+ * Max file size: 476.8MB, Timeout: 30.0s
15622
+ * @example https://storage.googleapis.com/falserverless/example_outputs/wan-25-i2v-output.mp4
15623
+ */
15624
+ video_url: string;
15625
+ }
15626
+
15627
+ export interface WorkflowUtilitiesTrimVideoOutput {
15628
+ /**
15629
+ * Original Duration
15630
+ * @description Duration of the original video in seconds
15631
+ * @example 5.041667
15632
+ */
15633
+ original_duration: number;
15634
+ /**
15635
+ * Trimmed Duration
15636
+ * @description Duration of the trimmed video in seconds
15637
+ * @example 2
15638
+ */
15639
+ trimmed_duration: number;
15640
+ /**
15641
+ * Video
15642
+ * @description The trimmed video
15643
+ * @example {
15644
+ * "file_size": 2550671,
15645
+ * "file_name": "output.mp4",
15646
+ * "content_type": "application/octet-stream",
15647
+ * "url": "https://v3b.fal.media/files/b/0a8dcc60/sIg2GuYYYrmYAa2lUVgum_output.mp4"
15648
+ * }
15649
+ */
15650
+ video: Components.File;
15651
+ }
15652
+
15396
15653
  export interface WorkflowUtilitiesInterleaveVideoInput {
15397
15654
  /**
15398
15655
  * Video Urls
@@ -16638,13 +16895,12 @@ export interface WanV22A14bVideoToVideoOutput {
16638
16895
  */
16639
16896
  seed: number;
16640
16897
  /**
16641
- * Video
16642
16898
  * @description The generated video file.
16643
16899
  * @example {
16644
16900
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/wan-2.2-v2v-output.mp4"
16645
16901
  * }
16646
16902
  */
16647
- video: Components.File;
16903
+ video: Components.File_1;
16648
16904
  }
16649
16905
 
16650
16906
  export interface WanV22A14bTextToVideoTurboInput {
@@ -16738,13 +16994,12 @@ export interface WanV22A14bTextToVideoTurboOutput {
16738
16994
  */
16739
16995
  seed: number;
16740
16996
  /**
16741
- * Video
16742
16997
  * @description The generated video file.
16743
16998
  * @example {
16744
16999
  * "url": "https://storage.googleapis.com/falserverless/gallery/wan-t2v-turbo.mp4"
16745
17000
  * }
16746
17001
  */
16747
- video: Components.File;
17002
+ video: Components.File_1;
16748
17003
  }
16749
17004
 
16750
17005
  export interface WanV22A14bTextToVideoLoraInput {
@@ -16906,7 +17161,7 @@ export interface WanV22A14bTextToVideoLoraInput {
16906
17161
  video_write_mode?: 'fast' | 'balanced' | 'small';
16907
17162
  }
16908
17163
 
16909
- export interface WanV22A14bTextToVideoLoraOutput extends SharedType_ec2 {}
17164
+ export interface WanV22A14bTextToVideoLoraOutput extends SharedType_382 {}
16910
17165
 
16911
17166
  export interface WanV22A14bTextToVideoInput {
16912
17167
  /**
@@ -17055,7 +17310,7 @@ export interface WanV22A14bTextToVideoInput {
17055
17310
  video_write_mode?: 'fast' | 'balanced' | 'small';
17056
17311
  }
17057
17312
 
17058
- export interface WanV22A14bTextToVideoOutput extends SharedType_ec2 {}
17313
+ export interface WanV22A14bTextToVideoOutput extends SharedType_382 {}
17059
17314
 
17060
17315
  export interface WanV22A14bTextToImageLoraInput {
17061
17316
  /**
@@ -17170,7 +17425,7 @@ export interface WanV22A14bTextToImageLoraInput {
17170
17425
  shift?: number;
17171
17426
  }
17172
17427
 
17173
- export interface WanV22A14bTextToImageLoraOutput extends SharedType_4f3 {}
17428
+ export interface WanV22A14bTextToImageLoraOutput extends SharedType_c0f {}
17174
17429
 
17175
17430
  export interface WanV22A14bTextToImageInput {
17176
17431
  /**
@@ -17265,7 +17520,7 @@ export interface WanV22A14bTextToImageInput {
17265
17520
  shift?: number;
17266
17521
  }
17267
17522
 
17268
- export interface WanV22A14bTextToImageOutput extends SharedType_4f3 {}
17523
+ export interface WanV22A14bTextToImageOutput extends SharedType_c0f {}
17269
17524
 
17270
17525
  export interface WanV22A14bImageToVideoTurboInput {
17271
17526
  /**
@@ -17366,13 +17621,12 @@ export interface WanV22A14bImageToVideoTurboOutput {
17366
17621
  */
17367
17622
  seed: number;
17368
17623
  /**
17369
- * Video
17370
17624
  * @description The generated video file.
17371
17625
  * @example {
17372
17626
  * "url": "https://storage.googleapis.com/falserverless/gallery/wan-i2v-turbo.mp4"
17373
17627
  * }
17374
17628
  */
17375
- video: Components.File;
17629
+ video: Components.File_1;
17376
17630
  }
17377
17631
 
17378
17632
  export interface WanV22A14bImageToVideoLoraInput {
@@ -17542,7 +17796,7 @@ export interface WanV22A14bImageToVideoLoraInput {
17542
17796
  video_write_mode?: 'fast' | 'balanced' | 'small';
17543
17797
  }
17544
17798
 
17545
- export interface WanV22A14bImageToVideoLoraOutput extends SharedType_bd3 {}
17799
+ export interface WanV22A14bImageToVideoLoraOutput extends SharedType_78a {}
17546
17800
 
17547
17801
  export interface WanV22A14bImageToVideoInput {
17548
17802
  /**
@@ -17699,7 +17953,7 @@ export interface WanV22A14bImageToVideoInput {
17699
17953
  video_write_mode?: 'fast' | 'balanced' | 'small';
17700
17954
  }
17701
17955
 
17702
- export interface WanV22A14bImageToVideoOutput extends SharedType_bd3 {}
17956
+ export interface WanV22A14bImageToVideoOutput extends SharedType_78a {}
17703
17957
 
17704
17958
  export interface WanV22A14bImageToImageInput {
17705
17959
  /**
@@ -17815,13 +18069,12 @@ export interface WanV22A14bImageToImageInput {
17815
18069
 
17816
18070
  export interface WanV22A14bImageToImageOutput {
17817
18071
  /**
17818
- * Image
17819
18072
  * @description The generated image file.
17820
18073
  * @example {
17821
18074
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/wan-image-to-image-output.png"
17822
18075
  * }
17823
18076
  */
17824
- image: Components.File;
18077
+ image: Components.File_1;
17825
18078
  /**
17826
18079
  * Prompt
17827
18080
  * @description The text prompt used for image generation.
@@ -17954,7 +18207,7 @@ export interface WanV225bTextToVideoFastWanInput {
17954
18207
  video_write_mode?: 'fast' | 'balanced' | 'small';
17955
18208
  }
17956
18209
 
17957
- export interface WanV225bTextToVideoFastWanOutput extends SharedType_b5a {}
18210
+ export interface WanV225bTextToVideoFastWanOutput extends SharedType_5c5 {}
17958
18211
 
17959
18212
  export interface WanV225bTextToVideoDistillInput {
17960
18213
  /**
@@ -18082,7 +18335,7 @@ export interface WanV225bTextToVideoDistillInput {
18082
18335
  video_write_mode?: 'fast' | 'balanced' | 'small';
18083
18336
  }
18084
18337
 
18085
- export interface WanV225bTextToVideoDistillOutput extends SharedType_b5a {}
18338
+ export interface WanV225bTextToVideoDistillOutput extends SharedType_5c5 {}
18086
18339
 
18087
18340
  export interface WanV225bTextToVideoInput {
18088
18341
  /**
@@ -18216,7 +18469,7 @@ export interface WanV225bTextToVideoInput {
18216
18469
  video_write_mode?: 'fast' | 'balanced' | 'small';
18217
18470
  }
18218
18471
 
18219
- export interface WanV225bTextToVideoOutput extends SharedType_b5a {}
18472
+ export interface WanV225bTextToVideoOutput extends SharedType_5c5 {}
18220
18473
 
18221
18474
  export interface WanV225bTextToImageInput {
18222
18475
  /**
@@ -18306,13 +18559,12 @@ export interface WanV225bTextToImageInput {
18306
18559
 
18307
18560
  export interface WanV225bTextToImageOutput {
18308
18561
  /**
18309
- * Image
18310
18562
  * @description The generated image file.
18311
18563
  * @example {
18312
18564
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/wan/small-t2i-output-2.png"
18313
18565
  * }
18314
18566
  */
18315
- image: Components.File;
18567
+ image: Components.File_1;
18316
18568
  /**
18317
18569
  * Seed
18318
18570
  * @description The seed used for generation.
@@ -18469,13 +18721,12 @@ export interface WanV225bImageToVideoOutput {
18469
18721
  */
18470
18722
  seed: number;
18471
18723
  /**
18472
- * Video
18473
18724
  * @description The generated video file.
18474
18725
  * @example {
18475
18726
  * "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-small-i2v-output.mp4"
18476
18727
  * }
18477
18728
  */
18478
- video: Components.File;
18729
+ video: Components.File_1;
18479
18730
  }
18480
18731
 
18481
18732
  export interface WanV2214bSpeechToVideoInput {
@@ -18585,7 +18836,6 @@ export interface WanV2214bSpeechToVideoInput {
18585
18836
 
18586
18837
  export interface WanV2214bSpeechToVideoOutput {
18587
18838
  /**
18588
- * Video
18589
18839
  * @description The generated video file.
18590
18840
  * @example {
18591
18841
  * "file_size": 4685303,
@@ -18594,17 +18844,14 @@ export interface WanV2214bSpeechToVideoOutput {
18594
18844
  * "url": "https://v3.fal.media/files/panda/f7tXRCjvwEcVlmxHuw8kO_2c7ab2540af44eceaf5ffde4e8d094ed.mp4"
18595
18845
  * }
18596
18846
  */
18597
- video: Components.File;
18847
+ video: Components.File_1;
18598
18848
  }
18599
18849
 
18600
18850
  export interface WanV2214bAnimateReplaceInput extends SharedType_cd2 {}
18601
18851
 
18602
18852
  export interface WanV2214bAnimateReplaceOutput {
18603
- /**
18604
- * Frames Zip
18605
- * @description ZIP archive of generated frames (if requested).
18606
- */
18607
- frames_zip?: Components.File;
18853
+ /** @description ZIP archive of generated frames (if requested). */
18854
+ frames_zip?: Components.File_1;
18608
18855
  /**
18609
18856
  * Prompt
18610
18857
  * @description The prompt used for generation (auto-generated by the model)
@@ -18618,23 +18865,19 @@ export interface WanV2214bAnimateReplaceOutput {
18618
18865
  */
18619
18866
  seed: number;
18620
18867
  /**
18621
- * Video
18622
18868
  * @description The generated video file.
18623
18869
  * @example {
18624
18870
  * "url": "https://v3b.fal.media/files/b/elephant/9Ofgiju3Peb3b5hriTuBH_wan_animate_output.mp4"
18625
18871
  * }
18626
18872
  */
18627
- video: Components.File;
18873
+ video: Components.File_1;
18628
18874
  }
18629
18875
 
18630
18876
  export interface WanV2214bAnimateMoveInput extends SharedType_cd2 {}
18631
18877
 
18632
18878
  export interface WanV2214bAnimateMoveOutput {
18633
- /**
18634
- * Frames Zip
18635
- * @description ZIP archive of generated frames (if requested).
18636
- */
18637
- frames_zip?: Components.File;
18879
+ /** @description ZIP archive of generated frames (if requested). */
18880
+ frames_zip?: Components.File_1;
18638
18881
  /**
18639
18882
  * Prompt
18640
18883
  * @description The prompt used for generation (auto-generated by the model)
@@ -18648,13 +18891,12 @@ export interface WanV2214bAnimateMoveOutput {
18648
18891
  */
18649
18892
  seed: number;
18650
18893
  /**
18651
- * Video
18652
18894
  * @description The generated video file.
18653
18895
  * @example {
18654
18896
  * "url": "https://v3b.fal.media/files/b/monkey/xjJYzO0jqMi7MxufJe5tx_wan_animate_output.mp4"
18655
18897
  * }
18656
18898
  */
18657
- video: Components.File;
18899
+ video: Components.File_1;
18658
18900
  }
18659
18901
 
18660
18902
  export interface WanVisionEnhancerInput {
@@ -21183,112 +21425,21 @@ export interface ViduReferenceToImageInput extends SharedType_79c {}
21183
21425
 
21184
21426
  export interface ViduReferenceToImageOutput extends SharedType_b81 {}
21185
21427
 
21186
- export interface ViduQ3TextToVideoInput {
21187
- /**
21188
- * Aspect Ratio
21189
- * @description The aspect ratio of the output video
21190
- * @default 16:9
21191
- * @enum {string}
21192
- */
21193
- aspect_ratio?: '16:9' | '9:16' | '4:3' | '3:4' | '1:1';
21194
- /**
21195
- * Audio
21196
- * @description Whether to use direct audio-video generation. When true, outputs video with sound.
21197
- * @default true
21198
- */
21199
- audio?: boolean;
21200
- /**
21201
- * Duration
21202
- * @description Duration of the video in seconds
21203
- * @default 5
21204
- */
21205
- duration?: number;
21206
- /**
21207
- * Prompt
21208
- * @description Text prompt for video generation, max 2000 characters
21209
- * @example In an ultra-realistic fashion photography style featuring light blue and pale amber tones, an astronaut in a spacesuit walks through the fog.
21210
- */
21211
- prompt: string;
21212
- /**
21213
- * Resolution
21214
- * @description Output video resolution
21215
- * @default 720p
21216
- * @enum {string}
21217
- */
21218
- resolution?: '360p' | '540p' | '720p' | '1080p';
21219
- /**
21220
- * Seed
21221
- * @description Random seed for reproducibility. If None, a random seed is chosen.
21222
- */
21223
- seed?: number;
21224
- }
21428
+ export interface ViduQ3TextToVideoTurboInput extends SharedType_95d {}
21225
21429
 
21226
- export interface ViduQ3TextToVideoOutput {
21227
- /**
21228
- * Video
21229
- * @description The generated video from text using the Q3 model
21230
- * @example {
21231
- * "url": "https://v3b.fal.media/files/b/0a8c915a/8dtkty-vyhotySs--cTKS_output.mp4"
21232
- * }
21233
- */
21234
- video: Components.File;
21235
- }
21430
+ export interface ViduQ3TextToVideoTurboOutput extends SharedType_0a0 {}
21236
21431
 
21237
- export interface ViduQ3ImageToVideoInput {
21238
- /**
21239
- * Audio
21240
- * @description Whether to use direct audio-video generation. When true, outputs video with sound (including dialogue and sound effects).
21241
- * @default true
21242
- */
21243
- audio?: boolean;
21244
- /**
21245
- * Duration
21246
- * @description Duration of the video in seconds (1-16 for Q3 models)
21247
- * @default 5
21248
- */
21249
- duration?: number;
21250
- /**
21251
- * End Image Url
21252
- * @description URL of the image to use as the ending frame. When provided, generates a transition video between start and end frames.
21253
- */
21254
- end_image_url?: string;
21255
- /**
21256
- * Image Url
21257
- * @description URL or base64 image to use as the starting frame
21258
- * @example https://prod-ss-images.s3.cn-northwest-1.amazonaws.com.cn/vidu-maas/template/image2video.png
21259
- */
21260
- image_url: string;
21261
- /**
21262
- * Prompt
21263
- * @description Text prompt for video generation, max 2000 characters
21264
- * @default
21265
- * @example The astronaut waved and the camera moved up.
21266
- */
21267
- prompt?: string;
21268
- /**
21269
- * Resolution
21270
- * @description Output video resolution. Note: 360p is not available when end_image_url is provided.
21271
- * @default 720p
21272
- * @enum {string}
21273
- */
21274
- resolution?: '360p' | '540p' | '720p' | '1080p';
21275
- /**
21276
- * Seed
21277
- * @description Random seed for reproducibility. If None, a random seed is chosen.
21278
- */
21279
- seed?: number;
21280
- }
21432
+ export interface ViduQ3TextToVideoInput extends SharedType_95d {}
21281
21433
 
21282
- export interface ViduQ3ImageToVideoOutput {
21283
- /**
21284
- * Video
21285
- * @description The generated video from image using the Q3 model
21286
- * @example {
21287
- * "url": "https://v3b.fal.media/files/b/0a8c9189/n9z3uUDPqmU2msAtqr25-_output.mp4"
21288
- * }
21289
- */
21290
- video: Components.File;
21291
- }
21434
+ export interface ViduQ3TextToVideoOutput extends SharedType_0a0 {}
21435
+
21436
+ export interface ViduQ3ImageToVideoTurboInput extends SharedType_015 {}
21437
+
21438
+ export interface ViduQ3ImageToVideoTurboOutput extends SharedType_c61 {}
21439
+
21440
+ export interface ViduQ3ImageToVideoInput extends SharedType_015 {}
21441
+
21442
+ export interface ViduQ3ImageToVideoOutput extends SharedType_c61 {}
21292
21443
 
21293
21444
  export interface ViduQ2VideoExtensionProInput {
21294
21445
  /**
@@ -22792,7 +22943,7 @@ export interface TrellisMultiOutput extends SharedType_0f9 {}
22792
22943
  export interface Trellis2Input {
22793
22944
  /**
22794
22945
  * Decimation Target
22795
- * @description Target vertex count for mesh simplification during export
22946
+ * @description Target number of vertices in the final mesh. Lower values produce smaller files but less detail. 500k is good for most uses, reduce to 20k-50k for web/mobile.
22796
22947
  * @default 500000
22797
22948
  */
22798
22949
  decimation_target?: number;
@@ -22804,17 +22955,19 @@ export interface Trellis2Input {
22804
22955
  image_url: string;
22805
22956
  /**
22806
22957
  * Remesh
22807
- * @description Run remeshing (slower; often improves topology)
22958
+ * @description Rebuild the mesh topology for cleaner triangles. Slower but usually produces better results for downstream use (animation, 3D printing, etc).
22808
22959
  * @default true
22809
22960
  */
22810
22961
  remesh?: boolean;
22811
22962
  /**
22812
22963
  * Remesh Band
22964
+ * @description Controls how far remeshing can move vertices from the original surface. Higher values allow more smoothing but may lose fine details.
22813
22965
  * @default 1
22814
22966
  */
22815
22967
  remesh_band?: number;
22816
22968
  /**
22817
22969
  * Remesh Project
22970
+ * @description How much to project remeshed vertices back onto the original surface. 0 = no projection (smoother), 1 = full projection (preserves detail).
22818
22971
  * @default 0
22819
22972
  */
22820
22973
  remesh_project?: number;
@@ -22832,67 +22985,79 @@ export interface Trellis2Input {
22832
22985
  seed?: number;
22833
22986
  /**
22834
22987
  * Shape Slat Guidance Rescale
22988
+ * @description Dampens artifacts from high guidance in the shape stage. Increase if you see noisy geometry.
22835
22989
  * @default 0.5
22836
22990
  */
22837
22991
  shape_slat_guidance_rescale?: number;
22838
22992
  /**
22839
22993
  * Shape Slat Guidance Strength
22994
+ * @description How closely the detailed geometry follows the input image. Higher values add more detail but may introduce noise.
22840
22995
  * @default 7.5
22841
22996
  */
22842
22997
  shape_slat_guidance_strength?: number;
22843
22998
  /**
22844
22999
  * Shape Slat Rescale T
23000
+ * @description Controls noise schedule sharpness for shape refinement. Higher values produce sharper geometric details.
22845
23001
  * @default 3
22846
23002
  */
22847
23003
  shape_slat_rescale_t?: number;
22848
23004
  /**
22849
23005
  * Shape Slat Sampling Steps
23006
+ * @description Number of denoising steps for shape refinement. More steps = slower but potentially smoother geometry.
22850
23007
  * @default 12
22851
23008
  */
22852
23009
  shape_slat_sampling_steps?: number;
22853
23010
  /**
22854
23011
  * Ss Guidance Rescale
23012
+ * @description Dampens artifacts from high guidance in stage 1. Lower values allow stronger guidance effects, higher values stabilize the output.
22855
23013
  * @default 0.7
22856
23014
  */
22857
23015
  ss_guidance_rescale?: number;
22858
23016
  /**
22859
23017
  * Ss Guidance Strength
23018
+ * @description How closely the initial 3D structure follows the input image. Higher values produce more faithful but potentially noisier results.
22860
23019
  * @default 7.5
22861
23020
  */
22862
23021
  ss_guidance_strength?: number;
22863
23022
  /**
22864
23023
  * Ss Rescale T
23024
+ * @description Controls noise schedule sharpness for structure generation. Higher values produce sharper transitions.
22865
23025
  * @default 5
22866
23026
  */
22867
23027
  ss_rescale_t?: number;
22868
23028
  /**
22869
23029
  * Ss Sampling Steps
23030
+ * @description Number of denoising steps for the initial structure. More steps = slower but potentially higher quality.
22870
23031
  * @default 12
22871
23032
  */
22872
23033
  ss_sampling_steps?: number;
22873
23034
  /**
22874
23035
  * Tex Slat Guidance Rescale
23036
+ * @description Dampens artifacts from high guidance in the texture stage. Increase if textures look noisy or have color banding.
22875
23037
  * @default 0
22876
23038
  */
22877
23039
  tex_slat_guidance_rescale?: number;
22878
23040
  /**
22879
23041
  * Tex Slat Guidance Strength
23042
+ * @description How closely the texture follows the input image colors. Higher values produce more vivid but potentially oversaturated textures.
22880
23043
  * @default 1
22881
23044
  */
22882
23045
  tex_slat_guidance_strength?: number;
22883
23046
  /**
22884
23047
  * Tex Slat Rescale T
23048
+ * @description Controls noise schedule sharpness for texture generation. Higher values produce sharper texture details.
22885
23049
  * @default 3
22886
23050
  */
22887
23051
  tex_slat_rescale_t?: number;
22888
23052
  /**
22889
23053
  * Tex Slat Sampling Steps
23054
+ * @description Number of denoising steps for texture generation. More steps = slower but potentially cleaner textures.
22890
23055
  * @default 12
22891
23056
  */
22892
23057
  tex_slat_sampling_steps?: number;
22893
23058
  /**
22894
23059
  * Texture Size
22895
- * @description Texture resolution
23060
+ * @description Resolution of the texture image baked onto the mesh. Higher values capture finer surface details but produce larger files.
22896
23061
  * @default 2048
22897
23062
  * @enum {integer}
22898
23063
  */
@@ -22901,13 +23066,12 @@ export interface Trellis2Input {
22901
23066
 
22902
23067
  export interface Trellis2Output {
22903
23068
  /**
22904
- * Model Glb
22905
23069
  * @description Generated 3D GLB file
22906
23070
  * @example {
22907
23071
  * "url": "https://v3b.fal.media/files/b/0a86b61d/DNmTkiWHUQ8k-rG6aussB_trellis2_68d6300f70f34d23b69a912b5fe60487.glb"
22908
23072
  * }
22909
23073
  */
22910
- model_glb: Components.File;
23074
+ model_glb: Components.File_1;
22911
23075
  }
22912
23076
 
22913
23077
  export interface TrellisInput {
@@ -23838,6 +24002,12 @@ export interface StarVectorInput {
23838
24002
  * @example https://storage.googleapis.com/falserverless/star-vector/sample-18.png
23839
24003
  */
23840
24004
  image_url: string;
24005
+ /**
24006
+ * Max Length
24007
+ * @description Maximum number of tokens for SVG generation. Increase for complex images.
24008
+ * @default 4000
24009
+ */
24010
+ max_length?: number;
23841
24011
  /**
23842
24012
  * Seed
23843
24013
  * @description seed to be used for generation
@@ -32914,7 +33084,7 @@ export interface PixverseSwapInput {
32914
33084
  image_url: string;
32915
33085
  /**
32916
33086
  * Keyframe Id
32917
- * @description The keyframe ID (from 1 to the last frame position)
33087
+ * @description The keyframe ID to use for face/object mapping. The input video is normalized to 24 FPS before processing, so keyframe 1 = first frame, keyframe 24 = 1 second in, etc. Valid range: 1 to (duration_seconds * 24).
32918
33088
  * @default 1
32919
33089
  */
32920
33090
  keyframe_id?: number;
@@ -32938,6 +33108,11 @@ export interface PixverseSwapInput {
32938
33108
  * @enum {string}
32939
33109
  */
32940
33110
  resolution?: '360p' | '540p' | '720p';
33111
+ /**
33112
+ * Seed
33113
+ * @description Random seed for generation
33114
+ */
33115
+ seed?: number;
32941
33116
  /**
32942
33117
  * Video Url
32943
33118
  * @description URL of the external video to swap
@@ -34007,6 +34182,103 @@ export interface PhotomakerOutput {
34007
34182
  seed: number;
34008
34183
  }
34009
34184
 
34185
+ export interface PersonaplexInput {
34186
+ /**
34187
+ * Audio Url
34188
+ * @description URL to the input audio file (user's speech).
34189
+ * @example https://v3b.fal.media/files/b/0a8dd5a2/hS140ygvRuxn-eY_qPhlv_assets_test_input_service.wav
34190
+ */
34191
+ audio_url: string;
34192
+ /**
34193
+ * Prompt
34194
+ * @description Text prompt describing the AI persona and conversation context.
34195
+ * @default You are a wise and friendly teacher. Answer questions or provide advice in a clear and engaging way.
34196
+ * @example You work for SwiftPlex Appliances which is a appliance repair company and your name is Farhod Toshmatov. Information: The dishwasher model is out of stock for replacement parts; we can use an alternative part with a 3-day delay. Labor cost remains $60 per hour.
34197
+ */
34198
+ prompt?: string;
34199
+ /**
34200
+ * Seed
34201
+ * @description Random seed for reproducibility.
34202
+ */
34203
+ seed?: number;
34204
+ /**
34205
+ * Temperature Audio
34206
+ * @description Audio sampling temperature. Higher values produce more diverse outputs.
34207
+ * @default 0.8
34208
+ */
34209
+ temperature_audio?: number;
34210
+ /**
34211
+ * Temperature Text
34212
+ * @description Text sampling temperature. Higher values produce more diverse outputs.
34213
+ * @default 0.7
34214
+ */
34215
+ temperature_text?: number;
34216
+ /**
34217
+ * Top K Audio
34218
+ * @description Top-K sampling for audio tokens.
34219
+ * @default 250
34220
+ */
34221
+ top_k_audio?: number;
34222
+ /**
34223
+ * Top K Text
34224
+ * @description Top-K sampling for text tokens.
34225
+ * @default 25
34226
+ */
34227
+ top_k_text?: number;
34228
+ /**
34229
+ * Voice
34230
+ * @description Voice ID for the AI response. NAT = natural, VAR = variety. F = female, M = male.
34231
+ * @default NATF2
34232
+ * @enum {string}
34233
+ */
34234
+ voice?:
34235
+ | 'NATF0'
34236
+ | 'NATF1'
34237
+ | 'NATF2'
34238
+ | 'NATF3'
34239
+ | 'NATM0'
34240
+ | 'NATM1'
34241
+ | 'NATM2'
34242
+ | 'NATM3'
34243
+ | 'VARF0'
34244
+ | 'VARF1'
34245
+ | 'VARF2'
34246
+ | 'VARF3'
34247
+ | 'VARF4'
34248
+ | 'VARM0'
34249
+ | 'VARM1'
34250
+ | 'VARM2'
34251
+ | 'VARM3'
34252
+ | 'VARM4';
34253
+ }
34254
+
34255
+ export interface PersonaplexOutput {
34256
+ /**
34257
+ * Audio
34258
+ * @description The generated AI response audio (WAV, 24kHz).
34259
+ * @example https://v3b.fal.media/files/b/0a8dd5a4/9vFUMdauY7vobpLtXq4Np_output.wav
34260
+ */
34261
+ audio: Components.File;
34262
+ /**
34263
+ * Duration
34264
+ * @description Duration of the generated audio in seconds.
34265
+ * @example 20
34266
+ */
34267
+ duration: number;
34268
+ /**
34269
+ * Seed
34270
+ * @description The seed used for generation.
34271
+ * @example 60112277
34272
+ */
34273
+ seed: number;
34274
+ /**
34275
+ * Text
34276
+ * @description Transcribed text of the AI's response.
34277
+ * @example Thank you for calling SwiftPlex Appliances. This is Farhod, how can I help you today? I understand your concern. The replacement part is out of stock, so we can't ship it right away. We can use an alternative part that will arrive in 3 days. Would you like to proceed with that?
34278
+ */
34279
+ text: string;
34280
+ }
34281
+
34010
34282
  export interface PasdInput {
34011
34283
  /**
34012
34284
  * Conditioning Scale
@@ -34238,12 +34510,13 @@ export interface OviImageToVideoOutput {
34238
34510
  */
34239
34511
  seed: number;
34240
34512
  /**
34513
+ * Video
34241
34514
  * @description The generated video file.
34242
34515
  * @example {
34243
34516
  * "url": "https://storage.googleapis.com/falserverless/example_inputs/ovi_i2v_output.mp4"
34244
34517
  * }
34245
34518
  */
34246
- video?: Components.File_1;
34519
+ video?: Components.File;
34247
34520
  }
34248
34521
 
34249
34522
  export interface OviInput {
@@ -34299,12 +34572,13 @@ export interface OviOutput {
34299
34572
  */
34300
34573
  seed: number;
34301
34574
  /**
34575
+ * Video
34302
34576
  * @description The generated video file.
34303
34577
  * @example {
34304
34578
  * "url": "https://storage.googleapis.com/falserverless/example_inputs/ovi_t2v_output.mp4"
34305
34579
  * }
34306
34580
  */
34307
- video?: Components.File_1;
34581
+ video?: Components.File;
34308
34582
  }
34309
34583
 
34310
34584
  export interface OrpheusTtsInput {
@@ -34616,9 +34890,7 @@ export interface OmnigenV1Input {
34616
34890
  seed?: number;
34617
34891
  /**
34618
34892
  * Sync Mode
34619
- * @description If set to true, the function will wait for the image to be generated and uploaded
34620
- * before returning the response. This will increase the latency of the function but
34621
- * it allows you to get the image directly in the response without going through the CDN.
34893
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
34622
34894
  * @default false
34623
34895
  */
34624
34896
  sync_mode?: boolean;
@@ -34985,180 +35257,21 @@ export interface NemotronAsrOutput {
34985
35257
  partial?: boolean;
34986
35258
  }
34987
35259
 
34988
- export interface NanoBananaEditInput extends SharedType_813 {}
35260
+ export interface NanoBananaEditInput extends SharedType_85d {}
34989
35261
 
34990
- export interface NanoBananaEditOutput extends SharedType_98c {}
35262
+ export interface NanoBananaEditOutput extends SharedType_876 {}
34991
35263
 
34992
- export interface NanoBananaProEditInput {
34993
- /**
34994
- * Aspect Ratio
34995
- * @description The aspect ratio of the generated image.
34996
- * @default auto
34997
- * @enum {string}
34998
- */
34999
- aspect_ratio?:
35000
- | 'auto'
35001
- | '21:9'
35002
- | '16:9'
35003
- | '3:2'
35004
- | '4:3'
35005
- | '5:4'
35006
- | '1:1'
35007
- | '4:5'
35008
- | '3:4'
35009
- | '2:3'
35010
- | '9:16';
35011
- /**
35012
- * Enable Web Search
35013
- * @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
35014
- * @default false
35015
- */
35016
- enable_web_search?: boolean;
35017
- /**
35018
- * Image URLs
35019
- * @description The URLs of the images to use for image-to-image generation or image editing.
35020
- * @example [
35021
- * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
35022
- * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
35023
- * ]
35024
- */
35025
- image_urls: string[];
35026
- /**
35027
- * Limit Generations
35028
- * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
35029
- * @default false
35030
- */
35031
- limit_generations?: boolean;
35032
- /**
35033
- * Number of Images
35034
- * @description The number of images to generate.
35035
- * @default 1
35036
- */
35037
- num_images?: number;
35038
- /**
35039
- * Output Format
35040
- * @description The format of the generated image.
35041
- * @default png
35042
- * @enum {string}
35043
- */
35044
- output_format?: 'jpeg' | 'png' | 'webp';
35045
- /**
35046
- * Prompt
35047
- * @description The prompt for image editing.
35048
- * @example make a photo of the man driving the car down the california coastline
35049
- */
35050
- prompt: string;
35051
- /**
35052
- * Resolution
35053
- * @description The resolution of the image to generate.
35054
- * @default 1K
35055
- * @enum {string}
35056
- */
35057
- resolution?: '1K' | '2K' | '4K';
35058
- /**
35059
- * Safety Tolerance
35060
- * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
35061
- * @default 4
35062
- * @enum {string}
35063
- */
35064
- safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
35065
- /**
35066
- * Seed
35067
- * @description The seed for the random number generator.
35068
- */
35069
- seed?: number;
35070
- /**
35071
- * Sync Mode
35072
- * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
35073
- * @default false
35074
- */
35075
- sync_mode?: boolean;
35076
- }
35264
+ export interface NanoBananaProEditInput extends SharedType_367 {}
35077
35265
 
35078
- export interface NanoBananaProEditOutput extends SharedType_98c {}
35266
+ export interface NanoBananaProEditOutput extends SharedType_876 {}
35079
35267
 
35080
- export interface NanoBananaProInput {
35081
- /**
35082
- * Aspect Ratio
35083
- * @description The aspect ratio of the generated image. Use "auto" to let the model decide based on the prompt.
35084
- * @default 1:1
35085
- * @enum {string}
35086
- */
35087
- aspect_ratio?:
35088
- | 'auto'
35089
- | '21:9'
35090
- | '16:9'
35091
- | '3:2'
35092
- | '4:3'
35093
- | '5:4'
35094
- | '1:1'
35095
- | '4:5'
35096
- | '3:4'
35097
- | '2:3'
35098
- | '9:16';
35099
- /**
35100
- * Enable Web Search
35101
- * @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
35102
- * @default false
35103
- */
35104
- enable_web_search?: boolean;
35105
- /**
35106
- * Limit Generations
35107
- * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
35108
- * @default false
35109
- */
35110
- limit_generations?: boolean;
35111
- /**
35112
- * Number of Images
35113
- * @description The number of images to generate.
35114
- * @default 1
35115
- */
35116
- num_images?: number;
35117
- /**
35118
- * Output Format
35119
- * @description The format of the generated image.
35120
- * @default png
35121
- * @enum {string}
35122
- */
35123
- output_format?: 'jpeg' | 'png' | 'webp';
35124
- /**
35125
- * Prompt
35126
- * @description The text prompt to generate an image from.
35127
- * @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
35128
- */
35129
- prompt: string;
35130
- /**
35131
- * Resolution
35132
- * @description The resolution of the image to generate.
35133
- * @default 1K
35134
- * @enum {string}
35135
- */
35136
- resolution?: '1K' | '2K' | '4K';
35137
- /**
35138
- * Safety Tolerance
35139
- * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
35140
- * @default 4
35141
- * @enum {string}
35142
- */
35143
- safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
35144
- /**
35145
- * Seed
35146
- * @description The seed for the random number generator.
35147
- */
35148
- seed?: number;
35149
- /**
35150
- * Sync Mode
35151
- * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
35152
- * @default false
35153
- */
35154
- sync_mode?: boolean;
35155
- }
35268
+ export interface NanoBananaProInput extends SharedType_8c2 {}
35156
35269
 
35157
- export interface NanoBananaProOutput extends SharedType_662 {}
35270
+ export interface NanoBananaProOutput extends SharedType_7b9 {}
35158
35271
 
35159
- export interface NanoBananaInput extends SharedType_97e {}
35272
+ export interface NanoBananaInput extends SharedType_f57 {}
35160
35273
 
35161
- export interface NanoBananaOutput extends SharedType_662 {}
35274
+ export interface NanoBananaOutput extends SharedType_7b9 {}
35162
35275
 
35163
35276
  export interface NafnetDenoiseInput {
35164
35277
  /**
@@ -37151,19 +37264,19 @@ export interface MinimaxMusicOutput {
37151
37264
 
37152
37265
  export interface MeshyV6TextTo3dInput extends SharedType_ec0 {}
37153
37266
 
37154
- export interface MeshyV6TextTo3dOutput extends SharedType_ae3 {}
37267
+ export interface MeshyV6TextTo3dOutput extends SharedType_e49 {}
37155
37268
 
37156
37269
  export interface MeshyV6ImageTo3dInput extends SharedType_9af {}
37157
37270
 
37158
- export interface MeshyV6ImageTo3dOutput extends SharedType_748 {}
37271
+ export interface MeshyV6ImageTo3dOutput extends SharedType_8f1 {}
37159
37272
 
37160
37273
  export interface MeshyV6PreviewTextTo3dInput extends SharedType_ec0 {}
37161
37274
 
37162
- export interface MeshyV6PreviewTextTo3dOutput extends SharedType_ae3 {}
37275
+ export interface MeshyV6PreviewTextTo3dOutput extends SharedType_e49 {}
37163
37276
 
37164
37277
  export interface MeshyV6PreviewImageTo3dInput extends SharedType_9af {}
37165
37278
 
37166
- export interface MeshyV6PreviewImageTo3dOutput extends SharedType_748 {}
37279
+ export interface MeshyV6PreviewImageTo3dOutput extends SharedType_8f1 {}
37167
37280
 
37168
37281
  export interface MeshyV5RetextureInput {
37169
37282
  /**
@@ -37236,17 +37349,17 @@ export interface MeshyV5RetextureOutput {
37236
37349
  * "content_type": "model/vnd.usdz+zip",
37237
37350
  * "url": "https://v3b.fal.media/files/b/panda/4ItUhLHiH4foEw30qcWZv_model.usdz"
37238
37351
  * },
37239
- * "glb": {
37240
- * "file_size": 4097640,
37241
- * "file_name": "model.glb",
37242
- * "content_type": "model/gltf-binary",
37243
- * "url": "https://v3b.fal.media/files/b/tiger/pU0TtsRTxXM6VnKEYTHSV_model.glb"
37244
- * },
37245
37352
  * "obj": {
37246
37353
  * "file_size": 2964508,
37247
37354
  * "file_name": "model.obj",
37248
37355
  * "content_type": "text/plain",
37249
37356
  * "url": "https://v3b.fal.media/files/b/zebra/M5aK_b6vKH7KeGCZoSLq7_model.obj"
37357
+ * },
37358
+ * "glb": {
37359
+ * "file_size": 4097640,
37360
+ * "file_name": "model.glb",
37361
+ * "content_type": "model/gltf-binary",
37362
+ * "url": "https://v3b.fal.media/files/b/tiger/pU0TtsRTxXM6VnKEYTHSV_model.glb"
37250
37363
  * }
37251
37364
  * }
37252
37365
  */
@@ -37467,17 +37580,17 @@ export interface MeshyV5MultiImageTo3dOutput {
37467
37580
  * "content_type": "model/vnd.usdz+zip",
37468
37581
  * "url": "https://v3b.fal.media/files/b/panda/fSGLGmtgzUjhepklN06Zw_model.usdz"
37469
37582
  * },
37470
- * "glb": {
37471
- * "file_size": 7875308,
37472
- * "file_name": "model.glb",
37473
- * "content_type": "model/gltf-binary",
37474
- * "url": "https://v3b.fal.media/files/b/tiger/62QMEQqZ3pjUds4DfuVtX_model.glb"
37475
- * },
37476
37583
  * "obj": {
37477
37584
  * "file_size": 2761323,
37478
37585
  * "file_name": "model.obj",
37479
37586
  * "content_type": "text/plain",
37480
37587
  * "url": "https://v3b.fal.media/files/b/koala/xmOnmSeePfuROe3pqHpf0_model.obj"
37588
+ * },
37589
+ * "glb": {
37590
+ * "file_size": 7875308,
37591
+ * "file_name": "model.glb",
37592
+ * "content_type": "model/gltf-binary",
37593
+ * "url": "https://v3b.fal.media/files/b/tiger/62QMEQqZ3pjUds4DfuVtX_model.glb"
37481
37594
  * }
37482
37595
  * }
37483
37596
  */
@@ -47840,9 +47953,7 @@ export interface LcmSd15I2iInput {
47840
47953
  strength?: number;
47841
47954
  /**
47842
47955
  * Sync Mode
47843
- * @description If set to true, the function will wait for the image to be generated and uploaded
47844
- * before returning the response. This will increase the latency of the function but
47845
- * it allows you to get the image directly in the response without going through the CDN.
47956
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
47846
47957
  * @default false
47847
47958
  */
47848
47959
  sync_mode?: boolean;
@@ -49907,7 +50018,7 @@ export interface KlingVideoV1StandardImageToVideoOutput extends SharedType_b2d {
49907
50018
 
49908
50019
  export interface KlingVideoV1StandardEffectsInput extends SharedType_9db {}
49909
50020
 
49910
- export interface KlingVideoV1StandardEffectsOutput extends SharedType_95d {}
50021
+ export interface KlingVideoV1StandardEffectsOutput extends SharedType_95d1 {}
49911
50022
 
49912
50023
  export interface KlingVideoV1StandardAiAvatarInput extends SharedType_c85 {}
49913
50024
 
@@ -49961,7 +50072,7 @@ export interface KlingVideoV16StandardElementsOutput extends SharedType_d98 {}
49961
50072
 
49962
50073
  export interface KlingVideoV16StandardEffectsInput extends SharedType_9db {}
49963
50074
 
49964
- export interface KlingVideoV16StandardEffectsOutput extends SharedType_95d {}
50075
+ export interface KlingVideoV16StandardEffectsOutput extends SharedType_95d1 {}
49965
50076
 
49966
50077
  export interface KlingVideoV16ProTextToVideoInput extends SharedType_a8a {}
49967
50078
 
@@ -50019,7 +50130,7 @@ export interface KlingVideoV16ProElementsOutput extends SharedType_d98 {}
50019
50130
 
50020
50131
  export interface KlingVideoV16ProEffectsInput extends SharedType_9db {}
50021
50132
 
50022
- export interface KlingVideoV16ProEffectsOutput extends SharedType_95d {}
50133
+ export interface KlingVideoV16ProEffectsOutput extends SharedType_95d1 {}
50023
50134
 
50024
50135
  export interface KlingVideoV15ProTextToVideoInput extends SharedType_a8a {}
50025
50136
 
@@ -50084,7 +50195,7 @@ export interface KlingVideoV15ProImageToVideoOutput extends SharedType_b2d {}
50084
50195
 
50085
50196
  export interface KlingVideoV15ProEffectsInput extends SharedType_9db {}
50086
50197
 
50087
- export interface KlingVideoV15ProEffectsOutput extends SharedType_95d {}
50198
+ export interface KlingVideoV15ProEffectsOutput extends SharedType_95d1 {}
50088
50199
 
50089
50200
  export interface KlingVideoO3StandardVideoToVideoReferenceInput {
50090
50201
  /**
@@ -54138,12 +54249,6 @@ export interface ImageAppsV2VirtualTryOnOutput {
54138
54249
  * ]
54139
54250
  */
54140
54251
  images: Components.Image_2[];
54141
- /**
54142
- * Inference Time Ms
54143
- * @description Total inference time in milliseconds
54144
- * @example 15234
54145
- */
54146
- inference_time_ms: number;
54147
54252
  }
54148
54253
 
54149
54254
  export interface ImageAppsV2TextureTransformInput {
@@ -54196,12 +54301,6 @@ export interface ImageAppsV2TextureTransformOutput {
54196
54301
  * ]
54197
54302
  */
54198
54303
  images: Components.Image_2[];
54199
- /**
54200
- * Inference Time Ms
54201
- * @description Total inference time in milliseconds
54202
- * @example 15234
54203
- */
54204
- inference_time_ms: number;
54205
54304
  }
54206
54305
 
54207
54306
  export interface ImageAppsV2StyleTransferInput {
@@ -54263,12 +54362,6 @@ export interface ImageAppsV2StyleTransferOutput {
54263
54362
  * ]
54264
54363
  */
54265
54364
  images: Components.Image_2[];
54266
- /**
54267
- * Inference Time Ms
54268
- * @description Total inference time in milliseconds
54269
- * @example 15234
54270
- */
54271
- inference_time_ms: number;
54272
54365
  }
54273
54366
 
54274
54367
  export interface ImageAppsV2RelightingInput {
@@ -54317,12 +54410,6 @@ export interface ImageAppsV2RelightingOutput {
54317
54410
  * ]
54318
54411
  */
54319
54412
  images: Components.Image_2[];
54320
- /**
54321
- * Inference Time Ms
54322
- * @description Total inference time in milliseconds
54323
- * @example 15234
54324
- */
54325
- inference_time_ms: number;
54326
54413
  }
54327
54414
 
54328
54415
  export interface ImageAppsV2ProductPhotographyInput {
@@ -54347,12 +54434,6 @@ export interface ImageAppsV2ProductPhotographyOutput {
54347
54434
  * ]
54348
54435
  */
54349
54436
  images: Components.Image_2[];
54350
- /**
54351
- * Inference Time Ms
54352
- * @description Total inference time in milliseconds
54353
- * @example 15234
54354
- */
54355
- inference_time_ms: number;
54356
54437
  }
54357
54438
 
54358
54439
  export interface ImageAppsV2ProductHoldingInput {
@@ -54383,12 +54464,6 @@ export interface ImageAppsV2ProductHoldingOutput {
54383
54464
  * ]
54384
54465
  */
54385
54466
  images: Components.Image_2[];
54386
- /**
54387
- * Inference Time Ms
54388
- * @description Total inference time in milliseconds
54389
- * @example 15234
54390
- */
54391
- inference_time_ms: number;
54392
54467
  }
54393
54468
 
54394
54469
  export interface ImageAppsV2PortraitEnhanceInput {
@@ -54413,12 +54488,6 @@ export interface ImageAppsV2PortraitEnhanceOutput {
54413
54488
  * ]
54414
54489
  */
54415
54490
  images: Components.Image_2[];
54416
- /**
54417
- * Inference Time Ms
54418
- * @description Total inference time in milliseconds
54419
- * @example 15234
54420
- */
54421
- inference_time_ms: number;
54422
54491
  }
54423
54492
 
54424
54493
  export interface ImageAppsV2PhotographyEffectsInput {
@@ -54465,12 +54534,6 @@ export interface ImageAppsV2PhotographyEffectsOutput {
54465
54534
  * ]
54466
54535
  */
54467
54536
  images: Components.Image_2[];
54468
- /**
54469
- * Inference Time Ms
54470
- * @description Total inference time in milliseconds
54471
- * @example 15234
54472
- */
54473
- inference_time_ms: number;
54474
54537
  }
54475
54538
 
54476
54539
  export interface ImageAppsV2PhotoRestorationInput {
@@ -54510,12 +54573,6 @@ export interface ImageAppsV2PhotoRestorationOutput {
54510
54573
  * ]
54511
54574
  */
54512
54575
  images: Components.Image_2[];
54513
- /**
54514
- * Inference Time Ms
54515
- * @description Total inference time in milliseconds
54516
- * @example 15234
54517
- */
54518
- inference_time_ms: number;
54519
54576
  }
54520
54577
 
54521
54578
  export interface ImageAppsV2PerspectiveInput {
@@ -54555,12 +54612,6 @@ export interface ImageAppsV2PerspectiveOutput {
54555
54612
  * ]
54556
54613
  */
54557
54614
  images: Components.Image_2[];
54558
- /**
54559
- * Inference Time Ms
54560
- * @description Total inference time in milliseconds
54561
- * @example 15234
54562
- */
54563
- inference_time_ms: number;
54564
54615
  }
54565
54616
 
54566
54617
  export interface ImageAppsV2OutpaintInput {
@@ -54673,12 +54724,6 @@ export interface ImageAppsV2ObjectRemovalOutput {
54673
54724
  * ]
54674
54725
  */
54675
54726
  images: Components.Image_2[];
54676
- /**
54677
- * Inference Time Ms
54678
- * @description Total inference time in milliseconds
54679
- * @example 15234
54680
- */
54681
- inference_time_ms: number;
54682
54727
  }
54683
54728
 
54684
54729
  export interface ImageAppsV2MakeupApplicationInput {
@@ -54726,12 +54771,6 @@ export interface ImageAppsV2MakeupApplicationOutput {
54726
54771
  * ]
54727
54772
  */
54728
54773
  images: Components.Image_2[];
54729
- /**
54730
- * Inference Time Ms
54731
- * @description Total inference time in milliseconds
54732
- * @example 15234
54733
- */
54734
- inference_time_ms: number;
54735
54774
  }
54736
54775
 
54737
54776
  export interface ImageAppsV2HeadshotPhotoInput {
@@ -54762,12 +54801,6 @@ export interface ImageAppsV2HeadshotPhotoOutput {
54762
54801
  * ]
54763
54802
  */
54764
54803
  images: Components.Image_2[];
54765
- /**
54766
- * Inference Time Ms
54767
- * @description Total inference time in milliseconds
54768
- * @example 15234
54769
- */
54770
- inference_time_ms: number;
54771
54804
  }
54772
54805
 
54773
54806
  export interface ImageAppsV2HairChangeInput {
@@ -54840,12 +54873,6 @@ export interface ImageAppsV2HairChangeOutput {
54840
54873
  * ]
54841
54874
  */
54842
54875
  images: Components.Image_2[];
54843
- /**
54844
- * Inference Time Ms
54845
- * @description Total inference time in milliseconds
54846
- * @example 15234
54847
- */
54848
- inference_time_ms: number;
54849
54876
  }
54850
54877
 
54851
54878
  export interface ImageAppsV2ExpressionChangeInput {
@@ -54896,12 +54923,6 @@ export interface ImageAppsV2ExpressionChangeOutput {
54896
54923
  * ]
54897
54924
  */
54898
54925
  images: Components.Image_2[];
54899
- /**
54900
- * Inference Time Ms
54901
- * @description Total inference time in milliseconds
54902
- * @example 15234
54903
- */
54904
- inference_time_ms: number;
54905
54926
  }
54906
54927
 
54907
54928
  export interface ImageAppsV2CityTeleportInput {
@@ -54967,12 +54988,6 @@ export interface ImageAppsV2CityTeleportOutput {
54967
54988
  * ]
54968
54989
  */
54969
54990
  images: Components.Image_2[];
54970
- /**
54971
- * Inference Time Ms
54972
- * @description Total inference time in milliseconds
54973
- * @example 15234
54974
- */
54975
- inference_time_ms: number;
54976
54991
  }
54977
54992
 
54978
54993
  export interface ImageAppsV2AgeModifyInput {
@@ -55007,12 +55022,6 @@ export interface ImageAppsV2AgeModifyOutput {
55007
55022
  * ]
55008
55023
  */
55009
55024
  images: Components.Image_2[];
55010
- /**
55011
- * Inference Time Ms
55012
- * @description Total inference time in milliseconds
55013
- * @example 15234
55014
- */
55015
- inference_time_ms: number;
55016
55025
  }
55017
55026
 
55018
55027
  export interface IllusionDiffusionInput {
@@ -59373,7 +59382,7 @@ export interface GptImage15EditOutput {
59373
59382
  * }
59374
59383
  * ]
59375
59384
  */
59376
- images: Components.ImageFile[];
59385
+ images: Components.ImageFile_1[];
59377
59386
  }
59378
59387
 
59379
59388
  export interface GptImage15Input {
@@ -59440,7 +59449,7 @@ export interface GptImage15Output {
59440
59449
  * }
59441
59450
  * ]
59442
59451
  */
59443
- images: Components.ImageFile[];
59452
+ images: Components.ImageFile_1[];
59444
59453
  }
59445
59454
 
59446
59455
  export interface GptImage1MiniEditInput {
@@ -59895,117 +59904,35 @@ export interface GhiblifyOutput {
59895
59904
  image: Components.Image_2;
59896
59905
  }
59897
59906
 
59898
- export interface GeminiFlashEditMultiInput {
59899
- /**
59900
- * Input Image Urls
59901
- * @description List of URLs of input images for editing
59902
- * @example [
59903
- * "https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png"
59904
- * ]
59905
- */
59906
- input_image_urls: string[];
59907
+ export interface GenfocusAllInFocusInput {
59907
59908
  /**
59908
- * Prompt
59909
- * @description The prompt for image generation or editing
59910
- * @example Make the car black
59909
+ * Enable Safety Checker
59910
+ * @description If set to true, the safety checker will be enabled.
59911
+ * @default true
59911
59912
  */
59912
- prompt: string;
59913
- }
59914
-
59915
- export interface GeminiFlashEditMultiOutput extends SharedType_1b2 {}
59916
-
59917
- export interface GeminiFlashEditInput {
59913
+ enable_safety_checker?: boolean;
59918
59914
  /**
59919
- * Image Url
59920
- * @description Optional URL of an input image for editing. If not provided, generates a new image.
59921
- * @example https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png
59915
+ * Image URL
59916
+ * @description The URL of the image to deblur and restore to all-in-focus.
59917
+ * @example https://v3b.fal.media/files/b/0a8e76a1/AOVtMeklSm7IGoCTMd2T3.jpg
59922
59918
  */
59923
59919
  image_url: string;
59924
59920
  /**
59925
- * Prompt
59926
- * @description The prompt for image generation or editing
59927
- * @example Make the car black
59928
- */
59929
- prompt: string;
59930
- }
59931
-
59932
- export interface GeminiFlashEditOutput extends SharedType_1b2 {}
59933
-
59934
- export interface Gemini3ProImagePreviewEditInput {
59935
- /**
59936
- * Aspect Ratio
59937
- * @description The aspect ratio of the generated image.
59938
- * @default auto
59939
- */
59940
- aspect_ratio?:
59941
- | 'auto'
59942
- | '21:9'
59943
- | '16:9'
59944
- | '3:2'
59945
- | '4:3'
59946
- | '5:4'
59947
- | '1:1'
59948
- | '4:5'
59949
- | '3:4'
59950
- | '2:3'
59951
- | '9:16';
59952
- /**
59953
- * Enable Web Search
59954
- * @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
59955
- * @default false
59956
- */
59957
- enable_web_search?: boolean;
59958
- /**
59959
- * Image URLs
59960
- * @description The URLs of the images to use for image-to-image generation or image editing.
59961
- * @example [
59962
- * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
59963
- * "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
59964
- * ]
59965
- */
59966
- image_urls: string[];
59967
- /**
59968
- * Limit Generations
59969
- * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
59970
- * @default false
59971
- */
59972
- limit_generations?: boolean;
59973
- /**
59974
- * Number of Images
59975
- * @description The number of images to generate.
59976
- * @default 1
59921
+ * Num Inference Steps
59922
+ * @description The number of inference steps to perform. Lower values are faster but may reduce quality.
59923
+ * @default 28
59977
59924
  */
59978
- num_images?: number;
59925
+ num_inference_steps?: number;
59979
59926
  /**
59980
59927
  * Output Format
59981
59928
  * @description The format of the generated image.
59982
- * @default png
59983
- * @enum {string}
59984
- */
59985
- output_format?: 'jpeg' | 'png' | 'webp';
59986
- /**
59987
- * Prompt
59988
- * @description The prompt for image editing.
59989
- * @example make a photo of the man driving the car down the california coastline
59990
- */
59991
- prompt: string;
59992
- /**
59993
- * Resolution
59994
- * @description The resolution of the image to generate.
59995
- * @default 1K
59996
- * @enum {string}
59997
- */
59998
- resolution?: '1K' | '2K' | '4K';
59999
- /**
60000
- * Safety Tolerance
60001
- * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
60002
- * @default 4
59929
+ * @default jpeg
60003
59930
  * @enum {string}
60004
59931
  */
60005
- safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
59932
+ output_format?: 'jpeg' | 'png';
60006
59933
  /**
60007
59934
  * Seed
60008
- * @description The seed for the random number generator.
59935
+ * @description The same seed and the same input given to the same version of the model will output the same image every time.
60009
59936
  */
60010
59937
  seed?: number;
60011
59938
  /**
@@ -60014,94 +59941,60 @@ export interface Gemini3ProImagePreviewEditInput {
60014
59941
  * @default false
60015
59942
  */
60016
59943
  sync_mode?: boolean;
60017
- }
60018
-
60019
- export interface Gemini3ProImagePreviewEditOutput {
60020
59944
  /**
60021
- * Description
60022
- * @description The description of the generated images.
60023
- */
60024
- description: string;
60025
- /**
60026
- * Images
60027
- * @description The edited images.
60028
- * @example [
60029
- * {
60030
- * "file_name": "nano-banana-multi-edit-output.png",
60031
- * "content_type": "image/png",
60032
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-multi-edit-output.png"
60033
- * }
60034
- * ]
59945
+ * Target Long Side
59946
+ * @description Resize the image so the longer side matches this value (in pixels). If not set, the original resolution is used (aligned to 16px). Recommended range: 512 to 2000.
59947
+ * @default 512
60035
59948
  */
60036
- images: Components.ImageFile_1[];
59949
+ target_long_side?: number;
60037
59950
  }
60038
59951
 
60039
- export interface Gemini3ProImagePreviewInput {
60040
- /**
60041
- * Aspect Ratio
60042
- * @description The aspect ratio of the generated image. Use "auto" to let the model decide based on the prompt.
60043
- * @default 1:1
60044
- */
60045
- aspect_ratio?:
60046
- | 'auto'
60047
- | '21:9'
60048
- | '16:9'
60049
- | '3:2'
60050
- | '4:3'
60051
- | '5:4'
60052
- | '1:1'
60053
- | '4:5'
60054
- | '3:4'
60055
- | '2:3'
60056
- | '9:16';
60057
- /**
60058
- * Enable Web Search
60059
- * @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
60060
- * @default false
60061
- */
60062
- enable_web_search?: boolean;
59952
+ export interface GenfocusAllInFocusOutput extends SharedType_5c2 {}
59953
+
59954
+ export interface GenfocusInput {
60063
59955
  /**
60064
- * Limit Generations
60065
- * @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
60066
- * @default false
59956
+ * Blur Strength (K)
59957
+ * @description Blur strength (K value). Higher values produce stronger bokeh/defocus. Set to 0 to only perform all-in-focus (deblur) estimation without bokeh.
59958
+ * @default 8
60067
59959
  */
60068
- limit_generations?: boolean;
59960
+ blur_strength?: number;
60069
59961
  /**
60070
- * Number of Images
60071
- * @description The number of images to generate.
60072
- * @default 1
59962
+ * Enable Safety Checker
59963
+ * @description If set to true, the safety checker will be enabled.
59964
+ * @default true
60073
59965
  */
60074
- num_images?: number;
59966
+ enable_safety_checker?: boolean;
60075
59967
  /**
60076
- * Output Format
60077
- * @description The format of the generated image.
60078
- * @default png
60079
- * @enum {string}
59968
+ * Focus Point
59969
+ * @description The [x, y] coordinates of the focus point as fractions of image dimensions (0.0 to 1.0). For example, [0.5, 0.5] means center. If not provided, the center of the image is used.
59970
+ * @default [
59971
+ * 0.5,
59972
+ * 0.5
59973
+ * ]
60080
59974
  */
60081
- output_format?: 'jpeg' | 'png' | 'webp';
59975
+ focus_point?: number[];
60082
59976
  /**
60083
- * Prompt
60084
- * @description The text prompt to generate an image from.
60085
- * @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
59977
+ * Image URL
59978
+ * @description The URL of the image to refocus.
59979
+ * @example https://v3b.fal.media/files/b/0a8e76a1/AOVtMeklSm7IGoCTMd2T3.jpg
60086
59980
  */
60087
- prompt: string;
59981
+ image_url: string;
60088
59982
  /**
60089
- * Resolution
60090
- * @description The resolution of the image to generate.
60091
- * @default 1K
60092
- * @enum {string}
59983
+ * Num Inference Steps
59984
+ * @description The number of inference steps to perform. Lower values are faster but may reduce quality.
59985
+ * @default 28
60093
59986
  */
60094
- resolution?: '1K' | '2K' | '4K';
59987
+ num_inference_steps?: number;
60095
59988
  /**
60096
- * Safety Tolerance
60097
- * @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
60098
- * @default 4
59989
+ * Output Format
59990
+ * @description The format of the generated image.
59991
+ * @default jpeg
60099
59992
  * @enum {string}
60100
59993
  */
60101
- safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
59994
+ output_format?: 'jpeg' | 'png';
60102
59995
  /**
60103
59996
  * Seed
60104
- * @description The seed for the random number generator.
59997
+ * @description The same seed and the same input given to the same version of the model will output the same image every time.
60105
59998
  */
60106
59999
  seed?: number;
60107
60000
  /**
@@ -60110,35 +60003,67 @@ export interface Gemini3ProImagePreviewInput {
60110
60003
  * @default false
60111
60004
  */
60112
60005
  sync_mode?: boolean;
60113
- }
60114
-
60115
- export interface Gemini3ProImagePreviewOutput {
60116
60006
  /**
60117
- * Description
60118
- * @description The description of the generated images.
60007
+ * Target Long Side
60008
+ * @description Resize the image so the longer side matches this value (in pixels). If not set, the original resolution is used (aligned to 16px). Recommended range: 512 to 2000.
60009
+ * @default 512
60119
60010
  */
60120
- description: string;
60011
+ target_long_side?: number;
60012
+ }
60013
+
60014
+ export interface GenfocusOutput extends SharedType_5c2 {}
60015
+
60016
+ export interface GeminiFlashEditMultiInput {
60121
60017
  /**
60122
- * Images
60123
- * @description The generated images.
60018
+ * Input Image Urls
60019
+ * @description List of URLs of input images for editing
60124
60020
  * @example [
60125
- * {
60126
- * "file_name": "nano-banana-t2i-output.png",
60127
- * "content_type": "image/png",
60128
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-t2i-output.png"
60129
- * }
60021
+ * "https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png"
60130
60022
  * ]
60131
60023
  */
60132
- images: Components.ImageFile_1[];
60024
+ input_image_urls: string[];
60025
+ /**
60026
+ * Prompt
60027
+ * @description The prompt for image generation or editing
60028
+ * @example Make the car black
60029
+ */
60030
+ prompt: string;
60031
+ }
60032
+
60033
+ export interface GeminiFlashEditMultiOutput extends SharedType_1b2 {}
60034
+
60035
+ export interface GeminiFlashEditInput {
60036
+ /**
60037
+ * Image Url
60038
+ * @description Optional URL of an input image for editing. If not provided, generates a new image.
60039
+ * @example https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png
60040
+ */
60041
+ image_url: string;
60042
+ /**
60043
+ * Prompt
60044
+ * @description The prompt for image generation or editing
60045
+ * @example Make the car black
60046
+ */
60047
+ prompt: string;
60133
60048
  }
60134
60049
 
60135
- export interface Gemini25FlashImageEditInput extends SharedType_813 {}
60050
+ export interface GeminiFlashEditOutput extends SharedType_1b2 {}
60051
+
60052
+ export interface Gemini3ProImagePreviewEditInput extends SharedType_367 {}
60136
60053
 
60137
- export interface Gemini25FlashImageEditOutput extends SharedType_98c {}
60054
+ export interface Gemini3ProImagePreviewEditOutput extends SharedType_876 {}
60138
60055
 
60139
- export interface Gemini25FlashImageInput extends SharedType_97e {}
60056
+ export interface Gemini3ProImagePreviewInput extends SharedType_8c2 {}
60140
60057
 
60141
- export interface Gemini25FlashImageOutput extends SharedType_662 {}
60058
+ export interface Gemini3ProImagePreviewOutput extends SharedType_7b9 {}
60059
+
60060
+ export interface Gemini25FlashImageEditInput extends SharedType_85d {}
60061
+
60062
+ export interface Gemini25FlashImageEditOutput extends SharedType_876 {}
60063
+
60064
+ export interface Gemini25FlashImageInput extends SharedType_f57 {}
60065
+
60066
+ export interface Gemini25FlashImageOutput extends SharedType_7b9 {}
60142
60067
 
60143
60068
  export interface FramepackFlf2vInput {
60144
60069
  /**
@@ -68382,7 +68307,7 @@ export interface Flux2LoraGalleryVirtualTryonOutput {
68382
68307
  * }
68383
68308
  * ]
68384
68309
  */
68385
- images: Components.Image_2[];
68310
+ images: Components.Image[];
68386
68311
  /**
68387
68312
  * Prompt
68388
68313
  * @description The prompt used for generation
@@ -68486,7 +68411,7 @@ export interface Flux2LoraGallerySepiaVintageOutput {
68486
68411
  * }
68487
68412
  * ]
68488
68413
  */
68489
- images: Components.Image_2[];
68414
+ images: Components.Image[];
68490
68415
  /**
68491
68416
  * Prompt
68492
68417
  * @description The prompt used for generation
@@ -68590,7 +68515,7 @@ export interface Flux2LoraGallerySatelliteViewStyleOutput {
68590
68515
  * }
68591
68516
  * ]
68592
68517
  */
68593
- images: Components.Image_2[];
68518
+ images: Components.Image[];
68594
68519
  /**
68595
68520
  * Prompt
68596
68521
  * @description The prompt used for generation
@@ -68695,7 +68620,7 @@ export interface Flux2LoraGalleryRealismOutput {
68695
68620
  * }
68696
68621
  * ]
68697
68622
  */
68698
- images: Components.Image_2[];
68623
+ images: Components.Image[];
68699
68624
  /**
68700
68625
  * Prompt
68701
68626
  * @description The prompt used for generation
@@ -68816,7 +68741,7 @@ export interface Flux2LoraGalleryMultipleAnglesOutput {
68816
68741
  * }
68817
68742
  * ]
68818
68743
  */
68819
- images: Components.Image_2[];
68744
+ images: Components.Image[];
68820
68745
  /**
68821
68746
  * Prompt
68822
68747
  * @description The prompt used for generation
@@ -68921,7 +68846,7 @@ export interface Flux2LoraGalleryHdrStyleOutput {
68921
68846
  * }
68922
68847
  * ]
68923
68848
  */
68924
- images: Components.Image_2[];
68849
+ images: Components.Image[];
68925
68850
  /**
68926
68851
  * Prompt
68927
68852
  * @description The prompt used for generation
@@ -69033,7 +68958,7 @@ export interface Flux2LoraGalleryFaceToFullPortraitOutput {
69033
68958
  * }
69034
68959
  * ]
69035
68960
  */
69036
- images: Components.Image_2[];
68961
+ images: Components.Image[];
69037
68962
  /**
69038
68963
  * Prompt
69039
68964
  * @description The prompt used for generation
@@ -69137,7 +69062,7 @@ export interface Flux2LoraGalleryDigitalComicArtOutput {
69137
69062
  * }
69138
69063
  * ]
69139
69064
  */
69140
- images: Components.Image_2[];
69065
+ images: Components.Image[];
69141
69066
  /**
69142
69067
  * Prompt
69143
69068
  * @description The prompt used for generation
@@ -69241,7 +69166,7 @@ export interface Flux2LoraGalleryBallpointPenSketchOutput {
69241
69166
  * }
69242
69167
  * ]
69243
69168
  */
69244
- images: Components.Image_2[];
69169
+ images: Components.Image[];
69245
69170
  /**
69246
69171
  * Prompt
69247
69172
  * @description The prompt used for generation
@@ -69352,7 +69277,7 @@ export interface Flux2LoraGalleryApartmentStagingOutput {
69352
69277
  * }
69353
69278
  * ]
69354
69279
  */
69355
- images: Components.Image_2[];
69280
+ images: Components.Image[];
69356
69281
  /**
69357
69282
  * Prompt
69358
69283
  * @description The prompt used for generation
@@ -69465,7 +69390,7 @@ export interface Flux2LoraGalleryAddBackgroundOutput {
69465
69390
  * }
69466
69391
  * ]
69467
69392
  */
69468
- images: Components.Image_2[];
69393
+ images: Components.Image[];
69469
69394
  /**
69470
69395
  * Prompt
69471
69396
  * @description The prompt used for generation
@@ -73700,7 +73625,7 @@ export interface F5TtsOutput {
73700
73625
  * Generated Speech
73701
73626
  * @description The audio file containing the generated speech.
73702
73627
  */
73703
- audio_url: Components.AudioFile_3;
73628
+ audio_url: Components.AudioFile_2;
73704
73629
  }
73705
73630
 
73706
73631
  export interface FLiteTextureInput extends SharedType_b7d {}
@@ -75804,7 +75729,7 @@ export interface Deepfilternet3Output {
75804
75729
  * "bitrate": "192k"
75805
75730
  * }
75806
75731
  */
75807
- audio_file: Components.AudioFile_2;
75732
+ audio_file: Components.AudioFile;
75808
75733
  /**
75809
75734
  * Timings
75810
75735
  * @description Timings for each step in the pipeline.
@@ -79825,8 +79750,8 @@ export interface BagelEditOutput {
79825
79750
  * @description The edited images.
79826
79751
  * @example [
79827
79752
  * {
79828
- * "file_size": 423052,
79829
79753
  * "height": 1024,
79754
+ * "file_size": 423052,
79830
79755
  * "file_name": "hQnndOMvGSt2UsYAiV3vs.jpeg",
79831
79756
  * "content_type": "image/jpeg",
79832
79757
  * "url": "https://storage.googleapis.com/falserverless/bagel/hQnndOMvGSt2UsYAiV3vs.jpeg",
@@ -79889,8 +79814,8 @@ export interface BagelOutput {
79889
79814
  * @description The generated images.
79890
79815
  * @example [
79891
79816
  * {
79892
- * "file_size": 423052,
79893
79817
  * "height": 1024,
79818
+ * "file_size": 423052,
79894
79819
  * "file_name": "wRhCPSyiKTiLnnWvUpGIl.jpeg",
79895
79820
  * "content_type": "image/jpeg",
79896
79821
  * "url": "https://storage.googleapis.com/falserverless/bagel/wRhCPSyiKTiLnnWvUpGIl.jpeg",
@@ -80007,7 +79932,7 @@ export interface AuraSrInput {
80007
79932
  * @example 4
80008
79933
  * @enum {integer}
80009
79934
  */
80010
- upscaling_factor?: 4;
79935
+ upscale_factor?: 4;
80011
79936
  }
80012
79937
 
80013
79938
  export interface AuraSrOutput {
@@ -82462,7 +82387,7 @@ export interface BriaFiboEditEditInput {
82462
82387
  /**
82463
82388
  * Steps Num
82464
82389
  * @description Number of inference steps.
82465
- * @default 50
82390
+ * @default 30
82466
82391
  */
82467
82392
  steps_num?: number;
82468
82393
  /** @description The structured prompt to generate an image from. */