fal-endpoint-types 1.3.44 → 1.3.45

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1043,11 +1043,11 @@ export interface VeedFabric10TextOutput {
1043
1043
 
1044
1044
  export interface VeedFabric10FastInput extends SharedType_c45 {}
1045
1045
 
1046
- export interface VeedFabric10FastOutput extends SharedType_514 {}
1046
+ export interface VeedFabric10FastOutput extends SharedType_5141 {}
1047
1047
 
1048
1048
  export interface VeedFabric10Input extends SharedType_c45 {}
1049
1049
 
1050
- export interface VeedFabric10Output extends SharedType_514 {}
1050
+ export interface VeedFabric10Output extends SharedType_5141 {}
1051
1051
 
1052
1052
  export interface VeedAvatarsTextToVideoInput {
1053
1053
  /**
@@ -1763,19 +1763,6 @@ export interface SharedType_fc4 {
1763
1763
  textured_mesh?: boolean;
1764
1764
  }
1765
1765
 
1766
- export interface SharedType_fb8 {
1767
- /**
1768
- * @description The generated video
1769
- * @example {
1770
- * "url": "https://v3b.fal.media/files/b/0a875336/8p3rFiXtx3fE2TLoh59KP_output.mp4",
1771
- * "file_name": "output.mp4",
1772
- * "content_type": "video/mp4",
1773
- * "file_size": 35299865
1774
- * }
1775
- */
1776
- video: Components.File;
1777
- }
1778
-
1779
1766
  export interface SharedType_fb3 {
1780
1767
  /**
1781
1768
  * Duration
@@ -2287,6 +2274,19 @@ export interface SharedType_f50 {
2287
2274
  sync_mode?: boolean;
2288
2275
  }
2289
2276
 
2277
+ export interface SharedType_f37 {
2278
+ /**
2279
+ * @description The generated video.
2280
+ * @example {
2281
+ * "content_type": "video/mp4",
2282
+ * "url": "https://v3b.fal.media/files/b/0a8d04e2/idOb9V-Q9ujlggPSKqsfS_output.mp4",
2283
+ * "file_size": 13096952,
2284
+ * "file_name": "output.mp4"
2285
+ * }
2286
+ */
2287
+ video: Components.File;
2288
+ }
2289
+
2290
2290
  export interface SharedType_f11 {
2291
2291
  /**
2292
2292
  * Create Masks
@@ -3088,19 +3088,6 @@ export interface SharedType_e15 {
3088
3088
  zoom_factor?: number;
3089
3089
  }
3090
3090
 
3091
- export interface SharedType_e07 {
3092
- /**
3093
- * @description The generated video.
3094
- * @example {
3095
- * "url": "https://v3b.fal.media/files/b/kangaroo/3n_Lpxm_SjK5NYyBobRdS_output.mp4",
3096
- * "file_name": "output.mp4",
3097
- * "content_type": "video/mp4",
3098
- * "file_size": 28472159
3099
- * }
3100
- */
3101
- video: Components.File;
3102
- }
3103
-
3104
3091
  export interface SharedType_df1 {
3105
3092
  /**
3106
3093
  * Guidance Scale
@@ -3140,18 +3127,6 @@ export interface SharedType_df1 {
3140
3127
  textured_mesh?: boolean;
3141
3128
  }
3142
3129
 
3143
- export interface SharedType_dea {
3144
- /**
3145
- * @description The generated video
3146
- * @example {
3147
- * "url": "https://storage.googleapis.com/falserverless/kling/kling_ex.mp4.mp4",
3148
- * "file_name": "output.mp4",
3149
- * "content_type": "video/mp4"
3150
- * }
3151
- */
3152
- video: Components.File;
3153
- }
3154
-
3155
3130
  export interface SharedType_de61 {
3156
3131
  /**
3157
3132
  * Images
@@ -3188,38 +3163,17 @@ export interface SharedType_de6 {
3188
3163
  seed: number;
3189
3164
  }
3190
3165
 
3191
- export interface SharedType_dd2 {
3166
+ export interface SharedType_de0 {
3192
3167
  /**
3193
- * @description The generated SVG file containing vector glyphs.
3194
- * @example {
3195
- * "url": "https://v3b.fal.media/files/b/0a90bfc3/0f8Su2aQUzPQy4NZMDQpp_180c718ac5d8407092d0688ea2605f0c.svg",
3196
- * "content_type": "image/svg+xml",
3197
- * "file_name": "180c718ac5d8407092d0688ea2605f0c.svg",
3198
- * "file_size": 959
3199
- * }
3200
- */
3201
- image: Components.File;
3202
- /**
3203
- * Seed
3204
- * @description The seed used for generation.
3205
- * @example 1835785439
3206
- */
3207
- seed: number;
3208
- /**
3209
- * Svg Content
3210
- * @description Raw SVG content as a string.
3211
- */
3212
- svg_content: string;
3213
- /**
3214
- * Timings
3215
- * @description Timing breakdown of the generation process.
3168
+ * @description The generated video.
3216
3169
  * @example {
3217
- * "inference": 10.83
3170
+ * "content_type": "video/mp4",
3171
+ * "url": "https://v3b.fal.media/files/b/0a86603b/YAlbB2535l07BTy1wpDeI_output.mp4",
3172
+ * "file_size": 7533071,
3173
+ * "file_name": "output.mp4"
3218
3174
  * }
3219
3175
  */
3220
- timings: {
3221
- [key: string]: number;
3222
- };
3176
+ video: Components.File;
3223
3177
  }
3224
3178
 
3225
3179
  export interface SharedType_dce {
@@ -3909,61 +3863,17 @@ export interface SharedType_cd2 {
3909
3863
  video_write_mode?: 'fast' | 'balanced' | 'small';
3910
3864
  }
3911
3865
 
3912
- export interface SharedType_cc5 {
3913
- /**
3914
- * Aspect Ratio
3915
- * @description The aspect ratio of the generated video frame. If 'auto', the aspect ratio will be determined automatically based on the input video, and the closest aspect ratio to the input video will be used.
3916
- * @default auto
3917
- * @enum {string}
3918
- */
3919
- aspect_ratio?: 'auto' | '16:9' | '9:16' | '1:1';
3920
- /**
3921
- * Duration
3922
- * @description Video duration in seconds.
3923
- * @default 5
3924
- * @enum {string}
3925
- */
3926
- duration?: '3' | '4' | '5' | '6' | '7' | '8' | '9' | '10';
3866
+ export interface SharedType_ccf {
3927
3867
  /**
3928
- * Elements
3929
- * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2, etc. Maximum 4 total (elements + reference images) when using video.
3930
- * @example [
3931
- * {
3932
- * "reference_image_urls": [
3933
- * "https://v3b.fal.media/files/b/kangaroo/YMpmQkYt9xugpOTQyZW0O.png",
3934
- * "https://v3b.fal.media/files/b/zebra/d6ywajNyJ6bnpa_xBue-K.png"
3935
- * ],
3936
- * "frontal_image_url": "https://v3b.fal.media/files/b/panda/MQp-ghIqshvMZROKh9lW3.png"
3937
- * }
3938
- * ]
3939
- */
3940
- elements?: Components.OmniVideoElementInput[];
3941
- /**
3942
- * Image Urls
3943
- * @description Reference images for style/appearance. Reference in prompt as @Image1, @Image2, etc. Maximum 4 total (elements + reference images) when using video.
3944
- * @example []
3945
- */
3946
- image_urls?: string[];
3947
- /**
3948
- * Keep Audio
3949
- * @description Whether to keep the original audio from the video.
3950
- * @default false
3951
- */
3952
- keep_audio?: boolean;
3953
- /**
3954
- * Prompt
3955
- * @description Use @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
3956
- * @example Based on @Video1, generate the next shot. keep the style of the video
3957
- */
3958
- prompt: string;
3959
- /**
3960
- * Video Url
3961
- * @description Reference video URL. Only .mp4/.mov formats supported, 3-10 seconds duration, 720-2160px resolution, max 200MB.
3962
- *
3963
- * Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
3964
- * @example https://v3b.fal.media/files/b/panda/oVdiICFXY03Vbam-08Aj8_output.mp4
3868
+ * @description The generated video.
3869
+ * @example {
3870
+ * "content_type": "video/mp4",
3871
+ * "url": "https://v3b.fal.media/files/b/kangaroo/3n_Lpxm_SjK5NYyBobRdS_output.mp4",
3872
+ * "file_size": 28472159,
3873
+ * "file_name": "output.mp4"
3874
+ * }
3965
3875
  */
3966
- video_url: string;
3876
+ video: Components.File;
3967
3877
  }
3968
3878
 
3969
3879
  export interface SharedType_cbd {
@@ -4062,26 +3972,6 @@ export interface SharedType_cbd {
4062
3972
  sync_mode?: boolean;
4063
3973
  }
4064
3974
 
4065
- export interface SharedType_cbb {
4066
- /**
4067
- * Output
4068
- * @description Generated output from video processing
4069
- * @example that's the way I look at it and I don't know what you would say. Sooner or later the child gets run over.
4070
- * They seem to be too local, too provincial.
4071
- */
4072
- output: string;
4073
- /**
4074
- * @description Token usage information
4075
- * @example {
4076
- * "completion_tokens": 100,
4077
- * "total_tokens": 1100,
4078
- * "prompt_tokens": 1000,
4079
- * "cost": 0.0005
4080
- * }
4081
- */
4082
- usage: Components.UsageInfo;
4083
- }
4084
-
4085
3975
  export interface SharedType_cb4 {
4086
3976
  /**
4087
3977
  * Acceleration
@@ -4201,58 +4091,6 @@ export interface SharedType_cb0 {
4201
4091
  };
4202
4092
  }
4203
4093
 
4204
- export interface SharedType_ca4 {
4205
- /**
4206
- * Aspect Ratio
4207
- * @description The aspect ratio of the generated video frame.
4208
- * @default 16:9
4209
- * @enum {string}
4210
- */
4211
- aspect_ratio?: '16:9' | '9:16' | '1:1';
4212
- /**
4213
- * Duration
4214
- * @description Video duration in seconds.
4215
- * @default 5
4216
- * @enum {string}
4217
- */
4218
- duration?: '3' | '4' | '5' | '6' | '7' | '8' | '9' | '10';
4219
- /**
4220
- * Elements
4221
- * @description Elements (characters/objects) to include in the video. Reference in prompt as @Element1, @Element2, etc. Maximum 7 total (elements + reference images + start image).
4222
- * @example [
4223
- * {
4224
- * "reference_image_urls": [
4225
- * "https://v3b.fal.media/files/b/kangaroo/YMpmQkYt9xugpOTQyZW0O.png",
4226
- * "https://v3b.fal.media/files/b/zebra/d6ywajNyJ6bnpa_xBue-K.png"
4227
- * ],
4228
- * "frontal_image_url": "https://v3b.fal.media/files/b/panda/MQp-ghIqshvMZROKh9lW3.png"
4229
- * },
4230
- * {
4231
- * "reference_image_urls": [
4232
- * "https://v3b.fal.media/files/b/kangaroo/EBF4nWihspyv4pp6hgj7D.png"
4233
- * ],
4234
- * "frontal_image_url": "https://v3b.fal.media/files/b/koala/gSnsA7HJlgcaTyR5Ujj2H.png"
4235
- * }
4236
- * ]
4237
- */
4238
- elements?: Components.OmniVideoElementInput[];
4239
- /**
4240
- * Image Urls
4241
- * @description Additional reference images for style/appearance. Reference in prompt as @Image1, @Image2, etc. Maximum 7 total (elements + reference images + start image).
4242
- * @example [
4243
- * "https://v3b.fal.media/files/b/koala/v9COzzH23FGBYdGLgbK3u.png",
4244
- * "https://v3b.fal.media/files/b/elephant/5Is2huKQFSE7A7c5uUeUF.png"
4245
- * ]
4246
- */
4247
- image_urls?: string[];
4248
- /**
4249
- * Prompt
4250
- * @description Take @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
4251
- * @example Take @Image1 as the start frame. Start with a high-angle satellite view of the ancient greenhouse ruin surrounded by nature. The camera swoops down and flies inside the building, revealing the character from @Element1 standing in the sun-drenched center. The camera then seamlessly transitions into a smooth 180-degree orbit around the character, moving to the back view. As the open backpack comes into focus, the camera continues to push forward, zooming deep inside the bag to reveal the glowing stone from @Element2 nestled inside. Cinematic lighting, hopeful atmosphere, 35mm lens. Make sure to keep it as the style of @Image2.
4252
- */
4253
- prompt: string;
4254
- }
4255
-
4256
4094
  export interface SharedType_c97 {
4257
4095
  /**
4258
4096
  * @description The generated video from image using the Q3 model
@@ -4624,51 +4462,6 @@ export interface SharedType_bdf {
4624
4462
  video: Components.File;
4625
4463
  }
4626
4464
 
4627
- export interface SharedType_bda {
4628
- /**
4629
- * Elements
4630
- * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2, etc. Maximum 4 total (elements + reference images) when using video.
4631
- * @example [
4632
- * {
4633
- * "reference_image_urls": [
4634
- * "https://v3b.fal.media/files/b/kangaroo/YMpmQkYt9xugpOTQyZW0O.png",
4635
- * "https://v3b.fal.media/files/b/zebra/d6ywajNyJ6bnpa_xBue-K.png"
4636
- * ],
4637
- * "frontal_image_url": "https://v3b.fal.media/files/b/panda/MQp-ghIqshvMZROKh9lW3.png"
4638
- * }
4639
- * ]
4640
- */
4641
- elements?: Components.OmniVideoElementInput[];
4642
- /**
4643
- * Image Urls
4644
- * @description Reference images for style/appearance. Reference in prompt as @Image1, @Image2, etc. Maximum 4 total (elements + reference images) when using video.
4645
- * @example [
4646
- * "https://v3b.fal.media/files/b/lion/MKvhFko5_wYnfORYacNII_AgPt8v25Wt4oyKhjnhVK5.png"
4647
- * ]
4648
- */
4649
- image_urls?: string[];
4650
- /**
4651
- * Keep Audio
4652
- * @description Whether to keep the original audio from the video.
4653
- * @default false
4654
- */
4655
- keep_audio?: boolean;
4656
- /**
4657
- * Prompt
4658
- * @description Use @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
4659
- * @example Replace the character in the video with @Element1, maintaining the same movements and camera angles. Transform the landscape into @Image1
4660
- */
4661
- prompt: string;
4662
- /**
4663
- * Video Url
4664
- * @description Reference video URL. Only .mp4/.mov formats supported, 3-10 seconds duration, 720-2160px resolution, max 200MB.
4665
- *
4666
- * Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
4667
- * @example https://v3b.fal.media/files/b/rabbit/ku8_Wdpf-oTbGRq4lB5DU_output.mp4
4668
- */
4669
- video_url: string;
4670
- }
4671
-
4672
4465
  export interface SharedType_bb3 {
4673
4466
  /**
4674
4467
  * Aspect Ratio
@@ -5456,6 +5249,16 @@ export interface SharedType_aed {
5456
5249
  sync_mode?: boolean;
5457
5250
  }
5458
5251
 
5252
+ export interface SharedType_ada {
5253
+ /**
5254
+ * @description The generated music
5255
+ * @example {
5256
+ * "url": "https://v3b.fal.media/files/b/kangaroo/GJe3sQehFteUTmpDROzOs_output.mp3"
5257
+ * }
5258
+ */
5259
+ audio: Components.File;
5260
+ }
5261
+
5459
5262
  export interface SharedType_acc {
5460
5263
  /**
5461
5264
  * @description Generated 3D object file.
@@ -5693,6 +5496,19 @@ export interface SharedType_a6b {
5693
5496
  video: Components.File;
5694
5497
  }
5695
5498
 
5499
+ export interface SharedType_a63 {
5500
+ /**
5501
+ * @description The generated video.
5502
+ * @example {
5503
+ * "content_type": "video/mp4",
5504
+ * "url": "https://v3b.fal.media/files/b/panda/oVdiICFXY03Vbam-08Aj8_output.mp4",
5505
+ * "file_size": 47359974,
5506
+ * "file_name": "output.mp4"
5507
+ * }
5508
+ */
5509
+ video: Components.File;
5510
+ }
5511
+
5696
5512
  export interface SharedType_a58 {
5697
5513
  /**
5698
5514
  * Duration
@@ -5896,19 +5712,6 @@ export interface SharedType_a1f {
5896
5712
  };
5897
5713
  }
5898
5714
 
5899
- export interface SharedType_a10 {
5900
- /**
5901
- * @description The generated video.
5902
- * @example {
5903
- * "url": "https://v3b.fal.media/files/b/0a8d04e2/idOb9V-Q9ujlggPSKqsfS_output.mp4",
5904
- * "file_name": "output.mp4",
5905
- * "content_type": "video/mp4",
5906
- * "file_size": 13096952
5907
- * }
5908
- */
5909
- video: Components.File;
5910
- }
5911
-
5912
5715
  export interface SharedType_9fb {
5913
5716
  /**
5914
5717
  * Image Url
@@ -7100,6 +6903,76 @@ export interface SharedType_8c2 {
7100
6903
  video: Components.File;
7101
6904
  }
7102
6905
 
6906
+ export interface SharedType_8ac {
6907
+ /**
6908
+ * @description The generated video
6909
+ * @example {
6910
+ * "content_type": "video/mp4",
6911
+ * "url": "https://v3.fal.media/files/penguin/twy6u1yv09NvqsX0mMFM2_output.mp4",
6912
+ * "file_size": 3910577,
6913
+ * "file_name": "output.mp4"
6914
+ * }
6915
+ */
6916
+ video: Components.File;
6917
+ }
6918
+
6919
+ export interface SharedType_8aa {
6920
+ /**
6921
+ * Aspect Ratio
6922
+ * @description The aspect ratio of the generated video frame. If 'auto', the aspect ratio will be determined automatically based on the input video, and the closest aspect ratio to the input video will be used.
6923
+ * @default auto
6924
+ * @enum {string}
6925
+ */
6926
+ aspect_ratio?: 'auto' | '16:9' | '9:16' | '1:1';
6927
+ /**
6928
+ * Duration
6929
+ * @description Video duration in seconds.
6930
+ * @default 5
6931
+ * @enum {string}
6932
+ */
6933
+ duration?: '3' | '4' | '5' | '6' | '7' | '8' | '9' | '10';
6934
+ /**
6935
+ * Elements
6936
+ * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2, etc. Maximum 4 total (elements + reference images) when using video.
6937
+ * @example [
6938
+ * {
6939
+ * "frontal_image_url": "https://v3b.fal.media/files/b/panda/MQp-ghIqshvMZROKh9lW3.png",
6940
+ * "reference_image_urls": [
6941
+ * "https://v3b.fal.media/files/b/kangaroo/YMpmQkYt9xugpOTQyZW0O.png",
6942
+ * "https://v3b.fal.media/files/b/zebra/d6ywajNyJ6bnpa_xBue-K.png"
6943
+ * ]
6944
+ * }
6945
+ * ]
6946
+ */
6947
+ elements?: Components.OmniVideoElementInput[];
6948
+ /**
6949
+ * Image Urls
6950
+ * @description Reference images for style/appearance. Reference in prompt as @Image1, @Image2, etc. Maximum 4 total (elements + reference images) when using video.
6951
+ * @example []
6952
+ */
6953
+ image_urls?: string[];
6954
+ /**
6955
+ * Keep Audio
6956
+ * @description Whether to keep the original audio from the video.
6957
+ * @default false
6958
+ */
6959
+ keep_audio?: boolean;
6960
+ /**
6961
+ * Prompt
6962
+ * @description Use @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
6963
+ * @example Based on @Video1, generate the next shot. keep the style of the video
6964
+ */
6965
+ prompt: string;
6966
+ /**
6967
+ * Video Url
6968
+ * @description Reference video URL. Only .mp4/.mov formats supported, 3-10 seconds duration, 720-2160px resolution, max 200MB.
6969
+ *
6970
+ * Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
6971
+ * @example https://v3b.fal.media/files/b/panda/oVdiICFXY03Vbam-08Aj8_output.mp4
6972
+ */
6973
+ video_url: string;
6974
+ }
6975
+
7103
6976
  export interface SharedType_896 {
7104
6977
  /**
7105
6978
  * Aspect Ratio
@@ -7443,41 +7316,6 @@ export interface SharedType_842 {
7443
7316
  video: Components.File;
7444
7317
  }
7445
7318
 
7446
- export interface SharedType_836 {
7447
- /**
7448
- * Has Nsfw Concepts
7449
- * @description Whether the generated images contain NSFW concepts.
7450
- */
7451
- has_nsfw_concepts: boolean[];
7452
- /**
7453
- * Images
7454
- * @description The generated image(s).
7455
- * @example [
7456
- * {
7457
- * "height": 768,
7458
- * "width": 1024,
7459
- * "url": "https://v3b.fal.media/files/b/0a8e76a4/n8AJdfVTm-9msRC-sVQFn.jpg",
7460
- * "content_type": "image/jpeg"
7461
- * }
7462
- * ]
7463
- */
7464
- images: Components.Image_1[];
7465
- /**
7466
- * Prompt
7467
- * @description The prompt used for generating the image.
7468
- */
7469
- prompt: string;
7470
- /**
7471
- * Seed
7472
- * @description Seed of the generated Image. It will be the same value of the one passed in the
7473
- * input or the randomly generated that was used in case none was passed.
7474
- */
7475
- seed: number;
7476
- timings: {
7477
- [key: string]: number;
7478
- };
7479
- }
7480
-
7481
7319
  export interface SharedType_82f1 {
7482
7320
  /**
7483
7321
  * Has Nsfw Concepts
@@ -8389,19 +8227,6 @@ export interface SharedType_7ca {
8389
8227
  topology?: 'quad' | 'triangle';
8390
8228
  }
8391
8229
 
8392
- export interface SharedType_7c9 {
8393
- /**
8394
- * @description The generated video.
8395
- * @example {
8396
- * "url": "https://v3b.fal.media/files/b/0a86603b/YAlbB2535l07BTy1wpDeI_output.mp4",
8397
- * "file_name": "output.mp4",
8398
- * "content_type": "video/mp4",
8399
- * "file_size": 7533071
8400
- * }
8401
- */
8402
- video: Components.File;
8403
- }
8404
-
8405
8230
  export interface SharedType_7c6 {
8406
8231
  /**
8407
8232
  * Has Nsfw Concepts
@@ -8433,6 +8258,19 @@ export interface SharedType_7c2 {
8433
8258
  lottie_file: Components.File;
8434
8259
  }
8435
8260
 
8261
+ export interface SharedType_7ad {
8262
+ /**
8263
+ * @description The generated video.
8264
+ * @example {
8265
+ * "content_type": "video/mp4",
8266
+ * "url": "https://v3b.fal.media/files/b/0a8d0278/pgIO9yOXTFCTetKBsqDwX_output.mp4",
8267
+ * "file_size": 12037975,
8268
+ * "file_name": "output.mp4"
8269
+ * }
8270
+ */
8271
+ video: Components.File;
8272
+ }
8273
+
8436
8274
  export interface SharedType_79e {
8437
8275
  /**
8438
8276
  * @description The generated video file
@@ -8632,41 +8470,6 @@ export interface SharedType_74d {
8632
8470
  video: Components.VideoFile;
8633
8471
  }
8634
8472
 
8635
- export interface SharedType_735 {
8636
- /**
8637
- * Has Nsfw Concepts
8638
- * @description Whether the generated images contain NSFW concepts.
8639
- */
8640
- has_nsfw_concepts: boolean[];
8641
- /**
8642
- * Images
8643
- * @description The generated image files info.
8644
- * @example [
8645
- * {
8646
- * "width": 1024,
8647
- * "url": "https://v3.fal.media/files/zebra/mMW8_S5PeGuDXLTfIKCpG.png",
8648
- * "content_type": "image/jpeg",
8649
- * "height": 768
8650
- * }
8651
- * ]
8652
- */
8653
- images: Components.Image_1[];
8654
- /**
8655
- * Prompt
8656
- * @description The prompt used for generating the image.
8657
- */
8658
- prompt: string;
8659
- /**
8660
- * Seed
8661
- * @description Seed of the generated Image. It will be the same value of the one passed in the
8662
- * input or the randomly generated that was used in case none was passed.
8663
- */
8664
- seed: number;
8665
- timings: {
8666
- [key: string]: number;
8667
- };
8668
- }
8669
-
8670
8473
  export interface SharedType_732 {
8671
8474
  /**
8672
8475
  * Duration
@@ -8697,19 +8500,6 @@ export interface SharedType_732 {
8697
8500
  start_image_url: string;
8698
8501
  }
8699
8502
 
8700
- export interface SharedType_70f {
8701
- /**
8702
- * @description The generated video
8703
- * @example {
8704
- * "url": "https://v3.fal.media/files/penguin/twy6u1yv09NvqsX0mMFM2_output.mp4",
8705
- * "file_name": "output.mp4",
8706
- * "content_type": "video/mp4",
8707
- * "file_size": 3910577
8708
- * }
8709
- */
8710
- video: Components.File;
8711
- }
8712
-
8713
8503
  export interface SharedType_709 {
8714
8504
  /**
8715
8505
  * Duration
@@ -8744,6 +8534,34 @@ export interface SharedType_709 {
8744
8534
  video_url: string;
8745
8535
  }
8746
8536
 
8537
+ export interface SharedType_6fc {
8538
+ /**
8539
+ * Images
8540
+ * @description The edited image.
8541
+ * @example [
8542
+ * {
8543
+ * "height": 768,
8544
+ * "width": 1152,
8545
+ * "url": "https://v3b.fal.media/files/b/koala/2_gRhwfsnmNKYtZ_dveyV.jpg",
8546
+ * "content_type": "image/jpeg",
8547
+ * "file_name": "2_gRhwfsnmNKYtZ_dveyV.jpg"
8548
+ * }
8549
+ * ]
8550
+ */
8551
+ images: Components.ImageFile[];
8552
+ /**
8553
+ * Prompt
8554
+ * @description The prompt used for the inference.
8555
+ * @example The user wants to add a surfer to the wave in the illustration while preserving the original ukiyo-e woodblock art style. The surfer should be depicted mid-action, crouched low on a modern-style surfboard, carving through the crest of the wave with one arm extended for balance and the other gripping the board. Their wavy hair and athletic physique should match the dynamic motion. The background must remain unchanged, including the iconic Mount Fuji and the traditional Japanese text, to maintain the artwork's historical aesthetic and composition.
8556
+ */
8557
+ prompt: string;
8558
+ /**
8559
+ * Seed
8560
+ * @description The seed for the inference.
8561
+ */
8562
+ seed: number;
8563
+ }
8564
+
8747
8565
  export interface SharedType_6f3 {
8748
8566
  /**
8749
8567
  * Has Nsfw Concepts
@@ -9266,7 +9084,7 @@ export interface SharedType_66b {
9266
9084
  video: Components.File;
9267
9085
  }
9268
9086
 
9269
- export interface SharedType_663 {
9087
+ export interface SharedType_6631 {
9270
9088
  /**
9271
9089
  * Enable Safety Checker
9272
9090
  * @description If set to true, the safety checker will be enabled.
@@ -9331,6 +9149,41 @@ export interface SharedType_663 {
9331
9149
  sync_mode?: boolean;
9332
9150
  }
9333
9151
 
9152
+ export interface SharedType_663 {
9153
+ /**
9154
+ * Has Nsfw Concepts
9155
+ * @description Whether the generated images contain NSFW concepts.
9156
+ */
9157
+ has_nsfw_concepts: boolean[];
9158
+ /**
9159
+ * Images
9160
+ * @description The generated image(s).
9161
+ * @example [
9162
+ * {
9163
+ * "content_type": "image/jpeg",
9164
+ * "width": 1024,
9165
+ * "url": "https://v3b.fal.media/files/b/0a8e76a4/n8AJdfVTm-9msRC-sVQFn.jpg",
9166
+ * "height": 768
9167
+ * }
9168
+ * ]
9169
+ */
9170
+ images: Components.Image_1[];
9171
+ /**
9172
+ * Prompt
9173
+ * @description The prompt used for generating the image.
9174
+ */
9175
+ prompt: string;
9176
+ /**
9177
+ * Seed
9178
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
9179
+ * input or the randomly generated that was used in case none was passed.
9180
+ */
9181
+ seed: number;
9182
+ timings: {
9183
+ [key: string]: number;
9184
+ };
9185
+ }
9186
+
9334
9187
  export interface SharedType_65e {
9335
9188
  /**
9336
9189
  * Video
@@ -9482,6 +9335,51 @@ export interface SharedType_5d8 {
9482
9335
  seed?: number;
9483
9336
  }
9484
9337
 
9338
+ export interface SharedType_5d7 {
9339
+ /**
9340
+ * Elements
9341
+ * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2, etc. Maximum 4 total (elements + reference images) when using video.
9342
+ * @example [
9343
+ * {
9344
+ * "frontal_image_url": "https://v3b.fal.media/files/b/panda/MQp-ghIqshvMZROKh9lW3.png",
9345
+ * "reference_image_urls": [
9346
+ * "https://v3b.fal.media/files/b/kangaroo/YMpmQkYt9xugpOTQyZW0O.png",
9347
+ * "https://v3b.fal.media/files/b/zebra/d6ywajNyJ6bnpa_xBue-K.png"
9348
+ * ]
9349
+ * }
9350
+ * ]
9351
+ */
9352
+ elements?: Components.OmniVideoElementInput[];
9353
+ /**
9354
+ * Image Urls
9355
+ * @description Reference images for style/appearance. Reference in prompt as @Image1, @Image2, etc. Maximum 4 total (elements + reference images) when using video.
9356
+ * @example [
9357
+ * "https://v3b.fal.media/files/b/lion/MKvhFko5_wYnfORYacNII_AgPt8v25Wt4oyKhjnhVK5.png"
9358
+ * ]
9359
+ */
9360
+ image_urls?: string[];
9361
+ /**
9362
+ * Keep Audio
9363
+ * @description Whether to keep the original audio from the video.
9364
+ * @default false
9365
+ */
9366
+ keep_audio?: boolean;
9367
+ /**
9368
+ * Prompt
9369
+ * @description Use @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
9370
+ * @example Replace the character in the video with @Element1, maintaining the same movements and camera angles. Transform the landscape into @Image1
9371
+ */
9372
+ prompt: string;
9373
+ /**
9374
+ * Video Url
9375
+ * @description Reference video URL. Only .mp4/.mov formats supported, 3-10 seconds duration, 720-2160px resolution, max 200MB.
9376
+ *
9377
+ * Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
9378
+ * @example https://v3b.fal.media/files/b/rabbit/ku8_Wdpf-oTbGRq4lB5DU_output.mp4
9379
+ */
9380
+ video_url: string;
9381
+ }
9382
+
9485
9383
  export interface SharedType_5b3 {
9486
9384
  /**
9487
9385
  * @description The generated video file
@@ -10385,7 +10283,7 @@ export interface SharedType_51a {
10385
10283
  seed: number;
10386
10284
  }
10387
10285
 
10388
- export interface SharedType_514 {
10286
+ export interface SharedType_5141 {
10389
10287
  /**
10390
10288
  * @example {
10391
10289
  * "content_type": "video/mp4",
@@ -10395,6 +10293,19 @@ export interface SharedType_514 {
10395
10293
  video: Components.File;
10396
10294
  }
10397
10295
 
10296
+ export interface SharedType_514 {
10297
+ /**
10298
+ * @description The generated video
10299
+ * @example {
10300
+ * "content_type": "video/mp4",
10301
+ * "url": "https://v3b.fal.media/files/b/0a875336/8p3rFiXtx3fE2TLoh59KP_output.mp4",
10302
+ * "file_size": 35299865,
10303
+ * "file_name": "output.mp4"
10304
+ * }
10305
+ */
10306
+ video: Components.File;
10307
+ }
10308
+
10398
10309
  export interface SharedType_50b {
10399
10310
  /**
10400
10311
  * @description The generated video.
@@ -10485,6 +10396,58 @@ export interface SharedType_503 {
10485
10396
  seed?: number;
10486
10397
  }
10487
10398
 
10399
+ export interface SharedType_4fd {
10400
+ /**
10401
+ * Aspect Ratio
10402
+ * @description The aspect ratio of the generated video frame.
10403
+ * @default 16:9
10404
+ * @enum {string}
10405
+ */
10406
+ aspect_ratio?: '16:9' | '9:16' | '1:1';
10407
+ /**
10408
+ * Duration
10409
+ * @description Video duration in seconds.
10410
+ * @default 5
10411
+ * @enum {string}
10412
+ */
10413
+ duration?: '3' | '4' | '5' | '6' | '7' | '8' | '9' | '10';
10414
+ /**
10415
+ * Elements
10416
+ * @description Elements (characters/objects) to include in the video. Reference in prompt as @Element1, @Element2, etc. Maximum 7 total (elements + reference images + start image).
10417
+ * @example [
10418
+ * {
10419
+ * "frontal_image_url": "https://v3b.fal.media/files/b/panda/MQp-ghIqshvMZROKh9lW3.png",
10420
+ * "reference_image_urls": [
10421
+ * "https://v3b.fal.media/files/b/kangaroo/YMpmQkYt9xugpOTQyZW0O.png",
10422
+ * "https://v3b.fal.media/files/b/zebra/d6ywajNyJ6bnpa_xBue-K.png"
10423
+ * ]
10424
+ * },
10425
+ * {
10426
+ * "frontal_image_url": "https://v3b.fal.media/files/b/koala/gSnsA7HJlgcaTyR5Ujj2H.png",
10427
+ * "reference_image_urls": [
10428
+ * "https://v3b.fal.media/files/b/kangaroo/EBF4nWihspyv4pp6hgj7D.png"
10429
+ * ]
10430
+ * }
10431
+ * ]
10432
+ */
10433
+ elements?: Components.OmniVideoElementInput[];
10434
+ /**
10435
+ * Image Urls
10436
+ * @description Additional reference images for style/appearance. Reference in prompt as @Image1, @Image2, etc. Maximum 7 total (elements + reference images + start image).
10437
+ * @example [
10438
+ * "https://v3b.fal.media/files/b/koala/v9COzzH23FGBYdGLgbK3u.png",
10439
+ * "https://v3b.fal.media/files/b/elephant/5Is2huKQFSE7A7c5uUeUF.png"
10440
+ * ]
10441
+ */
10442
+ image_urls?: string[];
10443
+ /**
10444
+ * Prompt
10445
+ * @description Take @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
10446
+ * @example Take @Image1 as the start frame. Start with a high-angle satellite view of the ancient greenhouse ruin surrounded by nature. The camera swoops down and flies inside the building, revealing the character from @Element1 standing in the sun-drenched center. The camera then seamlessly transitions into a smooth 180-degree orbit around the character, moving to the back view. As the open backpack comes into focus, the camera continues to push forward, zooming deep inside the bag to reveal the glowing stone from @Element2 nestled inside. Cinematic lighting, hopeful atmosphere, 35mm lens. Make sure to keep it as the style of @Image2.
10447
+ */
10448
+ prompt: string;
10449
+ }
10450
+
10488
10451
  export interface SharedType_4ed {
10489
10452
  /**
10490
10453
  * @description The generated video
@@ -10511,6 +10474,26 @@ export interface SharedType_4eb {
10511
10474
  video: Components.File;
10512
10475
  }
10513
10476
 
10477
+ export interface SharedType_4cf {
10478
+ /**
10479
+ * Output
10480
+ * @description Generated output from video processing
10481
+ * @example that's the way I look at it and I don't know what you would say. Sooner or later the child gets run over.
10482
+ * They seem to be too local, too provincial.
10483
+ */
10484
+ output: string;
10485
+ /**
10486
+ * @description Token usage information
10487
+ * @example {
10488
+ * "cost": 0.0005,
10489
+ * "prompt_tokens": 1000,
10490
+ * "total_tokens": 1100,
10491
+ * "completion_tokens": 100
10492
+ * }
10493
+ */
10494
+ usage: Components.UsageInfo;
10495
+ }
10496
+
10514
10497
  export interface SharedType_4c9 {
10515
10498
  /**
10516
10499
  * Control Lora Image Url
@@ -11320,6 +11303,18 @@ export interface SharedType_411 {
11320
11303
  sync_mode?: boolean;
11321
11304
  }
11322
11305
 
11306
+ export interface SharedType_3db {
11307
+ /**
11308
+ * @description The generated video
11309
+ * @example {
11310
+ * "content_type": "video/mp4",
11311
+ * "url": "https://storage.googleapis.com/falserverless/kling/kling_ex.mp4.mp4",
11312
+ * "file_name": "output.mp4"
11313
+ * }
11314
+ */
11315
+ video: Components.File;
11316
+ }
11317
+
11323
11318
  export interface SharedType_3d6 {
11324
11319
  /**
11325
11320
  * Video
@@ -11710,19 +11705,6 @@ export interface SharedType_37c {
11710
11705
  video: Components.File;
11711
11706
  }
11712
11707
 
11713
- export interface SharedType_37b {
11714
- /**
11715
- * @description The generated video.
11716
- * @example {
11717
- * "url": "https://v3b.fal.media/files/b/koala/knryyyGF3ZVyMMrGr77CL_output.mp4",
11718
- * "file_name": "output.mp4",
11719
- * "content_type": "video/mp4",
11720
- * "file_size": 27588984
11721
- * }
11722
- */
11723
- video: Components.File;
11724
- }
11725
-
11726
11708
  export interface SharedType_377 {
11727
11709
  /**
11728
11710
  * Output
@@ -11859,36 +11841,43 @@ export interface SharedType_364 {
11859
11841
  video: Components.File;
11860
11842
  }
11861
11843
 
11862
- export interface SharedType_357 {
11863
- image: Components.Image;
11864
- }
11865
-
11866
- export interface SharedType_343 {
11844
+ export interface SharedType_360 {
11845
+ /**
11846
+ * Has Nsfw Concepts
11847
+ * @description Whether the generated images contain NSFW concepts.
11848
+ */
11849
+ has_nsfw_concepts: boolean[];
11867
11850
  /**
11868
11851
  * Images
11869
- * @description The edited image.
11852
+ * @description The generated image files info.
11870
11853
  * @example [
11871
11854
  * {
11872
- * "url": "https://v3b.fal.media/files/b/koala/2_gRhwfsnmNKYtZ_dveyV.jpg",
11873
11855
  * "height": 768,
11874
- * "width": 1152,
11875
- * "file_name": "2_gRhwfsnmNKYtZ_dveyV.jpg",
11876
- * "content_type": "image/jpeg"
11856
+ * "content_type": "image/jpeg",
11857
+ * "url": "https://v3.fal.media/files/zebra/mMW8_S5PeGuDXLTfIKCpG.png",
11858
+ * "width": 1024
11877
11859
  * }
11878
11860
  * ]
11879
11861
  */
11880
- images: Components.ImageFile[];
11862
+ images: Components.Image_1[];
11881
11863
  /**
11882
11864
  * Prompt
11883
- * @description The prompt used for the inference.
11884
- * @example The user wants to add a surfer to the wave in the illustration while preserving the original ukiyo-e woodblock art style. The surfer should be depicted mid-action, crouched low on a modern-style surfboard, carving through the crest of the wave with one arm extended for balance and the other gripping the board. Their wavy hair and athletic physique should match the dynamic motion. The background must remain unchanged, including the iconic Mount Fuji and the traditional Japanese text, to maintain the artwork's historical aesthetic and composition.
11865
+ * @description The prompt used for generating the image.
11885
11866
  */
11886
11867
  prompt: string;
11887
11868
  /**
11888
11869
  * Seed
11889
- * @description The seed for the inference.
11870
+ * @description Seed of the generated Image. It will be the same value of the one passed in the
11871
+ * input or the randomly generated that was used in case none was passed.
11890
11872
  */
11891
11873
  seed: number;
11874
+ timings: {
11875
+ [key: string]: number;
11876
+ };
11877
+ }
11878
+
11879
+ export interface SharedType_357 {
11880
+ image: Components.Image;
11892
11881
  }
11893
11882
 
11894
11883
  export interface SharedType_304 {
@@ -12969,19 +12958,6 @@ export interface SharedType_19f {
12969
12958
  thinking_level?: 'minimal' | 'high';
12970
12959
  }
12971
12960
 
12972
- export interface SharedType_198 {
12973
- /**
12974
- * @description The generated video.
12975
- * @example {
12976
- * "url": "https://v3b.fal.media/files/b/0a8d0278/pgIO9yOXTFCTetKBsqDwX_output.mp4",
12977
- * "file_name": "output.mp4",
12978
- * "content_type": "video/mp4",
12979
- * "file_size": 12037975
12980
- * }
12981
- */
12982
- video: Components.File;
12983
- }
12984
-
12985
12961
  export interface SharedType_197 {
12986
12962
  image: Components.Image;
12987
12963
  /**
@@ -13258,6 +13234,19 @@ export interface SharedType_129 {
13258
13234
  results: string;
13259
13235
  }
13260
13236
 
13237
+ export interface SharedType_11e {
13238
+ /**
13239
+ * @description The generated video.
13240
+ * @example {
13241
+ * "content_type": "video/mp4",
13242
+ * "url": "https://v3b.fal.media/files/b/koala/knryyyGF3ZVyMMrGr77CL_output.mp4",
13243
+ * "file_size": 27588984,
13244
+ * "file_name": "output.mp4"
13245
+ * }
13246
+ */
13247
+ video: Components.File;
13248
+ }
13249
+
13261
13250
  export interface SharedType_100 {
13262
13251
  /**
13263
13252
  * Prompt
@@ -13381,6 +13370,40 @@ export interface SharedType_0d8 {
13381
13370
  sync_mode?: boolean;
13382
13371
  }
13383
13372
 
13373
+ export interface SharedType_0d7 {
13374
+ /**
13375
+ * @description The generated SVG file containing vector glyphs.
13376
+ * @example {
13377
+ * "content_type": "image/svg+xml",
13378
+ * "file_size": 959,
13379
+ * "url": "https://v3b.fal.media/files/b/0a90bfc3/0f8Su2aQUzPQy4NZMDQpp_180c718ac5d8407092d0688ea2605f0c.svg",
13380
+ * "file_name": "180c718ac5d8407092d0688ea2605f0c.svg"
13381
+ * }
13382
+ */
13383
+ image: Components.File;
13384
+ /**
13385
+ * Seed
13386
+ * @description The seed used for generation.
13387
+ * @example 1835785439
13388
+ */
13389
+ seed: number;
13390
+ /**
13391
+ * Svg Content
13392
+ * @description Raw SVG content as a string.
13393
+ */
13394
+ svg_content: string;
13395
+ /**
13396
+ * Timings
13397
+ * @description Timing breakdown of the generation process.
13398
+ * @example {
13399
+ * "inference": 10.83
13400
+ * }
13401
+ */
13402
+ timings: {
13403
+ [key: string]: number;
13404
+ };
13405
+ }
13406
+
13384
13407
  export interface SharedType_0cd {
13385
13408
  /**
13386
13409
  * Acceleration
@@ -13531,19 +13554,6 @@ export interface SharedType_090 {
13531
13554
  objects: Record<string, never>[];
13532
13555
  }
13533
13556
 
13534
- export interface SharedType_089 {
13535
- /**
13536
- * @description The generated video.
13537
- * @example {
13538
- * "url": "https://v3b.fal.media/files/b/panda/oVdiICFXY03Vbam-08Aj8_output.mp4",
13539
- * "file_name": "output.mp4",
13540
- * "content_type": "video/mp4",
13541
- * "file_size": 47359974
13542
- * }
13543
- */
13544
- video: Components.File;
13545
- }
13546
-
13547
13557
  export interface SharedType_084 {
13548
13558
  /**
13549
13559
  * Acceleration
@@ -14192,7 +14202,7 @@ export interface RundiffusionfalRundiffusionPhotoFluxInput {
14192
14202
 
14193
14203
  export interface RundiffusionfalRundiffusionPhotoFluxOutput extends SharedType_7c6 {}
14194
14204
 
14195
- export interface RundiffusionfalJuggernautFluxProImageToImageInput extends SharedType_663 {}
14205
+ export interface RundiffusionfalJuggernautFluxProImageToImageInput extends SharedType_6631 {}
14196
14206
 
14197
14207
  export interface RundiffusionfalJuggernautFluxProImageToImageOutput extends SharedType_7c6 {}
14198
14208
 
@@ -14263,7 +14273,7 @@ export interface RundiffusionfalJuggernautFluxLightningInput {
14263
14273
 
14264
14274
  export interface RundiffusionfalJuggernautFluxLightningOutput extends SharedType_7c6 {}
14265
14275
 
14266
- export interface RundiffusionfalJuggernautFluxBaseImageToImageInput extends SharedType_663 {}
14276
+ export interface RundiffusionfalJuggernautFluxBaseImageToImageInput extends SharedType_6631 {}
14267
14277
 
14268
14278
  export interface RundiffusionfalJuggernautFluxBaseImageToImageOutput extends SharedType_7c6 {}
14269
14279
 
@@ -14734,10 +14744,10 @@ export interface OpenrouterRouterVisionOutput {
14734
14744
  /**
14735
14745
  * @description Token usage information
14736
14746
  * @example {
14737
- * "completion_tokens": 63,
14738
- * "total_tokens": 1403,
14747
+ * "cost": 0.0005595,
14739
14748
  * "prompt_tokens": 1340,
14740
- * "cost": 0.0005595
14749
+ * "total_tokens": 1403,
14750
+ * "completion_tokens": 63
14741
14751
  * }
14742
14752
  */
14743
14753
  usage: Components.UsageInfo;
@@ -14801,7 +14811,7 @@ export interface OpenrouterRouterVideoEnterpriseInput {
14801
14811
  video_urls?: string[];
14802
14812
  }
14803
14813
 
14804
- export interface OpenrouterRouterVideoEnterpriseOutput extends SharedType_cbb {}
14814
+ export interface OpenrouterRouterVideoEnterpriseOutput extends SharedType_4cf {}
14805
14815
 
14806
14816
  export interface OpenrouterRouterVideoInput {
14807
14817
  /**
@@ -14854,7 +14864,7 @@ export interface OpenrouterRouterVideoInput {
14854
14864
  video_urls?: string[];
14855
14865
  }
14856
14866
 
14857
- export interface OpenrouterRouterVideoOutput extends SharedType_cbb {}
14867
+ export interface OpenrouterRouterVideoOutput extends SharedType_4cf {}
14858
14868
 
14859
14869
  export interface OpenrouterRouterAudioInput {
14860
14870
  /**
@@ -14914,10 +14924,10 @@ export interface OpenrouterRouterAudioOutput {
14914
14924
  /**
14915
14925
  * @description Token usage information
14916
14926
  * @example {
14917
- * "completion_tokens": 50,
14918
- * "total_tokens": 550,
14927
+ * "cost": 0.0003,
14919
14928
  * "prompt_tokens": 500,
14920
- * "cost": 0.0003
14929
+ * "total_tokens": 550,
14930
+ * "completion_tokens": 50
14921
14931
  * }
14922
14932
  */
14923
14933
  usage: Components.UsageInfo;
@@ -14998,10 +15008,10 @@ export interface OpenrouterRouterOutput {
14998
15008
  /**
14999
15009
  * @description Token usage information
15000
15010
  * @example {
15001
- * "completion_tokens": 227,
15002
- * "total_tokens": 267,
15011
+ * "cost": 0.0005795,
15003
15012
  * "prompt_tokens": 40,
15004
- * "cost": 0.0005795
15013
+ * "total_tokens": 267,
15014
+ * "completion_tokens": 227
15005
15015
  * }
15006
15016
  */
15007
15017
  usage?: Components.UsageInfo;
@@ -24372,7 +24382,7 @@ export interface VecglypherImageToSvgInput {
24372
24382
  top_p?: number;
24373
24383
  }
24374
24384
 
24375
- export interface VecglypherImageToSvgOutput extends SharedType_dd2 {}
24385
+ export interface VecglypherImageToSvgOutput extends SharedType_0d7 {}
24376
24386
 
24377
24387
  export interface VecglypherInput {
24378
24388
  /**
@@ -24461,7 +24471,7 @@ export interface VecglypherInput {
24461
24471
  top_p?: number;
24462
24472
  }
24463
24473
 
24464
- export interface VecglypherOutput extends SharedType_dd2 {}
24474
+ export interface VecglypherOutput extends SharedType_0d7 {}
24465
24475
 
24466
24476
  export interface UsoInput {
24467
24477
  /**
@@ -25716,14 +25726,14 @@ export interface SyncLipsyncReact1Output {
25716
25726
  /**
25717
25727
  * @description The generated video with synchronized lip and facial movements.
25718
25728
  * @example {
25729
+ * "height": 1088,
25719
25730
  * "width": 1920,
25720
25731
  * "duration": 7.041667,
25721
- * "num_frames": 169,
25722
25732
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/react_1/output.mp4",
25723
25733
  * "file_name": "output.mp4",
25734
+ * "num_frames": 169,
25724
25735
  * "fps": 24,
25725
- * "content_type": "video/mp4",
25726
- * "height": 1088
25736
+ * "content_type": "video/mp4"
25727
25737
  * }
25728
25738
  */
25729
25739
  video: Components.VideoFile;
@@ -25888,10 +25898,10 @@ export interface StepxEdit2Output {
25888
25898
  * @description The generated images
25889
25899
  * @example [
25890
25900
  * {
25891
- * "width": 672,
25892
25901
  * "height": 1024,
25893
- * "url": "https://v3.fal.media/files/kangaroo/kFPr5gC_Rr9JZbTTakEMd.jpeg",
25894
- * "content_type": "image/jpeg"
25902
+ * "content_type": "image/jpeg",
25903
+ * "width": 672,
25904
+ * "url": "https://v3.fal.media/files/kangaroo/kFPr5gC_Rr9JZbTTakEMd.jpeg"
25895
25905
  * }
25896
25906
  * ]
25897
25907
  */
@@ -25996,9 +26006,9 @@ export interface Step1xEditOutput {
25996
26006
  * @example [
25997
26007
  * {
25998
26008
  * "width": 672,
25999
- * "url": "https://v3.fal.media/files/kangaroo/kFPr5gC_Rr9JZbTTakEMd.jpeg",
26000
26009
  * "content_type": "image/jpeg",
26001
- * "height": 1024
26010
+ * "height": 1024,
26011
+ * "url": "https://v3.fal.media/files/kangaroo/kFPr5gC_Rr9JZbTTakEMd.jpeg"
26002
26012
  * }
26003
26013
  * ]
26004
26014
  */
@@ -31864,7 +31874,7 @@ export interface QwenImageEditPlusLoraGalleryAddBackgroundOutput extends SharedT
31864
31874
 
31865
31875
  export interface QwenImageEditPlusLoraInput extends SharedType_479 {}
31866
31876
 
31867
- export interface QwenImageEditPlusLoraOutput extends SharedType_735 {}
31877
+ export interface QwenImageEditPlusLoraOutput extends SharedType_360 {}
31868
31878
 
31869
31879
  export interface QwenImageEditPlusInput extends SharedType_faf {}
31870
31880
 
@@ -32364,7 +32374,7 @@ export interface QwenImageEdit2509LoraGalleryAddBackgroundOutput extends SharedT
32364
32374
 
32365
32375
  export interface QwenImageEdit2509LoraInput extends SharedType_479 {}
32366
32376
 
32367
- export interface QwenImageEdit2509LoraOutput extends SharedType_735 {}
32377
+ export interface QwenImageEdit2509LoraOutput extends SharedType_360 {}
32368
32378
 
32369
32379
  export interface QwenImageEdit2509Input extends SharedType_faf {}
32370
32380
 
@@ -33296,12 +33306,12 @@ export interface Qwen3TtsVoiceDesign17bOutput {
33296
33306
  /**
33297
33307
  * @description The generated speech audio file.
33298
33308
  * @example {
33299
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/qwen3-tts/design_out.mp3",
33300
- * "sample_rate": 24000,
33301
- * "channels": 1,
33302
33309
  * "file_name": "rHFLVApz9Rdenm20UvnGf_FtjmMLBV.mp3",
33310
+ * "sample_rate": 24000,
33311
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/qwen3-tts/design_out.mp3",
33303
33312
  * "duration": 7.736875,
33304
- * "content_type": "audio/mpeg"
33313
+ * "content_type": "audio/mpeg",
33314
+ * "channels": 1
33305
33315
  * }
33306
33316
  */
33307
33317
  audio: Components.AudioFile;
@@ -33426,12 +33436,12 @@ export interface Qwen3TtsTextToSpeech17bOutput {
33426
33436
  /**
33427
33437
  * @description The generated speech audio file.
33428
33438
  * @example {
33429
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/qwen3-tts/tts_out.mp3",
33430
- * "sample_rate": 24000,
33431
- * "channels": 1,
33432
33439
  * "file_name": "n5Ynr2aFKUPw1QjLYjB_4_XEdHoD1K.mp3",
33440
+ * "sample_rate": 24000,
33441
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/qwen3-tts/tts_out.mp3",
33433
33442
  * "duration": 13.025333333333334,
33434
- * "content_type": "audio/mpeg"
33443
+ * "content_type": "audio/mpeg",
33444
+ * "channels": 1
33435
33445
  * }
33436
33446
  */
33437
33447
  audio: Components.AudioFile;
@@ -33554,12 +33564,12 @@ export interface Qwen3TtsTextToSpeech06bOutput {
33554
33564
  /**
33555
33565
  * @description The generated speech audio file.
33556
33566
  * @example {
33557
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/example_outputs/qwen3-tts/tts_out_06b.mp3",
33558
- * "sample_rate": 24000,
33559
- * "channels": 1,
33560
33567
  * "file_name": "n6Av3SeD5dFENf9-VmQ1v_is3jLh5h.mp3",
33568
+ * "sample_rate": 24000,
33569
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/example_outputs/qwen3-tts/tts_out_06b.mp3",
33561
33570
  * "duration": 9.816875,
33562
- * "content_type": "audio/mpeg"
33571
+ * "content_type": "audio/mpeg",
33572
+ * "channels": 1
33563
33573
  * }
33564
33574
  */
33565
33575
  audio: Components.AudioFile;
@@ -36453,6 +36463,199 @@ export interface PixverseExtendInput {
36453
36463
 
36454
36464
  export interface PixverseExtendOutput extends SharedType_4eb {}
36455
36465
 
36466
+ export interface PixverseC1TransitionInput {
36467
+ /**
36468
+ * Aspect Ratio
36469
+ * @description The aspect ratio of the generated video
36470
+ * @default 16:9
36471
+ * @enum {string}
36472
+ */
36473
+ aspect_ratio?: '16:9' | '4:3' | '1:1' | '3:4' | '9:16' | '2:3' | '3:2' | '21:9';
36474
+ /**
36475
+ * Duration
36476
+ * @description The duration of the generated video in seconds (1-15)
36477
+ * @default 5
36478
+ */
36479
+ duration?: number;
36480
+ /**
36481
+ * End Image Url
36482
+ * @description URL of the image to use as the last frame
36483
+ * @example https://v3.fal.media/files/kangaroo/RgedFs_WSnq5BgER7qDx1_ONrbTJ1YAGXz-9JnSsBoB_bdc8750387734bfe940319f469f7b0b2.jpg
36484
+ */
36485
+ end_image_url?: string;
36486
+ /**
36487
+ * First Image Url
36488
+ * @description URL of the image to use as the first frame
36489
+ * @example https://v3.fal.media/files/zebra/owQh2DAzk8UU7J02nr5RY_Co2P4boLv6meIZ5t9gKvL_8685da151df343ab8bf82165c928e2a5.jpg
36490
+ */
36491
+ first_image_url: string;
36492
+ /**
36493
+ * Generate Audio Switch
36494
+ * @description Enable audio generation (BGM, SFX, dialogue)
36495
+ * @default false
36496
+ */
36497
+ generate_audio_switch?: boolean;
36498
+ /**
36499
+ * Prompt
36500
+ * @description The prompt for the transition
36501
+ * @example Scene slowly transition into cat swimming under water
36502
+ */
36503
+ prompt: string;
36504
+ /**
36505
+ * Resolution
36506
+ * @description The resolution of the generated video
36507
+ * @default 720p
36508
+ * @enum {string}
36509
+ */
36510
+ resolution?: '360p' | '540p' | '720p' | '1080p';
36511
+ /**
36512
+ * Seed
36513
+ * @description The same seed and the same prompt given to the same version of the model
36514
+ * will output the same video every time.
36515
+ */
36516
+ seed?: number;
36517
+ }
36518
+
36519
+ export interface PixverseC1TransitionOutput extends SharedType_751 {}
36520
+
36521
+ export interface PixverseC1TextToVideoInput {
36522
+ /**
36523
+ * Aspect Ratio
36524
+ * @description The aspect ratio of the generated video
36525
+ * @default 16:9
36526
+ * @enum {string}
36527
+ */
36528
+ aspect_ratio?: '16:9' | '4:3' | '1:1' | '3:4' | '9:16' | '2:3' | '3:2' | '21:9';
36529
+ /**
36530
+ * Duration
36531
+ * @description The duration of the generated video in seconds (1-15)
36532
+ * @default 5
36533
+ */
36534
+ duration?: number;
36535
+ /**
36536
+ * Generate Audio Switch
36537
+ * @description Enable audio generation (BGM, SFX, dialogue)
36538
+ * @default false
36539
+ */
36540
+ generate_audio_switch?: boolean;
36541
+ /**
36542
+ * Prompt
36543
+ * @example Epic low-cut camera capture of a girl clad in ultraviolet threads, luminous diamond skin glistening under a vast moon's radiance, hyper-detailed
36544
+ */
36545
+ prompt: string;
36546
+ /**
36547
+ * Resolution
36548
+ * @description The resolution of the generated video
36549
+ * @default 720p
36550
+ * @enum {string}
36551
+ */
36552
+ resolution?: '360p' | '540p' | '720p' | '1080p';
36553
+ /**
36554
+ * Seed
36555
+ * @description The same seed and the same prompt given to the same version of the model
36556
+ * will output the same video every time.
36557
+ */
36558
+ seed?: number;
36559
+ }
36560
+
36561
+ export interface PixverseC1TextToVideoOutput extends SharedType_751 {}
36562
+
36563
+ export interface PixverseC1ReferenceToVideoInput {
36564
+ /**
36565
+ * Aspect Ratio
36566
+ * @description The aspect ratio of the generated video
36567
+ * @default 16:9
36568
+ * @enum {string}
36569
+ */
36570
+ aspect_ratio?: '16:9' | '4:3' | '1:1' | '3:4' | '9:16' | '2:3' | '3:2' | '21:9';
36571
+ /**
36572
+ * Duration
36573
+ * @description The duration of the generated video in seconds (1-15)
36574
+ * @default 5
36575
+ */
36576
+ duration?: number;
36577
+ /**
36578
+ * Generate Audio Switch
36579
+ * @description Enable audio generation (BGM, SFX, dialogue)
36580
+ * @default false
36581
+ */
36582
+ generate_audio_switch?: boolean;
36583
+ /**
36584
+ * Image References
36585
+ * @description List of reference images with type and name. Use @ref_name in the prompt to refer to each reference.
36586
+ * @example [
36587
+ * {
36588
+ * "ref_name": "character",
36589
+ * "image_url": "https://v3.fal.media/files/zebra/qL93Je8ezvzQgDOEzTjKF_KhGKZTEebZcDw6T5rwQPK_output.png",
36590
+ * "type": "subject"
36591
+ * }
36592
+ * ]
36593
+ */
36594
+ image_references: Components.ImageReference[];
36595
+ /**
36596
+ * Prompt
36597
+ * @description The prompt for the reference-to-video generation. Use @ref_name to refer to uploaded references.
36598
+ * @example A @character walks through a magical forest
36599
+ */
36600
+ prompt: string;
36601
+ /**
36602
+ * Resolution
36603
+ * @description The resolution of the generated video
36604
+ * @default 720p
36605
+ * @enum {string}
36606
+ */
36607
+ resolution?: '360p' | '540p' | '720p' | '1080p';
36608
+ /**
36609
+ * Seed
36610
+ * @description The same seed and the same prompt given to the same version of the model
36611
+ * will output the same video every time.
36612
+ */
36613
+ seed?: number;
36614
+ }
36615
+
36616
+ export interface PixverseC1ReferenceToVideoOutput extends SharedType_751 {}
36617
+
36618
+ export interface PixverseC1ImageToVideoInput {
36619
+ /**
36620
+ * Duration
36621
+ * @description The duration of the generated video in seconds (1-15)
36622
+ * @default 5
36623
+ */
36624
+ duration?: number;
36625
+ /**
36626
+ * Generate Audio Switch
36627
+ * @description Enable audio generation (BGM, SFX, dialogue)
36628
+ * @default false
36629
+ */
36630
+ generate_audio_switch?: boolean;
36631
+ /**
36632
+ * Image Url
36633
+ * @description URL of the image to use as the first frame
36634
+ * @example https://v3.fal.media/files/zebra/qL93Je8ezvzQgDOEzTjKF_KhGKZTEebZcDw6T5rwQPK_output.png
36635
+ */
36636
+ image_url: string;
36637
+ /**
36638
+ * Prompt
36639
+ * @example A woman warrior with her hammer walking with his glacier wolf.
36640
+ */
36641
+ prompt: string;
36642
+ /**
36643
+ * Resolution
36644
+ * @description The resolution of the generated video
36645
+ * @default 720p
36646
+ * @enum {string}
36647
+ */
36648
+ resolution?: '360p' | '540p' | '720p' | '1080p';
36649
+ /**
36650
+ * Seed
36651
+ * @description The same seed and the same prompt given to the same version of the model
36652
+ * will output the same video every time.
36653
+ */
36654
+ seed?: number;
36655
+ }
36656
+
36657
+ export interface PixverseC1ImageToVideoOutput extends SharedType_751 {}
36658
+
36456
36659
  export interface PixartSigmaInput {
36457
36660
  /**
36458
36661
  * Enable Safety Checker
@@ -37179,9 +37382,9 @@ export interface PiflowOutput {
37179
37382
  * @description The URLs of the generated images.
37180
37383
  * @example [
37181
37384
  * {
37385
+ * "url": "https://v3b.fal.media/files/b/monkey/hfFo8wc77eSDchDUDxFEi.jpg",
37182
37386
  * "height": 1024,
37183
37387
  * "content_type": "image/jpeg",
37184
- * "url": "https://v3b.fal.media/files/b/monkey/hfFo8wc77eSDchDUDxFEi.jpg",
37185
37388
  * "width": 1024
37186
37389
  * }
37187
37390
  * ]
@@ -37262,10 +37465,10 @@ export interface PhysicEditOutput {
37262
37465
  * @description The edited images.
37263
37466
  * @example [
37264
37467
  * {
37265
- * "height": 1152,
37266
37468
  * "url": "https://v3b.fal.media/files/b/0a91cf62/DdpKY15kkD6N6v66uJ0q1.jpg",
37267
- * "content_type": "image/jpeg",
37268
- * "width": 864
37469
+ * "width": 864,
37470
+ * "height": 1152,
37471
+ * "content_type": "image/jpeg"
37269
37472
  * }
37270
37473
  * ]
37271
37474
  */
@@ -37774,6 +37977,93 @@ export interface PersonaplexOutput {
37774
37977
  text: string;
37775
37978
  }
37776
37979
 
37980
+ export interface PatinaInput {
37981
+ /**
37982
+ * Enable Safety Checker
37983
+ * @description Enable the safety checker for images.
37984
+ * @default true
37985
+ */
37986
+ enable_safety_checker?: boolean;
37987
+ /**
37988
+ * Image URL
37989
+ * @description URL of the input image (photograph or render).
37990
+ * @example https://storage.googleapis.com/falserverless/gallery/patina-blog-hero-render.png
37991
+ */
37992
+ image_url: string;
37993
+ /**
37994
+ * Maps
37995
+ * @description Which PBR maps to predict. Defaults to all five.
37996
+ * @default [
37997
+ * "basecolor",
37998
+ * "normal",
37999
+ * "roughness",
38000
+ * "metalness",
38001
+ * "height"
38002
+ * ]
38003
+ */
38004
+ maps?: ('basecolor' | 'normal' | 'roughness' | 'metalness' | 'height')[];
38005
+ /**
38006
+ * Output Format
38007
+ * @description Output image format.
38008
+ * @default png
38009
+ * @enum {string}
38010
+ */
38011
+ output_format?: 'jpeg' | 'png' | 'webp';
38012
+ /**
38013
+ * Seed
38014
+ * @description Random seed for reproducible denoising. If not set, a random seed is used.
38015
+ */
38016
+ seed?: number;
38017
+ /**
38018
+ * Sync Mode
38019
+ * @description If True, return images as data URIs instead of CDN URLs.
38020
+ * @default false
38021
+ */
38022
+ sync_mode?: boolean;
38023
+ }
38024
+
38025
+ export interface PatinaOutput {
38026
+ /**
38027
+ * Images
38028
+ * @description Predicted PBR material maps.
38029
+ * @example [
38030
+ * {
38031
+ * "map_type": "basecolor",
38032
+ * "url": "https://storage.googleapis.com/falserverless/gallery/patina-blog-hero-basecolor.png"
38033
+ * },
38034
+ * {
38035
+ * "map_type": "normal",
38036
+ * "url": "https://storage.googleapis.com/falserverless/gallery/patina-blog-hero-normal.png"
38037
+ * },
38038
+ * {
38039
+ * "map_type": "roughness",
38040
+ * "url": "https://storage.googleapis.com/falserverless/gallery/patina-blog-hero-roughness.png"
38041
+ * },
38042
+ * {
38043
+ * "map_type": "metalness",
38044
+ * "url": "https://storage.googleapis.com/falserverless/gallery/patina-blog-hero-metalness.png"
38045
+ * },
38046
+ * {
38047
+ * "map_type": "height",
38048
+ * "url": "https://storage.googleapis.com/falserverless/gallery/patina-blog-hero-height.png"
38049
+ * }
38050
+ * ]
38051
+ */
38052
+ images: Components.MapImageFile[];
38053
+ /**
38054
+ * Seed
38055
+ * @description The seed used for denoising.
38056
+ */
38057
+ seed: number;
38058
+ /**
38059
+ * Timings
38060
+ * @description Timing breakdown (seconds).
38061
+ */
38062
+ timings?: {
38063
+ [key: string]: number;
38064
+ };
38065
+ }
38066
+
37777
38067
  export interface PasdInput {
37778
38068
  /**
37779
38069
  * Conditioning Scale
@@ -38198,9 +38488,9 @@ export interface OnerewardOutput {
38198
38488
  * @example [
38199
38489
  * {
38200
38490
  * "url": "https://v3b.fal.media/files/b/0a90fdd0/wrnlNlrLJxbLDKsW2W2jn.jpg",
38201
- * "content_type": "image/jpeg",
38202
38491
  * "width": 1024,
38203
- * "height": 1024
38492
+ * "height": 1024,
38493
+ * "content_type": "image/jpeg"
38204
38494
  * }
38205
38495
  * ]
38206
38496
  */
@@ -39199,8 +39489,8 @@ export interface MultishotMasterOutput {
39199
39489
  * @description The generated multi-shot narrative video.
39200
39490
  * @example {
39201
39491
  * "file_size": 837114,
39202
- * "url": "https://v3b.fal.media/files/b/0a8fb858/x6xEQNBP9m61inmR6p_WH_output.mp4",
39203
39492
  * "file_name": "output.mp4",
39493
+ * "url": "https://v3b.fal.media/files/b/0a8fb858/x6xEQNBP9m61inmR6p_WH_output.mp4",
39204
39494
  * "content_type": "video/mp4"
39205
39495
  * }
39206
39496
  */
@@ -40937,6 +41227,70 @@ export interface MinimaxHailuo02FastImageToVideoOutput {
40937
41227
  video: Components.File;
40938
41228
  }
40939
41229
 
41230
+ export interface MinimaxMusicV26Input {
41231
+ audio_setting?: Components.AudioSetting25;
41232
+ /**
41233
+ * Is Instrumental
41234
+ * @description When true, generates vocal-free instrumental music.
41235
+ * @default false
41236
+ */
41237
+ is_instrumental?: boolean;
41238
+ /**
41239
+ * Lyrics
41240
+ * @description Lyrics of the song. Use \n to separate lines. Supports structure tags: [Intro], [Verse], [Pre Chorus], [Chorus], [Post Chorus], [Hook], [Bridge], [Interlude], [Transition], [Build Up], [Break], [Inst], [Solo], [Outro]. Max 1000 characters. Required when is_instrumental is false.
41241
+ * @default
41242
+ * @example [verse]
41243
+ * Streetlights flicker, the night breeze sighs
41244
+ * Shadows stretch as I walk alone
41245
+ * [chorus]
41246
+ * Wandering, longing, where should I go
41247
+ */
41248
+ lyrics?: string;
41249
+ /**
41250
+ * Prompt
41251
+ * @description A description of the music style, mood, genre, and scenario. 10-300 characters.
41252
+ * @example City Pop, 80s retro, groovy synth bass, warm female vocal, 104 BPM, nostalgic urban night
41253
+ */
41254
+ prompt: string;
41255
+ }
41256
+
41257
+ export interface MinimaxMusicV26Output extends SharedType_ada {}
41258
+
41259
+ export interface MinimaxMusicV25Input {
41260
+ audio_setting?: Components.AudioSetting25;
41261
+ /**
41262
+ * Is Instrumental
41263
+ * @description When true, generates vocal-free instrumental music.
41264
+ * @default false
41265
+ */
41266
+ is_instrumental?: boolean;
41267
+ /**
41268
+ * Lyrics
41269
+ * @description Lyrics of the song. Use \n to separate lines. Supports structure tags: [Intro], [Verse], [Pre Chorus], [Chorus], [Bridge], [Outro], [Interlude], [Hook], [Break], [Solo], [Inst]. Max 3500 characters. Required when is_instrumental is false and lyrics_optimizer is false.
41270
+ * @default
41271
+ * @example [verse]
41272
+ * Streetlights flicker, the night breeze sighs
41273
+ * Shadows stretch as I walk alone
41274
+ * [chorus]
41275
+ * Wandering, longing, where should I go
41276
+ */
41277
+ lyrics?: string;
41278
+ /**
41279
+ * Lyrics Optimizer
41280
+ * @description When true and lyrics is empty, auto-generates lyrics from the prompt.
41281
+ * @default false
41282
+ */
41283
+ lyrics_optimizer?: boolean;
41284
+ /**
41285
+ * Prompt
41286
+ * @description A description of the music style, mood, genre, and scenario. Max 2000 characters.
41287
+ * @example Indie folk, melancholic, introspective, longing, solitary walk, coffee shop
41288
+ */
41289
+ prompt: string;
41290
+ }
41291
+
41292
+ export interface MinimaxMusicV25Output extends SharedType_ada {}
41293
+
40940
41294
  export interface MinimaxMusicV2Input {
40941
41295
  audio_setting?: Components.AudioSetting_1;
40942
41296
  /**
@@ -44744,9 +45098,9 @@ export interface Ltx23TextToVideoFastOutput {
44744
45098
  /**
44745
45099
  * @description The generated video file
44746
45100
  * @example {
44747
- * "url": "https://v3b.fal.media/files/b/0a90dde7/7ufClzCcdMD-up1rau1jF_6xx4WLgg.mp4",
45101
+ * "file_name": "7ufClzCcdMD-up1rau1jF_6xx4WLgg.mp4",
44748
45102
  * "content_type": "video/mp4",
44749
- * "file_name": "7ufClzCcdMD-up1rau1jF_6xx4WLgg.mp4"
45103
+ * "url": "https://v3b.fal.media/files/b/0a90dde7/7ufClzCcdMD-up1rau1jF_6xx4WLgg.mp4"
44750
45104
  * }
44751
45105
  */
44752
45106
  video: Components.VideoFile;
@@ -44799,9 +45153,9 @@ export interface Ltx23TextToVideoOutput {
44799
45153
  /**
44800
45154
  * @description The generated video file
44801
45155
  * @example {
44802
- * "url": "https://v3b.fal.media/files/b/0a90dd3a/EG257tf7j9UfVc-jHF8lI_0MWboc13.mp4",
45156
+ * "file_name": "EG257tf7j9UfVc-jHF8lI_0MWboc13.mp4",
44803
45157
  * "content_type": "video/mp4",
44804
- * "file_name": "EG257tf7j9UfVc-jHF8lI_0MWboc13.mp4"
45158
+ * "url": "https://v3b.fal.media/files/b/0a90dd3a/EG257tf7j9UfVc-jHF8lI_0MWboc13.mp4"
44805
45159
  * }
44806
45160
  */
44807
45161
  video: Components.VideoFile;
@@ -44845,9 +45199,9 @@ export interface Ltx23RetakeVideoOutput {
44845
45199
  /**
44846
45200
  * @description The generated video file
44847
45201
  * @example {
44848
- * "url": "https://v3b.fal.media/files/b/0a90e00f/xvl9SlF6b7a8327wKJDvm_JW6crfr3.mp4",
45202
+ * "file_name": "xvl9SlF6b7a8327wKJDvm_JW6crfr3.mp4",
44849
45203
  * "content_type": "video/mp4",
44850
- * "file_name": "xvl9SlF6b7a8327wKJDvm_JW6crfr3.mp4"
45204
+ * "url": "https://v3b.fal.media/files/b/0a90e00f/xvl9SlF6b7a8327wKJDvm_JW6crfr3.mp4"
44851
45205
  * }
44852
45206
  */
44853
45207
  video: Components.VideoFile;
@@ -44911,9 +45265,9 @@ export interface Ltx23ImageToVideoFastOutput {
44911
45265
  /**
44912
45266
  * @description The generated video file
44913
45267
  * @example {
44914
- * "url": "https://v3b.fal.media/files/b/0a90dd0b/JCP0MJDH3ioOFzRd7GsGi_aB0OxLmA.mp4",
45268
+ * "file_name": "JCP0MJDH3ioOFzRd7GsGi_aB0OxLmA.mp4",
44915
45269
  * "content_type": "video/mp4",
44916
- * "file_name": "JCP0MJDH3ioOFzRd7GsGi_aB0OxLmA.mp4"
45270
+ * "url": "https://v3b.fal.media/files/b/0a90dd0b/JCP0MJDH3ioOFzRd7GsGi_aB0OxLmA.mp4"
44917
45271
  * }
44918
45272
  */
44919
45273
  video: Components.VideoFile;
@@ -44977,9 +45331,9 @@ export interface Ltx23ImageToVideoOutput {
44977
45331
  /**
44978
45332
  * @description The generated video file
44979
45333
  * @example {
44980
- * "url": "https://v3b.fal.media/files/b/0a90dd31/VUCTD0YYJIp_K4qQkkneL_wMVDWNRn.mp4",
45334
+ * "file_name": "VUCTD0YYJIp_K4qQkkneL_wMVDWNRn.mp4",
44981
45335
  * "content_type": "video/mp4",
44982
- * "file_name": "VUCTD0YYJIp_K4qQkkneL_wMVDWNRn.mp4"
45336
+ * "url": "https://v3b.fal.media/files/b/0a90dd31/VUCTD0YYJIp_K4qQkkneL_wMVDWNRn.mp4"
44983
45337
  * }
44984
45338
  */
44985
45339
  video: Components.VideoFile;
@@ -45022,9 +45376,9 @@ export interface Ltx23ExtendVideoOutput {
45022
45376
  /**
45023
45377
  * @description The extended video file
45024
45378
  * @example {
45025
- * "url": "https://v3b.fal.media/files/b/0a90e020/sQT2ORTL4ISEeLv3mRdF0_rGO62gwz.mp4",
45379
+ * "file_name": "sQT2ORTL4ISEeLv3mRdF0_rGO62gwz.mp4",
45026
45380
  * "content_type": "video/mp4",
45027
- * "file_name": "sQT2ORTL4ISEeLv3mRdF0_rGO62gwz.mp4"
45381
+ * "url": "https://v3b.fal.media/files/b/0a90e020/sQT2ORTL4ISEeLv3mRdF0_rGO62gwz.mp4"
45028
45382
  * }
45029
45383
  */
45030
45384
  video: Components.VideoFile;
@@ -45067,9 +45421,9 @@ export interface Ltx23AudioToVideoOutput {
45067
45421
  /**
45068
45422
  * @description The generated video file
45069
45423
  * @example {
45070
- * "url": "https://v3b.fal.media/files/b/0a90dfec/qqP0ClY8E3d83X8y0zLse_shyIe9XU.mp4",
45424
+ * "file_name": "qqP0ClY8E3d83X8y0zLse_shyIe9XU.mp4",
45071
45425
  * "content_type": "video/mp4",
45072
- * "file_name": "qqP0ClY8E3d83X8y0zLse_shyIe9XU.mp4"
45426
+ * "url": "https://v3b.fal.media/files/b/0a90dfec/qqP0ClY8E3d83X8y0zLse_shyIe9XU.mp4"
45073
45427
  * }
45074
45428
  */
45075
45429
  video: Components.VideoFile;
@@ -48795,7 +49149,7 @@ export interface LoraInpaintInput {
48795
49149
  negative_prompt?: string;
48796
49150
  /**
48797
49151
  * Noise Strength
48798
- * @description The amount of noise to add to noise image for image. Only used if the image_url is provided. 1.0 is complete noise and 0 is no noise.
49152
+ * @description The amount of noise to add to noise image for image. Only used if the image_url is provided. 1.0 is complete noise and values near 0 are minimal noise.
48799
49153
  * @default 0.5
48800
49154
  */
48801
49155
  noise_strength?: number;
@@ -49059,7 +49413,7 @@ export interface LoraImageToImageInput {
49059
49413
  negative_prompt?: string;
49060
49414
  /**
49061
49415
  * Noise Strength
49062
- * @description The amount of noise to add to noise image for image. Only used if the image_url is provided. 1.0 is complete noise and 0 is no noise.
49416
+ * @description The amount of noise to add to noise image for image. Only used if the image_url is provided. 1.0 is complete noise and values near 0 are minimal noise.
49063
49417
  * @default 0.5
49064
49418
  */
49065
49419
  noise_strength?: number;
@@ -52727,10 +53081,10 @@ export interface KlingVideoV3StandardTextToVideoOutput {
52727
53081
  /**
52728
53082
  * @description The generated video
52729
53083
  * @example {
52730
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/standard-t2v/out.mp4",
52731
- * "file_name": "output.mp4",
52732
53084
  * "content_type": "video/mp4",
52733
- * "file_size": 6797486
53085
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/standard-t2v/out.mp4",
53086
+ * "file_size": 6797486,
53087
+ * "file_name": "output.mp4"
52734
53088
  * }
52735
53089
  */
52736
53090
  video: Components.File;
@@ -52806,10 +53160,10 @@ export interface KlingVideoV3StandardImageToVideoInput {
52806
53160
  * @description Elements (characters/objects) to include in the video. Each example can either be an image set (frontal + reference images) or a video. Reference in prompt as @Element1, @Element2, etc.
52807
53161
  * @example [
52808
53162
  * {
53163
+ * "frontal_image_url": "https://v3b.fal.media/files/b/0a8cfd5f/-kZL-ha3Iuelku5IHXC-A_glasses.png",
52809
53164
  * "reference_image_urls": [
52810
53165
  * "https://v3b.fal.media/files/b/0a8cfd62/psPCmzrD1y9vDgdyNfKAL_glasses_back.png"
52811
- * ],
52812
- * "frontal_image_url": "https://v3b.fal.media/files/b/0a8cfd5f/-kZL-ha3Iuelku5IHXC-A_glasses.png"
53166
+ * ]
52813
53167
  * },
52814
53168
  * {
52815
53169
  * "video_url": "https://v3b.fal.media/files/b/0a8cfd66/b03SOiQvKLlFx_jqdNZ9z_child_video.mp4"
@@ -52864,10 +53218,10 @@ export interface KlingVideoV3StandardImageToVideoOutput {
52864
53218
  /**
52865
53219
  * @description The generated video
52866
53220
  * @example {
52867
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/standard-i2v/out.mp4",
52868
- * "file_name": "out.mp4",
52869
53221
  * "content_type": "video/mp4",
52870
- * "file_size": 3149129
53222
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/standard-i2v/out.mp4",
53223
+ * "file_size": 3149129,
53224
+ * "file_name": "out.mp4"
52871
53225
  * }
52872
53226
  */
52873
53227
  video: Components.File;
@@ -52931,10 +53285,10 @@ export interface KlingVideoV3ProTextToVideoOutput {
52931
53285
  /**
52932
53286
  * @description The generated video
52933
53287
  * @example {
52934
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/pro-t2v/out.mp4",
52935
- * "file_name": "output.mp4",
52936
53288
  * "content_type": "video/mp4",
52937
- * "file_size": 8062911
53289
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/pro-t2v/out.mp4",
53290
+ * "file_size": 8062911,
53291
+ * "file_name": "output.mp4"
52938
53292
  * }
52939
53293
  */
52940
53294
  video: Components.File;
@@ -53010,10 +53364,10 @@ export interface KlingVideoV3ProImageToVideoInput {
53010
53364
  * @description Elements (characters/objects) to include in the video. Each example can either be an image set (frontal + reference images) or a video. Reference in prompt as @Element1, @Element2, etc.
53011
53365
  * @example [
53012
53366
  * {
53367
+ * "frontal_image_url": "https://v3b.fal.media/files/b/0a8cfd5f/-kZL-ha3Iuelku5IHXC-A_glasses.png",
53013
53368
  * "reference_image_urls": [
53014
53369
  * "https://v3b.fal.media/files/b/0a8cfd62/psPCmzrD1y9vDgdyNfKAL_glasses_back.png"
53015
- * ],
53016
- * "frontal_image_url": "https://v3b.fal.media/files/b/0a8cfd5f/-kZL-ha3Iuelku5IHXC-A_glasses.png"
53370
+ * ]
53017
53371
  * },
53018
53372
  * {
53019
53373
  * "video_url": "https://v3b.fal.media/files/b/0a8cfd66/b03SOiQvKLlFx_jqdNZ9z_child_video.mp4"
@@ -53068,10 +53422,10 @@ export interface KlingVideoV3ProImageToVideoOutput {
53068
53422
  /**
53069
53423
  * @description The generated video
53070
53424
  * @example {
53071
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/pro-i2v/out.mp4",
53072
- * "file_name": "out.mp4",
53073
53425
  * "content_type": "video/mp4",
53074
- * "file_size": 8431922
53426
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-v3/pro-i2v/out.mp4",
53427
+ * "file_size": 8431922,
53428
+ * "file_name": "out.mp4"
53075
53429
  * }
53076
53430
  */
53077
53431
  video: Components.File;
@@ -53166,7 +53520,7 @@ export interface KlingVideoV2MasterImageToVideoOutput {
53166
53520
 
53167
53521
  export interface KlingVideoV26StandardMotionControlInput extends SharedType_778 {}
53168
53522
 
53169
- export interface KlingVideoV26StandardMotionControlOutput extends SharedType_fb8 {}
53523
+ export interface KlingVideoV26StandardMotionControlOutput extends SharedType_514 {}
53170
53524
 
53171
53525
  export interface KlingVideoV26ProTextToVideoInput {
53172
53526
  /**
@@ -53212,10 +53566,10 @@ export interface KlingVideoV26ProTextToVideoOutput {
53212
53566
  /**
53213
53567
  * @description The generated video
53214
53568
  * @example {
53215
- * "url": "https://v3b.fal.media/files/b/0a84ab71/8hPbLs7n59WhWY-BN69yX_output.mp4",
53216
- * "file_name": "output.mp4",
53217
53569
  * "content_type": "video/mp4",
53218
- * "file_size": 8195664
53570
+ * "url": "https://v3b.fal.media/files/b/0a84ab71/8hPbLs7n59WhWY-BN69yX_output.mp4",
53571
+ * "file_size": 8195664,
53572
+ * "file_name": "output.mp4"
53219
53573
  * }
53220
53574
  */
53221
53575
  video: Components.File;
@@ -53223,7 +53577,7 @@ export interface KlingVideoV26ProTextToVideoOutput {
53223
53577
 
53224
53578
  export interface KlingVideoV26ProMotionControlInput extends SharedType_778 {}
53225
53579
 
53226
- export interface KlingVideoV26ProMotionControlOutput extends SharedType_fb8 {}
53580
+ export interface KlingVideoV26ProMotionControlOutput extends SharedType_514 {}
53227
53581
 
53228
53582
  export interface KlingVideoV26ProImageToVideoInput {
53229
53583
  /**
@@ -53271,10 +53625,10 @@ export interface KlingVideoV26ProImageToVideoOutput {
53271
53625
  /**
53272
53626
  * @description The generated video
53273
53627
  * @example {
53274
- * "url": "https://v3b.fal.media/files/b/0a84ab51/Qr1twf8UgtD5rZHpNXC2P_output.mp4",
53275
- * "file_name": "output.mp4",
53276
53628
  * "content_type": "video/mp4",
53277
- * "file_size": 11814817
53629
+ * "url": "https://v3b.fal.media/files/b/0a84ab51/Qr1twf8UgtD5rZHpNXC2P_output.mp4",
53630
+ * "file_size": 11814817,
53631
+ * "file_name": "output.mp4"
53278
53632
  * }
53279
53633
  */
53280
53634
  video: Components.File;
@@ -53767,7 +54121,7 @@ export interface KlingVideoV1StandardImageToVideoOutput extends SharedType_94e {
53767
54121
 
53768
54122
  export interface KlingVideoV1StandardEffectsInput extends SharedType_9db {}
53769
54123
 
53770
- export interface KlingVideoV1StandardEffectsOutput extends SharedType_dea {}
54124
+ export interface KlingVideoV1StandardEffectsOutput extends SharedType_3db {}
53771
54125
 
53772
54126
  export interface KlingVideoV1StandardAiAvatarInput extends SharedType_c85 {}
53773
54127
 
@@ -53817,11 +54171,11 @@ export interface KlingVideoV16StandardImageToVideoOutput extends SharedType_94e
53817
54171
 
53818
54172
  export interface KlingVideoV16StandardElementsInput extends SharedType_6bb {}
53819
54173
 
53820
- export interface KlingVideoV16StandardElementsOutput extends SharedType_70f {}
54174
+ export interface KlingVideoV16StandardElementsOutput extends SharedType_8ac {}
53821
54175
 
53822
54176
  export interface KlingVideoV16StandardEffectsInput extends SharedType_9db {}
53823
54177
 
53824
- export interface KlingVideoV16StandardEffectsOutput extends SharedType_dea {}
54178
+ export interface KlingVideoV16StandardEffectsOutput extends SharedType_3db {}
53825
54179
 
53826
54180
  export interface KlingVideoV16ProTextToVideoInput extends SharedType_a8a {}
53827
54181
 
@@ -53875,11 +54229,11 @@ export interface KlingVideoV16ProImageToVideoOutput extends SharedType_94e {}
53875
54229
 
53876
54230
  export interface KlingVideoV16ProElementsInput extends SharedType_6bb {}
53877
54231
 
53878
- export interface KlingVideoV16ProElementsOutput extends SharedType_70f {}
54232
+ export interface KlingVideoV16ProElementsOutput extends SharedType_8ac {}
53879
54233
 
53880
54234
  export interface KlingVideoV16ProEffectsInput extends SharedType_9db {}
53881
54235
 
53882
- export interface KlingVideoV16ProEffectsOutput extends SharedType_dea {}
54236
+ export interface KlingVideoV16ProEffectsOutput extends SharedType_3db {}
53883
54237
 
53884
54238
  export interface KlingVideoV15ProTextToVideoInput extends SharedType_a8a {}
53885
54239
 
@@ -53944,7 +54298,7 @@ export interface KlingVideoV15ProImageToVideoOutput extends SharedType_94e {}
53944
54298
 
53945
54299
  export interface KlingVideoV15ProEffectsInput extends SharedType_9db {}
53946
54300
 
53947
- export interface KlingVideoV15ProEffectsOutput extends SharedType_dea {}
54301
+ export interface KlingVideoV15ProEffectsOutput extends SharedType_3db {}
53948
54302
 
53949
54303
  export interface KlingVideoO3StandardVideoToVideoReferenceInput {
53950
54304
  /**
@@ -53966,16 +54320,16 @@ export interface KlingVideoO3StandardVideoToVideoReferenceInput {
53966
54320
  * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2.
53967
54321
  * @example [
53968
54322
  * {
54323
+ * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-reference/element1_front.png",
53969
54324
  * "reference_image_urls": [
53970
54325
  * "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-reference/element1_reference1.png"
53971
- * ],
53972
- * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-reference/element1_front.png"
54326
+ * ]
53973
54327
  * },
53974
54328
  * {
54329
+ * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-reference/element2_front.png",
53975
54330
  * "reference_image_urls": [
53976
54331
  * "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-reference/element2_reference1.png"
53977
- * ],
53978
- * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-reference/element2_front.png"
54332
+ * ]
53979
54333
  * }
53980
54334
  * ]
53981
54335
  */
@@ -54018,10 +54372,10 @@ export interface KlingVideoO3StandardVideoToVideoReferenceOutput {
54018
54372
  /**
54019
54373
  * @description The generated video.
54020
54374
  * @example {
54021
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/standard-v2v-reference/output.mp4",
54022
- * "file_name": "output.mp4",
54023
54375
  * "content_type": "video/mp4",
54024
- * "file_size": 7992288
54376
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/standard-v2v-reference/output.mp4",
54377
+ * "file_size": 7992288,
54378
+ * "file_name": "output.mp4"
54025
54379
  * }
54026
54380
  */
54027
54381
  video: Components.File;
@@ -54033,10 +54387,10 @@ export interface KlingVideoO3StandardVideoToVideoEditInput {
54033
54387
  * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2.
54034
54388
  * @example [
54035
54389
  * {
54390
+ * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-edit/element1_front.jpg",
54036
54391
  * "reference_image_urls": [
54037
54392
  * "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-edit/standard_element1_reference1.jpg"
54038
- * ],
54039
- * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/standard-v2v-edit/element1_front.jpg"
54393
+ * ]
54040
54394
  * }
54041
54395
  * ]
54042
54396
  */
@@ -54078,10 +54432,10 @@ export interface KlingVideoO3StandardVideoToVideoEditOutput {
54078
54432
  /**
54079
54433
  * @description The generated video.
54080
54434
  * @example {
54081
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/standard-v2v-edit/output.mp4",
54082
- * "file_name": "output.mp4",
54083
54435
  * "content_type": "video/mp4",
54084
- * "file_size": 4322769
54436
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/standard-v2v-edit/output.mp4",
54437
+ * "file_size": 4322769,
54438
+ * "file_name": "output.mp4"
54085
54439
  * }
54086
54440
  */
54087
54441
  video: Components.File;
@@ -54089,7 +54443,7 @@ export interface KlingVideoO3StandardVideoToVideoEditOutput {
54089
54443
 
54090
54444
  export interface KlingVideoO3StandardTextToVideoInput extends SharedType_25d {}
54091
54445
 
54092
- export interface KlingVideoO3StandardTextToVideoOutput extends SharedType_a10 {}
54446
+ export interface KlingVideoO3StandardTextToVideoOutput extends SharedType_f37 {}
54093
54447
 
54094
54448
  export interface KlingVideoO3StandardReferenceToVideoInput {
54095
54449
  /**
@@ -54112,16 +54466,16 @@ export interface KlingVideoO3StandardReferenceToVideoInput {
54112
54466
  * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2.
54113
54467
  * @example [
54114
54468
  * {
54469
+ * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b64/-ddhNV-utVpHj_1uGfuY-_i6viYke3.png",
54115
54470
  * "reference_image_urls": [
54116
54471
  * "https://v3b.fal.media/files/b/0a8d1b5e/Gl1qUHJeTG63vAGtQmGM-_S3kPW32v.png"
54117
- * ],
54118
- * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b64/-ddhNV-utVpHj_1uGfuY-_i6viYke3.png"
54472
+ * ]
54119
54473
  * },
54120
54474
  * {
54475
+ * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b66/cWe8LC84I_OF6ee9ZtnsO_nY9Hw5UB.png",
54121
54476
  * "reference_image_urls": [
54122
54477
  * "https://v3b.fal.media/files/b/0a8d1b60/Dt21s8LElZdSccIGDC7ec_Nuaraa9P.png"
54123
- * ],
54124
- * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b66/cWe8LC84I_OF6ee9ZtnsO_nY9Hw5UB.png"
54478
+ * ]
54125
54479
  * }
54126
54480
  * ]
54127
54481
  */
@@ -54176,10 +54530,10 @@ export interface KlingVideoO3StandardReferenceToVideoOutput {
54176
54530
  /**
54177
54531
  * @description The generated video.
54178
54532
  * @example {
54179
- * "url": "https://v3b.fal.media/files/b/0a8d200d/ejCxI5DalzOPlP4yf6uP3_output.mp4",
54180
- * "file_name": "output.mp4",
54181
54533
  * "content_type": "video/mp4",
54182
- * "file_size": 3192162
54534
+ * "url": "https://v3b.fal.media/files/b/0a8d200d/ejCxI5DalzOPlP4yf6uP3_output.mp4",
54535
+ * "file_size": 3192162,
54536
+ * "file_name": "output.mp4"
54183
54537
  * }
54184
54538
  */
54185
54539
  video: Components.File;
@@ -54187,7 +54541,7 @@ export interface KlingVideoO3StandardReferenceToVideoOutput {
54187
54541
 
54188
54542
  export interface KlingVideoO3StandardImageToVideoInput extends SharedType_816 {}
54189
54543
 
54190
- export interface KlingVideoO3StandardImageToVideoOutput extends SharedType_198 {}
54544
+ export interface KlingVideoO3StandardImageToVideoOutput extends SharedType_7ad {}
54191
54545
 
54192
54546
  export interface KlingVideoO3ProVideoToVideoReferenceInput {
54193
54547
  /**
@@ -54209,10 +54563,10 @@ export interface KlingVideoO3ProVideoToVideoReferenceInput {
54209
54563
  * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2.
54210
54564
  * @example [
54211
54565
  * {
54566
+ * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/pro-v2v-reference/element1_front.png",
54212
54567
  * "reference_image_urls": [
54213
54568
  * "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/pro-v2v-reference/element1_reference1.png"
54214
- * ],
54215
- * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/pro-v2v-reference/element1_front.png"
54569
+ * ]
54216
54570
  * }
54217
54571
  * ]
54218
54572
  */
@@ -54258,10 +54612,10 @@ export interface KlingVideoO3ProVideoToVideoReferenceOutput {
54258
54612
  /**
54259
54613
  * @description The generated video.
54260
54614
  * @example {
54261
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/pro-v2v-reference/output.mp4",
54262
- * "file_name": "output.mp4",
54263
54615
  * "content_type": "video/mp4",
54264
- * "file_size": 15809901
54616
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/pro-v2v-reference/output.mp4",
54617
+ * "file_size": 15809901,
54618
+ * "file_name": "output.mp4"
54265
54619
  * }
54266
54620
  */
54267
54621
  video: Components.File;
@@ -54273,10 +54627,10 @@ export interface KlingVideoO3ProVideoToVideoEditInput {
54273
54627
  * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2.
54274
54628
  * @example [
54275
54629
  * {
54630
+ * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/pro-v2v-edit/element1_front.png",
54276
54631
  * "reference_image_urls": [
54277
54632
  * "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/pro-v2v-edit/element1_reference1.png"
54278
- * ],
54279
- * "frontal_image_url": "https://storage.googleapis.com/falserverless/example_inputs/kling-o3/pro-v2v-edit/element1_front.png"
54633
+ * ]
54280
54634
  * }
54281
54635
  * ]
54282
54636
  */
@@ -54320,10 +54674,10 @@ export interface KlingVideoO3ProVideoToVideoEditOutput {
54320
54674
  /**
54321
54675
  * @description The generated video.
54322
54676
  * @example {
54323
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/pro-v2v-edit/output.mp4",
54324
- * "file_name": "output.mp4",
54325
54677
  * "content_type": "video/mp4",
54326
- * "file_size": 4322769
54678
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/kling-o3/pro-v2v-edit/output.mp4",
54679
+ * "file_size": 4322769,
54680
+ * "file_name": "output.mp4"
54327
54681
  * }
54328
54682
  */
54329
54683
  video: Components.File;
@@ -54331,7 +54685,7 @@ export interface KlingVideoO3ProVideoToVideoEditOutput {
54331
54685
 
54332
54686
  export interface KlingVideoO3ProTextToVideoInput extends SharedType_25d {}
54333
54687
 
54334
- export interface KlingVideoO3ProTextToVideoOutput extends SharedType_a10 {}
54688
+ export interface KlingVideoO3ProTextToVideoOutput extends SharedType_f37 {}
54335
54689
 
54336
54690
  export interface KlingVideoO3ProReferenceToVideoInput {
54337
54691
  /**
@@ -54354,16 +54708,16 @@ export interface KlingVideoO3ProReferenceToVideoInput {
54354
54708
  * @description Elements (characters/objects) to include. Reference in prompt as @Element1, @Element2.
54355
54709
  * @example [
54356
54710
  * {
54711
+ * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b2e/yiHiZP1Now0V5JC5_OClE_PaKOtOGJ.png",
54357
54712
  * "reference_image_urls": [
54358
54713
  * "https://v3b.fal.media/files/b/0a8d1b1a/eZfSbcQ58EzD_l2SEbevg_F3U9GMLK.png"
54359
- * ],
54360
- * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b2e/yiHiZP1Now0V5JC5_OClE_PaKOtOGJ.png"
54714
+ * ]
54361
54715
  * },
54362
54716
  * {
54717
+ * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b19/_eIj7GjmI5zgQkMN936YJ_2f3hZ7Xb.png",
54363
54718
  * "reference_image_urls": [
54364
54719
  * "https://v3b.fal.media/files/b/0a8d1b3c/_ZE2iIjkb-Eun3WXXGP4x_TSG1ELBo.png"
54365
- * ],
54366
- * "frontal_image_url": "https://v3b.fal.media/files/b/0a8d1b19/_eIj7GjmI5zgQkMN936YJ_2f3hZ7Xb.png"
54720
+ * ]
54367
54721
  * }
54368
54722
  * ]
54369
54723
  */
@@ -54415,10 +54769,10 @@ export interface KlingVideoO3ProReferenceToVideoOutput {
54415
54769
  /**
54416
54770
  * @description The generated video.
54417
54771
  * @example {
54418
- * "url": "https://v3b.fal.media/files/b/0a8d1c8a/ZxdKrvPb3CQEmeuS-u_kU_output.mp4",
54419
- * "file_name": "output.mp4",
54420
54772
  * "content_type": "video/mp4",
54421
- * "file_size": 18468404
54773
+ * "url": "https://v3b.fal.media/files/b/0a8d1c8a/ZxdKrvPb3CQEmeuS-u_kU_output.mp4",
54774
+ * "file_size": 18468404,
54775
+ * "file_name": "output.mp4"
54422
54776
  * }
54423
54777
  */
54424
54778
  video: Components.File;
@@ -54426,39 +54780,39 @@ export interface KlingVideoO3ProReferenceToVideoOutput {
54426
54780
 
54427
54781
  export interface KlingVideoO3ProImageToVideoInput extends SharedType_816 {}
54428
54782
 
54429
- export interface KlingVideoO3ProImageToVideoOutput extends SharedType_198 {}
54783
+ export interface KlingVideoO3ProImageToVideoOutput extends SharedType_7ad {}
54430
54784
 
54431
- export interface KlingVideoO1VideoToVideoReferenceInput extends SharedType_cc5 {}
54785
+ export interface KlingVideoO1VideoToVideoReferenceInput extends SharedType_8aa {}
54432
54786
 
54433
- export interface KlingVideoO1VideoToVideoReferenceOutput extends SharedType_e07 {}
54787
+ export interface KlingVideoO1VideoToVideoReferenceOutput extends SharedType_ccf {}
54434
54788
 
54435
- export interface KlingVideoO1VideoToVideoEditInput extends SharedType_bda {}
54789
+ export interface KlingVideoO1VideoToVideoEditInput extends SharedType_5d7 {}
54436
54790
 
54437
- export interface KlingVideoO1VideoToVideoEditOutput extends SharedType_7c9 {}
54791
+ export interface KlingVideoO1VideoToVideoEditOutput extends SharedType_de0 {}
54438
54792
 
54439
- export interface KlingVideoO1StandardVideoToVideoReferenceInput extends SharedType_cc5 {}
54793
+ export interface KlingVideoO1StandardVideoToVideoReferenceInput extends SharedType_8aa {}
54440
54794
 
54441
- export interface KlingVideoO1StandardVideoToVideoReferenceOutput extends SharedType_e07 {}
54795
+ export interface KlingVideoO1StandardVideoToVideoReferenceOutput extends SharedType_ccf {}
54442
54796
 
54443
- export interface KlingVideoO1StandardVideoToVideoEditInput extends SharedType_bda {}
54797
+ export interface KlingVideoO1StandardVideoToVideoEditInput extends SharedType_5d7 {}
54444
54798
 
54445
- export interface KlingVideoO1StandardVideoToVideoEditOutput extends SharedType_7c9 {}
54799
+ export interface KlingVideoO1StandardVideoToVideoEditOutput extends SharedType_de0 {}
54446
54800
 
54447
- export interface KlingVideoO1StandardReferenceToVideoInput extends SharedType_ca4 {}
54801
+ export interface KlingVideoO1StandardReferenceToVideoInput extends SharedType_4fd {}
54448
54802
 
54449
- export interface KlingVideoO1StandardReferenceToVideoOutput extends SharedType_089 {}
54803
+ export interface KlingVideoO1StandardReferenceToVideoOutput extends SharedType_a63 {}
54450
54804
 
54451
54805
  export interface KlingVideoO1StandardImageToVideoInput extends SharedType_732 {}
54452
54806
 
54453
- export interface KlingVideoO1StandardImageToVideoOutput extends SharedType_37b {}
54807
+ export interface KlingVideoO1StandardImageToVideoOutput extends SharedType_11e {}
54454
54808
 
54455
- export interface KlingVideoO1ReferenceToVideoInput extends SharedType_ca4 {}
54809
+ export interface KlingVideoO1ReferenceToVideoInput extends SharedType_4fd {}
54456
54810
 
54457
- export interface KlingVideoO1ReferenceToVideoOutput extends SharedType_089 {}
54811
+ export interface KlingVideoO1ReferenceToVideoOutput extends SharedType_a63 {}
54458
54812
 
54459
54813
  export interface KlingVideoO1ImageToVideoInput extends SharedType_732 {}
54460
54814
 
54461
- export interface KlingVideoO1ImageToVideoOutput extends SharedType_37b {}
54815
+ export interface KlingVideoO1ImageToVideoOutput extends SharedType_11e {}
54462
54816
 
54463
54817
  export interface KlingVideoLipsyncTextToVideoInput {
54464
54818
  /**
@@ -55230,10 +55584,10 @@ export interface JoyaiImageEditOutput {
55230
55584
  * @description The edited output image.
55231
55585
  * @example [
55232
55586
  * {
55233
- * "width": 1024,
55587
+ * "url": "https://v3b.fal.media/files/b/0a951738/yWOL7Oie3FQfzM9_HD5BJ.jpg",
55234
55588
  * "content_type": "image/jpeg",
55235
55589
  * "height": 1024,
55236
- * "url": "https://v3b.fal.media/files/b/0a951738/yWOL7Oie3FQfzM9_HD5BJ.jpg"
55590
+ * "width": 1024
55237
55591
  * }
55238
55592
  * ]
55239
55593
  */
@@ -61642,16 +61996,15 @@ export interface HunyuanPortraitInput {
61642
61996
 
61643
61997
  export interface HunyuanPortraitOutput {
61644
61998
  /**
61645
- * Video
61646
61999
  * @description The generated video with the portrait animation.
61647
62000
  * @example {
61648
62001
  * "file_size": 5485412,
61649
62002
  * "file_name": "output_with_audio.mp4",
61650
- * "content_type": "video/mp4",
61651
- * "url": "https://v3.fal.media/files/tiger/9H3vkuqNoYcKr6OBGj3Mr_3b01c7a4-802c-4697-b8fb-ad86bd0eba41_output_with_audio.mp4"
62003
+ * "url": "https://v3.fal.media/files/tiger/9H3vkuqNoYcKr6OBGj3Mr_3b01c7a4-802c-4697-b8fb-ad86bd0eba41_output_with_audio.mp4",
62004
+ * "content_type": "video/mp4"
61652
62005
  * }
61653
62006
  */
61654
- video: Components.File_1;
62007
+ video: Components.File;
61655
62008
  }
61656
62009
 
61657
62010
  export interface HunyuanPartInput {
@@ -61900,9 +62253,9 @@ export interface HunyuanImageV3InstructTextToImageOutput {
61900
62253
  * @description A list of the generated images.
61901
62254
  * @example [
61902
62255
  * {
61903
- * "height": 1024,
61904
- * "width": 1024,
61905
62256
  * "url": "https://v3b.fal.media/files/b/0a8c37ef/niSZAZpVwwixexdtx1AAf.png",
62257
+ * "width": 1024,
62258
+ * "height": 1024,
61906
62259
  * "content_type": "image/png"
61907
62260
  * }
61908
62261
  * ]
@@ -61990,9 +62343,9 @@ export interface HunyuanImageV3InstructEditOutput {
61990
62343
  * @description A list of the generated images.
61991
62344
  * @example [
61992
62345
  * {
61993
- * "height": 1024,
61994
- * "width": 1024,
61995
62346
  * "url": "https://v3b.fal.media/files/b/0a8c3d5f/K7ecUzkFaxU-5hvzdhvG-_SOvArpR4.png",
62347
+ * "width": 1024,
62348
+ * "height": 1024,
61996
62349
  * "content_type": "image/png"
61997
62350
  * }
61998
62351
  * ]
@@ -64364,7 +64717,7 @@ export interface GenfocusAllInFocusInput {
64364
64717
  target_long_side?: number;
64365
64718
  }
64366
64719
 
64367
- export interface GenfocusAllInFocusOutput extends SharedType_836 {}
64720
+ export interface GenfocusAllInFocusOutput extends SharedType_663 {}
64368
64721
 
64369
64722
  export interface GenfocusInput {
64370
64723
  /**
@@ -64426,7 +64779,7 @@ export interface GenfocusInput {
64426
64779
  target_long_side?: number;
64427
64780
  }
64428
64781
 
64429
- export interface GenfocusOutput extends SharedType_836 {}
64782
+ export interface GenfocusOutput extends SharedType_663 {}
64430
64783
 
64431
64784
  export interface GeminiTtsInput {
64432
64785
  /**
@@ -64720,8 +65073,8 @@ export interface Gemini3ProImagePreviewEditOutput {
64720
65073
  * @example [
64721
65074
  * {
64722
65075
  * "file_name": "nano-banana-pro-edit-output.png",
64723
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-pro-edit-output.png",
64724
- * "content_type": "image/png"
65076
+ * "content_type": "image/png",
65077
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-pro-edit-output.png"
64725
65078
  * }
64726
65079
  * ]
64727
65080
  */
@@ -64742,8 +65095,8 @@ export interface Gemini3ProImagePreviewOutput {
64742
65095
  * @example [
64743
65096
  * {
64744
65097
  * "file_name": "nano-banana-pro-t2i-output.png",
64745
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-pro-t2i-output.png",
64746
- * "content_type": "image/png"
65098
+ * "content_type": "image/png",
65099
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-pro-t2i-output.png"
64747
65100
  * }
64748
65101
  * ]
64749
65102
  */
@@ -78612,221 +78965,221 @@ export interface ElevenlabsSpeechToTextScribeV2Output {
78612
78965
  * Words
78613
78966
  * @description Word-level transcription details
78614
78967
  * @example {
78968
+ * "text": "Hey,",
78615
78969
  * "start": 0.079,
78970
+ * "end": 0.539,
78616
78971
  * "speaker_id": "speaker_0",
78617
- * "text": "Hey,",
78618
- * "type": "word",
78619
- * "end": 0.539
78972
+ * "type": "word"
78620
78973
  * }
78621
78974
  * @example {
78975
+ * "text": " ",
78622
78976
  * "start": 0.539,
78977
+ * "end": 0.599,
78623
78978
  * "speaker_id": "speaker_0",
78624
- * "text": " ",
78625
- * "type": "spacing",
78626
- * "end": 0.599
78979
+ * "type": "spacing"
78627
78980
  * }
78628
78981
  * @example {
78982
+ * "text": "this",
78629
78983
  * "start": 0.599,
78984
+ * "end": 0.679,
78630
78985
  * "speaker_id": "speaker_0",
78631
- * "text": "this",
78632
- * "type": "word",
78633
- * "end": 0.679
78986
+ * "type": "word"
78634
78987
  * }
78635
78988
  * @example {
78989
+ * "text": " ",
78636
78990
  * "start": 0.679,
78991
+ * "end": 0.739,
78637
78992
  * "speaker_id": "speaker_0",
78638
- * "text": " ",
78639
- * "type": "spacing",
78640
- * "end": 0.739
78993
+ * "type": "spacing"
78641
78994
  * }
78642
78995
  * @example {
78996
+ * "text": "is",
78643
78997
  * "start": 0.739,
78998
+ * "end": 0.799,
78644
78999
  * "speaker_id": "speaker_0",
78645
- * "text": "is",
78646
- * "type": "word",
78647
- * "end": 0.799
79000
+ * "type": "word"
78648
79001
  * }
78649
79002
  * @example {
79003
+ * "text": " ",
78650
79004
  * "start": 0.799,
79005
+ * "end": 0.939,
78651
79006
  * "speaker_id": "speaker_0",
78652
- * "text": " ",
78653
- * "type": "spacing",
78654
- * "end": 0.939
79007
+ * "type": "spacing"
78655
79008
  * }
78656
79009
  * @example {
79010
+ * "text": "a",
78657
79011
  * "start": 0.939,
79012
+ * "end": 0.939,
78658
79013
  * "speaker_id": "speaker_0",
78659
- * "text": "a",
78660
- * "type": "word",
78661
- * "end": 0.939
79014
+ * "type": "word"
78662
79015
  * }
78663
79016
  * @example {
79017
+ * "text": " ",
78664
79018
  * "start": 0.939,
79019
+ * "end": 0.959,
78665
79020
  * "speaker_id": "speaker_0",
78666
- * "text": " ",
78667
- * "type": "spacing",
78668
- * "end": 0.959
79021
+ * "type": "spacing"
78669
79022
  * }
78670
79023
  * @example {
79024
+ * "text": "test",
78671
79025
  * "start": 0.959,
79026
+ * "end": 1.179,
78672
79027
  * "speaker_id": "speaker_0",
78673
- * "text": "test",
78674
- * "type": "word",
78675
- * "end": 1.179
79028
+ * "type": "word"
78676
79029
  * }
78677
79030
  * @example {
79031
+ * "text": " ",
78678
79032
  * "start": 1.179,
79033
+ * "end": 1.219,
78679
79034
  * "speaker_id": "speaker_0",
78680
- * "text": " ",
78681
- * "type": "spacing",
78682
- * "end": 1.219
79035
+ * "type": "spacing"
78683
79036
  * }
78684
79037
  * @example {
79038
+ * "text": "recording",
78685
79039
  * "start": 1.22,
79040
+ * "end": 1.719,
78686
79041
  * "speaker_id": "speaker_0",
78687
- * "text": "recording",
78688
- * "type": "word",
78689
- * "end": 1.719
79042
+ * "type": "word"
78690
79043
  * }
78691
79044
  * @example {
79045
+ * "text": " ",
78692
79046
  * "start": 1.719,
79047
+ * "end": 1.719,
78693
79048
  * "speaker_id": "speaker_0",
78694
- * "text": " ",
78695
- * "type": "spacing",
78696
- * "end": 1.719
79049
+ * "type": "spacing"
78697
79050
  * }
78698
79051
  * @example {
79052
+ * "text": "for",
78699
79053
  * "start": 1.719,
79054
+ * "end": 1.86,
78700
79055
  * "speaker_id": "speaker_0",
78701
- * "text": "for",
78702
- * "type": "word",
78703
- * "end": 1.86
79056
+ * "type": "word"
78704
79057
  * }
78705
79058
  * @example {
79059
+ * "text": " ",
78706
79060
  * "start": 1.86,
79061
+ * "end": 1.879,
78707
79062
  * "speaker_id": "speaker_0",
78708
- * "text": " ",
78709
- * "type": "spacing",
78710
- * "end": 1.879
79063
+ * "type": "spacing"
78711
79064
  * }
78712
79065
  * @example {
79066
+ * "text": "Scribe",
78713
79067
  * "start": 1.879,
79068
+ * "end": 2.24,
78714
79069
  * "speaker_id": "speaker_0",
78715
- * "text": "Scribe",
78716
- * "type": "word",
78717
- * "end": 2.24
79070
+ * "type": "word"
78718
79071
  * }
78719
79072
  * @example {
79073
+ * "text": " ",
78720
79074
  * "start": 2.24,
79075
+ * "end": 2.319,
78721
79076
  * "speaker_id": "speaker_0",
78722
- * "text": " ",
78723
- * "type": "spacing",
78724
- * "end": 2.319
79077
+ * "type": "spacing"
78725
79078
  * }
78726
79079
  * @example {
79080
+ * "text": "version",
78727
79081
  * "start": 2.319,
79082
+ * "end": 2.759,
78728
79083
  * "speaker_id": "speaker_0",
78729
- * "text": "version",
78730
- * "type": "word",
78731
- * "end": 2.759
79084
+ * "type": "word"
78732
79085
  * }
78733
79086
  * @example {
79087
+ * "text": " ",
78734
79088
  * "start": 2.759,
79089
+ * "end": 2.779,
78735
79090
  * "speaker_id": "speaker_0",
78736
- * "text": " ",
78737
- * "type": "spacing",
78738
- * "end": 2.779
79091
+ * "type": "spacing"
78739
79092
  * }
78740
79093
  * @example {
79094
+ * "text": "two,",
78741
79095
  * "start": 2.779,
79096
+ * "end": 3.379,
78742
79097
  * "speaker_id": "speaker_0",
78743
- * "text": "two,",
78744
- * "type": "word",
78745
- * "end": 3.379
79098
+ * "type": "word"
78746
79099
  * }
78747
79100
  * @example {
79101
+ * "text": " ",
78748
79102
  * "start": 3.379,
79103
+ * "end": 3.399,
78749
79104
  * "speaker_id": "speaker_0",
78750
- * "text": " ",
78751
- * "type": "spacing",
78752
- * "end": 3.399
79105
+ * "type": "spacing"
78753
79106
  * }
78754
79107
  * @example {
79108
+ * "text": "which",
78755
79109
  * "start": 3.399,
79110
+ * "end": 3.519,
78756
79111
  * "speaker_id": "speaker_0",
78757
- * "text": "which",
78758
- * "type": "word",
78759
- * "end": 3.519
79112
+ * "type": "word"
78760
79113
  * }
78761
79114
  * @example {
79115
+ * "text": " ",
78762
79116
  * "start": 3.519,
79117
+ * "end": 3.539,
78763
79118
  * "speaker_id": "speaker_0",
78764
- * "text": " ",
78765
- * "type": "spacing",
78766
- * "end": 3.539
79119
+ * "type": "spacing"
78767
79120
  * }
78768
79121
  * @example {
79122
+ * "text": "is",
78769
79123
  * "start": 3.539,
79124
+ * "end": 3.659,
78770
79125
  * "speaker_id": "speaker_0",
78771
- * "text": "is",
78772
- * "type": "word",
78773
- * "end": 3.659
79126
+ * "type": "word"
78774
79127
  * }
78775
79128
  * @example {
79129
+ * "text": " ",
78776
79130
  * "start": 3.659,
79131
+ * "end": 3.699,
78777
79132
  * "speaker_id": "speaker_0",
78778
- * "text": " ",
78779
- * "type": "spacing",
78780
- * "end": 3.699
79133
+ * "type": "spacing"
78781
79134
  * }
78782
79135
  * @example {
79136
+ * "text": "now",
78783
79137
  * "start": 3.699,
79138
+ * "end": 3.839,
78784
79139
  * "speaker_id": "speaker_0",
78785
- * "text": "now",
78786
- * "type": "word",
78787
- * "end": 3.839
79140
+ * "type": "word"
78788
79141
  * }
78789
79142
  * @example {
79143
+ * "text": " ",
78790
79144
  * "start": 3.839,
79145
+ * "end": 3.839,
78791
79146
  * "speaker_id": "speaker_0",
78792
- * "text": " ",
78793
- * "type": "spacing",
78794
- * "end": 3.839
79147
+ * "type": "spacing"
78795
79148
  * }
78796
79149
  * @example {
79150
+ * "text": "available",
78797
79151
  * "start": 3.839,
79152
+ * "end": 4.319,
78798
79153
  * "speaker_id": "speaker_0",
78799
- * "text": "available",
78800
- * "type": "word",
78801
- * "end": 4.319
79154
+ * "type": "word"
78802
79155
  * }
78803
79156
  * @example {
79157
+ * "text": " ",
78804
79158
  * "start": 4.319,
79159
+ * "end": 4.339,
78805
79160
  * "speaker_id": "speaker_0",
78806
- * "text": " ",
78807
- * "type": "spacing",
78808
- * "end": 4.339
79161
+ * "type": "spacing"
78809
79162
  * }
78810
79163
  * @example {
79164
+ * "text": "on",
78811
79165
  * "start": 4.339,
79166
+ * "end": 4.579,
78812
79167
  * "speaker_id": "speaker_0",
78813
- * "text": "on",
78814
- * "type": "word",
78815
- * "end": 4.579
79168
+ * "type": "word"
78816
79169
  * }
78817
79170
  * @example {
79171
+ * "text": " ",
78818
79172
  * "start": 4.579,
79173
+ * "end": 4.599,
78819
79174
  * "speaker_id": "speaker_0",
78820
- * "text": " ",
78821
- * "type": "spacing",
78822
- * "end": 4.599
79175
+ * "type": "spacing"
78823
79176
  * }
78824
79177
  * @example {
79178
+ * "text": "fal.ai.",
78825
79179
  * "start": 4.599,
79180
+ * "end": 5.699,
78826
79181
  * "speaker_id": "speaker_0",
78827
- * "text": "fal.ai.",
78828
- * "type": "word",
78829
- * "end": 5.699
79182
+ * "type": "word"
78830
79183
  * }
78831
79184
  */
78832
79185
  words: Components.TranscriptionWord[];
@@ -79006,9 +79359,9 @@ export interface ElevenlabsMusicOutput {
79006
79359
  /**
79007
79360
  * @description The generated music audio file in MP3 format
79008
79361
  * @example {
79009
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/elevenlabs/music_generated.mp3",
79010
79362
  * "content_type": "audio/mpeg",
79011
- * "file_name": "music_generated.mp3"
79363
+ * "file_name": "music_generated.mp3",
79364
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/elevenlabs/music_generated.mp3"
79012
79365
  * }
79013
79366
  */
79014
79367
  audio: Components.File;
@@ -79068,10 +79421,10 @@ export interface ElevenlabsDubbingOutput {
79068
79421
  /**
79069
79422
  * @description The dubbed video file. Will be populated if video_url was provided in the request.
79070
79423
  * @example {
79071
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/elevenlabs/e11_dubbing_out.mp4",
79424
+ * "file_size": 1344041,
79072
79425
  * "content_type": "video/mp4",
79073
79426
  * "file_name": "e11_dubbing_out.mp4",
79074
- * "file_size": 1344041
79427
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/elevenlabs/e11_dubbing_out.mp4"
79075
79428
  * }
79076
79429
  */
79077
79430
  video: Components.File;
@@ -79270,13 +79623,13 @@ export interface EdittoOutput {
79270
79623
  /**
79271
79624
  * @description The generated image to video file.
79272
79625
  * @example {
79273
- * "width": 832,
79274
- * "duration": 4.86,
79626
+ * "content_type": "video/mp4",
79275
79627
  * "fps": 15,
79276
- * "file_name": "example_out.mp4",
79628
+ * "duration": 4.86,
79629
+ * "width": 832,
79277
79630
  * "url": "https://storage.googleapis.com/falserverless/example_outputs/editto/example_out.mp4",
79278
79631
  * "height": 480,
79279
- * "content_type": "video/mp4",
79632
+ * "file_name": "example_out.mp4",
79280
79633
  * "num_frames": 73
79281
79634
  * }
79282
79635
  */
@@ -80107,12 +80460,12 @@ export interface Deepfilternet3Output {
80107
80460
  * @description The audio file that was enhanced.
80108
80461
  * @example {
80109
80462
  * "sample_rate": 48000,
80110
- * "duration": 6.9544375,
80111
80463
  * "bitrate": "192k",
80112
- * "url": "https://v3b.fal.media/files/b/0a8a4024/-2cD9CyGEjYsyVQ5lEERh_9qwIkJjf.mp3",
80464
+ * "file_name": "-2cD9CyGEjYsyVQ5lEERh_9qwIkJjf.mp3",
80113
80465
  * "channels": 1,
80114
80466
  * "content_type": "audio/mpeg",
80115
- * "file_name": "-2cD9CyGEjYsyVQ5lEERh_9qwIkJjf.mp3"
80467
+ * "url": "https://v3b.fal.media/files/b/0a8a4024/-2cD9CyGEjYsyVQ5lEERh_9qwIkJjf.mp3",
80468
+ * "duration": 6.9544375
80116
80469
  * }
80117
80470
  */
80118
80471
  audio_file: Components.AudioFile;
@@ -80450,9 +80803,9 @@ export interface CreatifyAuroraOutput {
80450
80803
  /**
80451
80804
  * @description The generated video file.
80452
80805
  * @example {
80806
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/creatify/aurora/output.mp4",
80453
80807
  * "content_type": "video/mp4",
80454
- * "file_name": "output.mp4",
80455
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/creatify/aurora/output.mp4"
80808
+ * "file_name": "output.mp4"
80456
80809
  * }
80457
80810
  */
80458
80811
  video: Components.VideoFile;
@@ -80881,10 +81234,10 @@ export interface Cogview4Output {
80881
81234
  * @description The generated images
80882
81235
  * @example [
80883
81236
  * {
80884
- * "url": "https://v3.fal.media/files/tiger/rN6_PpE-o8QlSecqFku6h.jpeg",
80885
81237
  * "width": 1024,
80886
- * "height": 768,
80887
- * "content_type": "image/jpeg"
81238
+ * "content_type": "image/jpeg",
81239
+ * "url": "https://v3.fal.media/files/tiger/rN6_PpE-o8QlSecqFku6h.jpeg",
81240
+ * "height": 768
80888
81241
  * }
80889
81242
  * ]
80890
81243
  */
@@ -81190,11 +81543,11 @@ export interface CodeformerOutput {
81190
81543
  /**
81191
81544
  * @description The generated image file info.
81192
81545
  * @example {
81546
+ * "width": 512,
81547
+ * "height": 512,
81193
81548
  * "url": "https://storage.googleapis.com/falserverless/model_tests/codeformer/codeformer_restored_1.jpeg",
81194
- * "file_size": 423052,
81195
81549
  * "content_type": "image/png",
81196
- * "height": 512,
81197
- * "width": 512,
81550
+ * "file_size": 423052,
81198
81551
  * "file_name": "36d3ca4791a647678b2ff01a35c87f5a.png"
81199
81552
  * }
81200
81553
  */
@@ -81353,7 +81706,7 @@ export interface ChronoEditLoraGalleryUpscalerInput {
81353
81706
  upscale_factor?: number;
81354
81707
  }
81355
81708
 
81356
- export interface ChronoEditLoraGalleryUpscalerOutput extends SharedType_343 {}
81709
+ export interface ChronoEditLoraGalleryUpscalerOutput extends SharedType_6fc {}
81357
81710
 
81358
81711
  export interface ChronoEditLoraGalleryPaintbrushInput {
81359
81712
  /**
@@ -81434,7 +81787,7 @@ export interface ChronoEditLoraGalleryPaintbrushInput {
81434
81787
  turbo_mode?: boolean;
81435
81788
  }
81436
81789
 
81437
- export interface ChronoEditLoraGalleryPaintbrushOutput extends SharedType_343 {}
81790
+ export interface ChronoEditLoraGalleryPaintbrushOutput extends SharedType_6fc {}
81438
81791
 
81439
81792
  export interface ChronoEditLoraInput {
81440
81793
  /**
@@ -81531,11 +81884,11 @@ export interface ChronoEditLoraOutput {
81531
81884
  * @description The edited image.
81532
81885
  * @example [
81533
81886
  * {
81887
+ * "height": 768,
81534
81888
  * "file_name": "2_gRhwfsnmNKYtZ_dveyV.jpg",
81535
81889
  * "width": 1152,
81536
- * "url": "https://v3b.fal.media/files/b/koala/2_gRhwfsnmNKYtZ_dveyV.jpg",
81537
81890
  * "content_type": "image/jpeg",
81538
- * "height": 768
81891
+ * "url": "https://v3b.fal.media/files/b/koala/2_gRhwfsnmNKYtZ_dveyV.jpg"
81539
81892
  * }
81540
81893
  * ]
81541
81894
  */
@@ -81642,11 +81995,11 @@ export interface ChronoEditOutput {
81642
81995
  * @description The edited image.
81643
81996
  * @example [
81644
81997
  * {
81645
- * "url": "https://v3b.fal.media/files/b/koala/2_gRhwfsnmNKYtZ_dveyV.jpg",
81998
+ * "height": 768,
81646
81999
  * "width": 1152,
81647
82000
  * "file_name": "2_gRhwfsnmNKYtZ_dveyV.jpg",
81648
- * "content_type": "image/jpeg",
81649
- * "height": 768
82001
+ * "url": "https://v3b.fal.media/files/b/koala/2_gRhwfsnmNKYtZ_dveyV.jpg",
82002
+ * "content_type": "image/jpeg"
81650
82003
  * }
81651
82004
  * ]
81652
82005
  */
@@ -84292,10 +84645,10 @@ export interface BitdanceOutput {
84292
84645
  * @description The generated image files info.
84293
84646
  * @example [
84294
84647
  * {
84295
- * "url": "https://v3b.fal.media/files/b/0a8f4d70/JxZtUW5HcyjcZAEOSzJEl.jpg",
84296
- * "width": 1024,
84648
+ * "content_type": "image/jpeg",
84297
84649
  * "height": 1024,
84298
- * "content_type": "image/jpeg"
84650
+ * "url": "https://v3b.fal.media/files/b/0a8f4d70/JxZtUW5HcyjcZAEOSzJEl.jpg",
84651
+ * "width": 1024
84299
84652
  * }
84300
84653
  * ]
84301
84654
  */
@@ -84401,14 +84754,14 @@ export interface BirefnetV2VideoOutput {
84401
84754
  /**
84402
84755
  * @description Video with background removed
84403
84756
  * @example {
84757
+ * "height": 1080,
84758
+ * "num_frames": 192,
84759
+ * "file_name": "birefnet-video-output.webm",
84760
+ * "fps": 24,
84761
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/birefnet-video-output.webm",
84404
84762
  * "width": 1920,
84405
84763
  * "content_type": "video/webm",
84406
- * "duration": 8,
84407
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/birefnet-video-output.webm",
84408
- * "fps": 24,
84409
- * "file_name": "birefnet-video-output.webm",
84410
- * "num_frames": 192,
84411
- * "height": 1080
84764
+ * "duration": 8
84412
84765
  * }
84413
84766
  */
84414
84767
  video: Components.VideoFile;
@@ -84493,10 +84846,10 @@ export interface BirefnetV2Output {
84493
84846
  /**
84494
84847
  * @description Image with background removed. When `mask_only` is `True`, this contains the segmentation mask instead.
84495
84848
  * @example {
84496
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/birefnet-output.png",
84497
- * "content_type": "image/png",
84498
84849
  * "height": 1024,
84850
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/birefnet-output.png",
84499
84851
  * "width": 1024,
84852
+ * "content_type": "image/png",
84500
84853
  * "file_name": "birefnet-output.png"
84501
84854
  * }
84502
84855
  */
@@ -84565,10 +84918,10 @@ export interface BirefnetOutput {
84565
84918
  /**
84566
84919
  * @description Image with background removed
84567
84920
  * @example {
84568
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/birefnet-output.png",
84569
- * "content_type": "image/png",
84570
84921
  * "height": 1024,
84922
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/birefnet-output.png",
84571
84923
  * "width": 1024,
84924
+ * "content_type": "image/png",
84572
84925
  * "file_name": "birefnet-output.png"
84573
84926
  * }
84574
84927
  */
@@ -86432,14 +86785,14 @@ export interface ClarityaiCrystalVideoUpscalerOutput {
86432
86785
  /**
86433
86786
  * @description URL to the upscaled video
86434
86787
  * @example {
86435
- * "file_name": "w0VQQvPdwvV2GSCtRTMzh_hDH8SPrB.mp4",
86436
86788
  * "fps": 23.130193905817176,
86437
- * "url": "https://storage.googleapis.com/falserverless/example_outputs/crystal_upscaler/video_upscaling/video_out.mp4",
86438
- * "content_type": "video/mp4",
86439
86789
  * "duration": 13.056527,
86440
- * "width": 4096,
86790
+ * "file_name": "w0VQQvPdwvV2GSCtRTMzh_hDH8SPrB.mp4",
86441
86791
  * "num_frames": 302,
86442
- * "height": 2160
86792
+ * "width": 4096,
86793
+ * "height": 2160,
86794
+ * "content_type": "video/mp4",
86795
+ * "url": "https://storage.googleapis.com/falserverless/example_outputs/crystal_upscaler/video_upscaling/video_out.mp4"
86443
86796
  * }
86444
86797
  */
86445
86798
  video: Components.VideoFile;