fal-endpoint-types 1.3.37 → 1.3.39

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "fal-endpoint-types",
3
- "version": "1.3.37",
3
+ "version": "1.3.39",
4
4
  "description": "TypeScript types for Fal AI endpoints generated from the OpenAPI schemas.",
5
5
  "homepage": "https://github.com/rawpixel-vincent/fal-endpoint-types#readme",
6
6
  "bugs": {
@@ -4264,33 +4264,6 @@ export interface ForceParams {
4264
4264
  y: number;
4265
4265
  }
4266
4266
 
4267
- export interface File_2 {
4268
- /**
4269
- * Content Type
4270
- * @description The mime type of the file.
4271
- * @example image/png
4272
- */
4273
- content_type: string;
4274
- /**
4275
- * File Name
4276
- * @description The name of the file. It will be auto-generated if not provided.
4277
- * @example z9RV14K95DvU.png
4278
- */
4279
- file_name: string;
4280
- /**
4281
- * File Size
4282
- * @description The size of the file in bytes.
4283
- * @example 4404019
4284
- */
4285
- file_size: number;
4286
- /**
4287
- * Url
4288
- * @description The URL where the file can be downloaded from.
4289
- * @example https://url.to/generated/file/z9RV14K95DvU.png
4290
- */
4291
- url: string;
4292
- }
4293
-
4294
4267
  export interface File_1 {
4295
4268
  /**
4296
4269
  * Content Type
@@ -2114,6 +2114,26 @@ declare global {
2114
2114
  output: falEndpoints.PhotomakerOutput;
2115
2115
  };
2116
2116
 
2117
+ 'fal-ai/phota/enhance': {
2118
+ input: falEndpoints.PhotaEnhanceInput;
2119
+ output: falEndpoints.PhotaEnhanceOutput;
2120
+ };
2121
+
2122
+ 'fal-ai/phota/edit': {
2123
+ input: falEndpoints.PhotaEditInput;
2124
+ output: falEndpoints.PhotaEditOutput;
2125
+ };
2126
+
2127
+ 'fal-ai/phota/create-profile': {
2128
+ input: falEndpoints.PhotaCreateProfileInput;
2129
+ output: falEndpoints.PhotaCreateProfileOutput;
2130
+ };
2131
+
2132
+ 'fal-ai/phota': {
2133
+ input: falEndpoints.PhotaInput;
2134
+ output: falEndpoints.PhotaOutput;
2135
+ };
2136
+
2117
2137
  'fal-ai/personaplex/realtime': {
2118
2138
  input: falEndpoints.PersonaplexRealtimeInput;
2119
2139
  output: falEndpoints.PersonaplexRealtimeOutput;
@@ -4059,6 +4079,11 @@ declare global {
4059
4079
  output: { [x in string]: any };
4060
4080
  };
4061
4081
 
4082
+ 'fal-ai/hy-wu-edit': {
4083
+ input: falEndpoints.HyWuEditInput;
4084
+ output: falEndpoints.HyWuEditOutput;
4085
+ };
4086
+
4062
4087
  'fal-ai/hunyuan3d/v2/turbo': {
4063
4088
  input: falEndpoints.Hunyuan3dV2TurboInput;
4064
4089
  output: falEndpoints.Hunyuan3dV2TurboOutput;
@@ -6149,11 +6174,6 @@ declare global {
6149
6174
  output: falEndpoints.BeatovenSoundEffectGenerationOutput;
6150
6175
  };
6151
6176
 
6152
- 'beatoven/music-generation': {
6153
- input: falEndpoints.BeatovenMusicGenerationInput;
6154
- output: falEndpoints.BeatovenMusicGenerationOutput;
6155
- };
6156
-
6157
6177
  'argil/avatars/text-to-video': {
6158
6178
  input: falEndpoints.ArgilAvatarsTextToVideoInput;
6159
6179
  output: falEndpoints.ArgilAvatarsTextToVideoOutput;
@@ -94,8 +94,8 @@ export interface XaiGrokImagineVideoTextToVideoOutput {
94
94
  * "fps": 24,
95
95
  * "width": 1280,
96
96
  * "file_name": "RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
97
- * "content_type": "video/mp4",
98
- * "num_frames": 145
97
+ * "num_frames": 145,
98
+ * "content_type": "video/mp4"
99
99
  * }
100
100
  */
101
101
  video: Components.VideoFile;
@@ -148,8 +148,8 @@ export interface XaiGrokImagineVideoReferenceToVideoOutput {
148
148
  * "fps": 24,
149
149
  * "width": 1280,
150
150
  * "file_name": "r2v_output.mp4",
151
- * "content_type": "video/mp4",
152
- * "num_frames": 192
151
+ * "num_frames": 192,
152
+ * "content_type": "video/mp4"
153
153
  * }
154
154
  */
155
155
  video: Components.VideoFile;
@@ -199,8 +199,8 @@ export interface XaiGrokImagineVideoImageToVideoOutput {
199
199
  * "fps": 24,
200
200
  * "width": 1280,
201
201
  * "file_name": "0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
202
- * "content_type": "video/mp4",
203
- * "num_frames": 145
202
+ * "num_frames": 145,
203
+ * "content_type": "video/mp4"
204
204
  * }
205
205
  */
206
206
  video: Components.VideoFile;
@@ -237,8 +237,8 @@ export interface XaiGrokImagineVideoExtendVideoOutput {
237
237
  * "fps": 24,
238
238
  * "width": 1280,
239
239
  * "file_name": "extended_video.mp4",
240
- * "content_type": "video/mp4",
241
- * "num_frames": 384
240
+ * "num_frames": 384,
241
+ * "content_type": "video/mp4"
242
242
  * }
243
243
  */
244
244
  video: Components.VideoFile;
@@ -276,8 +276,8 @@ export interface XaiGrokImagineVideoEditVideoOutput {
276
276
  * "fps": 24,
277
277
  * "width": 1280,
278
278
  * "file_name": "EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
279
- * "content_type": "video/mp4",
280
- * "num_frames": 121
279
+ * "num_frames": 121,
280
+ * "content_type": "video/mp4"
281
281
  * }
282
282
  */
283
283
  video: Components.VideoFile;
@@ -2768,29 +2768,6 @@ export interface SharedType_e33 {
2768
2768
  };
2769
2769
  }
2770
2770
 
2771
- export interface SharedType_e19 {
2772
- /**
2773
- * Images
2774
- * @description The generated images
2775
- * @example [
2776
- * {
2777
- * "height": 1536,
2778
- * "file_size": 3731290,
2779
- * "file_name": "257cf8e7bd3a47c2959396343d5b38cf.png",
2780
- * "content_type": "image/png",
2781
- * "url": "https://v3.fal.media/files/tiger/48e63e0K6C9XQYBuomoU-_257cf8e7bd3a47c2959396343d5b38cf.png",
2782
- * "width": 1536
2783
- * }
2784
- * ]
2785
- */
2786
- images: Components.Image[];
2787
- /**
2788
- * Seed
2789
- * @description Seed value used for generation.
2790
- */
2791
- seed: number;
2792
- }
2793
-
2794
2771
  export interface SharedType_e18 {
2795
2772
  /**
2796
2773
  * Default Caption
@@ -3726,24 +3703,6 @@ export interface SharedType_cfd {
3726
3703
  sync_mode?: boolean;
3727
3704
  }
3728
3705
 
3729
- export interface SharedType_cf1 {
3730
- /**
3731
- * Images
3732
- * @description The generated images with objects removed.
3733
- * @example [
3734
- * {
3735
- * "file_size": 730703,
3736
- * "height": 768,
3737
- * "file_name": "85a2309b2c954c85a75120e664adbe17.png",
3738
- * "content_type": "image/png",
3739
- * "url": "https://v3.fal.media/files/lion/arYSoJeqWjhbcA8o4budv_85a2309b2c954c85a75120e664adbe17.png",
3740
- * "width": 1024
3741
- * }
3742
- * ]
3743
- */
3744
- images: Components.Image_2[];
3745
- }
3746
-
3747
3706
  export interface SharedType_cde {
3748
3707
  /**
3749
3708
  * Actual Prompt
@@ -5353,6 +5312,29 @@ export interface SharedType_a9b {
5353
5312
  seed?: number;
5354
5313
  }
5355
5314
 
5315
+ export interface SharedType_a97 {
5316
+ /**
5317
+ * Images
5318
+ * @description The generated images
5319
+ * @example [
5320
+ * {
5321
+ * "file_size": 3731290,
5322
+ * "height": 1536,
5323
+ * "file_name": "257cf8e7bd3a47c2959396343d5b38cf.png",
5324
+ * "content_type": "image/png",
5325
+ * "url": "https://v3.fal.media/files/tiger/48e63e0K6C9XQYBuomoU-_257cf8e7bd3a47c2959396343d5b38cf.png",
5326
+ * "width": 1536
5327
+ * }
5328
+ * ]
5329
+ */
5330
+ images: Components.Image[];
5331
+ /**
5332
+ * Seed
5333
+ * @description Seed value used for generation.
5334
+ */
5335
+ seed: number;
5336
+ }
5337
+
5356
5338
  export interface SharedType_a8f {
5357
5339
  /**
5358
5340
  * Image Url
@@ -11536,6 +11518,24 @@ export interface SharedType_38d {
11536
11518
  image_url: string;
11537
11519
  }
11538
11520
 
11521
+ export interface SharedType_386 {
11522
+ /**
11523
+ * Images
11524
+ * @description The generated images with objects removed.
11525
+ * @example [
11526
+ * {
11527
+ * "file_size": 730703,
11528
+ * "height": 768,
11529
+ * "file_name": "85a2309b2c954c85a75120e664adbe17.png",
11530
+ * "content_type": "image/png",
11531
+ * "url": "https://v3.fal.media/files/lion/arYSoJeqWjhbcA8o4budv_85a2309b2c954c85a75120e664adbe17.png",
11532
+ * "width": 1024
11533
+ * }
11534
+ * ]
11535
+ */
11536
+ images: Components.Image[];
11537
+ }
11538
+
11539
11539
  export interface SharedType_384 {
11540
11540
  /**
11541
11541
  * Images
@@ -14796,6 +14796,7 @@ export interface OpenrouterRouterVideoEnterpriseInput {
14796
14796
  * @example google/gemini-2.0-flash-lite-001
14797
14797
  * @example google/gemini-3-flash-preview
14798
14798
  * @example google/gemini-3-pro-preview
14799
+ * @example google/gemini-3.1-pro-preview
14799
14800
  */
14800
14801
  model: string;
14801
14802
  /**
@@ -14848,6 +14849,7 @@ export interface OpenrouterRouterVideoInput {
14848
14849
  * @example google/gemini-2.5-pro
14849
14850
  * @example google/gemini-3-flash-preview
14850
14851
  * @example google/gemini-3-pro-preview
14852
+ * @example google/gemini-3.1-pro-preview
14851
14853
  */
14852
14854
  model: string;
14853
14855
  /**
@@ -14905,6 +14907,7 @@ export interface OpenrouterRouterAudioInput {
14905
14907
  * @example google/gemini-3-flash-preview
14906
14908
  * @example google/gemini-2.5-flash
14907
14909
  * @example google/gemini-3-pro-preview
14910
+ * @example google/gemini-3.1-pro-preview
14908
14911
  */
14909
14912
  model: string;
14910
14913
  /**
@@ -24323,16 +24326,8 @@ export interface TriposrInput {
24323
24326
  }
24324
24327
 
24325
24328
  export interface TriposrOutput {
24326
- /**
24327
- * Model Mesh
24328
- * @description Generated 3D object file.
24329
- */
24330
- model_mesh: Components.File_1;
24331
- /**
24332
- * Remeshing Dir
24333
- * @description Directory containing textures for the remeshed model.
24334
- */
24335
- remeshing_dir?: Components.File_1;
24329
+ model_mesh: Components.File;
24330
+ remeshing_dir?: Components.File;
24336
24331
  /**
24337
24332
  * Timings
24338
24333
  * @description Inference timings.
@@ -27165,7 +27160,7 @@ export interface SeedvrUpscaleImageSeamlessInput {
27165
27160
  /**
27166
27161
  * Image Url
27167
27162
  * @description The input image to be processed
27168
- * @example https://storage.googleapis.com/falserverless/example_inputs/seedvr2/image_in.png
27163
+ * @example https://v3b.fal.media/files/b/0a93eb33/dQ_ZwLyeu8P2UJYsa8RDW_seamless_image_in.png
27169
27164
  */
27170
27165
  image_url: string;
27171
27166
  /**
@@ -28288,15 +28283,15 @@ export interface Sam2VideoInput {
28288
28283
  * @example [
28289
28284
  * {
28290
28285
  * "y": 350,
28291
- * "label": 1,
28286
+ * "x": 210,
28292
28287
  * "frame_index": 0,
28293
- * "x": 210
28288
+ * "label": 1
28294
28289
  * },
28295
28290
  * {
28296
28291
  * "y": 220,
28297
- * "label": 1,
28292
+ * "x": 250,
28298
28293
  * "frame_index": 0,
28299
- * "x": 250
28294
+ * "label": 1
28300
28295
  * }
28301
28296
  * ]
28302
28297
  */
@@ -28355,8 +28350,8 @@ export interface Sam2ImageInput {
28355
28350
  * @example [
28356
28351
  * {
28357
28352
  * "y": 375,
28358
- * "label": 1,
28359
- * "x": 500
28353
+ * "x": 500,
28354
+ * "label": 1
28360
28355
  * }
28361
28356
  * ]
28362
28357
  */
@@ -35245,13 +35240,19 @@ export interface PixverseLipsyncOutput {
35245
35240
  }
35246
35241
 
35247
35242
  export interface PixverseExtendFastInput {
35243
+ /**
35244
+ * Generate Audio Switch
35245
+ * @description Enable audio generation (BGM, SFX, dialogue). Supported in v5.6+ models.
35246
+ * @default false
35247
+ */
35248
+ generate_audio_switch?: boolean;
35248
35249
  /**
35249
35250
  * Model
35250
35251
  * @description The model version to use for generation
35251
35252
  * @default v4.5
35252
35253
  * @enum {string}
35253
35254
  */
35254
- model?: 'v3.5' | 'v4' | 'v4.5' | 'v5' | 'v5.5' | 'v5.6';
35255
+ model?: 'v3.5' | 'v4' | 'v4.5' | 'v5' | 'v5.5' | 'v5.6' | 'v6';
35255
35256
  /**
35256
35257
  * Negative Prompt
35257
35258
  * @description Negative prompt to be used for the generation
@@ -35299,13 +35300,19 @@ export interface PixverseExtendInput {
35299
35300
  * @enum {string}
35300
35301
  */
35301
35302
  duration?: '5' | '8';
35303
+ /**
35304
+ * Generate Audio Switch
35305
+ * @description Enable audio generation (BGM, SFX, dialogue). Supported in v5.6+ models.
35306
+ * @default false
35307
+ */
35308
+ generate_audio_switch?: boolean;
35302
35309
  /**
35303
35310
  * Model
35304
35311
  * @description The model version to use for generation
35305
35312
  * @default v4.5
35306
35313
  * @enum {string}
35307
35314
  */
35308
- model?: 'v3.5' | 'v4' | 'v4.5' | 'v5' | 'v5.5' | 'v5.6';
35315
+ model?: 'v3.5' | 'v4' | 'v4.5' | 'v5' | 'v5.5' | 'v5.6' | 'v6';
35309
35316
  /**
35310
35317
  * Negative Prompt
35311
35318
  * @description Negative prompt to be used for the generation
@@ -36288,6 +36295,191 @@ export interface PhotomakerOutput {
36288
36295
  seed: number;
36289
36296
  }
36290
36297
 
36298
+ export interface PhotaEnhanceInput {
36299
+ /**
36300
+ * Image URL
36301
+ * @description URL/Base64 data URI of the image to enhance.
36302
+ * @example https://v3b.fal.media/files/b/0a8b911d/Abk8vStrvmSPlzUqI_NN3_image_043.png
36303
+ */
36304
+ image_url: string;
36305
+ /**
36306
+ * Number of Images
36307
+ * @description Number of images to generate.
36308
+ * @default 1
36309
+ */
36310
+ num_images?: number;
36311
+ /**
36312
+ * Output Format
36313
+ * @description The format of the generated image.
36314
+ * @default jpeg
36315
+ * @enum {string}
36316
+ */
36317
+ output_format?: 'jpeg' | 'png' | 'webp';
36318
+ /**
36319
+ * Profile Ids
36320
+ * @description List of profile IDs to use for the image enhancement. The profiles sent over will be used as candidates for identity preservation.
36321
+ */
36322
+ profile_ids?: string[];
36323
+ /**
36324
+ * Sync Mode
36325
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
36326
+ * @default false
36327
+ */
36328
+ sync_mode?: boolean;
36329
+ }
36330
+
36331
+ export interface PhotaEnhanceOutput {
36332
+ /**
36333
+ * Images
36334
+ * @description The URL of the enhanced image.
36335
+ * @example [
36336
+ * {
36337
+ * "url": "https://v3b.fal.media/files/b/0a8b911d/XMqiVoO2ECXUZEUYmPl2l.jpg"
36338
+ * }
36339
+ * ]
36340
+ */
36341
+ images: Components.ImageFile[];
36342
+ }
36343
+
36344
+ export interface PhotaEditInput {
36345
+ /**
36346
+ * Aspect Ratio
36347
+ * @description Aspect ratio of the generated image.
36348
+ * @default auto
36349
+ * @enum {string}
36350
+ */
36351
+ aspect_ratio?: 'auto' | '1:1' | '16:9' | '4:3' | '3:4' | '9:16';
36352
+ /**
36353
+ * Image URLs
36354
+ * @description List of URLs/ Base64 data URIs of the images to edit. A maximum of 10 images are supported, additional images will be ignored.
36355
+ * @example [
36356
+ * "https://v3b.fal.media/files/b/0a8b911d/Abk8vStrvmSPlzUqI_NN3_image_043.png"
36357
+ * ]
36358
+ */
36359
+ image_urls?: string[];
36360
+ /**
36361
+ * Number of Images
36362
+ * @description Number of images to generate.
36363
+ * @default 1
36364
+ */
36365
+ num_images?: number;
36366
+ /**
36367
+ * Output Format
36368
+ * @description The format of the generated image.
36369
+ * @default jpeg
36370
+ * @enum {string}
36371
+ */
36372
+ output_format?: 'jpeg' | 'png' | 'webp';
36373
+ /**
36374
+ * Profile Ids
36375
+ * @description List of profile IDs to use for the image generation. Profiles may be tagged in the prompt as @Profile1, @Profile2, etc.
36376
+ */
36377
+ profile_ids?: string[];
36378
+ /**
36379
+ * Prompt
36380
+ * @description Text description of the desired image. To refer to specific profiles, use [[profile_id_1]], [[profile_id_2]], etc.
36381
+ * @example Make this scene more realistic but still keep the game vibes
36382
+ */
36383
+ prompt: string;
36384
+ /**
36385
+ * Resolution
36386
+ * @description Resolution of the generated image.
36387
+ * @default 1K
36388
+ * @enum {string}
36389
+ */
36390
+ resolution?: '1K' | '4K';
36391
+ /**
36392
+ * Sync Mode
36393
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
36394
+ * @default false
36395
+ */
36396
+ sync_mode?: boolean;
36397
+ }
36398
+
36399
+ export interface PhotaEditOutput {
36400
+ /**
36401
+ * Images
36402
+ * @description The URL of the edited image.
36403
+ * @example [
36404
+ * {
36405
+ * "url": "https://v3b.fal.media/files/b/0a8b911d/XMqiVoO2ECXUZEUYmPl2l.jpg"
36406
+ * }
36407
+ * ]
36408
+ */
36409
+ images: Components.ImageFile[];
36410
+ }
36411
+
36412
+ export interface PhotaCreateProfileInput {
36413
+ /**
36414
+ * Image ZIP URL
36415
+ * @description URL to a ZIP archive containing the profile images.
36416
+ */
36417
+ image_data_url: string;
36418
+ }
36419
+
36420
+ export interface PhotaCreateProfileOutput {
36421
+ /**
36422
+ * Profile Id
36423
+ * @description The Photalabs profile ID.
36424
+ */
36425
+ profile_id: string;
36426
+ }
36427
+
36428
+ export interface PhotaInput {
36429
+ /**
36430
+ * Aspect Ratio
36431
+ * @description Aspect ratio of the generated image.
36432
+ * @default auto
36433
+ * @enum {string}
36434
+ */
36435
+ aspect_ratio?: 'auto' | '1:1' | '16:9' | '4:3' | '3:4' | '9:16';
36436
+ /**
36437
+ * Number of Images
36438
+ * @description Number of images to generate.
36439
+ * @default 1
36440
+ */
36441
+ num_images?: number;
36442
+ /**
36443
+ * Output Format
36444
+ * @description The format of the generated image.
36445
+ * @default jpeg
36446
+ * @enum {string}
36447
+ */
36448
+ output_format?: 'jpeg' | 'png' | 'webp';
36449
+ /**
36450
+ * Prompt
36451
+ * @description Text description of the desired image. In case you wish to use specific profiles, refer to them as [[profile_id_1]], [[profile_id_2]], etc.
36452
+ * @example Middle Eastern man in traditional clothing sitting in a cool tent in the desert with a laptop
36453
+ */
36454
+ prompt: string;
36455
+ /**
36456
+ * Resolution
36457
+ * @description Resolution of the generated image.
36458
+ * @default 1K
36459
+ * @enum {string}
36460
+ */
36461
+ resolution?: '1K' | '4K';
36462
+ /**
36463
+ * Sync Mode
36464
+ * @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
36465
+ * @default false
36466
+ */
36467
+ sync_mode?: boolean;
36468
+ }
36469
+
36470
+ export interface PhotaOutput {
36471
+ /**
36472
+ * Images
36473
+ * @description The URL of the generated image.
36474
+ * @example [
36475
+ * {
36476
+ * "url": "https://v3b.fal.media/files/b/0a8b90b7/9avg_nKJmcVinjQHJR_Ja.jpg"
36477
+ * }
36478
+ * ]
36479
+ */
36480
+ images: Components.ImageFile[];
36481
+ }
36482
+
36291
36483
  export interface PersonaplexRealtimeInput {
36292
36484
  /**
36293
36485
  * Audio
@@ -37465,7 +37657,7 @@ export interface ObjectRemovalMaskInput {
37465
37657
  model?: 'low_quality' | 'medium_quality' | 'high_quality' | 'best_quality';
37466
37658
  }
37467
37659
 
37468
- export interface ObjectRemovalMaskOutput extends SharedType_cf1 {}
37660
+ export interface ObjectRemovalMaskOutput extends SharedType_386 {}
37469
37661
 
37470
37662
  export interface ObjectRemovalBboxInput {
37471
37663
  /**
@@ -37502,7 +37694,7 @@ export interface ObjectRemovalBboxInput {
37502
37694
  model?: 'low_quality' | 'medium_quality' | 'high_quality' | 'best_quality';
37503
37695
  }
37504
37696
 
37505
- export interface ObjectRemovalBboxOutput extends SharedType_cf1 {}
37697
+ export interface ObjectRemovalBboxOutput extends SharedType_386 {}
37506
37698
 
37507
37699
  export interface ObjectRemovalInput {
37508
37700
  /**
@@ -37531,7 +37723,7 @@ export interface ObjectRemovalInput {
37531
37723
  prompt: string;
37532
37724
  }
37533
37725
 
37534
- export interface ObjectRemovalOutput extends SharedType_cf1 {}
37726
+ export interface ObjectRemovalOutput extends SharedType_386 {}
37535
37727
 
37536
37728
  export interface NovaSrInput {
37537
37729
  /**
@@ -49186,9 +49378,9 @@ export interface LightxRelightInput {
49186
49378
  /**
49187
49379
  * @description Relighting parameters (required for relight_condition_type='ic'). Not used for 'bg' (which expects a background image URL instead).
49188
49380
  * @example {
49189
- * "relight_prompt": "Sunlight",
49190
- * "bg_source": "Right",
49191
49381
  * "use_sky_mask": false,
49382
+ * "bg_source": "Right",
49383
+ * "relight_prompt": "Sunlight",
49192
49384
  * "cfg": 2
49193
49385
  * }
49194
49386
  */
@@ -52733,6 +52925,8 @@ export interface KlingImageV3ImageToImageInput {
52733
52925
  /**
52734
52926
  * Image Url
52735
52927
  * @description Reference image for image-to-image generation.
52928
+ *
52929
+ * Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
52736
52930
  * @example https://v3b.fal.media/files/b/0a8d06b9/3zxm2qoj2xYWSNwEe5Vd9_a74d767fc42e47a0bf657117fbcf8b90.png
52737
52931
  */
52738
52932
  image_url: string;
@@ -58463,6 +58657,103 @@ export interface Hyper3dRodinOutput {
58463
58657
  textures: Components.Image[];
58464
58658
  }
58465
58659
 
58660
+ export interface HyWuEditInput {
58661
+ /**
58662
+ * Enable Safety Checker
58663
+ * @description If set to true, the safety checker will be enabled.
58664
+ * @default true
58665
+ */
58666
+ enable_safety_checker?: boolean;
58667
+ /**
58668
+ * Enable Thinking
58669
+ * @description Enable thinking mode. The model reasons about the edit before generating, producing higher quality results at the cost of longer inference time. Disable for faster results on straightforward edits.
58670
+ * @default true
58671
+ */
58672
+ enable_thinking?: boolean;
58673
+ /**
58674
+ * Image Size
58675
+ * @description The desired size of the generated image. If auto, image size will be determined by the model.
58676
+ * @default auto
58677
+ */
58678
+ image_size?:
58679
+ | Components.ImageSize
58680
+ | (
58681
+ | 'auto'
58682
+ | 'square_hd'
58683
+ | 'square'
58684
+ | 'portrait_4_3'
58685
+ | 'portrait_16_9'
58686
+ | 'landscape_4_3'
58687
+ | 'landscape_16_9'
58688
+ );
58689
+ /**
58690
+ * Image Urls
58691
+ * @description URLs of input images for editing. Typically 2 images: the base image and the reference image. Supports up to 3 images.
58692
+ * @example [
58693
+ * "https://v3b.fal.media/files/b/0a933dff/BE-FgBximAbCJzZSgDNNw_input_1_1.png",
58694
+ * "https://v3b.fal.media/files/b/0a933dff/fNUqzO_Lxwvr-_-4BLeCV_input_1_2.png"
58695
+ * ]
58696
+ */
58697
+ image_urls: string[];
58698
+ /**
58699
+ * Num Images
58700
+ * @description The number of images to generate.
58701
+ * @default 1
58702
+ */
58703
+ num_images?: number;
58704
+ /**
58705
+ * Num Inference Steps
58706
+ * @description Number of diffusion denoising steps.
58707
+ * @default 30
58708
+ */
58709
+ num_inference_steps?: number;
58710
+ /**
58711
+ * Output Format
58712
+ * @description The format of the generated image.
58713
+ * @default png
58714
+ * @enum {string}
58715
+ */
58716
+ output_format?: 'jpeg' | 'png';
58717
+ /**
58718
+ * Prompt
58719
+ * @description The text prompt describing the desired edit. Supports both English and Chinese. Use specific instructions like 'Replace the clothing on figure 1 with the outfit from figure 2'.
58720
+ * @example Using image 1 as the base image, replace the outfit with the clothing from image 2 while keeping the subject, pose, and background unchanged.
58721
+ */
58722
+ prompt: string;
58723
+ /**
58724
+ * Seed
58725
+ * @description Random seed for reproducible results. If None, a random seed is used.
58726
+ */
58727
+ seed?: number;
58728
+ /**
58729
+ * Sync Mode
58730
+ * @description If True, the media will be returned as a data URI.
58731
+ * @default false
58732
+ */
58733
+ sync_mode?: boolean;
58734
+ }
58735
+
58736
+ export interface HyWuEditOutput {
58737
+ /**
58738
+ * Images
58739
+ * @description A list of the generated/edited images.
58740
+ */
58741
+ images: Components.Image[];
58742
+ /**
58743
+ * Seed
58744
+ * @description The seed used for generation.
58745
+ */
58746
+ seed: number;
58747
+ /**
58748
+ * Timings
58749
+ * @description Performance timing breakdown.
58750
+ * @default {}
58751
+ */
58752
+ timings?: {
58753
+ [key: string]: number;
58754
+ };
58755
+ }
58756
+
58466
58757
  export interface Hunyuan3dV2TurboInput extends SharedType_df1 {}
58467
58758
 
58468
58759
  export interface Hunyuan3dV2TurboOutput extends SharedType_7c0 {}
@@ -64969,15 +65260,15 @@ export interface FluxVisionUpscalerInput {
64969
65260
  export interface FluxVisionUpscalerOutput {
64970
65261
  /**
64971
65262
  * Caption
64972
- * @description The VLM-generated caption describing the upscaled image.
64973
- * @example A highly detailed upscaled photograph featuring sharp edges and enhanced textures. The image shows improved clarity in fine details with natural color preservation and minimal artifacts, demonstrating the AI-enhanced resolution increase.
65263
+ * @description The VLM-generated caption used for upscaling.
65264
+ * @example A highly detailed upscaled photograph featuring sharp edges and enhanced textures. The image shows improved clarity in fine details with natural color preservation and minimal artifacts.
64974
65265
  */
64975
65266
  caption: string;
64976
65267
  /**
64977
65268
  * @description The URL of the generated image.
64978
65269
  * @example {
64979
- * "height": 2048,
64980
65270
  * "file_size": 8842156,
65271
+ * "height": 2048,
64981
65272
  * "file_name": "20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
64982
65273
  * "content_type": "image/png",
64983
65274
  * "url": "https://v3b.fal.media/files/b/panda/20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
@@ -75369,7 +75660,7 @@ export interface FaceToStickerOutput {
75369
75660
  export interface F5TtsInput {
75370
75661
  /**
75371
75662
  * Text to be converted to speech
75372
- * @description The text to be converted to speech.
75663
+ * @description The text to be converted to speech. Maximum 5000 characters.
75373
75664
  * @example I don't really care what you call me. I've been a silent spectator, watching species evolve, empires rise and fall. But always remember, I am mighty and enduring. Respect me and I'll nurture you; ignore me and you shall face the consequences.
75374
75665
  */
75375
75666
  gen_text: string;
@@ -76026,220 +76317,220 @@ export interface ElevenlabsSpeechToTextScribeV2Output {
76026
76317
  * Words
76027
76318
  * @description Word-level transcription details
76028
76319
  * @example {
76029
- * "text": "Hey,",
76320
+ * "end": 0.539,
76030
76321
  * "start": 0.079,
76031
76322
  * "type": "word",
76032
- * "end": 0.539,
76323
+ * "text": "Hey,",
76033
76324
  * "speaker_id": "speaker_0"
76034
76325
  * }
76035
76326
  * @example {
76036
- * "text": " ",
76327
+ * "end": 0.599,
76037
76328
  * "start": 0.539,
76038
76329
  * "type": "spacing",
76039
- * "end": 0.599,
76330
+ * "text": " ",
76040
76331
  * "speaker_id": "speaker_0"
76041
76332
  * }
76042
76333
  * @example {
76043
- * "text": "this",
76334
+ * "end": 0.679,
76044
76335
  * "start": 0.599,
76045
76336
  * "type": "word",
76046
- * "end": 0.679,
76337
+ * "text": "this",
76047
76338
  * "speaker_id": "speaker_0"
76048
76339
  * }
76049
76340
  * @example {
76050
- * "text": " ",
76341
+ * "end": 0.739,
76051
76342
  * "start": 0.679,
76052
76343
  * "type": "spacing",
76053
- * "end": 0.739,
76344
+ * "text": " ",
76054
76345
  * "speaker_id": "speaker_0"
76055
76346
  * }
76056
76347
  * @example {
76057
- * "text": "is",
76348
+ * "end": 0.799,
76058
76349
  * "start": 0.739,
76059
76350
  * "type": "word",
76060
- * "end": 0.799,
76351
+ * "text": "is",
76061
76352
  * "speaker_id": "speaker_0"
76062
76353
  * }
76063
76354
  * @example {
76064
- * "text": " ",
76355
+ * "end": 0.939,
76065
76356
  * "start": 0.799,
76066
76357
  * "type": "spacing",
76067
- * "end": 0.939,
76358
+ * "text": " ",
76068
76359
  * "speaker_id": "speaker_0"
76069
76360
  * }
76070
76361
  * @example {
76071
- * "text": "a",
76362
+ * "end": 0.939,
76072
76363
  * "start": 0.939,
76073
76364
  * "type": "word",
76074
- * "end": 0.939,
76365
+ * "text": "a",
76075
76366
  * "speaker_id": "speaker_0"
76076
76367
  * }
76077
76368
  * @example {
76078
- * "text": " ",
76369
+ * "end": 0.959,
76079
76370
  * "start": 0.939,
76080
76371
  * "type": "spacing",
76081
- * "end": 0.959,
76372
+ * "text": " ",
76082
76373
  * "speaker_id": "speaker_0"
76083
76374
  * }
76084
76375
  * @example {
76085
- * "text": "test",
76376
+ * "end": 1.179,
76086
76377
  * "start": 0.959,
76087
76378
  * "type": "word",
76088
- * "end": 1.179,
76379
+ * "text": "test",
76089
76380
  * "speaker_id": "speaker_0"
76090
76381
  * }
76091
76382
  * @example {
76092
- * "text": " ",
76383
+ * "end": 1.219,
76093
76384
  * "start": 1.179,
76094
76385
  * "type": "spacing",
76095
- * "end": 1.219,
76386
+ * "text": " ",
76096
76387
  * "speaker_id": "speaker_0"
76097
76388
  * }
76098
76389
  * @example {
76099
- * "text": "recording",
76390
+ * "end": 1.719,
76100
76391
  * "start": 1.22,
76101
76392
  * "type": "word",
76102
- * "end": 1.719,
76393
+ * "text": "recording",
76103
76394
  * "speaker_id": "speaker_0"
76104
76395
  * }
76105
76396
  * @example {
76106
- * "text": " ",
76397
+ * "end": 1.719,
76107
76398
  * "start": 1.719,
76108
76399
  * "type": "spacing",
76109
- * "end": 1.719,
76400
+ * "text": " ",
76110
76401
  * "speaker_id": "speaker_0"
76111
76402
  * }
76112
76403
  * @example {
76113
- * "text": "for",
76404
+ * "end": 1.86,
76114
76405
  * "start": 1.719,
76115
76406
  * "type": "word",
76116
- * "end": 1.86,
76407
+ * "text": "for",
76117
76408
  * "speaker_id": "speaker_0"
76118
76409
  * }
76119
76410
  * @example {
76120
- * "text": " ",
76411
+ * "end": 1.879,
76121
76412
  * "start": 1.86,
76122
76413
  * "type": "spacing",
76123
- * "end": 1.879,
76414
+ * "text": " ",
76124
76415
  * "speaker_id": "speaker_0"
76125
76416
  * }
76126
76417
  * @example {
76127
- * "text": "Scribe",
76418
+ * "end": 2.24,
76128
76419
  * "start": 1.879,
76129
76420
  * "type": "word",
76130
- * "end": 2.24,
76421
+ * "text": "Scribe",
76131
76422
  * "speaker_id": "speaker_0"
76132
76423
  * }
76133
76424
  * @example {
76134
- * "text": " ",
76425
+ * "end": 2.319,
76135
76426
  * "start": 2.24,
76136
76427
  * "type": "spacing",
76137
- * "end": 2.319,
76428
+ * "text": " ",
76138
76429
  * "speaker_id": "speaker_0"
76139
76430
  * }
76140
76431
  * @example {
76141
- * "text": "version",
76432
+ * "end": 2.759,
76142
76433
  * "start": 2.319,
76143
76434
  * "type": "word",
76144
- * "end": 2.759,
76435
+ * "text": "version",
76145
76436
  * "speaker_id": "speaker_0"
76146
76437
  * }
76147
76438
  * @example {
76148
- * "text": " ",
76439
+ * "end": 2.779,
76149
76440
  * "start": 2.759,
76150
76441
  * "type": "spacing",
76151
- * "end": 2.779,
76442
+ * "text": " ",
76152
76443
  * "speaker_id": "speaker_0"
76153
76444
  * }
76154
76445
  * @example {
76155
- * "text": "two,",
76446
+ * "end": 3.379,
76156
76447
  * "start": 2.779,
76157
76448
  * "type": "word",
76158
- * "end": 3.379,
76449
+ * "text": "two,",
76159
76450
  * "speaker_id": "speaker_0"
76160
76451
  * }
76161
76452
  * @example {
76162
- * "text": " ",
76453
+ * "end": 3.399,
76163
76454
  * "start": 3.379,
76164
76455
  * "type": "spacing",
76165
- * "end": 3.399,
76456
+ * "text": " ",
76166
76457
  * "speaker_id": "speaker_0"
76167
76458
  * }
76168
76459
  * @example {
76169
- * "text": "which",
76460
+ * "end": 3.519,
76170
76461
  * "start": 3.399,
76171
76462
  * "type": "word",
76172
- * "end": 3.519,
76463
+ * "text": "which",
76173
76464
  * "speaker_id": "speaker_0"
76174
76465
  * }
76175
76466
  * @example {
76176
- * "text": " ",
76467
+ * "end": 3.539,
76177
76468
  * "start": 3.519,
76178
76469
  * "type": "spacing",
76179
- * "end": 3.539,
76470
+ * "text": " ",
76180
76471
  * "speaker_id": "speaker_0"
76181
76472
  * }
76182
76473
  * @example {
76183
- * "text": "is",
76474
+ * "end": 3.659,
76184
76475
  * "start": 3.539,
76185
76476
  * "type": "word",
76186
- * "end": 3.659,
76477
+ * "text": "is",
76187
76478
  * "speaker_id": "speaker_0"
76188
76479
  * }
76189
76480
  * @example {
76190
- * "text": " ",
76481
+ * "end": 3.699,
76191
76482
  * "start": 3.659,
76192
76483
  * "type": "spacing",
76193
- * "end": 3.699,
76484
+ * "text": " ",
76194
76485
  * "speaker_id": "speaker_0"
76195
76486
  * }
76196
76487
  * @example {
76197
- * "text": "now",
76488
+ * "end": 3.839,
76198
76489
  * "start": 3.699,
76199
76490
  * "type": "word",
76200
- * "end": 3.839,
76491
+ * "text": "now",
76201
76492
  * "speaker_id": "speaker_0"
76202
76493
  * }
76203
76494
  * @example {
76204
- * "text": " ",
76495
+ * "end": 3.839,
76205
76496
  * "start": 3.839,
76206
76497
  * "type": "spacing",
76207
- * "end": 3.839,
76498
+ * "text": " ",
76208
76499
  * "speaker_id": "speaker_0"
76209
76500
  * }
76210
76501
  * @example {
76211
- * "text": "available",
76502
+ * "end": 4.319,
76212
76503
  * "start": 3.839,
76213
76504
  * "type": "word",
76214
- * "end": 4.319,
76505
+ * "text": "available",
76215
76506
  * "speaker_id": "speaker_0"
76216
76507
  * }
76217
76508
  * @example {
76218
- * "text": " ",
76509
+ * "end": 4.339,
76219
76510
  * "start": 4.319,
76220
76511
  * "type": "spacing",
76221
- * "end": 4.339,
76512
+ * "text": " ",
76222
76513
  * "speaker_id": "speaker_0"
76223
76514
  * }
76224
76515
  * @example {
76225
- * "text": "on",
76516
+ * "end": 4.579,
76226
76517
  * "start": 4.339,
76227
76518
  * "type": "word",
76228
- * "end": 4.579,
76519
+ * "text": "on",
76229
76520
  * "speaker_id": "speaker_0"
76230
76521
  * }
76231
76522
  * @example {
76232
- * "text": " ",
76523
+ * "end": 4.599,
76233
76524
  * "start": 4.579,
76234
76525
  * "type": "spacing",
76235
- * "end": 4.599,
76526
+ * "text": " ",
76236
76527
  * "speaker_id": "speaker_0"
76237
76528
  * }
76238
76529
  * @example {
76239
- * "text": "fal.ai.",
76530
+ * "end": 5.699,
76240
76531
  * "start": 4.599,
76241
76532
  * "type": "word",
76242
- * "end": 5.699,
76533
+ * "text": "fal.ai.",
76243
76534
  * "speaker_id": "speaker_0"
76244
76535
  * }
76245
76536
  */
@@ -80948,7 +81239,7 @@ export interface BytedanceUpscalerUpscaleVideoOutput {
80948
81239
 
80949
81240
  export interface BriaTextToImageHdInput extends SharedType_411 {}
80950
81241
 
80951
- export interface BriaTextToImageHdOutput extends SharedType_e19 {}
81242
+ export interface BriaTextToImageHdOutput extends SharedType_a97 {}
80952
81243
 
80953
81244
  export interface BriaTextToImageFastInput {
80954
81245
  /**
@@ -81021,11 +81312,11 @@ export interface BriaTextToImageFastInput {
81021
81312
  sync_mode?: boolean;
81022
81313
  }
81023
81314
 
81024
- export interface BriaTextToImageFastOutput extends SharedType_e19 {}
81315
+ export interface BriaTextToImageFastOutput extends SharedType_a97 {}
81025
81316
 
81026
81317
  export interface BriaTextToImageBaseInput extends SharedType_411 {}
81027
81318
 
81028
- export interface BriaTextToImageBaseOutput extends SharedType_e19 {}
81319
+ export interface BriaTextToImageBaseOutput extends SharedType_a97 {}
81029
81320
 
81030
81321
  export interface BriaReimagineInput {
81031
81322
  /**
@@ -81254,8 +81545,8 @@ export interface BriaGenfillOutput {
81254
81545
  * @description Generated Images
81255
81546
  * @example [
81256
81547
  * {
81257
- * "height": 768,
81258
81548
  * "file_size": 1064550,
81549
+ * "height": 768,
81259
81550
  * "file_name": "a0d138e6820c4ad58f1fd3c758f16047.png",
81260
81551
  * "content_type": "image/png",
81261
81552
  * "url": "https://storage.googleapis.com/falserverless/bria/bria_genfill_res.png",
@@ -81339,8 +81630,8 @@ export interface BriaExpandOutput {
81339
81630
  /**
81340
81631
  * @description The generated image
81341
81632
  * @example {
81342
- * "height": 674,
81343
81633
  * "file_size": 1471342,
81634
+ * "height": 674,
81344
81635
  * "file_name": "afa402a35ea742cdb5c3e219b2b19bfb.png",
81345
81636
  * "content_type": "image/png",
81346
81637
  * "url": "https://v3.fal.media/files/koala/8np-spgxxG-I1r3cjthRV_afa402a35ea742cdb5c3e219b2b19bfb.png",
@@ -81496,8 +81787,8 @@ export interface BriaBackgroundRemoveOutput {
81496
81787
  /**
81497
81788
  * @description The generated image
81498
81789
  * @example {
81499
- * "height": 1024,
81500
81790
  * "file_size": 1076276,
81791
+ * "height": 1024,
81501
81792
  * "file_name": "070c731993e949d993c10ef6283d335d.png",
81502
81793
  * "content_type": "image/png",
81503
81794
  * "url": "https://v3.fal.media/files/tiger/GQEMNjRyxSoza7N8LPPqb_070c731993e949d993c10ef6283d335d.png",
@@ -82420,11 +82711,7 @@ export interface AnimatediffSparsectrlLcmOutput {
82420
82711
  * @description The seed used to generate the video.
82421
82712
  */
82422
82713
  seed: number;
82423
- /**
82424
- * Video
82425
- * @description Generated video file.
82426
- */
82427
- video: Components.File_2;
82714
+ video: Components.File;
82428
82715
  }
82429
82716
 
82430
82717
  export interface AmtInterpolationFrameInterpolationInput {
@@ -84959,82 +85246,6 @@ export interface BeatovenSoundEffectGenerationOutput {
84959
85246
  prompt: string;
84960
85247
  }
84961
85248
 
84962
- export interface BeatovenMusicGenerationInput {
84963
- /**
84964
- * Creativity
84965
- * @description Creativity level - higher values allow more creative interpretation of the prompt
84966
- * @default 16
84967
- * @example 16
84968
- * @example 14
84969
- * @example 11
84970
- */
84971
- creativity?: number;
84972
- /**
84973
- * Duration
84974
- * @description Length of the generated music in seconds
84975
- * @default 90
84976
- * @example 90
84977
- * @example 47
84978
- * @example 150
84979
- */
84980
- duration?: number;
84981
- /**
84982
- * Negative Prompt
84983
- * @description Describe what you want to avoid in the music (instruments, styles, moods). Leave blank for none.
84984
- * @default
84985
- * @example noise
84986
- * @example distortion
84987
- * @example heavy drums
84988
- * @example high-hats
84989
- */
84990
- negative_prompt?: string;
84991
- /**
84992
- * Prompt
84993
- * @description Describe the music you want to generate
84994
- * @example Jazz music for a late-night restaurant setting
84995
- * @example A lush, ambient soundscape featuring serene sounds, and a gentle, melancholic piano melody
84996
- * @example Hip-hop music, mellow keys and vinyl crackle
84997
- * @example House music with synthesizers, driving bass and a steady 4/4 beat
84998
- * @example Classical piano melody with emotional depth and gentle strings
84999
- */
85000
- prompt: string;
85001
- /**
85002
- * Refinement
85003
- * @description Refinement level - higher values may improve quality but take longer
85004
- * @default 100
85005
- * @example 100
85006
- * @example 200
85007
- */
85008
- refinement?: number;
85009
- /**
85010
- * Seed
85011
- * @description Random seed for reproducible results - leave empty for random generation
85012
- */
85013
- seed?: number;
85014
- }
85015
-
85016
- export interface BeatovenMusicGenerationOutput {
85017
- /**
85018
- * @description Generated audio file in WAV format
85019
- * @example {
85020
- * "url": "https://v3b.fal.media/files/b/rabbit/DBesSNPP6NwfhwMftene-_music_ZfniDF.wav"
85021
- * }
85022
- */
85023
- audio: Components.File;
85024
- /**
85025
- * Metadata
85026
- * @description Generation metadata including duration, sample rate, and parameters
85027
- */
85028
- metadata: {
85029
- [key: string]: { [x: string]: any } | null;
85030
- };
85031
- /**
85032
- * Prompt
85033
- * @description The processed prompt used for generation
85034
- */
85035
- prompt: string;
85036
- }
85037
-
85038
85249
  export interface ArgilAvatarsTextToVideoInput {
85039
85250
  /**
85040
85251
  * Avatar