fal-endpoint-types 1.3.24 → 1.3.26

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "fal-endpoint-types",
3
- "version": "1.3.24",
3
+ "version": "1.3.26",
4
4
  "description": "TypeScript types for Fal AI endpoints generated from the OpenAPI schemas.",
5
5
  "homepage": "https://github.com/rawpixel-vincent/fal-endpoint-types#readme",
6
6
  "bugs": {
@@ -26,14 +26,18 @@ export interface WhisperChunk {
26
26
  * Timestamp
27
27
  * @description Start and end timestamp of the chunk
28
28
  */
29
- timestamp: { [x: string]: any }[];
29
+ timestamp: [number, number];
30
30
  }
31
31
 
32
32
  export interface VoiceSetting {
33
+ /**
34
+ * Custom Voice Id
35
+ * @description Custom cloned voice ID. If provided, this will override the voice_id field.
36
+ */
37
+ custom_voice_id?: string;
33
38
  /**
34
39
  * Emotion
35
40
  * @description Emotion of the generated speech
36
- * @enum {string}
37
41
  */
38
42
  emotion?: 'happy' | 'sad' | 'angry' | 'fearful' | 'disgusted' | 'surprised' | 'neutral';
39
43
  /**
@@ -247,7 +251,7 @@ export interface VideoFile {
247
251
  width?: number;
248
252
  }
249
253
 
250
- export interface VideoConditioningInput_2 {
254
+ export interface VideoConditioningInput_3 {
251
255
  /**
252
256
  * Conditioning Type
253
257
  * @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
@@ -313,7 +317,7 @@ export interface VideoConditioningInput_2 {
313
317
  video_url: string;
314
318
  }
315
319
 
316
- export interface VideoConditioningInput_1 {
320
+ export interface VideoConditioningInput_2 {
317
321
  /**
318
322
  * Start Frame Num
319
323
  * @description Frame number of the video from which the conditioning starts. Must be a multiple of 8.
@@ -326,7 +330,7 @@ export interface VideoConditioningInput_1 {
326
330
  video_url: string;
327
331
  }
328
332
 
329
- export interface VideoConditioningInput {
333
+ export interface VideoConditioningInput_1 {
330
334
  /**
331
335
  * Conditioning Type
332
336
  * @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
@@ -392,6 +396,71 @@ export interface VideoConditioningInput {
392
396
  video_url: string;
393
397
  }
394
398
 
399
+ export interface VideoConditioningInput {
400
+ /**
401
+ * Conditioning Type
402
+ * @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
403
+ * @default rgb
404
+ * @example rgb
405
+ */
406
+ conditioning_type?: 'rgb' | 'depth' | 'pose' | 'canny';
407
+ /**
408
+ * Limit Number of Frames
409
+ * @description Whether to limit the number of frames used from the video. If True, the `max_num_frames` parameter will be used to limit the number of frames.
410
+ * @default false
411
+ */
412
+ limit_num_frames?: boolean;
413
+ /**
414
+ * Maximum Number of Frames
415
+ * @description Maximum number of frames to use from the video. If None, all frames will be used.
416
+ * @default 1441
417
+ * @example 1441
418
+ */
419
+ max_num_frames?: number;
420
+ /**
421
+ * Preprocess
422
+ * @description Whether to preprocess the video. If True, the video will be preprocessed to match the conditioning type. This is a no-op for RGB conditioning.
423
+ * @default false
424
+ */
425
+ preprocess?: boolean;
426
+ /**
427
+ * Resample FPS
428
+ * @description Whether to resample the video to a specific FPS. If True, the `target_fps` parameter will be used to resample the video.
429
+ * @default false
430
+ */
431
+ resample_fps?: boolean;
432
+ /**
433
+ * Reverse Video
434
+ * @description Whether to reverse the video. This is useful for tasks where the video conditioning should be applied in reverse order.
435
+ * @default false
436
+ */
437
+ reverse_video?: boolean;
438
+ /**
439
+ * Start Frame Number
440
+ * @description Frame number of the video from which the conditioning starts. Must be a multiple of 8.
441
+ * @default 0
442
+ */
443
+ start_frame_num?: number;
444
+ /**
445
+ * Strength
446
+ * @description Strength of the conditioning. 0.0 means no conditioning, 1.0 means full conditioning.
447
+ * @default 1
448
+ */
449
+ strength?: number;
450
+ /**
451
+ * Target FPS
452
+ * @description Target FPS to resample the video to. Only relevant if `resample_fps` is True.
453
+ * @default 24
454
+ * @example 24
455
+ */
456
+ target_fps?: number;
457
+ /**
458
+ * Video URL
459
+ * @description URL of video to use as conditioning
460
+ */
461
+ video_url: string;
462
+ }
463
+
395
464
  export interface VideoCondition {
396
465
  /**
397
466
  * Start Frame Number
@@ -761,22 +830,22 @@ export interface TextureFiles {
761
830
  * Base Color
762
831
  * @description Base color texture
763
832
  */
764
- base_color: File_1;
833
+ base_color: File;
765
834
  /**
766
835
  * Metallic
767
836
  * @description Metallic texture (PBR)
768
837
  */
769
- metallic?: File_1;
838
+ metallic?: File;
770
839
  /**
771
840
  * Normal
772
841
  * @description Normal texture (PBR)
773
842
  */
774
- normal?: File_1;
843
+ normal?: File;
775
844
  /**
776
845
  * Roughness
777
846
  * @description Roughness texture (PBR)
778
847
  */
779
- roughness?: File_1;
848
+ roughness?: File;
780
849
  }
781
850
 
782
851
  export interface TextRender {
@@ -998,6 +1067,21 @@ export interface SigmasInput {
998
1067
  method?: 'default' | 'array';
999
1068
  }
1000
1069
 
1070
+ export interface Shot {
1071
+ /**
1072
+ * Num Frames
1073
+ * @description Number of frames for this shot. Each shot must have between 17 and 100 frames.
1074
+ * @default 65
1075
+ */
1076
+ num_frames?: number;
1077
+ /**
1078
+ * Prompt
1079
+ * @description Caption describing what happens in this shot, including subject actions, camera angles, and scene details.
1080
+ * @example Subject 1 is smiling and gesturing with his right hand while facing the camera. Subject 2 is seen from behind, facing Subject 1. The scene is indoors in a well-decorated room, with a dark door and ornate glass visible in the background. The camera is static, providing a medium close-up shot of Subject 1.
1081
+ */
1082
+ prompt: string;
1083
+ }
1084
+
1001
1085
  export interface SemanticImageInput {
1002
1086
  /**
1003
1087
  * Hypothesis
@@ -1374,7 +1458,6 @@ export interface PointPromptBase {
1374
1458
  /**
1375
1459
  * Label
1376
1460
  * @description 1 for foreground, 0 for background
1377
- * @enum {integer}
1378
1461
  */
1379
1462
  label?: 0 | 1;
1380
1463
  /**
@@ -1431,7 +1514,6 @@ export interface PointPrompt {
1431
1514
  /**
1432
1515
  * Label
1433
1516
  * @description 1 for foreground, 0 for background
1434
- * @enum {integer}
1435
1517
  */
1436
1518
  label?: 0 | 1;
1437
1519
  /**
@@ -1634,24 +1716,6 @@ export interface Object {
1634
1716
  y_min: number;
1635
1717
  }
1636
1718
 
1637
- export interface NovaSRTimings {
1638
- /**
1639
- * Inference
1640
- * @description Time taken to run the inference in seconds.
1641
- */
1642
- inference: number;
1643
- /**
1644
- * Postprocess
1645
- * @description Time taken to postprocess the audio in seconds.
1646
- */
1647
- postprocess: number;
1648
- /**
1649
- * Preprocess
1650
- * @description Time taken to preprocess the audio in seconds.
1651
- */
1652
- preprocess: number;
1653
- }
1654
-
1655
1719
  export interface MusicSection {
1656
1720
  /**
1657
1721
  * Duration Ms
@@ -1719,22 +1783,22 @@ export interface ModelUrls_2 {
1719
1783
  * Fbx
1720
1784
  * @description FBX format 3D model
1721
1785
  */
1722
- fbx?: File_1;
1786
+ fbx?: File;
1723
1787
  /**
1724
1788
  * Glb
1725
1789
  * @description GLB format 3D model
1726
1790
  */
1727
- glb?: File_1;
1791
+ glb?: File;
1728
1792
  /**
1729
1793
  * Obj
1730
1794
  * @description OBJ format 3D model
1731
1795
  */
1732
- obj?: File_1;
1796
+ obj?: File;
1733
1797
  /**
1734
1798
  * Usdz
1735
1799
  * @description USDZ format 3D model
1736
1800
  */
1737
- usdz?: File_1;
1801
+ usdz?: File;
1738
1802
  }
1739
1803
 
1740
1804
  export interface ModelUrls_1 {
@@ -1742,32 +1806,32 @@ export interface ModelUrls_1 {
1742
1806
  * Fbx
1743
1807
  * @description FBX format 3D model
1744
1808
  */
1745
- fbx?: File_1;
1809
+ fbx?: File;
1746
1810
  /**
1747
1811
  * Glb
1748
1812
  * @description GLB format 3D model
1749
1813
  */
1750
- glb?: File_1;
1814
+ glb?: File;
1751
1815
  /**
1752
1816
  * Mtl
1753
1817
  * @description MTL material file for OBJ model
1754
1818
  */
1755
- mtl?: File_1;
1819
+ mtl?: File;
1756
1820
  /**
1757
1821
  * Obj
1758
1822
  * @description OBJ format 3D model
1759
1823
  */
1760
- obj?: File_1;
1824
+ obj?: File;
1761
1825
  /**
1762
1826
  * Texture
1763
1827
  * @description Texture image for the 3D model
1764
1828
  */
1765
- texture?: File_1;
1829
+ texture?: File;
1766
1830
  /**
1767
1831
  * Usdz
1768
1832
  * @description USDZ format 3D model
1769
1833
  */
1770
- usdz?: File_1;
1834
+ usdz?: File;
1771
1835
  }
1772
1836
 
1773
1837
  export interface ModelUrls {
@@ -1775,32 +1839,32 @@ export interface ModelUrls {
1775
1839
  * Blend
1776
1840
  * @description Blender format 3D model
1777
1841
  */
1778
- blend?: File_1;
1842
+ blend?: File;
1779
1843
  /**
1780
1844
  * Fbx
1781
1845
  * @description FBX format 3D model
1782
1846
  */
1783
- fbx?: File_1;
1847
+ fbx?: File;
1784
1848
  /**
1785
1849
  * Glb
1786
1850
  * @description GLB format 3D model
1787
1851
  */
1788
- glb?: File_1;
1852
+ glb?: File;
1789
1853
  /**
1790
1854
  * Obj
1791
1855
  * @description OBJ format 3D model
1792
1856
  */
1793
- obj?: File_1;
1857
+ obj?: File;
1794
1858
  /**
1795
1859
  * Stl
1796
1860
  * @description STL format 3D model
1797
1861
  */
1798
- stl?: File_1;
1862
+ stl?: File;
1799
1863
  /**
1800
1864
  * Usdz
1801
1865
  * @description USDZ format 3D model
1802
1866
  */
1803
- usdz?: File_1;
1867
+ usdz?: File;
1804
1868
  }
1805
1869
 
1806
1870
  export interface MaskMetadata {
@@ -2136,6 +2200,24 @@ export interface Lighting {
2136
2200
  shadows?: string;
2137
2201
  }
2138
2202
 
2203
+ export interface LavaSRTimings {
2204
+ /**
2205
+ * Inference
2206
+ * @description Time taken to run the inference in seconds.
2207
+ */
2208
+ inference: number;
2209
+ /**
2210
+ * Postprocess
2211
+ * @description Time taken to postprocess the audio in seconds.
2212
+ */
2213
+ postprocess: number;
2214
+ /**
2215
+ * Preprocess
2216
+ * @description Time taken to preprocess the audio in seconds.
2217
+ */
2218
+ preprocess: number;
2219
+ }
2220
+
2139
2221
  export interface KlingV3MultiPromptElement {
2140
2222
  /**
2141
2223
  * Duration
@@ -2755,6 +2837,12 @@ export interface File_1 {
2755
2837
  * @example image/png
2756
2838
  */
2757
2839
  content_type?: string;
2840
+ /**
2841
+ * File Data
2842
+ * Format: binary
2843
+ * @description File data
2844
+ */
2845
+ file_data?: string;
2758
2846
  /**
2759
2847
  * File Name
2760
2848
  * @description The name of the file. It will be auto-generated if not provided.
@@ -2781,12 +2869,6 @@ export interface File {
2781
2869
  * @example image/png
2782
2870
  */
2783
2871
  content_type?: string;
2784
- /**
2785
- * File Data
2786
- * Format: binary
2787
- * @description File data
2788
- */
2789
- file_data?: string;
2790
2872
  /**
2791
2873
  * File Name
2792
2874
  * @description The name of the file. It will be auto-generated if not provided.
@@ -2909,7 +2991,7 @@ export interface EmotionalStrengths {
2909
2991
  surprised?: number;
2910
2992
  }
2911
2993
 
2912
- export interface Embedding_1 {
2994
+ export interface Embedding_2 {
2913
2995
  /**
2914
2996
  * Path
2915
2997
  * @description URL or the path to the embedding weights.
@@ -2926,6 +3008,31 @@ export interface Embedding_1 {
2926
3008
  tokens?: string[];
2927
3009
  }
2928
3010
 
3011
+ export interface Embedding_1 {
3012
+ /**
3013
+ * Force
3014
+ * @description If set to true, the embedding will be forced to be used.
3015
+ * @default false
3016
+ */
3017
+ force?: boolean;
3018
+ /**
3019
+ * Path
3020
+ * @description URL or the path to the embedding weights.
3021
+ * @example https://civitai.com/api/download/models/135931
3022
+ * @example https://filebin.net/3chfqasxpqu21y8n/my-custom-lora-v1.safetensors
3023
+ */
3024
+ path: string;
3025
+ /**
3026
+ * Tokens
3027
+ * @description The list of tokens to use for the embedding.
3028
+ * @default [
3029
+ * "<s0>",
3030
+ * "<s1>"
3031
+ * ]
3032
+ */
3033
+ tokens?: string[];
3034
+ }
3035
+
2929
3036
  export interface Embedding {
2930
3037
  /**
2931
3038
  * Path
@@ -3025,7 +3132,7 @@ export interface DiarizationSegment {
3025
3132
  * Timestamp
3026
3133
  * @description Start and end timestamp of the segment
3027
3134
  */
3028
- timestamp: { [x: string]: any }[];
3135
+ timestamp: [number, number];
3029
3136
  }
3030
3137
 
3031
3138
  export interface DialogueBlock {
@@ -3745,12 +3852,6 @@ export interface AudioFile {
3745
3852
  * @description The duration of the audio
3746
3853
  */
3747
3854
  duration?: number;
3748
- /**
3749
- * File Data
3750
- * Format: binary
3751
- * @description File data
3752
- */
3753
- file_data?: string;
3754
3855
  /**
3755
3856
  * File Name
3756
3857
  * @description The name of the file. It will be auto-generated if not provided.
@@ -624,6 +624,11 @@ declare global {
624
624
  output: { [x in string]: any };
625
625
  };
626
626
 
627
+ 'fal-ai/wan-motion': {
628
+ input: falEndpoints.WanMotionInput;
629
+ output: falEndpoints.WanMotionOutput;
630
+ };
631
+
627
632
  'fal-ai/wan-i2v-lora': {
628
633
  input: falEndpoints.WanI2vLoraInput;
629
634
  output: falEndpoints.WanI2vLoraOutput;
@@ -704,6 +709,16 @@ declare global {
704
709
  output: falEndpoints.Wan22VaceFunA14bDepthOutput;
705
710
  };
706
711
 
712
+ 'fal-ai/wan-22-trainer/t2v-a14b': {
713
+ input: falEndpoints.Wan22TrainerT2vA14bInput;
714
+ output: falEndpoints.Wan22TrainerT2vA14bOutput;
715
+ };
716
+
717
+ 'fal-ai/wan-22-trainer/i2v-a14b': {
718
+ input: falEndpoints.Wan22TrainerI2vA14bInput;
719
+ output: falEndpoints.Wan22TrainerI2vA14bOutput;
720
+ };
721
+
707
722
  'fal-ai/wan-22-image-trainer': {
708
723
  input: falEndpoints.Wan22ImageTrainerInput;
709
724
  output: falEndpoints.Wan22ImageTrainerOutput;
@@ -2169,6 +2184,11 @@ declare global {
2169
2184
  output: falEndpoints.MusetalkOutput;
2170
2185
  };
2171
2186
 
2187
+ 'fal-ai/multishot-master': {
2188
+ input: falEndpoints.MultishotMasterInput;
2189
+ output: falEndpoints.MultishotMasterOutput;
2190
+ };
2191
+
2172
2192
  'fal-ai/moondream3-preview/segment': {
2173
2193
  input: { [x in string]: any };
2174
2194
  output: { [x in string]: any };
@@ -2910,8 +2930,8 @@ declare global {
2910
2930
  };
2911
2931
 
2912
2932
  'fal-ai/live-portrait/image': {
2913
- input: { [x in string]: any };
2914
- output: { [x in string]: any };
2933
+ input: falEndpoints.LivePortraitImageInput;
2934
+ output: falEndpoints.LivePortraitImageOutput;
2915
2935
  };
2916
2936
 
2917
2937
  'fal-ai/live-portrait': {
@@ -2964,6 +2984,11 @@ declare global {
2964
2984
  output: falEndpoints.LayerDiffusionOutput;
2965
2985
  };
2966
2986
 
2987
+ 'fal-ai/lava-sr': {
2988
+ input: falEndpoints.LavaSrInput;
2989
+ output: falEndpoints.LavaSrOutput;
2990
+ };
2991
+
2967
2992
  'fal-ai/latentsync': {
2968
2993
  input: falEndpoints.LatentsyncInput;
2969
2994
  output: falEndpoints.LatentsyncOutput;
@@ -5244,6 +5269,26 @@ declare global {
5244
5269
  output: falEndpoints.CreatifyAuroraOutput;
5245
5270
  };
5246
5271
 
5272
+ 'fal-ai/cosmos-predict-2.5/video-to-video': {
5273
+ input: falEndpoints.CosmosPredict25VideoToVideoInput;
5274
+ output: falEndpoints.CosmosPredict25VideoToVideoOutput;
5275
+ };
5276
+
5277
+ 'fal-ai/cosmos-predict-2.5/text-to-video': {
5278
+ input: falEndpoints.CosmosPredict25TextToVideoInput;
5279
+ output: falEndpoints.CosmosPredict25TextToVideoOutput;
5280
+ };
5281
+
5282
+ 'fal-ai/cosmos-predict-2.5/image-to-video': {
5283
+ input: falEndpoints.CosmosPredict25ImageToVideoInput;
5284
+ output: falEndpoints.CosmosPredict25ImageToVideoOutput;
5285
+ };
5286
+
5287
+ 'fal-ai/cosmos-predict-2.5/distilled/text-to-video': {
5288
+ input: falEndpoints.CosmosPredict25DistilledTextToVideoInput;
5289
+ output: falEndpoints.CosmosPredict25DistilledTextToVideoOutput;
5290
+ };
5291
+
5247
5292
  'fal-ai/controlnext': {
5248
5293
  input: falEndpoints.ControlnextInput;
5249
5294
  output: falEndpoints.ControlnextOutput;
@@ -5344,6 +5389,16 @@ declare global {
5344
5389
  output: falEndpoints.BytedanceVideoStylizeOutput;
5345
5390
  };
5346
5391
 
5392
+ 'fal-ai/bytedance/seedream/v5/lite/text-to-image': {
5393
+ input: falEndpoints.BytedanceSeedreamV5LiteTextToImageInput;
5394
+ output: falEndpoints.BytedanceSeedreamV5LiteTextToImageOutput;
5395
+ };
5396
+
5397
+ 'fal-ai/bytedance/seedream/v5/lite/edit': {
5398
+ input: falEndpoints.BytedanceSeedreamV5LiteEditInput;
5399
+ output: falEndpoints.BytedanceSeedreamV5LiteEditOutput;
5400
+ };
5401
+
5347
5402
  'fal-ai/bytedance/seedream/v4/text-to-image': {
5348
5403
  input: falEndpoints.BytedanceSeedreamV4TextToImageInput;
5349
5404
  output: falEndpoints.BytedanceSeedreamV4TextToImageOutput;
@@ -5489,6 +5544,11 @@ declare global {
5489
5544
  output: falEndpoints.BriaBackgroundRemoveOutput;
5490
5545
  };
5491
5546
 
5547
+ 'fal-ai/bitdance': {
5548
+ input: falEndpoints.BitdanceInput;
5549
+ output: falEndpoints.BitdanceOutput;
5550
+ };
5551
+
5492
5552
  'fal-ai/birefnet/v2/video': {
5493
5553
  input: falEndpoints.BirefnetV2VideoInput;
5494
5554
  output: falEndpoints.BirefnetV2VideoOutput;
@@ -5629,6 +5689,11 @@ declare global {
5629
5689
  output: falEndpoints.DecartLucyRestyleOutput;
5630
5690
  };
5631
5691
 
5692
+ 'decart/lucy-i2v': {
5693
+ input: falEndpoints.DecartLucyI2vInput;
5694
+ output: falEndpoints.DecartLucyI2vOutput;
5695
+ };
5696
+
5632
5697
  'decart/lucy-edit/pro': {
5633
5698
  input: falEndpoints.DecartLucyEditProInput;
5634
5699
  output: falEndpoints.DecartLucyEditProOutput;
@@ -5644,11 +5709,6 @@ declare global {
5644
5709
  output: falEndpoints.DecartLucyEditDevOutput;
5645
5710
  };
5646
5711
 
5647
- 'decart/lucy-14b/image-to-video': {
5648
- input: falEndpoints.DecartLucy14bImageToVideoInput;
5649
- output: falEndpoints.DecartLucy14bImageToVideoOutput;
5650
- };
5651
-
5652
5712
  'clarityai/crystal-video-upscaler': {
5653
5713
  input: falEndpoints.ClarityaiCrystalVideoUpscalerInput;
5654
5714
  output: falEndpoints.ClarityaiCrystalVideoUpscalerOutput;
@@ -5704,6 +5764,11 @@ declare global {
5704
5764
  output: falEndpoints.BriaVideoBackgroundRemovalOutput;
5705
5765
  };
5706
5766
 
5767
+ 'bria/upscale/creative': {
5768
+ input: falEndpoints.BriaUpscaleCreativeInput;
5769
+ output: falEndpoints.BriaUpscaleCreativeOutput;
5770
+ };
5771
+
5707
5772
  'bria/text-to-image/3.2': {
5708
5773
  input: falEndpoints.BriaTextToImage32Input;
5709
5774
  output: falEndpoints.BriaTextToImage32Output;