fal-endpoint-types 1.3.24 → 1.3.25

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "fal-endpoint-types",
3
- "version": "1.3.24",
3
+ "version": "1.3.25",
4
4
  "description": "TypeScript types for Fal AI endpoints generated from the OpenAPI schemas.",
5
5
  "homepage": "https://github.com/rawpixel-vincent/fal-endpoint-types#readme",
6
6
  "bugs": {
@@ -30,10 +30,14 @@ export interface WhisperChunk {
30
30
  }
31
31
 
32
32
  export interface VoiceSetting {
33
+ /**
34
+ * Custom Voice Id
35
+ * @description Custom cloned voice ID. If provided, this will override the voice_id field.
36
+ */
37
+ custom_voice_id?: string;
33
38
  /**
34
39
  * Emotion
35
40
  * @description Emotion of the generated speech
36
- * @enum {string}
37
41
  */
38
42
  emotion?: 'happy' | 'sad' | 'angry' | 'fearful' | 'disgusted' | 'surprised' | 'neutral';
39
43
  /**
@@ -247,7 +251,7 @@ export interface VideoFile {
247
251
  width?: number;
248
252
  }
249
253
 
250
- export interface VideoConditioningInput_2 {
254
+ export interface VideoConditioningInput_3 {
251
255
  /**
252
256
  * Conditioning Type
253
257
  * @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
@@ -313,7 +317,7 @@ export interface VideoConditioningInput_2 {
313
317
  video_url: string;
314
318
  }
315
319
 
316
- export interface VideoConditioningInput_1 {
320
+ export interface VideoConditioningInput_2 {
317
321
  /**
318
322
  * Start Frame Num
319
323
  * @description Frame number of the video from which the conditioning starts. Must be a multiple of 8.
@@ -326,7 +330,7 @@ export interface VideoConditioningInput_1 {
326
330
  video_url: string;
327
331
  }
328
332
 
329
- export interface VideoConditioningInput {
333
+ export interface VideoConditioningInput_1 {
330
334
  /**
331
335
  * Conditioning Type
332
336
  * @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
@@ -392,6 +396,71 @@ export interface VideoConditioningInput {
392
396
  video_url: string;
393
397
  }
394
398
 
399
+ export interface VideoConditioningInput {
400
+ /**
401
+ * Conditioning Type
402
+ * @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
403
+ * @default rgb
404
+ * @example rgb
405
+ */
406
+ conditioning_type?: 'rgb' | 'depth' | 'pose' | 'canny';
407
+ /**
408
+ * Limit Number of Frames
409
+ * @description Whether to limit the number of frames used from the video. If True, the `max_num_frames` parameter will be used to limit the number of frames.
410
+ * @default false
411
+ */
412
+ limit_num_frames?: boolean;
413
+ /**
414
+ * Maximum Number of Frames
415
+ * @description Maximum number of frames to use from the video. If None, all frames will be used.
416
+ * @default 1441
417
+ * @example 1441
418
+ */
419
+ max_num_frames?: number;
420
+ /**
421
+ * Preprocess
422
+ * @description Whether to preprocess the video. If True, the video will be preprocessed to match the conditioning type. This is a no-op for RGB conditioning.
423
+ * @default false
424
+ */
425
+ preprocess?: boolean;
426
+ /**
427
+ * Resample FPS
428
+ * @description Whether to resample the video to a specific FPS. If True, the `target_fps` parameter will be used to resample the video.
429
+ * @default false
430
+ */
431
+ resample_fps?: boolean;
432
+ /**
433
+ * Reverse Video
434
+ * @description Whether to reverse the video. This is useful for tasks where the video conditioning should be applied in reverse order.
435
+ * @default false
436
+ */
437
+ reverse_video?: boolean;
438
+ /**
439
+ * Start Frame Number
440
+ * @description Frame number of the video from which the conditioning starts. Must be a multiple of 8.
441
+ * @default 0
442
+ */
443
+ start_frame_num?: number;
444
+ /**
445
+ * Strength
446
+ * @description Strength of the conditioning. 0.0 means no conditioning, 1.0 means full conditioning.
447
+ * @default 1
448
+ */
449
+ strength?: number;
450
+ /**
451
+ * Target FPS
452
+ * @description Target FPS to resample the video to. Only relevant if `resample_fps` is True.
453
+ * @default 24
454
+ * @example 24
455
+ */
456
+ target_fps?: number;
457
+ /**
458
+ * Video URL
459
+ * @description URL of video to use as conditioning
460
+ */
461
+ video_url: string;
462
+ }
463
+
395
464
  export interface VideoCondition {
396
465
  /**
397
466
  * Start Frame Number
@@ -761,22 +830,22 @@ export interface TextureFiles {
761
830
  * Base Color
762
831
  * @description Base color texture
763
832
  */
764
- base_color: File_1;
833
+ base_color: File;
765
834
  /**
766
835
  * Metallic
767
836
  * @description Metallic texture (PBR)
768
837
  */
769
- metallic?: File_1;
838
+ metallic?: File;
770
839
  /**
771
840
  * Normal
772
841
  * @description Normal texture (PBR)
773
842
  */
774
- normal?: File_1;
843
+ normal?: File;
775
844
  /**
776
845
  * Roughness
777
846
  * @description Roughness texture (PBR)
778
847
  */
779
- roughness?: File_1;
848
+ roughness?: File;
780
849
  }
781
850
 
782
851
  export interface TextRender {
@@ -1374,7 +1443,6 @@ export interface PointPromptBase {
1374
1443
  /**
1375
1444
  * Label
1376
1445
  * @description 1 for foreground, 0 for background
1377
- * @enum {integer}
1378
1446
  */
1379
1447
  label?: 0 | 1;
1380
1448
  /**
@@ -1431,7 +1499,6 @@ export interface PointPrompt {
1431
1499
  /**
1432
1500
  * Label
1433
1501
  * @description 1 for foreground, 0 for background
1434
- * @enum {integer}
1435
1502
  */
1436
1503
  label?: 0 | 1;
1437
1504
  /**
@@ -1719,22 +1786,22 @@ export interface ModelUrls_2 {
1719
1786
  * Fbx
1720
1787
  * @description FBX format 3D model
1721
1788
  */
1722
- fbx?: File_1;
1789
+ fbx?: File;
1723
1790
  /**
1724
1791
  * Glb
1725
1792
  * @description GLB format 3D model
1726
1793
  */
1727
- glb?: File_1;
1794
+ glb?: File;
1728
1795
  /**
1729
1796
  * Obj
1730
1797
  * @description OBJ format 3D model
1731
1798
  */
1732
- obj?: File_1;
1799
+ obj?: File;
1733
1800
  /**
1734
1801
  * Usdz
1735
1802
  * @description USDZ format 3D model
1736
1803
  */
1737
- usdz?: File_1;
1804
+ usdz?: File;
1738
1805
  }
1739
1806
 
1740
1807
  export interface ModelUrls_1 {
@@ -1742,32 +1809,32 @@ export interface ModelUrls_1 {
1742
1809
  * Fbx
1743
1810
  * @description FBX format 3D model
1744
1811
  */
1745
- fbx?: File_1;
1812
+ fbx?: File;
1746
1813
  /**
1747
1814
  * Glb
1748
1815
  * @description GLB format 3D model
1749
1816
  */
1750
- glb?: File_1;
1817
+ glb?: File;
1751
1818
  /**
1752
1819
  * Mtl
1753
1820
  * @description MTL material file for OBJ model
1754
1821
  */
1755
- mtl?: File_1;
1822
+ mtl?: File;
1756
1823
  /**
1757
1824
  * Obj
1758
1825
  * @description OBJ format 3D model
1759
1826
  */
1760
- obj?: File_1;
1827
+ obj?: File;
1761
1828
  /**
1762
1829
  * Texture
1763
1830
  * @description Texture image for the 3D model
1764
1831
  */
1765
- texture?: File_1;
1832
+ texture?: File;
1766
1833
  /**
1767
1834
  * Usdz
1768
1835
  * @description USDZ format 3D model
1769
1836
  */
1770
- usdz?: File_1;
1837
+ usdz?: File;
1771
1838
  }
1772
1839
 
1773
1840
  export interface ModelUrls {
@@ -1775,32 +1842,32 @@ export interface ModelUrls {
1775
1842
  * Blend
1776
1843
  * @description Blender format 3D model
1777
1844
  */
1778
- blend?: File_1;
1845
+ blend?: File;
1779
1846
  /**
1780
1847
  * Fbx
1781
1848
  * @description FBX format 3D model
1782
1849
  */
1783
- fbx?: File_1;
1850
+ fbx?: File;
1784
1851
  /**
1785
1852
  * Glb
1786
1853
  * @description GLB format 3D model
1787
1854
  */
1788
- glb?: File_1;
1855
+ glb?: File;
1789
1856
  /**
1790
1857
  * Obj
1791
1858
  * @description OBJ format 3D model
1792
1859
  */
1793
- obj?: File_1;
1860
+ obj?: File;
1794
1861
  /**
1795
1862
  * Stl
1796
1863
  * @description STL format 3D model
1797
1864
  */
1798
- stl?: File_1;
1865
+ stl?: File;
1799
1866
  /**
1800
1867
  * Usdz
1801
1868
  * @description USDZ format 3D model
1802
1869
  */
1803
- usdz?: File_1;
1870
+ usdz?: File;
1804
1871
  }
1805
1872
 
1806
1873
  export interface MaskMetadata {
@@ -2155,8 +2222,6 @@ export interface KlingV3ImageElementInput {
2155
2222
  /**
2156
2223
  * Frontal Image Url
2157
2224
  * @description The frontal image of the element (main view).
2158
- *
2159
- * Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
2160
2225
  */
2161
2226
  frontal_image_url?: string;
2162
2227
  /**
@@ -2170,8 +2235,6 @@ export interface KlingV3ComboElementInput {
2170
2235
  /**
2171
2236
  * Frontal Image Url
2172
2237
  * @description The frontal image of the element (main view).
2173
- *
2174
- * Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
2175
2238
  */
2176
2239
  frontal_image_url?: string;
2177
2240
  /**
@@ -2182,8 +2245,6 @@ export interface KlingV3ComboElementInput {
2182
2245
  /**
2183
2246
  * Video Url
2184
2247
  * @description The video URL of the element. A request can only have one element with a video.
2185
- *
2186
- * Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
2187
2248
  */
2188
2249
  video_url?: string;
2189
2250
  }
@@ -2755,6 +2816,12 @@ export interface File_1 {
2755
2816
  * @example image/png
2756
2817
  */
2757
2818
  content_type?: string;
2819
+ /**
2820
+ * File Data
2821
+ * Format: binary
2822
+ * @description File data
2823
+ */
2824
+ file_data?: string;
2758
2825
  /**
2759
2826
  * File Name
2760
2827
  * @description The name of the file. It will be auto-generated if not provided.
@@ -2781,12 +2848,6 @@ export interface File {
2781
2848
  * @example image/png
2782
2849
  */
2783
2850
  content_type?: string;
2784
- /**
2785
- * File Data
2786
- * Format: binary
2787
- * @description File data
2788
- */
2789
- file_data?: string;
2790
2851
  /**
2791
2852
  * File Name
2792
2853
  * @description The name of the file. It will be auto-generated if not provided.
@@ -3745,12 +3806,6 @@ export interface AudioFile {
3745
3806
  * @description The duration of the audio
3746
3807
  */
3747
3808
  duration?: number;
3748
- /**
3749
- * File Data
3750
- * Format: binary
3751
- * @description File data
3752
- */
3753
- file_data?: string;
3754
3809
  /**
3755
3810
  * File Name
3756
3811
  * @description The name of the file. It will be auto-generated if not provided.
@@ -2039,6 +2039,11 @@ declare global {
2039
2039
  output: falEndpoints.PhotomakerOutput;
2040
2040
  };
2041
2041
 
2042
+ 'fal-ai/personaplex/realtime': {
2043
+ input: falEndpoints.PersonaplexRealtimeInput;
2044
+ output: falEndpoints.PersonaplexRealtimeOutput;
2045
+ };
2046
+
2042
2047
  'fal-ai/personaplex': {
2043
2048
  input: falEndpoints.PersonaplexInput;
2044
2049
  output: falEndpoints.PersonaplexOutput;
@@ -2910,8 +2915,8 @@ declare global {
2910
2915
  };
2911
2916
 
2912
2917
  'fal-ai/live-portrait/image': {
2913
- input: { [x in string]: any };
2914
- output: { [x in string]: any };
2918
+ input: falEndpoints.LivePortraitImageInput;
2919
+ output: falEndpoints.LivePortraitImageOutput;
2915
2920
  };
2916
2921
 
2917
2922
  'fal-ai/live-portrait': {
@@ -5489,6 +5494,11 @@ declare global {
5489
5494
  output: falEndpoints.BriaBackgroundRemoveOutput;
5490
5495
  };
5491
5496
 
5497
+ 'fal-ai/bitdance': {
5498
+ input: falEndpoints.BitdanceInput;
5499
+ output: falEndpoints.BitdanceOutput;
5500
+ };
5501
+
5492
5502
  'fal-ai/birefnet/v2/video': {
5493
5503
  input: falEndpoints.BirefnetV2VideoInput;
5494
5504
  output: falEndpoints.BirefnetV2VideoOutput;
@@ -5824,6 +5834,11 @@ declare global {
5824
5834
  output: falEndpoints.BriaBria_video_eraserEraseKeypointsOutput;
5825
5835
  };
5826
5836
 
5837
+ 'bria/aesthetics/upscaler': {
5838
+ input: { [x in string]: any };
5839
+ output: { [x in string]: any };
5840
+ };
5841
+
5827
5842
  'beatoven/sound-effect-generation': {
5828
5843
  input: falEndpoints.BeatovenSoundEffectGenerationInput;
5829
5844
  output: falEndpoints.BeatovenSoundEffectGenerationOutput;