fal-endpoint-types 1.3.24 → 1.3.26
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/types/fal/endpoints/components.d.ts +160 -59
- package/types/fal/endpoints/index.d.ts +72 -7
- package/types/fal/endpoints/schema.d.ts +6490 -3841
package/package.json
CHANGED
|
@@ -26,14 +26,18 @@ export interface WhisperChunk {
|
|
|
26
26
|
* Timestamp
|
|
27
27
|
* @description Start and end timestamp of the chunk
|
|
28
28
|
*/
|
|
29
|
-
timestamp:
|
|
29
|
+
timestamp: [number, number];
|
|
30
30
|
}
|
|
31
31
|
|
|
32
32
|
export interface VoiceSetting {
|
|
33
|
+
/**
|
|
34
|
+
* Custom Voice Id
|
|
35
|
+
* @description Custom cloned voice ID. If provided, this will override the voice_id field.
|
|
36
|
+
*/
|
|
37
|
+
custom_voice_id?: string;
|
|
33
38
|
/**
|
|
34
39
|
* Emotion
|
|
35
40
|
* @description Emotion of the generated speech
|
|
36
|
-
* @enum {string}
|
|
37
41
|
*/
|
|
38
42
|
emotion?: 'happy' | 'sad' | 'angry' | 'fearful' | 'disgusted' | 'surprised' | 'neutral';
|
|
39
43
|
/**
|
|
@@ -247,7 +251,7 @@ export interface VideoFile {
|
|
|
247
251
|
width?: number;
|
|
248
252
|
}
|
|
249
253
|
|
|
250
|
-
export interface
|
|
254
|
+
export interface VideoConditioningInput_3 {
|
|
251
255
|
/**
|
|
252
256
|
* Conditioning Type
|
|
253
257
|
* @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
|
|
@@ -313,7 +317,7 @@ export interface VideoConditioningInput_2 {
|
|
|
313
317
|
video_url: string;
|
|
314
318
|
}
|
|
315
319
|
|
|
316
|
-
export interface
|
|
320
|
+
export interface VideoConditioningInput_2 {
|
|
317
321
|
/**
|
|
318
322
|
* Start Frame Num
|
|
319
323
|
* @description Frame number of the video from which the conditioning starts. Must be a multiple of 8.
|
|
@@ -326,7 +330,7 @@ export interface VideoConditioningInput_1 {
|
|
|
326
330
|
video_url: string;
|
|
327
331
|
}
|
|
328
332
|
|
|
329
|
-
export interface
|
|
333
|
+
export interface VideoConditioningInput_1 {
|
|
330
334
|
/**
|
|
331
335
|
* Conditioning Type
|
|
332
336
|
* @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
|
|
@@ -392,6 +396,71 @@ export interface VideoConditioningInput {
|
|
|
392
396
|
video_url: string;
|
|
393
397
|
}
|
|
394
398
|
|
|
399
|
+
export interface VideoConditioningInput {
|
|
400
|
+
/**
|
|
401
|
+
* Conditioning Type
|
|
402
|
+
* @description Type of conditioning this video provides. This is relevant to ensure in-context LoRA weights are applied correctly, as well as selecting the correct preprocessing pipeline, when enabled.
|
|
403
|
+
* @default rgb
|
|
404
|
+
* @example rgb
|
|
405
|
+
*/
|
|
406
|
+
conditioning_type?: 'rgb' | 'depth' | 'pose' | 'canny';
|
|
407
|
+
/**
|
|
408
|
+
* Limit Number of Frames
|
|
409
|
+
* @description Whether to limit the number of frames used from the video. If True, the `max_num_frames` parameter will be used to limit the number of frames.
|
|
410
|
+
* @default false
|
|
411
|
+
*/
|
|
412
|
+
limit_num_frames?: boolean;
|
|
413
|
+
/**
|
|
414
|
+
* Maximum Number of Frames
|
|
415
|
+
* @description Maximum number of frames to use from the video. If None, all frames will be used.
|
|
416
|
+
* @default 1441
|
|
417
|
+
* @example 1441
|
|
418
|
+
*/
|
|
419
|
+
max_num_frames?: number;
|
|
420
|
+
/**
|
|
421
|
+
* Preprocess
|
|
422
|
+
* @description Whether to preprocess the video. If True, the video will be preprocessed to match the conditioning type. This is a no-op for RGB conditioning.
|
|
423
|
+
* @default false
|
|
424
|
+
*/
|
|
425
|
+
preprocess?: boolean;
|
|
426
|
+
/**
|
|
427
|
+
* Resample FPS
|
|
428
|
+
* @description Whether to resample the video to a specific FPS. If True, the `target_fps` parameter will be used to resample the video.
|
|
429
|
+
* @default false
|
|
430
|
+
*/
|
|
431
|
+
resample_fps?: boolean;
|
|
432
|
+
/**
|
|
433
|
+
* Reverse Video
|
|
434
|
+
* @description Whether to reverse the video. This is useful for tasks where the video conditioning should be applied in reverse order.
|
|
435
|
+
* @default false
|
|
436
|
+
*/
|
|
437
|
+
reverse_video?: boolean;
|
|
438
|
+
/**
|
|
439
|
+
* Start Frame Number
|
|
440
|
+
* @description Frame number of the video from which the conditioning starts. Must be a multiple of 8.
|
|
441
|
+
* @default 0
|
|
442
|
+
*/
|
|
443
|
+
start_frame_num?: number;
|
|
444
|
+
/**
|
|
445
|
+
* Strength
|
|
446
|
+
* @description Strength of the conditioning. 0.0 means no conditioning, 1.0 means full conditioning.
|
|
447
|
+
* @default 1
|
|
448
|
+
*/
|
|
449
|
+
strength?: number;
|
|
450
|
+
/**
|
|
451
|
+
* Target FPS
|
|
452
|
+
* @description Target FPS to resample the video to. Only relevant if `resample_fps` is True.
|
|
453
|
+
* @default 24
|
|
454
|
+
* @example 24
|
|
455
|
+
*/
|
|
456
|
+
target_fps?: number;
|
|
457
|
+
/**
|
|
458
|
+
* Video URL
|
|
459
|
+
* @description URL of video to use as conditioning
|
|
460
|
+
*/
|
|
461
|
+
video_url: string;
|
|
462
|
+
}
|
|
463
|
+
|
|
395
464
|
export interface VideoCondition {
|
|
396
465
|
/**
|
|
397
466
|
* Start Frame Number
|
|
@@ -761,22 +830,22 @@ export interface TextureFiles {
|
|
|
761
830
|
* Base Color
|
|
762
831
|
* @description Base color texture
|
|
763
832
|
*/
|
|
764
|
-
base_color:
|
|
833
|
+
base_color: File;
|
|
765
834
|
/**
|
|
766
835
|
* Metallic
|
|
767
836
|
* @description Metallic texture (PBR)
|
|
768
837
|
*/
|
|
769
|
-
metallic?:
|
|
838
|
+
metallic?: File;
|
|
770
839
|
/**
|
|
771
840
|
* Normal
|
|
772
841
|
* @description Normal texture (PBR)
|
|
773
842
|
*/
|
|
774
|
-
normal?:
|
|
843
|
+
normal?: File;
|
|
775
844
|
/**
|
|
776
845
|
* Roughness
|
|
777
846
|
* @description Roughness texture (PBR)
|
|
778
847
|
*/
|
|
779
|
-
roughness?:
|
|
848
|
+
roughness?: File;
|
|
780
849
|
}
|
|
781
850
|
|
|
782
851
|
export interface TextRender {
|
|
@@ -998,6 +1067,21 @@ export interface SigmasInput {
|
|
|
998
1067
|
method?: 'default' | 'array';
|
|
999
1068
|
}
|
|
1000
1069
|
|
|
1070
|
+
export interface Shot {
|
|
1071
|
+
/**
|
|
1072
|
+
* Num Frames
|
|
1073
|
+
* @description Number of frames for this shot. Each shot must have between 17 and 100 frames.
|
|
1074
|
+
* @default 65
|
|
1075
|
+
*/
|
|
1076
|
+
num_frames?: number;
|
|
1077
|
+
/**
|
|
1078
|
+
* Prompt
|
|
1079
|
+
* @description Caption describing what happens in this shot, including subject actions, camera angles, and scene details.
|
|
1080
|
+
* @example Subject 1 is smiling and gesturing with his right hand while facing the camera. Subject 2 is seen from behind, facing Subject 1. The scene is indoors in a well-decorated room, with a dark door and ornate glass visible in the background. The camera is static, providing a medium close-up shot of Subject 1.
|
|
1081
|
+
*/
|
|
1082
|
+
prompt: string;
|
|
1083
|
+
}
|
|
1084
|
+
|
|
1001
1085
|
export interface SemanticImageInput {
|
|
1002
1086
|
/**
|
|
1003
1087
|
* Hypothesis
|
|
@@ -1374,7 +1458,6 @@ export interface PointPromptBase {
|
|
|
1374
1458
|
/**
|
|
1375
1459
|
* Label
|
|
1376
1460
|
* @description 1 for foreground, 0 for background
|
|
1377
|
-
* @enum {integer}
|
|
1378
1461
|
*/
|
|
1379
1462
|
label?: 0 | 1;
|
|
1380
1463
|
/**
|
|
@@ -1431,7 +1514,6 @@ export interface PointPrompt {
|
|
|
1431
1514
|
/**
|
|
1432
1515
|
* Label
|
|
1433
1516
|
* @description 1 for foreground, 0 for background
|
|
1434
|
-
* @enum {integer}
|
|
1435
1517
|
*/
|
|
1436
1518
|
label?: 0 | 1;
|
|
1437
1519
|
/**
|
|
@@ -1634,24 +1716,6 @@ export interface Object {
|
|
|
1634
1716
|
y_min: number;
|
|
1635
1717
|
}
|
|
1636
1718
|
|
|
1637
|
-
export interface NovaSRTimings {
|
|
1638
|
-
/**
|
|
1639
|
-
* Inference
|
|
1640
|
-
* @description Time taken to run the inference in seconds.
|
|
1641
|
-
*/
|
|
1642
|
-
inference: number;
|
|
1643
|
-
/**
|
|
1644
|
-
* Postprocess
|
|
1645
|
-
* @description Time taken to postprocess the audio in seconds.
|
|
1646
|
-
*/
|
|
1647
|
-
postprocess: number;
|
|
1648
|
-
/**
|
|
1649
|
-
* Preprocess
|
|
1650
|
-
* @description Time taken to preprocess the audio in seconds.
|
|
1651
|
-
*/
|
|
1652
|
-
preprocess: number;
|
|
1653
|
-
}
|
|
1654
|
-
|
|
1655
1719
|
export interface MusicSection {
|
|
1656
1720
|
/**
|
|
1657
1721
|
* Duration Ms
|
|
@@ -1719,22 +1783,22 @@ export interface ModelUrls_2 {
|
|
|
1719
1783
|
* Fbx
|
|
1720
1784
|
* @description FBX format 3D model
|
|
1721
1785
|
*/
|
|
1722
|
-
fbx?:
|
|
1786
|
+
fbx?: File;
|
|
1723
1787
|
/**
|
|
1724
1788
|
* Glb
|
|
1725
1789
|
* @description GLB format 3D model
|
|
1726
1790
|
*/
|
|
1727
|
-
glb?:
|
|
1791
|
+
glb?: File;
|
|
1728
1792
|
/**
|
|
1729
1793
|
* Obj
|
|
1730
1794
|
* @description OBJ format 3D model
|
|
1731
1795
|
*/
|
|
1732
|
-
obj?:
|
|
1796
|
+
obj?: File;
|
|
1733
1797
|
/**
|
|
1734
1798
|
* Usdz
|
|
1735
1799
|
* @description USDZ format 3D model
|
|
1736
1800
|
*/
|
|
1737
|
-
usdz?:
|
|
1801
|
+
usdz?: File;
|
|
1738
1802
|
}
|
|
1739
1803
|
|
|
1740
1804
|
export interface ModelUrls_1 {
|
|
@@ -1742,32 +1806,32 @@ export interface ModelUrls_1 {
|
|
|
1742
1806
|
* Fbx
|
|
1743
1807
|
* @description FBX format 3D model
|
|
1744
1808
|
*/
|
|
1745
|
-
fbx?:
|
|
1809
|
+
fbx?: File;
|
|
1746
1810
|
/**
|
|
1747
1811
|
* Glb
|
|
1748
1812
|
* @description GLB format 3D model
|
|
1749
1813
|
*/
|
|
1750
|
-
glb?:
|
|
1814
|
+
glb?: File;
|
|
1751
1815
|
/**
|
|
1752
1816
|
* Mtl
|
|
1753
1817
|
* @description MTL material file for OBJ model
|
|
1754
1818
|
*/
|
|
1755
|
-
mtl?:
|
|
1819
|
+
mtl?: File;
|
|
1756
1820
|
/**
|
|
1757
1821
|
* Obj
|
|
1758
1822
|
* @description OBJ format 3D model
|
|
1759
1823
|
*/
|
|
1760
|
-
obj?:
|
|
1824
|
+
obj?: File;
|
|
1761
1825
|
/**
|
|
1762
1826
|
* Texture
|
|
1763
1827
|
* @description Texture image for the 3D model
|
|
1764
1828
|
*/
|
|
1765
|
-
texture?:
|
|
1829
|
+
texture?: File;
|
|
1766
1830
|
/**
|
|
1767
1831
|
* Usdz
|
|
1768
1832
|
* @description USDZ format 3D model
|
|
1769
1833
|
*/
|
|
1770
|
-
usdz?:
|
|
1834
|
+
usdz?: File;
|
|
1771
1835
|
}
|
|
1772
1836
|
|
|
1773
1837
|
export interface ModelUrls {
|
|
@@ -1775,32 +1839,32 @@ export interface ModelUrls {
|
|
|
1775
1839
|
* Blend
|
|
1776
1840
|
* @description Blender format 3D model
|
|
1777
1841
|
*/
|
|
1778
|
-
blend?:
|
|
1842
|
+
blend?: File;
|
|
1779
1843
|
/**
|
|
1780
1844
|
* Fbx
|
|
1781
1845
|
* @description FBX format 3D model
|
|
1782
1846
|
*/
|
|
1783
|
-
fbx?:
|
|
1847
|
+
fbx?: File;
|
|
1784
1848
|
/**
|
|
1785
1849
|
* Glb
|
|
1786
1850
|
* @description GLB format 3D model
|
|
1787
1851
|
*/
|
|
1788
|
-
glb?:
|
|
1852
|
+
glb?: File;
|
|
1789
1853
|
/**
|
|
1790
1854
|
* Obj
|
|
1791
1855
|
* @description OBJ format 3D model
|
|
1792
1856
|
*/
|
|
1793
|
-
obj?:
|
|
1857
|
+
obj?: File;
|
|
1794
1858
|
/**
|
|
1795
1859
|
* Stl
|
|
1796
1860
|
* @description STL format 3D model
|
|
1797
1861
|
*/
|
|
1798
|
-
stl?:
|
|
1862
|
+
stl?: File;
|
|
1799
1863
|
/**
|
|
1800
1864
|
* Usdz
|
|
1801
1865
|
* @description USDZ format 3D model
|
|
1802
1866
|
*/
|
|
1803
|
-
usdz?:
|
|
1867
|
+
usdz?: File;
|
|
1804
1868
|
}
|
|
1805
1869
|
|
|
1806
1870
|
export interface MaskMetadata {
|
|
@@ -2136,6 +2200,24 @@ export interface Lighting {
|
|
|
2136
2200
|
shadows?: string;
|
|
2137
2201
|
}
|
|
2138
2202
|
|
|
2203
|
+
export interface LavaSRTimings {
|
|
2204
|
+
/**
|
|
2205
|
+
* Inference
|
|
2206
|
+
* @description Time taken to run the inference in seconds.
|
|
2207
|
+
*/
|
|
2208
|
+
inference: number;
|
|
2209
|
+
/**
|
|
2210
|
+
* Postprocess
|
|
2211
|
+
* @description Time taken to postprocess the audio in seconds.
|
|
2212
|
+
*/
|
|
2213
|
+
postprocess: number;
|
|
2214
|
+
/**
|
|
2215
|
+
* Preprocess
|
|
2216
|
+
* @description Time taken to preprocess the audio in seconds.
|
|
2217
|
+
*/
|
|
2218
|
+
preprocess: number;
|
|
2219
|
+
}
|
|
2220
|
+
|
|
2139
2221
|
export interface KlingV3MultiPromptElement {
|
|
2140
2222
|
/**
|
|
2141
2223
|
* Duration
|
|
@@ -2755,6 +2837,12 @@ export interface File_1 {
|
|
|
2755
2837
|
* @example image/png
|
|
2756
2838
|
*/
|
|
2757
2839
|
content_type?: string;
|
|
2840
|
+
/**
|
|
2841
|
+
* File Data
|
|
2842
|
+
* Format: binary
|
|
2843
|
+
* @description File data
|
|
2844
|
+
*/
|
|
2845
|
+
file_data?: string;
|
|
2758
2846
|
/**
|
|
2759
2847
|
* File Name
|
|
2760
2848
|
* @description The name of the file. It will be auto-generated if not provided.
|
|
@@ -2781,12 +2869,6 @@ export interface File {
|
|
|
2781
2869
|
* @example image/png
|
|
2782
2870
|
*/
|
|
2783
2871
|
content_type?: string;
|
|
2784
|
-
/**
|
|
2785
|
-
* File Data
|
|
2786
|
-
* Format: binary
|
|
2787
|
-
* @description File data
|
|
2788
|
-
*/
|
|
2789
|
-
file_data?: string;
|
|
2790
2872
|
/**
|
|
2791
2873
|
* File Name
|
|
2792
2874
|
* @description The name of the file. It will be auto-generated if not provided.
|
|
@@ -2909,7 +2991,7 @@ export interface EmotionalStrengths {
|
|
|
2909
2991
|
surprised?: number;
|
|
2910
2992
|
}
|
|
2911
2993
|
|
|
2912
|
-
export interface
|
|
2994
|
+
export interface Embedding_2 {
|
|
2913
2995
|
/**
|
|
2914
2996
|
* Path
|
|
2915
2997
|
* @description URL or the path to the embedding weights.
|
|
@@ -2926,6 +3008,31 @@ export interface Embedding_1 {
|
|
|
2926
3008
|
tokens?: string[];
|
|
2927
3009
|
}
|
|
2928
3010
|
|
|
3011
|
+
export interface Embedding_1 {
|
|
3012
|
+
/**
|
|
3013
|
+
* Force
|
|
3014
|
+
* @description If set to true, the embedding will be forced to be used.
|
|
3015
|
+
* @default false
|
|
3016
|
+
*/
|
|
3017
|
+
force?: boolean;
|
|
3018
|
+
/**
|
|
3019
|
+
* Path
|
|
3020
|
+
* @description URL or the path to the embedding weights.
|
|
3021
|
+
* @example https://civitai.com/api/download/models/135931
|
|
3022
|
+
* @example https://filebin.net/3chfqasxpqu21y8n/my-custom-lora-v1.safetensors
|
|
3023
|
+
*/
|
|
3024
|
+
path: string;
|
|
3025
|
+
/**
|
|
3026
|
+
* Tokens
|
|
3027
|
+
* @description The list of tokens to use for the embedding.
|
|
3028
|
+
* @default [
|
|
3029
|
+
* "<s0>",
|
|
3030
|
+
* "<s1>"
|
|
3031
|
+
* ]
|
|
3032
|
+
*/
|
|
3033
|
+
tokens?: string[];
|
|
3034
|
+
}
|
|
3035
|
+
|
|
2929
3036
|
export interface Embedding {
|
|
2930
3037
|
/**
|
|
2931
3038
|
* Path
|
|
@@ -3025,7 +3132,7 @@ export interface DiarizationSegment {
|
|
|
3025
3132
|
* Timestamp
|
|
3026
3133
|
* @description Start and end timestamp of the segment
|
|
3027
3134
|
*/
|
|
3028
|
-
timestamp:
|
|
3135
|
+
timestamp: [number, number];
|
|
3029
3136
|
}
|
|
3030
3137
|
|
|
3031
3138
|
export interface DialogueBlock {
|
|
@@ -3745,12 +3852,6 @@ export interface AudioFile {
|
|
|
3745
3852
|
* @description The duration of the audio
|
|
3746
3853
|
*/
|
|
3747
3854
|
duration?: number;
|
|
3748
|
-
/**
|
|
3749
|
-
* File Data
|
|
3750
|
-
* Format: binary
|
|
3751
|
-
* @description File data
|
|
3752
|
-
*/
|
|
3753
|
-
file_data?: string;
|
|
3754
3855
|
/**
|
|
3755
3856
|
* File Name
|
|
3756
3857
|
* @description The name of the file. It will be auto-generated if not provided.
|
|
@@ -624,6 +624,11 @@ declare global {
|
|
|
624
624
|
output: { [x in string]: any };
|
|
625
625
|
};
|
|
626
626
|
|
|
627
|
+
'fal-ai/wan-motion': {
|
|
628
|
+
input: falEndpoints.WanMotionInput;
|
|
629
|
+
output: falEndpoints.WanMotionOutput;
|
|
630
|
+
};
|
|
631
|
+
|
|
627
632
|
'fal-ai/wan-i2v-lora': {
|
|
628
633
|
input: falEndpoints.WanI2vLoraInput;
|
|
629
634
|
output: falEndpoints.WanI2vLoraOutput;
|
|
@@ -704,6 +709,16 @@ declare global {
|
|
|
704
709
|
output: falEndpoints.Wan22VaceFunA14bDepthOutput;
|
|
705
710
|
};
|
|
706
711
|
|
|
712
|
+
'fal-ai/wan-22-trainer/t2v-a14b': {
|
|
713
|
+
input: falEndpoints.Wan22TrainerT2vA14bInput;
|
|
714
|
+
output: falEndpoints.Wan22TrainerT2vA14bOutput;
|
|
715
|
+
};
|
|
716
|
+
|
|
717
|
+
'fal-ai/wan-22-trainer/i2v-a14b': {
|
|
718
|
+
input: falEndpoints.Wan22TrainerI2vA14bInput;
|
|
719
|
+
output: falEndpoints.Wan22TrainerI2vA14bOutput;
|
|
720
|
+
};
|
|
721
|
+
|
|
707
722
|
'fal-ai/wan-22-image-trainer': {
|
|
708
723
|
input: falEndpoints.Wan22ImageTrainerInput;
|
|
709
724
|
output: falEndpoints.Wan22ImageTrainerOutput;
|
|
@@ -2169,6 +2184,11 @@ declare global {
|
|
|
2169
2184
|
output: falEndpoints.MusetalkOutput;
|
|
2170
2185
|
};
|
|
2171
2186
|
|
|
2187
|
+
'fal-ai/multishot-master': {
|
|
2188
|
+
input: falEndpoints.MultishotMasterInput;
|
|
2189
|
+
output: falEndpoints.MultishotMasterOutput;
|
|
2190
|
+
};
|
|
2191
|
+
|
|
2172
2192
|
'fal-ai/moondream3-preview/segment': {
|
|
2173
2193
|
input: { [x in string]: any };
|
|
2174
2194
|
output: { [x in string]: any };
|
|
@@ -2910,8 +2930,8 @@ declare global {
|
|
|
2910
2930
|
};
|
|
2911
2931
|
|
|
2912
2932
|
'fal-ai/live-portrait/image': {
|
|
2913
|
-
input:
|
|
2914
|
-
output:
|
|
2933
|
+
input: falEndpoints.LivePortraitImageInput;
|
|
2934
|
+
output: falEndpoints.LivePortraitImageOutput;
|
|
2915
2935
|
};
|
|
2916
2936
|
|
|
2917
2937
|
'fal-ai/live-portrait': {
|
|
@@ -2964,6 +2984,11 @@ declare global {
|
|
|
2964
2984
|
output: falEndpoints.LayerDiffusionOutput;
|
|
2965
2985
|
};
|
|
2966
2986
|
|
|
2987
|
+
'fal-ai/lava-sr': {
|
|
2988
|
+
input: falEndpoints.LavaSrInput;
|
|
2989
|
+
output: falEndpoints.LavaSrOutput;
|
|
2990
|
+
};
|
|
2991
|
+
|
|
2967
2992
|
'fal-ai/latentsync': {
|
|
2968
2993
|
input: falEndpoints.LatentsyncInput;
|
|
2969
2994
|
output: falEndpoints.LatentsyncOutput;
|
|
@@ -5244,6 +5269,26 @@ declare global {
|
|
|
5244
5269
|
output: falEndpoints.CreatifyAuroraOutput;
|
|
5245
5270
|
};
|
|
5246
5271
|
|
|
5272
|
+
'fal-ai/cosmos-predict-2.5/video-to-video': {
|
|
5273
|
+
input: falEndpoints.CosmosPredict25VideoToVideoInput;
|
|
5274
|
+
output: falEndpoints.CosmosPredict25VideoToVideoOutput;
|
|
5275
|
+
};
|
|
5276
|
+
|
|
5277
|
+
'fal-ai/cosmos-predict-2.5/text-to-video': {
|
|
5278
|
+
input: falEndpoints.CosmosPredict25TextToVideoInput;
|
|
5279
|
+
output: falEndpoints.CosmosPredict25TextToVideoOutput;
|
|
5280
|
+
};
|
|
5281
|
+
|
|
5282
|
+
'fal-ai/cosmos-predict-2.5/image-to-video': {
|
|
5283
|
+
input: falEndpoints.CosmosPredict25ImageToVideoInput;
|
|
5284
|
+
output: falEndpoints.CosmosPredict25ImageToVideoOutput;
|
|
5285
|
+
};
|
|
5286
|
+
|
|
5287
|
+
'fal-ai/cosmos-predict-2.5/distilled/text-to-video': {
|
|
5288
|
+
input: falEndpoints.CosmosPredict25DistilledTextToVideoInput;
|
|
5289
|
+
output: falEndpoints.CosmosPredict25DistilledTextToVideoOutput;
|
|
5290
|
+
};
|
|
5291
|
+
|
|
5247
5292
|
'fal-ai/controlnext': {
|
|
5248
5293
|
input: falEndpoints.ControlnextInput;
|
|
5249
5294
|
output: falEndpoints.ControlnextOutput;
|
|
@@ -5344,6 +5389,16 @@ declare global {
|
|
|
5344
5389
|
output: falEndpoints.BytedanceVideoStylizeOutput;
|
|
5345
5390
|
};
|
|
5346
5391
|
|
|
5392
|
+
'fal-ai/bytedance/seedream/v5/lite/text-to-image': {
|
|
5393
|
+
input: falEndpoints.BytedanceSeedreamV5LiteTextToImageInput;
|
|
5394
|
+
output: falEndpoints.BytedanceSeedreamV5LiteTextToImageOutput;
|
|
5395
|
+
};
|
|
5396
|
+
|
|
5397
|
+
'fal-ai/bytedance/seedream/v5/lite/edit': {
|
|
5398
|
+
input: falEndpoints.BytedanceSeedreamV5LiteEditInput;
|
|
5399
|
+
output: falEndpoints.BytedanceSeedreamV5LiteEditOutput;
|
|
5400
|
+
};
|
|
5401
|
+
|
|
5347
5402
|
'fal-ai/bytedance/seedream/v4/text-to-image': {
|
|
5348
5403
|
input: falEndpoints.BytedanceSeedreamV4TextToImageInput;
|
|
5349
5404
|
output: falEndpoints.BytedanceSeedreamV4TextToImageOutput;
|
|
@@ -5489,6 +5544,11 @@ declare global {
|
|
|
5489
5544
|
output: falEndpoints.BriaBackgroundRemoveOutput;
|
|
5490
5545
|
};
|
|
5491
5546
|
|
|
5547
|
+
'fal-ai/bitdance': {
|
|
5548
|
+
input: falEndpoints.BitdanceInput;
|
|
5549
|
+
output: falEndpoints.BitdanceOutput;
|
|
5550
|
+
};
|
|
5551
|
+
|
|
5492
5552
|
'fal-ai/birefnet/v2/video': {
|
|
5493
5553
|
input: falEndpoints.BirefnetV2VideoInput;
|
|
5494
5554
|
output: falEndpoints.BirefnetV2VideoOutput;
|
|
@@ -5629,6 +5689,11 @@ declare global {
|
|
|
5629
5689
|
output: falEndpoints.DecartLucyRestyleOutput;
|
|
5630
5690
|
};
|
|
5631
5691
|
|
|
5692
|
+
'decart/lucy-i2v': {
|
|
5693
|
+
input: falEndpoints.DecartLucyI2vInput;
|
|
5694
|
+
output: falEndpoints.DecartLucyI2vOutput;
|
|
5695
|
+
};
|
|
5696
|
+
|
|
5632
5697
|
'decart/lucy-edit/pro': {
|
|
5633
5698
|
input: falEndpoints.DecartLucyEditProInput;
|
|
5634
5699
|
output: falEndpoints.DecartLucyEditProOutput;
|
|
@@ -5644,11 +5709,6 @@ declare global {
|
|
|
5644
5709
|
output: falEndpoints.DecartLucyEditDevOutput;
|
|
5645
5710
|
};
|
|
5646
5711
|
|
|
5647
|
-
'decart/lucy-14b/image-to-video': {
|
|
5648
|
-
input: falEndpoints.DecartLucy14bImageToVideoInput;
|
|
5649
|
-
output: falEndpoints.DecartLucy14bImageToVideoOutput;
|
|
5650
|
-
};
|
|
5651
|
-
|
|
5652
5712
|
'clarityai/crystal-video-upscaler': {
|
|
5653
5713
|
input: falEndpoints.ClarityaiCrystalVideoUpscalerInput;
|
|
5654
5714
|
output: falEndpoints.ClarityaiCrystalVideoUpscalerOutput;
|
|
@@ -5704,6 +5764,11 @@ declare global {
|
|
|
5704
5764
|
output: falEndpoints.BriaVideoBackgroundRemovalOutput;
|
|
5705
5765
|
};
|
|
5706
5766
|
|
|
5767
|
+
'bria/upscale/creative': {
|
|
5768
|
+
input: falEndpoints.BriaUpscaleCreativeInput;
|
|
5769
|
+
output: falEndpoints.BriaUpscaleCreativeOutput;
|
|
5770
|
+
};
|
|
5771
|
+
|
|
5707
5772
|
'bria/text-to-image/3.2': {
|
|
5708
5773
|
input: falEndpoints.BriaTextToImage32Input;
|
|
5709
5774
|
output: falEndpoints.BriaTextToImage32Output;
|