fal-endpoint-types 1.3.25 → 1.3.26
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/types/fal/endpoints/components.d.ts +67 -21
- package/types/fal/endpoints/index.d.ts +65 -15
- package/types/fal/endpoints/schema.d.ts +3304 -1558
package/package.json
CHANGED
|
@@ -26,7 +26,7 @@ export interface WhisperChunk {
|
|
|
26
26
|
* Timestamp
|
|
27
27
|
* @description Start and end timestamp of the chunk
|
|
28
28
|
*/
|
|
29
|
-
timestamp:
|
|
29
|
+
timestamp: [number, number];
|
|
30
30
|
}
|
|
31
31
|
|
|
32
32
|
export interface VoiceSetting {
|
|
@@ -1067,6 +1067,21 @@ export interface SigmasInput {
|
|
|
1067
1067
|
method?: 'default' | 'array';
|
|
1068
1068
|
}
|
|
1069
1069
|
|
|
1070
|
+
export interface Shot {
|
|
1071
|
+
/**
|
|
1072
|
+
* Num Frames
|
|
1073
|
+
* @description Number of frames for this shot. Each shot must have between 17 and 100 frames.
|
|
1074
|
+
* @default 65
|
|
1075
|
+
*/
|
|
1076
|
+
num_frames?: number;
|
|
1077
|
+
/**
|
|
1078
|
+
* Prompt
|
|
1079
|
+
* @description Caption describing what happens in this shot, including subject actions, camera angles, and scene details.
|
|
1080
|
+
* @example Subject 1 is smiling and gesturing with his right hand while facing the camera. Subject 2 is seen from behind, facing Subject 1. The scene is indoors in a well-decorated room, with a dark door and ornate glass visible in the background. The camera is static, providing a medium close-up shot of Subject 1.
|
|
1081
|
+
*/
|
|
1082
|
+
prompt: string;
|
|
1083
|
+
}
|
|
1084
|
+
|
|
1070
1085
|
export interface SemanticImageInput {
|
|
1071
1086
|
/**
|
|
1072
1087
|
* Hypothesis
|
|
@@ -1701,24 +1716,6 @@ export interface Object {
|
|
|
1701
1716
|
y_min: number;
|
|
1702
1717
|
}
|
|
1703
1718
|
|
|
1704
|
-
export interface NovaSRTimings {
|
|
1705
|
-
/**
|
|
1706
|
-
* Inference
|
|
1707
|
-
* @description Time taken to run the inference in seconds.
|
|
1708
|
-
*/
|
|
1709
|
-
inference: number;
|
|
1710
|
-
/**
|
|
1711
|
-
* Postprocess
|
|
1712
|
-
* @description Time taken to postprocess the audio in seconds.
|
|
1713
|
-
*/
|
|
1714
|
-
postprocess: number;
|
|
1715
|
-
/**
|
|
1716
|
-
* Preprocess
|
|
1717
|
-
* @description Time taken to preprocess the audio in seconds.
|
|
1718
|
-
*/
|
|
1719
|
-
preprocess: number;
|
|
1720
|
-
}
|
|
1721
|
-
|
|
1722
1719
|
export interface MusicSection {
|
|
1723
1720
|
/**
|
|
1724
1721
|
* Duration Ms
|
|
@@ -2203,6 +2200,24 @@ export interface Lighting {
|
|
|
2203
2200
|
shadows?: string;
|
|
2204
2201
|
}
|
|
2205
2202
|
|
|
2203
|
+
export interface LavaSRTimings {
|
|
2204
|
+
/**
|
|
2205
|
+
* Inference
|
|
2206
|
+
* @description Time taken to run the inference in seconds.
|
|
2207
|
+
*/
|
|
2208
|
+
inference: number;
|
|
2209
|
+
/**
|
|
2210
|
+
* Postprocess
|
|
2211
|
+
* @description Time taken to postprocess the audio in seconds.
|
|
2212
|
+
*/
|
|
2213
|
+
postprocess: number;
|
|
2214
|
+
/**
|
|
2215
|
+
* Preprocess
|
|
2216
|
+
* @description Time taken to preprocess the audio in seconds.
|
|
2217
|
+
*/
|
|
2218
|
+
preprocess: number;
|
|
2219
|
+
}
|
|
2220
|
+
|
|
2206
2221
|
export interface KlingV3MultiPromptElement {
|
|
2207
2222
|
/**
|
|
2208
2223
|
* Duration
|
|
@@ -2222,6 +2237,8 @@ export interface KlingV3ImageElementInput {
|
|
|
2222
2237
|
/**
|
|
2223
2238
|
* Frontal Image Url
|
|
2224
2239
|
* @description The frontal image of the element (main view).
|
|
2240
|
+
*
|
|
2241
|
+
* Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
|
|
2225
2242
|
*/
|
|
2226
2243
|
frontal_image_url?: string;
|
|
2227
2244
|
/**
|
|
@@ -2235,6 +2252,8 @@ export interface KlingV3ComboElementInput {
|
|
|
2235
2252
|
/**
|
|
2236
2253
|
* Frontal Image Url
|
|
2237
2254
|
* @description The frontal image of the element (main view).
|
|
2255
|
+
*
|
|
2256
|
+
* Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
|
|
2238
2257
|
*/
|
|
2239
2258
|
frontal_image_url?: string;
|
|
2240
2259
|
/**
|
|
@@ -2245,6 +2264,8 @@ export interface KlingV3ComboElementInput {
|
|
|
2245
2264
|
/**
|
|
2246
2265
|
* Video Url
|
|
2247
2266
|
* @description The video URL of the element. A request can only have one element with a video.
|
|
2267
|
+
*
|
|
2268
|
+
* Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
|
|
2248
2269
|
*/
|
|
2249
2270
|
video_url?: string;
|
|
2250
2271
|
}
|
|
@@ -2970,7 +2991,7 @@ export interface EmotionalStrengths {
|
|
|
2970
2991
|
surprised?: number;
|
|
2971
2992
|
}
|
|
2972
2993
|
|
|
2973
|
-
export interface
|
|
2994
|
+
export interface Embedding_2 {
|
|
2974
2995
|
/**
|
|
2975
2996
|
* Path
|
|
2976
2997
|
* @description URL or the path to the embedding weights.
|
|
@@ -2987,6 +3008,31 @@ export interface Embedding_1 {
|
|
|
2987
3008
|
tokens?: string[];
|
|
2988
3009
|
}
|
|
2989
3010
|
|
|
3011
|
+
export interface Embedding_1 {
|
|
3012
|
+
/**
|
|
3013
|
+
* Force
|
|
3014
|
+
* @description If set to true, the embedding will be forced to be used.
|
|
3015
|
+
* @default false
|
|
3016
|
+
*/
|
|
3017
|
+
force?: boolean;
|
|
3018
|
+
/**
|
|
3019
|
+
* Path
|
|
3020
|
+
* @description URL or the path to the embedding weights.
|
|
3021
|
+
* @example https://civitai.com/api/download/models/135931
|
|
3022
|
+
* @example https://filebin.net/3chfqasxpqu21y8n/my-custom-lora-v1.safetensors
|
|
3023
|
+
*/
|
|
3024
|
+
path: string;
|
|
3025
|
+
/**
|
|
3026
|
+
* Tokens
|
|
3027
|
+
* @description The list of tokens to use for the embedding.
|
|
3028
|
+
* @default [
|
|
3029
|
+
* "<s0>",
|
|
3030
|
+
* "<s1>"
|
|
3031
|
+
* ]
|
|
3032
|
+
*/
|
|
3033
|
+
tokens?: string[];
|
|
3034
|
+
}
|
|
3035
|
+
|
|
2990
3036
|
export interface Embedding {
|
|
2991
3037
|
/**
|
|
2992
3038
|
* Path
|
|
@@ -3086,7 +3132,7 @@ export interface DiarizationSegment {
|
|
|
3086
3132
|
* Timestamp
|
|
3087
3133
|
* @description Start and end timestamp of the segment
|
|
3088
3134
|
*/
|
|
3089
|
-
timestamp:
|
|
3135
|
+
timestamp: [number, number];
|
|
3090
3136
|
}
|
|
3091
3137
|
|
|
3092
3138
|
export interface DialogueBlock {
|
|
@@ -624,6 +624,11 @@ declare global {
|
|
|
624
624
|
output: { [x in string]: any };
|
|
625
625
|
};
|
|
626
626
|
|
|
627
|
+
'fal-ai/wan-motion': {
|
|
628
|
+
input: falEndpoints.WanMotionInput;
|
|
629
|
+
output: falEndpoints.WanMotionOutput;
|
|
630
|
+
};
|
|
631
|
+
|
|
627
632
|
'fal-ai/wan-i2v-lora': {
|
|
628
633
|
input: falEndpoints.WanI2vLoraInput;
|
|
629
634
|
output: falEndpoints.WanI2vLoraOutput;
|
|
@@ -704,6 +709,16 @@ declare global {
|
|
|
704
709
|
output: falEndpoints.Wan22VaceFunA14bDepthOutput;
|
|
705
710
|
};
|
|
706
711
|
|
|
712
|
+
'fal-ai/wan-22-trainer/t2v-a14b': {
|
|
713
|
+
input: falEndpoints.Wan22TrainerT2vA14bInput;
|
|
714
|
+
output: falEndpoints.Wan22TrainerT2vA14bOutput;
|
|
715
|
+
};
|
|
716
|
+
|
|
717
|
+
'fal-ai/wan-22-trainer/i2v-a14b': {
|
|
718
|
+
input: falEndpoints.Wan22TrainerI2vA14bInput;
|
|
719
|
+
output: falEndpoints.Wan22TrainerI2vA14bOutput;
|
|
720
|
+
};
|
|
721
|
+
|
|
707
722
|
'fal-ai/wan-22-image-trainer': {
|
|
708
723
|
input: falEndpoints.Wan22ImageTrainerInput;
|
|
709
724
|
output: falEndpoints.Wan22ImageTrainerOutput;
|
|
@@ -2039,11 +2054,6 @@ declare global {
|
|
|
2039
2054
|
output: falEndpoints.PhotomakerOutput;
|
|
2040
2055
|
};
|
|
2041
2056
|
|
|
2042
|
-
'fal-ai/personaplex/realtime': {
|
|
2043
|
-
input: falEndpoints.PersonaplexRealtimeInput;
|
|
2044
|
-
output: falEndpoints.PersonaplexRealtimeOutput;
|
|
2045
|
-
};
|
|
2046
|
-
|
|
2047
2057
|
'fal-ai/personaplex': {
|
|
2048
2058
|
input: falEndpoints.PersonaplexInput;
|
|
2049
2059
|
output: falEndpoints.PersonaplexOutput;
|
|
@@ -2174,6 +2184,11 @@ declare global {
|
|
|
2174
2184
|
output: falEndpoints.MusetalkOutput;
|
|
2175
2185
|
};
|
|
2176
2186
|
|
|
2187
|
+
'fal-ai/multishot-master': {
|
|
2188
|
+
input: falEndpoints.MultishotMasterInput;
|
|
2189
|
+
output: falEndpoints.MultishotMasterOutput;
|
|
2190
|
+
};
|
|
2191
|
+
|
|
2177
2192
|
'fal-ai/moondream3-preview/segment': {
|
|
2178
2193
|
input: { [x in string]: any };
|
|
2179
2194
|
output: { [x in string]: any };
|
|
@@ -2969,6 +2984,11 @@ declare global {
|
|
|
2969
2984
|
output: falEndpoints.LayerDiffusionOutput;
|
|
2970
2985
|
};
|
|
2971
2986
|
|
|
2987
|
+
'fal-ai/lava-sr': {
|
|
2988
|
+
input: falEndpoints.LavaSrInput;
|
|
2989
|
+
output: falEndpoints.LavaSrOutput;
|
|
2990
|
+
};
|
|
2991
|
+
|
|
2972
2992
|
'fal-ai/latentsync': {
|
|
2973
2993
|
input: falEndpoints.LatentsyncInput;
|
|
2974
2994
|
output: falEndpoints.LatentsyncOutput;
|
|
@@ -5249,6 +5269,26 @@ declare global {
|
|
|
5249
5269
|
output: falEndpoints.CreatifyAuroraOutput;
|
|
5250
5270
|
};
|
|
5251
5271
|
|
|
5272
|
+
'fal-ai/cosmos-predict-2.5/video-to-video': {
|
|
5273
|
+
input: falEndpoints.CosmosPredict25VideoToVideoInput;
|
|
5274
|
+
output: falEndpoints.CosmosPredict25VideoToVideoOutput;
|
|
5275
|
+
};
|
|
5276
|
+
|
|
5277
|
+
'fal-ai/cosmos-predict-2.5/text-to-video': {
|
|
5278
|
+
input: falEndpoints.CosmosPredict25TextToVideoInput;
|
|
5279
|
+
output: falEndpoints.CosmosPredict25TextToVideoOutput;
|
|
5280
|
+
};
|
|
5281
|
+
|
|
5282
|
+
'fal-ai/cosmos-predict-2.5/image-to-video': {
|
|
5283
|
+
input: falEndpoints.CosmosPredict25ImageToVideoInput;
|
|
5284
|
+
output: falEndpoints.CosmosPredict25ImageToVideoOutput;
|
|
5285
|
+
};
|
|
5286
|
+
|
|
5287
|
+
'fal-ai/cosmos-predict-2.5/distilled/text-to-video': {
|
|
5288
|
+
input: falEndpoints.CosmosPredict25DistilledTextToVideoInput;
|
|
5289
|
+
output: falEndpoints.CosmosPredict25DistilledTextToVideoOutput;
|
|
5290
|
+
};
|
|
5291
|
+
|
|
5252
5292
|
'fal-ai/controlnext': {
|
|
5253
5293
|
input: falEndpoints.ControlnextInput;
|
|
5254
5294
|
output: falEndpoints.ControlnextOutput;
|
|
@@ -5349,6 +5389,16 @@ declare global {
|
|
|
5349
5389
|
output: falEndpoints.BytedanceVideoStylizeOutput;
|
|
5350
5390
|
};
|
|
5351
5391
|
|
|
5392
|
+
'fal-ai/bytedance/seedream/v5/lite/text-to-image': {
|
|
5393
|
+
input: falEndpoints.BytedanceSeedreamV5LiteTextToImageInput;
|
|
5394
|
+
output: falEndpoints.BytedanceSeedreamV5LiteTextToImageOutput;
|
|
5395
|
+
};
|
|
5396
|
+
|
|
5397
|
+
'fal-ai/bytedance/seedream/v5/lite/edit': {
|
|
5398
|
+
input: falEndpoints.BytedanceSeedreamV5LiteEditInput;
|
|
5399
|
+
output: falEndpoints.BytedanceSeedreamV5LiteEditOutput;
|
|
5400
|
+
};
|
|
5401
|
+
|
|
5352
5402
|
'fal-ai/bytedance/seedream/v4/text-to-image': {
|
|
5353
5403
|
input: falEndpoints.BytedanceSeedreamV4TextToImageInput;
|
|
5354
5404
|
output: falEndpoints.BytedanceSeedreamV4TextToImageOutput;
|
|
@@ -5639,6 +5689,11 @@ declare global {
|
|
|
5639
5689
|
output: falEndpoints.DecartLucyRestyleOutput;
|
|
5640
5690
|
};
|
|
5641
5691
|
|
|
5692
|
+
'decart/lucy-i2v': {
|
|
5693
|
+
input: falEndpoints.DecartLucyI2vInput;
|
|
5694
|
+
output: falEndpoints.DecartLucyI2vOutput;
|
|
5695
|
+
};
|
|
5696
|
+
|
|
5642
5697
|
'decart/lucy-edit/pro': {
|
|
5643
5698
|
input: falEndpoints.DecartLucyEditProInput;
|
|
5644
5699
|
output: falEndpoints.DecartLucyEditProOutput;
|
|
@@ -5654,11 +5709,6 @@ declare global {
|
|
|
5654
5709
|
output: falEndpoints.DecartLucyEditDevOutput;
|
|
5655
5710
|
};
|
|
5656
5711
|
|
|
5657
|
-
'decart/lucy-14b/image-to-video': {
|
|
5658
|
-
input: falEndpoints.DecartLucy14bImageToVideoInput;
|
|
5659
|
-
output: falEndpoints.DecartLucy14bImageToVideoOutput;
|
|
5660
|
-
};
|
|
5661
|
-
|
|
5662
5712
|
'clarityai/crystal-video-upscaler': {
|
|
5663
5713
|
input: falEndpoints.ClarityaiCrystalVideoUpscalerInput;
|
|
5664
5714
|
output: falEndpoints.ClarityaiCrystalVideoUpscalerOutput;
|
|
@@ -5714,6 +5764,11 @@ declare global {
|
|
|
5714
5764
|
output: falEndpoints.BriaVideoBackgroundRemovalOutput;
|
|
5715
5765
|
};
|
|
5716
5766
|
|
|
5767
|
+
'bria/upscale/creative': {
|
|
5768
|
+
input: falEndpoints.BriaUpscaleCreativeInput;
|
|
5769
|
+
output: falEndpoints.BriaUpscaleCreativeOutput;
|
|
5770
|
+
};
|
|
5771
|
+
|
|
5717
5772
|
'bria/text-to-image/3.2': {
|
|
5718
5773
|
input: falEndpoints.BriaTextToImage32Input;
|
|
5719
5774
|
output: falEndpoints.BriaTextToImage32Output;
|
|
@@ -5834,11 +5889,6 @@ declare global {
|
|
|
5834
5889
|
output: falEndpoints.BriaBria_video_eraserEraseKeypointsOutput;
|
|
5835
5890
|
};
|
|
5836
5891
|
|
|
5837
|
-
'bria/aesthetics/upscaler': {
|
|
5838
|
-
input: { [x in string]: any };
|
|
5839
|
-
output: { [x in string]: any };
|
|
5840
|
-
};
|
|
5841
|
-
|
|
5842
5892
|
'beatoven/sound-effect-generation': {
|
|
5843
5893
|
input: falEndpoints.BeatovenSoundEffectGenerationInput;
|
|
5844
5894
|
output: falEndpoints.BeatovenSoundEffectGenerationOutput;
|