fal-endpoint-types 1.3.25 → 1.3.26

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "fal-endpoint-types",
3
- "version": "1.3.25",
3
+ "version": "1.3.26",
4
4
  "description": "TypeScript types for Fal AI endpoints generated from the OpenAPI schemas.",
5
5
  "homepage": "https://github.com/rawpixel-vincent/fal-endpoint-types#readme",
6
6
  "bugs": {
@@ -26,7 +26,7 @@ export interface WhisperChunk {
26
26
  * Timestamp
27
27
  * @description Start and end timestamp of the chunk
28
28
  */
29
- timestamp: { [x: string]: any }[];
29
+ timestamp: [number, number];
30
30
  }
31
31
 
32
32
  export interface VoiceSetting {
@@ -1067,6 +1067,21 @@ export interface SigmasInput {
1067
1067
  method?: 'default' | 'array';
1068
1068
  }
1069
1069
 
1070
+ export interface Shot {
1071
+ /**
1072
+ * Num Frames
1073
+ * @description Number of frames for this shot. Each shot must have between 17 and 100 frames.
1074
+ * @default 65
1075
+ */
1076
+ num_frames?: number;
1077
+ /**
1078
+ * Prompt
1079
+ * @description Caption describing what happens in this shot, including subject actions, camera angles, and scene details.
1080
+ * @example Subject 1 is smiling and gesturing with his right hand while facing the camera. Subject 2 is seen from behind, facing Subject 1. The scene is indoors in a well-decorated room, with a dark door and ornate glass visible in the background. The camera is static, providing a medium close-up shot of Subject 1.
1081
+ */
1082
+ prompt: string;
1083
+ }
1084
+
1070
1085
  export interface SemanticImageInput {
1071
1086
  /**
1072
1087
  * Hypothesis
@@ -1701,24 +1716,6 @@ export interface Object {
1701
1716
  y_min: number;
1702
1717
  }
1703
1718
 
1704
- export interface NovaSRTimings {
1705
- /**
1706
- * Inference
1707
- * @description Time taken to run the inference in seconds.
1708
- */
1709
- inference: number;
1710
- /**
1711
- * Postprocess
1712
- * @description Time taken to postprocess the audio in seconds.
1713
- */
1714
- postprocess: number;
1715
- /**
1716
- * Preprocess
1717
- * @description Time taken to preprocess the audio in seconds.
1718
- */
1719
- preprocess: number;
1720
- }
1721
-
1722
1719
  export interface MusicSection {
1723
1720
  /**
1724
1721
  * Duration Ms
@@ -2203,6 +2200,24 @@ export interface Lighting {
2203
2200
  shadows?: string;
2204
2201
  }
2205
2202
 
2203
+ export interface LavaSRTimings {
2204
+ /**
2205
+ * Inference
2206
+ * @description Time taken to run the inference in seconds.
2207
+ */
2208
+ inference: number;
2209
+ /**
2210
+ * Postprocess
2211
+ * @description Time taken to postprocess the audio in seconds.
2212
+ */
2213
+ postprocess: number;
2214
+ /**
2215
+ * Preprocess
2216
+ * @description Time taken to preprocess the audio in seconds.
2217
+ */
2218
+ preprocess: number;
2219
+ }
2220
+
2206
2221
  export interface KlingV3MultiPromptElement {
2207
2222
  /**
2208
2223
  * Duration
@@ -2222,6 +2237,8 @@ export interface KlingV3ImageElementInput {
2222
2237
  /**
2223
2238
  * Frontal Image Url
2224
2239
  * @description The frontal image of the element (main view).
2240
+ *
2241
+ * Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
2225
2242
  */
2226
2243
  frontal_image_url?: string;
2227
2244
  /**
@@ -2235,6 +2252,8 @@ export interface KlingV3ComboElementInput {
2235
2252
  /**
2236
2253
  * Frontal Image Url
2237
2254
  * @description The frontal image of the element (main view).
2255
+ *
2256
+ * Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
2238
2257
  */
2239
2258
  frontal_image_url?: string;
2240
2259
  /**
@@ -2245,6 +2264,8 @@ export interface KlingV3ComboElementInput {
2245
2264
  /**
2246
2265
  * Video Url
2247
2266
  * @description The video URL of the element. A request can only have one element with a video.
2267
+ *
2268
+ * Max file size: 200.0MB, Min width: 720px, Min height: 720px, Max width: 2160px, Max height: 2160px, Min duration: 3.0s, Max duration: 10.05s, Min FPS: 24.0, Max FPS: 60.0, Timeout: 30.0s
2248
2269
  */
2249
2270
  video_url?: string;
2250
2271
  }
@@ -2970,7 +2991,7 @@ export interface EmotionalStrengths {
2970
2991
  surprised?: number;
2971
2992
  }
2972
2993
 
2973
- export interface Embedding_1 {
2994
+ export interface Embedding_2 {
2974
2995
  /**
2975
2996
  * Path
2976
2997
  * @description URL or the path to the embedding weights.
@@ -2987,6 +3008,31 @@ export interface Embedding_1 {
2987
3008
  tokens?: string[];
2988
3009
  }
2989
3010
 
3011
+ export interface Embedding_1 {
3012
+ /**
3013
+ * Force
3014
+ * @description If set to true, the embedding will be forced to be used.
3015
+ * @default false
3016
+ */
3017
+ force?: boolean;
3018
+ /**
3019
+ * Path
3020
+ * @description URL or the path to the embedding weights.
3021
+ * @example https://civitai.com/api/download/models/135931
3022
+ * @example https://filebin.net/3chfqasxpqu21y8n/my-custom-lora-v1.safetensors
3023
+ */
3024
+ path: string;
3025
+ /**
3026
+ * Tokens
3027
+ * @description The list of tokens to use for the embedding.
3028
+ * @default [
3029
+ * "<s0>",
3030
+ * "<s1>"
3031
+ * ]
3032
+ */
3033
+ tokens?: string[];
3034
+ }
3035
+
2990
3036
  export interface Embedding {
2991
3037
  /**
2992
3038
  * Path
@@ -3086,7 +3132,7 @@ export interface DiarizationSegment {
3086
3132
  * Timestamp
3087
3133
  * @description Start and end timestamp of the segment
3088
3134
  */
3089
- timestamp: { [x: string]: any }[];
3135
+ timestamp: [number, number];
3090
3136
  }
3091
3137
 
3092
3138
  export interface DialogueBlock {
@@ -624,6 +624,11 @@ declare global {
624
624
  output: { [x in string]: any };
625
625
  };
626
626
 
627
+ 'fal-ai/wan-motion': {
628
+ input: falEndpoints.WanMotionInput;
629
+ output: falEndpoints.WanMotionOutput;
630
+ };
631
+
627
632
  'fal-ai/wan-i2v-lora': {
628
633
  input: falEndpoints.WanI2vLoraInput;
629
634
  output: falEndpoints.WanI2vLoraOutput;
@@ -704,6 +709,16 @@ declare global {
704
709
  output: falEndpoints.Wan22VaceFunA14bDepthOutput;
705
710
  };
706
711
 
712
+ 'fal-ai/wan-22-trainer/t2v-a14b': {
713
+ input: falEndpoints.Wan22TrainerT2vA14bInput;
714
+ output: falEndpoints.Wan22TrainerT2vA14bOutput;
715
+ };
716
+
717
+ 'fal-ai/wan-22-trainer/i2v-a14b': {
718
+ input: falEndpoints.Wan22TrainerI2vA14bInput;
719
+ output: falEndpoints.Wan22TrainerI2vA14bOutput;
720
+ };
721
+
707
722
  'fal-ai/wan-22-image-trainer': {
708
723
  input: falEndpoints.Wan22ImageTrainerInput;
709
724
  output: falEndpoints.Wan22ImageTrainerOutput;
@@ -2039,11 +2054,6 @@ declare global {
2039
2054
  output: falEndpoints.PhotomakerOutput;
2040
2055
  };
2041
2056
 
2042
- 'fal-ai/personaplex/realtime': {
2043
- input: falEndpoints.PersonaplexRealtimeInput;
2044
- output: falEndpoints.PersonaplexRealtimeOutput;
2045
- };
2046
-
2047
2057
  'fal-ai/personaplex': {
2048
2058
  input: falEndpoints.PersonaplexInput;
2049
2059
  output: falEndpoints.PersonaplexOutput;
@@ -2174,6 +2184,11 @@ declare global {
2174
2184
  output: falEndpoints.MusetalkOutput;
2175
2185
  };
2176
2186
 
2187
+ 'fal-ai/multishot-master': {
2188
+ input: falEndpoints.MultishotMasterInput;
2189
+ output: falEndpoints.MultishotMasterOutput;
2190
+ };
2191
+
2177
2192
  'fal-ai/moondream3-preview/segment': {
2178
2193
  input: { [x in string]: any };
2179
2194
  output: { [x in string]: any };
@@ -2969,6 +2984,11 @@ declare global {
2969
2984
  output: falEndpoints.LayerDiffusionOutput;
2970
2985
  };
2971
2986
 
2987
+ 'fal-ai/lava-sr': {
2988
+ input: falEndpoints.LavaSrInput;
2989
+ output: falEndpoints.LavaSrOutput;
2990
+ };
2991
+
2972
2992
  'fal-ai/latentsync': {
2973
2993
  input: falEndpoints.LatentsyncInput;
2974
2994
  output: falEndpoints.LatentsyncOutput;
@@ -5249,6 +5269,26 @@ declare global {
5249
5269
  output: falEndpoints.CreatifyAuroraOutput;
5250
5270
  };
5251
5271
 
5272
+ 'fal-ai/cosmos-predict-2.5/video-to-video': {
5273
+ input: falEndpoints.CosmosPredict25VideoToVideoInput;
5274
+ output: falEndpoints.CosmosPredict25VideoToVideoOutput;
5275
+ };
5276
+
5277
+ 'fal-ai/cosmos-predict-2.5/text-to-video': {
5278
+ input: falEndpoints.CosmosPredict25TextToVideoInput;
5279
+ output: falEndpoints.CosmosPredict25TextToVideoOutput;
5280
+ };
5281
+
5282
+ 'fal-ai/cosmos-predict-2.5/image-to-video': {
5283
+ input: falEndpoints.CosmosPredict25ImageToVideoInput;
5284
+ output: falEndpoints.CosmosPredict25ImageToVideoOutput;
5285
+ };
5286
+
5287
+ 'fal-ai/cosmos-predict-2.5/distilled/text-to-video': {
5288
+ input: falEndpoints.CosmosPredict25DistilledTextToVideoInput;
5289
+ output: falEndpoints.CosmosPredict25DistilledTextToVideoOutput;
5290
+ };
5291
+
5252
5292
  'fal-ai/controlnext': {
5253
5293
  input: falEndpoints.ControlnextInput;
5254
5294
  output: falEndpoints.ControlnextOutput;
@@ -5349,6 +5389,16 @@ declare global {
5349
5389
  output: falEndpoints.BytedanceVideoStylizeOutput;
5350
5390
  };
5351
5391
 
5392
+ 'fal-ai/bytedance/seedream/v5/lite/text-to-image': {
5393
+ input: falEndpoints.BytedanceSeedreamV5LiteTextToImageInput;
5394
+ output: falEndpoints.BytedanceSeedreamV5LiteTextToImageOutput;
5395
+ };
5396
+
5397
+ 'fal-ai/bytedance/seedream/v5/lite/edit': {
5398
+ input: falEndpoints.BytedanceSeedreamV5LiteEditInput;
5399
+ output: falEndpoints.BytedanceSeedreamV5LiteEditOutput;
5400
+ };
5401
+
5352
5402
  'fal-ai/bytedance/seedream/v4/text-to-image': {
5353
5403
  input: falEndpoints.BytedanceSeedreamV4TextToImageInput;
5354
5404
  output: falEndpoints.BytedanceSeedreamV4TextToImageOutput;
@@ -5639,6 +5689,11 @@ declare global {
5639
5689
  output: falEndpoints.DecartLucyRestyleOutput;
5640
5690
  };
5641
5691
 
5692
+ 'decart/lucy-i2v': {
5693
+ input: falEndpoints.DecartLucyI2vInput;
5694
+ output: falEndpoints.DecartLucyI2vOutput;
5695
+ };
5696
+
5642
5697
  'decart/lucy-edit/pro': {
5643
5698
  input: falEndpoints.DecartLucyEditProInput;
5644
5699
  output: falEndpoints.DecartLucyEditProOutput;
@@ -5654,11 +5709,6 @@ declare global {
5654
5709
  output: falEndpoints.DecartLucyEditDevOutput;
5655
5710
  };
5656
5711
 
5657
- 'decart/lucy-14b/image-to-video': {
5658
- input: falEndpoints.DecartLucy14bImageToVideoInput;
5659
- output: falEndpoints.DecartLucy14bImageToVideoOutput;
5660
- };
5661
-
5662
5712
  'clarityai/crystal-video-upscaler': {
5663
5713
  input: falEndpoints.ClarityaiCrystalVideoUpscalerInput;
5664
5714
  output: falEndpoints.ClarityaiCrystalVideoUpscalerOutput;
@@ -5714,6 +5764,11 @@ declare global {
5714
5764
  output: falEndpoints.BriaVideoBackgroundRemovalOutput;
5715
5765
  };
5716
5766
 
5767
+ 'bria/upscale/creative': {
5768
+ input: falEndpoints.BriaUpscaleCreativeInput;
5769
+ output: falEndpoints.BriaUpscaleCreativeOutput;
5770
+ };
5771
+
5717
5772
  'bria/text-to-image/3.2': {
5718
5773
  input: falEndpoints.BriaTextToImage32Input;
5719
5774
  output: falEndpoints.BriaTextToImage32Output;
@@ -5834,11 +5889,6 @@ declare global {
5834
5889
  output: falEndpoints.BriaBria_video_eraserEraseKeypointsOutput;
5835
5890
  };
5836
5891
 
5837
- 'bria/aesthetics/upscaler': {
5838
- input: { [x in string]: any };
5839
- output: { [x in string]: any };
5840
- };
5841
-
5842
5892
  'beatoven/sound-effect-generation': {
5843
5893
  input: falEndpoints.BeatovenSoundEffectGenerationInput;
5844
5894
  output: falEndpoints.BeatovenSoundEffectGenerationOutput;