fal-endpoint-types 1.3.32 → 1.3.33
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json
CHANGED
|
@@ -1204,6 +1204,11 @@ declare global {
|
|
|
1204
1204
|
output: falEndpoints.Sora2ImageToVideoOutput;
|
|
1205
1205
|
};
|
|
1206
1206
|
|
|
1207
|
+
'fal-ai/sora-2/characters': {
|
|
1208
|
+
input: falEndpoints.Sora2CharactersInput;
|
|
1209
|
+
output: falEndpoints.Sora2CharactersOutput;
|
|
1210
|
+
};
|
|
1211
|
+
|
|
1207
1212
|
'fal-ai/smart-turn': {
|
|
1208
1213
|
input: falEndpoints.SmartTurnInput;
|
|
1209
1214
|
output: falEndpoints.SmartTurnOutput;
|
|
@@ -1831,100 +1831,6 @@ export interface SharedType_faf {
|
|
|
1831
1831
|
sync_mode?: boolean;
|
|
1832
1832
|
}
|
|
1833
1833
|
|
|
1834
|
-
export interface SharedType_fa0 {
|
|
1835
|
-
/**
|
|
1836
|
-
* Control Lora Image Url
|
|
1837
|
-
* @description The image to use for control lora. This is used to control the style of the generated image.
|
|
1838
|
-
* @example https://v3.fal.media/files/kangaroo/Cb7BeM7G4DauK_lWjzY3N_Celeb6.jpg
|
|
1839
|
-
*/
|
|
1840
|
-
control_lora_image_url: string;
|
|
1841
|
-
/**
|
|
1842
|
-
* Control Lora Strength
|
|
1843
|
-
* @description The strength of the control lora.
|
|
1844
|
-
* @default 1
|
|
1845
|
-
*/
|
|
1846
|
-
control_lora_strength?: number;
|
|
1847
|
-
/**
|
|
1848
|
-
* Enable Safety Checker
|
|
1849
|
-
* @description If set to true, the safety checker will be enabled.
|
|
1850
|
-
* @default true
|
|
1851
|
-
*/
|
|
1852
|
-
enable_safety_checker?: boolean;
|
|
1853
|
-
/**
|
|
1854
|
-
* Guidance scale (CFG)
|
|
1855
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
1856
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
1857
|
-
* @default 3.5
|
|
1858
|
-
*/
|
|
1859
|
-
guidance_scale?: number;
|
|
1860
|
-
/**
|
|
1861
|
-
* Image Size
|
|
1862
|
-
* @description The size of the generated image.
|
|
1863
|
-
* @default landscape_4_3
|
|
1864
|
-
*/
|
|
1865
|
-
image_size?:
|
|
1866
|
-
| Components.ImageSize
|
|
1867
|
-
| (
|
|
1868
|
-
| 'square_hd'
|
|
1869
|
-
| 'square'
|
|
1870
|
-
| 'portrait_4_3'
|
|
1871
|
-
| 'portrait_16_9'
|
|
1872
|
-
| 'landscape_4_3'
|
|
1873
|
-
| 'landscape_16_9'
|
|
1874
|
-
);
|
|
1875
|
-
/**
|
|
1876
|
-
* Loras
|
|
1877
|
-
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
1878
|
-
* and they will be merged together to generate the final image.
|
|
1879
|
-
* @default []
|
|
1880
|
-
*/
|
|
1881
|
-
loras?: Components.LoraWeight[];
|
|
1882
|
-
/**
|
|
1883
|
-
* Num Images
|
|
1884
|
-
* @description The number of images to generate.
|
|
1885
|
-
* @default 1
|
|
1886
|
-
*/
|
|
1887
|
-
num_images?: number;
|
|
1888
|
-
/**
|
|
1889
|
-
* Num Inference Steps
|
|
1890
|
-
* @description The number of inference steps to perform.
|
|
1891
|
-
* @default 28
|
|
1892
|
-
*/
|
|
1893
|
-
num_inference_steps?: number;
|
|
1894
|
-
/**
|
|
1895
|
-
* Output Format
|
|
1896
|
-
* @description The format of the generated image.
|
|
1897
|
-
* @default jpeg
|
|
1898
|
-
* @enum {string}
|
|
1899
|
-
*/
|
|
1900
|
-
output_format?: 'jpeg' | 'png';
|
|
1901
|
-
/**
|
|
1902
|
-
* Preprocess Depth
|
|
1903
|
-
* @description If set to true, the input image will be preprocessed to extract depth information.
|
|
1904
|
-
* This is useful for generating depth maps from images.
|
|
1905
|
-
* @default true
|
|
1906
|
-
*/
|
|
1907
|
-
preprocess_depth?: boolean;
|
|
1908
|
-
/**
|
|
1909
|
-
* Prompt
|
|
1910
|
-
* @description The prompt to generate an image from.
|
|
1911
|
-
* @example Extreme close-up of a single tiger eye, direct frontal view. Detailed iris and pupil. Sharp focus on eye texture and color. Natural lighting to capture authentic eye shine and depth. The word "FLUX" is painted over it in big, white brush strokes with visible texture.
|
|
1912
|
-
*/
|
|
1913
|
-
prompt: string;
|
|
1914
|
-
/**
|
|
1915
|
-
* Seed
|
|
1916
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
1917
|
-
* will output the same image every time.
|
|
1918
|
-
*/
|
|
1919
|
-
seed?: number;
|
|
1920
|
-
/**
|
|
1921
|
-
* Sync Mode
|
|
1922
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
1923
|
-
* @default false
|
|
1924
|
-
*/
|
|
1925
|
-
sync_mode?: boolean;
|
|
1926
|
-
}
|
|
1927
|
-
|
|
1928
1834
|
export interface SharedType_f7a {
|
|
1929
1835
|
config_file: Components.File;
|
|
1930
1836
|
debug_dataset?: Components.File;
|
|
@@ -6368,6 +6274,105 @@ export interface SharedType_913 {
|
|
|
6368
6274
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
6369
6275
|
}
|
|
6370
6276
|
|
|
6277
|
+
export interface SharedType_90c {
|
|
6278
|
+
/**
|
|
6279
|
+
* Acceleration
|
|
6280
|
+
* @description Acceleration level for image generation. 'regular' balances speed and quality.
|
|
6281
|
+
* @default none
|
|
6282
|
+
* @enum {string}
|
|
6283
|
+
*/
|
|
6284
|
+
acceleration?: 'none' | 'regular';
|
|
6285
|
+
/**
|
|
6286
|
+
* Enable Safety Checker
|
|
6287
|
+
* @description If set to true, the safety checker will be enabled.
|
|
6288
|
+
* @default true
|
|
6289
|
+
*/
|
|
6290
|
+
enable_safety_checker?: boolean;
|
|
6291
|
+
/**
|
|
6292
|
+
* Guidance scale (CFG)
|
|
6293
|
+
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
6294
|
+
* the model to stick to your prompt when looking for a related image to show you.
|
|
6295
|
+
* @default 3.5
|
|
6296
|
+
*/
|
|
6297
|
+
guidance_scale?: number;
|
|
6298
|
+
/**
|
|
6299
|
+
* Image Size
|
|
6300
|
+
* @description The size of the generated image.
|
|
6301
|
+
*/
|
|
6302
|
+
image_size?:
|
|
6303
|
+
| Components.ImageSize
|
|
6304
|
+
| (
|
|
6305
|
+
| 'square_hd'
|
|
6306
|
+
| 'square'
|
|
6307
|
+
| 'portrait_4_3'
|
|
6308
|
+
| 'portrait_16_9'
|
|
6309
|
+
| 'landscape_4_3'
|
|
6310
|
+
| 'landscape_16_9'
|
|
6311
|
+
);
|
|
6312
|
+
/**
|
|
6313
|
+
* Image Url
|
|
6314
|
+
* @description URL of image to use for inpainting. or img2img
|
|
6315
|
+
* @example https://storage.googleapis.com/falserverless/example_inputs/dog.png
|
|
6316
|
+
*/
|
|
6317
|
+
image_url: string;
|
|
6318
|
+
/**
|
|
6319
|
+
* Loras
|
|
6320
|
+
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
6321
|
+
* and they will be merged together to generate the final image.
|
|
6322
|
+
* @default []
|
|
6323
|
+
*/
|
|
6324
|
+
loras?: Components.LoraWeight[];
|
|
6325
|
+
/**
|
|
6326
|
+
* Mask Url
|
|
6327
|
+
* @description The mask to area to Inpaint in.
|
|
6328
|
+
* @example https://storage.googleapis.com/falserverless/example_inputs/dog_mask.png
|
|
6329
|
+
*/
|
|
6330
|
+
mask_url: string;
|
|
6331
|
+
/**
|
|
6332
|
+
* Num Images
|
|
6333
|
+
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
6334
|
+
* @default 1
|
|
6335
|
+
*/
|
|
6336
|
+
num_images?: number;
|
|
6337
|
+
/**
|
|
6338
|
+
* Num Inference Steps
|
|
6339
|
+
* @description The number of inference steps to perform.
|
|
6340
|
+
* @default 28
|
|
6341
|
+
*/
|
|
6342
|
+
num_inference_steps?: number;
|
|
6343
|
+
/**
|
|
6344
|
+
* Output Format
|
|
6345
|
+
* @description The format of the generated image.
|
|
6346
|
+
* @default jpeg
|
|
6347
|
+
* @enum {string}
|
|
6348
|
+
*/
|
|
6349
|
+
output_format?: 'jpeg' | 'png';
|
|
6350
|
+
/**
|
|
6351
|
+
* Prompt
|
|
6352
|
+
* @description The prompt to generate an image from.
|
|
6353
|
+
* @example A photo of a lion sitting on a stone bench
|
|
6354
|
+
*/
|
|
6355
|
+
prompt: string;
|
|
6356
|
+
/**
|
|
6357
|
+
* Seed
|
|
6358
|
+
* @description The same seed and the same prompt given to the same version of the model
|
|
6359
|
+
* will output the same image every time.
|
|
6360
|
+
*/
|
|
6361
|
+
seed?: number;
|
|
6362
|
+
/**
|
|
6363
|
+
* Strength
|
|
6364
|
+
* @description The strength to use for inpainting/image-to-image. Only used if the image_url is provided. 1.0 is completely remakes the image while 0.0 preserves the original.
|
|
6365
|
+
* @default 0.85
|
|
6366
|
+
*/
|
|
6367
|
+
strength?: number;
|
|
6368
|
+
/**
|
|
6369
|
+
* Sync Mode
|
|
6370
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
6371
|
+
* @default false
|
|
6372
|
+
*/
|
|
6373
|
+
sync_mode?: boolean;
|
|
6374
|
+
}
|
|
6375
|
+
|
|
6371
6376
|
export interface SharedType_905 {
|
|
6372
6377
|
/**
|
|
6373
6378
|
* Aspect Ratio
|
|
@@ -10090,6 +10095,105 @@ export interface SharedType_4d6 {
|
|
|
10090
10095
|
};
|
|
10091
10096
|
}
|
|
10092
10097
|
|
|
10098
|
+
export interface SharedType_4c9 {
|
|
10099
|
+
/**
|
|
10100
|
+
* Control Lora Image Url
|
|
10101
|
+
* @description The image to use for control lora. This is used to control the style of the generated image.
|
|
10102
|
+
* @example https://v3.fal.media/files/kangaroo/Cb7BeM7G4DauK_lWjzY3N_Celeb6.jpg
|
|
10103
|
+
*/
|
|
10104
|
+
control_lora_image_url: string;
|
|
10105
|
+
/**
|
|
10106
|
+
* Control Lora Strength
|
|
10107
|
+
* @description The strength of the control lora.
|
|
10108
|
+
* @default 1
|
|
10109
|
+
*/
|
|
10110
|
+
control_lora_strength?: number;
|
|
10111
|
+
/**
|
|
10112
|
+
* Enable Safety Checker
|
|
10113
|
+
* @description If set to true, the safety checker will be enabled.
|
|
10114
|
+
* @default true
|
|
10115
|
+
*/
|
|
10116
|
+
enable_safety_checker?: boolean;
|
|
10117
|
+
/**
|
|
10118
|
+
* Guidance scale (CFG)
|
|
10119
|
+
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
10120
|
+
* the model to stick to your prompt when looking for a related image to show you.
|
|
10121
|
+
* @default 3.5
|
|
10122
|
+
*/
|
|
10123
|
+
guidance_scale?: number;
|
|
10124
|
+
/**
|
|
10125
|
+
* Image Size
|
|
10126
|
+
* @description The size of the generated image.
|
|
10127
|
+
* @default landscape_4_3
|
|
10128
|
+
*/
|
|
10129
|
+
image_size?:
|
|
10130
|
+
| Components.ImageSize
|
|
10131
|
+
| (
|
|
10132
|
+
| 'square_hd'
|
|
10133
|
+
| 'square'
|
|
10134
|
+
| 'portrait_4_3'
|
|
10135
|
+
| 'portrait_16_9'
|
|
10136
|
+
| 'landscape_4_3'
|
|
10137
|
+
| 'landscape_16_9'
|
|
10138
|
+
);
|
|
10139
|
+
/**
|
|
10140
|
+
* Image Url
|
|
10141
|
+
* @description URL of image to use for image-to-image generation.
|
|
10142
|
+
*/
|
|
10143
|
+
image_url?: string;
|
|
10144
|
+
/**
|
|
10145
|
+
* Loras
|
|
10146
|
+
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
10147
|
+
* and they will be merged together to generate the final image.
|
|
10148
|
+
* @default []
|
|
10149
|
+
*/
|
|
10150
|
+
loras?: Components.LoraWeight[];
|
|
10151
|
+
/**
|
|
10152
|
+
* Num Images
|
|
10153
|
+
* @description The number of images to generate.
|
|
10154
|
+
* @default 1
|
|
10155
|
+
*/
|
|
10156
|
+
num_images?: number;
|
|
10157
|
+
/**
|
|
10158
|
+
* Num Inference Steps
|
|
10159
|
+
* @description The number of inference steps to perform.
|
|
10160
|
+
* @default 28
|
|
10161
|
+
*/
|
|
10162
|
+
num_inference_steps?: number;
|
|
10163
|
+
/**
|
|
10164
|
+
* Output Format
|
|
10165
|
+
* @description The format of the generated image.
|
|
10166
|
+
* @default jpeg
|
|
10167
|
+
* @enum {string}
|
|
10168
|
+
*/
|
|
10169
|
+
output_format?: 'jpeg' | 'png';
|
|
10170
|
+
/**
|
|
10171
|
+
* Preprocess Depth
|
|
10172
|
+
* @description If set to true, the input image will be preprocessed to extract depth information.
|
|
10173
|
+
* This is useful for generating depth maps from images.
|
|
10174
|
+
* @default true
|
|
10175
|
+
*/
|
|
10176
|
+
preprocess_depth?: boolean;
|
|
10177
|
+
/**
|
|
10178
|
+
* Prompt
|
|
10179
|
+
* @description The prompt to generate an image from.
|
|
10180
|
+
* @example Extreme close-up of a single tiger eye, direct frontal view. Detailed iris and pupil. Sharp focus on eye texture and color. Natural lighting to capture authentic eye shine and depth. The word "FLUX" is painted over it in big, white brush strokes with visible texture.
|
|
10181
|
+
*/
|
|
10182
|
+
prompt: string;
|
|
10183
|
+
/**
|
|
10184
|
+
* Seed
|
|
10185
|
+
* @description The same seed and the same prompt given to the same version of the model
|
|
10186
|
+
* will output the same image every time.
|
|
10187
|
+
*/
|
|
10188
|
+
seed?: number;
|
|
10189
|
+
/**
|
|
10190
|
+
* Sync Mode
|
|
10191
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
10192
|
+
* @default false
|
|
10193
|
+
*/
|
|
10194
|
+
sync_mode?: boolean;
|
|
10195
|
+
}
|
|
10196
|
+
|
|
10093
10197
|
export interface SharedType_4c2 {
|
|
10094
10198
|
/**
|
|
10095
10199
|
* Description
|
|
@@ -10326,6 +10430,99 @@ export interface SharedType_4a0 {
|
|
|
10326
10430
|
sync_mode?: boolean;
|
|
10327
10431
|
}
|
|
10328
10432
|
|
|
10433
|
+
export interface SharedType_49b {
|
|
10434
|
+
/**
|
|
10435
|
+
* Acceleration
|
|
10436
|
+
* @description Acceleration level for image generation. 'regular' balances speed and quality.
|
|
10437
|
+
* @default none
|
|
10438
|
+
* @enum {string}
|
|
10439
|
+
*/
|
|
10440
|
+
acceleration?: 'none' | 'regular';
|
|
10441
|
+
/**
|
|
10442
|
+
* Enable Safety Checker
|
|
10443
|
+
* @description If set to true, the safety checker will be enabled.
|
|
10444
|
+
* @default true
|
|
10445
|
+
*/
|
|
10446
|
+
enable_safety_checker?: boolean;
|
|
10447
|
+
/**
|
|
10448
|
+
* Guidance scale (CFG)
|
|
10449
|
+
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
10450
|
+
* the model to stick to your prompt when looking for a related image to show you.
|
|
10451
|
+
* @default 3.5
|
|
10452
|
+
*/
|
|
10453
|
+
guidance_scale?: number;
|
|
10454
|
+
/**
|
|
10455
|
+
* Image Size
|
|
10456
|
+
* @description The size of the generated image.
|
|
10457
|
+
*/
|
|
10458
|
+
image_size?:
|
|
10459
|
+
| Components.ImageSize
|
|
10460
|
+
| (
|
|
10461
|
+
| 'square_hd'
|
|
10462
|
+
| 'square'
|
|
10463
|
+
| 'portrait_4_3'
|
|
10464
|
+
| 'portrait_16_9'
|
|
10465
|
+
| 'landscape_4_3'
|
|
10466
|
+
| 'landscape_16_9'
|
|
10467
|
+
);
|
|
10468
|
+
/**
|
|
10469
|
+
* Image Url
|
|
10470
|
+
* @description URL of image to use for inpainting. or img2img
|
|
10471
|
+
* @example https://storage.googleapis.com/falserverless/example_inputs/dog.png
|
|
10472
|
+
*/
|
|
10473
|
+
image_url: string;
|
|
10474
|
+
/**
|
|
10475
|
+
* Loras
|
|
10476
|
+
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
10477
|
+
* and they will be merged together to generate the final image.
|
|
10478
|
+
* @default []
|
|
10479
|
+
*/
|
|
10480
|
+
loras?: Components.LoraWeight[];
|
|
10481
|
+
/**
|
|
10482
|
+
* Num Images
|
|
10483
|
+
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
10484
|
+
* @default 1
|
|
10485
|
+
*/
|
|
10486
|
+
num_images?: number;
|
|
10487
|
+
/**
|
|
10488
|
+
* Num Inference Steps
|
|
10489
|
+
* @description The number of inference steps to perform.
|
|
10490
|
+
* @default 28
|
|
10491
|
+
*/
|
|
10492
|
+
num_inference_steps?: number;
|
|
10493
|
+
/**
|
|
10494
|
+
* Output Format
|
|
10495
|
+
* @description The format of the generated image.
|
|
10496
|
+
* @default jpeg
|
|
10497
|
+
* @enum {string}
|
|
10498
|
+
*/
|
|
10499
|
+
output_format?: 'jpeg' | 'png';
|
|
10500
|
+
/**
|
|
10501
|
+
* Prompt
|
|
10502
|
+
* @description The prompt to generate an image from.
|
|
10503
|
+
* @example A photo of a lion sitting on a stone bench
|
|
10504
|
+
*/
|
|
10505
|
+
prompt: string;
|
|
10506
|
+
/**
|
|
10507
|
+
* Seed
|
|
10508
|
+
* @description The same seed and the same prompt given to the same version of the model
|
|
10509
|
+
* will output the same image every time.
|
|
10510
|
+
*/
|
|
10511
|
+
seed?: number;
|
|
10512
|
+
/**
|
|
10513
|
+
* Strength
|
|
10514
|
+
* @description The strength to use for inpainting/image-to-image. Only used if the image_url is provided. 1.0 is completely remakes the image while 0.0 preserves the original.
|
|
10515
|
+
* @default 0.85
|
|
10516
|
+
*/
|
|
10517
|
+
strength?: number;
|
|
10518
|
+
/**
|
|
10519
|
+
* Sync Mode
|
|
10520
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
10521
|
+
* @default false
|
|
10522
|
+
*/
|
|
10523
|
+
sync_mode?: boolean;
|
|
10524
|
+
}
|
|
10525
|
+
|
|
10329
10526
|
export interface SharedType_484 {
|
|
10330
10527
|
/**
|
|
10331
10528
|
* Seed
|
|
@@ -12103,81 +12300,6 @@ export interface SharedType_20d {
|
|
|
12103
12300
|
seed: number;
|
|
12104
12301
|
}
|
|
12105
12302
|
|
|
12106
|
-
export interface SharedType_207 {
|
|
12107
|
-
/**
|
|
12108
|
-
* Enable Safety Checker
|
|
12109
|
-
* @description If set to true, the safety checker will be enabled.
|
|
12110
|
-
* @default true
|
|
12111
|
-
*/
|
|
12112
|
-
enable_safety_checker?: boolean;
|
|
12113
|
-
/**
|
|
12114
|
-
* Guidance scale (CFG)
|
|
12115
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
12116
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
12117
|
-
* @default 3.5
|
|
12118
|
-
*/
|
|
12119
|
-
guidance_scale?: number;
|
|
12120
|
-
/**
|
|
12121
|
-
* Image Size
|
|
12122
|
-
* @description The size of the generated image.
|
|
12123
|
-
* @default landscape_4_3
|
|
12124
|
-
*/
|
|
12125
|
-
image_size?:
|
|
12126
|
-
| Components.ImageSize
|
|
12127
|
-
| (
|
|
12128
|
-
| 'square_hd'
|
|
12129
|
-
| 'square'
|
|
12130
|
-
| 'portrait_4_3'
|
|
12131
|
-
| 'portrait_16_9'
|
|
12132
|
-
| 'landscape_4_3'
|
|
12133
|
-
| 'landscape_16_9'
|
|
12134
|
-
);
|
|
12135
|
-
/**
|
|
12136
|
-
* Loras
|
|
12137
|
-
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
12138
|
-
* and they will be merged together to generate the final image.
|
|
12139
|
-
* @default []
|
|
12140
|
-
*/
|
|
12141
|
-
loras?: Components.LoraWeight[];
|
|
12142
|
-
/**
|
|
12143
|
-
* Num Images
|
|
12144
|
-
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
12145
|
-
* @default 1
|
|
12146
|
-
*/
|
|
12147
|
-
num_images?: number;
|
|
12148
|
-
/**
|
|
12149
|
-
* Num Inference Steps
|
|
12150
|
-
* @description The number of inference steps to perform.
|
|
12151
|
-
* @default 28
|
|
12152
|
-
*/
|
|
12153
|
-
num_inference_steps?: number;
|
|
12154
|
-
/**
|
|
12155
|
-
* Output Format
|
|
12156
|
-
* @description The format of the generated image.
|
|
12157
|
-
* @default jpeg
|
|
12158
|
-
* @enum {string}
|
|
12159
|
-
*/
|
|
12160
|
-
output_format?: 'jpeg' | 'png';
|
|
12161
|
-
/**
|
|
12162
|
-
* Prompt
|
|
12163
|
-
* @description The prompt to generate an image from.
|
|
12164
|
-
* @example Extreme close-up of a single tiger eye, direct frontal view. Detailed iris and pupil. Sharp focus on eye texture and color. Natural lighting to capture authentic eye shine and depth. The word "FLUX" is painted over it in big, white brush strokes with visible texture.
|
|
12165
|
-
*/
|
|
12166
|
-
prompt: string;
|
|
12167
|
-
/**
|
|
12168
|
-
* Seed
|
|
12169
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
12170
|
-
* will output the same image every time.
|
|
12171
|
-
*/
|
|
12172
|
-
seed?: number;
|
|
12173
|
-
/**
|
|
12174
|
-
* Sync Mode
|
|
12175
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
12176
|
-
* @default false
|
|
12177
|
-
*/
|
|
12178
|
-
sync_mode?: boolean;
|
|
12179
|
-
}
|
|
12180
|
-
|
|
12181
12303
|
export interface SharedType_1eb {
|
|
12182
12304
|
/**
|
|
12183
12305
|
* @description The generated video file.
|
|
@@ -16448,9 +16570,7 @@ export interface WorkflowUtilitiesAutoSubtitleOutput {
|
|
|
16448
16570
|
* Transcription Metadata
|
|
16449
16571
|
* @description Additional transcription metadata from ElevenLabs (language, segments, etc.)
|
|
16450
16572
|
*/
|
|
16451
|
-
transcription_metadata?:
|
|
16452
|
-
[key: string]: { [x: string]: any } | null;
|
|
16453
|
-
};
|
|
16573
|
+
transcription_metadata?: Record<string, number>;
|
|
16454
16574
|
/**
|
|
16455
16575
|
* @description The video with automatic subtitles
|
|
16456
16576
|
* @example {
|
|
@@ -16465,9 +16585,7 @@ export interface WorkflowUtilitiesAutoSubtitleOutput {
|
|
|
16465
16585
|
* Words
|
|
16466
16586
|
* @description Word-level timing information from transcription service
|
|
16467
16587
|
*/
|
|
16468
|
-
words?:
|
|
16469
|
-
[key: string]: { [x: string]: any } | null;
|
|
16470
|
-
}[];
|
|
16588
|
+
words?: Record<string, never>[];
|
|
16471
16589
|
}
|
|
16472
16590
|
|
|
16473
16591
|
export interface WorkflowUtilitiesAudioCompressorInput {
|
|
@@ -26044,6 +26162,33 @@ export interface Sora2ImageToVideoOutput {
|
|
|
26044
26162
|
video_id: string;
|
|
26045
26163
|
}
|
|
26046
26164
|
|
|
26165
|
+
export interface Sora2CharactersInput {
|
|
26166
|
+
/**
|
|
26167
|
+
* Name
|
|
26168
|
+
* @description Name for the character (1–80 characters). Refer to this name in prompts when using the character.
|
|
26169
|
+
*/
|
|
26170
|
+
name: string;
|
|
26171
|
+
/**
|
|
26172
|
+
* Video URL
|
|
26173
|
+
* @description URL of an MP4 video (minimum 720p, max ~2.67:1 aspect ratio) to define the character. Videos exceeding 1080p are automatically scaled down. Non-standard aspect ratios are automatically padded to 16:9 (landscape) or 9:16 (portrait). Videos longer than 4 seconds are trimmed to the first 4 seconds.
|
|
26174
|
+
*/
|
|
26175
|
+
video_url: string;
|
|
26176
|
+
}
|
|
26177
|
+
|
|
26178
|
+
export interface Sora2CharactersOutput {
|
|
26179
|
+
/**
|
|
26180
|
+
* Character ID
|
|
26181
|
+
* @description API character ID (format char_...). Use this in character_ids when generating video.
|
|
26182
|
+
* @example char_123
|
|
26183
|
+
*/
|
|
26184
|
+
id: string;
|
|
26185
|
+
/**
|
|
26186
|
+
* Name
|
|
26187
|
+
* @description The character name
|
|
26188
|
+
*/
|
|
26189
|
+
name: string;
|
|
26190
|
+
}
|
|
26191
|
+
|
|
26047
26192
|
export interface SmartTurnInput {
|
|
26048
26193
|
/**
|
|
26049
26194
|
* Audio Url
|
|
@@ -32002,7 +32147,6 @@ export interface PshumanInput {
|
|
|
32002
32147
|
|
|
32003
32148
|
export interface PshumanOutput {
|
|
32004
32149
|
/**
|
|
32005
|
-
* Model Obj
|
|
32006
32150
|
* @description The generated 3D model in OBJ format.
|
|
32007
32151
|
* @example {
|
|
32008
32152
|
* "file_name": "VGSdkXIgccoKhHs_JtXTa_result_clr_scale4_image.obj",
|
|
@@ -32010,9 +32154,8 @@ export interface PshumanOutput {
|
|
|
32010
32154
|
* "url": "https://storage.googleapis.com/falserverless/model_tests/video_models/VGSdkXIgccoKhHs_JtXTa_result_clr_scale4_image.obj"
|
|
32011
32155
|
* }
|
|
32012
32156
|
*/
|
|
32013
|
-
model_obj: Components.
|
|
32157
|
+
model_obj: Components.File;
|
|
32014
32158
|
/**
|
|
32015
|
-
* Preview Image
|
|
32016
32159
|
* @description A preview image showing the input and the generated multi-view outputs.
|
|
32017
32160
|
* @example {
|
|
32018
32161
|
* "file_name": "WCN_SkT2-RwsGHlxCVHyn_image_preview.png",
|
|
@@ -32020,7 +32163,7 @@ export interface PshumanOutput {
|
|
|
32020
32163
|
* "url": "https://storage.googleapis.com/falserverless/model_tests/video_models/WCN_SkT2-RwsGHlxCVHyn_image_preview.png"
|
|
32021
32164
|
* }
|
|
32022
32165
|
*/
|
|
32023
|
-
preview_image: Components.
|
|
32166
|
+
preview_image: Components.File;
|
|
32024
32167
|
}
|
|
32025
32168
|
|
|
32026
32169
|
export interface PostProcessingVignetteInput {
|
|
@@ -36564,13 +36707,13 @@ export interface NovaSrOutput {
|
|
|
36564
36707
|
/**
|
|
36565
36708
|
* @description The enhanced audio file.
|
|
36566
36709
|
* @example {
|
|
36567
|
-
* "
|
|
36710
|
+
* "bitrate": "192k",
|
|
36568
36711
|
* "duration": 12.283291666666667,
|
|
36569
36712
|
* "url": "https://v3b.fal.media/files/b/0a8a3f1a/lTKExJu-R6ZJdnFlpzEeq_TxmNTNhl.mp3",
|
|
36570
36713
|
* "file_name": "lTKExJu-R6ZJdnFlpzEeq_TxmNTNhl.mp3",
|
|
36571
36714
|
* "sample_rate": 48000,
|
|
36572
36715
|
* "content_type": "audio/mpeg",
|
|
36573
|
-
* "
|
|
36716
|
+
* "channels": 1
|
|
36574
36717
|
* }
|
|
36575
36718
|
*/
|
|
36576
36719
|
audio: Components.AudioFile;
|
|
@@ -42091,7 +42234,7 @@ export interface Ltx2ImageToVideoOutput extends SharedType_62e {}
|
|
|
42091
42234
|
export interface Ltx2ExtendVideoInput {
|
|
42092
42235
|
/**
|
|
42093
42236
|
* Context
|
|
42094
|
-
* @description Number of seconds from the input video to use as context for the extension (maximum 20 seconds). If not provided, defaults to maximize available context within the 505 frame limit.
|
|
42237
|
+
* @description Number of seconds from the input video to use as context for the extension (minimum 1 second, maximum 20 seconds). If not provided, defaults to maximize available context within the 505 frame limit.
|
|
42095
42238
|
*/
|
|
42096
42239
|
context?: number;
|
|
42097
42240
|
/**
|
|
@@ -42462,7 +42605,7 @@ export interface Ltx23ImageToVideoOutput {
|
|
|
42462
42605
|
export interface Ltx23ExtendVideoInput {
|
|
42463
42606
|
/**
|
|
42464
42607
|
* Context
|
|
42465
|
-
* @description Number of seconds from the input video to use as context for the extension (maximum 20 seconds). If not provided, defaults to maximize available context within the 505 frame limit.
|
|
42608
|
+
* @description Number of seconds from the input video to use as context for the extension (minimum 1 second, maximum 20 seconds). If not provided, defaults to maximize available context within the 505 frame limit.
|
|
42466
42609
|
*/
|
|
42467
42610
|
context?: number;
|
|
42468
42611
|
/**
|
|
@@ -50253,13 +50396,6 @@ export interface KlingVideoV3StandardMotionControlOutput {
|
|
|
50253
50396
|
}
|
|
50254
50397
|
|
|
50255
50398
|
export interface KlingVideoV3StandardImageToVideoInput {
|
|
50256
|
-
/**
|
|
50257
|
-
* Aspect Ratio
|
|
50258
|
-
* @description The aspect ratio of the generated video frame
|
|
50259
|
-
* @default 16:9
|
|
50260
|
-
* @enum {string}
|
|
50261
|
-
*/
|
|
50262
|
-
aspect_ratio?: '16:9' | '9:16' | '1:1';
|
|
50263
50399
|
/**
|
|
50264
50400
|
* Cfg Scale
|
|
50265
50401
|
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
@@ -50474,13 +50610,6 @@ export interface KlingVideoV3ProMotionControlOutput {
|
|
|
50474
50610
|
}
|
|
50475
50611
|
|
|
50476
50612
|
export interface KlingVideoV3ProImageToVideoInput {
|
|
50477
|
-
/**
|
|
50478
|
-
* Aspect Ratio
|
|
50479
|
-
* @description The aspect ratio of the generated video frame
|
|
50480
|
-
* @default 16:9
|
|
50481
|
-
* @enum {string}
|
|
50482
|
-
*/
|
|
50483
|
-
aspect_ratio?: '16:9' | '9:16' | '1:1';
|
|
50484
50613
|
/**
|
|
50485
50614
|
* Cfg Scale
|
|
50486
50615
|
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
@@ -59812,8 +59941,8 @@ export interface Hunyuan_worldOutput {
|
|
|
59812
59941
|
/**
|
|
59813
59942
|
* @description The generated panorama image.
|
|
59814
59943
|
* @example {
|
|
59815
|
-
* "file_size": 2738127,
|
|
59816
59944
|
* "height": 960,
|
|
59945
|
+
* "file_size": 2738127,
|
|
59817
59946
|
* "file_name": "5db7925423b44f2a98098cd8f7cad7ec.png",
|
|
59818
59947
|
* "content_type": "image/png",
|
|
59819
59948
|
* "url": "https://v3.fal.media/files/kangaroo/P2AmXuLlyDIsivqjV_rAr_5db7925423b44f2a98098cd8f7cad7ec.png",
|
|
@@ -64196,9 +64325,7 @@ export interface FluxPulidInput {
|
|
|
64196
64325
|
start_step?: number;
|
|
64197
64326
|
/**
|
|
64198
64327
|
* Sync Mode
|
|
64199
|
-
* @description If
|
|
64200
|
-
* before returning the response. This will increase the latency of the function but
|
|
64201
|
-
* it allows you to get the image directly in the response without going through the CDN.
|
|
64328
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
64202
64329
|
* @default false
|
|
64203
64330
|
*/
|
|
64204
64331
|
sync_mode?: boolean;
|
|
@@ -64756,199 +64883,11 @@ export interface FluxLoraStreamInput extends SharedType_c5e {}
|
|
|
64756
64883
|
|
|
64757
64884
|
export interface FluxLoraStreamOutput extends SharedType_7c6 {}
|
|
64758
64885
|
|
|
64759
|
-
export interface FluxLoraInpaintingInput {
|
|
64760
|
-
/**
|
|
64761
|
-
* Acceleration
|
|
64762
|
-
* @description Acceleration level for image generation. 'regular' balances speed and quality.
|
|
64763
|
-
* @default none
|
|
64764
|
-
* @enum {string}
|
|
64765
|
-
*/
|
|
64766
|
-
acceleration?: 'none' | 'regular';
|
|
64767
|
-
/**
|
|
64768
|
-
* Enable Safety Checker
|
|
64769
|
-
* @description If set to true, the safety checker will be enabled.
|
|
64770
|
-
* @default true
|
|
64771
|
-
*/
|
|
64772
|
-
enable_safety_checker?: boolean;
|
|
64773
|
-
/**
|
|
64774
|
-
* Guidance scale (CFG)
|
|
64775
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
64776
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
64777
|
-
* @default 3.5
|
|
64778
|
-
*/
|
|
64779
|
-
guidance_scale?: number;
|
|
64780
|
-
/**
|
|
64781
|
-
* Image Size
|
|
64782
|
-
* @description The size of the generated image.
|
|
64783
|
-
*/
|
|
64784
|
-
image_size?:
|
|
64785
|
-
| Components.ImageSize
|
|
64786
|
-
| (
|
|
64787
|
-
| 'square_hd'
|
|
64788
|
-
| 'square'
|
|
64789
|
-
| 'portrait_4_3'
|
|
64790
|
-
| 'portrait_16_9'
|
|
64791
|
-
| 'landscape_4_3'
|
|
64792
|
-
| 'landscape_16_9'
|
|
64793
|
-
);
|
|
64794
|
-
/**
|
|
64795
|
-
* Image Url
|
|
64796
|
-
* @description URL of image to use for inpainting. or img2img
|
|
64797
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/dog.png
|
|
64798
|
-
*/
|
|
64799
|
-
image_url: string;
|
|
64800
|
-
/**
|
|
64801
|
-
* Loras
|
|
64802
|
-
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
64803
|
-
* and they will be merged together to generate the final image.
|
|
64804
|
-
* @default []
|
|
64805
|
-
*/
|
|
64806
|
-
loras?: Components.LoraWeight[];
|
|
64807
|
-
/**
|
|
64808
|
-
* Mask Url
|
|
64809
|
-
* @description The mask to area to Inpaint in.
|
|
64810
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/dog_mask.png
|
|
64811
|
-
*/
|
|
64812
|
-
mask_url: string;
|
|
64813
|
-
/**
|
|
64814
|
-
* Num Images
|
|
64815
|
-
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
64816
|
-
* @default 1
|
|
64817
|
-
*/
|
|
64818
|
-
num_images?: number;
|
|
64819
|
-
/**
|
|
64820
|
-
* Num Inference Steps
|
|
64821
|
-
* @description The number of inference steps to perform.
|
|
64822
|
-
* @default 28
|
|
64823
|
-
*/
|
|
64824
|
-
num_inference_steps?: number;
|
|
64825
|
-
/**
|
|
64826
|
-
* Output Format
|
|
64827
|
-
* @description The format of the generated image.
|
|
64828
|
-
* @default jpeg
|
|
64829
|
-
* @enum {string}
|
|
64830
|
-
*/
|
|
64831
|
-
output_format?: 'jpeg' | 'png';
|
|
64832
|
-
/**
|
|
64833
|
-
* Prompt
|
|
64834
|
-
* @description The prompt to generate an image from.
|
|
64835
|
-
* @example A photo of a lion sitting on a stone bench
|
|
64836
|
-
*/
|
|
64837
|
-
prompt: string;
|
|
64838
|
-
/**
|
|
64839
|
-
* Seed
|
|
64840
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
64841
|
-
* will output the same image every time.
|
|
64842
|
-
*/
|
|
64843
|
-
seed?: number;
|
|
64844
|
-
/**
|
|
64845
|
-
* Strength
|
|
64846
|
-
* @description The strength to use for inpainting/image-to-image. Only used if the image_url is provided. 1.0 is completely remakes the image while 0.0 preserves the original.
|
|
64847
|
-
* @default 0.85
|
|
64848
|
-
*/
|
|
64849
|
-
strength?: number;
|
|
64850
|
-
/**
|
|
64851
|
-
* Sync Mode
|
|
64852
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
64853
|
-
* @default false
|
|
64854
|
-
*/
|
|
64855
|
-
sync_mode?: boolean;
|
|
64856
|
-
}
|
|
64886
|
+
export interface FluxLoraInpaintingInput extends SharedType_90c {}
|
|
64857
64887
|
|
|
64858
64888
|
export interface FluxLoraInpaintingOutput extends SharedType_7c6 {}
|
|
64859
64889
|
|
|
64860
|
-
export interface FluxLoraImageToImageInput {
|
|
64861
|
-
/**
|
|
64862
|
-
* Acceleration
|
|
64863
|
-
* @description Acceleration level for image generation. 'regular' balances speed and quality.
|
|
64864
|
-
* @default none
|
|
64865
|
-
* @enum {string}
|
|
64866
|
-
*/
|
|
64867
|
-
acceleration?: 'none' | 'regular';
|
|
64868
|
-
/**
|
|
64869
|
-
* Enable Safety Checker
|
|
64870
|
-
* @description If set to true, the safety checker will be enabled.
|
|
64871
|
-
* @default true
|
|
64872
|
-
*/
|
|
64873
|
-
enable_safety_checker?: boolean;
|
|
64874
|
-
/**
|
|
64875
|
-
* Guidance scale (CFG)
|
|
64876
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
64877
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
64878
|
-
* @default 3.5
|
|
64879
|
-
*/
|
|
64880
|
-
guidance_scale?: number;
|
|
64881
|
-
/**
|
|
64882
|
-
* Image Size
|
|
64883
|
-
* @description The size of the generated image.
|
|
64884
|
-
*/
|
|
64885
|
-
image_size?:
|
|
64886
|
-
| Components.ImageSize
|
|
64887
|
-
| (
|
|
64888
|
-
| 'square_hd'
|
|
64889
|
-
| 'square'
|
|
64890
|
-
| 'portrait_4_3'
|
|
64891
|
-
| 'portrait_16_9'
|
|
64892
|
-
| 'landscape_4_3'
|
|
64893
|
-
| 'landscape_16_9'
|
|
64894
|
-
);
|
|
64895
|
-
/**
|
|
64896
|
-
* Image Url
|
|
64897
|
-
* @description URL of image to use for inpainting. or img2img
|
|
64898
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/dog.png
|
|
64899
|
-
*/
|
|
64900
|
-
image_url: string;
|
|
64901
|
-
/**
|
|
64902
|
-
* Loras
|
|
64903
|
-
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
64904
|
-
* and they will be merged together to generate the final image.
|
|
64905
|
-
* @default []
|
|
64906
|
-
*/
|
|
64907
|
-
loras?: Components.LoraWeight[];
|
|
64908
|
-
/**
|
|
64909
|
-
* Num Images
|
|
64910
|
-
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
64911
|
-
* @default 1
|
|
64912
|
-
*/
|
|
64913
|
-
num_images?: number;
|
|
64914
|
-
/**
|
|
64915
|
-
* Num Inference Steps
|
|
64916
|
-
* @description The number of inference steps to perform.
|
|
64917
|
-
* @default 28
|
|
64918
|
-
*/
|
|
64919
|
-
num_inference_steps?: number;
|
|
64920
|
-
/**
|
|
64921
|
-
* Output Format
|
|
64922
|
-
* @description The format of the generated image.
|
|
64923
|
-
* @default jpeg
|
|
64924
|
-
* @enum {string}
|
|
64925
|
-
*/
|
|
64926
|
-
output_format?: 'jpeg' | 'png';
|
|
64927
|
-
/**
|
|
64928
|
-
* Prompt
|
|
64929
|
-
* @description The prompt to generate an image from.
|
|
64930
|
-
* @example A photo of a lion sitting on a stone bench
|
|
64931
|
-
*/
|
|
64932
|
-
prompt: string;
|
|
64933
|
-
/**
|
|
64934
|
-
* Seed
|
|
64935
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
64936
|
-
* will output the same image every time.
|
|
64937
|
-
*/
|
|
64938
|
-
seed?: number;
|
|
64939
|
-
/**
|
|
64940
|
-
* Strength
|
|
64941
|
-
* @description The strength to use for inpainting/image-to-image. Only used if the image_url is provided. 1.0 is completely remakes the image while 0.0 preserves the original.
|
|
64942
|
-
* @default 0.85
|
|
64943
|
-
*/
|
|
64944
|
-
strength?: number;
|
|
64945
|
-
/**
|
|
64946
|
-
* Sync Mode
|
|
64947
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
64948
|
-
* @default false
|
|
64949
|
-
*/
|
|
64950
|
-
sync_mode?: boolean;
|
|
64951
|
-
}
|
|
64890
|
+
export interface FluxLoraImageToImageInput extends SharedType_49b {}
|
|
64952
64891
|
|
|
64953
64892
|
export interface FluxLoraImageToImageOutput extends SharedType_7c6 {}
|
|
64954
64893
|
|
|
@@ -65312,193 +65251,19 @@ export interface FluxKreaTrainerInput extends SharedType_f11 {}
|
|
|
65312
65251
|
|
|
65313
65252
|
export interface FluxKreaTrainerOutput extends SharedType_baf1 {}
|
|
65314
65253
|
|
|
65315
|
-
export interface FluxKreaLoraStreamInput extends
|
|
65254
|
+
export interface FluxKreaLoraStreamInput extends SharedType_c5e {}
|
|
65316
65255
|
|
|
65317
65256
|
export interface FluxKreaLoraStreamOutput extends SharedType_7c6 {}
|
|
65318
65257
|
|
|
65319
|
-
export interface FluxKreaLoraInpaintingInput {
|
|
65320
|
-
/**
|
|
65321
|
-
* Enable Safety Checker
|
|
65322
|
-
* @description If set to true, the safety checker will be enabled.
|
|
65323
|
-
* @default true
|
|
65324
|
-
*/
|
|
65325
|
-
enable_safety_checker?: boolean;
|
|
65326
|
-
/**
|
|
65327
|
-
* Guidance scale (CFG)
|
|
65328
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
65329
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
65330
|
-
* @default 3.5
|
|
65331
|
-
*/
|
|
65332
|
-
guidance_scale?: number;
|
|
65333
|
-
/**
|
|
65334
|
-
* Image Size
|
|
65335
|
-
* @description The size of the generated image.
|
|
65336
|
-
*/
|
|
65337
|
-
image_size?:
|
|
65338
|
-
| Components.ImageSize
|
|
65339
|
-
| (
|
|
65340
|
-
| 'square_hd'
|
|
65341
|
-
| 'square'
|
|
65342
|
-
| 'portrait_4_3'
|
|
65343
|
-
| 'portrait_16_9'
|
|
65344
|
-
| 'landscape_4_3'
|
|
65345
|
-
| 'landscape_16_9'
|
|
65346
|
-
);
|
|
65347
|
-
/**
|
|
65348
|
-
* Image Url
|
|
65349
|
-
* @description URL of image to use for inpainting. or img2img
|
|
65350
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/dog.png
|
|
65351
|
-
*/
|
|
65352
|
-
image_url: string;
|
|
65353
|
-
/**
|
|
65354
|
-
* Loras
|
|
65355
|
-
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
65356
|
-
* and they will be merged together to generate the final image.
|
|
65357
|
-
* @default []
|
|
65358
|
-
*/
|
|
65359
|
-
loras?: Components.LoraWeight[];
|
|
65360
|
-
/**
|
|
65361
|
-
* Mask Url
|
|
65362
|
-
* @description The mask to area to Inpaint in.
|
|
65363
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/dog_mask.png
|
|
65364
|
-
*/
|
|
65365
|
-
mask_url: string;
|
|
65366
|
-
/**
|
|
65367
|
-
* Num Images
|
|
65368
|
-
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
65369
|
-
* @default 1
|
|
65370
|
-
*/
|
|
65371
|
-
num_images?: number;
|
|
65372
|
-
/**
|
|
65373
|
-
* Num Inference Steps
|
|
65374
|
-
* @description The number of inference steps to perform.
|
|
65375
|
-
* @default 28
|
|
65376
|
-
*/
|
|
65377
|
-
num_inference_steps?: number;
|
|
65378
|
-
/**
|
|
65379
|
-
* Output Format
|
|
65380
|
-
* @description The format of the generated image.
|
|
65381
|
-
* @default jpeg
|
|
65382
|
-
* @enum {string}
|
|
65383
|
-
*/
|
|
65384
|
-
output_format?: 'jpeg' | 'png';
|
|
65385
|
-
/**
|
|
65386
|
-
* Prompt
|
|
65387
|
-
* @description The prompt to generate an image from.
|
|
65388
|
-
* @example A photo of a lion sitting on a stone bench
|
|
65389
|
-
*/
|
|
65390
|
-
prompt: string;
|
|
65391
|
-
/**
|
|
65392
|
-
* Seed
|
|
65393
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
65394
|
-
* will output the same image every time.
|
|
65395
|
-
*/
|
|
65396
|
-
seed?: number;
|
|
65397
|
-
/**
|
|
65398
|
-
* Strength
|
|
65399
|
-
* @description The strength to use for inpainting/image-to-image. Only used if the image_url is provided. 1.0 is completely remakes the image while 0.0 preserves the original.
|
|
65400
|
-
* @default 0.85
|
|
65401
|
-
*/
|
|
65402
|
-
strength?: number;
|
|
65403
|
-
/**
|
|
65404
|
-
* Sync Mode
|
|
65405
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
65406
|
-
* @default false
|
|
65407
|
-
*/
|
|
65408
|
-
sync_mode?: boolean;
|
|
65409
|
-
}
|
|
65258
|
+
export interface FluxKreaLoraInpaintingInput extends SharedType_90c {}
|
|
65410
65259
|
|
|
65411
65260
|
export interface FluxKreaLoraInpaintingOutput extends SharedType_7c6 {}
|
|
65412
65261
|
|
|
65413
|
-
export interface FluxKreaLoraImageToImageInput {
|
|
65414
|
-
/**
|
|
65415
|
-
* Enable Safety Checker
|
|
65416
|
-
* @description If set to true, the safety checker will be enabled.
|
|
65417
|
-
* @default true
|
|
65418
|
-
*/
|
|
65419
|
-
enable_safety_checker?: boolean;
|
|
65420
|
-
/**
|
|
65421
|
-
* Guidance scale (CFG)
|
|
65422
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
65423
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
65424
|
-
* @default 3.5
|
|
65425
|
-
*/
|
|
65426
|
-
guidance_scale?: number;
|
|
65427
|
-
/**
|
|
65428
|
-
* Image Size
|
|
65429
|
-
* @description The size of the generated image.
|
|
65430
|
-
*/
|
|
65431
|
-
image_size?:
|
|
65432
|
-
| Components.ImageSize
|
|
65433
|
-
| (
|
|
65434
|
-
| 'square_hd'
|
|
65435
|
-
| 'square'
|
|
65436
|
-
| 'portrait_4_3'
|
|
65437
|
-
| 'portrait_16_9'
|
|
65438
|
-
| 'landscape_4_3'
|
|
65439
|
-
| 'landscape_16_9'
|
|
65440
|
-
);
|
|
65441
|
-
/**
|
|
65442
|
-
* Image Url
|
|
65443
|
-
* @description URL of image to use for inpainting. or img2img
|
|
65444
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/dog.png
|
|
65445
|
-
*/
|
|
65446
|
-
image_url: string;
|
|
65447
|
-
/**
|
|
65448
|
-
* Loras
|
|
65449
|
-
* @description The LoRAs to use for the image generation. You can use any number of LoRAs
|
|
65450
|
-
* and they will be merged together to generate the final image.
|
|
65451
|
-
* @default []
|
|
65452
|
-
*/
|
|
65453
|
-
loras?: Components.LoraWeight[];
|
|
65454
|
-
/**
|
|
65455
|
-
* Num Images
|
|
65456
|
-
* @description The number of images to generate. This is always set to 1 for streaming output.
|
|
65457
|
-
* @default 1
|
|
65458
|
-
*/
|
|
65459
|
-
num_images?: number;
|
|
65460
|
-
/**
|
|
65461
|
-
* Num Inference Steps
|
|
65462
|
-
* @description The number of inference steps to perform.
|
|
65463
|
-
* @default 28
|
|
65464
|
-
*/
|
|
65465
|
-
num_inference_steps?: number;
|
|
65466
|
-
/**
|
|
65467
|
-
* Output Format
|
|
65468
|
-
* @description The format of the generated image.
|
|
65469
|
-
* @default jpeg
|
|
65470
|
-
* @enum {string}
|
|
65471
|
-
*/
|
|
65472
|
-
output_format?: 'jpeg' | 'png';
|
|
65473
|
-
/**
|
|
65474
|
-
* Prompt
|
|
65475
|
-
* @description The prompt to generate an image from.
|
|
65476
|
-
* @example A photo of a lion sitting on a stone bench
|
|
65477
|
-
*/
|
|
65478
|
-
prompt: string;
|
|
65479
|
-
/**
|
|
65480
|
-
* Seed
|
|
65481
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
65482
|
-
* will output the same image every time.
|
|
65483
|
-
*/
|
|
65484
|
-
seed?: number;
|
|
65485
|
-
/**
|
|
65486
|
-
* Strength
|
|
65487
|
-
* @description The strength to use for inpainting/image-to-image. Only used if the image_url is provided. 1.0 is completely remakes the image while 0.0 preserves the original.
|
|
65488
|
-
* @default 0.85
|
|
65489
|
-
*/
|
|
65490
|
-
strength?: number;
|
|
65491
|
-
/**
|
|
65492
|
-
* Sync Mode
|
|
65493
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
65494
|
-
* @default false
|
|
65495
|
-
*/
|
|
65496
|
-
sync_mode?: boolean;
|
|
65497
|
-
}
|
|
65262
|
+
export interface FluxKreaLoraImageToImageInput extends SharedType_49b {}
|
|
65498
65263
|
|
|
65499
65264
|
export interface FluxKreaLoraImageToImageOutput extends SharedType_7c6 {}
|
|
65500
65265
|
|
|
65501
|
-
export interface FluxKreaLoraInput extends
|
|
65266
|
+
export interface FluxKreaLoraInput extends SharedType_c5e {}
|
|
65502
65267
|
|
|
65503
65268
|
export interface FluxKreaLoraOutput extends SharedType_7c6 {}
|
|
65504
65269
|
|
|
@@ -67237,7 +67002,7 @@ export interface FluxControlLoraDepthImageToImageInput extends SharedType_dce {}
|
|
|
67237
67002
|
|
|
67238
67003
|
export interface FluxControlLoraDepthImageToImageOutput extends SharedType_7c6 {}
|
|
67239
67004
|
|
|
67240
|
-
export interface FluxControlLoraDepthInput extends
|
|
67005
|
+
export interface FluxControlLoraDepthInput extends SharedType_4c9 {}
|
|
67241
67006
|
|
|
67242
67007
|
export interface FluxControlLoraDepthOutput extends SharedType_7c6 {}
|
|
67243
67008
|
|
|
@@ -67245,7 +67010,7 @@ export interface FluxControlLoraCannyImageToImageInput extends SharedType_dce {}
|
|
|
67245
67010
|
|
|
67246
67011
|
export interface FluxControlLoraCannyImageToImageOutput extends SharedType_7c6 {}
|
|
67247
67012
|
|
|
67248
|
-
export interface FluxControlLoraCannyInput extends
|
|
67013
|
+
export interface FluxControlLoraCannyInput extends SharedType_4c9 {}
|
|
67249
67014
|
|
|
67250
67015
|
export interface FluxControlLoraCannyOutput extends SharedType_7c6 {}
|
|
67251
67016
|
|
|
@@ -77911,8 +77676,8 @@ export interface CodeformerOutput {
|
|
|
77911
77676
|
/**
|
|
77912
77677
|
* @description The generated image file info.
|
|
77913
77678
|
* @example {
|
|
77914
|
-
* "file_size": 423052,
|
|
77915
77679
|
* "height": 512,
|
|
77680
|
+
* "file_size": 423052,
|
|
77916
77681
|
* "file_name": "36d3ca4791a647678b2ff01a35c87f5a.png",
|
|
77917
77682
|
* "content_type": "image/png",
|
|
77918
77683
|
* "url": "https://storage.googleapis.com/falserverless/model_tests/codeformer/codeformer_restored_1.jpeg",
|
|
@@ -81366,8 +81131,8 @@ export interface BagelEditOutput {
|
|
|
81366
81131
|
* @description The edited images.
|
|
81367
81132
|
* @example [
|
|
81368
81133
|
* {
|
|
81369
|
-
* "file_size": 423052,
|
|
81370
81134
|
* "height": 1024,
|
|
81135
|
+
* "file_size": 423052,
|
|
81371
81136
|
* "file_name": "hQnndOMvGSt2UsYAiV3vs.jpeg",
|
|
81372
81137
|
* "content_type": "image/jpeg",
|
|
81373
81138
|
* "url": "https://storage.googleapis.com/falserverless/bagel/hQnndOMvGSt2UsYAiV3vs.jpeg",
|
|
@@ -81429,8 +81194,8 @@ export interface BagelOutput {
|
|
|
81429
81194
|
* @description The generated images.
|
|
81430
81195
|
* @example [
|
|
81431
81196
|
* {
|
|
81432
|
-
* "file_size": 423052,
|
|
81433
81197
|
* "height": 1024,
|
|
81198
|
+
* "file_size": 423052,
|
|
81434
81199
|
* "file_name": "wRhCPSyiKTiLnnWvUpGIl.jpeg",
|
|
81435
81200
|
* "content_type": "image/jpeg",
|
|
81436
81201
|
* "url": "https://storage.googleapis.com/falserverless/bagel/wRhCPSyiKTiLnnWvUpGIl.jpeg",
|
|
@@ -81500,7 +81265,7 @@ export interface AutoCaptionInput {
|
|
|
81500
81265
|
txt_font?: string;
|
|
81501
81266
|
/**
|
|
81502
81267
|
* Video Url
|
|
81503
|
-
* @description URL to the .mp4 video with audio. Only videos of size <
|
|
81268
|
+
* @description URL to the .mp4 video with audio. Only videos of size <400MB are allowed.
|
|
81504
81269
|
*/
|
|
81505
81270
|
video_url: string;
|
|
81506
81271
|
}
|
|
@@ -83069,8 +82834,8 @@ export interface ClarityaiCrystalVideoUpscalerOutput {
|
|
|
83069
82834
|
* "height": 2160,
|
|
83070
82835
|
* "duration": 13.056527,
|
|
83071
82836
|
* "url": "https://storage.googleapis.com/falserverless/example_outputs/crystal_upscaler/video_upscaling/video_out.mp4",
|
|
83072
|
-
* "width": 4096,
|
|
83073
82837
|
* "fps": 23.130193905817176,
|
|
82838
|
+
* "width": 4096,
|
|
83074
82839
|
* "file_name": "w0VQQvPdwvV2GSCtRTMzh_hDH8SPrB.mp4",
|
|
83075
82840
|
* "num_frames": 302,
|
|
83076
82841
|
* "content_type": "video/mp4"
|
|
@@ -84039,6 +83804,7 @@ export interface BriaFiboEditEditInput {
|
|
|
84039
83804
|
* @default
|
|
84040
83805
|
*/
|
|
84041
83806
|
negative_prompt?: string;
|
|
83807
|
+
original_vgl?: Components.StructuredInstruction;
|
|
84042
83808
|
/**
|
|
84043
83809
|
* Seed
|
|
84044
83810
|
* @description Random seed for reproducibility.
|