fal-endpoint-types 1.3.33 → 1.3.35
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/types/fal/endpoints/components.d.ts +88 -19
- package/types/fal/endpoints/index.d.ts +60 -5
- package/types/fal/endpoints/schema.d.ts +1759 -1011
|
@@ -1,5 +1,60 @@
|
|
|
1
1
|
import type * as Components from './components.js';
|
|
2
2
|
|
|
3
|
+
export interface XaiTtsV1Input {
|
|
4
|
+
/**
|
|
5
|
+
* Language
|
|
6
|
+
* @description BCP-47 language code or 'auto' for automatic detection. Supported: en, zh, fr, de, hi, id, it, ja, ko, pt-BR, pt-PT, ru, es-MX, es-ES, tr, vi, bn, ar-EG, ar-SA, ar-AE.
|
|
7
|
+
* @default auto
|
|
8
|
+
* @enum {string}
|
|
9
|
+
*/
|
|
10
|
+
language?:
|
|
11
|
+
| 'auto'
|
|
12
|
+
| 'en'
|
|
13
|
+
| 'ar-EG'
|
|
14
|
+
| 'ar-SA'
|
|
15
|
+
| 'ar-AE'
|
|
16
|
+
| 'bn'
|
|
17
|
+
| 'zh'
|
|
18
|
+
| 'fr'
|
|
19
|
+
| 'de'
|
|
20
|
+
| 'hi'
|
|
21
|
+
| 'id'
|
|
22
|
+
| 'it'
|
|
23
|
+
| 'ja'
|
|
24
|
+
| 'ko'
|
|
25
|
+
| 'pt-BR'
|
|
26
|
+
| 'pt-PT'
|
|
27
|
+
| 'ru'
|
|
28
|
+
| 'es-MX'
|
|
29
|
+
| 'es-ES'
|
|
30
|
+
| 'tr'
|
|
31
|
+
| 'vi';
|
|
32
|
+
output_format?: Components.OutputFormat;
|
|
33
|
+
/**
|
|
34
|
+
* Text
|
|
35
|
+
* @description The text to convert to speech. Maximum 15,000 characters. Supports speech tags for expressive delivery: inline tags like [laugh], [pause], [sigh] and wrapping tags like <whisper>text</whisper>, <slow>text</slow>.
|
|
36
|
+
* @example Hello! This is xAI text to speech, brought to you by Fal AI.
|
|
37
|
+
*/
|
|
38
|
+
text: string;
|
|
39
|
+
/**
|
|
40
|
+
* Voice
|
|
41
|
+
* @description Voice to use for synthesis. eve: energetic, upbeat. ara: warm, friendly. rex: confident, clear. sal: smooth, balanced. leo: authoritative, strong.
|
|
42
|
+
* @default eve
|
|
43
|
+
* @enum {string}
|
|
44
|
+
*/
|
|
45
|
+
voice?: 'eve' | 'ara' | 'rex' | 'sal' | 'leo';
|
|
46
|
+
}
|
|
47
|
+
|
|
48
|
+
export interface XaiTtsV1Output {
|
|
49
|
+
/**
|
|
50
|
+
* @description The generated audio file.
|
|
51
|
+
* @example {
|
|
52
|
+
* "url": "https://v3b.fal.media/files/b/0a92750b/exZJCm6TDejS5xIulJs2r_xai_tts_output.mp3"
|
|
53
|
+
* }
|
|
54
|
+
*/
|
|
55
|
+
audio: Components.File;
|
|
56
|
+
}
|
|
57
|
+
|
|
3
58
|
export interface XaiGrokImagineVideoTextToVideoInput {
|
|
4
59
|
/**
|
|
5
60
|
* Aspect Ratio
|
|
@@ -39,8 +94,8 @@ export interface XaiGrokImagineVideoTextToVideoOutput {
|
|
|
39
94
|
* "fps": 24,
|
|
40
95
|
* "width": 1280,
|
|
41
96
|
* "file_name": "RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
|
|
42
|
-
* "
|
|
43
|
-
* "
|
|
97
|
+
* "num_frames": 145,
|
|
98
|
+
* "content_type": "video/mp4"
|
|
44
99
|
* }
|
|
45
100
|
*/
|
|
46
101
|
video: Components.VideoFile;
|
|
@@ -90,8 +145,8 @@ export interface XaiGrokImagineVideoImageToVideoOutput {
|
|
|
90
145
|
* "fps": 24,
|
|
91
146
|
* "width": 1280,
|
|
92
147
|
* "file_name": "0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
|
|
93
|
-
* "
|
|
94
|
-
* "
|
|
148
|
+
* "num_frames": 145,
|
|
149
|
+
* "content_type": "video/mp4"
|
|
95
150
|
* }
|
|
96
151
|
*/
|
|
97
152
|
video: Components.VideoFile;
|
|
@@ -129,8 +184,8 @@ export interface XaiGrokImagineVideoEditVideoOutput {
|
|
|
129
184
|
* "fps": 24,
|
|
130
185
|
* "width": 1280,
|
|
131
186
|
* "file_name": "EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
|
|
132
|
-
* "
|
|
133
|
-
* "
|
|
187
|
+
* "num_frames": 121,
|
|
188
|
+
* "content_type": "video/mp4"
|
|
134
189
|
* }
|
|
135
190
|
*/
|
|
136
191
|
video: Components.VideoFile;
|
|
@@ -1527,38 +1582,6 @@ export interface SharedType_fe5 {
|
|
|
1527
1582
|
};
|
|
1528
1583
|
}
|
|
1529
1584
|
|
|
1530
|
-
export interface SharedType_fda {
|
|
1531
|
-
/**
|
|
1532
|
-
* Masks
|
|
1533
|
-
* @description Dictionary of label: mask image
|
|
1534
|
-
* @example [
|
|
1535
|
-
* {
|
|
1536
|
-
* "height": 1200,
|
|
1537
|
-
* "file_size": 15724,
|
|
1538
|
-
* "file_name": "019c3c1e3c50446e9996f709d36debb4.png",
|
|
1539
|
-
* "content_type": "image/png",
|
|
1540
|
-
* "url": "https://v3.fal.media/files/monkey/6ITmhHQJ-69s-UxajrY5T_019c3c1e3c50446e9996f709d36debb4.png",
|
|
1541
|
-
* "width": 1800
|
|
1542
|
-
* },
|
|
1543
|
-
* {
|
|
1544
|
-
* "height": 1200,
|
|
1545
|
-
* "file_size": 14905,
|
|
1546
|
-
* "file_name": "0a1522ca410942c7ad6c73efa15b3549.png",
|
|
1547
|
-
* "content_type": "image/png",
|
|
1548
|
-
* "url": "https://v3.fal.media/files/monkey/IljtMxahoo9-7SUpx0fth_0a1522ca410942c7ad6c73efa15b3549.png",
|
|
1549
|
-
* "width": 1800
|
|
1550
|
-
* }
|
|
1551
|
-
* ]
|
|
1552
|
-
*/
|
|
1553
|
-
masks: Components.Image[];
|
|
1554
|
-
/**
|
|
1555
|
-
* Output
|
|
1556
|
-
* @description Generated output
|
|
1557
|
-
* @example <p> A white pickup truck </p> [SEG] is parked on the side of <p> the red building </p> [SEG] , creating a unique and eye-catching contrast.<|im_end|>
|
|
1558
|
-
*/
|
|
1559
|
-
output: string;
|
|
1560
|
-
}
|
|
1561
|
-
|
|
1562
1585
|
export interface SharedType_fd1 {
|
|
1563
1586
|
/**
|
|
1564
1587
|
* @description The generated video
|
|
@@ -1886,6 +1909,102 @@ export interface SharedType_f67 {
|
|
|
1886
1909
|
video_url: string;
|
|
1887
1910
|
}
|
|
1888
1911
|
|
|
1912
|
+
export interface SharedType_f66 {
|
|
1913
|
+
/**
|
|
1914
|
+
* Acceleration
|
|
1915
|
+
* @description The acceleration level to use for image generation.
|
|
1916
|
+
* @default regular
|
|
1917
|
+
* @enum {string}
|
|
1918
|
+
*/
|
|
1919
|
+
acceleration?: 'none' | 'regular' | 'high';
|
|
1920
|
+
/**
|
|
1921
|
+
* Enable Safety Checker
|
|
1922
|
+
* @description If set to true, the safety checker will be enabled.
|
|
1923
|
+
* @default true
|
|
1924
|
+
*/
|
|
1925
|
+
enable_safety_checker?: boolean;
|
|
1926
|
+
/**
|
|
1927
|
+
* Guidance Scale
|
|
1928
|
+
* @description Guidance scale for classifier-free guidance.
|
|
1929
|
+
* @default 5
|
|
1930
|
+
*/
|
|
1931
|
+
guidance_scale?: number;
|
|
1932
|
+
/**
|
|
1933
|
+
* Image Size
|
|
1934
|
+
* @description The size of the generated image. If not provided, uses the input image size.
|
|
1935
|
+
* @example {
|
|
1936
|
+
* "height": 1152,
|
|
1937
|
+
* "width": 2016
|
|
1938
|
+
* }
|
|
1939
|
+
*/
|
|
1940
|
+
image_size?:
|
|
1941
|
+
| Components.ImageSize
|
|
1942
|
+
| (
|
|
1943
|
+
| 'square_hd'
|
|
1944
|
+
| 'square'
|
|
1945
|
+
| 'portrait_4_3'
|
|
1946
|
+
| 'portrait_16_9'
|
|
1947
|
+
| 'landscape_4_3'
|
|
1948
|
+
| 'landscape_16_9'
|
|
1949
|
+
);
|
|
1950
|
+
/**
|
|
1951
|
+
* Image URLs
|
|
1952
|
+
* @description The URLs of the images for editing. A maximum of 4 images are allowed.
|
|
1953
|
+
* @example [
|
|
1954
|
+
* "https://storage.googleapis.com/falserverless/example_inputs/flux2_dev_edit_input.png"
|
|
1955
|
+
* ]
|
|
1956
|
+
*/
|
|
1957
|
+
image_urls: string[];
|
|
1958
|
+
/**
|
|
1959
|
+
* Loras
|
|
1960
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
1961
|
+
* @default []
|
|
1962
|
+
*/
|
|
1963
|
+
loras?: Components.LoRAInput[];
|
|
1964
|
+
/**
|
|
1965
|
+
* Negative Prompt
|
|
1966
|
+
* @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
|
|
1967
|
+
* @default
|
|
1968
|
+
*/
|
|
1969
|
+
negative_prompt?: string;
|
|
1970
|
+
/**
|
|
1971
|
+
* Number of Images
|
|
1972
|
+
* @description The number of images to generate.
|
|
1973
|
+
* @default 1
|
|
1974
|
+
*/
|
|
1975
|
+
num_images?: number;
|
|
1976
|
+
/**
|
|
1977
|
+
* Number of Inference Steps
|
|
1978
|
+
* @description The number of inference steps to perform.
|
|
1979
|
+
* @default 28
|
|
1980
|
+
*/
|
|
1981
|
+
num_inference_steps?: number;
|
|
1982
|
+
/**
|
|
1983
|
+
* Output Format
|
|
1984
|
+
* @description The format of the generated image.
|
|
1985
|
+
* @default png
|
|
1986
|
+
* @enum {string}
|
|
1987
|
+
*/
|
|
1988
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
1989
|
+
/**
|
|
1990
|
+
* Prompt
|
|
1991
|
+
* @description The prompt to edit the image.
|
|
1992
|
+
* @example Change his clothes to casual suit and tie
|
|
1993
|
+
*/
|
|
1994
|
+
prompt: string;
|
|
1995
|
+
/**
|
|
1996
|
+
* Seed
|
|
1997
|
+
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
1998
|
+
*/
|
|
1999
|
+
seed?: number;
|
|
2000
|
+
/**
|
|
2001
|
+
* Sync Mode
|
|
2002
|
+
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
2003
|
+
* @default false
|
|
2004
|
+
*/
|
|
2005
|
+
sync_mode?: boolean;
|
|
2006
|
+
}
|
|
2007
|
+
|
|
1889
2008
|
export interface SharedType_f63 {
|
|
1890
2009
|
/**
|
|
1891
2010
|
* Acceleration
|
|
@@ -2505,91 +2624,6 @@ export interface SharedType_e43 {
|
|
|
2505
2624
|
image: Components.Image_2;
|
|
2506
2625
|
}
|
|
2507
2626
|
|
|
2508
|
-
export interface SharedType_e34 {
|
|
2509
|
-
/**
|
|
2510
|
-
* Acceleration
|
|
2511
|
-
* @description The acceleration level to use for image generation.
|
|
2512
|
-
* @default regular
|
|
2513
|
-
* @enum {string}
|
|
2514
|
-
*/
|
|
2515
|
-
acceleration?: 'none' | 'regular' | 'high';
|
|
2516
|
-
/**
|
|
2517
|
-
* Enable Safety Checker
|
|
2518
|
-
* @description If set to true, the safety checker will be enabled.
|
|
2519
|
-
* @default true
|
|
2520
|
-
*/
|
|
2521
|
-
enable_safety_checker?: boolean;
|
|
2522
|
-
/**
|
|
2523
|
-
* Guidance Scale
|
|
2524
|
-
* @description Guidance scale for classifier-free guidance.
|
|
2525
|
-
* @default 5
|
|
2526
|
-
*/
|
|
2527
|
-
guidance_scale?: number;
|
|
2528
|
-
/**
|
|
2529
|
-
* Image Size
|
|
2530
|
-
* @description The size of the image to generate.
|
|
2531
|
-
* @default landscape_4_3
|
|
2532
|
-
*/
|
|
2533
|
-
image_size?:
|
|
2534
|
-
| Components.ImageSize
|
|
2535
|
-
| (
|
|
2536
|
-
| 'square_hd'
|
|
2537
|
-
| 'square'
|
|
2538
|
-
| 'portrait_4_3'
|
|
2539
|
-
| 'portrait_16_9'
|
|
2540
|
-
| 'landscape_4_3'
|
|
2541
|
-
| 'landscape_16_9'
|
|
2542
|
-
);
|
|
2543
|
-
/**
|
|
2544
|
-
* Loras
|
|
2545
|
-
* @description List of LoRA weights to apply (maximum 3).
|
|
2546
|
-
* @default []
|
|
2547
|
-
*/
|
|
2548
|
-
loras?: Components.LoRAInput_1[];
|
|
2549
|
-
/**
|
|
2550
|
-
* Negative Prompt
|
|
2551
|
-
* @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
|
|
2552
|
-
* @default
|
|
2553
|
-
*/
|
|
2554
|
-
negative_prompt?: string;
|
|
2555
|
-
/**
|
|
2556
|
-
* Number of Images
|
|
2557
|
-
* @description The number of images to generate.
|
|
2558
|
-
* @default 1
|
|
2559
|
-
*/
|
|
2560
|
-
num_images?: number;
|
|
2561
|
-
/**
|
|
2562
|
-
* Number of Inference Steps
|
|
2563
|
-
* @description The number of inference steps to perform.
|
|
2564
|
-
* @default 28
|
|
2565
|
-
*/
|
|
2566
|
-
num_inference_steps?: number;
|
|
2567
|
-
/**
|
|
2568
|
-
* Output Format
|
|
2569
|
-
* @description The format of the generated image.
|
|
2570
|
-
* @default png
|
|
2571
|
-
* @enum {string}
|
|
2572
|
-
*/
|
|
2573
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
2574
|
-
/**
|
|
2575
|
-
* Prompt
|
|
2576
|
-
* @description The prompt to generate an image from.
|
|
2577
|
-
* @example A serene Japanese garden with cherry blossoms, koi pond, and traditional wooden bridge at golden hour
|
|
2578
|
-
*/
|
|
2579
|
-
prompt: string;
|
|
2580
|
-
/**
|
|
2581
|
-
* Seed
|
|
2582
|
-
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
2583
|
-
*/
|
|
2584
|
-
seed?: number;
|
|
2585
|
-
/**
|
|
2586
|
-
* Sync Mode
|
|
2587
|
-
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
2588
|
-
* @default false
|
|
2589
|
-
*/
|
|
2590
|
-
sync_mode?: boolean;
|
|
2591
|
-
}
|
|
2592
|
-
|
|
2593
2627
|
export interface SharedType_e33 {
|
|
2594
2628
|
/**
|
|
2595
2629
|
* Has Nsfw Concepts
|
|
@@ -2625,29 +2659,6 @@ export interface SharedType_e33 {
|
|
|
2625
2659
|
};
|
|
2626
2660
|
}
|
|
2627
2661
|
|
|
2628
|
-
export interface SharedType_e19 {
|
|
2629
|
-
/**
|
|
2630
|
-
* Images
|
|
2631
|
-
* @description The generated images
|
|
2632
|
-
* @example [
|
|
2633
|
-
* {
|
|
2634
|
-
* "height": 1536,
|
|
2635
|
-
* "file_size": 3731290,
|
|
2636
|
-
* "file_name": "257cf8e7bd3a47c2959396343d5b38cf.png",
|
|
2637
|
-
* "content_type": "image/png",
|
|
2638
|
-
* "url": "https://v3.fal.media/files/tiger/48e63e0K6C9XQYBuomoU-_257cf8e7bd3a47c2959396343d5b38cf.png",
|
|
2639
|
-
* "width": 1536
|
|
2640
|
-
* }
|
|
2641
|
-
* ]
|
|
2642
|
-
*/
|
|
2643
|
-
images: Components.Image[];
|
|
2644
|
-
/**
|
|
2645
|
-
* Seed
|
|
2646
|
-
* @description Seed value used for generation.
|
|
2647
|
-
*/
|
|
2648
|
-
seed: number;
|
|
2649
|
-
}
|
|
2650
|
-
|
|
2651
2662
|
export interface SharedType_e18 {
|
|
2652
2663
|
/**
|
|
2653
2664
|
* Default Caption
|
|
@@ -3234,24 +3245,36 @@ export interface SharedType_d94 {
|
|
|
3234
3245
|
video: Components.File;
|
|
3235
3246
|
}
|
|
3236
3247
|
|
|
3237
|
-
export interface
|
|
3248
|
+
export interface SharedType_d72 {
|
|
3238
3249
|
/**
|
|
3239
|
-
*
|
|
3240
|
-
* @description
|
|
3241
|
-
* @example https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/car.jpg
|
|
3242
|
-
* @example http://ecx.images-amazon.com/images/I/51UUzBDAMsL.jpg
|
|
3250
|
+
* Has Nsfw Concepts
|
|
3251
|
+
* @description Whether the generated images contain NSFW concepts.
|
|
3243
3252
|
*/
|
|
3244
|
-
|
|
3253
|
+
has_nsfw_concepts: boolean[];
|
|
3245
3254
|
/**
|
|
3246
|
-
*
|
|
3247
|
-
* @
|
|
3248
|
-
*
|
|
3249
|
-
*
|
|
3250
|
-
*
|
|
3251
|
-
*
|
|
3252
|
-
*
|
|
3255
|
+
* Images
|
|
3256
|
+
* @description The generated images
|
|
3257
|
+
* @example [
|
|
3258
|
+
* {
|
|
3259
|
+
* "url": "https://v3b.fal.media/files/b/0a8a69c7/gU9ZgfFC9oAZjpIoveAac.png"
|
|
3260
|
+
* }
|
|
3261
|
+
* ]
|
|
3253
3262
|
*/
|
|
3254
|
-
|
|
3263
|
+
images: Components.ImageFile[];
|
|
3264
|
+
/**
|
|
3265
|
+
* Prompt
|
|
3266
|
+
* @description The prompt used for generating the image.
|
|
3267
|
+
*/
|
|
3268
|
+
prompt: string;
|
|
3269
|
+
/**
|
|
3270
|
+
* Seed
|
|
3271
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
3272
|
+
* input or the randomly generated that was used in case none was passed.
|
|
3273
|
+
*/
|
|
3274
|
+
seed: number;
|
|
3275
|
+
timings: {
|
|
3276
|
+
[key: string]: number;
|
|
3277
|
+
};
|
|
3255
3278
|
}
|
|
3256
3279
|
|
|
3257
3280
|
export interface SharedType_d66 {
|
|
@@ -3494,20 +3517,14 @@ export interface SharedType_d35 {
|
|
|
3494
3517
|
seed?: number;
|
|
3495
3518
|
}
|
|
3496
3519
|
|
|
3497
|
-
export interface
|
|
3498
|
-
/**
|
|
3499
|
-
* Seed
|
|
3500
|
-
* @description The seed used for generation.
|
|
3501
|
-
*/
|
|
3502
|
-
seed: number;
|
|
3520
|
+
export interface SharedType_d2a {
|
|
3503
3521
|
/**
|
|
3504
|
-
*
|
|
3505
|
-
* @description The generated video file.
|
|
3522
|
+
* @description The generated speech audio
|
|
3506
3523
|
* @example {
|
|
3507
|
-
* "url": "https://
|
|
3524
|
+
* "url": "https://v3.fal.media/files/kangaroo/RQ_pxc7oPdueYqWUqEbPE_tmpjnzvvzx_.wav"
|
|
3508
3525
|
* }
|
|
3509
3526
|
*/
|
|
3510
|
-
|
|
3527
|
+
audio: Components.File;
|
|
3511
3528
|
}
|
|
3512
3529
|
|
|
3513
3530
|
export interface SharedType_d1a {
|
|
@@ -3863,6 +3880,35 @@ export interface SharedType_cbd {
|
|
|
3863
3880
|
sync_mode?: boolean;
|
|
3864
3881
|
}
|
|
3865
3882
|
|
|
3883
|
+
export interface SharedType_cb6 {
|
|
3884
|
+
/**
|
|
3885
|
+
* Prompt
|
|
3886
|
+
* @description The prompt used for the generation.
|
|
3887
|
+
* @example A woman stands still amid a busy neon-lit street at night. The camera slowly dollies in toward her face as people blur past, their motion emphasizing her calm presence. City lights flicker and reflections shift across her denim jacket.
|
|
3888
|
+
*/
|
|
3889
|
+
prompt: string;
|
|
3890
|
+
/**
|
|
3891
|
+
* Seed
|
|
3892
|
+
* @description The seed used for the random number generator.
|
|
3893
|
+
* @example 2078003885
|
|
3894
|
+
*/
|
|
3895
|
+
seed: number;
|
|
3896
|
+
/**
|
|
3897
|
+
* @description The generated video.
|
|
3898
|
+
* @example {
|
|
3899
|
+
* "height": 704,
|
|
3900
|
+
* "duration": 6.44,
|
|
3901
|
+
* "url": "https://v3b.fal.media/files/b/0a894013/N9lnMTq7W3uMC0lOQg845_BknRPV8I.mp4",
|
|
3902
|
+
* "fps": 25,
|
|
3903
|
+
* "width": 1248,
|
|
3904
|
+
* "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
|
|
3905
|
+
* "num_frames": 161,
|
|
3906
|
+
* "content_type": "video/mp4"
|
|
3907
|
+
* }
|
|
3908
|
+
*/
|
|
3909
|
+
video: Components.VideoFile;
|
|
3910
|
+
}
|
|
3911
|
+
|
|
3866
3912
|
export interface SharedType_cb4 {
|
|
3867
3913
|
/**
|
|
3868
3914
|
* Acceleration
|
|
@@ -4466,6 +4512,91 @@ export interface SharedType_bb3 {
|
|
|
4466
4512
|
resolution?: '540p' | '720p' | '1080p';
|
|
4467
4513
|
}
|
|
4468
4514
|
|
|
4515
|
+
export interface SharedType_bb2 {
|
|
4516
|
+
/**
|
|
4517
|
+
* Acceleration
|
|
4518
|
+
* @description The acceleration level to use for image generation.
|
|
4519
|
+
* @default regular
|
|
4520
|
+
* @enum {string}
|
|
4521
|
+
*/
|
|
4522
|
+
acceleration?: 'none' | 'regular' | 'high';
|
|
4523
|
+
/**
|
|
4524
|
+
* Enable Safety Checker
|
|
4525
|
+
* @description If set to true, the safety checker will be enabled.
|
|
4526
|
+
* @default true
|
|
4527
|
+
*/
|
|
4528
|
+
enable_safety_checker?: boolean;
|
|
4529
|
+
/**
|
|
4530
|
+
* Guidance Scale
|
|
4531
|
+
* @description Guidance scale for classifier-free guidance.
|
|
4532
|
+
* @default 5
|
|
4533
|
+
*/
|
|
4534
|
+
guidance_scale?: number;
|
|
4535
|
+
/**
|
|
4536
|
+
* Image Size
|
|
4537
|
+
* @description The size of the image to generate.
|
|
4538
|
+
* @default landscape_4_3
|
|
4539
|
+
*/
|
|
4540
|
+
image_size?:
|
|
4541
|
+
| Components.ImageSize
|
|
4542
|
+
| (
|
|
4543
|
+
| 'square_hd'
|
|
4544
|
+
| 'square'
|
|
4545
|
+
| 'portrait_4_3'
|
|
4546
|
+
| 'portrait_16_9'
|
|
4547
|
+
| 'landscape_4_3'
|
|
4548
|
+
| 'landscape_16_9'
|
|
4549
|
+
);
|
|
4550
|
+
/**
|
|
4551
|
+
* Loras
|
|
4552
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
4553
|
+
* @default []
|
|
4554
|
+
*/
|
|
4555
|
+
loras?: Components.LoRAInput[];
|
|
4556
|
+
/**
|
|
4557
|
+
* Negative Prompt
|
|
4558
|
+
* @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
|
|
4559
|
+
* @default
|
|
4560
|
+
*/
|
|
4561
|
+
negative_prompt?: string;
|
|
4562
|
+
/**
|
|
4563
|
+
* Number of Images
|
|
4564
|
+
* @description The number of images to generate.
|
|
4565
|
+
* @default 1
|
|
4566
|
+
*/
|
|
4567
|
+
num_images?: number;
|
|
4568
|
+
/**
|
|
4569
|
+
* Number of Inference Steps
|
|
4570
|
+
* @description The number of inference steps to perform.
|
|
4571
|
+
* @default 28
|
|
4572
|
+
*/
|
|
4573
|
+
num_inference_steps?: number;
|
|
4574
|
+
/**
|
|
4575
|
+
* Output Format
|
|
4576
|
+
* @description The format of the generated image.
|
|
4577
|
+
* @default png
|
|
4578
|
+
* @enum {string}
|
|
4579
|
+
*/
|
|
4580
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
4581
|
+
/**
|
|
4582
|
+
* Prompt
|
|
4583
|
+
* @description The prompt to generate an image from.
|
|
4584
|
+
* @example A serene Japanese garden with cherry blossoms, koi pond, and traditional wooden bridge at golden hour
|
|
4585
|
+
*/
|
|
4586
|
+
prompt: string;
|
|
4587
|
+
/**
|
|
4588
|
+
* Seed
|
|
4589
|
+
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
4590
|
+
*/
|
|
4591
|
+
seed?: number;
|
|
4592
|
+
/**
|
|
4593
|
+
* Sync Mode
|
|
4594
|
+
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
4595
|
+
* @default false
|
|
4596
|
+
*/
|
|
4597
|
+
sync_mode?: boolean;
|
|
4598
|
+
}
|
|
4599
|
+
|
|
4469
4600
|
export interface SharedType_bb0 {
|
|
4470
4601
|
/**
|
|
4471
4602
|
* Image URL
|
|
@@ -4612,76 +4743,6 @@ export interface SharedType_b89 {
|
|
|
4612
4743
|
image: Components.Image;
|
|
4613
4744
|
}
|
|
4614
4745
|
|
|
4615
|
-
export interface SharedType_b7d {
|
|
4616
|
-
/**
|
|
4617
|
-
* Enable Safety Checker
|
|
4618
|
-
* @description If set to true, the safety checker will be enabled.
|
|
4619
|
-
* @default true
|
|
4620
|
-
*/
|
|
4621
|
-
enable_safety_checker?: boolean;
|
|
4622
|
-
/**
|
|
4623
|
-
* Guidance scale (CFG)
|
|
4624
|
-
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
4625
|
-
* the model to stick to your prompt when looking for a related image to show you.
|
|
4626
|
-
* @default 3.5
|
|
4627
|
-
*/
|
|
4628
|
-
guidance_scale?: number;
|
|
4629
|
-
/**
|
|
4630
|
-
* Image Size
|
|
4631
|
-
* @description The size of the generated image.
|
|
4632
|
-
* @default landscape_4_3
|
|
4633
|
-
*/
|
|
4634
|
-
image_size?:
|
|
4635
|
-
| Components.ImageSize
|
|
4636
|
-
| (
|
|
4637
|
-
| 'square_hd'
|
|
4638
|
-
| 'square'
|
|
4639
|
-
| 'portrait_4_3'
|
|
4640
|
-
| 'portrait_16_9'
|
|
4641
|
-
| 'landscape_4_3'
|
|
4642
|
-
| 'landscape_16_9'
|
|
4643
|
-
);
|
|
4644
|
-
/**
|
|
4645
|
-
* Negative prompt
|
|
4646
|
-
* @description Negative Prompt for generation.
|
|
4647
|
-
* @default
|
|
4648
|
-
* @example Blurry, out of focus, low resolution, bad anatomy, ugly, deformed, poorly drawn, extra limbs
|
|
4649
|
-
*/
|
|
4650
|
-
negative_prompt?: string;
|
|
4651
|
-
/**
|
|
4652
|
-
* Num Images
|
|
4653
|
-
* @description The number of images to generate.
|
|
4654
|
-
* @default 1
|
|
4655
|
-
*/
|
|
4656
|
-
num_images?: number;
|
|
4657
|
-
/**
|
|
4658
|
-
* Num Inference Steps
|
|
4659
|
-
* @description The number of inference steps to perform.
|
|
4660
|
-
* @default 28
|
|
4661
|
-
*/
|
|
4662
|
-
num_inference_steps?: number;
|
|
4663
|
-
/**
|
|
4664
|
-
* Prompt
|
|
4665
|
-
* @description The prompt to generate an image from.
|
|
4666
|
-
* @example Mount Fuji at sunset, with the iconic snow-capped peak silhouetted against a vibrant orange and purple sky. A tranquil lake in the foreground perfectly reflects the mountain and colorful sky. A few traditional Japanese cherry blossom trees frame the scene, with their delicate pink petals visible in the foreground.
|
|
4667
|
-
*/
|
|
4668
|
-
prompt: string;
|
|
4669
|
-
/**
|
|
4670
|
-
* Seed
|
|
4671
|
-
* @description The same seed and the same prompt given to the same version of the model
|
|
4672
|
-
* will output the same image every time.
|
|
4673
|
-
*/
|
|
4674
|
-
seed?: number;
|
|
4675
|
-
/**
|
|
4676
|
-
* Sync Mode
|
|
4677
|
-
* @description If set to true, the function will wait for the image to be generated and uploaded
|
|
4678
|
-
* before returning the response. This will increase the latency of the function but
|
|
4679
|
-
* it allows you to get the image directly in the response without going through the CDN.
|
|
4680
|
-
* @default false
|
|
4681
|
-
*/
|
|
4682
|
-
sync_mode?: boolean;
|
|
4683
|
-
}
|
|
4684
|
-
|
|
4685
4746
|
export interface SharedType_b71 {
|
|
4686
4747
|
model_mesh: Components.File;
|
|
4687
4748
|
/**
|
|
@@ -4831,6 +4892,16 @@ export interface SharedType_b35 {
|
|
|
4831
4892
|
sync_mode?: boolean;
|
|
4832
4893
|
}
|
|
4833
4894
|
|
|
4895
|
+
export interface SharedType_b23 {
|
|
4896
|
+
/**
|
|
4897
|
+
* Seed
|
|
4898
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
4899
|
+
* input or the randomly generated that was used in case none was passed.
|
|
4900
|
+
*/
|
|
4901
|
+
seed: number;
|
|
4902
|
+
video: Components.File;
|
|
4903
|
+
}
|
|
4904
|
+
|
|
4834
4905
|
export interface SharedType_b08 {
|
|
4835
4906
|
/**
|
|
4836
4907
|
* Acceleration
|
|
@@ -5159,6 +5230,29 @@ export interface SharedType_a9b {
|
|
|
5159
5230
|
seed?: number;
|
|
5160
5231
|
}
|
|
5161
5232
|
|
|
5233
|
+
export interface SharedType_a97 {
|
|
5234
|
+
/**
|
|
5235
|
+
* Images
|
|
5236
|
+
* @description The generated images
|
|
5237
|
+
* @example [
|
|
5238
|
+
* {
|
|
5239
|
+
* "file_size": 3731290,
|
|
5240
|
+
* "height": 1536,
|
|
5241
|
+
* "file_name": "257cf8e7bd3a47c2959396343d5b38cf.png",
|
|
5242
|
+
* "content_type": "image/png",
|
|
5243
|
+
* "url": "https://v3.fal.media/files/tiger/48e63e0K6C9XQYBuomoU-_257cf8e7bd3a47c2959396343d5b38cf.png",
|
|
5244
|
+
* "width": 1536
|
|
5245
|
+
* }
|
|
5246
|
+
* ]
|
|
5247
|
+
*/
|
|
5248
|
+
images: Components.Image[];
|
|
5249
|
+
/**
|
|
5250
|
+
* Seed
|
|
5251
|
+
* @description Seed value used for generation.
|
|
5252
|
+
*/
|
|
5253
|
+
seed: number;
|
|
5254
|
+
}
|
|
5255
|
+
|
|
5162
5256
|
export interface SharedType_a8f {
|
|
5163
5257
|
/**
|
|
5164
5258
|
* Image Url
|
|
@@ -5224,6 +5318,38 @@ export interface SharedType_a77 {
|
|
|
5224
5318
|
audio_url: string;
|
|
5225
5319
|
}
|
|
5226
5320
|
|
|
5321
|
+
export interface SharedType_a75 {
|
|
5322
|
+
/**
|
|
5323
|
+
* Has Nsfw Concepts
|
|
5324
|
+
* @description Whether the generated images contain NSFW concepts.
|
|
5325
|
+
*/
|
|
5326
|
+
has_nsfw_concepts: boolean[];
|
|
5327
|
+
/**
|
|
5328
|
+
* Images
|
|
5329
|
+
* @description The generated images
|
|
5330
|
+
* @example [
|
|
5331
|
+
* {
|
|
5332
|
+
* "url": "https://v3b.fal.media/files/b/0a8a6905/DuNRy1OODaGrlUGEfl9SX.png"
|
|
5333
|
+
* }
|
|
5334
|
+
* ]
|
|
5335
|
+
*/
|
|
5336
|
+
images: Components.ImageFile[];
|
|
5337
|
+
/**
|
|
5338
|
+
* Prompt
|
|
5339
|
+
* @description The prompt used for generating the image.
|
|
5340
|
+
*/
|
|
5341
|
+
prompt: string;
|
|
5342
|
+
/**
|
|
5343
|
+
* Seed
|
|
5344
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
5345
|
+
* input or the randomly generated that was used in case none was passed.
|
|
5346
|
+
*/
|
|
5347
|
+
seed: number;
|
|
5348
|
+
timings: {
|
|
5349
|
+
[key: string]: number;
|
|
5350
|
+
};
|
|
5351
|
+
}
|
|
5352
|
+
|
|
5227
5353
|
export interface SharedType_a58 {
|
|
5228
5354
|
/**
|
|
5229
5355
|
* Duration
|
|
@@ -5427,35 +5553,6 @@ export interface SharedType_a1f {
|
|
|
5427
5553
|
};
|
|
5428
5554
|
}
|
|
5429
5555
|
|
|
5430
|
-
export interface SharedType_a00 {
|
|
5431
|
-
/**
|
|
5432
|
-
* Prompt
|
|
5433
|
-
* @description The prompt used for the generation.
|
|
5434
|
-
* @example Continue the scene naturally, maintaining the same style and motion.
|
|
5435
|
-
*/
|
|
5436
|
-
prompt: string;
|
|
5437
|
-
/**
|
|
5438
|
-
* Seed
|
|
5439
|
-
* @description The seed used for the random number generator.
|
|
5440
|
-
* @example 866232447
|
|
5441
|
-
*/
|
|
5442
|
-
seed: number;
|
|
5443
|
-
/**
|
|
5444
|
-
* @description The generated video.
|
|
5445
|
-
* @example {
|
|
5446
|
-
* "height": 704,
|
|
5447
|
-
* "duration": 10.28,
|
|
5448
|
-
* "url": "https://v3b.fal.media/files/b/0a88289e/CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
|
|
5449
|
-
* "fps": 25,
|
|
5450
|
-
* "width": 1248,
|
|
5451
|
-
* "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
|
|
5452
|
-
* "content_type": "video/mp4",
|
|
5453
|
-
* "num_frames": 257
|
|
5454
|
-
* }
|
|
5455
|
-
*/
|
|
5456
|
-
video: Components.VideoFile;
|
|
5457
|
-
}
|
|
5458
|
-
|
|
5459
5556
|
export interface SharedType_9f4 {
|
|
5460
5557
|
/**
|
|
5461
5558
|
* Audio Url
|
|
@@ -5858,12 +5955,81 @@ export interface SharedType_97e {
|
|
|
5858
5955
|
voice_setting?: Components.VoiceSetting;
|
|
5859
5956
|
}
|
|
5860
5957
|
|
|
5861
|
-
export interface
|
|
5958
|
+
export interface SharedType_974 {
|
|
5862
5959
|
/**
|
|
5863
|
-
*
|
|
5864
|
-
* @description
|
|
5960
|
+
* Enable Safety Checker
|
|
5961
|
+
* @description If set to true, the safety checker will be enabled.
|
|
5962
|
+
* @default true
|
|
5865
5963
|
*/
|
|
5866
|
-
|
|
5964
|
+
enable_safety_checker?: boolean;
|
|
5965
|
+
/**
|
|
5966
|
+
* Image Size
|
|
5967
|
+
* @description The size of the generated image. If not provided, uses the input image size.
|
|
5968
|
+
* @example {
|
|
5969
|
+
* "height": 1152,
|
|
5970
|
+
* "width": 2016
|
|
5971
|
+
* }
|
|
5972
|
+
*/
|
|
5973
|
+
image_size?:
|
|
5974
|
+
| Components.ImageSize
|
|
5975
|
+
| (
|
|
5976
|
+
| 'square_hd'
|
|
5977
|
+
| 'square'
|
|
5978
|
+
| 'portrait_4_3'
|
|
5979
|
+
| 'portrait_16_9'
|
|
5980
|
+
| 'landscape_4_3'
|
|
5981
|
+
| 'landscape_16_9'
|
|
5982
|
+
);
|
|
5983
|
+
/**
|
|
5984
|
+
* Image URLs
|
|
5985
|
+
* @description The URLs of the images for editing. A maximum of 4 images are allowed.
|
|
5986
|
+
* @example [
|
|
5987
|
+
* "https://v3b.fal.media/files/b/0a8a69d5/kkXxFfj1QeVtw35kxy5Py_1a7e3511-bd2c-46be-923a-8e6be2496f12.png"
|
|
5988
|
+
* ]
|
|
5989
|
+
*/
|
|
5990
|
+
image_urls: string[];
|
|
5991
|
+
/**
|
|
5992
|
+
* Loras
|
|
5993
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
5994
|
+
* @default []
|
|
5995
|
+
*/
|
|
5996
|
+
loras?: Components.LoRAInput[];
|
|
5997
|
+
/**
|
|
5998
|
+
* Number of Images
|
|
5999
|
+
* @description The number of images to generate.
|
|
6000
|
+
* @default 1
|
|
6001
|
+
*/
|
|
6002
|
+
num_images?: number;
|
|
6003
|
+
/**
|
|
6004
|
+
* Number of Inference Steps
|
|
6005
|
+
* @description The number of inference steps to perform.
|
|
6006
|
+
* @default 4
|
|
6007
|
+
*/
|
|
6008
|
+
num_inference_steps?: number;
|
|
6009
|
+
/**
|
|
6010
|
+
* Output Format
|
|
6011
|
+
* @description The format of the generated image.
|
|
6012
|
+
* @default png
|
|
6013
|
+
* @enum {string}
|
|
6014
|
+
*/
|
|
6015
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
6016
|
+
/**
|
|
6017
|
+
* Prompt
|
|
6018
|
+
* @description The prompt to edit the image.
|
|
6019
|
+
* @example Turn this into a realistic image
|
|
6020
|
+
*/
|
|
6021
|
+
prompt: string;
|
|
6022
|
+
/**
|
|
6023
|
+
* Seed
|
|
6024
|
+
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
6025
|
+
*/
|
|
6026
|
+
seed?: number;
|
|
6027
|
+
/**
|
|
6028
|
+
* Sync Mode
|
|
6029
|
+
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
6030
|
+
* @default false
|
|
6031
|
+
*/
|
|
6032
|
+
sync_mode?: boolean;
|
|
5867
6033
|
}
|
|
5868
6034
|
|
|
5869
6035
|
export interface SharedType_95d {
|
|
@@ -5979,6 +6145,38 @@ export interface SharedType_94e {
|
|
|
5979
6145
|
video: Components.File;
|
|
5980
6146
|
}
|
|
5981
6147
|
|
|
6148
|
+
export interface SharedType_938 {
|
|
6149
|
+
/**
|
|
6150
|
+
* Has Nsfw Concepts
|
|
6151
|
+
* @description Whether the generated images contain NSFW concepts.
|
|
6152
|
+
*/
|
|
6153
|
+
has_nsfw_concepts: boolean[];
|
|
6154
|
+
/**
|
|
6155
|
+
* Images
|
|
6156
|
+
* @description The edited images
|
|
6157
|
+
* @example [
|
|
6158
|
+
* {
|
|
6159
|
+
* "url": "https://v3b.fal.media/files/b/0a8a69d6/M73KvDgfEgIM77t4mFsS2.png"
|
|
6160
|
+
* }
|
|
6161
|
+
* ]
|
|
6162
|
+
*/
|
|
6163
|
+
images: Components.ImageFile[];
|
|
6164
|
+
/**
|
|
6165
|
+
* Prompt
|
|
6166
|
+
* @description The prompt used for generating the image.
|
|
6167
|
+
*/
|
|
6168
|
+
prompt: string;
|
|
6169
|
+
/**
|
|
6170
|
+
* Seed
|
|
6171
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
6172
|
+
* input or the randomly generated that was used in case none was passed.
|
|
6173
|
+
*/
|
|
6174
|
+
seed: number;
|
|
6175
|
+
timings: {
|
|
6176
|
+
[key: string]: number;
|
|
6177
|
+
};
|
|
6178
|
+
}
|
|
6179
|
+
|
|
5982
6180
|
export interface SharedType_92f {
|
|
5983
6181
|
/**
|
|
5984
6182
|
* Has Nsfw Concepts
|
|
@@ -6825,6 +7023,35 @@ export interface SharedType_85d {
|
|
|
6825
7023
|
sync_mode?: boolean;
|
|
6826
7024
|
}
|
|
6827
7025
|
|
|
7026
|
+
export interface SharedType_857 {
|
|
7027
|
+
/**
|
|
7028
|
+
* Prompt
|
|
7029
|
+
* @description The prompt used for the generation.
|
|
7030
|
+
* @example A cowboy walking through a dusty town at high noon, camera following from behind, cinematic depth, realistic lighting, western mood, 4K film grain.
|
|
7031
|
+
*/
|
|
7032
|
+
prompt: string;
|
|
7033
|
+
/**
|
|
7034
|
+
* Seed
|
|
7035
|
+
* @description The seed used for the random number generator.
|
|
7036
|
+
* @example 149063119
|
|
7037
|
+
*/
|
|
7038
|
+
seed: number;
|
|
7039
|
+
/**
|
|
7040
|
+
* @description The generated video.
|
|
7041
|
+
* @example {
|
|
7042
|
+
* "height": 704,
|
|
7043
|
+
* "duration": 6.44,
|
|
7044
|
+
* "url": "https://v3b.fal.media/files/b/0a8824b1/sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
|
|
7045
|
+
* "fps": 25,
|
|
7046
|
+
* "width": 1248,
|
|
7047
|
+
* "file_name": "sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
|
|
7048
|
+
* "num_frames": 161,
|
|
7049
|
+
* "content_type": "video/mp4"
|
|
7050
|
+
* }
|
|
7051
|
+
*/
|
|
7052
|
+
video: Components.VideoFile;
|
|
7053
|
+
}
|
|
7054
|
+
|
|
6828
7055
|
export interface SharedType_844 {
|
|
6829
7056
|
/**
|
|
6830
7057
|
* Prompt
|
|
@@ -7780,6 +8007,10 @@ export interface SharedType_7c6 {
|
|
|
7780
8007
|
};
|
|
7781
8008
|
}
|
|
7782
8009
|
|
|
8010
|
+
export interface SharedType_7c2 {
|
|
8011
|
+
lottie_file: Components.File;
|
|
8012
|
+
}
|
|
8013
|
+
|
|
7783
8014
|
export interface SharedType_7c0 {
|
|
7784
8015
|
/**
|
|
7785
8016
|
* @description Generated 3D object file.
|
|
@@ -8124,102 +8355,6 @@ export interface SharedType_709 {
|
|
|
8124
8355
|
video_url: string;
|
|
8125
8356
|
}
|
|
8126
8357
|
|
|
8127
|
-
export interface SharedType_6f9 {
|
|
8128
|
-
/**
|
|
8129
|
-
* Acceleration
|
|
8130
|
-
* @description The acceleration level to use for image generation.
|
|
8131
|
-
* @default regular
|
|
8132
|
-
* @enum {string}
|
|
8133
|
-
*/
|
|
8134
|
-
acceleration?: 'none' | 'regular' | 'high';
|
|
8135
|
-
/**
|
|
8136
|
-
* Enable Safety Checker
|
|
8137
|
-
* @description If set to true, the safety checker will be enabled.
|
|
8138
|
-
* @default true
|
|
8139
|
-
*/
|
|
8140
|
-
enable_safety_checker?: boolean;
|
|
8141
|
-
/**
|
|
8142
|
-
* Guidance Scale
|
|
8143
|
-
* @description Guidance scale for classifier-free guidance.
|
|
8144
|
-
* @default 5
|
|
8145
|
-
*/
|
|
8146
|
-
guidance_scale?: number;
|
|
8147
|
-
/**
|
|
8148
|
-
* Image Size
|
|
8149
|
-
* @description The size of the generated image. If not provided, uses the input image size.
|
|
8150
|
-
* @example {
|
|
8151
|
-
* "height": 1152,
|
|
8152
|
-
* "width": 2016
|
|
8153
|
-
* }
|
|
8154
|
-
*/
|
|
8155
|
-
image_size?:
|
|
8156
|
-
| Components.ImageSize
|
|
8157
|
-
| (
|
|
8158
|
-
| 'square_hd'
|
|
8159
|
-
| 'square'
|
|
8160
|
-
| 'portrait_4_3'
|
|
8161
|
-
| 'portrait_16_9'
|
|
8162
|
-
| 'landscape_4_3'
|
|
8163
|
-
| 'landscape_16_9'
|
|
8164
|
-
);
|
|
8165
|
-
/**
|
|
8166
|
-
* Image URLs
|
|
8167
|
-
* @description The URLs of the images for editing. A maximum of 4 images are allowed.
|
|
8168
|
-
* @example [
|
|
8169
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/flux2_dev_edit_input.png"
|
|
8170
|
-
* ]
|
|
8171
|
-
*/
|
|
8172
|
-
image_urls: string[];
|
|
8173
|
-
/**
|
|
8174
|
-
* Loras
|
|
8175
|
-
* @description List of LoRA weights to apply (maximum 3).
|
|
8176
|
-
* @default []
|
|
8177
|
-
*/
|
|
8178
|
-
loras?: Components.LoRAInput_1[];
|
|
8179
|
-
/**
|
|
8180
|
-
* Negative Prompt
|
|
8181
|
-
* @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
|
|
8182
|
-
* @default
|
|
8183
|
-
*/
|
|
8184
|
-
negative_prompt?: string;
|
|
8185
|
-
/**
|
|
8186
|
-
* Number of Images
|
|
8187
|
-
* @description The number of images to generate.
|
|
8188
|
-
* @default 1
|
|
8189
|
-
*/
|
|
8190
|
-
num_images?: number;
|
|
8191
|
-
/**
|
|
8192
|
-
* Number of Inference Steps
|
|
8193
|
-
* @description The number of inference steps to perform.
|
|
8194
|
-
* @default 28
|
|
8195
|
-
*/
|
|
8196
|
-
num_inference_steps?: number;
|
|
8197
|
-
/**
|
|
8198
|
-
* Output Format
|
|
8199
|
-
* @description The format of the generated image.
|
|
8200
|
-
* @default png
|
|
8201
|
-
* @enum {string}
|
|
8202
|
-
*/
|
|
8203
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
8204
|
-
/**
|
|
8205
|
-
* Prompt
|
|
8206
|
-
* @description The prompt to edit the image.
|
|
8207
|
-
* @example Change his clothes to casual suit and tie
|
|
8208
|
-
*/
|
|
8209
|
-
prompt: string;
|
|
8210
|
-
/**
|
|
8211
|
-
* Seed
|
|
8212
|
-
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
8213
|
-
*/
|
|
8214
|
-
seed?: number;
|
|
8215
|
-
/**
|
|
8216
|
-
* Sync Mode
|
|
8217
|
-
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
8218
|
-
* @default false
|
|
8219
|
-
*/
|
|
8220
|
-
sync_mode?: boolean;
|
|
8221
|
-
}
|
|
8222
|
-
|
|
8223
8358
|
export interface SharedType_6cd {
|
|
8224
8359
|
/**
|
|
8225
8360
|
* Aspect Ratio
|
|
@@ -8821,35 +8956,6 @@ export interface SharedType_62e {
|
|
|
8821
8956
|
video: Components.VideoFile;
|
|
8822
8957
|
}
|
|
8823
8958
|
|
|
8824
|
-
export interface SharedType_61b {
|
|
8825
|
-
/**
|
|
8826
|
-
* Prompt
|
|
8827
|
-
* @description The prompt used for the generation.
|
|
8828
|
-
* @example A cowboy walking through a dusty town at high noon, camera following from behind, cinematic depth, realistic lighting, western mood, 4K film grain.
|
|
8829
|
-
*/
|
|
8830
|
-
prompt: string;
|
|
8831
|
-
/**
|
|
8832
|
-
* Seed
|
|
8833
|
-
* @description The seed used for the random number generator.
|
|
8834
|
-
* @example 149063119
|
|
8835
|
-
*/
|
|
8836
|
-
seed: number;
|
|
8837
|
-
/**
|
|
8838
|
-
* @description The generated video.
|
|
8839
|
-
* @example {
|
|
8840
|
-
* "height": 704,
|
|
8841
|
-
* "duration": 6.44,
|
|
8842
|
-
* "url": "https://v3b.fal.media/files/b/0a8824b1/sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
|
|
8843
|
-
* "fps": 25,
|
|
8844
|
-
* "width": 1248,
|
|
8845
|
-
* "file_name": "sdm0KfmenrlywesfzY1Y1_if6euPp1.mp4",
|
|
8846
|
-
* "content_type": "video/mp4",
|
|
8847
|
-
* "num_frames": 161
|
|
8848
|
-
* }
|
|
8849
|
-
*/
|
|
8850
|
-
video: Components.VideoFile;
|
|
8851
|
-
}
|
|
8852
|
-
|
|
8853
8959
|
export interface SharedType_618 {
|
|
8854
8960
|
/**
|
|
8855
8961
|
* Video
|
|
@@ -9264,6 +9370,24 @@ export interface SharedType_595 {
|
|
|
9264
9370
|
video: Components.File_1;
|
|
9265
9371
|
}
|
|
9266
9372
|
|
|
9373
|
+
export interface SharedType_58d {
|
|
9374
|
+
/**
|
|
9375
|
+
* Seed
|
|
9376
|
+
* @description The seed used for generation.
|
|
9377
|
+
*/
|
|
9378
|
+
seed: number;
|
|
9379
|
+
/**
|
|
9380
|
+
* @description The generated video file.
|
|
9381
|
+
* @example {
|
|
9382
|
+
* "file_size": 797478,
|
|
9383
|
+
* "file_name": "6c9dd31e1d9a4482877747a52a661a0a.mp4",
|
|
9384
|
+
* "content_type": "application/octet-stream",
|
|
9385
|
+
* "url": "https://v3.fal.media/files/elephant/-huMN0zTaXmBr2CqzCMps_6c9dd31e1d9a4482877747a52a661a0a.mp4"
|
|
9386
|
+
* }
|
|
9387
|
+
*/
|
|
9388
|
+
video: Components.File;
|
|
9389
|
+
}
|
|
9390
|
+
|
|
9267
9391
|
export interface SharedType_57e {
|
|
9268
9392
|
/**
|
|
9269
9393
|
* Acceleration
|
|
@@ -9528,35 +9652,6 @@ export interface SharedType_576 {
|
|
|
9528
9652
|
};
|
|
9529
9653
|
}
|
|
9530
9654
|
|
|
9531
|
-
export interface SharedType_567 {
|
|
9532
|
-
/**
|
|
9533
|
-
* Prompt
|
|
9534
|
-
* @description The prompt used for the generation.
|
|
9535
|
-
* @example A woman stands still amid a busy neon-lit street at night. The camera slowly dollies in toward her face as people blur past, their motion emphasizing her calm presence. City lights flicker and reflections shift across her denim jacket.
|
|
9536
|
-
*/
|
|
9537
|
-
prompt: string;
|
|
9538
|
-
/**
|
|
9539
|
-
* Seed
|
|
9540
|
-
* @description The seed used for the random number generator.
|
|
9541
|
-
* @example 2078003885
|
|
9542
|
-
*/
|
|
9543
|
-
seed: number;
|
|
9544
|
-
/**
|
|
9545
|
-
* @description The generated video.
|
|
9546
|
-
* @example {
|
|
9547
|
-
* "height": 704,
|
|
9548
|
-
* "duration": 6.44,
|
|
9549
|
-
* "url": "https://v3b.fal.media/files/b/0a894013/N9lnMTq7W3uMC0lOQg845_BknRPV8I.mp4",
|
|
9550
|
-
* "fps": 25,
|
|
9551
|
-
* "width": 1248,
|
|
9552
|
-
* "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
|
|
9553
|
-
* "content_type": "video/mp4",
|
|
9554
|
-
* "num_frames": 161
|
|
9555
|
-
* }
|
|
9556
|
-
*/
|
|
9557
|
-
video: Components.VideoFile;
|
|
9558
|
-
}
|
|
9559
|
-
|
|
9560
9655
|
export interface SharedType_554 {
|
|
9561
9656
|
boundingbox_frames_zip?: Components.File;
|
|
9562
9657
|
/**
|
|
@@ -10031,35 +10126,6 @@ export interface SharedType_503 {
|
|
|
10031
10126
|
seed?: number;
|
|
10032
10127
|
}
|
|
10033
10128
|
|
|
10034
|
-
export interface SharedType_4ed {
|
|
10035
|
-
/**
|
|
10036
|
-
* Prompt
|
|
10037
|
-
* @description The prompt used for the generation.
|
|
10038
|
-
* @example black-and-white video, a cowboy walks through a dusty town, film grain
|
|
10039
|
-
*/
|
|
10040
|
-
prompt: string;
|
|
10041
|
-
/**
|
|
10042
|
-
* Seed
|
|
10043
|
-
* @description The seed used for the random number generator.
|
|
10044
|
-
* @example 1490631192028410600
|
|
10045
|
-
*/
|
|
10046
|
-
seed: number;
|
|
10047
|
-
/**
|
|
10048
|
-
* @description The generated video.
|
|
10049
|
-
* @example {
|
|
10050
|
-
* "height": 704,
|
|
10051
|
-
* "duration": 6.44,
|
|
10052
|
-
* "url": "https://v3b.fal.media/files/b/0a895ed5/SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
|
|
10053
|
-
* "fps": 25,
|
|
10054
|
-
* "width": 1248,
|
|
10055
|
-
* "file_name": "SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
|
|
10056
|
-
* "content_type": "video/mp4",
|
|
10057
|
-
* "num_frames": 161
|
|
10058
|
-
* }
|
|
10059
|
-
*/
|
|
10060
|
-
video: Components.VideoFile;
|
|
10061
|
-
}
|
|
10062
|
-
|
|
10063
10129
|
export interface SharedType_4d6 {
|
|
10064
10130
|
/**
|
|
10065
10131
|
* Has Nsfw Concepts
|
|
@@ -10523,20 +10589,6 @@ export interface SharedType_49b {
|
|
|
10523
10589
|
sync_mode?: boolean;
|
|
10524
10590
|
}
|
|
10525
10591
|
|
|
10526
|
-
export interface SharedType_484 {
|
|
10527
|
-
/**
|
|
10528
|
-
* Seed
|
|
10529
|
-
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
10530
|
-
* input or the randomly generated that was used in case none was passed.
|
|
10531
|
-
*/
|
|
10532
|
-
seed: number;
|
|
10533
|
-
/**
|
|
10534
|
-
* Video
|
|
10535
|
-
* @description The generated video file.
|
|
10536
|
-
*/
|
|
10537
|
-
video: Components.File_1;
|
|
10538
|
-
}
|
|
10539
|
-
|
|
10540
10592
|
export interface SharedType_480 {
|
|
10541
10593
|
/**
|
|
10542
10594
|
* @description The generated video from image using the Q2 model
|
|
@@ -11098,6 +11150,35 @@ export interface SharedType_3be {
|
|
|
11098
11150
|
video: Components.File;
|
|
11099
11151
|
}
|
|
11100
11152
|
|
|
11153
|
+
export interface SharedType_3b0 {
|
|
11154
|
+
/**
|
|
11155
|
+
* Prompt
|
|
11156
|
+
* @description The prompt used for the generation.
|
|
11157
|
+
* @example black-and-white video, a cowboy walks through a dusty town, film grain
|
|
11158
|
+
*/
|
|
11159
|
+
prompt: string;
|
|
11160
|
+
/**
|
|
11161
|
+
* Seed
|
|
11162
|
+
* @description The seed used for the random number generator.
|
|
11163
|
+
* @example 1490631192028410600
|
|
11164
|
+
*/
|
|
11165
|
+
seed: number;
|
|
11166
|
+
/**
|
|
11167
|
+
* @description The generated video.
|
|
11168
|
+
* @example {
|
|
11169
|
+
* "height": 704,
|
|
11170
|
+
* "duration": 6.44,
|
|
11171
|
+
* "url": "https://v3b.fal.media/files/b/0a895ed5/SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
|
|
11172
|
+
* "fps": 25,
|
|
11173
|
+
* "width": 1248,
|
|
11174
|
+
* "file_name": "SaTGe87IpMUMiSq33w5Qb_RoCJFZhc.mp4",
|
|
11175
|
+
* "num_frames": 161,
|
|
11176
|
+
* "content_type": "video/mp4"
|
|
11177
|
+
* }
|
|
11178
|
+
*/
|
|
11179
|
+
video: Components.VideoFile;
|
|
11180
|
+
}
|
|
11181
|
+
|
|
11101
11182
|
export interface SharedType_397 {
|
|
11102
11183
|
/**
|
|
11103
11184
|
* Acceleration
|
|
@@ -11309,6 +11390,44 @@ export interface SharedType_397 {
|
|
|
11309
11390
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
11310
11391
|
}
|
|
11311
11392
|
|
|
11393
|
+
export interface SharedType_390 {
|
|
11394
|
+
/**
|
|
11395
|
+
* Has Nsfw Concepts
|
|
11396
|
+
* @description Whether the generated images contain NSFW concepts.
|
|
11397
|
+
*/
|
|
11398
|
+
has_nsfw_concepts: boolean[];
|
|
11399
|
+
/**
|
|
11400
|
+
* Images
|
|
11401
|
+
* @description The generated image files info.
|
|
11402
|
+
* @example [
|
|
11403
|
+
* {
|
|
11404
|
+
* "height": 2048,
|
|
11405
|
+
* "content_type": "image/png",
|
|
11406
|
+
* "url": "https://v3b.fal.media/files/b/0a9276c3/6KxxMb0bPmfNLH3DqcuSf_qRUe8cIL.png",
|
|
11407
|
+
* "width": 2048
|
|
11408
|
+
* }
|
|
11409
|
+
* ]
|
|
11410
|
+
*/
|
|
11411
|
+
images: Components.ImageFile[];
|
|
11412
|
+
/**
|
|
11413
|
+
* Prompt
|
|
11414
|
+
* @description The prompt used for generating the image.
|
|
11415
|
+
*/
|
|
11416
|
+
prompt: string;
|
|
11417
|
+
/**
|
|
11418
|
+
* Seed
|
|
11419
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the input or the randomly generated that was used in case none was passed.
|
|
11420
|
+
*/
|
|
11421
|
+
seed: number;
|
|
11422
|
+
/**
|
|
11423
|
+
* Timings
|
|
11424
|
+
* @description The timings of the generation process.
|
|
11425
|
+
*/
|
|
11426
|
+
timings: {
|
|
11427
|
+
[key: string]: number;
|
|
11428
|
+
};
|
|
11429
|
+
}
|
|
11430
|
+
|
|
11312
11431
|
export interface SharedType_38d {
|
|
11313
11432
|
/**
|
|
11314
11433
|
* Image Url
|
|
@@ -11380,6 +11499,35 @@ export interface SharedType_371 {
|
|
|
11380
11499
|
prompt_optimizer?: boolean;
|
|
11381
11500
|
}
|
|
11382
11501
|
|
|
11502
|
+
export interface SharedType_369 {
|
|
11503
|
+
/**
|
|
11504
|
+
* Prompt
|
|
11505
|
+
* @description The prompt used for the generation.
|
|
11506
|
+
* @example Continue the scene naturally, maintaining the same style and motion.
|
|
11507
|
+
*/
|
|
11508
|
+
prompt: string;
|
|
11509
|
+
/**
|
|
11510
|
+
* Seed
|
|
11511
|
+
* @description The seed used for the random number generator.
|
|
11512
|
+
* @example 866232447
|
|
11513
|
+
*/
|
|
11514
|
+
seed: number;
|
|
11515
|
+
/**
|
|
11516
|
+
* @description The generated video.
|
|
11517
|
+
* @example {
|
|
11518
|
+
* "height": 704,
|
|
11519
|
+
* "duration": 10.28,
|
|
11520
|
+
* "url": "https://v3b.fal.media/files/b/0a88289e/CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
|
|
11521
|
+
* "fps": 25,
|
|
11522
|
+
* "width": 1248,
|
|
11523
|
+
* "file_name": "CJcQGDrxOSRg2YFl5GNDt_glXPMoji.mp4",
|
|
11524
|
+
* "num_frames": 257,
|
|
11525
|
+
* "content_type": "video/mp4"
|
|
11526
|
+
* }
|
|
11527
|
+
*/
|
|
11528
|
+
video: Components.VideoFile;
|
|
11529
|
+
}
|
|
11530
|
+
|
|
11383
11531
|
export interface SharedType_367 {
|
|
11384
11532
|
/**
|
|
11385
11533
|
* Aspect Ratio
|
|
@@ -11589,6 +11737,26 @@ export interface SharedType_304 {
|
|
|
11589
11737
|
voice: Components.TextVoice;
|
|
11590
11738
|
}
|
|
11591
11739
|
|
|
11740
|
+
export interface SharedType_2ff {
|
|
11741
|
+
/**
|
|
11742
|
+
* Image Url
|
|
11743
|
+
* @description The URL of the image to be processed.
|
|
11744
|
+
* @example https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/car.jpg
|
|
11745
|
+
* @example http://ecx.images-amazon.com/images/I/51UUzBDAMsL.jpg
|
|
11746
|
+
*/
|
|
11747
|
+
image_url: string;
|
|
11748
|
+
/**
|
|
11749
|
+
* @description The user input coordinates
|
|
11750
|
+
* @example {
|
|
11751
|
+
* "y2": 200,
|
|
11752
|
+
* "x2": 200,
|
|
11753
|
+
* "x1": 100,
|
|
11754
|
+
* "y1": 100
|
|
11755
|
+
* }
|
|
11756
|
+
*/
|
|
11757
|
+
region: Components.Region;
|
|
11758
|
+
}
|
|
11759
|
+
|
|
11592
11760
|
export interface SharedType_2e6 {
|
|
11593
11761
|
/**
|
|
11594
11762
|
* Aspect Ratio
|
|
@@ -12184,6 +12352,74 @@ export interface SharedType_21d {
|
|
|
12184
12352
|
turn_off_cfg_start_si?: number;
|
|
12185
12353
|
}
|
|
12186
12354
|
|
|
12355
|
+
export interface SharedType_219 {
|
|
12356
|
+
/**
|
|
12357
|
+
* Enable Safety Checker
|
|
12358
|
+
* @description If set to true, the safety checker will be enabled.
|
|
12359
|
+
* @default true
|
|
12360
|
+
*/
|
|
12361
|
+
enable_safety_checker?: boolean;
|
|
12362
|
+
/**
|
|
12363
|
+
* Guidance scale (CFG)
|
|
12364
|
+
* @description The CFG (Classifier Free Guidance) scale is a measure of how close you want
|
|
12365
|
+
* the model to stick to your prompt when looking for a related image to show you.
|
|
12366
|
+
* @default 3.5
|
|
12367
|
+
*/
|
|
12368
|
+
guidance_scale?: number;
|
|
12369
|
+
/**
|
|
12370
|
+
* Image Size
|
|
12371
|
+
* @description The size of the generated image.
|
|
12372
|
+
* @default landscape_4_3
|
|
12373
|
+
*/
|
|
12374
|
+
image_size?:
|
|
12375
|
+
| Components.ImageSize
|
|
12376
|
+
| (
|
|
12377
|
+
| 'square_hd'
|
|
12378
|
+
| 'square'
|
|
12379
|
+
| 'portrait_4_3'
|
|
12380
|
+
| 'portrait_16_9'
|
|
12381
|
+
| 'landscape_4_3'
|
|
12382
|
+
| 'landscape_16_9'
|
|
12383
|
+
);
|
|
12384
|
+
/**
|
|
12385
|
+
* Negative prompt
|
|
12386
|
+
* @description Negative Prompt for generation.
|
|
12387
|
+
* @default
|
|
12388
|
+
* @example Blurry, out of focus, low resolution, bad anatomy, ugly, deformed, poorly drawn, extra limbs
|
|
12389
|
+
*/
|
|
12390
|
+
negative_prompt?: string;
|
|
12391
|
+
/**
|
|
12392
|
+
* Num Images
|
|
12393
|
+
* @description The number of images to generate.
|
|
12394
|
+
* @default 1
|
|
12395
|
+
*/
|
|
12396
|
+
num_images?: number;
|
|
12397
|
+
/**
|
|
12398
|
+
* Num Inference Steps
|
|
12399
|
+
* @description The number of inference steps to perform.
|
|
12400
|
+
* @default 28
|
|
12401
|
+
*/
|
|
12402
|
+
num_inference_steps?: number;
|
|
12403
|
+
/**
|
|
12404
|
+
* Prompt
|
|
12405
|
+
* @description The prompt to generate an image from.
|
|
12406
|
+
* @example Mount Fuji at sunset, with the iconic snow-capped peak silhouetted against a vibrant orange and purple sky. A tranquil lake in the foreground perfectly reflects the mountain and colorful sky. A few traditional Japanese cherry blossom trees frame the scene, with their delicate pink petals visible in the foreground.
|
|
12407
|
+
*/
|
|
12408
|
+
prompt: string;
|
|
12409
|
+
/**
|
|
12410
|
+
* Seed
|
|
12411
|
+
* @description The same seed and the same prompt given to the same version of the model
|
|
12412
|
+
* will output the same image every time.
|
|
12413
|
+
*/
|
|
12414
|
+
seed?: number;
|
|
12415
|
+
/**
|
|
12416
|
+
* Sync Mode
|
|
12417
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
12418
|
+
* @default false
|
|
12419
|
+
*/
|
|
12420
|
+
sync_mode?: boolean;
|
|
12421
|
+
}
|
|
12422
|
+
|
|
12187
12423
|
export interface SharedType_215 {
|
|
12188
12424
|
/**
|
|
12189
12425
|
* Acceleration
|
|
@@ -12778,7 +13014,7 @@ export interface SharedType_129 {
|
|
|
12778
13014
|
results: string;
|
|
12779
13015
|
}
|
|
12780
13016
|
|
|
12781
|
-
export interface
|
|
13017
|
+
export interface SharedType_1001 {
|
|
12782
13018
|
/**
|
|
12783
13019
|
* Prompt
|
|
12784
13020
|
* @description The prompt used for generating the video.
|
|
@@ -12796,6 +13032,38 @@ export interface SharedType_100 {
|
|
|
12796
13032
|
video: Components.File;
|
|
12797
13033
|
}
|
|
12798
13034
|
|
|
13035
|
+
export interface SharedType_100 {
|
|
13036
|
+
/**
|
|
13037
|
+
* Masks
|
|
13038
|
+
* @description Dictionary of label: mask image
|
|
13039
|
+
* @example [
|
|
13040
|
+
* {
|
|
13041
|
+
* "file_size": 15724,
|
|
13042
|
+
* "height": 1200,
|
|
13043
|
+
* "file_name": "019c3c1e3c50446e9996f709d36debb4.png",
|
|
13044
|
+
* "content_type": "image/png",
|
|
13045
|
+
* "url": "https://v3.fal.media/files/monkey/6ITmhHQJ-69s-UxajrY5T_019c3c1e3c50446e9996f709d36debb4.png",
|
|
13046
|
+
* "width": 1800
|
|
13047
|
+
* },
|
|
13048
|
+
* {
|
|
13049
|
+
* "file_size": 14905,
|
|
13050
|
+
* "height": 1200,
|
|
13051
|
+
* "file_name": "0a1522ca410942c7ad6c73efa15b3549.png",
|
|
13052
|
+
* "content_type": "image/png",
|
|
13053
|
+
* "url": "https://v3.fal.media/files/monkey/IljtMxahoo9-7SUpx0fth_0a1522ca410942c7ad6c73efa15b3549.png",
|
|
13054
|
+
* "width": 1800
|
|
13055
|
+
* }
|
|
13056
|
+
* ]
|
|
13057
|
+
*/
|
|
13058
|
+
masks: Components.Image[];
|
|
13059
|
+
/**
|
|
13060
|
+
* Output
|
|
13061
|
+
* @description Generated output
|
|
13062
|
+
* @example <p> A white pickup truck </p> [SEG] is parked on the side of <p> the red building </p> [SEG] , creating a unique and eye-catching contrast.<|im_end|>
|
|
13063
|
+
*/
|
|
13064
|
+
output: string;
|
|
13065
|
+
}
|
|
13066
|
+
|
|
12799
13067
|
export interface SharedType_0ff {
|
|
12800
13068
|
/**
|
|
12801
13069
|
* @description The generated speaker embedding file in safetensors format.
|
|
@@ -12990,6 +13258,24 @@ export interface SharedType_0cd {
|
|
|
12990
13258
|
sync_mode?: boolean;
|
|
12991
13259
|
}
|
|
12992
13260
|
|
|
13261
|
+
export interface SharedType_0c6 {
|
|
13262
|
+
/**
|
|
13263
|
+
* Seed
|
|
13264
|
+
* @description The seed used for generation.
|
|
13265
|
+
*/
|
|
13266
|
+
seed: number;
|
|
13267
|
+
/**
|
|
13268
|
+
* @description The generated video file.
|
|
13269
|
+
* @example {
|
|
13270
|
+
* "file_size": 515275,
|
|
13271
|
+
* "file_name": "74af6c0bdd6041c3b1130d54885e3eee.mp4",
|
|
13272
|
+
* "content_type": "application/octet-stream",
|
|
13273
|
+
* "url": "https://v3.fal.media/files/kangaroo/z6VqUwNTwzuWa6YE1g7In_74af6c0bdd6041c3b1130d54885e3eee.mp4"
|
|
13274
|
+
* }
|
|
13275
|
+
*/
|
|
13276
|
+
video: Components.File;
|
|
13277
|
+
}
|
|
13278
|
+
|
|
12993
13279
|
export interface SharedType_0c5 {
|
|
12994
13280
|
/**
|
|
12995
13281
|
* Images
|
|
@@ -14907,7 +15193,7 @@ export interface ZonosOutput {
|
|
|
14907
15193
|
audio: Components.File_1;
|
|
14908
15194
|
}
|
|
14909
15195
|
|
|
14910
|
-
export interface
|
|
15196
|
+
export interface ZImageTurboTilingLoraInput {
|
|
14911
15197
|
/**
|
|
14912
15198
|
* Acceleration
|
|
14913
15199
|
* @description The acceleration level to use.
|
|
@@ -14929,8 +15215,8 @@ export interface ZImageTurboLoraInput {
|
|
|
14929
15215
|
enable_safety_checker?: boolean;
|
|
14930
15216
|
/**
|
|
14931
15217
|
* Image Size
|
|
14932
|
-
* @description The size of the generated image.
|
|
14933
|
-
* @default
|
|
15218
|
+
* @description The size of the generated image. Use 'auto' to match the input image size (or 1024x1024 if no image).
|
|
15219
|
+
* @default square_hd
|
|
14934
15220
|
*/
|
|
14935
15221
|
image_size?:
|
|
14936
15222
|
| Components.ImageSize
|
|
@@ -14941,13 +15227,24 @@ export interface ZImageTurboLoraInput {
|
|
|
14941
15227
|
| 'portrait_16_9'
|
|
14942
15228
|
| 'landscape_4_3'
|
|
14943
15229
|
| 'landscape_16_9'
|
|
15230
|
+
| 'auto'
|
|
14944
15231
|
);
|
|
15232
|
+
/**
|
|
15233
|
+
* Image URL
|
|
15234
|
+
* @description URL of an image for image-to-image or inpainting. When provided without mask_image_url, performs image-to-image; with mask_image_url, performs inpainting.
|
|
15235
|
+
*/
|
|
15236
|
+
image_url?: string;
|
|
14945
15237
|
/**
|
|
14946
15238
|
* Loras
|
|
14947
15239
|
* @description List of LoRA weights to apply (maximum 3).
|
|
14948
15240
|
* @default []
|
|
14949
15241
|
*/
|
|
14950
15242
|
loras?: Components.LoRAInput_2[];
|
|
15243
|
+
/**
|
|
15244
|
+
* Mask Image URL
|
|
15245
|
+
* @description URL of a mask image for inpainting. White regions are regenerated, black regions are preserved. Requires image_url.
|
|
15246
|
+
*/
|
|
15247
|
+
mask_image_url?: string;
|
|
14951
15248
|
/**
|
|
14952
15249
|
* Number of Images
|
|
14953
15250
|
* @description The number of images to generate.
|
|
@@ -14970,7 +15267,7 @@ export interface ZImageTurboLoraInput {
|
|
|
14970
15267
|
/**
|
|
14971
15268
|
* Prompt
|
|
14972
15269
|
* @description The prompt to generate an image from.
|
|
14973
|
-
* @example A hyper-realistic,
|
|
15270
|
+
* @example A hyper-realistic, high-resolution 4k texture of an ancient weathered brick wall heavily overgrown with lush green moss and soft lichens. The bricks are aged, featuring deep earthy tones, natural cracks, and gritty textures. Vibrant emerald moss fills the mortar lines and spills over the rough surfaces of the stones. Uniform, flat cinematic lighting ensures no harsh shadows, highlighting the intricate organic details and damp stone surfaces. The composition is a perfectly balanced overhead view, showcasing a rich tapestry of botanical growth and masonry craftsmanship with professional clarity and hyper-detailed grit.
|
|
14974
15271
|
*/
|
|
14975
15272
|
prompt: string;
|
|
14976
15273
|
/**
|
|
@@ -14979,17 +15276,42 @@ export interface ZImageTurboLoraInput {
|
|
|
14979
15276
|
* will output the same image every time.
|
|
14980
15277
|
*/
|
|
14981
15278
|
seed?: number;
|
|
15279
|
+
/**
|
|
15280
|
+
* Strength
|
|
15281
|
+
* @description How much to transform the input image. Only used when image_url is provided.
|
|
15282
|
+
* @default 0.6
|
|
15283
|
+
*/
|
|
15284
|
+
strength?: number;
|
|
14982
15285
|
/**
|
|
14983
15286
|
* Sync Mode
|
|
14984
15287
|
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
14985
15288
|
* @default false
|
|
14986
15289
|
*/
|
|
14987
15290
|
sync_mode?: boolean;
|
|
15291
|
+
/**
|
|
15292
|
+
* Tile Size
|
|
15293
|
+
* @description Tile size in latent space (64 = 512px, 128 = 1024px, 256 = 2048px).
|
|
15294
|
+
* @default 128
|
|
15295
|
+
*/
|
|
15296
|
+
tile_size?: number;
|
|
15297
|
+
/**
|
|
15298
|
+
* Tile Stride
|
|
15299
|
+
* @description Tile stride in latent space. (32 = 256px, 64 = 512px, 128 = 1024px).
|
|
15300
|
+
* @default 64
|
|
15301
|
+
*/
|
|
15302
|
+
tile_stride?: number;
|
|
15303
|
+
/**
|
|
15304
|
+
* Tiling Mode
|
|
15305
|
+
* @description Tiling direction: 'both' (omnidirectional), 'horizontal' (left-right only), 'vertical' (top-bottom only).
|
|
15306
|
+
* @default both
|
|
15307
|
+
* @enum {string}
|
|
15308
|
+
*/
|
|
15309
|
+
tiling_mode?: 'both' | 'horizontal' | 'vertical';
|
|
14988
15310
|
}
|
|
14989
15311
|
|
|
14990
|
-
export interface
|
|
15312
|
+
export interface ZImageTurboTilingLoraOutput extends SharedType_390 {}
|
|
14991
15313
|
|
|
14992
|
-
export interface
|
|
15314
|
+
export interface ZImageTurboTilingInput {
|
|
14993
15315
|
/**
|
|
14994
15316
|
* Acceleration
|
|
14995
15317
|
* @description The acceleration level to use.
|
|
@@ -14998,23 +15320,117 @@ export interface ZImageTurboInpaintLoraInput {
|
|
|
14998
15320
|
*/
|
|
14999
15321
|
acceleration?: 'none' | 'regular' | 'high';
|
|
15000
15322
|
/**
|
|
15001
|
-
*
|
|
15002
|
-
* @description
|
|
15003
|
-
* @default
|
|
15323
|
+
* Enable Prompt Expansion
|
|
15324
|
+
* @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
|
|
15325
|
+
* @default false
|
|
15004
15326
|
*/
|
|
15005
|
-
|
|
15327
|
+
enable_prompt_expansion?: boolean;
|
|
15006
15328
|
/**
|
|
15007
|
-
*
|
|
15008
|
-
* @description
|
|
15009
|
-
* @default
|
|
15329
|
+
* Enable Safety Checker
|
|
15330
|
+
* @description If set to true, the safety checker will be enabled.
|
|
15331
|
+
* @default true
|
|
15010
15332
|
*/
|
|
15011
|
-
|
|
15333
|
+
enable_safety_checker?: boolean;
|
|
15012
15334
|
/**
|
|
15013
|
-
*
|
|
15014
|
-
* @description The
|
|
15015
|
-
* @default
|
|
15335
|
+
* Image Size
|
|
15336
|
+
* @description The size of the generated image. Use 'auto' to match the input image size (or 1024x1024 if no image).
|
|
15337
|
+
* @default square_hd
|
|
15016
15338
|
*/
|
|
15017
|
-
|
|
15339
|
+
image_size?:
|
|
15340
|
+
| Components.ImageSize
|
|
15341
|
+
| (
|
|
15342
|
+
| 'square_hd'
|
|
15343
|
+
| 'square'
|
|
15344
|
+
| 'portrait_4_3'
|
|
15345
|
+
| 'portrait_16_9'
|
|
15346
|
+
| 'landscape_4_3'
|
|
15347
|
+
| 'landscape_16_9'
|
|
15348
|
+
| 'auto'
|
|
15349
|
+
);
|
|
15350
|
+
/**
|
|
15351
|
+
* Image URL
|
|
15352
|
+
* @description URL of an image for image-to-image or inpainting. When provided without mask_image_url, performs image-to-image; with mask_image_url, performs inpainting.
|
|
15353
|
+
*/
|
|
15354
|
+
image_url?: string;
|
|
15355
|
+
/**
|
|
15356
|
+
* Mask Image URL
|
|
15357
|
+
* @description URL of a mask image for inpainting. White regions are regenerated, black regions are preserved. Requires image_url.
|
|
15358
|
+
*/
|
|
15359
|
+
mask_image_url?: string;
|
|
15360
|
+
/**
|
|
15361
|
+
* Number of Images
|
|
15362
|
+
* @description The number of images to generate.
|
|
15363
|
+
* @default 1
|
|
15364
|
+
*/
|
|
15365
|
+
num_images?: number;
|
|
15366
|
+
/**
|
|
15367
|
+
* Number of Inference Steps
|
|
15368
|
+
* @description The number of inference steps to perform.
|
|
15369
|
+
* @default 8
|
|
15370
|
+
*/
|
|
15371
|
+
num_inference_steps?: number;
|
|
15372
|
+
/**
|
|
15373
|
+
* Output Format
|
|
15374
|
+
* @description The format of the generated image.
|
|
15375
|
+
* @default png
|
|
15376
|
+
* @enum {string}
|
|
15377
|
+
*/
|
|
15378
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
15379
|
+
/**
|
|
15380
|
+
* Prompt
|
|
15381
|
+
* @description The prompt to generate an image from.
|
|
15382
|
+
* @example A hyper-realistic, high-resolution 4k texture of an ancient weathered brick wall heavily overgrown with lush green moss and soft lichens. The bricks are aged, featuring deep earthy tones, natural cracks, and gritty textures. Vibrant emerald moss fills the mortar lines and spills over the rough surfaces of the stones. Uniform, flat cinematic lighting ensures no harsh shadows, highlighting the intricate organic details and damp stone surfaces. The composition is a perfectly balanced overhead view, showcasing a rich tapestry of botanical growth and masonry craftsmanship with professional clarity and hyper-detailed grit.
|
|
15383
|
+
*/
|
|
15384
|
+
prompt: string;
|
|
15385
|
+
/**
|
|
15386
|
+
* Seed
|
|
15387
|
+
* @description The same seed and the same prompt given to the same version of the model
|
|
15388
|
+
* will output the same image every time.
|
|
15389
|
+
*/
|
|
15390
|
+
seed?: number;
|
|
15391
|
+
/**
|
|
15392
|
+
* Strength
|
|
15393
|
+
* @description How much to transform the input image. Only used when image_url is provided.
|
|
15394
|
+
* @default 0.6
|
|
15395
|
+
*/
|
|
15396
|
+
strength?: number;
|
|
15397
|
+
/**
|
|
15398
|
+
* Sync Mode
|
|
15399
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
15400
|
+
* @default false
|
|
15401
|
+
*/
|
|
15402
|
+
sync_mode?: boolean;
|
|
15403
|
+
/**
|
|
15404
|
+
* Tile Size
|
|
15405
|
+
* @description Tile size in latent space (64 = 512px, 128 = 1024px, 256 = 2048px).
|
|
15406
|
+
* @default 128
|
|
15407
|
+
*/
|
|
15408
|
+
tile_size?: number;
|
|
15409
|
+
/**
|
|
15410
|
+
* Tile Stride
|
|
15411
|
+
* @description Tile stride in latent space. (32 = 256px, 64 = 512px, 128 = 1024px).
|
|
15412
|
+
* @default 64
|
|
15413
|
+
*/
|
|
15414
|
+
tile_stride?: number;
|
|
15415
|
+
/**
|
|
15416
|
+
* Tiling Mode
|
|
15417
|
+
* @description Tiling direction: 'both' (omnidirectional), 'horizontal' (left-right only), 'vertical' (top-bottom only).
|
|
15418
|
+
* @default both
|
|
15419
|
+
* @enum {string}
|
|
15420
|
+
*/
|
|
15421
|
+
tiling_mode?: 'both' | 'horizontal' | 'vertical';
|
|
15422
|
+
}
|
|
15423
|
+
|
|
15424
|
+
export interface ZImageTurboTilingOutput extends SharedType_390 {}
|
|
15425
|
+
|
|
15426
|
+
export interface ZImageTurboLoraInput {
|
|
15427
|
+
/**
|
|
15428
|
+
* Acceleration
|
|
15429
|
+
* @description The acceleration level to use.
|
|
15430
|
+
* @default regular
|
|
15431
|
+
* @enum {string}
|
|
15432
|
+
*/
|
|
15433
|
+
acceleration?: 'none' | 'regular' | 'high';
|
|
15018
15434
|
/**
|
|
15019
15435
|
* Enable Prompt Expansion
|
|
15020
15436
|
* @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
|
|
@@ -15030,7 +15446,7 @@ export interface ZImageTurboInpaintLoraInput {
|
|
|
15030
15446
|
/**
|
|
15031
15447
|
* Image Size
|
|
15032
15448
|
* @description The size of the generated image.
|
|
15033
|
-
* @default
|
|
15449
|
+
* @default landscape_4_3
|
|
15034
15450
|
*/
|
|
15035
15451
|
image_size?:
|
|
15036
15452
|
| Components.ImageSize
|
|
@@ -15041,26 +15457,13 @@ export interface ZImageTurboInpaintLoraInput {
|
|
|
15041
15457
|
| 'portrait_16_9'
|
|
15042
15458
|
| 'landscape_4_3'
|
|
15043
15459
|
| 'landscape_16_9'
|
|
15044
|
-
| 'auto'
|
|
15045
15460
|
);
|
|
15046
|
-
/**
|
|
15047
|
-
* Image URL
|
|
15048
|
-
* @description URL of Image for Inpaint generation.
|
|
15049
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/inpaint-input.jpg
|
|
15050
|
-
*/
|
|
15051
|
-
image_url: string;
|
|
15052
15461
|
/**
|
|
15053
15462
|
* Loras
|
|
15054
15463
|
* @description List of LoRA weights to apply (maximum 3).
|
|
15055
15464
|
* @default []
|
|
15056
15465
|
*/
|
|
15057
15466
|
loras?: Components.LoRAInput_2[];
|
|
15058
|
-
/**
|
|
15059
|
-
* Mask Image URL
|
|
15060
|
-
* @description URL of Mask for Inpaint generation.
|
|
15061
|
-
* @example https://storage.googleapis.com/falserverless/whls/z-image-inpaint-mask.jpg
|
|
15062
|
-
*/
|
|
15063
|
-
mask_image_url: string;
|
|
15064
15467
|
/**
|
|
15065
15468
|
* Number of Images
|
|
15066
15469
|
* @description The number of images to generate.
|
|
@@ -15083,7 +15486,7 @@ export interface ZImageTurboInpaintLoraInput {
|
|
|
15083
15486
|
/**
|
|
15084
15487
|
* Prompt
|
|
15085
15488
|
* @description The prompt to generate an image from.
|
|
15086
|
-
* @example A
|
|
15489
|
+
* @example A hyper-realistic, close-up portrait of a tribal elder from the Omo Valley, painted with intricate white chalk patterns and adorned with a headdress made of dried flowers, seed pods, and rusted bottle caps. The focus is razor-sharp on the texture of the skin, showing every pore, wrinkle, and scar that tells a story of survival. The background is a blurred, smoky hut interior, with the warm glow of a cooking fire reflecting in the subject's dark, soulful eyes. Shot on a Leica M6 with Kodak Portra 400 film grain aesthetic.
|
|
15087
15490
|
*/
|
|
15088
15491
|
prompt: string;
|
|
15089
15492
|
/**
|
|
@@ -15092,12 +15495,6 @@ export interface ZImageTurboInpaintLoraInput {
|
|
|
15092
15495
|
* will output the same image every time.
|
|
15093
15496
|
*/
|
|
15094
15497
|
seed?: number;
|
|
15095
|
-
/**
|
|
15096
|
-
* Strength
|
|
15097
|
-
* @description The strength of the inpaint conditioning.
|
|
15098
|
-
* @default 1
|
|
15099
|
-
*/
|
|
15100
|
-
strength?: number;
|
|
15101
15498
|
/**
|
|
15102
15499
|
* Sync Mode
|
|
15103
15500
|
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
@@ -15106,9 +15503,9 @@ export interface ZImageTurboInpaintLoraInput {
|
|
|
15106
15503
|
sync_mode?: boolean;
|
|
15107
15504
|
}
|
|
15108
15505
|
|
|
15109
|
-
export interface
|
|
15506
|
+
export interface ZImageTurboLoraOutput extends SharedType_352 {}
|
|
15110
15507
|
|
|
15111
|
-
export interface
|
|
15508
|
+
export interface ZImageTurboInpaintLoraInput {
|
|
15112
15509
|
/**
|
|
15113
15510
|
* Acceleration
|
|
15114
15511
|
* @description The acceleration level to use.
|
|
@@ -15168,6 +15565,12 @@ export interface ZImageTurboInpaintInput {
|
|
|
15168
15565
|
* @example https://storage.googleapis.com/falserverless/example_inputs/inpaint-input.jpg
|
|
15169
15566
|
*/
|
|
15170
15567
|
image_url: string;
|
|
15568
|
+
/**
|
|
15569
|
+
* Loras
|
|
15570
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
15571
|
+
* @default []
|
|
15572
|
+
*/
|
|
15573
|
+
loras?: Components.LoRAInput_2[];
|
|
15171
15574
|
/**
|
|
15172
15575
|
* Mask Image URL
|
|
15173
15576
|
* @description URL of Mask for Inpaint generation.
|
|
@@ -15219,9 +15622,9 @@ export interface ZImageTurboInpaintInput {
|
|
|
15219
15622
|
sync_mode?: boolean;
|
|
15220
15623
|
}
|
|
15221
15624
|
|
|
15222
|
-
export interface
|
|
15625
|
+
export interface ZImageTurboInpaintLoraOutput extends SharedType_18d {}
|
|
15223
15626
|
|
|
15224
|
-
export interface
|
|
15627
|
+
export interface ZImageTurboInpaintInput {
|
|
15225
15628
|
/**
|
|
15226
15629
|
* Acceleration
|
|
15227
15630
|
* @description The acceleration level to use.
|
|
@@ -15229,6 +15632,24 @@ export interface ZImageTurboImageToImageLoraInput {
|
|
|
15229
15632
|
* @enum {string}
|
|
15230
15633
|
*/
|
|
15231
15634
|
acceleration?: 'none' | 'regular' | 'high';
|
|
15635
|
+
/**
|
|
15636
|
+
* Control End
|
|
15637
|
+
* @description The end of the controlnet conditioning.
|
|
15638
|
+
* @default 0.8
|
|
15639
|
+
*/
|
|
15640
|
+
control_end?: number;
|
|
15641
|
+
/**
|
|
15642
|
+
* Control Scale
|
|
15643
|
+
* @description The scale of the controlnet conditioning.
|
|
15644
|
+
* @default 0.75
|
|
15645
|
+
*/
|
|
15646
|
+
control_scale?: number;
|
|
15647
|
+
/**
|
|
15648
|
+
* Control Start
|
|
15649
|
+
* @description The start of the controlnet conditioning.
|
|
15650
|
+
* @default 0
|
|
15651
|
+
*/
|
|
15652
|
+
control_start?: number;
|
|
15232
15653
|
/**
|
|
15233
15654
|
* Enable Prompt Expansion
|
|
15234
15655
|
* @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
|
|
@@ -15259,16 +15680,16 @@ export interface ZImageTurboImageToImageLoraInput {
|
|
|
15259
15680
|
);
|
|
15260
15681
|
/**
|
|
15261
15682
|
* Image URL
|
|
15262
|
-
* @description URL of Image for
|
|
15263
|
-
* @example https://storage.googleapis.com/falserverless/example_inputs/
|
|
15683
|
+
* @description URL of Image for Inpaint generation.
|
|
15684
|
+
* @example https://storage.googleapis.com/falserverless/example_inputs/inpaint-input.jpg
|
|
15264
15685
|
*/
|
|
15265
15686
|
image_url: string;
|
|
15266
15687
|
/**
|
|
15267
|
-
*
|
|
15268
|
-
* @description
|
|
15269
|
-
* @
|
|
15688
|
+
* Mask Image URL
|
|
15689
|
+
* @description URL of Mask for Inpaint generation.
|
|
15690
|
+
* @example https://storage.googleapis.com/falserverless/whls/z-image-inpaint-mask.jpg
|
|
15270
15691
|
*/
|
|
15271
|
-
|
|
15692
|
+
mask_image_url: string;
|
|
15272
15693
|
/**
|
|
15273
15694
|
* Number of Images
|
|
15274
15695
|
* @description The number of images to generate.
|
|
@@ -15302,8 +15723,8 @@ export interface ZImageTurboImageToImageLoraInput {
|
|
|
15302
15723
|
seed?: number;
|
|
15303
15724
|
/**
|
|
15304
15725
|
* Strength
|
|
15305
|
-
* @description The strength of the
|
|
15306
|
-
* @default
|
|
15726
|
+
* @description The strength of the inpaint conditioning.
|
|
15727
|
+
* @default 1
|
|
15307
15728
|
*/
|
|
15308
15729
|
strength?: number;
|
|
15309
15730
|
/**
|
|
@@ -15314,9 +15735,104 @@ export interface ZImageTurboImageToImageLoraInput {
|
|
|
15314
15735
|
sync_mode?: boolean;
|
|
15315
15736
|
}
|
|
15316
15737
|
|
|
15317
|
-
export interface
|
|
15738
|
+
export interface ZImageTurboInpaintOutput extends SharedType_18d {}
|
|
15318
15739
|
|
|
15319
|
-
export interface
|
|
15740
|
+
export interface ZImageTurboImageToImageLoraInput {
|
|
15741
|
+
/**
|
|
15742
|
+
* Acceleration
|
|
15743
|
+
* @description The acceleration level to use.
|
|
15744
|
+
* @default regular
|
|
15745
|
+
* @enum {string}
|
|
15746
|
+
*/
|
|
15747
|
+
acceleration?: 'none' | 'regular' | 'high';
|
|
15748
|
+
/**
|
|
15749
|
+
* Enable Prompt Expansion
|
|
15750
|
+
* @description Whether to enable prompt expansion. Note: this will increase the price by 0.0025 credits per request.
|
|
15751
|
+
* @default false
|
|
15752
|
+
*/
|
|
15753
|
+
enable_prompt_expansion?: boolean;
|
|
15754
|
+
/**
|
|
15755
|
+
* Enable Safety Checker
|
|
15756
|
+
* @description If set to true, the safety checker will be enabled.
|
|
15757
|
+
* @default true
|
|
15758
|
+
*/
|
|
15759
|
+
enable_safety_checker?: boolean;
|
|
15760
|
+
/**
|
|
15761
|
+
* Image Size
|
|
15762
|
+
* @description The size of the generated image.
|
|
15763
|
+
* @default auto
|
|
15764
|
+
*/
|
|
15765
|
+
image_size?:
|
|
15766
|
+
| Components.ImageSize
|
|
15767
|
+
| (
|
|
15768
|
+
| 'square_hd'
|
|
15769
|
+
| 'square'
|
|
15770
|
+
| 'portrait_4_3'
|
|
15771
|
+
| 'portrait_16_9'
|
|
15772
|
+
| 'landscape_4_3'
|
|
15773
|
+
| 'landscape_16_9'
|
|
15774
|
+
| 'auto'
|
|
15775
|
+
);
|
|
15776
|
+
/**
|
|
15777
|
+
* Image URL
|
|
15778
|
+
* @description URL of Image for Image-to-Image generation.
|
|
15779
|
+
* @example https://storage.googleapis.com/falserverless/example_inputs/z-image-turbo-i2i-input.png
|
|
15780
|
+
*/
|
|
15781
|
+
image_url: string;
|
|
15782
|
+
/**
|
|
15783
|
+
* Loras
|
|
15784
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
15785
|
+
* @default []
|
|
15786
|
+
*/
|
|
15787
|
+
loras?: Components.LoRAInput_2[];
|
|
15788
|
+
/**
|
|
15789
|
+
* Number of Images
|
|
15790
|
+
* @description The number of images to generate.
|
|
15791
|
+
* @default 1
|
|
15792
|
+
*/
|
|
15793
|
+
num_images?: number;
|
|
15794
|
+
/**
|
|
15795
|
+
* Number of Inference Steps
|
|
15796
|
+
* @description The number of inference steps to perform.
|
|
15797
|
+
* @default 8
|
|
15798
|
+
*/
|
|
15799
|
+
num_inference_steps?: number;
|
|
15800
|
+
/**
|
|
15801
|
+
* Output Format
|
|
15802
|
+
* @description The format of the generated image.
|
|
15803
|
+
* @default png
|
|
15804
|
+
* @enum {string}
|
|
15805
|
+
*/
|
|
15806
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
15807
|
+
/**
|
|
15808
|
+
* Prompt
|
|
15809
|
+
* @description The prompt to generate an image from.
|
|
15810
|
+
* @example A young Asian woman with long, vibrant purple hair stands on a sunlit sandy beach, posing confidently with her left hand resting on her hip. She gazes directly at the camera with a neutral expression. A sleek black ribbon bow is tied neatly on the right side of her head, just above her ear. She wears a flowing white cotton dress with a fitted bodice and a flared skirt that reaches mid-calf, slightly lifted by a gentle sea breeze. The beach behind her features fine, pale golden sand with subtle footprints, leading to calm turquoise waves under a clear blue sky with soft, wispy clouds. The lighting is natural daylight, casting soft shadows to her left, indicating late afternoon sun. The horizon line is visible in the background, with a faint silhouette of distant dunes. Her skin tone is fair with a natural glow, and her facial features are delicately defined. The composition is centered on her figure, framed from mid-thigh up, with shallow depth of field blurring the distant waves slightly.
|
|
15811
|
+
*/
|
|
15812
|
+
prompt: string;
|
|
15813
|
+
/**
|
|
15814
|
+
* Seed
|
|
15815
|
+
* @description The same seed and the same prompt given to the same version of the model
|
|
15816
|
+
* will output the same image every time.
|
|
15817
|
+
*/
|
|
15818
|
+
seed?: number;
|
|
15819
|
+
/**
|
|
15820
|
+
* Strength
|
|
15821
|
+
* @description The strength of the image-to-image conditioning.
|
|
15822
|
+
* @default 0.6
|
|
15823
|
+
*/
|
|
15824
|
+
strength?: number;
|
|
15825
|
+
/**
|
|
15826
|
+
* Sync Mode
|
|
15827
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
15828
|
+
* @default false
|
|
15829
|
+
*/
|
|
15830
|
+
sync_mode?: boolean;
|
|
15831
|
+
}
|
|
15832
|
+
|
|
15833
|
+
export interface ZImageTurboImageToImageLoraOutput extends SharedType_6b3 {}
|
|
15834
|
+
|
|
15835
|
+
export interface ZImageTurboImageToImageInput {
|
|
15320
15836
|
/**
|
|
15321
15837
|
* Acceleration
|
|
15322
15838
|
* @description The acceleration level to use.
|
|
@@ -20089,7 +20605,21 @@ export interface WanVace13bInput {
|
|
|
20089
20605
|
video_url?: string;
|
|
20090
20606
|
}
|
|
20091
20607
|
|
|
20092
|
-
export interface WanVace13bOutput
|
|
20608
|
+
export interface WanVace13bOutput {
|
|
20609
|
+
/**
|
|
20610
|
+
* Seed
|
|
20611
|
+
* @description The seed used for generation.
|
|
20612
|
+
*/
|
|
20613
|
+
seed: number;
|
|
20614
|
+
/**
|
|
20615
|
+
* Video
|
|
20616
|
+
* @description The generated video file.
|
|
20617
|
+
* @example {
|
|
20618
|
+
* "url": "https://storage.googleapis.com/falserverless/vace/out_video_vace.mp4"
|
|
20619
|
+
* }
|
|
20620
|
+
*/
|
|
20621
|
+
video: Components.File_1;
|
|
20622
|
+
}
|
|
20093
20623
|
|
|
20094
20624
|
export interface WanVaceInput {
|
|
20095
20625
|
/**
|
|
@@ -20202,7 +20732,20 @@ export interface WanVaceInput {
|
|
|
20202
20732
|
video_url?: string;
|
|
20203
20733
|
}
|
|
20204
20734
|
|
|
20205
|
-
export interface WanVaceOutput
|
|
20735
|
+
export interface WanVaceOutput {
|
|
20736
|
+
/**
|
|
20737
|
+
* Seed
|
|
20738
|
+
* @description The seed used for generation.
|
|
20739
|
+
*/
|
|
20740
|
+
seed: number;
|
|
20741
|
+
/**
|
|
20742
|
+
* @description The generated video file.
|
|
20743
|
+
* @example {
|
|
20744
|
+
* "url": "https://storage.googleapis.com/falserverless/vace/out_video_vace.mp4"
|
|
20745
|
+
* }
|
|
20746
|
+
*/
|
|
20747
|
+
video: Components.File;
|
|
20748
|
+
}
|
|
20206
20749
|
|
|
20207
20750
|
export interface WanTrainerT2v14bInput extends SharedType_5951 {}
|
|
20208
20751
|
|
|
@@ -20257,7 +20800,7 @@ export interface WanT2vLoraInput {
|
|
|
20257
20800
|
* @description LoRA weights to be used in the inference.
|
|
20258
20801
|
* @default []
|
|
20259
20802
|
*/
|
|
20260
|
-
loras?: Components.
|
|
20803
|
+
loras?: Components.LoraWeight_5[];
|
|
20261
20804
|
/**
|
|
20262
20805
|
* Negative Prompt
|
|
20263
20806
|
* @description Negative prompt for video generation.
|
|
@@ -20572,7 +21115,7 @@ export interface WanI2vLoraInput {
|
|
|
20572
21115
|
* @description LoRA weights to be used in the inference.
|
|
20573
21116
|
* @default []
|
|
20574
21117
|
*/
|
|
20575
|
-
loras?: Components.
|
|
21118
|
+
loras?: Components.LoraWeight_5[];
|
|
20576
21119
|
/**
|
|
20577
21120
|
* Negative Prompt
|
|
20578
21121
|
* @description Negative prompt for video generation.
|
|
@@ -24153,13 +24696,12 @@ export interface ThinksoundAudioInput extends SharedType_4bc {}
|
|
|
24153
24696
|
|
|
24154
24697
|
export interface ThinksoundAudioOutput {
|
|
24155
24698
|
/**
|
|
24156
|
-
* Audio
|
|
24157
24699
|
* @description The generated audio file.
|
|
24158
24700
|
* @example {
|
|
24159
24701
|
* "url": "https://storage.googleapis.com/falserverless/example_outputs/thinksound-audio.wav"
|
|
24160
24702
|
* }
|
|
24161
24703
|
*/
|
|
24162
|
-
audio: Components.
|
|
24704
|
+
audio: Components.File;
|
|
24163
24705
|
/**
|
|
24164
24706
|
* Prompt
|
|
24165
24707
|
* @description The prompt used to generate the audio.
|
|
@@ -24178,13 +24720,12 @@ export interface ThinksoundOutput {
|
|
|
24178
24720
|
*/
|
|
24179
24721
|
prompt: string;
|
|
24180
24722
|
/**
|
|
24181
|
-
* Video
|
|
24182
24723
|
* @description The generated video with audio.
|
|
24183
24724
|
* @example {
|
|
24184
24725
|
* "url": "https://storage.googleapis.com/falserverless/example_outputs/thinksound-output.mp4"
|
|
24185
24726
|
* }
|
|
24186
24727
|
*/
|
|
24187
|
-
video: Components.
|
|
24728
|
+
video: Components.File;
|
|
24188
24729
|
}
|
|
24189
24730
|
|
|
24190
24731
|
export interface TheraInput {
|
|
@@ -24834,7 +25375,6 @@ export interface SteadyDancerOutput {
|
|
|
24834
25375
|
*/
|
|
24835
25376
|
seed: number;
|
|
24836
25377
|
/**
|
|
24837
|
-
* Video
|
|
24838
25378
|
* @description The generated dance animation video.
|
|
24839
25379
|
* @example {
|
|
24840
25380
|
* "file_size": 7772111,
|
|
@@ -24843,7 +25383,7 @@ export interface SteadyDancerOutput {
|
|
|
24843
25383
|
* "url": "https://v3b.fal.media/files/b/0a87871b/ll5ps0ZyBgxBkuWz-fHcT_output_with_audio.mp4"
|
|
24844
25384
|
* }
|
|
24845
25385
|
*/
|
|
24846
|
-
video: Components.
|
|
25386
|
+
video: Components.File;
|
|
24847
25387
|
}
|
|
24848
25388
|
|
|
24849
25389
|
export interface StarVectorInput {
|
|
@@ -25360,6 +25900,7 @@ export interface StableDiffusionV15Input {
|
|
|
25360
25900
|
* Seed
|
|
25361
25901
|
* @description The same seed and the same prompt given to the same version of Stable Diffusion
|
|
25362
25902
|
* will output the same image every time.
|
|
25903
|
+
* @default null
|
|
25363
25904
|
*/
|
|
25364
25905
|
seed?: number;
|
|
25365
25906
|
/**
|
|
@@ -25604,11 +26145,10 @@ export interface StableAvatarInput {
|
|
|
25604
26145
|
|
|
25605
26146
|
export interface StableAvatarOutput {
|
|
25606
26147
|
/**
|
|
25607
|
-
* Video
|
|
25608
26148
|
* @description The generated video file.
|
|
25609
26149
|
* @example https://storage.googleapis.com/falserverless/example_outputs/stable-avatar-output.mp4
|
|
25610
26150
|
*/
|
|
25611
|
-
video: Components.
|
|
26151
|
+
video: Components.File;
|
|
25612
26152
|
}
|
|
25613
26153
|
|
|
25614
26154
|
export interface StableAudio25TextToAudioInput {
|
|
@@ -26275,92 +26815,6 @@ export interface SkyreelsI2vOutput {
|
|
|
26275
26815
|
video: Components.File;
|
|
26276
26816
|
}
|
|
26277
26817
|
|
|
26278
|
-
export interface SkyRaccoonInput {
|
|
26279
|
-
/**
|
|
26280
|
-
* Enable Prompt Expansion
|
|
26281
|
-
* @description Whether to enable prompt expansion.
|
|
26282
|
-
* @default false
|
|
26283
|
-
* @example false
|
|
26284
|
-
*/
|
|
26285
|
-
enable_prompt_expansion?: boolean;
|
|
26286
|
-
/**
|
|
26287
|
-
* Enable Safety Checker
|
|
26288
|
-
* @description If set to true, the safety checker will be enabled.
|
|
26289
|
-
* @default false
|
|
26290
|
-
* @example true
|
|
26291
|
-
*/
|
|
26292
|
-
enable_safety_checker?: boolean;
|
|
26293
|
-
/**
|
|
26294
|
-
* Image Size
|
|
26295
|
-
* @description The size of the generated image.
|
|
26296
|
-
* @default {
|
|
26297
|
-
* "height": 1024,
|
|
26298
|
-
* "width": 1024
|
|
26299
|
-
* }
|
|
26300
|
-
*/
|
|
26301
|
-
image_size?:
|
|
26302
|
-
| Components.ImageSize
|
|
26303
|
-
| (
|
|
26304
|
-
| 'square_hd'
|
|
26305
|
-
| 'square'
|
|
26306
|
-
| 'portrait_4_3'
|
|
26307
|
-
| 'portrait_16_9'
|
|
26308
|
-
| 'landscape_4_3'
|
|
26309
|
-
| 'landscape_16_9'
|
|
26310
|
-
);
|
|
26311
|
-
/**
|
|
26312
|
-
* Loras
|
|
26313
|
-
* @description LoRA weights to be used in the inference.
|
|
26314
|
-
* @default []
|
|
26315
|
-
*/
|
|
26316
|
-
loras?: Components.LoraWeight_3[];
|
|
26317
|
-
/**
|
|
26318
|
-
* Negative Prompt
|
|
26319
|
-
* @description Negative prompt for video generation.
|
|
26320
|
-
* @default bright colors, overexposed, static, blurred details, subtitles, style, artwork, painting, picture, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, malformed limbs, fused fingers, still picture, cluttered background, three legs, many people in the background, walking backwards
|
|
26321
|
-
* @example bright colors, overexposed, static, blurred details, subtitles, style, artwork, painting, picture, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, malformed limbs, fused fingers, still picture, cluttered background, three legs, many people in the background, walking backwards
|
|
26322
|
-
*/
|
|
26323
|
-
negative_prompt?: string;
|
|
26324
|
-
/**
|
|
26325
|
-
* Num Inference Steps
|
|
26326
|
-
* @description Number of inference steps for sampling. Higher values give better quality but take longer.
|
|
26327
|
-
* @default 30
|
|
26328
|
-
*/
|
|
26329
|
-
num_inference_steps?: number;
|
|
26330
|
-
/**
|
|
26331
|
-
* Prompt
|
|
26332
|
-
* @description The text prompt to guide video generation.
|
|
26333
|
-
* @example A stylish woman walks down a Tokyo street filled with warm glowing neon and animated city signage. She wears a black leather jacket, a long red dress, and black boots, and carries a black purse.
|
|
26334
|
-
*/
|
|
26335
|
-
prompt: string;
|
|
26336
|
-
/**
|
|
26337
|
-
* Seed
|
|
26338
|
-
* @description Random seed for reproducibility. If None, a random seed is chosen.
|
|
26339
|
-
*/
|
|
26340
|
-
seed?: number;
|
|
26341
|
-
/**
|
|
26342
|
-
* Turbo Mode
|
|
26343
|
-
* @description If true, the image will be generated faster with no noticeable degradation in the visual quality.
|
|
26344
|
-
* @default true
|
|
26345
|
-
*/
|
|
26346
|
-
turbo_mode?: boolean;
|
|
26347
|
-
}
|
|
26348
|
-
|
|
26349
|
-
export interface SkyRaccoonOutput {
|
|
26350
|
-
/**
|
|
26351
|
-
* @description The generated image file.
|
|
26352
|
-
* @example {
|
|
26353
|
-
* "url": "https://v3.fal.media/files/kangaroo/xP17uR1w0uHA1T2W2k6Gi_video-1752605675.png"
|
|
26354
|
-
* }
|
|
26355
|
-
*/
|
|
26356
|
-
image: Components.File;
|
|
26357
|
-
/**
|
|
26358
|
-
* Seed
|
|
26359
|
-
* @description The seed used for generation.
|
|
26360
|
-
*/
|
|
26361
|
-
seed: number;
|
|
26362
|
-
}
|
|
26363
|
-
|
|
26364
26818
|
export interface SileroVadInput {
|
|
26365
26819
|
/**
|
|
26366
26820
|
* Audio URL
|
|
@@ -27219,7 +27673,7 @@ export interface Sd15DepthControlnetInput {
|
|
|
27219
27673
|
* @description The list of LoRA weights to use.
|
|
27220
27674
|
* @default []
|
|
27221
27675
|
*/
|
|
27222
|
-
loras?: Components.
|
|
27676
|
+
loras?: Components.LoraWeight_3[];
|
|
27223
27677
|
/**
|
|
27224
27678
|
* Negative Prompt
|
|
27225
27679
|
* @description The negative prompt to use. Use it to address details that you don't want
|
|
@@ -27544,15 +27998,15 @@ export interface Sam2VideoInput {
|
|
|
27544
27998
|
* @example [
|
|
27545
27999
|
* {
|
|
27546
28000
|
* "y": 350,
|
|
27547
|
-
* "
|
|
28001
|
+
* "x": 210,
|
|
27548
28002
|
* "frame_index": 0,
|
|
27549
|
-
* "
|
|
28003
|
+
* "label": 1
|
|
27550
28004
|
* },
|
|
27551
28005
|
* {
|
|
27552
28006
|
* "y": 220,
|
|
27553
|
-
* "
|
|
28007
|
+
* "x": 250,
|
|
27554
28008
|
* "frame_index": 0,
|
|
27555
|
-
* "
|
|
28009
|
+
* "label": 1
|
|
27556
28010
|
* }
|
|
27557
28011
|
* ]
|
|
27558
28012
|
*/
|
|
@@ -27611,8 +28065,8 @@ export interface Sam2ImageInput {
|
|
|
27611
28065
|
* @example [
|
|
27612
28066
|
* {
|
|
27613
28067
|
* "y": 375,
|
|
27614
|
-
* "
|
|
27615
|
-
* "
|
|
28068
|
+
* "x": 500,
|
|
28069
|
+
* "label": 1
|
|
27616
28070
|
* }
|
|
27617
28071
|
* ]
|
|
27618
28072
|
*/
|
|
@@ -28052,6 +28506,13 @@ export interface Sam3VideoInput {
|
|
|
28052
28506
|
* @description [DEPRECATED] Use 'prompt' instead. Kept for backward compatibility.
|
|
28053
28507
|
*/
|
|
28054
28508
|
text_prompt?: string;
|
|
28509
|
+
/**
|
|
28510
|
+
* Video Output Type
|
|
28511
|
+
* @description The output type of the generated video.
|
|
28512
|
+
* @default X264 (.mp4)
|
|
28513
|
+
* @enum {string}
|
|
28514
|
+
*/
|
|
28515
|
+
video_output_type?: 'X264 (.mp4)' | 'VP9 (.webm)';
|
|
28055
28516
|
/**
|
|
28056
28517
|
* Video Url
|
|
28057
28518
|
* @description The URL of the video to be segmented.
|
|
@@ -28409,7 +28870,7 @@ export interface Sa2va8bVideoOutput extends SharedType_800 {}
|
|
|
28409
28870
|
|
|
28410
28871
|
export interface Sa2va8bImageInput extends SharedType_f51 {}
|
|
28411
28872
|
|
|
28412
|
-
export interface Sa2va8bImageOutput extends
|
|
28873
|
+
export interface Sa2va8bImageOutput extends SharedType_100 {}
|
|
28413
28874
|
|
|
28414
28875
|
export interface Sa2va4bVideoInput extends SharedType_5f5 {}
|
|
28415
28876
|
|
|
@@ -28417,7 +28878,7 @@ export interface Sa2va4bVideoOutput extends SharedType_800 {}
|
|
|
28417
28878
|
|
|
28418
28879
|
export interface Sa2va4bImageInput extends SharedType_f51 {}
|
|
28419
28880
|
|
|
28420
|
-
export interface Sa2va4bImageOutput extends
|
|
28881
|
+
export interface Sa2va4bImageOutput extends SharedType_100 {}
|
|
28421
28882
|
|
|
28422
28883
|
export interface RifeVideoInput {
|
|
28423
28884
|
/**
|
|
@@ -28558,13 +29019,12 @@ export interface RetoucherInput {
|
|
|
28558
29019
|
|
|
28559
29020
|
export interface RetoucherOutput {
|
|
28560
29021
|
/**
|
|
28561
|
-
* Image
|
|
28562
29022
|
* @description The generated image file info.
|
|
28563
29023
|
* @example {
|
|
28564
29024
|
* "url": "https://storage.googleapis.com/falserverless/model_tests/retoucher/retoucher_example_output.png"
|
|
28565
29025
|
* }
|
|
28566
29026
|
*/
|
|
28567
|
-
image: Components.
|
|
29027
|
+
image: Components.Image;
|
|
28568
29028
|
/**
|
|
28569
29029
|
* Seed
|
|
28570
29030
|
* @description The seed used for the generation.
|
|
@@ -36059,15 +36519,7 @@ export interface OrpheusTtsInput {
|
|
|
36059
36519
|
voice?: 'tara' | 'leah' | 'jess' | 'leo' | 'dan' | 'mia' | 'zac' | 'zoe';
|
|
36060
36520
|
}
|
|
36061
36521
|
|
|
36062
|
-
export interface OrpheusTtsOutput {
|
|
36063
|
-
/**
|
|
36064
|
-
* @description The generated speech audio
|
|
36065
|
-
* @example {
|
|
36066
|
-
* "url": "https://v3.fal.media/files/kangaroo/RQ_pxc7oPdueYqWUqEbPE_tmpjnzvvzx_.wav"
|
|
36067
|
-
* }
|
|
36068
|
-
*/
|
|
36069
|
-
audio: Components.File;
|
|
36070
|
-
}
|
|
36522
|
+
export interface OrpheusTtsOutput extends SharedType_d2a {}
|
|
36071
36523
|
|
|
36072
36524
|
export interface OnerewardInput {
|
|
36073
36525
|
/**
|
|
@@ -36266,6 +36718,122 @@ export interface OmnipartOutput {
|
|
|
36266
36718
|
seed: number;
|
|
36267
36719
|
}
|
|
36268
36720
|
|
|
36721
|
+
export interface OmnilottieVideoToLottieInput {
|
|
36722
|
+
/**
|
|
36723
|
+
* Max Tokens
|
|
36724
|
+
* @description Maximum number of Lottie tokens to generate.
|
|
36725
|
+
* @default 4096
|
|
36726
|
+
*/
|
|
36727
|
+
max_tokens?: number;
|
|
36728
|
+
/**
|
|
36729
|
+
* Prompt
|
|
36730
|
+
* @description Optional text description guiding the conversion.
|
|
36731
|
+
*/
|
|
36732
|
+
prompt?: string;
|
|
36733
|
+
/**
|
|
36734
|
+
* Temperature
|
|
36735
|
+
* @description Sampling temperature for generation.
|
|
36736
|
+
* @default 0.9
|
|
36737
|
+
*/
|
|
36738
|
+
temperature?: number;
|
|
36739
|
+
/**
|
|
36740
|
+
* Top K
|
|
36741
|
+
* @description Top-k sampling parameter.
|
|
36742
|
+
* @default 5
|
|
36743
|
+
*/
|
|
36744
|
+
top_k?: number;
|
|
36745
|
+
/**
|
|
36746
|
+
* Top P
|
|
36747
|
+
* @description Nucleus sampling probability threshold.
|
|
36748
|
+
* @default 0.25
|
|
36749
|
+
*/
|
|
36750
|
+
top_p?: number;
|
|
36751
|
+
/**
|
|
36752
|
+
* Video Url
|
|
36753
|
+
* @description URL of the video to convert into a Lottie animation.
|
|
36754
|
+
* @example https://storage.googleapis.com/falserverless/gallery/0dd62535b46d6f8ecf42f22dc0e148cd.mp4
|
|
36755
|
+
*/
|
|
36756
|
+
video_url: string;
|
|
36757
|
+
}
|
|
36758
|
+
|
|
36759
|
+
export interface OmnilottieVideoToLottieOutput extends SharedType_7c2 {}
|
|
36760
|
+
|
|
36761
|
+
export interface OmnilottieImageToLottieInput {
|
|
36762
|
+
/**
|
|
36763
|
+
* Image Url
|
|
36764
|
+
* @description URL of the reference image to animate.
|
|
36765
|
+
* @example https://storage.googleapis.com/falserverless/gallery/00de75e2c031cb3fc3f472e356aba5b6.png
|
|
36766
|
+
*/
|
|
36767
|
+
image_url: string;
|
|
36768
|
+
/**
|
|
36769
|
+
* Max Tokens
|
|
36770
|
+
* @description Maximum number of Lottie tokens to generate.
|
|
36771
|
+
* @default 4096
|
|
36772
|
+
*/
|
|
36773
|
+
max_tokens?: number;
|
|
36774
|
+
/**
|
|
36775
|
+
* Prompt
|
|
36776
|
+
* @description Text description guiding the animation of the image.
|
|
36777
|
+
* @example A cyan-colored, elongated shape resembling a smiley face, consisting of two pink oval eyes with navy blue circular pupils positioned symmetrically. The shape rotates gradually around its horizontal axis, revealing small animated sparklesâpink and tealâthat appear sporadically near the eyes, enhancing the dynamic visual effect.
|
|
36778
|
+
*/
|
|
36779
|
+
prompt: string;
|
|
36780
|
+
/**
|
|
36781
|
+
* Temperature
|
|
36782
|
+
* @description Sampling temperature for generation.
|
|
36783
|
+
* @default 0.9
|
|
36784
|
+
*/
|
|
36785
|
+
temperature?: number;
|
|
36786
|
+
/**
|
|
36787
|
+
* Top K
|
|
36788
|
+
* @description Top-k sampling parameter.
|
|
36789
|
+
* @default 5
|
|
36790
|
+
*/
|
|
36791
|
+
top_k?: number;
|
|
36792
|
+
/**
|
|
36793
|
+
* Top P
|
|
36794
|
+
* @description Nucleus sampling probability threshold.
|
|
36795
|
+
* @default 0.25
|
|
36796
|
+
*/
|
|
36797
|
+
top_p?: number;
|
|
36798
|
+
}
|
|
36799
|
+
|
|
36800
|
+
export interface OmnilottieImageToLottieOutput extends SharedType_7c2 {}
|
|
36801
|
+
|
|
36802
|
+
export interface OmnilottieInput {
|
|
36803
|
+
/**
|
|
36804
|
+
* Max Tokens
|
|
36805
|
+
* @description Maximum number of Lottie tokens to generate.
|
|
36806
|
+
* @default 4096
|
|
36807
|
+
*/
|
|
36808
|
+
max_tokens?: number;
|
|
36809
|
+
/**
|
|
36810
|
+
* Prompt
|
|
36811
|
+
* @description Text description of the Lottie animation to generate.
|
|
36812
|
+
* @example A red ball bouncing up and down
|
|
36813
|
+
*/
|
|
36814
|
+
prompt: string;
|
|
36815
|
+
/**
|
|
36816
|
+
* Temperature
|
|
36817
|
+
* @description Sampling temperature for generation.
|
|
36818
|
+
* @default 0.9
|
|
36819
|
+
*/
|
|
36820
|
+
temperature?: number;
|
|
36821
|
+
/**
|
|
36822
|
+
* Top K
|
|
36823
|
+
* @description Top-k sampling parameter.
|
|
36824
|
+
* @default 5
|
|
36825
|
+
*/
|
|
36826
|
+
top_k?: number;
|
|
36827
|
+
/**
|
|
36828
|
+
* Top P
|
|
36829
|
+
* @description Nucleus sampling probability threshold.
|
|
36830
|
+
* @default 0.25
|
|
36831
|
+
*/
|
|
36832
|
+
top_p?: number;
|
|
36833
|
+
}
|
|
36834
|
+
|
|
36835
|
+
export interface OmnilottieOutput extends SharedType_7c2 {}
|
|
36836
|
+
|
|
36269
36837
|
export interface OmnigenV2Input {
|
|
36270
36838
|
/**
|
|
36271
36839
|
* Cfg Range End
|
|
@@ -37043,10 +37611,10 @@ export interface Moondream3PreviewQueryOutput {
|
|
|
37043
37611
|
* @description Usage information for the request
|
|
37044
37612
|
* @example {
|
|
37045
37613
|
* "output_tokens": 23,
|
|
37046
|
-
* "
|
|
37614
|
+
* "prefill_time_ms": 54.45315001998097,
|
|
37047
37615
|
* "input_tokens": 737,
|
|
37048
37616
|
* "ttft_ms": 91.87838807702065,
|
|
37049
|
-
* "
|
|
37617
|
+
* "decode_time_ms": 811.5944429300725
|
|
37050
37618
|
* }
|
|
37051
37619
|
*/
|
|
37052
37620
|
usage_info: Components.UsageInfo_1;
|
|
@@ -37143,10 +37711,10 @@ export interface Moondream3PreviewPointOutput {
|
|
|
37143
37711
|
* @description Usage information for the request
|
|
37144
37712
|
* @example {
|
|
37145
37713
|
* "output_tokens": 23,
|
|
37146
|
-
* "
|
|
37714
|
+
* "prefill_time_ms": 54.45315001998097,
|
|
37147
37715
|
* "input_tokens": 737,
|
|
37148
37716
|
* "ttft_ms": 91.87838807702065,
|
|
37149
|
-
* "
|
|
37717
|
+
* "decode_time_ms": 811.5944429300725
|
|
37150
37718
|
* }
|
|
37151
37719
|
*/
|
|
37152
37720
|
usage_info: Components.UsageInfo_1;
|
|
@@ -37211,10 +37779,10 @@ export interface Moondream3PreviewDetectOutput {
|
|
|
37211
37779
|
* @description Usage information for the request
|
|
37212
37780
|
* @example {
|
|
37213
37781
|
* "output_tokens": 23,
|
|
37214
|
-
* "
|
|
37782
|
+
* "prefill_time_ms": 54.45315001998097,
|
|
37215
37783
|
* "input_tokens": 737,
|
|
37216
37784
|
* "ttft_ms": 91.87838807702065,
|
|
37217
|
-
* "
|
|
37785
|
+
* "decode_time_ms": 811.5944429300725
|
|
37218
37786
|
* }
|
|
37219
37787
|
*/
|
|
37220
37788
|
usage_info: Components.UsageInfo_1;
|
|
@@ -37263,10 +37831,10 @@ export interface Moondream3PreviewCaptionOutput {
|
|
|
37263
37831
|
* @description Usage information for the request
|
|
37264
37832
|
* @example {
|
|
37265
37833
|
* "output_tokens": 23,
|
|
37266
|
-
* "
|
|
37834
|
+
* "prefill_time_ms": 54.45315001998097,
|
|
37267
37835
|
* "input_tokens": 737,
|
|
37268
37836
|
* "ttft_ms": 91.87838807702065,
|
|
37269
|
-
* "
|
|
37837
|
+
* "decode_time_ms": 811.5944429300725
|
|
37270
37838
|
* }
|
|
37271
37839
|
*/
|
|
37272
37840
|
usage_info: Components.UsageInfo_1;
|
|
@@ -40350,6 +40918,12 @@ export interface Ltx2VideoTrainerInput {
|
|
|
40350
40918
|
* @example false
|
|
40351
40919
|
*/
|
|
40352
40920
|
auto_scale_input?: boolean;
|
|
40921
|
+
/**
|
|
40922
|
+
* Debug Dataset
|
|
40923
|
+
* @description When enabled, the trainer returns a downloadable archive of your preprocessed training data for manual inspection. Use this to verify that your videos, images, and captions were processed correctly before committing to a full training run.
|
|
40924
|
+
* @default false
|
|
40925
|
+
*/
|
|
40926
|
+
debug_dataset?: boolean;
|
|
40353
40927
|
/**
|
|
40354
40928
|
* First Frame Conditioning P
|
|
40355
40929
|
* @description Probability of conditioning on the first frame during training. Higher values improve image-to-video performance.
|
|
@@ -40512,6 +41086,12 @@ export interface Ltx2V2vTrainerInput {
|
|
|
40512
41086
|
* @example false
|
|
40513
41087
|
*/
|
|
40514
41088
|
auto_scale_input?: boolean;
|
|
41089
|
+
/**
|
|
41090
|
+
* Debug Dataset
|
|
41091
|
+
* @description When enabled, the trainer returns a downloadable archive of your preprocessed training data for manual inspection. Use this to verify that your videos, images, and captions were processed correctly before committing to a full training run.
|
|
41092
|
+
* @default false
|
|
41093
|
+
*/
|
|
41094
|
+
debug_dataset?: boolean;
|
|
40515
41095
|
/**
|
|
40516
41096
|
* First Frame Conditioning P
|
|
40517
41097
|
* @description Probability of conditioning on the first frame during training. Lower values work better for video-to-video transformation.
|
|
@@ -42797,7 +43377,7 @@ export interface Ltx219bVideoToVideoLoraInput {
|
|
|
42797
43377
|
* LoRAs
|
|
42798
43378
|
* @description The LoRAs to use for the generation.
|
|
42799
43379
|
*/
|
|
42800
|
-
loras: Components.
|
|
43380
|
+
loras: Components.LoRAInput_1[];
|
|
42801
43381
|
/**
|
|
42802
43382
|
* Match Input FPS
|
|
42803
43383
|
* @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
|
|
@@ -42910,7 +43490,7 @@ export interface Ltx219bVideoToVideoLoraInput {
|
|
|
42910
43490
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
42911
43491
|
}
|
|
42912
43492
|
|
|
42913
|
-
export interface Ltx219bVideoToVideoLoraOutput extends
|
|
43493
|
+
export interface Ltx219bVideoToVideoLoraOutput extends SharedType_3b0 {}
|
|
42914
43494
|
|
|
42915
43495
|
export interface Ltx219bVideoToVideoInput {
|
|
42916
43496
|
/**
|
|
@@ -43132,7 +43712,7 @@ export interface Ltx219bVideoToVideoInput {
|
|
|
43132
43712
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
43133
43713
|
}
|
|
43134
43714
|
|
|
43135
|
-
export interface Ltx219bVideoToVideoOutput extends
|
|
43715
|
+
export interface Ltx219bVideoToVideoOutput extends SharedType_3b0 {}
|
|
43136
43716
|
|
|
43137
43717
|
export interface Ltx219bTextToVideoLoraInput {
|
|
43138
43718
|
/**
|
|
@@ -43199,7 +43779,7 @@ export interface Ltx219bTextToVideoLoraInput {
|
|
|
43199
43779
|
* LoRAs
|
|
43200
43780
|
* @description The LoRAs to use for the generation.
|
|
43201
43781
|
*/
|
|
43202
|
-
loras: Components.
|
|
43782
|
+
loras: Components.LoRAInput_1[];
|
|
43203
43783
|
/**
|
|
43204
43784
|
* Negative Prompt
|
|
43205
43785
|
* @description The negative prompt to generate the video from.
|
|
@@ -43279,7 +43859,7 @@ export interface Ltx219bTextToVideoLoraInput {
|
|
|
43279
43859
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
43280
43860
|
}
|
|
43281
43861
|
|
|
43282
|
-
export interface Ltx219bTextToVideoLoraOutput extends
|
|
43862
|
+
export interface Ltx219bTextToVideoLoraOutput extends SharedType_857 {}
|
|
43283
43863
|
|
|
43284
43864
|
export interface Ltx219bTextToVideoInput {
|
|
43285
43865
|
/**
|
|
@@ -43421,7 +44001,7 @@ export interface Ltx219bTextToVideoInput {
|
|
|
43421
44001
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
43422
44002
|
}
|
|
43423
44003
|
|
|
43424
|
-
export interface Ltx219bTextToVideoOutput extends
|
|
44004
|
+
export interface Ltx219bTextToVideoOutput extends SharedType_857 {}
|
|
43425
44005
|
|
|
43426
44006
|
export interface Ltx219bImageToVideoLoraInput {
|
|
43427
44007
|
/**
|
|
@@ -43518,7 +44098,7 @@ export interface Ltx219bImageToVideoLoraInput {
|
|
|
43518
44098
|
* LoRAs
|
|
43519
44099
|
* @description The LoRAs to use for the generation.
|
|
43520
44100
|
*/
|
|
43521
|
-
loras: Components.
|
|
44101
|
+
loras: Components.LoRAInput_1[];
|
|
43522
44102
|
/**
|
|
43523
44103
|
* Negative Prompt
|
|
43524
44104
|
* @description The negative prompt to generate the video from.
|
|
@@ -43599,7 +44179,7 @@ export interface Ltx219bImageToVideoLoraInput {
|
|
|
43599
44179
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
43600
44180
|
}
|
|
43601
44181
|
|
|
43602
|
-
export interface Ltx219bImageToVideoLoraOutput extends
|
|
44182
|
+
export interface Ltx219bImageToVideoLoraOutput extends SharedType_369 {}
|
|
43603
44183
|
|
|
43604
44184
|
export interface Ltx219bImageToVideoInput {
|
|
43605
44185
|
/**
|
|
@@ -43772,7 +44352,7 @@ export interface Ltx219bImageToVideoInput {
|
|
|
43772
44352
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
43773
44353
|
}
|
|
43774
44354
|
|
|
43775
|
-
export interface Ltx219bImageToVideoOutput extends
|
|
44355
|
+
export interface Ltx219bImageToVideoOutput extends SharedType_369 {}
|
|
43776
44356
|
|
|
43777
44357
|
export interface Ltx219bExtendVideoLoraInput {
|
|
43778
44358
|
/**
|
|
@@ -43863,7 +44443,7 @@ export interface Ltx219bExtendVideoLoraInput {
|
|
|
43863
44443
|
* LoRAs
|
|
43864
44444
|
* @description The LoRAs to use for the generation.
|
|
43865
44445
|
*/
|
|
43866
|
-
loras: Components.
|
|
44446
|
+
loras: Components.LoRAInput_1[];
|
|
43867
44447
|
/**
|
|
43868
44448
|
* Match Input FPS
|
|
43869
44449
|
* @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
|
|
@@ -43968,7 +44548,7 @@ export interface Ltx219bExtendVideoLoraInput {
|
|
|
43968
44548
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
43969
44549
|
}
|
|
43970
44550
|
|
|
43971
|
-
export interface Ltx219bExtendVideoLoraOutput extends
|
|
44551
|
+
export interface Ltx219bExtendVideoLoraOutput extends SharedType_cb6 {}
|
|
43972
44552
|
|
|
43973
44553
|
export interface Ltx219bExtendVideoInput {
|
|
43974
44554
|
/**
|
|
@@ -44159,7 +44739,7 @@ export interface Ltx219bExtendVideoInput {
|
|
|
44159
44739
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
44160
44740
|
}
|
|
44161
44741
|
|
|
44162
|
-
export interface Ltx219bExtendVideoOutput extends
|
|
44742
|
+
export interface Ltx219bExtendVideoOutput extends SharedType_cb6 {}
|
|
44163
44743
|
|
|
44164
44744
|
export interface Ltx219bDistilledVideoToVideoLoraInput {
|
|
44165
44745
|
/**
|
|
@@ -44267,7 +44847,7 @@ export interface Ltx219bDistilledVideoToVideoLoraInput {
|
|
|
44267
44847
|
* LoRAs
|
|
44268
44848
|
* @description The LoRAs to use for the generation.
|
|
44269
44849
|
*/
|
|
44270
|
-
loras: Components.
|
|
44850
|
+
loras: Components.LoRAInput_1[];
|
|
44271
44851
|
/**
|
|
44272
44852
|
* Match Input FPS
|
|
44273
44853
|
* @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
|
|
@@ -44374,7 +44954,7 @@ export interface Ltx219bDistilledVideoToVideoLoraInput {
|
|
|
44374
44954
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
44375
44955
|
}
|
|
44376
44956
|
|
|
44377
|
-
export interface Ltx219bDistilledVideoToVideoLoraOutput extends
|
|
44957
|
+
export interface Ltx219bDistilledVideoToVideoLoraOutput extends SharedType_3b0 {}
|
|
44378
44958
|
|
|
44379
44959
|
export interface Ltx219bDistilledVideoToVideoInput {
|
|
44380
44960
|
/**
|
|
@@ -44584,7 +45164,7 @@ export interface Ltx219bDistilledVideoToVideoInput {
|
|
|
44584
45164
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
44585
45165
|
}
|
|
44586
45166
|
|
|
44587
|
-
export interface Ltx219bDistilledVideoToVideoOutput extends
|
|
45167
|
+
export interface Ltx219bDistilledVideoToVideoOutput extends SharedType_3b0 {}
|
|
44588
45168
|
|
|
44589
45169
|
export interface Ltx219bDistilledTextToVideoLoraInput {
|
|
44590
45170
|
/**
|
|
@@ -44645,7 +45225,7 @@ export interface Ltx219bDistilledTextToVideoLoraInput {
|
|
|
44645
45225
|
* LoRAs
|
|
44646
45226
|
* @description The LoRAs to use for the generation.
|
|
44647
45227
|
*/
|
|
44648
|
-
loras: Components.
|
|
45228
|
+
loras: Components.LoRAInput_1[];
|
|
44649
45229
|
/**
|
|
44650
45230
|
* Negative Prompt
|
|
44651
45231
|
* @description The negative prompt to generate the video from.
|
|
@@ -44719,7 +45299,7 @@ export interface Ltx219bDistilledTextToVideoLoraInput {
|
|
|
44719
45299
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
44720
45300
|
}
|
|
44721
45301
|
|
|
44722
|
-
export interface Ltx219bDistilledTextToVideoLoraOutput extends
|
|
45302
|
+
export interface Ltx219bDistilledTextToVideoLoraOutput extends SharedType_857 {}
|
|
44723
45303
|
|
|
44724
45304
|
export interface Ltx219bDistilledTextToVideoInput {
|
|
44725
45305
|
/**
|
|
@@ -44849,7 +45429,7 @@ export interface Ltx219bDistilledTextToVideoInput {
|
|
|
44849
45429
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
44850
45430
|
}
|
|
44851
45431
|
|
|
44852
|
-
export interface Ltx219bDistilledTextToVideoOutput extends
|
|
45432
|
+
export interface Ltx219bDistilledTextToVideoOutput extends SharedType_857 {}
|
|
44853
45433
|
|
|
44854
45434
|
export interface Ltx219bDistilledImageToVideoLoraInput {
|
|
44855
45435
|
/**
|
|
@@ -44940,7 +45520,7 @@ export interface Ltx219bDistilledImageToVideoLoraInput {
|
|
|
44940
45520
|
* LoRAs
|
|
44941
45521
|
* @description The LoRAs to use for the generation.
|
|
44942
45522
|
*/
|
|
44943
|
-
loras: Components.
|
|
45523
|
+
loras: Components.LoRAInput_1[];
|
|
44944
45524
|
/**
|
|
44945
45525
|
* Negative Prompt
|
|
44946
45526
|
* @description The negative prompt to generate the video from.
|
|
@@ -45015,7 +45595,7 @@ export interface Ltx219bDistilledImageToVideoLoraInput {
|
|
|
45015
45595
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
45016
45596
|
}
|
|
45017
45597
|
|
|
45018
|
-
export interface Ltx219bDistilledImageToVideoLoraOutput extends
|
|
45598
|
+
export interface Ltx219bDistilledImageToVideoLoraOutput extends SharedType_369 {}
|
|
45019
45599
|
|
|
45020
45600
|
export interface Ltx219bDistilledImageToVideoInput {
|
|
45021
45601
|
/**
|
|
@@ -45176,7 +45756,7 @@ export interface Ltx219bDistilledImageToVideoInput {
|
|
|
45176
45756
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
45177
45757
|
}
|
|
45178
45758
|
|
|
45179
|
-
export interface Ltx219bDistilledImageToVideoOutput extends
|
|
45759
|
+
export interface Ltx219bDistilledImageToVideoOutput extends SharedType_369 {}
|
|
45180
45760
|
|
|
45181
45761
|
export interface Ltx219bDistilledExtendVideoLoraInput {
|
|
45182
45762
|
/**
|
|
@@ -45261,7 +45841,7 @@ export interface Ltx219bDistilledExtendVideoLoraInput {
|
|
|
45261
45841
|
* LoRAs
|
|
45262
45842
|
* @description The LoRAs to use for the generation.
|
|
45263
45843
|
*/
|
|
45264
|
-
loras: Components.
|
|
45844
|
+
loras: Components.LoRAInput_1[];
|
|
45265
45845
|
/**
|
|
45266
45846
|
* Match Input FPS
|
|
45267
45847
|
* @description When true, match the output FPS to the input video's FPS instead of using the default target FPS.
|
|
@@ -45360,7 +45940,7 @@ export interface Ltx219bDistilledExtendVideoLoraInput {
|
|
|
45360
45940
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
45361
45941
|
}
|
|
45362
45942
|
|
|
45363
|
-
export interface Ltx219bDistilledExtendVideoLoraOutput extends
|
|
45943
|
+
export interface Ltx219bDistilledExtendVideoLoraOutput extends SharedType_cb6 {}
|
|
45364
45944
|
|
|
45365
45945
|
export interface Ltx219bDistilledExtendVideoInput {
|
|
45366
45946
|
/**
|
|
@@ -45539,7 +46119,7 @@ export interface Ltx219bDistilledExtendVideoInput {
|
|
|
45539
46119
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
45540
46120
|
}
|
|
45541
46121
|
|
|
45542
|
-
export interface Ltx219bDistilledExtendVideoOutput extends
|
|
46122
|
+
export interface Ltx219bDistilledExtendVideoOutput extends SharedType_cb6 {}
|
|
45543
46123
|
|
|
45544
46124
|
export interface Ltx219bDistilledAudioToVideoLoraInput {
|
|
45545
46125
|
/**
|
|
@@ -45629,7 +46209,7 @@ export interface Ltx219bDistilledAudioToVideoLoraInput {
|
|
|
45629
46209
|
* LoRAs
|
|
45630
46210
|
* @description The LoRAs to use for the generation.
|
|
45631
46211
|
*/
|
|
45632
|
-
loras: Components.
|
|
46212
|
+
loras: Components.LoRAInput_1[];
|
|
45633
46213
|
/**
|
|
45634
46214
|
* Match Audio Length
|
|
45635
46215
|
* @description When enabled, the number of frames will be calculated based on the audio duration and FPS. When disabled, use the specified num_frames.
|
|
@@ -45984,7 +46564,7 @@ export interface Ltx219bAudioToVideoLoraInput {
|
|
|
45984
46564
|
* LoRAs
|
|
45985
46565
|
* @description The LoRAs to use for the generation.
|
|
45986
46566
|
*/
|
|
45987
|
-
loras: Components.
|
|
46567
|
+
loras: Components.LoRAInput_1[];
|
|
45988
46568
|
/**
|
|
45989
46569
|
* Match Audio Length
|
|
45990
46570
|
* @description When enabled, the number of frames will be calculated based on the audio duration and FPS. When disabled, use the specified num_frames.
|
|
@@ -53389,23 +53969,7 @@ export interface InfinitalkSingleTextInput {
|
|
|
53389
53969
|
| 'Bill';
|
|
53390
53970
|
}
|
|
53391
53971
|
|
|
53392
|
-
export interface InfinitalkSingleTextOutput {
|
|
53393
|
-
/**
|
|
53394
|
-
* Seed
|
|
53395
|
-
* @description The seed used for generation.
|
|
53396
|
-
*/
|
|
53397
|
-
seed: number;
|
|
53398
|
-
/**
|
|
53399
|
-
* @description The generated video file.
|
|
53400
|
-
* @example {
|
|
53401
|
-
* "file_size": 797478,
|
|
53402
|
-
* "file_name": "6c9dd31e1d9a4482877747a52a661a0a.mp4",
|
|
53403
|
-
* "content_type": "application/octet-stream",
|
|
53404
|
-
* "url": "https://v3.fal.media/files/elephant/-huMN0zTaXmBr2CqzCMps_6c9dd31e1d9a4482877747a52a661a0a.mp4"
|
|
53405
|
-
* }
|
|
53406
|
-
*/
|
|
53407
|
-
video: Components.File;
|
|
53408
|
-
}
|
|
53972
|
+
export interface InfinitalkSingleTextOutput extends SharedType_58d {}
|
|
53409
53973
|
|
|
53410
53974
|
export interface InfinitalkInput {
|
|
53411
53975
|
/**
|
|
@@ -53454,23 +54018,7 @@ export interface InfinitalkInput {
|
|
|
53454
54018
|
seed?: number;
|
|
53455
54019
|
}
|
|
53456
54020
|
|
|
53457
|
-
export interface InfinitalkOutput {
|
|
53458
|
-
/**
|
|
53459
|
-
* Seed
|
|
53460
|
-
* @description The seed used for generation.
|
|
53461
|
-
*/
|
|
53462
|
-
seed: number;
|
|
53463
|
-
/**
|
|
53464
|
-
* @description The generated video file.
|
|
53465
|
-
* @example {
|
|
53466
|
-
* "file_size": 515275,
|
|
53467
|
-
* "file_name": "74af6c0bdd6041c3b1130d54885e3eee.mp4",
|
|
53468
|
-
* "content_type": "application/octet-stream",
|
|
53469
|
-
* "url": "https://v3.fal.media/files/kangaroo/z6VqUwNTwzuWa6YE1g7In_74af6c0bdd6041c3b1130d54885e3eee.mp4"
|
|
53470
|
-
* }
|
|
53471
|
-
*/
|
|
53472
|
-
video: Components.File;
|
|
53473
|
-
}
|
|
54021
|
+
export interface InfinitalkOutput extends SharedType_0c6 {}
|
|
53474
54022
|
|
|
53475
54023
|
export interface IndexTts2TextToSpeechInput {
|
|
53476
54024
|
/**
|
|
@@ -53545,13 +54093,7 @@ export interface ImageutilsRembgInput {
|
|
|
53545
54093
|
sync_mode?: boolean;
|
|
53546
54094
|
}
|
|
53547
54095
|
|
|
53548
|
-
export interface ImageutilsRembgOutput {
|
|
53549
|
-
/**
|
|
53550
|
-
* Image
|
|
53551
|
-
* @description Background removed image.
|
|
53552
|
-
*/
|
|
53553
|
-
image: Components.Image_2;
|
|
53554
|
-
}
|
|
54096
|
+
export interface ImageutilsRembgOutput extends SharedType_357 {}
|
|
53555
54097
|
|
|
53556
54098
|
export interface ImageutilsNsfwInput {
|
|
53557
54099
|
/**
|
|
@@ -53600,7 +54142,7 @@ export interface ImageutilsMarigoldDepthInput {
|
|
|
53600
54142
|
processing_res?: number;
|
|
53601
54143
|
}
|
|
53602
54144
|
|
|
53603
|
-
export interface ImageutilsMarigoldDepthOutput extends
|
|
54145
|
+
export interface ImageutilsMarigoldDepthOutput extends SharedType_357 {}
|
|
53604
54146
|
|
|
53605
54147
|
export interface ImageutilsDepthInput {
|
|
53606
54148
|
/**
|
|
@@ -53629,7 +54171,7 @@ export interface ImageutilsDepthInput {
|
|
|
53629
54171
|
image_url: string;
|
|
53630
54172
|
}
|
|
53631
54173
|
|
|
53632
|
-
export interface ImageutilsDepthOutput extends
|
|
54174
|
+
export interface ImageutilsDepthOutput extends SharedType_357 {}
|
|
53633
54175
|
|
|
53634
54176
|
export interface Imagen4PreviewUltraInput {
|
|
53635
54177
|
/**
|
|
@@ -58006,17 +58548,17 @@ export interface Hunyuan3dV3TextTo3dOutput {
|
|
|
58006
58548
|
/**
|
|
58007
58549
|
* @description URLs for different 3D model formats
|
|
58008
58550
|
* @example {
|
|
58009
|
-
* "glb": {
|
|
58010
|
-
* "file_size": 64724836,
|
|
58011
|
-
* "file_name": "model.glb",
|
|
58012
|
-
* "content_type": "model/gltf-binary",
|
|
58013
|
-
* "url": "https://v3b.fal.media/files/b/0a8686a8/1hPquv3AqqkfnqSM9fpmB_model.glb"
|
|
58014
|
-
* },
|
|
58015
58551
|
* "obj": {
|
|
58016
58552
|
* "file_size": 44084728,
|
|
58017
58553
|
* "file_name": "model.obj",
|
|
58018
58554
|
* "content_type": "text/plain",
|
|
58019
58555
|
* "url": "https://v3b.fal.media/files/b/0a8686a8/AVgdsVFrGAKGAFr4e2g56_model.obj"
|
|
58556
|
+
* },
|
|
58557
|
+
* "glb": {
|
|
58558
|
+
* "file_size": 64724836,
|
|
58559
|
+
* "file_name": "model.glb",
|
|
58560
|
+
* "content_type": "model/gltf-binary",
|
|
58561
|
+
* "url": "https://v3b.fal.media/files/b/0a8686a8/1hPquv3AqqkfnqSM9fpmB_model.glb"
|
|
58020
58562
|
* }
|
|
58021
58563
|
* }
|
|
58022
58564
|
*/
|
|
@@ -58079,17 +58621,17 @@ export interface Hunyuan3dV3SketchTo3dOutput {
|
|
|
58079
58621
|
/**
|
|
58080
58622
|
* @description URLs for different 3D model formats
|
|
58081
58623
|
* @example {
|
|
58082
|
-
* "glb": {
|
|
58083
|
-
* "file_size": 30655724,
|
|
58084
|
-
* "file_name": "model.glb",
|
|
58085
|
-
* "content_type": "model/gltf-binary",
|
|
58086
|
-
* "url": "https://v3b.fal.media/files/b/0a8688bb/vd2SlBP92cZls3zG5EPbg_model.glb"
|
|
58087
|
-
* },
|
|
58088
58624
|
* "obj": {
|
|
58089
58625
|
* "file_size": 23418473,
|
|
58090
58626
|
* "file_name": "model.obj",
|
|
58091
58627
|
* "content_type": "text/plain",
|
|
58092
58628
|
* "url": "https://v3b.fal.media/files/b/0a8688bb/QNik1DVxzvj23YEF3vhs__model.obj"
|
|
58629
|
+
* },
|
|
58630
|
+
* "glb": {
|
|
58631
|
+
* "file_size": 30655724,
|
|
58632
|
+
* "file_name": "model.glb",
|
|
58633
|
+
* "content_type": "model/gltf-binary",
|
|
58634
|
+
* "url": "https://v3b.fal.media/files/b/0a8688bb/vd2SlBP92cZls3zG5EPbg_model.glb"
|
|
58093
58635
|
* }
|
|
58094
58636
|
* }
|
|
58095
58637
|
*/
|
|
@@ -58175,17 +58717,17 @@ export interface Hunyuan3dV3ImageTo3dOutput {
|
|
|
58175
58717
|
/**
|
|
58176
58718
|
* @description URLs for different 3D model formats
|
|
58177
58719
|
* @example {
|
|
58178
|
-
* "glb": {
|
|
58179
|
-
* "file_size": 64122888,
|
|
58180
|
-
* "file_name": "model.glb",
|
|
58181
|
-
* "content_type": "model/gltf-binary",
|
|
58182
|
-
* "url": "https://v3b.fal.media/files/b/0a8686ae/MQN_KtP32PbqtPr_VLcyp_model.glb"
|
|
58183
|
-
* },
|
|
58184
58720
|
* "obj": {
|
|
58185
58721
|
* "file_size": 42886419,
|
|
58186
58722
|
* "file_name": "model.obj",
|
|
58187
58723
|
* "content_type": "text/plain",
|
|
58188
58724
|
* "url": "https://v3b.fal.media/files/b/0a8686ad/ifdJskhUfQysq-NN20iQR_model.obj"
|
|
58725
|
+
* },
|
|
58726
|
+
* "glb": {
|
|
58727
|
+
* "file_size": 64122888,
|
|
58728
|
+
* "file_name": "model.glb",
|
|
58729
|
+
* "content_type": "model/gltf-binary",
|
|
58730
|
+
* "url": "https://v3b.fal.media/files/b/0a8686ae/MQN_KtP32PbqtPr_VLcyp_model.glb"
|
|
58189
58731
|
* }
|
|
58190
58732
|
* }
|
|
58191
58733
|
*/
|
|
@@ -59472,17 +60014,17 @@ export interface Hunyuan3dV31SmartTopologyOutput {
|
|
|
59472
60014
|
/**
|
|
59473
60015
|
* @description URLs for different 3D model formats
|
|
59474
60016
|
* @example {
|
|
59475
|
-
* "glb": {
|
|
59476
|
-
* "file_size": 206004,
|
|
59477
|
-
* "file_name": "model.glb",
|
|
59478
|
-
* "content_type": "model/gltf-binary",
|
|
59479
|
-
* "url": "https://v3b.fal.media/files/b/0a8c0ab4/eX-_x0Wv8fZL05l9CGp6Y_model.glb"
|
|
59480
|
-
* },
|
|
59481
60017
|
* "obj": {
|
|
59482
60018
|
* "file_size": 394409,
|
|
59483
60019
|
* "file_name": "model.obj",
|
|
59484
60020
|
* "content_type": "model/obj",
|
|
59485
60021
|
* "url": "https://v3b.fal.media/files/b/0a8c0ab4/tqMY5NJLnHjpwN8rQ15dj_model.obj"
|
|
60022
|
+
* },
|
|
60023
|
+
* "glb": {
|
|
60024
|
+
* "file_size": 206004,
|
|
60025
|
+
* "file_name": "model.glb",
|
|
60026
|
+
* "content_type": "model/gltf-binary",
|
|
60027
|
+
* "url": "https://v3b.fal.media/files/b/0a8c0ab4/eX-_x0Wv8fZL05l9CGp6Y_model.glb"
|
|
59486
60028
|
* }
|
|
59487
60029
|
* }
|
|
59488
60030
|
*/
|
|
@@ -59719,17 +60261,17 @@ export interface Hunyuan3dV31ProTextTo3dOutput {
|
|
|
59719
60261
|
* "content_type": "text/plain",
|
|
59720
60262
|
* "url": "https://v3b.fal.media/files/b/0a8c5482/ZxJepsEkhM67VSugmZ7QT_material.mtl"
|
|
59721
60263
|
* },
|
|
59722
|
-
* "glb": {
|
|
59723
|
-
* "file_size": 35833072,
|
|
59724
|
-
* "file_name": "model.glb",
|
|
59725
|
-
* "content_type": "model/gltf-binary",
|
|
59726
|
-
* "url": "https://v3b.fal.media/files/b/0a8c5483/z6sbpr5wBRjqgnQlJM2Ot_model.glb"
|
|
59727
|
-
* },
|
|
59728
60264
|
* "obj": {
|
|
59729
60265
|
* "file_size": 34755929,
|
|
59730
60266
|
* "file_name": "6168030a8817075aaa55c94cc5145000.obj",
|
|
59731
60267
|
* "content_type": "model/obj",
|
|
59732
60268
|
* "url": "https://v3b.fal.media/files/b/0a8c5482/ZzC1xlOftyGQxhDkbZzVW_6168030a8817075aaa55c94cc5145000.obj"
|
|
60269
|
+
* },
|
|
60270
|
+
* "glb": {
|
|
60271
|
+
* "file_size": 35833072,
|
|
60272
|
+
* "file_name": "model.glb",
|
|
60273
|
+
* "content_type": "model/gltf-binary",
|
|
60274
|
+
* "url": "https://v3b.fal.media/files/b/0a8c5483/z6sbpr5wBRjqgnQlJM2Ot_model.glb"
|
|
59733
60275
|
* }
|
|
59734
60276
|
* }
|
|
59735
60277
|
*/
|
|
@@ -59828,17 +60370,17 @@ export interface Hunyuan3dV31ProImageTo3dOutput {
|
|
|
59828
60370
|
/**
|
|
59829
60371
|
* @description URLs for different 3D model formats
|
|
59830
60372
|
* @example {
|
|
59831
|
-
* "glb": {
|
|
59832
|
-
* "file_size": 38554640,
|
|
59833
|
-
* "file_name": "model.glb",
|
|
59834
|
-
* "content_type": "model/gltf-binary",
|
|
59835
|
-
* "url": "https://v3b.fal.media/files/b/0a8c3187/jOeZmtBuhdQMkDu65AkdT_model.glb"
|
|
59836
|
-
* },
|
|
59837
60373
|
* "obj": {
|
|
59838
60374
|
* "file_size": 31447160,
|
|
59839
60375
|
* "file_name": "model.obj",
|
|
59840
60376
|
* "content_type": "model/obj",
|
|
59841
60377
|
* "url": "https://v3b.fal.media/files/b/0a8c3186/no-aBFEDnOuthILfv-wzs_model.obj"
|
|
60378
|
+
* },
|
|
60379
|
+
* "glb": {
|
|
60380
|
+
* "file_size": 38554640,
|
|
60381
|
+
* "file_name": "model.glb",
|
|
60382
|
+
* "content_type": "model/gltf-binary",
|
|
60383
|
+
* "url": "https://v3b.fal.media/files/b/0a8c3187/jOeZmtBuhdQMkDu65AkdT_model.glb"
|
|
59842
60384
|
* }
|
|
59843
60385
|
* }
|
|
59844
60386
|
*/
|
|
@@ -59941,8 +60483,8 @@ export interface Hunyuan_worldOutput {
|
|
|
59941
60483
|
/**
|
|
59942
60484
|
* @description The generated panorama image.
|
|
59943
60485
|
* @example {
|
|
59944
|
-
* "height": 960,
|
|
59945
60486
|
* "file_size": 2738127,
|
|
60487
|
+
* "height": 960,
|
|
59946
60488
|
* "file_name": "5db7925423b44f2a98098cd8f7cad7ec.png",
|
|
59947
60489
|
* "content_type": "image/png",
|
|
59948
60490
|
* "url": "https://v3.fal.media/files/kangaroo/P2AmXuLlyDIsivqjV_rAr_5db7925423b44f2a98098cd8f7cad7ec.png",
|
|
@@ -59991,7 +60533,7 @@ export interface HidreamI1FullImageToImageInput {
|
|
|
59991
60533
|
* @description A list of LoRAs to apply to the model. Each LoRA specifies its path, scale, and optional weight name.
|
|
59992
60534
|
* @default []
|
|
59993
60535
|
*/
|
|
59994
|
-
loras?: Components.
|
|
60536
|
+
loras?: Components.LoraWeight_5[];
|
|
59995
60537
|
/**
|
|
59996
60538
|
* Negative Prompt
|
|
59997
60539
|
* @description The negative prompt to use. Use it to address details that you don't want
|
|
@@ -60118,7 +60660,7 @@ export interface HidreamI1FullInput {
|
|
|
60118
60660
|
* @description A list of LoRAs to apply to the model. Each LoRA specifies its path, scale, and optional weight name.
|
|
60119
60661
|
* @default []
|
|
60120
60662
|
*/
|
|
60121
|
-
loras?: Components.
|
|
60663
|
+
loras?: Components.LoraWeight_5[];
|
|
60122
60664
|
/**
|
|
60123
60665
|
* Negative Prompt
|
|
60124
60666
|
* @description The negative prompt to use. Use it to address details that you don't want
|
|
@@ -60167,7 +60709,7 @@ export interface HidreamI1FullInput {
|
|
|
60167
60709
|
sync_mode?: boolean;
|
|
60168
60710
|
}
|
|
60169
60711
|
|
|
60170
|
-
export interface HidreamI1FullOutput extends
|
|
60712
|
+
export interface HidreamI1FullOutput extends SharedType_7c6 {}
|
|
60171
60713
|
|
|
60172
60714
|
export interface HidreamI1FastInput {
|
|
60173
60715
|
/**
|
|
@@ -61106,6 +61648,115 @@ export interface GotOcrV2Output {
|
|
|
61106
61648
|
outputs: string[];
|
|
61107
61649
|
}
|
|
61108
61650
|
|
|
61651
|
+
export interface GoalForceInput {
|
|
61652
|
+
/**
|
|
61653
|
+
* Enable Safety Checker
|
|
61654
|
+
* @description Enable safety checker for generated content.
|
|
61655
|
+
* @default true
|
|
61656
|
+
*/
|
|
61657
|
+
enable_safety_checker?: boolean;
|
|
61658
|
+
/**
|
|
61659
|
+
* @description Indirect goal force applied to a target object. The model generates physics-plausible interactions to move the target in the specified direction. Specify either this, projectile_force, or both.
|
|
61660
|
+
* @example {
|
|
61661
|
+
* "y": 0.18,
|
|
61662
|
+
* "mass": -1,
|
|
61663
|
+
* "angle": 0,
|
|
61664
|
+
* "magnitude": 350,
|
|
61665
|
+
* "x": 0.22
|
|
61666
|
+
* }
|
|
61667
|
+
* @example {
|
|
61668
|
+
* "y": 0.49,
|
|
61669
|
+
* "mass": 3.5,
|
|
61670
|
+
* "angle": 180,
|
|
61671
|
+
* "magnitude": 350,
|
|
61672
|
+
* "x": 0.57
|
|
61673
|
+
* }
|
|
61674
|
+
*/
|
|
61675
|
+
goal_force?: Components.ForceParams;
|
|
61676
|
+
/**
|
|
61677
|
+
* Guidance Scale
|
|
61678
|
+
* @description Classifier-free guidance scale.
|
|
61679
|
+
* @default 5
|
|
61680
|
+
* @example 5
|
|
61681
|
+
*/
|
|
61682
|
+
guidance_scale?: number;
|
|
61683
|
+
/**
|
|
61684
|
+
* Image Url
|
|
61685
|
+
* @description URL of the input image (first frame). The image will be resized to 832x480.
|
|
61686
|
+
* @default https://raw.githubusercontent.com/brown-palm/goal-force/2d0209dc27333bffdb3f06df657ae05b11407615/datasets/examples/human-object-interaction/images/_bulb_tool.png
|
|
61687
|
+
* @example https://raw.githubusercontent.com/brown-palm/goal-force/2d0209dc27333bffdb3f06df657ae05b11407615/datasets/examples/human-object-interaction/images/_bulb_tool.png
|
|
61688
|
+
*/
|
|
61689
|
+
image_url?: string;
|
|
61690
|
+
/**
|
|
61691
|
+
* Negative Prompt
|
|
61692
|
+
* @description Negative prompt for generation.
|
|
61693
|
+
* @default 色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走
|
|
61694
|
+
*/
|
|
61695
|
+
negative_prompt?: string;
|
|
61696
|
+
/**
|
|
61697
|
+
* Num Inference Steps
|
|
61698
|
+
* @description Number of denoising steps. Higher values produce better quality but take longer.
|
|
61699
|
+
* @default 50
|
|
61700
|
+
* @example 50
|
|
61701
|
+
*/
|
|
61702
|
+
num_inference_steps?: number;
|
|
61703
|
+
/**
|
|
61704
|
+
* @description Direct force applied to a projectile object, causing it to move along the force direction. Specify either this, goal_force, or both.
|
|
61705
|
+
* @example {
|
|
61706
|
+
* "y": 0.18,
|
|
61707
|
+
* "mass": -1,
|
|
61708
|
+
* "angle": 0,
|
|
61709
|
+
* "magnitude": 350,
|
|
61710
|
+
* "x": 0.22
|
|
61711
|
+
* }
|
|
61712
|
+
*/
|
|
61713
|
+
projectile_force?: Components.ForceParams;
|
|
61714
|
+
/**
|
|
61715
|
+
* Prompt
|
|
61716
|
+
* @description Text description of the scene and the desired physics interaction.
|
|
61717
|
+
* @default A hand pushes the hanging bulb, causing it to sway.
|
|
61718
|
+
* @example A hand pushes the hanging bulb, causing it to sway.
|
|
61719
|
+
* @example A soccer ball is kicked and rolls across the field.
|
|
61720
|
+
* @example A cue stick strikes a pool ball, causing it to roll across the table.
|
|
61721
|
+
* @example A golf club hits a golf ball.
|
|
61722
|
+
*/
|
|
61723
|
+
prompt?: string;
|
|
61724
|
+
/**
|
|
61725
|
+
* Seed
|
|
61726
|
+
* @description Random seed for reproducibility.
|
|
61727
|
+
*/
|
|
61728
|
+
seed?: number;
|
|
61729
|
+
}
|
|
61730
|
+
|
|
61731
|
+
export interface GoalForceOutput {
|
|
61732
|
+
/**
|
|
61733
|
+
* Seed
|
|
61734
|
+
* @description The seed used for generation.
|
|
61735
|
+
* @example 5
|
|
61736
|
+
*/
|
|
61737
|
+
seed: number;
|
|
61738
|
+
/**
|
|
61739
|
+
* Timings
|
|
61740
|
+
* @description Timing information.
|
|
61741
|
+
* @example {
|
|
61742
|
+
* "inference": 234.79
|
|
61743
|
+
* }
|
|
61744
|
+
*/
|
|
61745
|
+
timings: {
|
|
61746
|
+
[key: string]: number;
|
|
61747
|
+
};
|
|
61748
|
+
/**
|
|
61749
|
+
* @description The generated video.
|
|
61750
|
+
* @example {
|
|
61751
|
+
* "file_size": 502438,
|
|
61752
|
+
* "file_name": "tmpbrjvor4y.mp4",
|
|
61753
|
+
* "content_type": "application/octet-stream",
|
|
61754
|
+
* "url": "https://v3b.fal.media/files/b/0a91e900/S4Q1EK9vvZW_HATdXnbBo_tmpbrjvor4y.mp4"
|
|
61755
|
+
* }
|
|
61756
|
+
*/
|
|
61757
|
+
video: Components.VideoFile;
|
|
61758
|
+
}
|
|
61759
|
+
|
|
61109
61760
|
export interface GlmImageImageToImageInput {
|
|
61110
61761
|
/**
|
|
61111
61762
|
* Enable Prompt Expansion
|
|
@@ -61684,7 +62335,7 @@ export interface FooocusUpscaleOrVaryInput {
|
|
|
61684
62335
|
* }
|
|
61685
62336
|
* ]
|
|
61686
62337
|
*/
|
|
61687
|
-
loras?: Components.
|
|
62338
|
+
loras?: Components.LoraWeight_4[];
|
|
61688
62339
|
/**
|
|
61689
62340
|
* Mixing Image Prompt and Vary/Upscale
|
|
61690
62341
|
* @description Mixing Image Prompt and Vary/Upscale
|
|
@@ -61759,9 +62410,9 @@ export interface FooocusUpscaleOrVaryInput {
|
|
|
61759
62410
|
* Styles
|
|
61760
62411
|
* @description The style to use.
|
|
61761
62412
|
* @default [
|
|
62413
|
+
* "Fooocus Enhance",
|
|
61762
62414
|
* "Fooocus V2",
|
|
61763
|
-
* "Fooocus Sharp"
|
|
61764
|
-
* "Fooocus Enhance"
|
|
62415
|
+
* "Fooocus Sharp"
|
|
61765
62416
|
* ]
|
|
61766
62417
|
*/
|
|
61767
62418
|
styles?: (
|
|
@@ -62172,7 +62823,7 @@ export interface FooocusInpaintInput {
|
|
|
62172
62823
|
* }
|
|
62173
62824
|
* ]
|
|
62174
62825
|
*/
|
|
62175
|
-
loras?: Components.
|
|
62826
|
+
loras?: Components.LoraWeight_4[];
|
|
62176
62827
|
/**
|
|
62177
62828
|
* Mask Image Url
|
|
62178
62829
|
* @description The image to use as a mask for the generated image.
|
|
@@ -62268,9 +62919,9 @@ export interface FooocusInpaintInput {
|
|
|
62268
62919
|
* Styles
|
|
62269
62920
|
* @description The style to use.
|
|
62270
62921
|
* @default [
|
|
62922
|
+
* "Fooocus Enhance",
|
|
62271
62923
|
* "Fooocus V2",
|
|
62272
|
-
* "Fooocus Sharp"
|
|
62273
|
-
* "Fooocus Enhance"
|
|
62924
|
+
* "Fooocus Sharp"
|
|
62274
62925
|
* ]
|
|
62275
62926
|
*/
|
|
62276
62927
|
styles?: (
|
|
@@ -62585,10 +63236,10 @@ export interface FooocusImagePromptInput {
|
|
|
62585
63236
|
guidance_scale?: number;
|
|
62586
63237
|
/**
|
|
62587
63238
|
* @example {
|
|
62588
|
-
* "
|
|
63239
|
+
* "image_url": "https://storage.googleapis.com/falserverless/model_tests/fooocus/Pikachu.webp",
|
|
62589
63240
|
* "stop_at": 1,
|
|
62590
63241
|
* "type": "PyraCanny",
|
|
62591
|
-
* "
|
|
63242
|
+
* "weight": 1
|
|
62592
63243
|
* }
|
|
62593
63244
|
*/
|
|
62594
63245
|
image_prompt_1: Components.ImagePrompt;
|
|
@@ -62627,7 +63278,7 @@ export interface FooocusImagePromptInput {
|
|
|
62627
63278
|
* }
|
|
62628
63279
|
* ]
|
|
62629
63280
|
*/
|
|
62630
|
-
loras?: Components.
|
|
63281
|
+
loras?: Components.LoraWeight_4[];
|
|
62631
63282
|
/**
|
|
62632
63283
|
* Mask Image URL
|
|
62633
63284
|
* @description The image to use as a mask for the generated image.
|
|
@@ -62719,9 +63370,9 @@ export interface FooocusImagePromptInput {
|
|
|
62719
63370
|
* Styles
|
|
62720
63371
|
* @description The style to use.
|
|
62721
63372
|
* @default [
|
|
63373
|
+
* "Fooocus Enhance",
|
|
62722
63374
|
* "Fooocus V2",
|
|
62723
|
-
* "Fooocus Sharp"
|
|
62724
|
-
* "Fooocus Enhance"
|
|
63375
|
+
* "Fooocus Sharp"
|
|
62725
63376
|
* ]
|
|
62726
63377
|
*/
|
|
62727
63378
|
styles?: (
|
|
@@ -63098,7 +63749,7 @@ export interface FooocusInput {
|
|
|
63098
63749
|
* }
|
|
63099
63750
|
* ]
|
|
63100
63751
|
*/
|
|
63101
|
-
loras?: Components.
|
|
63752
|
+
loras?: Components.LoraWeight_4[];
|
|
63102
63753
|
/**
|
|
63103
63754
|
* Mask Image Url
|
|
63104
63755
|
* @description The image to use as a mask for the generated image.
|
|
@@ -63177,9 +63828,9 @@ export interface FooocusInput {
|
|
|
63177
63828
|
* Styles
|
|
63178
63829
|
* @description The style to use.
|
|
63179
63830
|
* @default [
|
|
63831
|
+
* "Fooocus Enhance",
|
|
63180
63832
|
* "Fooocus V2",
|
|
63181
|
-
* "Fooocus Sharp"
|
|
63182
|
-
* "Fooocus Enhance"
|
|
63833
|
+
* "Fooocus Sharp"
|
|
63183
63834
|
* ]
|
|
63184
63835
|
*/
|
|
63185
63836
|
styles?: (
|
|
@@ -64139,8 +64790,8 @@ export interface FluxVisionUpscalerOutput {
|
|
|
64139
64790
|
/**
|
|
64140
64791
|
* @description The URL of the generated image.
|
|
64141
64792
|
* @example {
|
|
64142
|
-
* "height": 2048,
|
|
64143
64793
|
* "file_size": 8842156,
|
|
64794
|
+
* "height": 2048,
|
|
64144
64795
|
* "file_name": "20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
|
|
64145
64796
|
* "content_type": "image/png",
|
|
64146
64797
|
* "url": "https://v3b.fal.media/files/b/panda/20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
|
|
@@ -67276,7 +67927,7 @@ export interface Flux2LoraEditInput {
|
|
|
67276
67927
|
* @description List of LoRA weights to apply (maximum 3). Each LoRA can be a URL, HuggingFace repo ID, or local path.
|
|
67277
67928
|
* @default []
|
|
67278
67929
|
*/
|
|
67279
|
-
loras?: Components.
|
|
67930
|
+
loras?: Components.LoRAInput[];
|
|
67280
67931
|
/**
|
|
67281
67932
|
* Number of Images
|
|
67282
67933
|
* @description The number of images to generate.
|
|
@@ -67394,7 +68045,7 @@ export interface Flux2LoraInput {
|
|
|
67394
68045
|
* @description List of LoRA weights to apply (maximum 3). Each LoRA can be a URL, HuggingFace repo ID, or local path.
|
|
67395
68046
|
* @default []
|
|
67396
68047
|
*/
|
|
67397
|
-
loras?: Components.
|
|
68048
|
+
loras?: Components.LoRAInput[];
|
|
67398
68049
|
/**
|
|
67399
68050
|
* Number of Images
|
|
67400
68051
|
* @description The number of images to generate.
|
|
@@ -67529,15 +68180,87 @@ export interface Flux2KleinRealtimeOutput {
|
|
|
67529
68180
|
seed: number;
|
|
67530
68181
|
}
|
|
67531
68182
|
|
|
68183
|
+
export interface Flux2Klein9bLoraInput {
|
|
68184
|
+
/**
|
|
68185
|
+
* Enable Safety Checker
|
|
68186
|
+
* @description If set to true, the safety checker will be enabled.
|
|
68187
|
+
* @default true
|
|
68188
|
+
*/
|
|
68189
|
+
enable_safety_checker?: boolean;
|
|
68190
|
+
/**
|
|
68191
|
+
* Image Size
|
|
68192
|
+
* @description The size of the image to generate.
|
|
68193
|
+
* @default landscape_4_3
|
|
68194
|
+
*/
|
|
68195
|
+
image_size?:
|
|
68196
|
+
| Components.ImageSize
|
|
68197
|
+
| (
|
|
68198
|
+
| 'square_hd'
|
|
68199
|
+
| 'square'
|
|
68200
|
+
| 'portrait_4_3'
|
|
68201
|
+
| 'portrait_16_9'
|
|
68202
|
+
| 'landscape_4_3'
|
|
68203
|
+
| 'landscape_16_9'
|
|
68204
|
+
);
|
|
68205
|
+
/**
|
|
68206
|
+
* Loras
|
|
68207
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
68208
|
+
* @default []
|
|
68209
|
+
*/
|
|
68210
|
+
loras?: Components.LoRAInput[];
|
|
68211
|
+
/**
|
|
68212
|
+
* Number of Images
|
|
68213
|
+
* @description The number of images to generate.
|
|
68214
|
+
* @default 1
|
|
68215
|
+
*/
|
|
68216
|
+
num_images?: number;
|
|
68217
|
+
/**
|
|
68218
|
+
* Number of Inference Steps
|
|
68219
|
+
* @description The number of inference steps to perform.
|
|
68220
|
+
* @default 4
|
|
68221
|
+
*/
|
|
68222
|
+
num_inference_steps?: number;
|
|
68223
|
+
/**
|
|
68224
|
+
* Output Format
|
|
68225
|
+
* @description The format of the generated image.
|
|
68226
|
+
* @default png
|
|
68227
|
+
* @enum {string}
|
|
68228
|
+
*/
|
|
68229
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
68230
|
+
/**
|
|
68231
|
+
* Prompt
|
|
68232
|
+
* @description The prompt to generate an image from.
|
|
68233
|
+
* @example A bright, surreal scene on a vast white salt flat under a clear blue sky, featuring two fluffy white alpacas standing in the foreground. Behind them sits a sleek white supercar with its scissor doors open, creating a dramatic futuristic silhouette. The sunlight is strong and crisp, casting sharp shadows on the ground and giving the image a clean, high-contrast cinematic look. Minimalist composition, playful luxury vibe, ultra sharp detail, wide-angle perspective, high resolution.
|
|
68234
|
+
*/
|
|
68235
|
+
prompt: string;
|
|
68236
|
+
/**
|
|
68237
|
+
* Seed
|
|
68238
|
+
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
68239
|
+
*/
|
|
68240
|
+
seed?: number;
|
|
68241
|
+
/**
|
|
68242
|
+
* Sync Mode
|
|
68243
|
+
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
68244
|
+
* @default false
|
|
68245
|
+
*/
|
|
68246
|
+
sync_mode?: boolean;
|
|
68247
|
+
}
|
|
68248
|
+
|
|
68249
|
+
export interface Flux2Klein9bLoraOutput extends SharedType_a75 {}
|
|
68250
|
+
|
|
68251
|
+
export interface Flux2Klein9bEditLoraInput extends SharedType_974 {}
|
|
68252
|
+
|
|
68253
|
+
export interface Flux2Klein9bEditLoraOutput extends SharedType_1651 {}
|
|
68254
|
+
|
|
67532
68255
|
export interface Flux2Klein9bEditInput extends SharedType_2db {}
|
|
67533
68256
|
|
|
67534
68257
|
export interface Flux2Klein9bEditOutput extends SharedType_1651 {}
|
|
67535
68258
|
|
|
67536
|
-
export interface Flux2Klein9bBaseLoraInput extends
|
|
68259
|
+
export interface Flux2Klein9bBaseLoraInput extends SharedType_bb2 {}
|
|
67537
68260
|
|
|
67538
68261
|
export interface Flux2Klein9bBaseLoraOutput extends SharedType_1651 {}
|
|
67539
68262
|
|
|
67540
|
-
export interface Flux2Klein9bBaseEditLoraInput extends
|
|
68263
|
+
export interface Flux2Klein9bBaseEditLoraInput extends SharedType_f66 {}
|
|
67541
68264
|
|
|
67542
68265
|
export interface Flux2Klein9bBaseEditLoraOutput extends SharedType_1651 {}
|
|
67543
68266
|
|
|
@@ -67614,8 +68337,119 @@ export interface Flux2Klein9bBaseEditInput {
|
|
|
67614
68337
|
output_format?: 'jpeg' | 'png' | 'webp';
|
|
67615
68338
|
/**
|
|
67616
68339
|
* Prompt
|
|
67617
|
-
* @description The prompt to edit the image.
|
|
67618
|
-
* @example Imagine a young woman. Use the Style from the Reference Image.
|
|
68340
|
+
* @description The prompt to edit the image.
|
|
68341
|
+
* @example Imagine a young woman. Use the Style from the Reference Image.
|
|
68342
|
+
*/
|
|
68343
|
+
prompt: string;
|
|
68344
|
+
/**
|
|
68345
|
+
* Seed
|
|
68346
|
+
* @description The seed to use for the generation. If not provided, a random seed will be used.
|
|
68347
|
+
*/
|
|
68348
|
+
seed?: number;
|
|
68349
|
+
/**
|
|
68350
|
+
* Sync Mode
|
|
68351
|
+
* @description If `True`, the media will be returned as a data URI. Output is not stored when this is True.
|
|
68352
|
+
* @default false
|
|
68353
|
+
*/
|
|
68354
|
+
sync_mode?: boolean;
|
|
68355
|
+
}
|
|
68356
|
+
|
|
68357
|
+
export interface Flux2Klein9bBaseEditOutput {
|
|
68358
|
+
/**
|
|
68359
|
+
* Has Nsfw Concepts
|
|
68360
|
+
* @description Whether the generated images contain NSFW concepts.
|
|
68361
|
+
*/
|
|
68362
|
+
has_nsfw_concepts: boolean[];
|
|
68363
|
+
/**
|
|
68364
|
+
* Images
|
|
68365
|
+
* @description The edited images
|
|
68366
|
+
* @example [
|
|
68367
|
+
* {
|
|
68368
|
+
* "url": "https://v3b.fal.media/files/b/0a8a69f1/HpSn20CQnEgVbFRD5E-Eh.png"
|
|
68369
|
+
* }
|
|
68370
|
+
* ]
|
|
68371
|
+
*/
|
|
68372
|
+
images: Components.ImageFile[];
|
|
68373
|
+
/**
|
|
68374
|
+
* Prompt
|
|
68375
|
+
* @description The prompt used for generating the image.
|
|
68376
|
+
*/
|
|
68377
|
+
prompt: string;
|
|
68378
|
+
/**
|
|
68379
|
+
* Seed
|
|
68380
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
68381
|
+
* input or the randomly generated that was used in case none was passed.
|
|
68382
|
+
*/
|
|
68383
|
+
seed: number;
|
|
68384
|
+
timings: {
|
|
68385
|
+
[key: string]: number;
|
|
68386
|
+
};
|
|
68387
|
+
}
|
|
68388
|
+
|
|
68389
|
+
export interface Flux2Klein9bBaseInput {
|
|
68390
|
+
/**
|
|
68391
|
+
* Acceleration
|
|
68392
|
+
* @description The acceleration level to use for image generation.
|
|
68393
|
+
* @default regular
|
|
68394
|
+
* @enum {string}
|
|
68395
|
+
*/
|
|
68396
|
+
acceleration?: 'none' | 'regular' | 'high';
|
|
68397
|
+
/**
|
|
68398
|
+
* Enable Safety Checker
|
|
68399
|
+
* @description If set to true, the safety checker will be enabled.
|
|
68400
|
+
* @default true
|
|
68401
|
+
*/
|
|
68402
|
+
enable_safety_checker?: boolean;
|
|
68403
|
+
/**
|
|
68404
|
+
* Guidance Scale
|
|
68405
|
+
* @description Guidance scale for classifier-free guidance.
|
|
68406
|
+
* @default 5
|
|
68407
|
+
*/
|
|
68408
|
+
guidance_scale?: number;
|
|
68409
|
+
/**
|
|
68410
|
+
* Image Size
|
|
68411
|
+
* @description The size of the image to generate.
|
|
68412
|
+
* @default landscape_4_3
|
|
68413
|
+
*/
|
|
68414
|
+
image_size?:
|
|
68415
|
+
| Components.ImageSize
|
|
68416
|
+
| (
|
|
68417
|
+
| 'square_hd'
|
|
68418
|
+
| 'square'
|
|
68419
|
+
| 'portrait_4_3'
|
|
68420
|
+
| 'portrait_16_9'
|
|
68421
|
+
| 'landscape_4_3'
|
|
68422
|
+
| 'landscape_16_9'
|
|
68423
|
+
);
|
|
68424
|
+
/**
|
|
68425
|
+
* Negative Prompt
|
|
68426
|
+
* @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
|
|
68427
|
+
* @default
|
|
68428
|
+
*/
|
|
68429
|
+
negative_prompt?: string;
|
|
68430
|
+
/**
|
|
68431
|
+
* Number of Images
|
|
68432
|
+
* @description The number of images to generate.
|
|
68433
|
+
* @default 1
|
|
68434
|
+
*/
|
|
68435
|
+
num_images?: number;
|
|
68436
|
+
/**
|
|
68437
|
+
* Number of Inference Steps
|
|
68438
|
+
* @description The number of inference steps to perform.
|
|
68439
|
+
* @default 28
|
|
68440
|
+
*/
|
|
68441
|
+
num_inference_steps?: number;
|
|
68442
|
+
/**
|
|
68443
|
+
* Output Format
|
|
68444
|
+
* @description The format of the generated image.
|
|
68445
|
+
* @default png
|
|
68446
|
+
* @enum {string}
|
|
68447
|
+
*/
|
|
68448
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
68449
|
+
/**
|
|
68450
|
+
* Prompt
|
|
68451
|
+
* @description The prompt to generate an image from.
|
|
68452
|
+
* @example A cyberpunk samurai standing in neon-lit Tokyo street at night, rain-soaked pavement reflecting holographic advertisements in pink and blue, steam rising from street vents, wearing futuristic armor with glowing accents, cinematic composition, Unreal Engine 5 quality, ray-traced reflections, shallow depth of field, hyper-detailed
|
|
67619
68453
|
*/
|
|
67620
68454
|
prompt: string;
|
|
67621
68455
|
/**
|
|
@@ -67631,7 +68465,7 @@ export interface Flux2Klein9bBaseEditInput {
|
|
|
67631
68465
|
sync_mode?: boolean;
|
|
67632
68466
|
}
|
|
67633
68467
|
|
|
67634
|
-
export interface
|
|
68468
|
+
export interface Flux2Klein9bBaseOutput {
|
|
67635
68469
|
/**
|
|
67636
68470
|
* Has Nsfw Concepts
|
|
67637
68471
|
* @description Whether the generated images contain NSFW concepts.
|
|
@@ -67639,10 +68473,10 @@ export interface Flux2Klein9bBaseEditOutput {
|
|
|
67639
68473
|
has_nsfw_concepts: boolean[];
|
|
67640
68474
|
/**
|
|
67641
68475
|
* Images
|
|
67642
|
-
* @description The
|
|
68476
|
+
* @description The generated images
|
|
67643
68477
|
* @example [
|
|
67644
68478
|
* {
|
|
67645
|
-
* "url": "https://v3b.fal.media/files/b/
|
|
68479
|
+
* "url": "https://v3b.fal.media/files/b/0a8a69bc/EaxKO6wroq3eaDb2Znfpo.png"
|
|
67646
68480
|
* }
|
|
67647
68481
|
* ]
|
|
67648
68482
|
*/
|
|
@@ -67663,26 +68497,13 @@ export interface Flux2Klein9bBaseEditOutput {
|
|
|
67663
68497
|
};
|
|
67664
68498
|
}
|
|
67665
68499
|
|
|
67666
|
-
export interface
|
|
67667
|
-
/**
|
|
67668
|
-
* Acceleration
|
|
67669
|
-
* @description The acceleration level to use for image generation.
|
|
67670
|
-
* @default regular
|
|
67671
|
-
* @enum {string}
|
|
67672
|
-
*/
|
|
67673
|
-
acceleration?: 'none' | 'regular' | 'high';
|
|
68500
|
+
export interface Flux2Klein9bInput {
|
|
67674
68501
|
/**
|
|
67675
68502
|
* Enable Safety Checker
|
|
67676
68503
|
* @description If set to true, the safety checker will be enabled.
|
|
67677
68504
|
* @default true
|
|
67678
68505
|
*/
|
|
67679
68506
|
enable_safety_checker?: boolean;
|
|
67680
|
-
/**
|
|
67681
|
-
* Guidance Scale
|
|
67682
|
-
* @description Guidance scale for classifier-free guidance.
|
|
67683
|
-
* @default 5
|
|
67684
|
-
*/
|
|
67685
|
-
guidance_scale?: number;
|
|
67686
68507
|
/**
|
|
67687
68508
|
* Image Size
|
|
67688
68509
|
* @description The size of the image to generate.
|
|
@@ -67698,12 +68519,6 @@ export interface Flux2Klein9bBaseInput {
|
|
|
67698
68519
|
| 'landscape_4_3'
|
|
67699
68520
|
| 'landscape_16_9'
|
|
67700
68521
|
);
|
|
67701
|
-
/**
|
|
67702
|
-
* Negative Prompt
|
|
67703
|
-
* @description Negative prompt for classifier-free guidance. Describes what to avoid in the image.
|
|
67704
|
-
* @default
|
|
67705
|
-
*/
|
|
67706
|
-
negative_prompt?: string;
|
|
67707
68522
|
/**
|
|
67708
68523
|
* Number of Images
|
|
67709
68524
|
* @description The number of images to generate.
|
|
@@ -67713,7 +68528,7 @@ export interface Flux2Klein9bBaseInput {
|
|
|
67713
68528
|
/**
|
|
67714
68529
|
* Number of Inference Steps
|
|
67715
68530
|
* @description The number of inference steps to perform.
|
|
67716
|
-
* @default
|
|
68531
|
+
* @default 4
|
|
67717
68532
|
*/
|
|
67718
68533
|
num_inference_steps?: number;
|
|
67719
68534
|
/**
|
|
@@ -67726,7 +68541,7 @@ export interface Flux2Klein9bBaseInput {
|
|
|
67726
68541
|
/**
|
|
67727
68542
|
* Prompt
|
|
67728
68543
|
* @description The prompt to generate an image from.
|
|
67729
|
-
* @example A
|
|
68544
|
+
* @example A bright, surreal scene on a vast white salt flat under a clear blue sky, featuring two fluffy white alpacas standing in the foreground. Behind them sits a sleek white supercar with its scissor doors open, creating a dramatic futuristic silhouette. The sunlight is strong and crisp, casting sharp shadows on the ground and giving the image a clean, high-contrast cinematic look. Minimalist composition, playful luxury vibe, ultra sharp detail, wide-angle perspective, high resolution.
|
|
67730
68545
|
*/
|
|
67731
68546
|
prompt: string;
|
|
67732
68547
|
/**
|
|
@@ -67742,39 +68557,9 @@ export interface Flux2Klein9bBaseInput {
|
|
|
67742
68557
|
sync_mode?: boolean;
|
|
67743
68558
|
}
|
|
67744
68559
|
|
|
67745
|
-
export interface
|
|
67746
|
-
/**
|
|
67747
|
-
* Has Nsfw Concepts
|
|
67748
|
-
* @description Whether the generated images contain NSFW concepts.
|
|
67749
|
-
*/
|
|
67750
|
-
has_nsfw_concepts: boolean[];
|
|
67751
|
-
/**
|
|
67752
|
-
* Images
|
|
67753
|
-
* @description The generated images
|
|
67754
|
-
* @example [
|
|
67755
|
-
* {
|
|
67756
|
-
* "url": "https://v3b.fal.media/files/b/0a8a69bc/EaxKO6wroq3eaDb2Znfpo.png"
|
|
67757
|
-
* }
|
|
67758
|
-
* ]
|
|
67759
|
-
*/
|
|
67760
|
-
images: Components.ImageFile[];
|
|
67761
|
-
/**
|
|
67762
|
-
* Prompt
|
|
67763
|
-
* @description The prompt used for generating the image.
|
|
67764
|
-
*/
|
|
67765
|
-
prompt: string;
|
|
67766
|
-
/**
|
|
67767
|
-
* Seed
|
|
67768
|
-
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
67769
|
-
* input or the randomly generated that was used in case none was passed.
|
|
67770
|
-
*/
|
|
67771
|
-
seed: number;
|
|
67772
|
-
timings: {
|
|
67773
|
-
[key: string]: number;
|
|
67774
|
-
};
|
|
67775
|
-
}
|
|
68560
|
+
export interface Flux2Klein9bOutput extends SharedType_a75 {}
|
|
67776
68561
|
|
|
67777
|
-
export interface
|
|
68562
|
+
export interface Flux2Klein4bLoraInput {
|
|
67778
68563
|
/**
|
|
67779
68564
|
* Enable Safety Checker
|
|
67780
68565
|
* @description If set to true, the safety checker will be enabled.
|
|
@@ -67796,6 +68581,12 @@ export interface Flux2Klein9bInput {
|
|
|
67796
68581
|
| 'landscape_4_3'
|
|
67797
68582
|
| 'landscape_16_9'
|
|
67798
68583
|
);
|
|
68584
|
+
/**
|
|
68585
|
+
* Loras
|
|
68586
|
+
* @description List of LoRA weights to apply (maximum 3).
|
|
68587
|
+
* @default []
|
|
68588
|
+
*/
|
|
68589
|
+
loras?: Components.LoRAInput[];
|
|
67799
68590
|
/**
|
|
67800
68591
|
* Number of Images
|
|
67801
68592
|
* @description The number of images to generate.
|
|
@@ -67818,7 +68609,7 @@ export interface Flux2Klein9bInput {
|
|
|
67818
68609
|
/**
|
|
67819
68610
|
* Prompt
|
|
67820
68611
|
* @description The prompt to generate an image from.
|
|
67821
|
-
* @example
|
|
68612
|
+
* @example Cheetah in pre-sprint crouch, muscles tensed, gazelle visible in distance, African savanna golden grass, dust particles in air, National Geographic wildlife photography
|
|
67822
68613
|
*/
|
|
67823
68614
|
prompt: string;
|
|
67824
68615
|
/**
|
|
@@ -67834,77 +68625,21 @@ export interface Flux2Klein9bInput {
|
|
|
67834
68625
|
sync_mode?: boolean;
|
|
67835
68626
|
}
|
|
67836
68627
|
|
|
67837
|
-
export interface
|
|
67838
|
-
|
|
67839
|
-
|
|
67840
|
-
|
|
67841
|
-
|
|
67842
|
-
has_nsfw_concepts: boolean[];
|
|
67843
|
-
/**
|
|
67844
|
-
* Images
|
|
67845
|
-
* @description The generated images
|
|
67846
|
-
* @example [
|
|
67847
|
-
* {
|
|
67848
|
-
* "url": "https://v3b.fal.media/files/b/0a8a6905/DuNRy1OODaGrlUGEfl9SX.png"
|
|
67849
|
-
* }
|
|
67850
|
-
* ]
|
|
67851
|
-
*/
|
|
67852
|
-
images: Components.ImageFile[];
|
|
67853
|
-
/**
|
|
67854
|
-
* Prompt
|
|
67855
|
-
* @description The prompt used for generating the image.
|
|
67856
|
-
*/
|
|
67857
|
-
prompt: string;
|
|
67858
|
-
/**
|
|
67859
|
-
* Seed
|
|
67860
|
-
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
67861
|
-
* input or the randomly generated that was used in case none was passed.
|
|
67862
|
-
*/
|
|
67863
|
-
seed: number;
|
|
67864
|
-
timings: {
|
|
67865
|
-
[key: string]: number;
|
|
67866
|
-
};
|
|
67867
|
-
}
|
|
68628
|
+
export interface Flux2Klein4bLoraOutput extends SharedType_d72 {}
|
|
68629
|
+
|
|
68630
|
+
export interface Flux2Klein4bEditLoraInput extends SharedType_974 {}
|
|
68631
|
+
|
|
68632
|
+
export interface Flux2Klein4bEditLoraOutput extends SharedType_938 {}
|
|
67868
68633
|
|
|
67869
68634
|
export interface Flux2Klein4bEditInput extends SharedType_2db {}
|
|
67870
68635
|
|
|
67871
|
-
export interface Flux2Klein4bEditOutput {
|
|
67872
|
-
/**
|
|
67873
|
-
* Has Nsfw Concepts
|
|
67874
|
-
* @description Whether the generated images contain NSFW concepts.
|
|
67875
|
-
*/
|
|
67876
|
-
has_nsfw_concepts: boolean[];
|
|
67877
|
-
/**
|
|
67878
|
-
* Images
|
|
67879
|
-
* @description The edited images
|
|
67880
|
-
* @example [
|
|
67881
|
-
* {
|
|
67882
|
-
* "url": "https://v3b.fal.media/files/b/0a8a69d6/M73KvDgfEgIM77t4mFsS2.png"
|
|
67883
|
-
* }
|
|
67884
|
-
* ]
|
|
67885
|
-
*/
|
|
67886
|
-
images: Components.ImageFile[];
|
|
67887
|
-
/**
|
|
67888
|
-
* Prompt
|
|
67889
|
-
* @description The prompt used for generating the image.
|
|
67890
|
-
*/
|
|
67891
|
-
prompt: string;
|
|
67892
|
-
/**
|
|
67893
|
-
* Seed
|
|
67894
|
-
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
67895
|
-
* input or the randomly generated that was used in case none was passed.
|
|
67896
|
-
*/
|
|
67897
|
-
seed: number;
|
|
67898
|
-
timings: {
|
|
67899
|
-
[key: string]: number;
|
|
67900
|
-
};
|
|
67901
|
-
}
|
|
68636
|
+
export interface Flux2Klein4bEditOutput extends SharedType_938 {}
|
|
67902
68637
|
|
|
67903
|
-
export interface Flux2Klein4bBaseLoraInput extends
|
|
68638
|
+
export interface Flux2Klein4bBaseLoraInput extends SharedType_bb2 {}
|
|
67904
68639
|
|
|
67905
68640
|
export interface Flux2Klein4bBaseLoraOutput extends SharedType_1651 {}
|
|
67906
68641
|
|
|
67907
|
-
export interface Flux2Klein4bBaseEditLoraInput extends
|
|
68642
|
+
export interface Flux2Klein4bBaseEditLoraInput extends SharedType_f66 {}
|
|
67908
68643
|
|
|
67909
68644
|
export interface Flux2Klein4bBaseEditLoraOutput extends SharedType_1651 {}
|
|
67910
68645
|
|
|
@@ -68201,37 +68936,7 @@ export interface Flux2Klein4bInput {
|
|
|
68201
68936
|
sync_mode?: boolean;
|
|
68202
68937
|
}
|
|
68203
68938
|
|
|
68204
|
-
export interface Flux2Klein4bOutput {
|
|
68205
|
-
/**
|
|
68206
|
-
* Has Nsfw Concepts
|
|
68207
|
-
* @description Whether the generated images contain NSFW concepts.
|
|
68208
|
-
*/
|
|
68209
|
-
has_nsfw_concepts: boolean[];
|
|
68210
|
-
/**
|
|
68211
|
-
* Images
|
|
68212
|
-
* @description The generated images
|
|
68213
|
-
* @example [
|
|
68214
|
-
* {
|
|
68215
|
-
* "url": "https://v3b.fal.media/files/b/0a8a69c7/gU9ZgfFC9oAZjpIoveAac.png"
|
|
68216
|
-
* }
|
|
68217
|
-
* ]
|
|
68218
|
-
*/
|
|
68219
|
-
images: Components.ImageFile[];
|
|
68220
|
-
/**
|
|
68221
|
-
* Prompt
|
|
68222
|
-
* @description The prompt used for generating the image.
|
|
68223
|
-
*/
|
|
68224
|
-
prompt: string;
|
|
68225
|
-
/**
|
|
68226
|
-
* Seed
|
|
68227
|
-
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
68228
|
-
* input or the randomly generated that was used in case none was passed.
|
|
68229
|
-
*/
|
|
68230
|
-
seed: number;
|
|
68231
|
-
timings: {
|
|
68232
|
-
[key: string]: number;
|
|
68233
|
-
};
|
|
68234
|
-
}
|
|
68939
|
+
export interface Flux2Klein4bOutput extends SharedType_d72 {}
|
|
68235
68940
|
|
|
68236
68941
|
export interface Flux2FlashEditInput {
|
|
68237
68942
|
/**
|
|
@@ -71090,15 +71795,15 @@ export interface FloweditOutput {
|
|
|
71090
71795
|
seed: number;
|
|
71091
71796
|
}
|
|
71092
71797
|
|
|
71093
|
-
export interface Florence2LargeRegionToSegmentationInput extends
|
|
71798
|
+
export interface Florence2LargeRegionToSegmentationInput extends SharedType_2ff {}
|
|
71094
71799
|
|
|
71095
71800
|
export interface Florence2LargeRegionToSegmentationOutput extends SharedType_4aa {}
|
|
71096
71801
|
|
|
71097
|
-
export interface Florence2LargeRegionToDescriptionInput extends
|
|
71802
|
+
export interface Florence2LargeRegionToDescriptionInput extends SharedType_2ff {}
|
|
71098
71803
|
|
|
71099
71804
|
export interface Florence2LargeRegionToDescriptionOutput extends SharedType_129 {}
|
|
71100
71805
|
|
|
71101
|
-
export interface Florence2LargeRegionToCategoryInput extends
|
|
71806
|
+
export interface Florence2LargeRegionToCategoryInput extends SharedType_2ff {}
|
|
71102
71807
|
|
|
71103
71808
|
export interface Florence2LargeRegionToCategoryOutput extends SharedType_129 {}
|
|
71104
71809
|
|
|
@@ -72105,7 +72810,19 @@ export interface FastSvdTextToVideoInput {
|
|
|
72105
72810
|
);
|
|
72106
72811
|
}
|
|
72107
72812
|
|
|
72108
|
-
export interface FastSvdTextToVideoOutput
|
|
72813
|
+
export interface FastSvdTextToVideoOutput {
|
|
72814
|
+
/**
|
|
72815
|
+
* Seed
|
|
72816
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
72817
|
+
* input or the randomly generated that was used in case none was passed.
|
|
72818
|
+
*/
|
|
72819
|
+
seed: number;
|
|
72820
|
+
/**
|
|
72821
|
+
* Video
|
|
72822
|
+
* @description The generated video file.
|
|
72823
|
+
*/
|
|
72824
|
+
video: Components.File_1;
|
|
72825
|
+
}
|
|
72109
72826
|
|
|
72110
72827
|
export interface FastSvdLcmTextToVideoInput {
|
|
72111
72828
|
/**
|
|
@@ -72167,7 +72884,7 @@ export interface FastSvdLcmTextToVideoInput {
|
|
|
72167
72884
|
);
|
|
72168
72885
|
}
|
|
72169
72886
|
|
|
72170
|
-
export interface FastSvdLcmTextToVideoOutput extends
|
|
72887
|
+
export interface FastSvdLcmTextToVideoOutput extends SharedType_b23 {}
|
|
72171
72888
|
|
|
72172
72889
|
export interface FastSvdLcmInput {
|
|
72173
72890
|
/**
|
|
@@ -72217,7 +72934,7 @@ export interface FastSvdLcmInput {
|
|
|
72217
72934
|
steps?: number;
|
|
72218
72935
|
}
|
|
72219
72936
|
|
|
72220
|
-
export interface FastSvdLcmOutput extends
|
|
72937
|
+
export interface FastSvdLcmOutput extends SharedType_b23 {}
|
|
72221
72938
|
|
|
72222
72939
|
export interface FastSdxlInpaintingInput {
|
|
72223
72940
|
/**
|
|
@@ -72547,7 +73264,7 @@ export interface FastSdxlControlnetCannyInpaintingInput {
|
|
|
72547
73264
|
* @description The list of LoRA weights to use.
|
|
72548
73265
|
* @default []
|
|
72549
73266
|
*/
|
|
72550
|
-
loras?: Components.
|
|
73267
|
+
loras?: Components.LoraWeight_3[];
|
|
72551
73268
|
/**
|
|
72552
73269
|
* Mask Url
|
|
72553
73270
|
* @description The URL of the mask to use for inpainting.
|
|
@@ -72662,7 +73379,7 @@ export interface FastSdxlControlnetCannyImageToImageInput {
|
|
|
72662
73379
|
* @description The list of LoRA weights to use.
|
|
72663
73380
|
* @default []
|
|
72664
73381
|
*/
|
|
72665
|
-
loras?: Components.
|
|
73382
|
+
loras?: Components.LoraWeight_3[];
|
|
72666
73383
|
/**
|
|
72667
73384
|
* Negative Prompt
|
|
72668
73385
|
* @description The negative prompt to use.Use it to address details that you don't want
|
|
@@ -72771,7 +73488,7 @@ export interface FastSdxlControlnetCannyInput {
|
|
|
72771
73488
|
* @description The list of LoRA weights to use.
|
|
72772
73489
|
* @default []
|
|
72773
73490
|
*/
|
|
72774
|
-
loras?: Components.
|
|
73491
|
+
loras?: Components.LoraWeight_3[];
|
|
72775
73492
|
/**
|
|
72776
73493
|
* Negative Prompt
|
|
72777
73494
|
* @description The negative prompt to use. Use it to address details that you don't want
|
|
@@ -74505,11 +75222,11 @@ export interface F5TtsOutput {
|
|
|
74505
75222
|
audio_url: Components.AudioFile_2;
|
|
74506
75223
|
}
|
|
74507
75224
|
|
|
74508
|
-
export interface FLiteTextureInput extends
|
|
75225
|
+
export interface FLiteTextureInput extends SharedType_219 {}
|
|
74509
75226
|
|
|
74510
75227
|
export interface FLiteTextureOutput extends SharedType_7c6 {}
|
|
74511
75228
|
|
|
74512
|
-
export interface FLiteStandardInput extends
|
|
75229
|
+
export interface FLiteStandardInput extends SharedType_219 {}
|
|
74513
75230
|
|
|
74514
75231
|
export interface FLiteStandardOutput extends SharedType_7c6 {}
|
|
74515
75232
|
|
|
@@ -77474,7 +78191,7 @@ export interface Cogvideox5bVideoToVideoInput {
|
|
|
77474
78191
|
video_url: string;
|
|
77475
78192
|
}
|
|
77476
78193
|
|
|
77477
|
-
export interface Cogvideox5bVideoToVideoOutput extends
|
|
78194
|
+
export interface Cogvideox5bVideoToVideoOutput extends SharedType_1001 {}
|
|
77478
78195
|
|
|
77479
78196
|
export interface Cogvideox5bImageToVideoInput {
|
|
77480
78197
|
/**
|
|
@@ -77553,7 +78270,7 @@ export interface Cogvideox5bImageToVideoInput {
|
|
|
77553
78270
|
);
|
|
77554
78271
|
}
|
|
77555
78272
|
|
|
77556
|
-
export interface Cogvideox5bImageToVideoOutput extends
|
|
78273
|
+
export interface Cogvideox5bImageToVideoOutput extends SharedType_1001 {}
|
|
77557
78274
|
|
|
77558
78275
|
export interface Cogvideox5bInput {
|
|
77559
78276
|
/**
|
|
@@ -77626,7 +78343,7 @@ export interface Cogvideox5bInput {
|
|
|
77626
78343
|
);
|
|
77627
78344
|
}
|
|
77628
78345
|
|
|
77629
|
-
export interface Cogvideox5bOutput extends
|
|
78346
|
+
export interface Cogvideox5bOutput extends SharedType_1001 {}
|
|
77630
78347
|
|
|
77631
78348
|
export interface CodeformerInput {
|
|
77632
78349
|
/**
|
|
@@ -78108,7 +78825,6 @@ export interface ChatterboxTextToSpeechMultilingualInput {
|
|
|
78108
78825
|
/**
|
|
78109
78826
|
* Custom Audio Language
|
|
78110
78827
|
* @description If using a custom audio URL, specify the language of the audio here. Ignored if voice is not a custom url.
|
|
78111
|
-
* @enum {string}
|
|
78112
78828
|
*/
|
|
78113
78829
|
custom_audio_language?:
|
|
78114
78830
|
| 'english'
|
|
@@ -78191,13 +78907,12 @@ export interface ChatterboxTextToSpeechMultilingualInput {
|
|
|
78191
78907
|
|
|
78192
78908
|
export interface ChatterboxTextToSpeechMultilingualOutput {
|
|
78193
78909
|
/**
|
|
78194
|
-
* Audio
|
|
78195
78910
|
* @description The generated multilingual speech audio file
|
|
78196
78911
|
* @example {
|
|
78197
78912
|
* "url": "https://v3.fal.media/files/example/multilingual_speech_output.wav"
|
|
78198
78913
|
* }
|
|
78199
78914
|
*/
|
|
78200
|
-
audio: Components.
|
|
78915
|
+
audio: Components.File;
|
|
78201
78916
|
}
|
|
78202
78917
|
|
|
78203
78918
|
export interface ChatterboxTextToSpeechInput {
|
|
@@ -78237,7 +78952,7 @@ export interface ChatterboxTextToSpeechInput {
|
|
|
78237
78952
|
text: string;
|
|
78238
78953
|
}
|
|
78239
78954
|
|
|
78240
|
-
export interface ChatterboxTextToSpeechOutput extends
|
|
78955
|
+
export interface ChatterboxTextToSpeechOutput extends SharedType_d2a {}
|
|
78241
78956
|
|
|
78242
78957
|
export interface ChatterboxSpeechToSpeechInput {
|
|
78243
78958
|
/**
|
|
@@ -78253,16 +78968,7 @@ export interface ChatterboxSpeechToSpeechInput {
|
|
|
78253
78968
|
target_voice_audio_url?: string;
|
|
78254
78969
|
}
|
|
78255
78970
|
|
|
78256
|
-
export interface ChatterboxSpeechToSpeechOutput {
|
|
78257
|
-
/**
|
|
78258
|
-
* Audio
|
|
78259
|
-
* @description The generated speech audio
|
|
78260
|
-
* @example {
|
|
78261
|
-
* "url": "https://v3.fal.media/files/kangaroo/RQ_pxc7oPdueYqWUqEbPE_tmpjnzvvzx_.wav"
|
|
78262
|
-
* }
|
|
78263
|
-
*/
|
|
78264
|
-
audio: Components.File_1;
|
|
78265
|
-
}
|
|
78971
|
+
export interface ChatterboxSpeechToSpeechOutput extends SharedType_d2a {}
|
|
78266
78972
|
|
|
78267
78973
|
export interface ChainOfZoomInput {
|
|
78268
78974
|
/**
|
|
@@ -79868,6 +80574,84 @@ export interface BytedanceSeedanceV15ProImageToVideoOutput {
|
|
|
79868
80574
|
video: Components.File;
|
|
79869
80575
|
}
|
|
79870
80576
|
|
|
80577
|
+
export interface BytedanceSeedV2MiniInput {
|
|
80578
|
+
/**
|
|
80579
|
+
* Image Urls
|
|
80580
|
+
* @description URLs of images for visual understanding. Supported formats: JPEG, PNG, WebP. A maximum of 6 images is supported. Any additional images will be ignored.
|
|
80581
|
+
*/
|
|
80582
|
+
image_urls?: string[];
|
|
80583
|
+
/**
|
|
80584
|
+
* Max Completion Tokens
|
|
80585
|
+
* @description Controls the maximum length of the model's output, including both the model's response and its chain-of-thought content, measured in tokens.
|
|
80586
|
+
* @default 4096
|
|
80587
|
+
*/
|
|
80588
|
+
max_completion_tokens?: number;
|
|
80589
|
+
/**
|
|
80590
|
+
* Messages
|
|
80591
|
+
* @description Optional prior conversation history for multi-turn conversations. Pass back the `messages` field from a previous response to provide context. The current `prompt`, `image_urls`, `video_urls`, and `system_prompt` are always appended as the latest user turn.
|
|
80592
|
+
* @example []
|
|
80593
|
+
*/
|
|
80594
|
+
messages?: Components.Seed2MiniMessage[];
|
|
80595
|
+
/**
|
|
80596
|
+
* Prompt
|
|
80597
|
+
* @description The text prompt or question for the model.
|
|
80598
|
+
* @example What can you do?
|
|
80599
|
+
*/
|
|
80600
|
+
prompt: string;
|
|
80601
|
+
/**
|
|
80602
|
+
* Reasoning Effort
|
|
80603
|
+
* @description Controls the depth of reasoning before the model responds. Only applicable when `thinking` is `enabled` or `auto`. `minimal` for immediate response, `low` for faster response with light reasoning, `medium` for balanced speed and depth, `high` for deep analysis of complex issues.
|
|
80604
|
+
*/
|
|
80605
|
+
reasoning_effort?: 'minimal' | 'low' | 'medium' | 'high';
|
|
80606
|
+
/**
|
|
80607
|
+
* System Prompt
|
|
80608
|
+
* @description Optional system prompt to guide the model's behavior.
|
|
80609
|
+
*/
|
|
80610
|
+
system_prompt?: string;
|
|
80611
|
+
/**
|
|
80612
|
+
* Temperature
|
|
80613
|
+
* @description Controls randomness in the response. Lower values make output more focused and deterministic, higher values make it more creative.
|
|
80614
|
+
* @default 1
|
|
80615
|
+
*/
|
|
80616
|
+
temperature?: number;
|
|
80617
|
+
/**
|
|
80618
|
+
* Thinking
|
|
80619
|
+
* @description Controls the model's chain-of-thought reasoning. `enabled` always includes reasoning, `disabled` never includes reasoning, `auto` lets the model decide based on the query.
|
|
80620
|
+
* @default enabled
|
|
80621
|
+
* @enum {string}
|
|
80622
|
+
*/
|
|
80623
|
+
thinking?: 'enabled' | 'disabled' | 'auto';
|
|
80624
|
+
/**
|
|
80625
|
+
* Top P
|
|
80626
|
+
* @description Nucleus sampling parameter. The model considers tokens with top_p cumulative probability mass. Lower values narrow the token selection.
|
|
80627
|
+
* @default 0.7
|
|
80628
|
+
*/
|
|
80629
|
+
top_p?: number;
|
|
80630
|
+
/**
|
|
80631
|
+
* Video Urls
|
|
80632
|
+
* @description URLs of videos for video understanding. Supported formats: MP4, MOV. Audio comprehension is not supported. A maximum of 3 videos is supported. Any additional videos will be ignored.
|
|
80633
|
+
*/
|
|
80634
|
+
video_urls?: string[];
|
|
80635
|
+
}
|
|
80636
|
+
|
|
80637
|
+
export interface BytedanceSeedV2MiniOutput {
|
|
80638
|
+
/**
|
|
80639
|
+
* Messages
|
|
80640
|
+
* @description The full conversation history including the model's response. Pass this back as the `messages` input field to continue the conversation.
|
|
80641
|
+
*/
|
|
80642
|
+
messages: Components.Seed2MiniMessage[];
|
|
80643
|
+
/**
|
|
80644
|
+
* Output
|
|
80645
|
+
* @description The model's text response.
|
|
80646
|
+
*/
|
|
80647
|
+
output: string;
|
|
80648
|
+
/**
|
|
80649
|
+
* Reasoning Content
|
|
80650
|
+
* @description The model's chain-of-thought reasoning content. Only present when `thinking` is `enabled` or `auto`.
|
|
80651
|
+
*/
|
|
80652
|
+
reasoning_content?: string;
|
|
80653
|
+
}
|
|
80654
|
+
|
|
79871
80655
|
export interface BytedanceOmnihumanV15Input {
|
|
79872
80656
|
/**
|
|
79873
80657
|
* Audio Url
|
|
@@ -80086,7 +80870,7 @@ export interface BytedanceUpscalerUpscaleVideoOutput {
|
|
|
80086
80870
|
|
|
80087
80871
|
export interface BriaTextToImageHdInput extends SharedType_411 {}
|
|
80088
80872
|
|
|
80089
|
-
export interface BriaTextToImageHdOutput extends
|
|
80873
|
+
export interface BriaTextToImageHdOutput extends SharedType_a97 {}
|
|
80090
80874
|
|
|
80091
80875
|
export interface BriaTextToImageFastInput {
|
|
80092
80876
|
/**
|
|
@@ -80159,11 +80943,11 @@ export interface BriaTextToImageFastInput {
|
|
|
80159
80943
|
sync_mode?: boolean;
|
|
80160
80944
|
}
|
|
80161
80945
|
|
|
80162
|
-
export interface BriaTextToImageFastOutput extends
|
|
80946
|
+
export interface BriaTextToImageFastOutput extends SharedType_a97 {}
|
|
80163
80947
|
|
|
80164
80948
|
export interface BriaTextToImageBaseInput extends SharedType_411 {}
|
|
80165
80949
|
|
|
80166
|
-
export interface BriaTextToImageBaseOutput extends
|
|
80950
|
+
export interface BriaTextToImageBaseOutput extends SharedType_a97 {}
|
|
80167
80951
|
|
|
80168
80952
|
export interface BriaReimagineInput {
|
|
80169
80953
|
/**
|
|
@@ -80392,8 +81176,8 @@ export interface BriaGenfillOutput {
|
|
|
80392
81176
|
* @description Generated Images
|
|
80393
81177
|
* @example [
|
|
80394
81178
|
* {
|
|
80395
|
-
* "height": 768,
|
|
80396
81179
|
* "file_size": 1064550,
|
|
81180
|
+
* "height": 768,
|
|
80397
81181
|
* "file_name": "a0d138e6820c4ad58f1fd3c758f16047.png",
|
|
80398
81182
|
* "content_type": "image/png",
|
|
80399
81183
|
* "url": "https://storage.googleapis.com/falserverless/bria/bria_genfill_res.png",
|
|
@@ -80477,8 +81261,8 @@ export interface BriaExpandOutput {
|
|
|
80477
81261
|
/**
|
|
80478
81262
|
* @description The generated image
|
|
80479
81263
|
* @example {
|
|
80480
|
-
* "height": 674,
|
|
80481
81264
|
* "file_size": 1471342,
|
|
81265
|
+
* "height": 674,
|
|
80482
81266
|
* "file_name": "afa402a35ea742cdb5c3e219b2b19bfb.png",
|
|
80483
81267
|
* "content_type": "image/png",
|
|
80484
81268
|
* "url": "https://v3.fal.media/files/koala/8np-spgxxG-I1r3cjthRV_afa402a35ea742cdb5c3e219b2b19bfb.png",
|
|
@@ -80634,8 +81418,8 @@ export interface BriaBackgroundRemoveOutput {
|
|
|
80634
81418
|
/**
|
|
80635
81419
|
* @description The generated image
|
|
80636
81420
|
* @example {
|
|
80637
|
-
* "height": 1024,
|
|
80638
81421
|
* "file_size": 1076276,
|
|
81422
|
+
* "height": 1024,
|
|
80639
81423
|
* "file_name": "070c731993e949d993c10ef6283d335d.png",
|
|
80640
81424
|
* "content_type": "image/png",
|
|
80641
81425
|
* "url": "https://v3.fal.media/files/tiger/GQEMNjRyxSoza7N8LPPqb_070c731993e949d993c10ef6283d335d.png",
|
|
@@ -80837,8 +81621,8 @@ export interface BirefnetV2VideoOutput {
|
|
|
80837
81621
|
* "width": 1920,
|
|
80838
81622
|
* "fps": 24,
|
|
80839
81623
|
* "file_name": "birefnet-video-output.webm",
|
|
80840
|
-
* "
|
|
80841
|
-
* "
|
|
81624
|
+
* "num_frames": 192,
|
|
81625
|
+
* "content_type": "video/webm"
|
|
80842
81626
|
* }
|
|
80843
81627
|
*/
|
|
80844
81628
|
video: Components.VideoFile;
|
|
@@ -81685,24 +82469,7 @@ export interface AiAvatarSingleTextInput {
|
|
|
81685
82469
|
| 'Bill';
|
|
81686
82470
|
}
|
|
81687
82471
|
|
|
81688
|
-
export interface AiAvatarSingleTextOutput {
|
|
81689
|
-
/**
|
|
81690
|
-
* Seed
|
|
81691
|
-
* @description The seed used for generation.
|
|
81692
|
-
*/
|
|
81693
|
-
seed: number;
|
|
81694
|
-
/**
|
|
81695
|
-
* Video
|
|
81696
|
-
* @description The generated video file.
|
|
81697
|
-
* @example {
|
|
81698
|
-
* "file_size": 797478,
|
|
81699
|
-
* "file_name": "6c9dd31e1d9a4482877747a52a661a0a.mp4",
|
|
81700
|
-
* "content_type": "application/octet-stream",
|
|
81701
|
-
* "url": "https://v3.fal.media/files/elephant/-huMN0zTaXmBr2CqzCMps_6c9dd31e1d9a4482877747a52a661a0a.mp4"
|
|
81702
|
-
* }
|
|
81703
|
-
*/
|
|
81704
|
-
video: Components.File_1;
|
|
81705
|
-
}
|
|
82472
|
+
export interface AiAvatarSingleTextOutput extends SharedType_58d {}
|
|
81706
82473
|
|
|
81707
82474
|
export interface AiAvatarMultiTextInput {
|
|
81708
82475
|
/**
|
|
@@ -81818,7 +82585,6 @@ export interface AiAvatarMultiTextOutput {
|
|
|
81818
82585
|
*/
|
|
81819
82586
|
seed: number;
|
|
81820
82587
|
/**
|
|
81821
|
-
* Video
|
|
81822
82588
|
* @description The generated video file.
|
|
81823
82589
|
* @example {
|
|
81824
82590
|
* "file_size": 352679,
|
|
@@ -81827,7 +82593,7 @@ export interface AiAvatarMultiTextOutput {
|
|
|
81827
82593
|
* "url": "https://v3.fal.media/files/zebra/lKMkUvzCqKn-gHC0vyUPP_30b76b90c2164f9a926527497c20832b.mp4"
|
|
81828
82594
|
* }
|
|
81829
82595
|
*/
|
|
81830
|
-
video: Components.
|
|
82596
|
+
video: Components.File;
|
|
81831
82597
|
}
|
|
81832
82598
|
|
|
81833
82599
|
export interface AiAvatarMultiInput {
|
|
@@ -81896,7 +82662,6 @@ export interface AiAvatarMultiOutput {
|
|
|
81896
82662
|
*/
|
|
81897
82663
|
seed: number;
|
|
81898
82664
|
/**
|
|
81899
|
-
* Video
|
|
81900
82665
|
* @description The generated video file.
|
|
81901
82666
|
* @example {
|
|
81902
82667
|
* "file_size": 704757,
|
|
@@ -81905,7 +82670,7 @@ export interface AiAvatarMultiOutput {
|
|
|
81905
82670
|
* "url": "https://v3.fal.media/files/kangaroo/uAF7N-Ow8WwuvbFw8J4Br_ab27ac57e9464dbea1ef78f7a25469d2.mp4"
|
|
81906
82671
|
* }
|
|
81907
82672
|
*/
|
|
81908
|
-
video: Components.
|
|
82673
|
+
video: Components.File;
|
|
81909
82674
|
}
|
|
81910
82675
|
|
|
81911
82676
|
export interface AiAvatarInput {
|
|
@@ -81955,24 +82720,7 @@ export interface AiAvatarInput {
|
|
|
81955
82720
|
seed?: number;
|
|
81956
82721
|
}
|
|
81957
82722
|
|
|
81958
|
-
export interface AiAvatarOutput {
|
|
81959
|
-
/**
|
|
81960
|
-
* Seed
|
|
81961
|
-
* @description The seed used for generation.
|
|
81962
|
-
*/
|
|
81963
|
-
seed: number;
|
|
81964
|
-
/**
|
|
81965
|
-
* Video
|
|
81966
|
-
* @description The generated video file.
|
|
81967
|
-
* @example {
|
|
81968
|
-
* "file_size": 515275,
|
|
81969
|
-
* "file_name": "74af6c0bdd6041c3b1130d54885e3eee.mp4",
|
|
81970
|
-
* "content_type": "application/octet-stream",
|
|
81971
|
-
* "url": "https://v3.fal.media/files/kangaroo/z6VqUwNTwzuWa6YE1g7In_74af6c0bdd6041c3b1130d54885e3eee.mp4"
|
|
81972
|
-
* }
|
|
81973
|
-
*/
|
|
81974
|
-
video: Components.File_1;
|
|
81975
|
-
}
|
|
82723
|
+
export interface AiAvatarOutput extends SharedType_0c6 {}
|
|
81976
82724
|
|
|
81977
82725
|
export interface AceStepPromptToAudioInput {
|
|
81978
82726
|
/**
|
|
@@ -82837,8 +83585,8 @@ export interface ClarityaiCrystalVideoUpscalerOutput {
|
|
|
82837
83585
|
* "fps": 23.130193905817176,
|
|
82838
83586
|
* "width": 4096,
|
|
82839
83587
|
* "file_name": "w0VQQvPdwvV2GSCtRTMzh_hDH8SPrB.mp4",
|
|
82840
|
-
* "
|
|
82841
|
-
* "
|
|
83588
|
+
* "content_type": "video/mp4",
|
|
83589
|
+
* "num_frames": 302
|
|
82842
83590
|
* }
|
|
82843
83591
|
*/
|
|
82844
83592
|
video: Components.VideoFile;
|