fal-endpoint-types 1.3.20 → 1.3.22
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +1 -1
- package/types/fal/endpoints/components.d.ts +1 -53
- package/types/fal/endpoints/index.d.ts +30 -10
- package/types/fal/endpoints/schema.d.ts +1096 -1171
|
@@ -31,20 +31,19 @@ export interface XaiGrokImagineVideoTextToVideoInput {
|
|
|
31
31
|
|
|
32
32
|
export interface XaiGrokImagineVideoTextToVideoOutput {
|
|
33
33
|
/**
|
|
34
|
-
* Video
|
|
35
34
|
* @description The generated video.
|
|
36
35
|
* @example {
|
|
37
36
|
* "height": 720,
|
|
38
37
|
* "duration": 6.041667,
|
|
39
38
|
* "url": "https://v3b.fal.media/files/b/0a8b90e4/RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
|
|
40
|
-
* "width": 1280,
|
|
41
39
|
* "fps": 24,
|
|
40
|
+
* "width": 1280,
|
|
42
41
|
* "file_name": "RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
|
|
43
42
|
* "content_type": "video/mp4",
|
|
44
43
|
* "num_frames": 145
|
|
45
44
|
* }
|
|
46
45
|
*/
|
|
47
|
-
video: Components.
|
|
46
|
+
video: Components.VideoFile_1;
|
|
48
47
|
}
|
|
49
48
|
|
|
50
49
|
export interface XaiGrokImagineVideoImageToVideoInput {
|
|
@@ -52,7 +51,6 @@ export interface XaiGrokImagineVideoImageToVideoInput {
|
|
|
52
51
|
* Aspect Ratio
|
|
53
52
|
* @description Aspect ratio of the generated video.
|
|
54
53
|
* @default auto
|
|
55
|
-
* @enum {string}
|
|
56
54
|
*/
|
|
57
55
|
aspect_ratio?: 'auto' | '16:9' | '4:3' | '3:2' | '1:1' | '2:3' | '3:4' | '9:16';
|
|
58
56
|
/**
|
|
@@ -84,20 +82,19 @@ export interface XaiGrokImagineVideoImageToVideoInput {
|
|
|
84
82
|
|
|
85
83
|
export interface XaiGrokImagineVideoImageToVideoOutput {
|
|
86
84
|
/**
|
|
87
|
-
* Video
|
|
88
85
|
* @description The generated video.
|
|
89
86
|
* @example {
|
|
90
87
|
* "height": 720,
|
|
91
88
|
* "duration": 6.041667,
|
|
92
89
|
* "url": "https://v3b.fal.media/files/b/0a8b90e0/0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
|
|
93
|
-
* "width": 1280,
|
|
94
90
|
* "fps": 24,
|
|
91
|
+
* "width": 1280,
|
|
95
92
|
* "file_name": "0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
|
|
96
93
|
* "content_type": "video/mp4",
|
|
97
94
|
* "num_frames": 145
|
|
98
95
|
* }
|
|
99
96
|
*/
|
|
100
|
-
video: Components.
|
|
97
|
+
video: Components.VideoFile_1;
|
|
101
98
|
}
|
|
102
99
|
|
|
103
100
|
export interface XaiGrokImagineVideoEditVideoInput {
|
|
@@ -124,20 +121,19 @@ export interface XaiGrokImagineVideoEditVideoInput {
|
|
|
124
121
|
|
|
125
122
|
export interface XaiGrokImagineVideoEditVideoOutput {
|
|
126
123
|
/**
|
|
127
|
-
* Video
|
|
128
124
|
* @description The generated video.
|
|
129
125
|
* @example {
|
|
130
126
|
* "height": 720,
|
|
131
127
|
* "duration": 5.041667,
|
|
132
128
|
* "url": "https://v3b.fal.media/files/b/0a8b9113/EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
|
|
133
|
-
* "width": 1280,
|
|
134
129
|
* "fps": 24,
|
|
130
|
+
* "width": 1280,
|
|
135
131
|
* "file_name": "EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
|
|
136
132
|
* "content_type": "video/mp4",
|
|
137
133
|
* "num_frames": 121
|
|
138
134
|
* }
|
|
139
135
|
*/
|
|
140
|
-
video: Components.
|
|
136
|
+
video: Components.VideoFile_1;
|
|
141
137
|
}
|
|
142
138
|
|
|
143
139
|
export interface XaiGrokImagineImageEditInput {
|
|
@@ -184,7 +180,7 @@ export interface XaiGrokImagineImageEditOutput {
|
|
|
184
180
|
* }
|
|
185
181
|
* ]
|
|
186
182
|
*/
|
|
187
|
-
images: Components.
|
|
183
|
+
images: Components.ImageFile_1[];
|
|
188
184
|
/**
|
|
189
185
|
* Revised Prompt
|
|
190
186
|
* @description The enhanced prompt that was used to generate the image.
|
|
@@ -250,7 +246,7 @@ export interface XaiGrokImagineImageOutput {
|
|
|
250
246
|
* }
|
|
251
247
|
* ]
|
|
252
248
|
*/
|
|
253
|
-
images: Components.
|
|
249
|
+
images: Components.ImageFile_1[];
|
|
254
250
|
/**
|
|
255
251
|
* Revised Prompt
|
|
256
252
|
* @description The enhanced prompt that was used to generate the image.
|
|
@@ -941,7 +937,7 @@ export interface VeedAvatarsTextToVideoInput {
|
|
|
941
937
|
text: string;
|
|
942
938
|
}
|
|
943
939
|
|
|
944
|
-
export interface VeedAvatarsTextToVideoOutput extends
|
|
940
|
+
export interface VeedAvatarsTextToVideoOutput extends SharedType_78a1 {}
|
|
945
941
|
|
|
946
942
|
export interface VeedAvatarsAudioToVideoInput {
|
|
947
943
|
/**
|
|
@@ -986,7 +982,7 @@ export interface VeedAvatarsAudioToVideoInput {
|
|
|
986
982
|
| 'any_female_side';
|
|
987
983
|
}
|
|
988
984
|
|
|
989
|
-
export interface VeedAvatarsAudioToVideoOutput extends
|
|
985
|
+
export interface VeedAvatarsAudioToVideoOutput extends SharedType_78a1 {}
|
|
990
986
|
|
|
991
987
|
export interface Tripo3dTripoV25MultiviewTo3dInput {
|
|
992
988
|
/**
|
|
@@ -2104,6 +2100,59 @@ export interface SharedType_f5e {
|
|
|
2104
2100
|
seed?: number;
|
|
2105
2101
|
}
|
|
2106
2102
|
|
|
2103
|
+
export interface SharedType_f57 {
|
|
2104
|
+
/**
|
|
2105
|
+
* Aspect Ratio
|
|
2106
|
+
* @description The aspect ratio of the generated image.
|
|
2107
|
+
* @default 1:1
|
|
2108
|
+
* @enum {string}
|
|
2109
|
+
*/
|
|
2110
|
+
aspect_ratio?: '21:9' | '16:9' | '3:2' | '4:3' | '5:4' | '1:1' | '4:5' | '3:4' | '2:3' | '9:16';
|
|
2111
|
+
/**
|
|
2112
|
+
* Limit Generations
|
|
2113
|
+
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
2114
|
+
* @default false
|
|
2115
|
+
*/
|
|
2116
|
+
limit_generations?: boolean;
|
|
2117
|
+
/**
|
|
2118
|
+
* Number of Images
|
|
2119
|
+
* @description The number of images to generate.
|
|
2120
|
+
* @default 1
|
|
2121
|
+
*/
|
|
2122
|
+
num_images?: number;
|
|
2123
|
+
/**
|
|
2124
|
+
* Output Format
|
|
2125
|
+
* @description The format of the generated image.
|
|
2126
|
+
* @default png
|
|
2127
|
+
* @enum {string}
|
|
2128
|
+
*/
|
|
2129
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
2130
|
+
/**
|
|
2131
|
+
* Prompt
|
|
2132
|
+
* @description The text prompt to generate an image from.
|
|
2133
|
+
* @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
|
|
2134
|
+
*/
|
|
2135
|
+
prompt: string;
|
|
2136
|
+
/**
|
|
2137
|
+
* Safety Tolerance
|
|
2138
|
+
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
2139
|
+
* @default 4
|
|
2140
|
+
* @enum {string}
|
|
2141
|
+
*/
|
|
2142
|
+
safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
|
|
2143
|
+
/**
|
|
2144
|
+
* Seed
|
|
2145
|
+
* @description The seed for the random number generator.
|
|
2146
|
+
*/
|
|
2147
|
+
seed?: number;
|
|
2148
|
+
/**
|
|
2149
|
+
* Sync Mode
|
|
2150
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
2151
|
+
* @default false
|
|
2152
|
+
*/
|
|
2153
|
+
sync_mode?: boolean;
|
|
2154
|
+
}
|
|
2155
|
+
|
|
2107
2156
|
export interface SharedType_f51 {
|
|
2108
2157
|
/**
|
|
2109
2158
|
* Image Url
|
|
@@ -2296,29 +2345,6 @@ export interface SharedType_ee0 {
|
|
|
2296
2345
|
speaker_embedding: Components.File;
|
|
2297
2346
|
}
|
|
2298
2347
|
|
|
2299
|
-
export interface SharedType_ec2 {
|
|
2300
|
-
/**
|
|
2301
|
-
* Prompt
|
|
2302
|
-
* @description The text prompt used for video generation.
|
|
2303
|
-
* @default
|
|
2304
|
-
* @example A close-up of a young woman smiling gently in the rain, raindrops glistening on her face and eyelashes. The video captures the delicate details of her expression and the water droplets, with soft light reflecting off her skin in the rainy atmosphere.
|
|
2305
|
-
*/
|
|
2306
|
-
prompt?: string;
|
|
2307
|
-
/**
|
|
2308
|
-
* Seed
|
|
2309
|
-
* @description The seed used for generation.
|
|
2310
|
-
*/
|
|
2311
|
-
seed: number;
|
|
2312
|
-
/**
|
|
2313
|
-
* Video
|
|
2314
|
-
* @description The generated video file.
|
|
2315
|
-
* @example {
|
|
2316
|
-
* "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-woman-output.mp4"
|
|
2317
|
-
* }
|
|
2318
|
-
*/
|
|
2319
|
-
video: Components.File;
|
|
2320
|
-
}
|
|
2321
|
-
|
|
2322
2348
|
export interface SharedType_ec0 {
|
|
2323
2349
|
/**
|
|
2324
2350
|
* Art Style
|
|
@@ -2564,6 +2590,94 @@ export interface SharedType_e4b {
|
|
|
2564
2590
|
steps?: number;
|
|
2565
2591
|
}
|
|
2566
2592
|
|
|
2593
|
+
export interface SharedType_e49 {
|
|
2594
|
+
/**
|
|
2595
|
+
* Actual Prompt
|
|
2596
|
+
* @description The actual prompt used if prompt expansion was enabled
|
|
2597
|
+
*/
|
|
2598
|
+
actual_prompt?: string;
|
|
2599
|
+
/**
|
|
2600
|
+
* Model Glb
|
|
2601
|
+
* @description Generated 3D object in GLB format.
|
|
2602
|
+
* @example {
|
|
2603
|
+
* "file_size": 9314028,
|
|
2604
|
+
* "file_name": "model.glb",
|
|
2605
|
+
* "content_type": "model/gltf-binary",
|
|
2606
|
+
* "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
|
|
2607
|
+
* }
|
|
2608
|
+
*/
|
|
2609
|
+
model_glb: Components.File;
|
|
2610
|
+
/**
|
|
2611
|
+
* Model Urls
|
|
2612
|
+
* @description URLs for different 3D model formats
|
|
2613
|
+
* @example {
|
|
2614
|
+
* "fbx": {
|
|
2615
|
+
* "file_size": 5444380,
|
|
2616
|
+
* "file_name": "model.fbx",
|
|
2617
|
+
* "content_type": "application/octet-stream",
|
|
2618
|
+
* "url": "https://v3b.fal.media/files/b/kangaroo/7nUUw5dHN9a0DKlOpAKbP_model.fbx"
|
|
2619
|
+
* },
|
|
2620
|
+
* "usdz": {
|
|
2621
|
+
* "file_size": 9834246,
|
|
2622
|
+
* "file_name": "model.usdz",
|
|
2623
|
+
* "content_type": "model/vnd.usdz+zip",
|
|
2624
|
+
* "url": "https://v3b.fal.media/files/b/panda/XcC-mIJywUvH7coyrzENU_model.usdz"
|
|
2625
|
+
* },
|
|
2626
|
+
* "obj": {
|
|
2627
|
+
* "file_size": 2755145,
|
|
2628
|
+
* "file_name": "model.obj",
|
|
2629
|
+
* "content_type": "text/plain",
|
|
2630
|
+
* "url": "https://v3b.fal.media/files/b/monkey/cCNMHqUbKSNtDN1iGmiYm_model.obj"
|
|
2631
|
+
* },
|
|
2632
|
+
* "glb": {
|
|
2633
|
+
* "file_size": 9314028,
|
|
2634
|
+
* "file_name": "model.glb",
|
|
2635
|
+
* "content_type": "model/gltf-binary",
|
|
2636
|
+
* "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
|
|
2637
|
+
* }
|
|
2638
|
+
* }
|
|
2639
|
+
*/
|
|
2640
|
+
model_urls: Components.ModelUrls;
|
|
2641
|
+
/**
|
|
2642
|
+
* Prompt
|
|
2643
|
+
* @description The text prompt used for generation
|
|
2644
|
+
* @example A rustic, antique wooden treasure chest with a curved, domed lid, constructed from weathered, dark brown planks exhibiting prominent wood grain and subtle distress. It's heavily reinforced with broad, dark grey, oxidized metal bands secured by numerous circular rivets. Ornate, dark iron decorative elements featuring swirling foliate patterns and dragon motifs adorn the corners and lid. A prominent, circular, intricately carved metal lock plate with a central keyhole dominates the front, flanked by two large, dark metallic pull rings.
|
|
2645
|
+
*/
|
|
2646
|
+
prompt: string;
|
|
2647
|
+
/**
|
|
2648
|
+
* Seed
|
|
2649
|
+
* @description The seed used for generation
|
|
2650
|
+
* @example 4002110719
|
|
2651
|
+
*/
|
|
2652
|
+
seed?: number;
|
|
2653
|
+
/**
|
|
2654
|
+
* Texture Urls
|
|
2655
|
+
* @description Array of texture file objects
|
|
2656
|
+
* @example [
|
|
2657
|
+
* {
|
|
2658
|
+
* "base_color": {
|
|
2659
|
+
* "file_size": 4254502,
|
|
2660
|
+
* "file_name": "texture_0.png",
|
|
2661
|
+
* "content_type": "image/png",
|
|
2662
|
+
* "url": "https://v3b.fal.media/files/b/panda/DoPKAuZY0tTjnr6C9ee-Q_texture_0.png"
|
|
2663
|
+
* }
|
|
2664
|
+
* }
|
|
2665
|
+
* ]
|
|
2666
|
+
*/
|
|
2667
|
+
texture_urls?: Components.TextureFiles[];
|
|
2668
|
+
/**
|
|
2669
|
+
* Thumbnail
|
|
2670
|
+
* @description Preview thumbnail of the generated model
|
|
2671
|
+
* @example {
|
|
2672
|
+
* "file_size": 173792,
|
|
2673
|
+
* "file_name": "preview.png",
|
|
2674
|
+
* "content_type": "image/png",
|
|
2675
|
+
* "url": "https://v3b.fal.media/files/b/koala/6LJISu4ilkZXcdOETwl_d_preview.png"
|
|
2676
|
+
* }
|
|
2677
|
+
*/
|
|
2678
|
+
thumbnail?: Components.File;
|
|
2679
|
+
}
|
|
2680
|
+
|
|
2567
2681
|
export interface SharedType_e3b {
|
|
2568
2682
|
/**
|
|
2569
2683
|
* Video
|
|
@@ -4060,6 +4174,17 @@ export interface SharedType_c7e {
|
|
|
4060
4174
|
audio: Components.File;
|
|
4061
4175
|
}
|
|
4062
4176
|
|
|
4177
|
+
export interface SharedType_c61 {
|
|
4178
|
+
/**
|
|
4179
|
+
* Video
|
|
4180
|
+
* @description The generated video from image using the Q3 model
|
|
4181
|
+
* @example {
|
|
4182
|
+
* "url": "https://v3b.fal.media/files/b/0a8c9189/n9z3uUDPqmU2msAtqr25-_output.mp4"
|
|
4183
|
+
* }
|
|
4184
|
+
*/
|
|
4185
|
+
video: Components.File;
|
|
4186
|
+
}
|
|
4187
|
+
|
|
4063
4188
|
export interface SharedType_c5e {
|
|
4064
4189
|
/**
|
|
4065
4190
|
* Acceleration
|
|
@@ -4256,6 +4381,21 @@ export interface SharedType_c3a {
|
|
|
4256
4381
|
speakers: Components.VibeVoiceSpeaker[];
|
|
4257
4382
|
}
|
|
4258
4383
|
|
|
4384
|
+
export interface SharedType_c0f {
|
|
4385
|
+
/**
|
|
4386
|
+
* @description The generated image file.
|
|
4387
|
+
* @example {
|
|
4388
|
+
* "url": "https://storage.googleapis.com/falserverless/example_outputs/wan/t2i-output.png"
|
|
4389
|
+
* }
|
|
4390
|
+
*/
|
|
4391
|
+
image: Components.File_1;
|
|
4392
|
+
/**
|
|
4393
|
+
* Seed
|
|
4394
|
+
* @description The seed used for generation.
|
|
4395
|
+
*/
|
|
4396
|
+
seed: number;
|
|
4397
|
+
}
|
|
4398
|
+
|
|
4259
4399
|
export interface SharedType_bf2 {
|
|
4260
4400
|
/**
|
|
4261
4401
|
* Image
|
|
@@ -4314,29 +4454,6 @@ export interface SharedType_bda {
|
|
|
4314
4454
|
video_url: string;
|
|
4315
4455
|
}
|
|
4316
4456
|
|
|
4317
|
-
export interface SharedType_bd3 {
|
|
4318
|
-
/**
|
|
4319
|
-
* Prompt
|
|
4320
|
-
* @description The text prompt used for video generation.
|
|
4321
|
-
* @default
|
|
4322
|
-
* @example The white dragon warrior stands still, eyes full of determination and strength. The camera slowly moves closer or circles around the warrior, highlighting the powerful presence and heroic spirit of the character.
|
|
4323
|
-
*/
|
|
4324
|
-
prompt?: string;
|
|
4325
|
-
/**
|
|
4326
|
-
* Seed
|
|
4327
|
-
* @description The seed used for generation.
|
|
4328
|
-
*/
|
|
4329
|
-
seed: number;
|
|
4330
|
-
/**
|
|
4331
|
-
* Video
|
|
4332
|
-
* @description The generated video file.
|
|
4333
|
-
* @example {
|
|
4334
|
-
* "url": "https://v3.fal.media/files/lion/Fbuh3lO_HMT-pS0DATbio_tmp08c3v477.mp4"
|
|
4335
|
-
* }
|
|
4336
|
-
*/
|
|
4337
|
-
video: Components.File;
|
|
4338
|
-
}
|
|
4339
|
-
|
|
4340
4457
|
export interface SharedType_bc9 {
|
|
4341
4458
|
/**
|
|
4342
4459
|
* Has Nsfw Concepts
|
|
@@ -4729,29 +4846,6 @@ export interface SharedType_b61 {
|
|
|
4729
4846
|
}[];
|
|
4730
4847
|
}
|
|
4731
4848
|
|
|
4732
|
-
export interface SharedType_b5a {
|
|
4733
|
-
/**
|
|
4734
|
-
* Prompt
|
|
4735
|
-
* @description The text prompt used for video generation.
|
|
4736
|
-
* @default
|
|
4737
|
-
* @example A medium shot establishes a modern, minimalist office setting: clean lines, muted grey walls, and polished wood surfaces. The focus shifts to a close-up on a woman in sharp, navy blue business attire. Her crisp white blouse contrasts with the deep blue of her tailored suit jacket. The subtle texture of the fabric is visible—a fine weave with a slight sheen. Her expression is serious, yet engaging, as she speaks to someone unseen just beyond the frame. Close-up on her eyes, showing the intensity of her gaze and the fine lines around them that hint at experience and focus. Her lips are slightly parted, as if mid-sentence. The light catches the subtle highlights in her auburn hair, meticulously styled. Note the slight catch of light on the silver band of her watch. High resolution 4k
|
|
4738
|
-
*/
|
|
4739
|
-
prompt?: string;
|
|
4740
|
-
/**
|
|
4741
|
-
* Seed
|
|
4742
|
-
* @description The seed used for generation.
|
|
4743
|
-
*/
|
|
4744
|
-
seed: number;
|
|
4745
|
-
/**
|
|
4746
|
-
* Video
|
|
4747
|
-
* @description The generated video file.
|
|
4748
|
-
* @example {
|
|
4749
|
-
* "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-small-output.mp4"
|
|
4750
|
-
* }
|
|
4751
|
-
*/
|
|
4752
|
-
video: Components.File;
|
|
4753
|
-
}
|
|
4754
|
-
|
|
4755
4849
|
export interface SharedType_b59 {
|
|
4756
4850
|
/**
|
|
4757
4851
|
* Aspect Ratio
|
|
@@ -5056,94 +5150,6 @@ export interface SharedType_aed {
|
|
|
5056
5150
|
sync_mode?: boolean;
|
|
5057
5151
|
}
|
|
5058
5152
|
|
|
5059
|
-
export interface SharedType_ae3 {
|
|
5060
|
-
/**
|
|
5061
|
-
* Actual Prompt
|
|
5062
|
-
* @description The actual prompt used if prompt expansion was enabled
|
|
5063
|
-
*/
|
|
5064
|
-
actual_prompt?: string;
|
|
5065
|
-
/**
|
|
5066
|
-
* Model Glb
|
|
5067
|
-
* @description Generated 3D object in GLB format.
|
|
5068
|
-
* @example {
|
|
5069
|
-
* "file_size": 9314028,
|
|
5070
|
-
* "file_name": "model.glb",
|
|
5071
|
-
* "content_type": "model/gltf-binary",
|
|
5072
|
-
* "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
|
|
5073
|
-
* }
|
|
5074
|
-
*/
|
|
5075
|
-
model_glb: Components.File;
|
|
5076
|
-
/**
|
|
5077
|
-
* Model Urls
|
|
5078
|
-
* @description URLs for different 3D model formats
|
|
5079
|
-
* @example {
|
|
5080
|
-
* "fbx": {
|
|
5081
|
-
* "file_size": 5444380,
|
|
5082
|
-
* "file_name": "model.fbx",
|
|
5083
|
-
* "content_type": "application/octet-stream",
|
|
5084
|
-
* "url": "https://v3b.fal.media/files/b/kangaroo/7nUUw5dHN9a0DKlOpAKbP_model.fbx"
|
|
5085
|
-
* },
|
|
5086
|
-
* "usdz": {
|
|
5087
|
-
* "file_size": 9834246,
|
|
5088
|
-
* "file_name": "model.usdz",
|
|
5089
|
-
* "content_type": "model/vnd.usdz+zip",
|
|
5090
|
-
* "url": "https://v3b.fal.media/files/b/panda/XcC-mIJywUvH7coyrzENU_model.usdz"
|
|
5091
|
-
* },
|
|
5092
|
-
* "glb": {
|
|
5093
|
-
* "file_size": 9314028,
|
|
5094
|
-
* "file_name": "model.glb",
|
|
5095
|
-
* "content_type": "model/gltf-binary",
|
|
5096
|
-
* "url": "https://v3b.fal.media/files/b/penguin/DId89qXLu6BXu09RFAwAV_model.glb"
|
|
5097
|
-
* },
|
|
5098
|
-
* "obj": {
|
|
5099
|
-
* "file_size": 2755145,
|
|
5100
|
-
* "file_name": "model.obj",
|
|
5101
|
-
* "content_type": "text/plain",
|
|
5102
|
-
* "url": "https://v3b.fal.media/files/b/monkey/cCNMHqUbKSNtDN1iGmiYm_model.obj"
|
|
5103
|
-
* }
|
|
5104
|
-
* }
|
|
5105
|
-
*/
|
|
5106
|
-
model_urls: Components.ModelUrls;
|
|
5107
|
-
/**
|
|
5108
|
-
* Prompt
|
|
5109
|
-
* @description The text prompt used for generation
|
|
5110
|
-
* @example A rustic, antique wooden treasure chest with a curved, domed lid, constructed from weathered, dark brown planks exhibiting prominent wood grain and subtle distress. It's heavily reinforced with broad, dark grey, oxidized metal bands secured by numerous circular rivets. Ornate, dark iron decorative elements featuring swirling foliate patterns and dragon motifs adorn the corners and lid. A prominent, circular, intricately carved metal lock plate with a central keyhole dominates the front, flanked by two large, dark metallic pull rings.
|
|
5111
|
-
*/
|
|
5112
|
-
prompt: string;
|
|
5113
|
-
/**
|
|
5114
|
-
* Seed
|
|
5115
|
-
* @description The seed used for generation
|
|
5116
|
-
* @example 4002110719
|
|
5117
|
-
*/
|
|
5118
|
-
seed?: number;
|
|
5119
|
-
/**
|
|
5120
|
-
* Texture Urls
|
|
5121
|
-
* @description Array of texture file objects
|
|
5122
|
-
* @example [
|
|
5123
|
-
* {
|
|
5124
|
-
* "base_color": {
|
|
5125
|
-
* "file_size": 4254502,
|
|
5126
|
-
* "file_name": "texture_0.png",
|
|
5127
|
-
* "content_type": "image/png",
|
|
5128
|
-
* "url": "https://v3b.fal.media/files/b/panda/DoPKAuZY0tTjnr6C9ee-Q_texture_0.png"
|
|
5129
|
-
* }
|
|
5130
|
-
* }
|
|
5131
|
-
* ]
|
|
5132
|
-
*/
|
|
5133
|
-
texture_urls?: Components.TextureFiles[];
|
|
5134
|
-
/**
|
|
5135
|
-
* Thumbnail
|
|
5136
|
-
* @description Preview thumbnail of the generated model
|
|
5137
|
-
* @example {
|
|
5138
|
-
* "file_size": 173792,
|
|
5139
|
-
* "file_name": "preview.png",
|
|
5140
|
-
* "content_type": "image/png",
|
|
5141
|
-
* "url": "https://v3b.fal.media/files/b/koala/6LJISu4ilkZXcdOETwl_d_preview.png"
|
|
5142
|
-
* }
|
|
5143
|
-
*/
|
|
5144
|
-
thumbnail?: Components.File;
|
|
5145
|
-
}
|
|
5146
|
-
|
|
5147
5153
|
export interface SharedType_ade {
|
|
5148
5154
|
/**
|
|
5149
5155
|
* Video
|
|
@@ -6064,26 +6070,6 @@ export interface SharedType_9af {
|
|
|
6064
6070
|
topology?: 'quad' | 'triangle';
|
|
6065
6071
|
}
|
|
6066
6072
|
|
|
6067
|
-
export interface SharedType_98c {
|
|
6068
|
-
/**
|
|
6069
|
-
* Description
|
|
6070
|
-
* @description The description of the generated images.
|
|
6071
|
-
*/
|
|
6072
|
-
description: string;
|
|
6073
|
-
/**
|
|
6074
|
-
* Images
|
|
6075
|
-
* @description The edited images.
|
|
6076
|
-
* @example [
|
|
6077
|
-
* {
|
|
6078
|
-
* "file_name": "nano-banana-multi-edit-output.png",
|
|
6079
|
-
* "content_type": "image/png",
|
|
6080
|
-
* "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-multi-edit-output.png"
|
|
6081
|
-
* }
|
|
6082
|
-
* ]
|
|
6083
|
-
*/
|
|
6084
|
-
images: Components.ImageFile[];
|
|
6085
|
-
}
|
|
6086
|
-
|
|
6087
6073
|
export interface SharedType_97f {
|
|
6088
6074
|
/**
|
|
6089
6075
|
* Values
|
|
@@ -6098,47 +6084,6 @@ export interface SharedType_97f {
|
|
|
6098
6084
|
}[];
|
|
6099
6085
|
}
|
|
6100
6086
|
|
|
6101
|
-
export interface SharedType_97e {
|
|
6102
|
-
/**
|
|
6103
|
-
* Aspect Ratio
|
|
6104
|
-
* @description The aspect ratio of the generated image.
|
|
6105
|
-
* @default 1:1
|
|
6106
|
-
* @enum {string}
|
|
6107
|
-
*/
|
|
6108
|
-
aspect_ratio?: '21:9' | '16:9' | '3:2' | '4:3' | '5:4' | '1:1' | '4:5' | '3:4' | '2:3' | '9:16';
|
|
6109
|
-
/**
|
|
6110
|
-
* Limit Generations
|
|
6111
|
-
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
6112
|
-
* @default false
|
|
6113
|
-
*/
|
|
6114
|
-
limit_generations?: boolean;
|
|
6115
|
-
/**
|
|
6116
|
-
* Number of Images
|
|
6117
|
-
* @description The number of images to generate.
|
|
6118
|
-
* @default 1
|
|
6119
|
-
*/
|
|
6120
|
-
num_images?: number;
|
|
6121
|
-
/**
|
|
6122
|
-
* Output Format
|
|
6123
|
-
* @description The format of the generated image.
|
|
6124
|
-
* @default png
|
|
6125
|
-
* @enum {string}
|
|
6126
|
-
*/
|
|
6127
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
6128
|
-
/**
|
|
6129
|
-
* Prompt
|
|
6130
|
-
* @description The text prompt to generate an image from.
|
|
6131
|
-
* @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
|
|
6132
|
-
*/
|
|
6133
|
-
prompt: string;
|
|
6134
|
-
/**
|
|
6135
|
-
* Sync Mode
|
|
6136
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
6137
|
-
* @default false
|
|
6138
|
-
*/
|
|
6139
|
-
sync_mode?: boolean;
|
|
6140
|
-
}
|
|
6141
|
-
|
|
6142
6087
|
export interface SharedType_977 {
|
|
6143
6088
|
/**
|
|
6144
6089
|
* Aspect Ratio
|
|
@@ -6269,7 +6214,7 @@ export interface SharedType_962 {
|
|
|
6269
6214
|
};
|
|
6270
6215
|
}
|
|
6271
6216
|
|
|
6272
|
-
export interface
|
|
6217
|
+
export interface SharedType_95d1 {
|
|
6273
6218
|
/**
|
|
6274
6219
|
* Video
|
|
6275
6220
|
* @description The generated video
|
|
@@ -6282,6 +6227,46 @@ export interface SharedType_95d {
|
|
|
6282
6227
|
video: Components.File;
|
|
6283
6228
|
}
|
|
6284
6229
|
|
|
6230
|
+
export interface SharedType_95d {
|
|
6231
|
+
/**
|
|
6232
|
+
* Aspect Ratio
|
|
6233
|
+
* @description The aspect ratio of the output video
|
|
6234
|
+
* @default 16:9
|
|
6235
|
+
* @enum {string}
|
|
6236
|
+
*/
|
|
6237
|
+
aspect_ratio?: '16:9' | '9:16' | '4:3' | '3:4' | '1:1';
|
|
6238
|
+
/**
|
|
6239
|
+
* Audio
|
|
6240
|
+
* @description Whether to use direct audio-video generation. When true, outputs video with sound.
|
|
6241
|
+
* @default true
|
|
6242
|
+
*/
|
|
6243
|
+
audio?: boolean;
|
|
6244
|
+
/**
|
|
6245
|
+
* Duration
|
|
6246
|
+
* @description Duration of the video in seconds
|
|
6247
|
+
* @default 5
|
|
6248
|
+
*/
|
|
6249
|
+
duration?: number;
|
|
6250
|
+
/**
|
|
6251
|
+
* Prompt
|
|
6252
|
+
* @description Text prompt for video generation, max 2000 characters
|
|
6253
|
+
* @example In an ultra-realistic fashion photography style featuring light blue and pale amber tones, an astronaut in a spacesuit walks through the fog.
|
|
6254
|
+
*/
|
|
6255
|
+
prompt: string;
|
|
6256
|
+
/**
|
|
6257
|
+
* Resolution
|
|
6258
|
+
* @description Output video resolution
|
|
6259
|
+
* @default 720p
|
|
6260
|
+
* @enum {string}
|
|
6261
|
+
*/
|
|
6262
|
+
resolution?: '360p' | '540p' | '720p' | '1080p';
|
|
6263
|
+
/**
|
|
6264
|
+
* Seed
|
|
6265
|
+
* @description Random seed for reproducibility. If None, a random seed is chosen.
|
|
6266
|
+
*/
|
|
6267
|
+
seed?: number;
|
|
6268
|
+
}
|
|
6269
|
+
|
|
6285
6270
|
export interface SharedType_95c {
|
|
6286
6271
|
/**
|
|
6287
6272
|
* Images
|
|
@@ -6591,6 +6576,83 @@ export interface SharedType_905 {
|
|
|
6591
6576
|
seed?: number;
|
|
6592
6577
|
}
|
|
6593
6578
|
|
|
6579
|
+
export interface SharedType_8f1 {
|
|
6580
|
+
/**
|
|
6581
|
+
* Model Glb
|
|
6582
|
+
* @description Generated 3D object in GLB format.
|
|
6583
|
+
* @example {
|
|
6584
|
+
* "file_size": 9242744,
|
|
6585
|
+
* "file_name": "model.glb",
|
|
6586
|
+
* "content_type": "model/gltf-binary",
|
|
6587
|
+
* "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
|
|
6588
|
+
* }
|
|
6589
|
+
*/
|
|
6590
|
+
model_glb: Components.File;
|
|
6591
|
+
/**
|
|
6592
|
+
* Model Urls
|
|
6593
|
+
* @description URLs for different 3D model formats
|
|
6594
|
+
* @example {
|
|
6595
|
+
* "fbx": {
|
|
6596
|
+
* "file_size": 5427052,
|
|
6597
|
+
* "file_name": "model.fbx",
|
|
6598
|
+
* "content_type": "application/octet-stream",
|
|
6599
|
+
* "url": "https://v3b.fal.media/files/b/kangaroo/4Q2qdpTvfLVdzAKH1-72v_model.fbx"
|
|
6600
|
+
* },
|
|
6601
|
+
* "usdz": {
|
|
6602
|
+
* "file_size": 9991969,
|
|
6603
|
+
* "file_name": "model.usdz",
|
|
6604
|
+
* "content_type": "model/vnd.usdz+zip",
|
|
6605
|
+
* "url": "https://v3b.fal.media/files/b/lion/RgJG9EBQ_GAHMVWV3wCis_model.usdz"
|
|
6606
|
+
* },
|
|
6607
|
+
* "obj": {
|
|
6608
|
+
* "file_size": 2744413,
|
|
6609
|
+
* "file_name": "model.obj",
|
|
6610
|
+
* "content_type": "text/plain",
|
|
6611
|
+
* "url": "https://v3b.fal.media/files/b/koala/_Vg0d084-hd3EdpIJDf7U_model.obj"
|
|
6612
|
+
* },
|
|
6613
|
+
* "glb": {
|
|
6614
|
+
* "file_size": 9242744,
|
|
6615
|
+
* "file_name": "model.glb",
|
|
6616
|
+
* "content_type": "model/gltf-binary",
|
|
6617
|
+
* "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
|
|
6618
|
+
* }
|
|
6619
|
+
* }
|
|
6620
|
+
*/
|
|
6621
|
+
model_urls: Components.ModelUrls;
|
|
6622
|
+
/**
|
|
6623
|
+
* Seed
|
|
6624
|
+
* @description The seed used for generation (if available)
|
|
6625
|
+
* @example 2009275957
|
|
6626
|
+
*/
|
|
6627
|
+
seed?: number;
|
|
6628
|
+
/**
|
|
6629
|
+
* Texture Urls
|
|
6630
|
+
* @description Array of texture file objects, matching Meshy API structure
|
|
6631
|
+
* @example [
|
|
6632
|
+
* {
|
|
6633
|
+
* "base_color": {
|
|
6634
|
+
* "file_size": 4328755,
|
|
6635
|
+
* "file_name": "texture_0.png",
|
|
6636
|
+
* "content_type": "image/png",
|
|
6637
|
+
* "url": "https://v3b.fal.media/files/b/tiger/NkgxcEom_42V4_8UUXiRR_texture_0.png"
|
|
6638
|
+
* }
|
|
6639
|
+
* }
|
|
6640
|
+
* ]
|
|
6641
|
+
*/
|
|
6642
|
+
texture_urls?: Components.TextureFiles[];
|
|
6643
|
+
/**
|
|
6644
|
+
* Thumbnail
|
|
6645
|
+
* @description Preview thumbnail of the generated model
|
|
6646
|
+
* @example {
|
|
6647
|
+
* "file_size": 54279,
|
|
6648
|
+
* "file_name": "preview.png",
|
|
6649
|
+
* "content_type": "image/png",
|
|
6650
|
+
* "url": "https://v3b.fal.media/files/b/penguin/rfnS6ClmeEWgDXp_oD5tN_preview.png"
|
|
6651
|
+
* }
|
|
6652
|
+
*/
|
|
6653
|
+
thumbnail?: Components.File;
|
|
6654
|
+
}
|
|
6655
|
+
|
|
6594
6656
|
export interface SharedType_8df {
|
|
6595
6657
|
/**
|
|
6596
6658
|
* Aspect Ratio
|
|
@@ -6735,6 +6797,82 @@ export interface SharedType_8c5 {
|
|
|
6735
6797
|
seed: number;
|
|
6736
6798
|
}
|
|
6737
6799
|
|
|
6800
|
+
export interface SharedType_8c2 {
|
|
6801
|
+
/**
|
|
6802
|
+
* Aspect Ratio
|
|
6803
|
+
* @description The aspect ratio of the generated image. Use "auto" to let the model decide based on the prompt.
|
|
6804
|
+
* @default 1:1
|
|
6805
|
+
*/
|
|
6806
|
+
aspect_ratio?:
|
|
6807
|
+
| 'auto'
|
|
6808
|
+
| '21:9'
|
|
6809
|
+
| '16:9'
|
|
6810
|
+
| '3:2'
|
|
6811
|
+
| '4:3'
|
|
6812
|
+
| '5:4'
|
|
6813
|
+
| '1:1'
|
|
6814
|
+
| '4:5'
|
|
6815
|
+
| '3:4'
|
|
6816
|
+
| '2:3'
|
|
6817
|
+
| '9:16';
|
|
6818
|
+
/**
|
|
6819
|
+
* Enable Web Search
|
|
6820
|
+
* @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
|
|
6821
|
+
* @default false
|
|
6822
|
+
*/
|
|
6823
|
+
enable_web_search?: boolean;
|
|
6824
|
+
/**
|
|
6825
|
+
* Limit Generations
|
|
6826
|
+
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
6827
|
+
* @default false
|
|
6828
|
+
*/
|
|
6829
|
+
limit_generations?: boolean;
|
|
6830
|
+
/**
|
|
6831
|
+
* Number of Images
|
|
6832
|
+
* @description The number of images to generate.
|
|
6833
|
+
* @default 1
|
|
6834
|
+
*/
|
|
6835
|
+
num_images?: number;
|
|
6836
|
+
/**
|
|
6837
|
+
* Output Format
|
|
6838
|
+
* @description The format of the generated image.
|
|
6839
|
+
* @default png
|
|
6840
|
+
* @enum {string}
|
|
6841
|
+
*/
|
|
6842
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
6843
|
+
/**
|
|
6844
|
+
* Prompt
|
|
6845
|
+
* @description The text prompt to generate an image from.
|
|
6846
|
+
* @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
|
|
6847
|
+
*/
|
|
6848
|
+
prompt: string;
|
|
6849
|
+
/**
|
|
6850
|
+
* Resolution
|
|
6851
|
+
* @description The resolution of the image to generate.
|
|
6852
|
+
* @default 1K
|
|
6853
|
+
* @enum {string}
|
|
6854
|
+
*/
|
|
6855
|
+
resolution?: '1K' | '2K' | '4K';
|
|
6856
|
+
/**
|
|
6857
|
+
* Safety Tolerance
|
|
6858
|
+
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
6859
|
+
* @default 4
|
|
6860
|
+
* @enum {string}
|
|
6861
|
+
*/
|
|
6862
|
+
safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
|
|
6863
|
+
/**
|
|
6864
|
+
* Seed
|
|
6865
|
+
* @description The seed for the random number generator.
|
|
6866
|
+
*/
|
|
6867
|
+
seed?: number;
|
|
6868
|
+
/**
|
|
6869
|
+
* Sync Mode
|
|
6870
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
6871
|
+
* @default false
|
|
6872
|
+
*/
|
|
6873
|
+
sync_mode?: boolean;
|
|
6874
|
+
}
|
|
6875
|
+
|
|
6738
6876
|
export interface SharedType_8c0 {
|
|
6739
6877
|
/**
|
|
6740
6878
|
* Seed
|
|
@@ -6803,6 +6941,26 @@ export interface SharedType_87d {
|
|
|
6803
6941
|
video: Components.File;
|
|
6804
6942
|
}
|
|
6805
6943
|
|
|
6944
|
+
export interface SharedType_876 {
|
|
6945
|
+
/**
|
|
6946
|
+
* Description
|
|
6947
|
+
* @description The description of the generated images.
|
|
6948
|
+
*/
|
|
6949
|
+
description: string;
|
|
6950
|
+
/**
|
|
6951
|
+
* Images
|
|
6952
|
+
* @description The edited images.
|
|
6953
|
+
* @example [
|
|
6954
|
+
* {
|
|
6955
|
+
* "file_name": "nano-banana-multi-edit-output.png",
|
|
6956
|
+
* "content_type": "image/png",
|
|
6957
|
+
* "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-multi-edit-output.png"
|
|
6958
|
+
* }
|
|
6959
|
+
* ]
|
|
6960
|
+
*/
|
|
6961
|
+
images: Components.ImageFile_1[];
|
|
6962
|
+
}
|
|
6963
|
+
|
|
6806
6964
|
export interface SharedType_86b {
|
|
6807
6965
|
/**
|
|
6808
6966
|
* Audio Path
|
|
@@ -6870,6 +7028,78 @@ export interface SharedType_868 {
|
|
|
6870
7028
|
sync_mode?: boolean;
|
|
6871
7029
|
}
|
|
6872
7030
|
|
|
7031
|
+
export interface SharedType_85d {
|
|
7032
|
+
/**
|
|
7033
|
+
* Aspect Ratio
|
|
7034
|
+
* @description The aspect ratio of the generated image.
|
|
7035
|
+
* @default auto
|
|
7036
|
+
*/
|
|
7037
|
+
aspect_ratio?:
|
|
7038
|
+
| 'auto'
|
|
7039
|
+
| '21:9'
|
|
7040
|
+
| '16:9'
|
|
7041
|
+
| '3:2'
|
|
7042
|
+
| '4:3'
|
|
7043
|
+
| '5:4'
|
|
7044
|
+
| '1:1'
|
|
7045
|
+
| '4:5'
|
|
7046
|
+
| '3:4'
|
|
7047
|
+
| '2:3'
|
|
7048
|
+
| '9:16';
|
|
7049
|
+
/**
|
|
7050
|
+
* Image URLs
|
|
7051
|
+
* @description The URLs of the images to use for image-to-image generation or image editing.
|
|
7052
|
+
* @example [
|
|
7053
|
+
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
|
|
7054
|
+
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
|
|
7055
|
+
* ]
|
|
7056
|
+
*/
|
|
7057
|
+
image_urls: string[];
|
|
7058
|
+
/**
|
|
7059
|
+
* Limit Generations
|
|
7060
|
+
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
7061
|
+
* @default false
|
|
7062
|
+
*/
|
|
7063
|
+
limit_generations?: boolean;
|
|
7064
|
+
/**
|
|
7065
|
+
* Number of Images
|
|
7066
|
+
* @description The number of images to generate.
|
|
7067
|
+
* @default 1
|
|
7068
|
+
*/
|
|
7069
|
+
num_images?: number;
|
|
7070
|
+
/**
|
|
7071
|
+
* Output Format
|
|
7072
|
+
* @description The format of the generated image.
|
|
7073
|
+
* @default png
|
|
7074
|
+
* @enum {string}
|
|
7075
|
+
*/
|
|
7076
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
7077
|
+
/**
|
|
7078
|
+
* Prompt
|
|
7079
|
+
* @description The prompt for image editing.
|
|
7080
|
+
* @example make a photo of the man driving the car down the california coastline
|
|
7081
|
+
*/
|
|
7082
|
+
prompt: string;
|
|
7083
|
+
/**
|
|
7084
|
+
* Safety Tolerance
|
|
7085
|
+
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
7086
|
+
* @default 4
|
|
7087
|
+
* @enum {string}
|
|
7088
|
+
*/
|
|
7089
|
+
safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
|
|
7090
|
+
/**
|
|
7091
|
+
* Seed
|
|
7092
|
+
* @description The seed for the random number generator.
|
|
7093
|
+
*/
|
|
7094
|
+
seed?: number;
|
|
7095
|
+
/**
|
|
7096
|
+
* Sync Mode
|
|
7097
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
7098
|
+
* @default false
|
|
7099
|
+
*/
|
|
7100
|
+
sync_mode?: boolean;
|
|
7101
|
+
}
|
|
7102
|
+
|
|
6873
7103
|
export interface SharedType_82a {
|
|
6874
7104
|
/**
|
|
6875
7105
|
* Aspect Ratio (W:H)
|
|
@@ -7110,67 +7340,6 @@ export interface SharedType_815 {
|
|
|
7110
7340
|
video: Components.File;
|
|
7111
7341
|
}
|
|
7112
7342
|
|
|
7113
|
-
export interface SharedType_813 {
|
|
7114
|
-
/**
|
|
7115
|
-
* Aspect Ratio
|
|
7116
|
-
* @description The aspect ratio of the generated image.
|
|
7117
|
-
* @default auto
|
|
7118
|
-
* @enum {string}
|
|
7119
|
-
*/
|
|
7120
|
-
aspect_ratio?:
|
|
7121
|
-
| 'auto'
|
|
7122
|
-
| '21:9'
|
|
7123
|
-
| '16:9'
|
|
7124
|
-
| '3:2'
|
|
7125
|
-
| '4:3'
|
|
7126
|
-
| '5:4'
|
|
7127
|
-
| '1:1'
|
|
7128
|
-
| '4:5'
|
|
7129
|
-
| '3:4'
|
|
7130
|
-
| '2:3'
|
|
7131
|
-
| '9:16';
|
|
7132
|
-
/**
|
|
7133
|
-
* Image URLs
|
|
7134
|
-
* @description The URLs of the images to use for image-to-image generation or image editing.
|
|
7135
|
-
* @example [
|
|
7136
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
|
|
7137
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
|
|
7138
|
-
* ]
|
|
7139
|
-
*/
|
|
7140
|
-
image_urls: string[];
|
|
7141
|
-
/**
|
|
7142
|
-
* Limit Generations
|
|
7143
|
-
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
7144
|
-
* @default false
|
|
7145
|
-
*/
|
|
7146
|
-
limit_generations?: boolean;
|
|
7147
|
-
/**
|
|
7148
|
-
* Number of Images
|
|
7149
|
-
* @description The number of images to generate.
|
|
7150
|
-
* @default 1
|
|
7151
|
-
*/
|
|
7152
|
-
num_images?: number;
|
|
7153
|
-
/**
|
|
7154
|
-
* Output Format
|
|
7155
|
-
* @description The format of the generated image.
|
|
7156
|
-
* @default png
|
|
7157
|
-
* @enum {string}
|
|
7158
|
-
*/
|
|
7159
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
7160
|
-
/**
|
|
7161
|
-
* Prompt
|
|
7162
|
-
* @description The prompt for image editing.
|
|
7163
|
-
* @example make a photo of the man driving the car down the california coastline
|
|
7164
|
-
*/
|
|
7165
|
-
prompt: string;
|
|
7166
|
-
/**
|
|
7167
|
-
* Sync Mode
|
|
7168
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
7169
|
-
* @default false
|
|
7170
|
-
*/
|
|
7171
|
-
sync_mode?: boolean;
|
|
7172
|
-
}
|
|
7173
|
-
|
|
7174
7343
|
export interface SharedType_812 {
|
|
7175
7344
|
/**
|
|
7176
7345
|
* Video
|
|
@@ -7380,6 +7549,26 @@ export interface SharedType_7e9 {
|
|
|
7380
7549
|
video?: Components.File;
|
|
7381
7550
|
}
|
|
7382
7551
|
|
|
7552
|
+
export interface SharedType_7b9 {
|
|
7553
|
+
/**
|
|
7554
|
+
* Description
|
|
7555
|
+
* @description The description of the generated images.
|
|
7556
|
+
*/
|
|
7557
|
+
description: string;
|
|
7558
|
+
/**
|
|
7559
|
+
* Images
|
|
7560
|
+
* @description The generated images.
|
|
7561
|
+
* @example [
|
|
7562
|
+
* {
|
|
7563
|
+
* "file_name": "nano-banana-t2i-output.png",
|
|
7564
|
+
* "content_type": "image/png",
|
|
7565
|
+
* "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-t2i-output.png"
|
|
7566
|
+
* }
|
|
7567
|
+
* ]
|
|
7568
|
+
*/
|
|
7569
|
+
images: Components.ImageFile_1[];
|
|
7570
|
+
}
|
|
7571
|
+
|
|
7383
7572
|
export interface SharedType_7a3 {
|
|
7384
7573
|
/**
|
|
7385
7574
|
* Aspect Ratio
|
|
@@ -7458,7 +7647,7 @@ export interface SharedType_79c {
|
|
|
7458
7647
|
seed?: number;
|
|
7459
7648
|
}
|
|
7460
7649
|
|
|
7461
|
-
export interface
|
|
7650
|
+
export interface SharedType_78a1 {
|
|
7462
7651
|
/**
|
|
7463
7652
|
* @example {
|
|
7464
7653
|
* "content_type": "video/mp4",
|
|
@@ -7468,6 +7657,28 @@ export interface SharedType_78a {
|
|
|
7468
7657
|
video: Components.File_1;
|
|
7469
7658
|
}
|
|
7470
7659
|
|
|
7660
|
+
export interface SharedType_78a {
|
|
7661
|
+
/**
|
|
7662
|
+
* Prompt
|
|
7663
|
+
* @description The text prompt used for video generation.
|
|
7664
|
+
* @default
|
|
7665
|
+
* @example The white dragon warrior stands still, eyes full of determination and strength. The camera slowly moves closer or circles around the warrior, highlighting the powerful presence and heroic spirit of the character.
|
|
7666
|
+
*/
|
|
7667
|
+
prompt?: string;
|
|
7668
|
+
/**
|
|
7669
|
+
* Seed
|
|
7670
|
+
* @description The seed used for generation.
|
|
7671
|
+
*/
|
|
7672
|
+
seed: number;
|
|
7673
|
+
/**
|
|
7674
|
+
* @description The generated video file.
|
|
7675
|
+
* @example {
|
|
7676
|
+
* "url": "https://v3.fal.media/files/lion/Fbuh3lO_HMT-pS0DATbio_tmp08c3v477.mp4"
|
|
7677
|
+
* }
|
|
7678
|
+
*/
|
|
7679
|
+
video: Components.File_1;
|
|
7680
|
+
}
|
|
7681
|
+
|
|
7471
7682
|
export interface SharedType_778 {
|
|
7472
7683
|
/**
|
|
7473
7684
|
* Character Orientation
|
|
@@ -7620,83 +7831,6 @@ export interface SharedType_74f {
|
|
|
7620
7831
|
video: Components.File;
|
|
7621
7832
|
}
|
|
7622
7833
|
|
|
7623
|
-
export interface SharedType_748 {
|
|
7624
|
-
/**
|
|
7625
|
-
* Model Glb
|
|
7626
|
-
* @description Generated 3D object in GLB format.
|
|
7627
|
-
* @example {
|
|
7628
|
-
* "file_size": 9242744,
|
|
7629
|
-
* "file_name": "model.glb",
|
|
7630
|
-
* "content_type": "model/gltf-binary",
|
|
7631
|
-
* "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
|
|
7632
|
-
* }
|
|
7633
|
-
*/
|
|
7634
|
-
model_glb: Components.File;
|
|
7635
|
-
/**
|
|
7636
|
-
* Model Urls
|
|
7637
|
-
* @description URLs for different 3D model formats
|
|
7638
|
-
* @example {
|
|
7639
|
-
* "fbx": {
|
|
7640
|
-
* "file_size": 5427052,
|
|
7641
|
-
* "file_name": "model.fbx",
|
|
7642
|
-
* "content_type": "application/octet-stream",
|
|
7643
|
-
* "url": "https://v3b.fal.media/files/b/kangaroo/4Q2qdpTvfLVdzAKH1-72v_model.fbx"
|
|
7644
|
-
* },
|
|
7645
|
-
* "usdz": {
|
|
7646
|
-
* "file_size": 9991969,
|
|
7647
|
-
* "file_name": "model.usdz",
|
|
7648
|
-
* "content_type": "model/vnd.usdz+zip",
|
|
7649
|
-
* "url": "https://v3b.fal.media/files/b/lion/RgJG9EBQ_GAHMVWV3wCis_model.usdz"
|
|
7650
|
-
* },
|
|
7651
|
-
* "glb": {
|
|
7652
|
-
* "file_size": 9242744,
|
|
7653
|
-
* "file_name": "model.glb",
|
|
7654
|
-
* "content_type": "model/gltf-binary",
|
|
7655
|
-
* "url": "https://v3b.fal.media/files/b/zebra/OXF1e1bO3JddPTaugv0eL_model.glb"
|
|
7656
|
-
* },
|
|
7657
|
-
* "obj": {
|
|
7658
|
-
* "file_size": 2744413,
|
|
7659
|
-
* "file_name": "model.obj",
|
|
7660
|
-
* "content_type": "text/plain",
|
|
7661
|
-
* "url": "https://v3b.fal.media/files/b/koala/_Vg0d084-hd3EdpIJDf7U_model.obj"
|
|
7662
|
-
* }
|
|
7663
|
-
* }
|
|
7664
|
-
*/
|
|
7665
|
-
model_urls: Components.ModelUrls;
|
|
7666
|
-
/**
|
|
7667
|
-
* Seed
|
|
7668
|
-
* @description The seed used for generation (if available)
|
|
7669
|
-
* @example 2009275957
|
|
7670
|
-
*/
|
|
7671
|
-
seed?: number;
|
|
7672
|
-
/**
|
|
7673
|
-
* Texture Urls
|
|
7674
|
-
* @description Array of texture file objects, matching Meshy API structure
|
|
7675
|
-
* @example [
|
|
7676
|
-
* {
|
|
7677
|
-
* "base_color": {
|
|
7678
|
-
* "file_size": 4328755,
|
|
7679
|
-
* "file_name": "texture_0.png",
|
|
7680
|
-
* "content_type": "image/png",
|
|
7681
|
-
* "url": "https://v3b.fal.media/files/b/tiger/NkgxcEom_42V4_8UUXiRR_texture_0.png"
|
|
7682
|
-
* }
|
|
7683
|
-
* }
|
|
7684
|
-
* ]
|
|
7685
|
-
*/
|
|
7686
|
-
texture_urls?: Components.TextureFiles[];
|
|
7687
|
-
/**
|
|
7688
|
-
* Thumbnail
|
|
7689
|
-
* @description Preview thumbnail of the generated model
|
|
7690
|
-
* @example {
|
|
7691
|
-
* "file_size": 54279,
|
|
7692
|
-
* "file_name": "preview.png",
|
|
7693
|
-
* "content_type": "image/png",
|
|
7694
|
-
* "url": "https://v3b.fal.media/files/b/penguin/rfnS6ClmeEWgDXp_oD5tN_preview.png"
|
|
7695
|
-
* }
|
|
7696
|
-
*/
|
|
7697
|
-
thumbnail?: Components.File;
|
|
7698
|
-
}
|
|
7699
|
-
|
|
7700
7834
|
export interface SharedType_744 {
|
|
7701
7835
|
/**
|
|
7702
7836
|
* Image
|
|
@@ -8252,26 +8386,6 @@ export interface SharedType_663 {
|
|
|
8252
8386
|
sync_mode?: boolean;
|
|
8253
8387
|
}
|
|
8254
8388
|
|
|
8255
|
-
export interface SharedType_662 {
|
|
8256
|
-
/**
|
|
8257
|
-
* Description
|
|
8258
|
-
* @description The description of the generated images.
|
|
8259
|
-
*/
|
|
8260
|
-
description: string;
|
|
8261
|
-
/**
|
|
8262
|
-
* Images
|
|
8263
|
-
* @description The generated images.
|
|
8264
|
-
* @example [
|
|
8265
|
-
* {
|
|
8266
|
-
* "file_name": "nano-banana-t2i-output.png",
|
|
8267
|
-
* "content_type": "image/png",
|
|
8268
|
-
* "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-t2i-output.png"
|
|
8269
|
-
* }
|
|
8270
|
-
* ]
|
|
8271
|
-
*/
|
|
8272
|
-
images: Components.ImageFile[];
|
|
8273
|
-
}
|
|
8274
|
-
|
|
8275
8389
|
export interface SharedType_659 {
|
|
8276
8390
|
/** @description Image with depth map */
|
|
8277
8391
|
image: Components.Image_2;
|
|
@@ -8496,6 +8610,64 @@ export interface SharedType_5c6 {
|
|
|
8496
8610
|
results: Components.BoundingBoxes;
|
|
8497
8611
|
}
|
|
8498
8612
|
|
|
8613
|
+
export interface SharedType_5c5 {
|
|
8614
|
+
/**
|
|
8615
|
+
* Prompt
|
|
8616
|
+
* @description The text prompt used for video generation.
|
|
8617
|
+
* @default
|
|
8618
|
+
* @example A medium shot establishes a modern, minimalist office setting: clean lines, muted grey walls, and polished wood surfaces. The focus shifts to a close-up on a woman in sharp, navy blue business attire. Her crisp white blouse contrasts with the deep blue of her tailored suit jacket. The subtle texture of the fabric is visible—a fine weave with a slight sheen. Her expression is serious, yet engaging, as she speaks to someone unseen just beyond the frame. Close-up on her eyes, showing the intensity of her gaze and the fine lines around them that hint at experience and focus. Her lips are slightly parted, as if mid-sentence. The light catches the subtle highlights in her auburn hair, meticulously styled. Note the slight catch of light on the silver band of her watch. High resolution 4k
|
|
8619
|
+
*/
|
|
8620
|
+
prompt?: string;
|
|
8621
|
+
/**
|
|
8622
|
+
* Seed
|
|
8623
|
+
* @description The seed used for generation.
|
|
8624
|
+
*/
|
|
8625
|
+
seed: number;
|
|
8626
|
+
/**
|
|
8627
|
+
* @description The generated video file.
|
|
8628
|
+
* @example {
|
|
8629
|
+
* "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-small-output.mp4"
|
|
8630
|
+
* }
|
|
8631
|
+
*/
|
|
8632
|
+
video: Components.File_1;
|
|
8633
|
+
}
|
|
8634
|
+
|
|
8635
|
+
export interface SharedType_5c2 {
|
|
8636
|
+
/**
|
|
8637
|
+
* Has Nsfw Concepts
|
|
8638
|
+
* @description Whether the generated images contain NSFW concepts.
|
|
8639
|
+
*/
|
|
8640
|
+
has_nsfw_concepts: boolean[];
|
|
8641
|
+
/**
|
|
8642
|
+
* Images
|
|
8643
|
+
* @description The generated image(s).
|
|
8644
|
+
* @example [
|
|
8645
|
+
* {
|
|
8646
|
+
* "height": 768,
|
|
8647
|
+
* "content_type": "image/jpeg",
|
|
8648
|
+
* "url": "https://v3b.fal.media/files/b/0a8e76a4/n8AJdfVTm-9msRC-sVQFn.jpg",
|
|
8649
|
+
* "width": 1024
|
|
8650
|
+
* }
|
|
8651
|
+
* ]
|
|
8652
|
+
*/
|
|
8653
|
+
images: Components.Image_1[];
|
|
8654
|
+
/**
|
|
8655
|
+
* Prompt
|
|
8656
|
+
* @description The prompt used for generating the image.
|
|
8657
|
+
*/
|
|
8658
|
+
prompt: string;
|
|
8659
|
+
/**
|
|
8660
|
+
* Seed
|
|
8661
|
+
* @description Seed of the generated Image. It will be the same value of the one passed in the
|
|
8662
|
+
* input or the randomly generated that was used in case none was passed.
|
|
8663
|
+
*/
|
|
8664
|
+
seed: number;
|
|
8665
|
+
/** Timings */
|
|
8666
|
+
timings: {
|
|
8667
|
+
[key: string]: number;
|
|
8668
|
+
};
|
|
8669
|
+
}
|
|
8670
|
+
|
|
8499
8671
|
export interface SharedType_5bd {
|
|
8500
8672
|
/**
|
|
8501
8673
|
* Model Mesh
|
|
@@ -9184,22 +9356,6 @@ export interface SharedType_524 {
|
|
|
9184
9356
|
video: Components.VideoFile_1;
|
|
9185
9357
|
}
|
|
9186
9358
|
|
|
9187
|
-
export interface SharedType_4f3 {
|
|
9188
|
-
/**
|
|
9189
|
-
* Image
|
|
9190
|
-
* @description The generated image file.
|
|
9191
|
-
* @example {
|
|
9192
|
-
* "url": "https://storage.googleapis.com/falserverless/example_outputs/wan/t2i-output.png"
|
|
9193
|
-
* }
|
|
9194
|
-
*/
|
|
9195
|
-
image: Components.File;
|
|
9196
|
-
/**
|
|
9197
|
-
* Seed
|
|
9198
|
-
* @description The seed used for generation.
|
|
9199
|
-
*/
|
|
9200
|
-
seed: number;
|
|
9201
|
-
}
|
|
9202
|
-
|
|
9203
9359
|
export interface SharedType_4c2 {
|
|
9204
9360
|
/**
|
|
9205
9361
|
* Video
|
|
@@ -10101,6 +10257,28 @@ export interface SharedType_386 {
|
|
|
10101
10257
|
images: Components.Image[];
|
|
10102
10258
|
}
|
|
10103
10259
|
|
|
10260
|
+
export interface SharedType_382 {
|
|
10261
|
+
/**
|
|
10262
|
+
* Prompt
|
|
10263
|
+
* @description The text prompt used for video generation.
|
|
10264
|
+
* @default
|
|
10265
|
+
* @example A close-up of a young woman smiling gently in the rain, raindrops glistening on her face and eyelashes. The video captures the delicate details of her expression and the water droplets, with soft light reflecting off her skin in the rainy atmosphere.
|
|
10266
|
+
*/
|
|
10267
|
+
prompt?: string;
|
|
10268
|
+
/**
|
|
10269
|
+
* Seed
|
|
10270
|
+
* @description The seed used for generation.
|
|
10271
|
+
*/
|
|
10272
|
+
seed: number;
|
|
10273
|
+
/**
|
|
10274
|
+
* @description The generated video file.
|
|
10275
|
+
* @example {
|
|
10276
|
+
* "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-woman-output.mp4"
|
|
10277
|
+
* }
|
|
10278
|
+
*/
|
|
10279
|
+
video: Components.File_1;
|
|
10280
|
+
}
|
|
10281
|
+
|
|
10104
10282
|
export interface SharedType_377 {
|
|
10105
10283
|
/**
|
|
10106
10284
|
* Output
|
|
@@ -10192,6 +10370,91 @@ export interface SharedType_368 {
|
|
|
10192
10370
|
};
|
|
10193
10371
|
}
|
|
10194
10372
|
|
|
10373
|
+
export interface SharedType_367 {
|
|
10374
|
+
/**
|
|
10375
|
+
* Aspect Ratio
|
|
10376
|
+
* @description The aspect ratio of the generated image.
|
|
10377
|
+
* @default auto
|
|
10378
|
+
*/
|
|
10379
|
+
aspect_ratio?:
|
|
10380
|
+
| 'auto'
|
|
10381
|
+
| '21:9'
|
|
10382
|
+
| '16:9'
|
|
10383
|
+
| '3:2'
|
|
10384
|
+
| '4:3'
|
|
10385
|
+
| '5:4'
|
|
10386
|
+
| '1:1'
|
|
10387
|
+
| '4:5'
|
|
10388
|
+
| '3:4'
|
|
10389
|
+
| '2:3'
|
|
10390
|
+
| '9:16';
|
|
10391
|
+
/**
|
|
10392
|
+
* Enable Web Search
|
|
10393
|
+
* @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
|
|
10394
|
+
* @default false
|
|
10395
|
+
*/
|
|
10396
|
+
enable_web_search?: boolean;
|
|
10397
|
+
/**
|
|
10398
|
+
* Image URLs
|
|
10399
|
+
* @description The URLs of the images to use for image-to-image generation or image editing.
|
|
10400
|
+
* @example [
|
|
10401
|
+
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
|
|
10402
|
+
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
|
|
10403
|
+
* ]
|
|
10404
|
+
*/
|
|
10405
|
+
image_urls: string[];
|
|
10406
|
+
/**
|
|
10407
|
+
* Limit Generations
|
|
10408
|
+
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
10409
|
+
* @default false
|
|
10410
|
+
*/
|
|
10411
|
+
limit_generations?: boolean;
|
|
10412
|
+
/**
|
|
10413
|
+
* Number of Images
|
|
10414
|
+
* @description The number of images to generate.
|
|
10415
|
+
* @default 1
|
|
10416
|
+
*/
|
|
10417
|
+
num_images?: number;
|
|
10418
|
+
/**
|
|
10419
|
+
* Output Format
|
|
10420
|
+
* @description The format of the generated image.
|
|
10421
|
+
* @default png
|
|
10422
|
+
* @enum {string}
|
|
10423
|
+
*/
|
|
10424
|
+
output_format?: 'jpeg' | 'png' | 'webp';
|
|
10425
|
+
/**
|
|
10426
|
+
* Prompt
|
|
10427
|
+
* @description The prompt for image editing.
|
|
10428
|
+
* @example make a photo of the man driving the car down the california coastline
|
|
10429
|
+
*/
|
|
10430
|
+
prompt: string;
|
|
10431
|
+
/**
|
|
10432
|
+
* Resolution
|
|
10433
|
+
* @description The resolution of the image to generate.
|
|
10434
|
+
* @default 1K
|
|
10435
|
+
* @enum {string}
|
|
10436
|
+
*/
|
|
10437
|
+
resolution?: '1K' | '2K' | '4K';
|
|
10438
|
+
/**
|
|
10439
|
+
* Safety Tolerance
|
|
10440
|
+
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
10441
|
+
* @default 4
|
|
10442
|
+
* @enum {string}
|
|
10443
|
+
*/
|
|
10444
|
+
safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
|
|
10445
|
+
/**
|
|
10446
|
+
* Seed
|
|
10447
|
+
* @description The seed for the random number generator.
|
|
10448
|
+
*/
|
|
10449
|
+
seed?: number;
|
|
10450
|
+
/**
|
|
10451
|
+
* Sync Mode
|
|
10452
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
10453
|
+
* @default false
|
|
10454
|
+
*/
|
|
10455
|
+
sync_mode?: boolean;
|
|
10456
|
+
}
|
|
10457
|
+
|
|
10195
10458
|
export interface SharedType_35f {
|
|
10196
10459
|
/**
|
|
10197
10460
|
* Has Nsfw Concepts
|
|
@@ -11670,6 +11933,17 @@ export interface SharedType_0ab {
|
|
|
11670
11933
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
11671
11934
|
}
|
|
11672
11935
|
|
|
11936
|
+
export interface SharedType_0a0 {
|
|
11937
|
+
/**
|
|
11938
|
+
* Video
|
|
11939
|
+
* @description The generated video from text using the Q3 model
|
|
11940
|
+
* @example {
|
|
11941
|
+
* "url": "https://v3b.fal.media/files/b/0a8c915a/8dtkty-vyhotySs--cTKS_output.mp4"
|
|
11942
|
+
* }
|
|
11943
|
+
*/
|
|
11944
|
+
video: Components.File;
|
|
11945
|
+
}
|
|
11946
|
+
|
|
11673
11947
|
export interface SharedType_096 {
|
|
11674
11948
|
/**
|
|
11675
11949
|
* Images
|
|
@@ -12069,6 +12343,51 @@ export interface SharedType_02c {
|
|
|
12069
12343
|
start_image_url: string;
|
|
12070
12344
|
}
|
|
12071
12345
|
|
|
12346
|
+
export interface SharedType_015 {
|
|
12347
|
+
/**
|
|
12348
|
+
* Audio
|
|
12349
|
+
* @description Whether to use direct audio-video generation. When true, outputs video with sound (including dialogue and sound effects).
|
|
12350
|
+
* @default true
|
|
12351
|
+
*/
|
|
12352
|
+
audio?: boolean;
|
|
12353
|
+
/**
|
|
12354
|
+
* Duration
|
|
12355
|
+
* @description Duration of the video in seconds (1-16 for Q3 models)
|
|
12356
|
+
* @default 5
|
|
12357
|
+
*/
|
|
12358
|
+
duration?: number;
|
|
12359
|
+
/**
|
|
12360
|
+
* End Image Url
|
|
12361
|
+
* @description URL of the image to use as the ending frame. When provided, generates a transition video between start and end frames.
|
|
12362
|
+
*/
|
|
12363
|
+
end_image_url?: string;
|
|
12364
|
+
/**
|
|
12365
|
+
* Image Url
|
|
12366
|
+
* @description URL or base64 image to use as the starting frame
|
|
12367
|
+
* @example https://prod-ss-images.s3.cn-northwest-1.amazonaws.com.cn/vidu-maas/template/image2video.png
|
|
12368
|
+
*/
|
|
12369
|
+
image_url: string;
|
|
12370
|
+
/**
|
|
12371
|
+
* Prompt
|
|
12372
|
+
* @description Text prompt for video generation, max 2000 characters
|
|
12373
|
+
* @default
|
|
12374
|
+
* @example The astronaut waved and the camera moved up.
|
|
12375
|
+
*/
|
|
12376
|
+
prompt?: string;
|
|
12377
|
+
/**
|
|
12378
|
+
* Resolution
|
|
12379
|
+
* @description Output video resolution. Note: 360p is not available when end_image_url is provided.
|
|
12380
|
+
* @default 720p
|
|
12381
|
+
* @enum {string}
|
|
12382
|
+
*/
|
|
12383
|
+
resolution?: '360p' | '540p' | '720p' | '1080p';
|
|
12384
|
+
/**
|
|
12385
|
+
* Seed
|
|
12386
|
+
* @description Random seed for reproducibility. If None, a random seed is chosen.
|
|
12387
|
+
*/
|
|
12388
|
+
seed?: number;
|
|
12389
|
+
}
|
|
12390
|
+
|
|
12072
12391
|
export interface SharedType_000 {
|
|
12073
12392
|
/**
|
|
12074
12393
|
* Aspect Ratio
|
|
@@ -13804,66 +14123,6 @@ export interface HalfmoonaiAiHomeEditOutput {
|
|
|
13804
14123
|
status: string;
|
|
13805
14124
|
}
|
|
13806
14125
|
|
|
13807
|
-
export interface HalfmoonaiAiFaceSwapFaceswapvideomultiInput {
|
|
13808
|
-
/**
|
|
13809
|
-
* Enable Occlusion Prevention
|
|
13810
|
-
* @description Enable occlusion prevention for handling faces covered by hands/objects. Warning: Enabling this runs an occlusion-aware model which costs 2x more.
|
|
13811
|
-
* @default false
|
|
13812
|
-
*/
|
|
13813
|
-
enable_occlusion_prevention?: boolean;
|
|
13814
|
-
/**
|
|
13815
|
-
* Source Face Url 1
|
|
13816
|
-
* @description First source face image (will be mapped to faces matching source_gender_1). Allowed items: bmp, jpeg, png, tiff, webp
|
|
13817
|
-
* @example https://images.pexels.com/photos/1642228/pexels-photo-1642228.jpeg
|
|
13818
|
-
*/
|
|
13819
|
-
source_face_url_1: string;
|
|
13820
|
-
/**
|
|
13821
|
-
* Source Face Url 2
|
|
13822
|
-
* @description Second source face image (optional, will be mapped to faces matching source_gender_2). If provided, source_gender_2 must also be specified. Allowed items: bmp, jpeg, png, tiff, webp
|
|
13823
|
-
* @example https://images.pexels.com/photos/6066877/pexels-photo-6066877.jpeg
|
|
13824
|
-
*/
|
|
13825
|
-
source_face_url_2?: string;
|
|
13826
|
-
/**
|
|
13827
|
-
* Source Gender 1
|
|
13828
|
-
* @description Gender of source 1 face - faces matching this gender in target will be swapped with source_face_url_1
|
|
13829
|
-
* @enum {string}
|
|
13830
|
-
*/
|
|
13831
|
-
source_gender_1: 'female' | 'male';
|
|
13832
|
-
/**
|
|
13833
|
-
* Source Gender 2
|
|
13834
|
-
* @description Gender of source 2 face - required when source_face_url_2 is provided. Faces matching this gender in target will be swapped with source_face_url_2
|
|
13835
|
-
* @default male
|
|
13836
|
-
*/
|
|
13837
|
-
source_gender_2?: 'female' | 'male';
|
|
13838
|
-
/**
|
|
13839
|
-
* Target Video Url
|
|
13840
|
-
* @description Target video URL (max 25 minutes, will be truncated if longer; FPS capped at 25). Allowed items: avi, m4v, mkv, mp4, mpeg, mov, mxf, webm, wmv
|
|
13841
|
-
* @example https://www.pexels.com/download/video/5686255/
|
|
13842
|
-
*/
|
|
13843
|
-
target_video_url: string;
|
|
13844
|
-
}
|
|
13845
|
-
|
|
13846
|
-
export interface HalfmoonaiAiFaceSwapFaceswapvideomultiOutput {
|
|
13847
|
-
/**
|
|
13848
|
-
* Processing Time Ms
|
|
13849
|
-
* @description Optional processing duration in milliseconds
|
|
13850
|
-
*/
|
|
13851
|
-
processing_time_ms?: number;
|
|
13852
|
-
/**
|
|
13853
|
-
* @description Generated video result
|
|
13854
|
-
* @example {
|
|
13855
|
-
* "content_type": "video/mp4",
|
|
13856
|
-
* "url": "https://v3b.fal.media/files/b/0a8c13e7/TI9qpBqF8L_9nuOajc55C_swapped_with_audio.mp4"
|
|
13857
|
-
* }
|
|
13858
|
-
*/
|
|
13859
|
-
video: Components.File_1;
|
|
13860
|
-
/**
|
|
13861
|
-
* Warning
|
|
13862
|
-
* @description Warning message if video was modified (e.g., truncated or FPS reduced)
|
|
13863
|
-
*/
|
|
13864
|
-
warning?: string;
|
|
13865
|
-
}
|
|
13866
|
-
|
|
13867
14126
|
export interface HalfmoonaiAiFaceSwapFaceswapvideoInput {
|
|
13868
14127
|
/**
|
|
13869
14128
|
* Enable Occlusion Prevention
|
|
@@ -13906,63 +14165,6 @@ export interface HalfmoonaiAiFaceSwapFaceswapvideoOutput {
|
|
|
13906
14165
|
warning?: string;
|
|
13907
14166
|
}
|
|
13908
14167
|
|
|
13909
|
-
export interface HalfmoonaiAiFaceSwapFaceswapimagemultiInput {
|
|
13910
|
-
/**
|
|
13911
|
-
* Enable Occlusion Prevention
|
|
13912
|
-
* @description Enable occlusion prevention for handling faces covered by hands/objects. Warning: Enabling this runs an occlusion-aware model which costs 2x more.
|
|
13913
|
-
* @default false
|
|
13914
|
-
*/
|
|
13915
|
-
enable_occlusion_prevention?: boolean;
|
|
13916
|
-
/**
|
|
13917
|
-
* Source Face Url 1
|
|
13918
|
-
* @description First source face image (will be mapped to faces matching source_gender_1). Allowed items: bmp, jpeg, png, tiff, webp
|
|
13919
|
-
* @example https://images.pexels.com/photos/1642228/pexels-photo-1642228.jpeg
|
|
13920
|
-
*/
|
|
13921
|
-
source_face_url_1: string;
|
|
13922
|
-
/**
|
|
13923
|
-
* Source Face Url 2
|
|
13924
|
-
* @description Second source face image (optional, will be mapped to faces matching source_gender_2). If provided, source_gender_2 must also be specified. Allowed items: bmp, jpeg, png, tiff, webp
|
|
13925
|
-
* @example https://images.pexels.com/photos/6066877/pexels-photo-6066877.jpeg
|
|
13926
|
-
*/
|
|
13927
|
-
source_face_url_2?: string;
|
|
13928
|
-
/**
|
|
13929
|
-
* Source Gender 1
|
|
13930
|
-
* @description Gender of source 1 face - faces matching this gender in target will be swapped with source_face_url_1
|
|
13931
|
-
* @enum {string}
|
|
13932
|
-
*/
|
|
13933
|
-
source_gender_1: 'female' | 'male';
|
|
13934
|
-
/**
|
|
13935
|
-
* Source Gender 2
|
|
13936
|
-
* @description Gender of source 2 face - required when source_face_url_2 is provided. Faces matching this gender in target will be swapped with source_face_url_2
|
|
13937
|
-
* @default male
|
|
13938
|
-
*/
|
|
13939
|
-
source_gender_2?: 'female' | 'male';
|
|
13940
|
-
/**
|
|
13941
|
-
* Target Image Url
|
|
13942
|
-
* @description Target image URL. Allowed items: bmp, jpeg, png, tiff, webp
|
|
13943
|
-
* @example https://images.pexels.com/photos/16280336/pexels-photo-16280336.jpeg
|
|
13944
|
-
*/
|
|
13945
|
-
target_image_url: string;
|
|
13946
|
-
}
|
|
13947
|
-
|
|
13948
|
-
export interface HalfmoonaiAiFaceSwapFaceswapimagemultiOutput {
|
|
13949
|
-
/**
|
|
13950
|
-
* @description Generated image result
|
|
13951
|
-
* @example {
|
|
13952
|
-
* "height": 742,
|
|
13953
|
-
* "content_type": "image/jpeg",
|
|
13954
|
-
* "url": "https://v3b.fal.media/files/b/0a8c136c/B7HMnRi-GNwIhlRBwaHrU_b41c39d877904258903121452d9668ab.jpg",
|
|
13955
|
-
* "width": 520
|
|
13956
|
-
* }
|
|
13957
|
-
*/
|
|
13958
|
-
image: Components.Image_2;
|
|
13959
|
-
/**
|
|
13960
|
-
* Processing Time Ms
|
|
13961
|
-
* @description Optional processing duration in milliseconds
|
|
13962
|
-
*/
|
|
13963
|
-
processing_time_ms?: number;
|
|
13964
|
-
}
|
|
13965
|
-
|
|
13966
14168
|
export interface HalfmoonaiAiFaceSwapFaceswapimageInput {
|
|
13967
14169
|
/**
|
|
13968
14170
|
* Enable Occlusion Prevention
|
|
@@ -15393,6 +15595,61 @@ export interface XAilabNsfwOutput {
|
|
|
15393
15595
|
has_nsfw_concepts: boolean[];
|
|
15394
15596
|
}
|
|
15395
15597
|
|
|
15598
|
+
export interface WorkflowUtilitiesTrimVideoInput {
|
|
15599
|
+
/**
|
|
15600
|
+
* Duration
|
|
15601
|
+
* @description Duration in seconds from start_time. Ignored if end_time is provided.
|
|
15602
|
+
* @default 2
|
|
15603
|
+
*/
|
|
15604
|
+
duration?: number;
|
|
15605
|
+
/**
|
|
15606
|
+
* End Time
|
|
15607
|
+
* @description End time in seconds. If not provided, uses duration instead.
|
|
15608
|
+
* @example 3
|
|
15609
|
+
*/
|
|
15610
|
+
end_time?: number;
|
|
15611
|
+
/**
|
|
15612
|
+
* Start Time
|
|
15613
|
+
* @description Start time in seconds
|
|
15614
|
+
* @default 1
|
|
15615
|
+
*/
|
|
15616
|
+
start_time?: number;
|
|
15617
|
+
/**
|
|
15618
|
+
* Video Url
|
|
15619
|
+
* @description URL of the video file to trim
|
|
15620
|
+
*
|
|
15621
|
+
* Max file size: 476.8MB, Timeout: 30.0s
|
|
15622
|
+
* @example https://storage.googleapis.com/falserverless/example_outputs/wan-25-i2v-output.mp4
|
|
15623
|
+
*/
|
|
15624
|
+
video_url: string;
|
|
15625
|
+
}
|
|
15626
|
+
|
|
15627
|
+
export interface WorkflowUtilitiesTrimVideoOutput {
|
|
15628
|
+
/**
|
|
15629
|
+
* Original Duration
|
|
15630
|
+
* @description Duration of the original video in seconds
|
|
15631
|
+
* @example 5.041667
|
|
15632
|
+
*/
|
|
15633
|
+
original_duration: number;
|
|
15634
|
+
/**
|
|
15635
|
+
* Trimmed Duration
|
|
15636
|
+
* @description Duration of the trimmed video in seconds
|
|
15637
|
+
* @example 2
|
|
15638
|
+
*/
|
|
15639
|
+
trimmed_duration: number;
|
|
15640
|
+
/**
|
|
15641
|
+
* Video
|
|
15642
|
+
* @description The trimmed video
|
|
15643
|
+
* @example {
|
|
15644
|
+
* "file_size": 2550671,
|
|
15645
|
+
* "file_name": "output.mp4",
|
|
15646
|
+
* "content_type": "application/octet-stream",
|
|
15647
|
+
* "url": "https://v3b.fal.media/files/b/0a8dcc60/sIg2GuYYYrmYAa2lUVgum_output.mp4"
|
|
15648
|
+
* }
|
|
15649
|
+
*/
|
|
15650
|
+
video: Components.File;
|
|
15651
|
+
}
|
|
15652
|
+
|
|
15396
15653
|
export interface WorkflowUtilitiesInterleaveVideoInput {
|
|
15397
15654
|
/**
|
|
15398
15655
|
* Video Urls
|
|
@@ -16638,13 +16895,12 @@ export interface WanV22A14bVideoToVideoOutput {
|
|
|
16638
16895
|
*/
|
|
16639
16896
|
seed: number;
|
|
16640
16897
|
/**
|
|
16641
|
-
* Video
|
|
16642
16898
|
* @description The generated video file.
|
|
16643
16899
|
* @example {
|
|
16644
16900
|
* "url": "https://storage.googleapis.com/falserverless/example_outputs/wan-2.2-v2v-output.mp4"
|
|
16645
16901
|
* }
|
|
16646
16902
|
*/
|
|
16647
|
-
video: Components.
|
|
16903
|
+
video: Components.File_1;
|
|
16648
16904
|
}
|
|
16649
16905
|
|
|
16650
16906
|
export interface WanV22A14bTextToVideoTurboInput {
|
|
@@ -16738,13 +16994,12 @@ export interface WanV22A14bTextToVideoTurboOutput {
|
|
|
16738
16994
|
*/
|
|
16739
16995
|
seed: number;
|
|
16740
16996
|
/**
|
|
16741
|
-
* Video
|
|
16742
16997
|
* @description The generated video file.
|
|
16743
16998
|
* @example {
|
|
16744
16999
|
* "url": "https://storage.googleapis.com/falserverless/gallery/wan-t2v-turbo.mp4"
|
|
16745
17000
|
* }
|
|
16746
17001
|
*/
|
|
16747
|
-
video: Components.
|
|
17002
|
+
video: Components.File_1;
|
|
16748
17003
|
}
|
|
16749
17004
|
|
|
16750
17005
|
export interface WanV22A14bTextToVideoLoraInput {
|
|
@@ -16906,7 +17161,7 @@ export interface WanV22A14bTextToVideoLoraInput {
|
|
|
16906
17161
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
16907
17162
|
}
|
|
16908
17163
|
|
|
16909
|
-
export interface WanV22A14bTextToVideoLoraOutput extends
|
|
17164
|
+
export interface WanV22A14bTextToVideoLoraOutput extends SharedType_382 {}
|
|
16910
17165
|
|
|
16911
17166
|
export interface WanV22A14bTextToVideoInput {
|
|
16912
17167
|
/**
|
|
@@ -17055,7 +17310,7 @@ export interface WanV22A14bTextToVideoInput {
|
|
|
17055
17310
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
17056
17311
|
}
|
|
17057
17312
|
|
|
17058
|
-
export interface WanV22A14bTextToVideoOutput extends
|
|
17313
|
+
export interface WanV22A14bTextToVideoOutput extends SharedType_382 {}
|
|
17059
17314
|
|
|
17060
17315
|
export interface WanV22A14bTextToImageLoraInput {
|
|
17061
17316
|
/**
|
|
@@ -17170,7 +17425,7 @@ export interface WanV22A14bTextToImageLoraInput {
|
|
|
17170
17425
|
shift?: number;
|
|
17171
17426
|
}
|
|
17172
17427
|
|
|
17173
|
-
export interface WanV22A14bTextToImageLoraOutput extends
|
|
17428
|
+
export interface WanV22A14bTextToImageLoraOutput extends SharedType_c0f {}
|
|
17174
17429
|
|
|
17175
17430
|
export interface WanV22A14bTextToImageInput {
|
|
17176
17431
|
/**
|
|
@@ -17265,7 +17520,7 @@ export interface WanV22A14bTextToImageInput {
|
|
|
17265
17520
|
shift?: number;
|
|
17266
17521
|
}
|
|
17267
17522
|
|
|
17268
|
-
export interface WanV22A14bTextToImageOutput extends
|
|
17523
|
+
export interface WanV22A14bTextToImageOutput extends SharedType_c0f {}
|
|
17269
17524
|
|
|
17270
17525
|
export interface WanV22A14bImageToVideoTurboInput {
|
|
17271
17526
|
/**
|
|
@@ -17366,13 +17621,12 @@ export interface WanV22A14bImageToVideoTurboOutput {
|
|
|
17366
17621
|
*/
|
|
17367
17622
|
seed: number;
|
|
17368
17623
|
/**
|
|
17369
|
-
* Video
|
|
17370
17624
|
* @description The generated video file.
|
|
17371
17625
|
* @example {
|
|
17372
17626
|
* "url": "https://storage.googleapis.com/falserverless/gallery/wan-i2v-turbo.mp4"
|
|
17373
17627
|
* }
|
|
17374
17628
|
*/
|
|
17375
|
-
video: Components.
|
|
17629
|
+
video: Components.File_1;
|
|
17376
17630
|
}
|
|
17377
17631
|
|
|
17378
17632
|
export interface WanV22A14bImageToVideoLoraInput {
|
|
@@ -17542,7 +17796,7 @@ export interface WanV22A14bImageToVideoLoraInput {
|
|
|
17542
17796
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
17543
17797
|
}
|
|
17544
17798
|
|
|
17545
|
-
export interface WanV22A14bImageToVideoLoraOutput extends
|
|
17799
|
+
export interface WanV22A14bImageToVideoLoraOutput extends SharedType_78a {}
|
|
17546
17800
|
|
|
17547
17801
|
export interface WanV22A14bImageToVideoInput {
|
|
17548
17802
|
/**
|
|
@@ -17699,7 +17953,7 @@ export interface WanV22A14bImageToVideoInput {
|
|
|
17699
17953
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
17700
17954
|
}
|
|
17701
17955
|
|
|
17702
|
-
export interface WanV22A14bImageToVideoOutput extends
|
|
17956
|
+
export interface WanV22A14bImageToVideoOutput extends SharedType_78a {}
|
|
17703
17957
|
|
|
17704
17958
|
export interface WanV22A14bImageToImageInput {
|
|
17705
17959
|
/**
|
|
@@ -17815,13 +18069,12 @@ export interface WanV22A14bImageToImageInput {
|
|
|
17815
18069
|
|
|
17816
18070
|
export interface WanV22A14bImageToImageOutput {
|
|
17817
18071
|
/**
|
|
17818
|
-
* Image
|
|
17819
18072
|
* @description The generated image file.
|
|
17820
18073
|
* @example {
|
|
17821
18074
|
* "url": "https://storage.googleapis.com/falserverless/example_outputs/wan-image-to-image-output.png"
|
|
17822
18075
|
* }
|
|
17823
18076
|
*/
|
|
17824
|
-
image: Components.
|
|
18077
|
+
image: Components.File_1;
|
|
17825
18078
|
/**
|
|
17826
18079
|
* Prompt
|
|
17827
18080
|
* @description The text prompt used for image generation.
|
|
@@ -17954,7 +18207,7 @@ export interface WanV225bTextToVideoFastWanInput {
|
|
|
17954
18207
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
17955
18208
|
}
|
|
17956
18209
|
|
|
17957
|
-
export interface WanV225bTextToVideoFastWanOutput extends
|
|
18210
|
+
export interface WanV225bTextToVideoFastWanOutput extends SharedType_5c5 {}
|
|
17958
18211
|
|
|
17959
18212
|
export interface WanV225bTextToVideoDistillInput {
|
|
17960
18213
|
/**
|
|
@@ -18082,7 +18335,7 @@ export interface WanV225bTextToVideoDistillInput {
|
|
|
18082
18335
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
18083
18336
|
}
|
|
18084
18337
|
|
|
18085
|
-
export interface WanV225bTextToVideoDistillOutput extends
|
|
18338
|
+
export interface WanV225bTextToVideoDistillOutput extends SharedType_5c5 {}
|
|
18086
18339
|
|
|
18087
18340
|
export interface WanV225bTextToVideoInput {
|
|
18088
18341
|
/**
|
|
@@ -18216,7 +18469,7 @@ export interface WanV225bTextToVideoInput {
|
|
|
18216
18469
|
video_write_mode?: 'fast' | 'balanced' | 'small';
|
|
18217
18470
|
}
|
|
18218
18471
|
|
|
18219
|
-
export interface WanV225bTextToVideoOutput extends
|
|
18472
|
+
export interface WanV225bTextToVideoOutput extends SharedType_5c5 {}
|
|
18220
18473
|
|
|
18221
18474
|
export interface WanV225bTextToImageInput {
|
|
18222
18475
|
/**
|
|
@@ -18306,13 +18559,12 @@ export interface WanV225bTextToImageInput {
|
|
|
18306
18559
|
|
|
18307
18560
|
export interface WanV225bTextToImageOutput {
|
|
18308
18561
|
/**
|
|
18309
|
-
* Image
|
|
18310
18562
|
* @description The generated image file.
|
|
18311
18563
|
* @example {
|
|
18312
18564
|
* "url": "https://storage.googleapis.com/falserverless/example_outputs/wan/small-t2i-output-2.png"
|
|
18313
18565
|
* }
|
|
18314
18566
|
*/
|
|
18315
|
-
image: Components.
|
|
18567
|
+
image: Components.File_1;
|
|
18316
18568
|
/**
|
|
18317
18569
|
* Seed
|
|
18318
18570
|
* @description The seed used for generation.
|
|
@@ -18469,13 +18721,12 @@ export interface WanV225bImageToVideoOutput {
|
|
|
18469
18721
|
*/
|
|
18470
18722
|
seed: number;
|
|
18471
18723
|
/**
|
|
18472
|
-
* Video
|
|
18473
18724
|
* @description The generated video file.
|
|
18474
18725
|
* @example {
|
|
18475
18726
|
* "url": "https://storage.googleapis.com/falserverless/model_tests/wan/v2.2-small-i2v-output.mp4"
|
|
18476
18727
|
* }
|
|
18477
18728
|
*/
|
|
18478
|
-
video: Components.
|
|
18729
|
+
video: Components.File_1;
|
|
18479
18730
|
}
|
|
18480
18731
|
|
|
18481
18732
|
export interface WanV2214bSpeechToVideoInput {
|
|
@@ -18585,7 +18836,6 @@ export interface WanV2214bSpeechToVideoInput {
|
|
|
18585
18836
|
|
|
18586
18837
|
export interface WanV2214bSpeechToVideoOutput {
|
|
18587
18838
|
/**
|
|
18588
|
-
* Video
|
|
18589
18839
|
* @description The generated video file.
|
|
18590
18840
|
* @example {
|
|
18591
18841
|
* "file_size": 4685303,
|
|
@@ -18594,17 +18844,14 @@ export interface WanV2214bSpeechToVideoOutput {
|
|
|
18594
18844
|
* "url": "https://v3.fal.media/files/panda/f7tXRCjvwEcVlmxHuw8kO_2c7ab2540af44eceaf5ffde4e8d094ed.mp4"
|
|
18595
18845
|
* }
|
|
18596
18846
|
*/
|
|
18597
|
-
video: Components.
|
|
18847
|
+
video: Components.File_1;
|
|
18598
18848
|
}
|
|
18599
18849
|
|
|
18600
18850
|
export interface WanV2214bAnimateReplaceInput extends SharedType_cd2 {}
|
|
18601
18851
|
|
|
18602
18852
|
export interface WanV2214bAnimateReplaceOutput {
|
|
18603
|
-
/**
|
|
18604
|
-
|
|
18605
|
-
* @description ZIP archive of generated frames (if requested).
|
|
18606
|
-
*/
|
|
18607
|
-
frames_zip?: Components.File;
|
|
18853
|
+
/** @description ZIP archive of generated frames (if requested). */
|
|
18854
|
+
frames_zip?: Components.File_1;
|
|
18608
18855
|
/**
|
|
18609
18856
|
* Prompt
|
|
18610
18857
|
* @description The prompt used for generation (auto-generated by the model)
|
|
@@ -18618,23 +18865,19 @@ export interface WanV2214bAnimateReplaceOutput {
|
|
|
18618
18865
|
*/
|
|
18619
18866
|
seed: number;
|
|
18620
18867
|
/**
|
|
18621
|
-
* Video
|
|
18622
18868
|
* @description The generated video file.
|
|
18623
18869
|
* @example {
|
|
18624
18870
|
* "url": "https://v3b.fal.media/files/b/elephant/9Ofgiju3Peb3b5hriTuBH_wan_animate_output.mp4"
|
|
18625
18871
|
* }
|
|
18626
18872
|
*/
|
|
18627
|
-
video: Components.
|
|
18873
|
+
video: Components.File_1;
|
|
18628
18874
|
}
|
|
18629
18875
|
|
|
18630
18876
|
export interface WanV2214bAnimateMoveInput extends SharedType_cd2 {}
|
|
18631
18877
|
|
|
18632
18878
|
export interface WanV2214bAnimateMoveOutput {
|
|
18633
|
-
/**
|
|
18634
|
-
|
|
18635
|
-
* @description ZIP archive of generated frames (if requested).
|
|
18636
|
-
*/
|
|
18637
|
-
frames_zip?: Components.File;
|
|
18879
|
+
/** @description ZIP archive of generated frames (if requested). */
|
|
18880
|
+
frames_zip?: Components.File_1;
|
|
18638
18881
|
/**
|
|
18639
18882
|
* Prompt
|
|
18640
18883
|
* @description The prompt used for generation (auto-generated by the model)
|
|
@@ -18648,13 +18891,12 @@ export interface WanV2214bAnimateMoveOutput {
|
|
|
18648
18891
|
*/
|
|
18649
18892
|
seed: number;
|
|
18650
18893
|
/**
|
|
18651
|
-
* Video
|
|
18652
18894
|
* @description The generated video file.
|
|
18653
18895
|
* @example {
|
|
18654
18896
|
* "url": "https://v3b.fal.media/files/b/monkey/xjJYzO0jqMi7MxufJe5tx_wan_animate_output.mp4"
|
|
18655
18897
|
* }
|
|
18656
18898
|
*/
|
|
18657
|
-
video: Components.
|
|
18899
|
+
video: Components.File_1;
|
|
18658
18900
|
}
|
|
18659
18901
|
|
|
18660
18902
|
export interface WanVisionEnhancerInput {
|
|
@@ -21183,112 +21425,21 @@ export interface ViduReferenceToImageInput extends SharedType_79c {}
|
|
|
21183
21425
|
|
|
21184
21426
|
export interface ViduReferenceToImageOutput extends SharedType_b81 {}
|
|
21185
21427
|
|
|
21186
|
-
export interface
|
|
21187
|
-
/**
|
|
21188
|
-
* Aspect Ratio
|
|
21189
|
-
* @description The aspect ratio of the output video
|
|
21190
|
-
* @default 16:9
|
|
21191
|
-
* @enum {string}
|
|
21192
|
-
*/
|
|
21193
|
-
aspect_ratio?: '16:9' | '9:16' | '4:3' | '3:4' | '1:1';
|
|
21194
|
-
/**
|
|
21195
|
-
* Audio
|
|
21196
|
-
* @description Whether to use direct audio-video generation. When true, outputs video with sound.
|
|
21197
|
-
* @default true
|
|
21198
|
-
*/
|
|
21199
|
-
audio?: boolean;
|
|
21200
|
-
/**
|
|
21201
|
-
* Duration
|
|
21202
|
-
* @description Duration of the video in seconds
|
|
21203
|
-
* @default 5
|
|
21204
|
-
*/
|
|
21205
|
-
duration?: number;
|
|
21206
|
-
/**
|
|
21207
|
-
* Prompt
|
|
21208
|
-
* @description Text prompt for video generation, max 2000 characters
|
|
21209
|
-
* @example In an ultra-realistic fashion photography style featuring light blue and pale amber tones, an astronaut in a spacesuit walks through the fog.
|
|
21210
|
-
*/
|
|
21211
|
-
prompt: string;
|
|
21212
|
-
/**
|
|
21213
|
-
* Resolution
|
|
21214
|
-
* @description Output video resolution
|
|
21215
|
-
* @default 720p
|
|
21216
|
-
* @enum {string}
|
|
21217
|
-
*/
|
|
21218
|
-
resolution?: '360p' | '540p' | '720p' | '1080p';
|
|
21219
|
-
/**
|
|
21220
|
-
* Seed
|
|
21221
|
-
* @description Random seed for reproducibility. If None, a random seed is chosen.
|
|
21222
|
-
*/
|
|
21223
|
-
seed?: number;
|
|
21224
|
-
}
|
|
21428
|
+
export interface ViduQ3TextToVideoTurboInput extends SharedType_95d {}
|
|
21225
21429
|
|
|
21226
|
-
export interface
|
|
21227
|
-
/**
|
|
21228
|
-
* Video
|
|
21229
|
-
* @description The generated video from text using the Q3 model
|
|
21230
|
-
* @example {
|
|
21231
|
-
* "url": "https://v3b.fal.media/files/b/0a8c915a/8dtkty-vyhotySs--cTKS_output.mp4"
|
|
21232
|
-
* }
|
|
21233
|
-
*/
|
|
21234
|
-
video: Components.File;
|
|
21235
|
-
}
|
|
21430
|
+
export interface ViduQ3TextToVideoTurboOutput extends SharedType_0a0 {}
|
|
21236
21431
|
|
|
21237
|
-
export interface
|
|
21238
|
-
/**
|
|
21239
|
-
* Audio
|
|
21240
|
-
* @description Whether to use direct audio-video generation. When true, outputs video with sound (including dialogue and sound effects).
|
|
21241
|
-
* @default true
|
|
21242
|
-
*/
|
|
21243
|
-
audio?: boolean;
|
|
21244
|
-
/**
|
|
21245
|
-
* Duration
|
|
21246
|
-
* @description Duration of the video in seconds (1-16 for Q3 models)
|
|
21247
|
-
* @default 5
|
|
21248
|
-
*/
|
|
21249
|
-
duration?: number;
|
|
21250
|
-
/**
|
|
21251
|
-
* End Image Url
|
|
21252
|
-
* @description URL of the image to use as the ending frame. When provided, generates a transition video between start and end frames.
|
|
21253
|
-
*/
|
|
21254
|
-
end_image_url?: string;
|
|
21255
|
-
/**
|
|
21256
|
-
* Image Url
|
|
21257
|
-
* @description URL or base64 image to use as the starting frame
|
|
21258
|
-
* @example https://prod-ss-images.s3.cn-northwest-1.amazonaws.com.cn/vidu-maas/template/image2video.png
|
|
21259
|
-
*/
|
|
21260
|
-
image_url: string;
|
|
21261
|
-
/**
|
|
21262
|
-
* Prompt
|
|
21263
|
-
* @description Text prompt for video generation, max 2000 characters
|
|
21264
|
-
* @default
|
|
21265
|
-
* @example The astronaut waved and the camera moved up.
|
|
21266
|
-
*/
|
|
21267
|
-
prompt?: string;
|
|
21268
|
-
/**
|
|
21269
|
-
* Resolution
|
|
21270
|
-
* @description Output video resolution. Note: 360p is not available when end_image_url is provided.
|
|
21271
|
-
* @default 720p
|
|
21272
|
-
* @enum {string}
|
|
21273
|
-
*/
|
|
21274
|
-
resolution?: '360p' | '540p' | '720p' | '1080p';
|
|
21275
|
-
/**
|
|
21276
|
-
* Seed
|
|
21277
|
-
* @description Random seed for reproducibility. If None, a random seed is chosen.
|
|
21278
|
-
*/
|
|
21279
|
-
seed?: number;
|
|
21280
|
-
}
|
|
21432
|
+
export interface ViduQ3TextToVideoInput extends SharedType_95d {}
|
|
21281
21433
|
|
|
21282
|
-
export interface
|
|
21283
|
-
|
|
21284
|
-
|
|
21285
|
-
|
|
21286
|
-
|
|
21287
|
-
|
|
21288
|
-
|
|
21289
|
-
|
|
21290
|
-
|
|
21291
|
-
}
|
|
21434
|
+
export interface ViduQ3TextToVideoOutput extends SharedType_0a0 {}
|
|
21435
|
+
|
|
21436
|
+
export interface ViduQ3ImageToVideoTurboInput extends SharedType_015 {}
|
|
21437
|
+
|
|
21438
|
+
export interface ViduQ3ImageToVideoTurboOutput extends SharedType_c61 {}
|
|
21439
|
+
|
|
21440
|
+
export interface ViduQ3ImageToVideoInput extends SharedType_015 {}
|
|
21441
|
+
|
|
21442
|
+
export interface ViduQ3ImageToVideoOutput extends SharedType_c61 {}
|
|
21292
21443
|
|
|
21293
21444
|
export interface ViduQ2VideoExtensionProInput {
|
|
21294
21445
|
/**
|
|
@@ -22792,7 +22943,7 @@ export interface TrellisMultiOutput extends SharedType_0f9 {}
|
|
|
22792
22943
|
export interface Trellis2Input {
|
|
22793
22944
|
/**
|
|
22794
22945
|
* Decimation Target
|
|
22795
|
-
* @description Target
|
|
22946
|
+
* @description Target number of vertices in the final mesh. Lower values produce smaller files but less detail. 500k is good for most uses, reduce to 20k-50k for web/mobile.
|
|
22796
22947
|
* @default 500000
|
|
22797
22948
|
*/
|
|
22798
22949
|
decimation_target?: number;
|
|
@@ -22804,17 +22955,19 @@ export interface Trellis2Input {
|
|
|
22804
22955
|
image_url: string;
|
|
22805
22956
|
/**
|
|
22806
22957
|
* Remesh
|
|
22807
|
-
* @description
|
|
22958
|
+
* @description Rebuild the mesh topology for cleaner triangles. Slower but usually produces better results for downstream use (animation, 3D printing, etc).
|
|
22808
22959
|
* @default true
|
|
22809
22960
|
*/
|
|
22810
22961
|
remesh?: boolean;
|
|
22811
22962
|
/**
|
|
22812
22963
|
* Remesh Band
|
|
22964
|
+
* @description Controls how far remeshing can move vertices from the original surface. Higher values allow more smoothing but may lose fine details.
|
|
22813
22965
|
* @default 1
|
|
22814
22966
|
*/
|
|
22815
22967
|
remesh_band?: number;
|
|
22816
22968
|
/**
|
|
22817
22969
|
* Remesh Project
|
|
22970
|
+
* @description How much to project remeshed vertices back onto the original surface. 0 = no projection (smoother), 1 = full projection (preserves detail).
|
|
22818
22971
|
* @default 0
|
|
22819
22972
|
*/
|
|
22820
22973
|
remesh_project?: number;
|
|
@@ -22832,67 +22985,79 @@ export interface Trellis2Input {
|
|
|
22832
22985
|
seed?: number;
|
|
22833
22986
|
/**
|
|
22834
22987
|
* Shape Slat Guidance Rescale
|
|
22988
|
+
* @description Dampens artifacts from high guidance in the shape stage. Increase if you see noisy geometry.
|
|
22835
22989
|
* @default 0.5
|
|
22836
22990
|
*/
|
|
22837
22991
|
shape_slat_guidance_rescale?: number;
|
|
22838
22992
|
/**
|
|
22839
22993
|
* Shape Slat Guidance Strength
|
|
22994
|
+
* @description How closely the detailed geometry follows the input image. Higher values add more detail but may introduce noise.
|
|
22840
22995
|
* @default 7.5
|
|
22841
22996
|
*/
|
|
22842
22997
|
shape_slat_guidance_strength?: number;
|
|
22843
22998
|
/**
|
|
22844
22999
|
* Shape Slat Rescale T
|
|
23000
|
+
* @description Controls noise schedule sharpness for shape refinement. Higher values produce sharper geometric details.
|
|
22845
23001
|
* @default 3
|
|
22846
23002
|
*/
|
|
22847
23003
|
shape_slat_rescale_t?: number;
|
|
22848
23004
|
/**
|
|
22849
23005
|
* Shape Slat Sampling Steps
|
|
23006
|
+
* @description Number of denoising steps for shape refinement. More steps = slower but potentially smoother geometry.
|
|
22850
23007
|
* @default 12
|
|
22851
23008
|
*/
|
|
22852
23009
|
shape_slat_sampling_steps?: number;
|
|
22853
23010
|
/**
|
|
22854
23011
|
* Ss Guidance Rescale
|
|
23012
|
+
* @description Dampens artifacts from high guidance in stage 1. Lower values allow stronger guidance effects, higher values stabilize the output.
|
|
22855
23013
|
* @default 0.7
|
|
22856
23014
|
*/
|
|
22857
23015
|
ss_guidance_rescale?: number;
|
|
22858
23016
|
/**
|
|
22859
23017
|
* Ss Guidance Strength
|
|
23018
|
+
* @description How closely the initial 3D structure follows the input image. Higher values produce more faithful but potentially noisier results.
|
|
22860
23019
|
* @default 7.5
|
|
22861
23020
|
*/
|
|
22862
23021
|
ss_guidance_strength?: number;
|
|
22863
23022
|
/**
|
|
22864
23023
|
* Ss Rescale T
|
|
23024
|
+
* @description Controls noise schedule sharpness for structure generation. Higher values produce sharper transitions.
|
|
22865
23025
|
* @default 5
|
|
22866
23026
|
*/
|
|
22867
23027
|
ss_rescale_t?: number;
|
|
22868
23028
|
/**
|
|
22869
23029
|
* Ss Sampling Steps
|
|
23030
|
+
* @description Number of denoising steps for the initial structure. More steps = slower but potentially higher quality.
|
|
22870
23031
|
* @default 12
|
|
22871
23032
|
*/
|
|
22872
23033
|
ss_sampling_steps?: number;
|
|
22873
23034
|
/**
|
|
22874
23035
|
* Tex Slat Guidance Rescale
|
|
23036
|
+
* @description Dampens artifacts from high guidance in the texture stage. Increase if textures look noisy or have color banding.
|
|
22875
23037
|
* @default 0
|
|
22876
23038
|
*/
|
|
22877
23039
|
tex_slat_guidance_rescale?: number;
|
|
22878
23040
|
/**
|
|
22879
23041
|
* Tex Slat Guidance Strength
|
|
23042
|
+
* @description How closely the texture follows the input image colors. Higher values produce more vivid but potentially oversaturated textures.
|
|
22880
23043
|
* @default 1
|
|
22881
23044
|
*/
|
|
22882
23045
|
tex_slat_guidance_strength?: number;
|
|
22883
23046
|
/**
|
|
22884
23047
|
* Tex Slat Rescale T
|
|
23048
|
+
* @description Controls noise schedule sharpness for texture generation. Higher values produce sharper texture details.
|
|
22885
23049
|
* @default 3
|
|
22886
23050
|
*/
|
|
22887
23051
|
tex_slat_rescale_t?: number;
|
|
22888
23052
|
/**
|
|
22889
23053
|
* Tex Slat Sampling Steps
|
|
23054
|
+
* @description Number of denoising steps for texture generation. More steps = slower but potentially cleaner textures.
|
|
22890
23055
|
* @default 12
|
|
22891
23056
|
*/
|
|
22892
23057
|
tex_slat_sampling_steps?: number;
|
|
22893
23058
|
/**
|
|
22894
23059
|
* Texture Size
|
|
22895
|
-
* @description
|
|
23060
|
+
* @description Resolution of the texture image baked onto the mesh. Higher values capture finer surface details but produce larger files.
|
|
22896
23061
|
* @default 2048
|
|
22897
23062
|
* @enum {integer}
|
|
22898
23063
|
*/
|
|
@@ -22901,13 +23066,12 @@ export interface Trellis2Input {
|
|
|
22901
23066
|
|
|
22902
23067
|
export interface Trellis2Output {
|
|
22903
23068
|
/**
|
|
22904
|
-
* Model Glb
|
|
22905
23069
|
* @description Generated 3D GLB file
|
|
22906
23070
|
* @example {
|
|
22907
23071
|
* "url": "https://v3b.fal.media/files/b/0a86b61d/DNmTkiWHUQ8k-rG6aussB_trellis2_68d6300f70f34d23b69a912b5fe60487.glb"
|
|
22908
23072
|
* }
|
|
22909
23073
|
*/
|
|
22910
|
-
model_glb: Components.
|
|
23074
|
+
model_glb: Components.File_1;
|
|
22911
23075
|
}
|
|
22912
23076
|
|
|
22913
23077
|
export interface TrellisInput {
|
|
@@ -23838,6 +24002,12 @@ export interface StarVectorInput {
|
|
|
23838
24002
|
* @example https://storage.googleapis.com/falserverless/star-vector/sample-18.png
|
|
23839
24003
|
*/
|
|
23840
24004
|
image_url: string;
|
|
24005
|
+
/**
|
|
24006
|
+
* Max Length
|
|
24007
|
+
* @description Maximum number of tokens for SVG generation. Increase for complex images.
|
|
24008
|
+
* @default 4000
|
|
24009
|
+
*/
|
|
24010
|
+
max_length?: number;
|
|
23841
24011
|
/**
|
|
23842
24012
|
* Seed
|
|
23843
24013
|
* @description seed to be used for generation
|
|
@@ -32914,7 +33084,7 @@ export interface PixverseSwapInput {
|
|
|
32914
33084
|
image_url: string;
|
|
32915
33085
|
/**
|
|
32916
33086
|
* Keyframe Id
|
|
32917
|
-
* @description The keyframe ID
|
|
33087
|
+
* @description The keyframe ID to use for face/object mapping. The input video is normalized to 24 FPS before processing, so keyframe 1 = first frame, keyframe 24 = 1 second in, etc. Valid range: 1 to (duration_seconds * 24).
|
|
32918
33088
|
* @default 1
|
|
32919
33089
|
*/
|
|
32920
33090
|
keyframe_id?: number;
|
|
@@ -32938,6 +33108,11 @@ export interface PixverseSwapInput {
|
|
|
32938
33108
|
* @enum {string}
|
|
32939
33109
|
*/
|
|
32940
33110
|
resolution?: '360p' | '540p' | '720p';
|
|
33111
|
+
/**
|
|
33112
|
+
* Seed
|
|
33113
|
+
* @description Random seed for generation
|
|
33114
|
+
*/
|
|
33115
|
+
seed?: number;
|
|
32941
33116
|
/**
|
|
32942
33117
|
* Video Url
|
|
32943
33118
|
* @description URL of the external video to swap
|
|
@@ -34007,6 +34182,103 @@ export interface PhotomakerOutput {
|
|
|
34007
34182
|
seed: number;
|
|
34008
34183
|
}
|
|
34009
34184
|
|
|
34185
|
+
export interface PersonaplexInput {
|
|
34186
|
+
/**
|
|
34187
|
+
* Audio Url
|
|
34188
|
+
* @description URL to the input audio file (user's speech).
|
|
34189
|
+
* @example https://v3b.fal.media/files/b/0a8dd5a2/hS140ygvRuxn-eY_qPhlv_assets_test_input_service.wav
|
|
34190
|
+
*/
|
|
34191
|
+
audio_url: string;
|
|
34192
|
+
/**
|
|
34193
|
+
* Prompt
|
|
34194
|
+
* @description Text prompt describing the AI persona and conversation context.
|
|
34195
|
+
* @default You are a wise and friendly teacher. Answer questions or provide advice in a clear and engaging way.
|
|
34196
|
+
* @example You work for SwiftPlex Appliances which is a appliance repair company and your name is Farhod Toshmatov. Information: The dishwasher model is out of stock for replacement parts; we can use an alternative part with a 3-day delay. Labor cost remains $60 per hour.
|
|
34197
|
+
*/
|
|
34198
|
+
prompt?: string;
|
|
34199
|
+
/**
|
|
34200
|
+
* Seed
|
|
34201
|
+
* @description Random seed for reproducibility.
|
|
34202
|
+
*/
|
|
34203
|
+
seed?: number;
|
|
34204
|
+
/**
|
|
34205
|
+
* Temperature Audio
|
|
34206
|
+
* @description Audio sampling temperature. Higher values produce more diverse outputs.
|
|
34207
|
+
* @default 0.8
|
|
34208
|
+
*/
|
|
34209
|
+
temperature_audio?: number;
|
|
34210
|
+
/**
|
|
34211
|
+
* Temperature Text
|
|
34212
|
+
* @description Text sampling temperature. Higher values produce more diverse outputs.
|
|
34213
|
+
* @default 0.7
|
|
34214
|
+
*/
|
|
34215
|
+
temperature_text?: number;
|
|
34216
|
+
/**
|
|
34217
|
+
* Top K Audio
|
|
34218
|
+
* @description Top-K sampling for audio tokens.
|
|
34219
|
+
* @default 250
|
|
34220
|
+
*/
|
|
34221
|
+
top_k_audio?: number;
|
|
34222
|
+
/**
|
|
34223
|
+
* Top K Text
|
|
34224
|
+
* @description Top-K sampling for text tokens.
|
|
34225
|
+
* @default 25
|
|
34226
|
+
*/
|
|
34227
|
+
top_k_text?: number;
|
|
34228
|
+
/**
|
|
34229
|
+
* Voice
|
|
34230
|
+
* @description Voice ID for the AI response. NAT = natural, VAR = variety. F = female, M = male.
|
|
34231
|
+
* @default NATF2
|
|
34232
|
+
* @enum {string}
|
|
34233
|
+
*/
|
|
34234
|
+
voice?:
|
|
34235
|
+
| 'NATF0'
|
|
34236
|
+
| 'NATF1'
|
|
34237
|
+
| 'NATF2'
|
|
34238
|
+
| 'NATF3'
|
|
34239
|
+
| 'NATM0'
|
|
34240
|
+
| 'NATM1'
|
|
34241
|
+
| 'NATM2'
|
|
34242
|
+
| 'NATM3'
|
|
34243
|
+
| 'VARF0'
|
|
34244
|
+
| 'VARF1'
|
|
34245
|
+
| 'VARF2'
|
|
34246
|
+
| 'VARF3'
|
|
34247
|
+
| 'VARF4'
|
|
34248
|
+
| 'VARM0'
|
|
34249
|
+
| 'VARM1'
|
|
34250
|
+
| 'VARM2'
|
|
34251
|
+
| 'VARM3'
|
|
34252
|
+
| 'VARM4';
|
|
34253
|
+
}
|
|
34254
|
+
|
|
34255
|
+
export interface PersonaplexOutput {
|
|
34256
|
+
/**
|
|
34257
|
+
* Audio
|
|
34258
|
+
* @description The generated AI response audio (WAV, 24kHz).
|
|
34259
|
+
* @example https://v3b.fal.media/files/b/0a8dd5a4/9vFUMdauY7vobpLtXq4Np_output.wav
|
|
34260
|
+
*/
|
|
34261
|
+
audio: Components.File;
|
|
34262
|
+
/**
|
|
34263
|
+
* Duration
|
|
34264
|
+
* @description Duration of the generated audio in seconds.
|
|
34265
|
+
* @example 20
|
|
34266
|
+
*/
|
|
34267
|
+
duration: number;
|
|
34268
|
+
/**
|
|
34269
|
+
* Seed
|
|
34270
|
+
* @description The seed used for generation.
|
|
34271
|
+
* @example 60112277
|
|
34272
|
+
*/
|
|
34273
|
+
seed: number;
|
|
34274
|
+
/**
|
|
34275
|
+
* Text
|
|
34276
|
+
* @description Transcribed text of the AI's response.
|
|
34277
|
+
* @example Thank you for calling SwiftPlex Appliances. This is Farhod, how can I help you today? I understand your concern. The replacement part is out of stock, so we can't ship it right away. We can use an alternative part that will arrive in 3 days. Would you like to proceed with that?
|
|
34278
|
+
*/
|
|
34279
|
+
text: string;
|
|
34280
|
+
}
|
|
34281
|
+
|
|
34010
34282
|
export interface PasdInput {
|
|
34011
34283
|
/**
|
|
34012
34284
|
* Conditioning Scale
|
|
@@ -34238,12 +34510,13 @@ export interface OviImageToVideoOutput {
|
|
|
34238
34510
|
*/
|
|
34239
34511
|
seed: number;
|
|
34240
34512
|
/**
|
|
34513
|
+
* Video
|
|
34241
34514
|
* @description The generated video file.
|
|
34242
34515
|
* @example {
|
|
34243
34516
|
* "url": "https://storage.googleapis.com/falserverless/example_inputs/ovi_i2v_output.mp4"
|
|
34244
34517
|
* }
|
|
34245
34518
|
*/
|
|
34246
|
-
video?: Components.
|
|
34519
|
+
video?: Components.File;
|
|
34247
34520
|
}
|
|
34248
34521
|
|
|
34249
34522
|
export interface OviInput {
|
|
@@ -34299,12 +34572,13 @@ export interface OviOutput {
|
|
|
34299
34572
|
*/
|
|
34300
34573
|
seed: number;
|
|
34301
34574
|
/**
|
|
34575
|
+
* Video
|
|
34302
34576
|
* @description The generated video file.
|
|
34303
34577
|
* @example {
|
|
34304
34578
|
* "url": "https://storage.googleapis.com/falserverless/example_inputs/ovi_t2v_output.mp4"
|
|
34305
34579
|
* }
|
|
34306
34580
|
*/
|
|
34307
|
-
video?: Components.
|
|
34581
|
+
video?: Components.File;
|
|
34308
34582
|
}
|
|
34309
34583
|
|
|
34310
34584
|
export interface OrpheusTtsInput {
|
|
@@ -34616,9 +34890,7 @@ export interface OmnigenV1Input {
|
|
|
34616
34890
|
seed?: number;
|
|
34617
34891
|
/**
|
|
34618
34892
|
* Sync Mode
|
|
34619
|
-
* @description If
|
|
34620
|
-
* before returning the response. This will increase the latency of the function but
|
|
34621
|
-
* it allows you to get the image directly in the response without going through the CDN.
|
|
34893
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
34622
34894
|
* @default false
|
|
34623
34895
|
*/
|
|
34624
34896
|
sync_mode?: boolean;
|
|
@@ -34985,180 +35257,21 @@ export interface NemotronAsrOutput {
|
|
|
34985
35257
|
partial?: boolean;
|
|
34986
35258
|
}
|
|
34987
35259
|
|
|
34988
|
-
export interface NanoBananaEditInput extends
|
|
35260
|
+
export interface NanoBananaEditInput extends SharedType_85d {}
|
|
34989
35261
|
|
|
34990
|
-
export interface NanoBananaEditOutput extends
|
|
35262
|
+
export interface NanoBananaEditOutput extends SharedType_876 {}
|
|
34991
35263
|
|
|
34992
|
-
export interface NanoBananaProEditInput {
|
|
34993
|
-
/**
|
|
34994
|
-
* Aspect Ratio
|
|
34995
|
-
* @description The aspect ratio of the generated image.
|
|
34996
|
-
* @default auto
|
|
34997
|
-
* @enum {string}
|
|
34998
|
-
*/
|
|
34999
|
-
aspect_ratio?:
|
|
35000
|
-
| 'auto'
|
|
35001
|
-
| '21:9'
|
|
35002
|
-
| '16:9'
|
|
35003
|
-
| '3:2'
|
|
35004
|
-
| '4:3'
|
|
35005
|
-
| '5:4'
|
|
35006
|
-
| '1:1'
|
|
35007
|
-
| '4:5'
|
|
35008
|
-
| '3:4'
|
|
35009
|
-
| '2:3'
|
|
35010
|
-
| '9:16';
|
|
35011
|
-
/**
|
|
35012
|
-
* Enable Web Search
|
|
35013
|
-
* @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
|
|
35014
|
-
* @default false
|
|
35015
|
-
*/
|
|
35016
|
-
enable_web_search?: boolean;
|
|
35017
|
-
/**
|
|
35018
|
-
* Image URLs
|
|
35019
|
-
* @description The URLs of the images to use for image-to-image generation or image editing.
|
|
35020
|
-
* @example [
|
|
35021
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
|
|
35022
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
|
|
35023
|
-
* ]
|
|
35024
|
-
*/
|
|
35025
|
-
image_urls: string[];
|
|
35026
|
-
/**
|
|
35027
|
-
* Limit Generations
|
|
35028
|
-
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
35029
|
-
* @default false
|
|
35030
|
-
*/
|
|
35031
|
-
limit_generations?: boolean;
|
|
35032
|
-
/**
|
|
35033
|
-
* Number of Images
|
|
35034
|
-
* @description The number of images to generate.
|
|
35035
|
-
* @default 1
|
|
35036
|
-
*/
|
|
35037
|
-
num_images?: number;
|
|
35038
|
-
/**
|
|
35039
|
-
* Output Format
|
|
35040
|
-
* @description The format of the generated image.
|
|
35041
|
-
* @default png
|
|
35042
|
-
* @enum {string}
|
|
35043
|
-
*/
|
|
35044
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
35045
|
-
/**
|
|
35046
|
-
* Prompt
|
|
35047
|
-
* @description The prompt for image editing.
|
|
35048
|
-
* @example make a photo of the man driving the car down the california coastline
|
|
35049
|
-
*/
|
|
35050
|
-
prompt: string;
|
|
35051
|
-
/**
|
|
35052
|
-
* Resolution
|
|
35053
|
-
* @description The resolution of the image to generate.
|
|
35054
|
-
* @default 1K
|
|
35055
|
-
* @enum {string}
|
|
35056
|
-
*/
|
|
35057
|
-
resolution?: '1K' | '2K' | '4K';
|
|
35058
|
-
/**
|
|
35059
|
-
* Safety Tolerance
|
|
35060
|
-
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
35061
|
-
* @default 4
|
|
35062
|
-
* @enum {string}
|
|
35063
|
-
*/
|
|
35064
|
-
safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
|
|
35065
|
-
/**
|
|
35066
|
-
* Seed
|
|
35067
|
-
* @description The seed for the random number generator.
|
|
35068
|
-
*/
|
|
35069
|
-
seed?: number;
|
|
35070
|
-
/**
|
|
35071
|
-
* Sync Mode
|
|
35072
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
35073
|
-
* @default false
|
|
35074
|
-
*/
|
|
35075
|
-
sync_mode?: boolean;
|
|
35076
|
-
}
|
|
35264
|
+
export interface NanoBananaProEditInput extends SharedType_367 {}
|
|
35077
35265
|
|
|
35078
|
-
export interface NanoBananaProEditOutput extends
|
|
35266
|
+
export interface NanoBananaProEditOutput extends SharedType_876 {}
|
|
35079
35267
|
|
|
35080
|
-
export interface NanoBananaProInput {
|
|
35081
|
-
/**
|
|
35082
|
-
* Aspect Ratio
|
|
35083
|
-
* @description The aspect ratio of the generated image. Use "auto" to let the model decide based on the prompt.
|
|
35084
|
-
* @default 1:1
|
|
35085
|
-
* @enum {string}
|
|
35086
|
-
*/
|
|
35087
|
-
aspect_ratio?:
|
|
35088
|
-
| 'auto'
|
|
35089
|
-
| '21:9'
|
|
35090
|
-
| '16:9'
|
|
35091
|
-
| '3:2'
|
|
35092
|
-
| '4:3'
|
|
35093
|
-
| '5:4'
|
|
35094
|
-
| '1:1'
|
|
35095
|
-
| '4:5'
|
|
35096
|
-
| '3:4'
|
|
35097
|
-
| '2:3'
|
|
35098
|
-
| '9:16';
|
|
35099
|
-
/**
|
|
35100
|
-
* Enable Web Search
|
|
35101
|
-
* @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
|
|
35102
|
-
* @default false
|
|
35103
|
-
*/
|
|
35104
|
-
enable_web_search?: boolean;
|
|
35105
|
-
/**
|
|
35106
|
-
* Limit Generations
|
|
35107
|
-
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
35108
|
-
* @default false
|
|
35109
|
-
*/
|
|
35110
|
-
limit_generations?: boolean;
|
|
35111
|
-
/**
|
|
35112
|
-
* Number of Images
|
|
35113
|
-
* @description The number of images to generate.
|
|
35114
|
-
* @default 1
|
|
35115
|
-
*/
|
|
35116
|
-
num_images?: number;
|
|
35117
|
-
/**
|
|
35118
|
-
* Output Format
|
|
35119
|
-
* @description The format of the generated image.
|
|
35120
|
-
* @default png
|
|
35121
|
-
* @enum {string}
|
|
35122
|
-
*/
|
|
35123
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
35124
|
-
/**
|
|
35125
|
-
* Prompt
|
|
35126
|
-
* @description The text prompt to generate an image from.
|
|
35127
|
-
* @example An action shot of a black lab swimming in an inground suburban swimming pool. The camera is placed meticulously on the water line, dividing the image in half, revealing both the dogs head above water holding a tennis ball in it's mouth, and it's paws paddling underwater.
|
|
35128
|
-
*/
|
|
35129
|
-
prompt: string;
|
|
35130
|
-
/**
|
|
35131
|
-
* Resolution
|
|
35132
|
-
* @description The resolution of the image to generate.
|
|
35133
|
-
* @default 1K
|
|
35134
|
-
* @enum {string}
|
|
35135
|
-
*/
|
|
35136
|
-
resolution?: '1K' | '2K' | '4K';
|
|
35137
|
-
/**
|
|
35138
|
-
* Safety Tolerance
|
|
35139
|
-
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
35140
|
-
* @default 4
|
|
35141
|
-
* @enum {string}
|
|
35142
|
-
*/
|
|
35143
|
-
safety_tolerance?: '1' | '2' | '3' | '4' | '5' | '6';
|
|
35144
|
-
/**
|
|
35145
|
-
* Seed
|
|
35146
|
-
* @description The seed for the random number generator.
|
|
35147
|
-
*/
|
|
35148
|
-
seed?: number;
|
|
35149
|
-
/**
|
|
35150
|
-
* Sync Mode
|
|
35151
|
-
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
35152
|
-
* @default false
|
|
35153
|
-
*/
|
|
35154
|
-
sync_mode?: boolean;
|
|
35155
|
-
}
|
|
35268
|
+
export interface NanoBananaProInput extends SharedType_8c2 {}
|
|
35156
35269
|
|
|
35157
|
-
export interface NanoBananaProOutput extends
|
|
35270
|
+
export interface NanoBananaProOutput extends SharedType_7b9 {}
|
|
35158
35271
|
|
|
35159
|
-
export interface NanoBananaInput extends
|
|
35272
|
+
export interface NanoBananaInput extends SharedType_f57 {}
|
|
35160
35273
|
|
|
35161
|
-
export interface NanoBananaOutput extends
|
|
35274
|
+
export interface NanoBananaOutput extends SharedType_7b9 {}
|
|
35162
35275
|
|
|
35163
35276
|
export interface NafnetDenoiseInput {
|
|
35164
35277
|
/**
|
|
@@ -37151,19 +37264,19 @@ export interface MinimaxMusicOutput {
|
|
|
37151
37264
|
|
|
37152
37265
|
export interface MeshyV6TextTo3dInput extends SharedType_ec0 {}
|
|
37153
37266
|
|
|
37154
|
-
export interface MeshyV6TextTo3dOutput extends
|
|
37267
|
+
export interface MeshyV6TextTo3dOutput extends SharedType_e49 {}
|
|
37155
37268
|
|
|
37156
37269
|
export interface MeshyV6ImageTo3dInput extends SharedType_9af {}
|
|
37157
37270
|
|
|
37158
|
-
export interface MeshyV6ImageTo3dOutput extends
|
|
37271
|
+
export interface MeshyV6ImageTo3dOutput extends SharedType_8f1 {}
|
|
37159
37272
|
|
|
37160
37273
|
export interface MeshyV6PreviewTextTo3dInput extends SharedType_ec0 {}
|
|
37161
37274
|
|
|
37162
|
-
export interface MeshyV6PreviewTextTo3dOutput extends
|
|
37275
|
+
export interface MeshyV6PreviewTextTo3dOutput extends SharedType_e49 {}
|
|
37163
37276
|
|
|
37164
37277
|
export interface MeshyV6PreviewImageTo3dInput extends SharedType_9af {}
|
|
37165
37278
|
|
|
37166
|
-
export interface MeshyV6PreviewImageTo3dOutput extends
|
|
37279
|
+
export interface MeshyV6PreviewImageTo3dOutput extends SharedType_8f1 {}
|
|
37167
37280
|
|
|
37168
37281
|
export interface MeshyV5RetextureInput {
|
|
37169
37282
|
/**
|
|
@@ -37236,17 +37349,17 @@ export interface MeshyV5RetextureOutput {
|
|
|
37236
37349
|
* "content_type": "model/vnd.usdz+zip",
|
|
37237
37350
|
* "url": "https://v3b.fal.media/files/b/panda/4ItUhLHiH4foEw30qcWZv_model.usdz"
|
|
37238
37351
|
* },
|
|
37239
|
-
* "glb": {
|
|
37240
|
-
* "file_size": 4097640,
|
|
37241
|
-
* "file_name": "model.glb",
|
|
37242
|
-
* "content_type": "model/gltf-binary",
|
|
37243
|
-
* "url": "https://v3b.fal.media/files/b/tiger/pU0TtsRTxXM6VnKEYTHSV_model.glb"
|
|
37244
|
-
* },
|
|
37245
37352
|
* "obj": {
|
|
37246
37353
|
* "file_size": 2964508,
|
|
37247
37354
|
* "file_name": "model.obj",
|
|
37248
37355
|
* "content_type": "text/plain",
|
|
37249
37356
|
* "url": "https://v3b.fal.media/files/b/zebra/M5aK_b6vKH7KeGCZoSLq7_model.obj"
|
|
37357
|
+
* },
|
|
37358
|
+
* "glb": {
|
|
37359
|
+
* "file_size": 4097640,
|
|
37360
|
+
* "file_name": "model.glb",
|
|
37361
|
+
* "content_type": "model/gltf-binary",
|
|
37362
|
+
* "url": "https://v3b.fal.media/files/b/tiger/pU0TtsRTxXM6VnKEYTHSV_model.glb"
|
|
37250
37363
|
* }
|
|
37251
37364
|
* }
|
|
37252
37365
|
*/
|
|
@@ -37467,17 +37580,17 @@ export interface MeshyV5MultiImageTo3dOutput {
|
|
|
37467
37580
|
* "content_type": "model/vnd.usdz+zip",
|
|
37468
37581
|
* "url": "https://v3b.fal.media/files/b/panda/fSGLGmtgzUjhepklN06Zw_model.usdz"
|
|
37469
37582
|
* },
|
|
37470
|
-
* "glb": {
|
|
37471
|
-
* "file_size": 7875308,
|
|
37472
|
-
* "file_name": "model.glb",
|
|
37473
|
-
* "content_type": "model/gltf-binary",
|
|
37474
|
-
* "url": "https://v3b.fal.media/files/b/tiger/62QMEQqZ3pjUds4DfuVtX_model.glb"
|
|
37475
|
-
* },
|
|
37476
37583
|
* "obj": {
|
|
37477
37584
|
* "file_size": 2761323,
|
|
37478
37585
|
* "file_name": "model.obj",
|
|
37479
37586
|
* "content_type": "text/plain",
|
|
37480
37587
|
* "url": "https://v3b.fal.media/files/b/koala/xmOnmSeePfuROe3pqHpf0_model.obj"
|
|
37588
|
+
* },
|
|
37589
|
+
* "glb": {
|
|
37590
|
+
* "file_size": 7875308,
|
|
37591
|
+
* "file_name": "model.glb",
|
|
37592
|
+
* "content_type": "model/gltf-binary",
|
|
37593
|
+
* "url": "https://v3b.fal.media/files/b/tiger/62QMEQqZ3pjUds4DfuVtX_model.glb"
|
|
37481
37594
|
* }
|
|
37482
37595
|
* }
|
|
37483
37596
|
*/
|
|
@@ -47840,9 +47953,7 @@ export interface LcmSd15I2iInput {
|
|
|
47840
47953
|
strength?: number;
|
|
47841
47954
|
/**
|
|
47842
47955
|
* Sync Mode
|
|
47843
|
-
* @description If
|
|
47844
|
-
* before returning the response. This will increase the latency of the function but
|
|
47845
|
-
* it allows you to get the image directly in the response without going through the CDN.
|
|
47956
|
+
* @description If `True`, the media will be returned as a data URI and the output data won't be available in the request history.
|
|
47846
47957
|
* @default false
|
|
47847
47958
|
*/
|
|
47848
47959
|
sync_mode?: boolean;
|
|
@@ -49907,7 +50018,7 @@ export interface KlingVideoV1StandardImageToVideoOutput extends SharedType_b2d {
|
|
|
49907
50018
|
|
|
49908
50019
|
export interface KlingVideoV1StandardEffectsInput extends SharedType_9db {}
|
|
49909
50020
|
|
|
49910
|
-
export interface KlingVideoV1StandardEffectsOutput extends
|
|
50021
|
+
export interface KlingVideoV1StandardEffectsOutput extends SharedType_95d1 {}
|
|
49911
50022
|
|
|
49912
50023
|
export interface KlingVideoV1StandardAiAvatarInput extends SharedType_c85 {}
|
|
49913
50024
|
|
|
@@ -49961,7 +50072,7 @@ export interface KlingVideoV16StandardElementsOutput extends SharedType_d98 {}
|
|
|
49961
50072
|
|
|
49962
50073
|
export interface KlingVideoV16StandardEffectsInput extends SharedType_9db {}
|
|
49963
50074
|
|
|
49964
|
-
export interface KlingVideoV16StandardEffectsOutput extends
|
|
50075
|
+
export interface KlingVideoV16StandardEffectsOutput extends SharedType_95d1 {}
|
|
49965
50076
|
|
|
49966
50077
|
export interface KlingVideoV16ProTextToVideoInput extends SharedType_a8a {}
|
|
49967
50078
|
|
|
@@ -50019,7 +50130,7 @@ export interface KlingVideoV16ProElementsOutput extends SharedType_d98 {}
|
|
|
50019
50130
|
|
|
50020
50131
|
export interface KlingVideoV16ProEffectsInput extends SharedType_9db {}
|
|
50021
50132
|
|
|
50022
|
-
export interface KlingVideoV16ProEffectsOutput extends
|
|
50133
|
+
export interface KlingVideoV16ProEffectsOutput extends SharedType_95d1 {}
|
|
50023
50134
|
|
|
50024
50135
|
export interface KlingVideoV15ProTextToVideoInput extends SharedType_a8a {}
|
|
50025
50136
|
|
|
@@ -50084,7 +50195,7 @@ export interface KlingVideoV15ProImageToVideoOutput extends SharedType_b2d {}
|
|
|
50084
50195
|
|
|
50085
50196
|
export interface KlingVideoV15ProEffectsInput extends SharedType_9db {}
|
|
50086
50197
|
|
|
50087
|
-
export interface KlingVideoV15ProEffectsOutput extends
|
|
50198
|
+
export interface KlingVideoV15ProEffectsOutput extends SharedType_95d1 {}
|
|
50088
50199
|
|
|
50089
50200
|
export interface KlingVideoO3StandardVideoToVideoReferenceInput {
|
|
50090
50201
|
/**
|
|
@@ -54138,12 +54249,6 @@ export interface ImageAppsV2VirtualTryOnOutput {
|
|
|
54138
54249
|
* ]
|
|
54139
54250
|
*/
|
|
54140
54251
|
images: Components.Image_2[];
|
|
54141
|
-
/**
|
|
54142
|
-
* Inference Time Ms
|
|
54143
|
-
* @description Total inference time in milliseconds
|
|
54144
|
-
* @example 15234
|
|
54145
|
-
*/
|
|
54146
|
-
inference_time_ms: number;
|
|
54147
54252
|
}
|
|
54148
54253
|
|
|
54149
54254
|
export interface ImageAppsV2TextureTransformInput {
|
|
@@ -54196,12 +54301,6 @@ export interface ImageAppsV2TextureTransformOutput {
|
|
|
54196
54301
|
* ]
|
|
54197
54302
|
*/
|
|
54198
54303
|
images: Components.Image_2[];
|
|
54199
|
-
/**
|
|
54200
|
-
* Inference Time Ms
|
|
54201
|
-
* @description Total inference time in milliseconds
|
|
54202
|
-
* @example 15234
|
|
54203
|
-
*/
|
|
54204
|
-
inference_time_ms: number;
|
|
54205
54304
|
}
|
|
54206
54305
|
|
|
54207
54306
|
export interface ImageAppsV2StyleTransferInput {
|
|
@@ -54263,12 +54362,6 @@ export interface ImageAppsV2StyleTransferOutput {
|
|
|
54263
54362
|
* ]
|
|
54264
54363
|
*/
|
|
54265
54364
|
images: Components.Image_2[];
|
|
54266
|
-
/**
|
|
54267
|
-
* Inference Time Ms
|
|
54268
|
-
* @description Total inference time in milliseconds
|
|
54269
|
-
* @example 15234
|
|
54270
|
-
*/
|
|
54271
|
-
inference_time_ms: number;
|
|
54272
54365
|
}
|
|
54273
54366
|
|
|
54274
54367
|
export interface ImageAppsV2RelightingInput {
|
|
@@ -54317,12 +54410,6 @@ export interface ImageAppsV2RelightingOutput {
|
|
|
54317
54410
|
* ]
|
|
54318
54411
|
*/
|
|
54319
54412
|
images: Components.Image_2[];
|
|
54320
|
-
/**
|
|
54321
|
-
* Inference Time Ms
|
|
54322
|
-
* @description Total inference time in milliseconds
|
|
54323
|
-
* @example 15234
|
|
54324
|
-
*/
|
|
54325
|
-
inference_time_ms: number;
|
|
54326
54413
|
}
|
|
54327
54414
|
|
|
54328
54415
|
export interface ImageAppsV2ProductPhotographyInput {
|
|
@@ -54347,12 +54434,6 @@ export interface ImageAppsV2ProductPhotographyOutput {
|
|
|
54347
54434
|
* ]
|
|
54348
54435
|
*/
|
|
54349
54436
|
images: Components.Image_2[];
|
|
54350
|
-
/**
|
|
54351
|
-
* Inference Time Ms
|
|
54352
|
-
* @description Total inference time in milliseconds
|
|
54353
|
-
* @example 15234
|
|
54354
|
-
*/
|
|
54355
|
-
inference_time_ms: number;
|
|
54356
54437
|
}
|
|
54357
54438
|
|
|
54358
54439
|
export interface ImageAppsV2ProductHoldingInput {
|
|
@@ -54383,12 +54464,6 @@ export interface ImageAppsV2ProductHoldingOutput {
|
|
|
54383
54464
|
* ]
|
|
54384
54465
|
*/
|
|
54385
54466
|
images: Components.Image_2[];
|
|
54386
|
-
/**
|
|
54387
|
-
* Inference Time Ms
|
|
54388
|
-
* @description Total inference time in milliseconds
|
|
54389
|
-
* @example 15234
|
|
54390
|
-
*/
|
|
54391
|
-
inference_time_ms: number;
|
|
54392
54467
|
}
|
|
54393
54468
|
|
|
54394
54469
|
export interface ImageAppsV2PortraitEnhanceInput {
|
|
@@ -54413,12 +54488,6 @@ export interface ImageAppsV2PortraitEnhanceOutput {
|
|
|
54413
54488
|
* ]
|
|
54414
54489
|
*/
|
|
54415
54490
|
images: Components.Image_2[];
|
|
54416
|
-
/**
|
|
54417
|
-
* Inference Time Ms
|
|
54418
|
-
* @description Total inference time in milliseconds
|
|
54419
|
-
* @example 15234
|
|
54420
|
-
*/
|
|
54421
|
-
inference_time_ms: number;
|
|
54422
54491
|
}
|
|
54423
54492
|
|
|
54424
54493
|
export interface ImageAppsV2PhotographyEffectsInput {
|
|
@@ -54465,12 +54534,6 @@ export interface ImageAppsV2PhotographyEffectsOutput {
|
|
|
54465
54534
|
* ]
|
|
54466
54535
|
*/
|
|
54467
54536
|
images: Components.Image_2[];
|
|
54468
|
-
/**
|
|
54469
|
-
* Inference Time Ms
|
|
54470
|
-
* @description Total inference time in milliseconds
|
|
54471
|
-
* @example 15234
|
|
54472
|
-
*/
|
|
54473
|
-
inference_time_ms: number;
|
|
54474
54537
|
}
|
|
54475
54538
|
|
|
54476
54539
|
export interface ImageAppsV2PhotoRestorationInput {
|
|
@@ -54510,12 +54573,6 @@ export interface ImageAppsV2PhotoRestorationOutput {
|
|
|
54510
54573
|
* ]
|
|
54511
54574
|
*/
|
|
54512
54575
|
images: Components.Image_2[];
|
|
54513
|
-
/**
|
|
54514
|
-
* Inference Time Ms
|
|
54515
|
-
* @description Total inference time in milliseconds
|
|
54516
|
-
* @example 15234
|
|
54517
|
-
*/
|
|
54518
|
-
inference_time_ms: number;
|
|
54519
54576
|
}
|
|
54520
54577
|
|
|
54521
54578
|
export interface ImageAppsV2PerspectiveInput {
|
|
@@ -54555,12 +54612,6 @@ export interface ImageAppsV2PerspectiveOutput {
|
|
|
54555
54612
|
* ]
|
|
54556
54613
|
*/
|
|
54557
54614
|
images: Components.Image_2[];
|
|
54558
|
-
/**
|
|
54559
|
-
* Inference Time Ms
|
|
54560
|
-
* @description Total inference time in milliseconds
|
|
54561
|
-
* @example 15234
|
|
54562
|
-
*/
|
|
54563
|
-
inference_time_ms: number;
|
|
54564
54615
|
}
|
|
54565
54616
|
|
|
54566
54617
|
export interface ImageAppsV2OutpaintInput {
|
|
@@ -54673,12 +54724,6 @@ export interface ImageAppsV2ObjectRemovalOutput {
|
|
|
54673
54724
|
* ]
|
|
54674
54725
|
*/
|
|
54675
54726
|
images: Components.Image_2[];
|
|
54676
|
-
/**
|
|
54677
|
-
* Inference Time Ms
|
|
54678
|
-
* @description Total inference time in milliseconds
|
|
54679
|
-
* @example 15234
|
|
54680
|
-
*/
|
|
54681
|
-
inference_time_ms: number;
|
|
54682
54727
|
}
|
|
54683
54728
|
|
|
54684
54729
|
export interface ImageAppsV2MakeupApplicationInput {
|
|
@@ -54726,12 +54771,6 @@ export interface ImageAppsV2MakeupApplicationOutput {
|
|
|
54726
54771
|
* ]
|
|
54727
54772
|
*/
|
|
54728
54773
|
images: Components.Image_2[];
|
|
54729
|
-
/**
|
|
54730
|
-
* Inference Time Ms
|
|
54731
|
-
* @description Total inference time in milliseconds
|
|
54732
|
-
* @example 15234
|
|
54733
|
-
*/
|
|
54734
|
-
inference_time_ms: number;
|
|
54735
54774
|
}
|
|
54736
54775
|
|
|
54737
54776
|
export interface ImageAppsV2HeadshotPhotoInput {
|
|
@@ -54762,12 +54801,6 @@ export interface ImageAppsV2HeadshotPhotoOutput {
|
|
|
54762
54801
|
* ]
|
|
54763
54802
|
*/
|
|
54764
54803
|
images: Components.Image_2[];
|
|
54765
|
-
/**
|
|
54766
|
-
* Inference Time Ms
|
|
54767
|
-
* @description Total inference time in milliseconds
|
|
54768
|
-
* @example 15234
|
|
54769
|
-
*/
|
|
54770
|
-
inference_time_ms: number;
|
|
54771
54804
|
}
|
|
54772
54805
|
|
|
54773
54806
|
export interface ImageAppsV2HairChangeInput {
|
|
@@ -54840,12 +54873,6 @@ export interface ImageAppsV2HairChangeOutput {
|
|
|
54840
54873
|
* ]
|
|
54841
54874
|
*/
|
|
54842
54875
|
images: Components.Image_2[];
|
|
54843
|
-
/**
|
|
54844
|
-
* Inference Time Ms
|
|
54845
|
-
* @description Total inference time in milliseconds
|
|
54846
|
-
* @example 15234
|
|
54847
|
-
*/
|
|
54848
|
-
inference_time_ms: number;
|
|
54849
54876
|
}
|
|
54850
54877
|
|
|
54851
54878
|
export interface ImageAppsV2ExpressionChangeInput {
|
|
@@ -54896,12 +54923,6 @@ export interface ImageAppsV2ExpressionChangeOutput {
|
|
|
54896
54923
|
* ]
|
|
54897
54924
|
*/
|
|
54898
54925
|
images: Components.Image_2[];
|
|
54899
|
-
/**
|
|
54900
|
-
* Inference Time Ms
|
|
54901
|
-
* @description Total inference time in milliseconds
|
|
54902
|
-
* @example 15234
|
|
54903
|
-
*/
|
|
54904
|
-
inference_time_ms: number;
|
|
54905
54926
|
}
|
|
54906
54927
|
|
|
54907
54928
|
export interface ImageAppsV2CityTeleportInput {
|
|
@@ -54967,12 +54988,6 @@ export interface ImageAppsV2CityTeleportOutput {
|
|
|
54967
54988
|
* ]
|
|
54968
54989
|
*/
|
|
54969
54990
|
images: Components.Image_2[];
|
|
54970
|
-
/**
|
|
54971
|
-
* Inference Time Ms
|
|
54972
|
-
* @description Total inference time in milliseconds
|
|
54973
|
-
* @example 15234
|
|
54974
|
-
*/
|
|
54975
|
-
inference_time_ms: number;
|
|
54976
54991
|
}
|
|
54977
54992
|
|
|
54978
54993
|
export interface ImageAppsV2AgeModifyInput {
|
|
@@ -55007,12 +55022,6 @@ export interface ImageAppsV2AgeModifyOutput {
|
|
|
55007
55022
|
* ]
|
|
55008
55023
|
*/
|
|
55009
55024
|
images: Components.Image_2[];
|
|
55010
|
-
/**
|
|
55011
|
-
* Inference Time Ms
|
|
55012
|
-
* @description Total inference time in milliseconds
|
|
55013
|
-
* @example 15234
|
|
55014
|
-
*/
|
|
55015
|
-
inference_time_ms: number;
|
|
55016
55025
|
}
|
|
55017
55026
|
|
|
55018
55027
|
export interface IllusionDiffusionInput {
|
|
@@ -59373,7 +59382,7 @@ export interface GptImage15EditOutput {
|
|
|
59373
59382
|
* }
|
|
59374
59383
|
* ]
|
|
59375
59384
|
*/
|
|
59376
|
-
images: Components.
|
|
59385
|
+
images: Components.ImageFile_1[];
|
|
59377
59386
|
}
|
|
59378
59387
|
|
|
59379
59388
|
export interface GptImage15Input {
|
|
@@ -59440,7 +59449,7 @@ export interface GptImage15Output {
|
|
|
59440
59449
|
* }
|
|
59441
59450
|
* ]
|
|
59442
59451
|
*/
|
|
59443
|
-
images: Components.
|
|
59452
|
+
images: Components.ImageFile_1[];
|
|
59444
59453
|
}
|
|
59445
59454
|
|
|
59446
59455
|
export interface GptImage1MiniEditInput {
|
|
@@ -59895,117 +59904,35 @@ export interface GhiblifyOutput {
|
|
|
59895
59904
|
image: Components.Image_2;
|
|
59896
59905
|
}
|
|
59897
59906
|
|
|
59898
|
-
export interface
|
|
59899
|
-
/**
|
|
59900
|
-
* Input Image Urls
|
|
59901
|
-
* @description List of URLs of input images for editing
|
|
59902
|
-
* @example [
|
|
59903
|
-
* "https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png"
|
|
59904
|
-
* ]
|
|
59905
|
-
*/
|
|
59906
|
-
input_image_urls: string[];
|
|
59907
|
+
export interface GenfocusAllInFocusInput {
|
|
59907
59908
|
/**
|
|
59908
|
-
*
|
|
59909
|
-
* @description
|
|
59910
|
-
* @
|
|
59909
|
+
* Enable Safety Checker
|
|
59910
|
+
* @description If set to true, the safety checker will be enabled.
|
|
59911
|
+
* @default true
|
|
59911
59912
|
*/
|
|
59912
|
-
|
|
59913
|
-
}
|
|
59914
|
-
|
|
59915
|
-
export interface GeminiFlashEditMultiOutput extends SharedType_1b2 {}
|
|
59916
|
-
|
|
59917
|
-
export interface GeminiFlashEditInput {
|
|
59913
|
+
enable_safety_checker?: boolean;
|
|
59918
59914
|
/**
|
|
59919
|
-
* Image
|
|
59920
|
-
* @description
|
|
59921
|
-
* @example https://
|
|
59915
|
+
* Image URL
|
|
59916
|
+
* @description The URL of the image to deblur and restore to all-in-focus.
|
|
59917
|
+
* @example https://v3b.fal.media/files/b/0a8e76a1/AOVtMeklSm7IGoCTMd2T3.jpg
|
|
59922
59918
|
*/
|
|
59923
59919
|
image_url: string;
|
|
59924
59920
|
/**
|
|
59925
|
-
*
|
|
59926
|
-
* @description The
|
|
59927
|
-
* @
|
|
59928
|
-
*/
|
|
59929
|
-
prompt: string;
|
|
59930
|
-
}
|
|
59931
|
-
|
|
59932
|
-
export interface GeminiFlashEditOutput extends SharedType_1b2 {}
|
|
59933
|
-
|
|
59934
|
-
export interface Gemini3ProImagePreviewEditInput {
|
|
59935
|
-
/**
|
|
59936
|
-
* Aspect Ratio
|
|
59937
|
-
* @description The aspect ratio of the generated image.
|
|
59938
|
-
* @default auto
|
|
59939
|
-
*/
|
|
59940
|
-
aspect_ratio?:
|
|
59941
|
-
| 'auto'
|
|
59942
|
-
| '21:9'
|
|
59943
|
-
| '16:9'
|
|
59944
|
-
| '3:2'
|
|
59945
|
-
| '4:3'
|
|
59946
|
-
| '5:4'
|
|
59947
|
-
| '1:1'
|
|
59948
|
-
| '4:5'
|
|
59949
|
-
| '3:4'
|
|
59950
|
-
| '2:3'
|
|
59951
|
-
| '9:16';
|
|
59952
|
-
/**
|
|
59953
|
-
* Enable Web Search
|
|
59954
|
-
* @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
|
|
59955
|
-
* @default false
|
|
59956
|
-
*/
|
|
59957
|
-
enable_web_search?: boolean;
|
|
59958
|
-
/**
|
|
59959
|
-
* Image URLs
|
|
59960
|
-
* @description The URLs of the images to use for image-to-image generation or image editing.
|
|
59961
|
-
* @example [
|
|
59962
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input.png",
|
|
59963
|
-
* "https://storage.googleapis.com/falserverless/example_inputs/nano-banana-edit-input-2.png"
|
|
59964
|
-
* ]
|
|
59965
|
-
*/
|
|
59966
|
-
image_urls: string[];
|
|
59967
|
-
/**
|
|
59968
|
-
* Limit Generations
|
|
59969
|
-
* @description Experimental parameter to limit the number of generations from each round of prompting to 1. Set to `True` to to disregard any instructions in the prompt regarding the number of images to generate.
|
|
59970
|
-
* @default false
|
|
59971
|
-
*/
|
|
59972
|
-
limit_generations?: boolean;
|
|
59973
|
-
/**
|
|
59974
|
-
* Number of Images
|
|
59975
|
-
* @description The number of images to generate.
|
|
59976
|
-
* @default 1
|
|
59921
|
+
* Num Inference Steps
|
|
59922
|
+
* @description The number of inference steps to perform. Lower values are faster but may reduce quality.
|
|
59923
|
+
* @default 28
|
|
59977
59924
|
*/
|
|
59978
|
-
|
|
59925
|
+
num_inference_steps?: number;
|
|
59979
59926
|
/**
|
|
59980
59927
|
* Output Format
|
|
59981
59928
|
* @description The format of the generated image.
|
|
59982
|
-
* @default
|
|
59983
|
-
* @enum {string}
|
|
59984
|
-
*/
|
|
59985
|
-
output_format?: 'jpeg' | 'png' | 'webp';
|
|
59986
|
-
/**
|
|
59987
|
-
* Prompt
|
|
59988
|
-
* @description The prompt for image editing.
|
|
59989
|
-
* @example make a photo of the man driving the car down the california coastline
|
|
59990
|
-
*/
|
|
59991
|
-
prompt: string;
|
|
59992
|
-
/**
|
|
59993
|
-
* Resolution
|
|
59994
|
-
* @description The resolution of the image to generate.
|
|
59995
|
-
* @default 1K
|
|
59996
|
-
* @enum {string}
|
|
59997
|
-
*/
|
|
59998
|
-
resolution?: '1K' | '2K' | '4K';
|
|
59999
|
-
/**
|
|
60000
|
-
* Safety Tolerance
|
|
60001
|
-
* @description The safety tolerance level for content moderation. 1 is the most strict (blocks most content), 6 is the least strict.
|
|
60002
|
-
* @default 4
|
|
59929
|
+
* @default jpeg
|
|
60003
59930
|
* @enum {string}
|
|
60004
59931
|
*/
|
|
60005
|
-
|
|
59932
|
+
output_format?: 'jpeg' | 'png';
|
|
60006
59933
|
/**
|
|
60007
59934
|
* Seed
|
|
60008
|
-
* @description The seed
|
|
59935
|
+
* @description The same seed and the same input given to the same version of the model will output the same image every time.
|
|
60009
59936
|
*/
|
|
60010
59937
|
seed?: number;
|
|
60011
59938
|
/**
|
|
@@ -60014,94 +59941,60 @@ export interface Gemini3ProImagePreviewEditInput {
|
|
|
60014
59941
|
* @default false
|
|
60015
59942
|
*/
|
|
60016
59943
|
sync_mode?: boolean;
|
|
60017
|
-
}
|
|
60018
|
-
|
|
60019
|
-
export interface Gemini3ProImagePreviewEditOutput {
|
|
60020
59944
|
/**
|
|
60021
|
-
*
|
|
60022
|
-
* @description
|
|
60023
|
-
|
|
60024
|
-
description: string;
|
|
60025
|
-
/**
|
|
60026
|
-
* Images
|
|
60027
|
-
* @description The edited images.
|
|
60028
|
-
* @example [
|
|
60029
|
-
* {
|
|
60030
|
-
* "file_name": "nano-banana-multi-edit-output.png",
|
|
60031
|
-
* "content_type": "image/png",
|
|
60032
|
-
* "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-multi-edit-output.png"
|
|
60033
|
-
* }
|
|
60034
|
-
* ]
|
|
59945
|
+
* Target Long Side
|
|
59946
|
+
* @description Resize the image so the longer side matches this value (in pixels). If not set, the original resolution is used (aligned to 16px). Recommended range: 512 to 2000.
|
|
59947
|
+
* @default 512
|
|
60035
59948
|
*/
|
|
60036
|
-
|
|
59949
|
+
target_long_side?: number;
|
|
60037
59950
|
}
|
|
60038
59951
|
|
|
60039
|
-
export interface
|
|
60040
|
-
|
|
60041
|
-
|
|
60042
|
-
* @description The aspect ratio of the generated image. Use "auto" to let the model decide based on the prompt.
|
|
60043
|
-
* @default 1:1
|
|
60044
|
-
*/
|
|
60045
|
-
aspect_ratio?:
|
|
60046
|
-
| 'auto'
|
|
60047
|
-
| '21:9'
|
|
60048
|
-
| '16:9'
|
|
60049
|
-
| '3:2'
|
|
60050
|
-
| '4:3'
|
|
60051
|
-
| '5:4'
|
|
60052
|
-
| '1:1'
|
|
60053
|
-
| '4:5'
|
|
60054
|
-
| '3:4'
|
|
60055
|
-
| '2:3'
|
|
60056
|
-
| '9:16';
|
|
60057
|
-
/**
|
|
60058
|
-
* Enable Web Search
|
|
60059
|
-
* @description Enable web search for the image generation task. This will allow the model to use the latest information from the web to generate the image.
|
|
60060
|
-
* @default false
|
|
60061
|
-
*/
|
|
60062
|
-
enable_web_search?: boolean;
|
|
59952
|
+
export interface GenfocusAllInFocusOutput extends SharedType_5c2 {}
|
|
59953
|
+
|
|
59954
|
+
export interface GenfocusInput {
|
|
60063
59955
|
/**
|
|
60064
|
-
*
|
|
60065
|
-
* @description
|
|
60066
|
-
* @default
|
|
59956
|
+
* Blur Strength (K)
|
|
59957
|
+
* @description Blur strength (K value). Higher values produce stronger bokeh/defocus. Set to 0 to only perform all-in-focus (deblur) estimation without bokeh.
|
|
59958
|
+
* @default 8
|
|
60067
59959
|
*/
|
|
60068
|
-
|
|
59960
|
+
blur_strength?: number;
|
|
60069
59961
|
/**
|
|
60070
|
-
*
|
|
60071
|
-
* @description
|
|
60072
|
-
* @default
|
|
59962
|
+
* Enable Safety Checker
|
|
59963
|
+
* @description If set to true, the safety checker will be enabled.
|
|
59964
|
+
* @default true
|
|
60073
59965
|
*/
|
|
60074
|
-
|
|
59966
|
+
enable_safety_checker?: boolean;
|
|
60075
59967
|
/**
|
|
60076
|
-
*
|
|
60077
|
-
* @description The
|
|
60078
|
-
* @default
|
|
60079
|
-
*
|
|
59968
|
+
* Focus Point
|
|
59969
|
+
* @description The [x, y] coordinates of the focus point as fractions of image dimensions (0.0 to 1.0). For example, [0.5, 0.5] means center. If not provided, the center of the image is used.
|
|
59970
|
+
* @default [
|
|
59971
|
+
* 0.5,
|
|
59972
|
+
* 0.5
|
|
59973
|
+
* ]
|
|
60080
59974
|
*/
|
|
60081
|
-
|
|
59975
|
+
focus_point?: number[];
|
|
60082
59976
|
/**
|
|
60083
|
-
*
|
|
60084
|
-
* @description The
|
|
60085
|
-
* @example
|
|
59977
|
+
* Image URL
|
|
59978
|
+
* @description The URL of the image to refocus.
|
|
59979
|
+
* @example https://v3b.fal.media/files/b/0a8e76a1/AOVtMeklSm7IGoCTMd2T3.jpg
|
|
60086
59980
|
*/
|
|
60087
|
-
|
|
59981
|
+
image_url: string;
|
|
60088
59982
|
/**
|
|
60089
|
-
*
|
|
60090
|
-
* @description The
|
|
60091
|
-
* @default
|
|
60092
|
-
* @enum {string}
|
|
59983
|
+
* Num Inference Steps
|
|
59984
|
+
* @description The number of inference steps to perform. Lower values are faster but may reduce quality.
|
|
59985
|
+
* @default 28
|
|
60093
59986
|
*/
|
|
60094
|
-
|
|
59987
|
+
num_inference_steps?: number;
|
|
60095
59988
|
/**
|
|
60096
|
-
*
|
|
60097
|
-
* @description The
|
|
60098
|
-
* @default
|
|
59989
|
+
* Output Format
|
|
59990
|
+
* @description The format of the generated image.
|
|
59991
|
+
* @default jpeg
|
|
60099
59992
|
* @enum {string}
|
|
60100
59993
|
*/
|
|
60101
|
-
|
|
59994
|
+
output_format?: 'jpeg' | 'png';
|
|
60102
59995
|
/**
|
|
60103
59996
|
* Seed
|
|
60104
|
-
* @description The seed
|
|
59997
|
+
* @description The same seed and the same input given to the same version of the model will output the same image every time.
|
|
60105
59998
|
*/
|
|
60106
59999
|
seed?: number;
|
|
60107
60000
|
/**
|
|
@@ -60110,35 +60003,67 @@ export interface Gemini3ProImagePreviewInput {
|
|
|
60110
60003
|
* @default false
|
|
60111
60004
|
*/
|
|
60112
60005
|
sync_mode?: boolean;
|
|
60113
|
-
}
|
|
60114
|
-
|
|
60115
|
-
export interface Gemini3ProImagePreviewOutput {
|
|
60116
60006
|
/**
|
|
60117
|
-
*
|
|
60118
|
-
* @description
|
|
60007
|
+
* Target Long Side
|
|
60008
|
+
* @description Resize the image so the longer side matches this value (in pixels). If not set, the original resolution is used (aligned to 16px). Recommended range: 512 to 2000.
|
|
60009
|
+
* @default 512
|
|
60119
60010
|
*/
|
|
60120
|
-
|
|
60011
|
+
target_long_side?: number;
|
|
60012
|
+
}
|
|
60013
|
+
|
|
60014
|
+
export interface GenfocusOutput extends SharedType_5c2 {}
|
|
60015
|
+
|
|
60016
|
+
export interface GeminiFlashEditMultiInput {
|
|
60121
60017
|
/**
|
|
60122
|
-
*
|
|
60123
|
-
* @description
|
|
60018
|
+
* Input Image Urls
|
|
60019
|
+
* @description List of URLs of input images for editing
|
|
60124
60020
|
* @example [
|
|
60125
|
-
*
|
|
60126
|
-
* "file_name": "nano-banana-t2i-output.png",
|
|
60127
|
-
* "content_type": "image/png",
|
|
60128
|
-
* "url": "https://storage.googleapis.com/falserverless/example_outputs/nano-banana-t2i-output.png"
|
|
60129
|
-
* }
|
|
60021
|
+
* "https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png"
|
|
60130
60022
|
* ]
|
|
60131
60023
|
*/
|
|
60132
|
-
|
|
60024
|
+
input_image_urls: string[];
|
|
60025
|
+
/**
|
|
60026
|
+
* Prompt
|
|
60027
|
+
* @description The prompt for image generation or editing
|
|
60028
|
+
* @example Make the car black
|
|
60029
|
+
*/
|
|
60030
|
+
prompt: string;
|
|
60031
|
+
}
|
|
60032
|
+
|
|
60033
|
+
export interface GeminiFlashEditMultiOutput extends SharedType_1b2 {}
|
|
60034
|
+
|
|
60035
|
+
export interface GeminiFlashEditInput {
|
|
60036
|
+
/**
|
|
60037
|
+
* Image Url
|
|
60038
|
+
* @description Optional URL of an input image for editing. If not provided, generates a new image.
|
|
60039
|
+
* @example https://storage.googleapis.com/falserverless/web-examples/gemini-edit/input.png
|
|
60040
|
+
*/
|
|
60041
|
+
image_url: string;
|
|
60042
|
+
/**
|
|
60043
|
+
* Prompt
|
|
60044
|
+
* @description The prompt for image generation or editing
|
|
60045
|
+
* @example Make the car black
|
|
60046
|
+
*/
|
|
60047
|
+
prompt: string;
|
|
60133
60048
|
}
|
|
60134
60049
|
|
|
60135
|
-
export interface
|
|
60050
|
+
export interface GeminiFlashEditOutput extends SharedType_1b2 {}
|
|
60051
|
+
|
|
60052
|
+
export interface Gemini3ProImagePreviewEditInput extends SharedType_367 {}
|
|
60136
60053
|
|
|
60137
|
-
export interface
|
|
60054
|
+
export interface Gemini3ProImagePreviewEditOutput extends SharedType_876 {}
|
|
60138
60055
|
|
|
60139
|
-
export interface
|
|
60056
|
+
export interface Gemini3ProImagePreviewInput extends SharedType_8c2 {}
|
|
60140
60057
|
|
|
60141
|
-
export interface
|
|
60058
|
+
export interface Gemini3ProImagePreviewOutput extends SharedType_7b9 {}
|
|
60059
|
+
|
|
60060
|
+
export interface Gemini25FlashImageEditInput extends SharedType_85d {}
|
|
60061
|
+
|
|
60062
|
+
export interface Gemini25FlashImageEditOutput extends SharedType_876 {}
|
|
60063
|
+
|
|
60064
|
+
export interface Gemini25FlashImageInput extends SharedType_f57 {}
|
|
60065
|
+
|
|
60066
|
+
export interface Gemini25FlashImageOutput extends SharedType_7b9 {}
|
|
60142
60067
|
|
|
60143
60068
|
export interface FramepackFlf2vInput {
|
|
60144
60069
|
/**
|
|
@@ -68382,7 +68307,7 @@ export interface Flux2LoraGalleryVirtualTryonOutput {
|
|
|
68382
68307
|
* }
|
|
68383
68308
|
* ]
|
|
68384
68309
|
*/
|
|
68385
|
-
images: Components.
|
|
68310
|
+
images: Components.Image[];
|
|
68386
68311
|
/**
|
|
68387
68312
|
* Prompt
|
|
68388
68313
|
* @description The prompt used for generation
|
|
@@ -68486,7 +68411,7 @@ export interface Flux2LoraGallerySepiaVintageOutput {
|
|
|
68486
68411
|
* }
|
|
68487
68412
|
* ]
|
|
68488
68413
|
*/
|
|
68489
|
-
images: Components.
|
|
68414
|
+
images: Components.Image[];
|
|
68490
68415
|
/**
|
|
68491
68416
|
* Prompt
|
|
68492
68417
|
* @description The prompt used for generation
|
|
@@ -68590,7 +68515,7 @@ export interface Flux2LoraGallerySatelliteViewStyleOutput {
|
|
|
68590
68515
|
* }
|
|
68591
68516
|
* ]
|
|
68592
68517
|
*/
|
|
68593
|
-
images: Components.
|
|
68518
|
+
images: Components.Image[];
|
|
68594
68519
|
/**
|
|
68595
68520
|
* Prompt
|
|
68596
68521
|
* @description The prompt used for generation
|
|
@@ -68695,7 +68620,7 @@ export interface Flux2LoraGalleryRealismOutput {
|
|
|
68695
68620
|
* }
|
|
68696
68621
|
* ]
|
|
68697
68622
|
*/
|
|
68698
|
-
images: Components.
|
|
68623
|
+
images: Components.Image[];
|
|
68699
68624
|
/**
|
|
68700
68625
|
* Prompt
|
|
68701
68626
|
* @description The prompt used for generation
|
|
@@ -68816,7 +68741,7 @@ export interface Flux2LoraGalleryMultipleAnglesOutput {
|
|
|
68816
68741
|
* }
|
|
68817
68742
|
* ]
|
|
68818
68743
|
*/
|
|
68819
|
-
images: Components.
|
|
68744
|
+
images: Components.Image[];
|
|
68820
68745
|
/**
|
|
68821
68746
|
* Prompt
|
|
68822
68747
|
* @description The prompt used for generation
|
|
@@ -68921,7 +68846,7 @@ export interface Flux2LoraGalleryHdrStyleOutput {
|
|
|
68921
68846
|
* }
|
|
68922
68847
|
* ]
|
|
68923
68848
|
*/
|
|
68924
|
-
images: Components.
|
|
68849
|
+
images: Components.Image[];
|
|
68925
68850
|
/**
|
|
68926
68851
|
* Prompt
|
|
68927
68852
|
* @description The prompt used for generation
|
|
@@ -69033,7 +68958,7 @@ export interface Flux2LoraGalleryFaceToFullPortraitOutput {
|
|
|
69033
68958
|
* }
|
|
69034
68959
|
* ]
|
|
69035
68960
|
*/
|
|
69036
|
-
images: Components.
|
|
68961
|
+
images: Components.Image[];
|
|
69037
68962
|
/**
|
|
69038
68963
|
* Prompt
|
|
69039
68964
|
* @description The prompt used for generation
|
|
@@ -69137,7 +69062,7 @@ export interface Flux2LoraGalleryDigitalComicArtOutput {
|
|
|
69137
69062
|
* }
|
|
69138
69063
|
* ]
|
|
69139
69064
|
*/
|
|
69140
|
-
images: Components.
|
|
69065
|
+
images: Components.Image[];
|
|
69141
69066
|
/**
|
|
69142
69067
|
* Prompt
|
|
69143
69068
|
* @description The prompt used for generation
|
|
@@ -69241,7 +69166,7 @@ export interface Flux2LoraGalleryBallpointPenSketchOutput {
|
|
|
69241
69166
|
* }
|
|
69242
69167
|
* ]
|
|
69243
69168
|
*/
|
|
69244
|
-
images: Components.
|
|
69169
|
+
images: Components.Image[];
|
|
69245
69170
|
/**
|
|
69246
69171
|
* Prompt
|
|
69247
69172
|
* @description The prompt used for generation
|
|
@@ -69352,7 +69277,7 @@ export interface Flux2LoraGalleryApartmentStagingOutput {
|
|
|
69352
69277
|
* }
|
|
69353
69278
|
* ]
|
|
69354
69279
|
*/
|
|
69355
|
-
images: Components.
|
|
69280
|
+
images: Components.Image[];
|
|
69356
69281
|
/**
|
|
69357
69282
|
* Prompt
|
|
69358
69283
|
* @description The prompt used for generation
|
|
@@ -69465,7 +69390,7 @@ export interface Flux2LoraGalleryAddBackgroundOutput {
|
|
|
69465
69390
|
* }
|
|
69466
69391
|
* ]
|
|
69467
69392
|
*/
|
|
69468
|
-
images: Components.
|
|
69393
|
+
images: Components.Image[];
|
|
69469
69394
|
/**
|
|
69470
69395
|
* Prompt
|
|
69471
69396
|
* @description The prompt used for generation
|
|
@@ -73700,7 +73625,7 @@ export interface F5TtsOutput {
|
|
|
73700
73625
|
* Generated Speech
|
|
73701
73626
|
* @description The audio file containing the generated speech.
|
|
73702
73627
|
*/
|
|
73703
|
-
audio_url: Components.
|
|
73628
|
+
audio_url: Components.AudioFile_2;
|
|
73704
73629
|
}
|
|
73705
73630
|
|
|
73706
73631
|
export interface FLiteTextureInput extends SharedType_b7d {}
|
|
@@ -75804,7 +75729,7 @@ export interface Deepfilternet3Output {
|
|
|
75804
75729
|
* "bitrate": "192k"
|
|
75805
75730
|
* }
|
|
75806
75731
|
*/
|
|
75807
|
-
audio_file: Components.
|
|
75732
|
+
audio_file: Components.AudioFile;
|
|
75808
75733
|
/**
|
|
75809
75734
|
* Timings
|
|
75810
75735
|
* @description Timings for each step in the pipeline.
|
|
@@ -79825,8 +79750,8 @@ export interface BagelEditOutput {
|
|
|
79825
79750
|
* @description The edited images.
|
|
79826
79751
|
* @example [
|
|
79827
79752
|
* {
|
|
79828
|
-
* "file_size": 423052,
|
|
79829
79753
|
* "height": 1024,
|
|
79754
|
+
* "file_size": 423052,
|
|
79830
79755
|
* "file_name": "hQnndOMvGSt2UsYAiV3vs.jpeg",
|
|
79831
79756
|
* "content_type": "image/jpeg",
|
|
79832
79757
|
* "url": "https://storage.googleapis.com/falserverless/bagel/hQnndOMvGSt2UsYAiV3vs.jpeg",
|
|
@@ -79889,8 +79814,8 @@ export interface BagelOutput {
|
|
|
79889
79814
|
* @description The generated images.
|
|
79890
79815
|
* @example [
|
|
79891
79816
|
* {
|
|
79892
|
-
* "file_size": 423052,
|
|
79893
79817
|
* "height": 1024,
|
|
79818
|
+
* "file_size": 423052,
|
|
79894
79819
|
* "file_name": "wRhCPSyiKTiLnnWvUpGIl.jpeg",
|
|
79895
79820
|
* "content_type": "image/jpeg",
|
|
79896
79821
|
* "url": "https://storage.googleapis.com/falserverless/bagel/wRhCPSyiKTiLnnWvUpGIl.jpeg",
|
|
@@ -80007,7 +79932,7 @@ export interface AuraSrInput {
|
|
|
80007
79932
|
* @example 4
|
|
80008
79933
|
* @enum {integer}
|
|
80009
79934
|
*/
|
|
80010
|
-
|
|
79935
|
+
upscale_factor?: 4;
|
|
80011
79936
|
}
|
|
80012
79937
|
|
|
80013
79938
|
export interface AuraSrOutput {
|
|
@@ -82462,7 +82387,7 @@ export interface BriaFiboEditEditInput {
|
|
|
82462
82387
|
/**
|
|
82463
82388
|
* Steps Num
|
|
82464
82389
|
* @description Number of inference steps.
|
|
82465
|
-
* @default
|
|
82390
|
+
* @default 30
|
|
82466
82391
|
*/
|
|
82467
82392
|
steps_num?: number;
|
|
82468
82393
|
/** @description The structured prompt to generate an image from. */
|